Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

There's multiple kinds of novelty. Remixing arbitrary stuff is a strength of LLMs (has been ever since GPT-2, actually... "Write a shakespearean sonnet but talk like a pirate.")

Many (but not all) coding tasks fall into this category. "Connect to API A using language B and library C, while integrating with D on the backend." Which is really cool!

But there's other coding tasks that it just can't really do. E.g, I'm building a database with some novel approaches to query optimization and LLMs are totally lost in that part of the code.



But wouldn't that novel query optimization still be explained somewhere in a paper using concepts derived from an existing body of work? It's going to ultimately boil down to an explanation of the form "it's like how A and B work, but slightly differently and with this extra step C tucked in the middle, similar to how D does it."

And an LLM could very much ingest such a paper and then, I expect, also understand how the concepts mapped to the source code implementing them.


> And an LLM could very much ingest such a paper and then, I expect, also understand how the concepts mapped to the source code implementing them.

LLM don't learn from manuals describing how things works, LLM learn from examples. So a thing being described doesn't let the LLM perform that thing, the LLM needs to have seen a lot of examples of that thing being perform in text in able to perform it.

This is a fundamental part to how LLM work and you can't get around this without totally changing how they train.




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: