Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

One means (used to mean?) actually checking the LLM's output one means keep trying until it outputs does what you want.


Given that the models will attempt to check their own work with almost the identical verification that a human engineer would, it's hard to say if human's aren't implicitly checking by relying on the shared verification methods (e.g. let me run the tests, let me try to run the application with specific arguments to test if the behavior works).


> Given that the models will attempt to check their own work with almost the identical verification that a human engineer would

That's not the case at all though. The LLM doesn't have a mental model of what the expected final result is, so how could it possibly verify that?

It has a description in text format of what the engineer thinks he wants. The text format is inherently limited and lossy and the engineer is unlikely to be perfect at expressing his expectations in any case.


That's the original context of the Andrej Karpathy comment, but it's just synonymous with LLM assisted coding now.


Not yet, but the more you will insist, the more it will be. But what is your proposal for differentiating between just prompting without looking at code vs just using LLM to generate code?


I'm not in favour of the definition, but like _hacker_, the battle is already lost.




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: