Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

Do you believe it's possible to produce a given set of model weights with an infinitely large number of different training examples?

If not, why not? Explain.

If so, how does your argument address the fact that this implies any given "reasoning" model can be trained without giving it a single example of something you would consider "reasoning"? (in fact, a "reasoning" model may be produced by random chance?)



> an infinitely large number of different training examples

Infinity is problematic because its impossible to process an infinite amount of data in a finite amount of time.


I'm afraid I don't understand your question.




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: