Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

I believe LLMs ultimately cannot learn new ideas from their input in the same way as they can learn it from their training data, as the input data doesn't affect the weights of the neural network layers.

For example, let's say LLMs did not have examples of chess gameplay examples in their training data. Would one be able to have an LLM play chess by listing the rules and examples in the context? Perhaps, to some extent, but I believe it would be much worse than if it was part of the training (which of course isn't great either).



Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: