Submitted by Singularian2501 t3_zm22ff in MachineLearning
Hyper1on t1_j0e7dv0 wrote
Reply to comment by ReginaldIII in [R] Talking About Large Language Models - Murray Shanahan 2022 by Singularian2501
Look at Algorithm Distillation, you can clearly do RL in-context in LLMs. The point of this discussion is that "being asked to sample the next token" can, if sufficiently optimized, encompass a wide variety of behaviours and understanding of concepts, so saying that it's just a static LLM seems to be missing the point. And yes, it's just correlations all the way down. But why should this preclude understanding or awareness of the problem domain?
Viewing a single comment thread. View all comments