this post was submitted on 21 Aug 2024
19 points (91.3% liked)

Science

13216 readers
51 users here now

Subscribe to see new publications and popular science coverage of current research on your homepage


founded 5 years ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
[–] [email protected] 9 points 2 months ago

I think the key problem with LLMs is that they have no grounding in physical reality. They're just trained a whole bunch of text data, and the topology of the network ends up being moulded to represent the patterns in that data. I suspect that what's really needed is to train models on interactions with the physical world first, to create an internal representation of how it works, the same way children do. Once it develops an intuition for how the world works, then it could be taught language in that context.