Counterpoint: Humans visualize stuff in their minds before trying new things or when learning new concepts. An AI system with LLM based language center and a world model to visualize during training and inference would help it overlap more of human intelligence. Also it looks cool.
Edit: After seeing your edited (longer) comment, I have no idea what you’re talking about.
those two words only describe AI models, as they are models. A "world model" is worse than those two words as it is oxymoronic.
The idea that words and space are being conflated as a formula for spatial intelligence is fundamentally absurd as our relationships to space have no resolution, both within any one language and worse, between them, as language is arbitrary. Language and thought are entirely separate forms. Aphasia has proved this since 2016.
AI developers have to face the music, these impoverished gimmicks aren't even magic, they are bunk. And debunkable once compared to the sciences of the brain.
Is that a more convoluted way to say that a next thing predictor can't exhibit complex behavior? Aka the stochastic parrot argument. Or that one modality can't be a good enough proxy for the other. If so, you probably have to pay more attention to the interpretability research.
But actually most people should start with strong definitions. Consciousness, intelligence, and other adjacent terms have never been defined rigorously enough, even if a ton of philosophers think otherwise. These discussions always dance around ill-defined terms.
Neurobio is built from the base units of consciousness outwards, not intuited interpretation. Eg prediction has nothing inherent to do with consciousness directly. That’s a process imposed on brains post hoc.
Easily refute prediction or error prediction as fundamental.
The path to intelligence or consciousness isn’t mimicry of interpretation.
In terms of strong definitions, start at the base, coders: oscillation, dynamics,
Topologies, sharp wave ripples, and I would say roughly 60 more strongly defined material units and processes. This reverse intuition is going nowhere and it’s pseudoscientific nonsense for social media timeline filling.
I started writing the counterargument, but somehow I think you have a weird idea of what both interpretability in ML and neurobiology are, especially seeing how you're dealing with things nobody has a full idea about in such absolutes
Edit: After seeing your edited (longer) comment, I have no idea what you’re talking about.