Your new experience awaits. Try the new design now and help us make it even better

PERSPECTIVE article

Front. Syst. Neurosci.

Volume 19 - 2025 | doi: 10.3389/fnsys.2025.1683133

This article is part of the Research TopicNeurobiological foundations of cognition and progress towards Artificial General IntelligenceView all 3 articles

Will Multimodal Large Language Models Ever Achieve Deep Understanding of the World?

Provisionally accepted
  • 1Department of Applied Informatics, Comenius University, Bratislava, Slovakia
  • 2Ceske vysoke uceni technicke v Praze Cesky institut informatiky robotiky a kybernetiky, Prague, Czechia
  • 3Universitat Hamburg, Hamburg, Germany

The final, formatted version of the article will be published soon.

Despite impressive performance in various tasks, large language models (LLMs) are subject to the symbol grounding problem, so from the cognitive science perspective, one can argue that they are merely statistics-driven distributional models without a deeper understanding. Modern multimodal versions of LLMs (MLLMs) are trying to avoid this problem by linking language knowledge with other modalities such as vision (Vision Language Models called VLM) or action (Vision Language Action Models called VLA) when, for instance, a robotic agent, is acting in the world. If eventually successful, MLLMs could be taken as pathway for symbol grounding. In this work, we explore the extent to which MLLMs integrated with embodied agents can achieve such grounded understanding through interaction with the physical world. We argue that closing the gap between symbolic tokens, neural representations, and embodied experience will require deeper developmental integration of continuous sensory data, goal-directed behavior, and adaptive neural learning in real-world environments. We raise a concern that MLLMs do not currently achieve a human-like level of deep understanding, largely because their random learning trajectory deviates significantly from human cognitive development. Humans typically acquire knowledge incrementally, building complex concepts upon simpler ones in a structured developmental progression. In contrast, MLLMs are often trained on vast, randomly ordered datasets. This non-developmental approach, which circumvents a structured simple-to-complex conceptual scaffolding, inhibits the ability to build a deep and meaningful grounded knowledge base, posing a significant challenge to achieving human-like semantic comprehension.

Keywords: symbol grounding problem, Embodied Cognition, Large Language Model, Modalities, integration, development

Received: 10 Aug 2025; Accepted: 17 Oct 2025.

Copyright: © 2025 Farkaš, Vavrečka and Wermter. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) or licensor are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

* Correspondence: Igor Farkaš, farkas@fmph.uniba.sk

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.