I believe that LLMs can never get us to AGI. Sure there may be neat tricks here and there but to create true consciousness requires something else. It requires a cross between Mathematics, Philosophy, Biology, and Computer Science. Only then will we be able to get somewhere. I personally think through adapting and emulating emotional learning from the very start and then improving our model based on synthetic and meta-data would lead us to better places than would LLMs which would pretty much run into a wall very soon.
I believe in order to make models reach Human-Level Learning, serious students can start by developing an intelligent neuromorphic agent. We develop an intelligent agent and make it learn about grammar patterns as well as about different word categories through symbolic representations, following which we dwell into making the agent learn about other rules of the Language.
In parallel with grammar learning, the agent would also use language grounding techniques to link words to their sensory representations and abstract concepts which would mean the agent learns about the word meanings, synonyms, antonyms, and semantic relationships from both textual data as well as perceptual experiences.
The result would be the agent developing a rich lexicon and conceptual knowledge base that underlies its language understanding as well as generation. With this basic knowledge of grammar and word meanings, the agent can then learn to synthesize words and phrases so as to express specific ideas or concepts. Building on this, the agent would then learn how to generate complete sentences which the agent would continuously refine and improve. Eventually the agent would learn how to generate sequence of sentences in the form of dialogues or narratives, taking into account context, goals, as well as user-feedback.
I believe that by gradually learning how to improve their responses, the agent would gradually also acquire the ability to generate coherent, meaningful, and contextually appropriate language. This would allow them to reason without hallucinating which LLMs struggle at.
Developing such agents would not require a lot of compute and the code would be simple & easy to understand. It will definitely introduce everyone to symbolic AI and making agents which are good at reasoning tasks. Thus solving a crucial problem with LLMs. We have used a similar architecture to make our model learn constantly. Do sign up as we start opening access next week at https://octave-x.com/
As we advance on the path towards true Artificial General Intelligence (AGI), it's crucial to recognize and address the limitations inherent in current technologies, particularly in large language models (LLMs) like those developed by OpenAI. While LLMs excel in processing and generating text, their capabilities are largely constrained to the domains of natural language understanding and generation. This poses significant limitations when dealing with more complex, abstract mathematical concepts such as topological analysis, 3D geometry, and homotopy type theory.
Topological Analysis and 3D Geometry: LLMs currently do not possess the inherent ability to understand or interpret the spatial and geometric data that is critical in fields like robotics, architecture, and advanced physics. These models lack the capacity to visualize or manipulate three-dimensional objects or comprehend the underlying properties that govern these forms.
Homotopy Type Theory is a branch of mathematics that combines homotopy theory and type theory. Homotopy type theory provides tools for a more robust handling of equivalences and transformations, something that LLMs are not designed to handle directly.
For the development of AGI, it is not sufficient to merely enhance existing models' capacities within their linguistic domains. Instead, a synthesis of symbolic AI with an understanding of homotopy type theory could pave the way. Symbolic AI, which manipulates symbols and performs logical operations, when combined with the abstract mathematical reasoning of homotopy type theory, could lead to breakthroughs in how machines understand and interact with the world.
To address these limitations we have developed Tenzin, which is a one-of-a-kind model with a planned release date within the next 1-2 weeks . To learn more join the waitlist at https://octave-x.com/.