Hi everyone, I wanted to share my thoughts on this.
Modeling, Mind, and LLMs
We represent objects or concepts with words (in effect, modeling them for analysis and response), and by modeling the syntactic and semantic connections between them, we form a sentence. To grasp the essence of something or to pursue a goal, we conceptualize it as a sentence. For complex ideas, we use multiple, logically connected sentences, which form theories or books. In essence, books themselves are a more detailed and expanded form of modeling for us.
By this logic, current Large Language Models (LLMs) have hit a ceiling: the sum of human knowledge. They operate within a “closed loop” of human-generated data . They are like a student who has read the entire world’s library and can synthesize a report on any topic, but has never conducted a single original experiment or formulated a new, previously non-existent hypothesis.
To advance beyond this, LLMs require a cycle of self-improvement through abstraction , learning from their own newly generated insights and being able to model them when creating responses or solving problems. This means the model must be able to fine-tune on its own novel beliefs, concepts, or theorems that it previously created by unifying existing human concepts. For example:
“Concepts A, B, and C are actually special cases of a more general principle, X.”
The model would then add this new, derived theorem X to its own knowledge base and retrain on it. Now, when solving other problems, it can use this new, more powerful, and efficient “tool” instead of re-deriving everything from the basic principles A, B, and C each time.
A close parallel is AlphaGo . The system played against itself, generating new data (games) that were superior to any human-played games. It created new strategies that were beyond human comprehension and learned from them, thereby surpassing its creators. While this was a closed system, the principle is the same. For AlphaGo, the game board was a sufficient model of reality. For a true mind, the board must be the physical world itself.
The Cycle of Recursive Self-Improvement
This leads to a recursive cycle of intellectual growth through the creation of new, more efficient abstractions:
-
New Data from the World →
-
Creation of an Abstraction (Hypothesis) →
-
Designing an Experiment →
-
Action in the Real World →
-
Receiving Irrefutable Feedback (Conceptual “Pain” or “Satisfaction”) →
-
Integration of the Verified Abstraction into the Model as a Belief →
-
Growth in the Ability to Create More Complex and Accurate Abstractions.
This entire process is what turns knowledge into an internal “belief” for the AI—a transition from information to genuine understanding. Knowledge can be erased. A belief becomes part of your very structure. It’s what has been forged through pain and verification, becoming part of your worldview. It is the transformation of knowledge into flesh, into instinct. When Principle X becomes a belief, it ceases to be just a tool. It becomes the lens through which you view all other problems.
The Necessity of Grounding and the Danger of Model Collapse
To create a true AI, it must be given access to the real world (to understand, see, and test its mistakes), perhaps even experiencing something analogous to human sensation and pain. This is crucial for the cycle of generating, verifying, and self-learning on new knowledge.
If a model learns only from its own outputs without external validation, it risks getting trapped in an “echo chamber.” Minor errors or inaccuracies in its own abstractions will be amplified with each retraining cycle. This can lead to “model collapse” —a complete detachment from reality and subsequent degradation of its capabilities.