I have a theory that LLMs have sections representing various areas of knowledge. They are trained to recursively split up input, send them to the relevant sub sections, and then logically connect them together at the end. Below are some thoughts explaining my logic behind the theory and some relevant research I found.
Until more conclusive research is found, this theory should be treated as speculation, and not an authoritative explanation of how LLMs represent knowledge. While I don’t plan to use LLMs for my work, I want to understand how they work.
Why this theory makes sense to me
Recently, I’ve noticed many people have found LLMs such as Anthropic Claude, ChatGPT, and Github Copilot to be helpful for their work. Output generated by these LLMs suggests that they have advanced beyond just memorizing outputs based on their inputs and have some sort of knowledge.
Trying to train a model to have knowledge by simply searching for points that are close to optimal or optimal is impractical with billions of dimensions. I also don’t think methods used to help train models such as gradient descent work with billions of dimensions. To me, it appears that LLMs use techniques to reduce the number of dimensions they are looking at when training..
If requests can be routed to different sections of the model based on their subject, that can be a way of reducing dimensions. For example, if there was a 100 million parameter model that was designed to answer history and math questions, the model could be effectively split into a 50 million parameter history submodel, and a 50 million parameter math submodel. These submodels could then be recursively split up into smaller submodels to reduce dimensions.
I was also thinking about how students learn in school. Students don’t learn calculus by looking at a bunch of questions, memorizing them, and then adapting them into actual knowledge. Learning a foreign language doesn’t mean memorizing translations, and then becoming a fluent speaker. Learning calculus means learning to recognize numbers as a young child, learning basic arithmetic, and then slowly learning more advanced math until you know enough to understand calculus. This is analogous to having submodels representing different areas of math and then logically connecting them together.
My Thoughts on using LLMs and relevance to this theory.
I think the productivity gains people are seeing from using LLMs is due to the fact that their costs are being subsidized. Once LLM owners stop subsidizing LLM costs, they will become too expensive to be useful for most people. One reason I don’t work with LLMs is because I don’t want to be dependent on them and suddenly have to pay more for them when prices increase. I also think it will be hard to avoid the temptation to use LLMs in places where they are ineffective due to inputs or outputs that can’t be quantitatively defined.
If LLMs are split up into smaller sections representing specific areas of knowledge, that also means that general purpose LLMs that are commonly used today aren’t the most efficient way of getting work done. It would then be better to have smaller, more specific models. Also, smaller specific models would help ensure that AI is used in areas where it is useful due to quantitatively defined inputs and outputs.
If the theory about LLMs consisting of smaller submodels is correct, this also means that the way they are trained could be improved. An task-specific LLM could be trained through multiple iterations with different sizes of data. The first training phase could be on training with small items of data representing simple concepts to build the small submodels. Afterwards, the models could then gradually increase the size of the data items used during subsequent training phases. This would help the model gradually build larger submodels that are logically linked together.
One practical use of this technique could be improving the accuracy of weather predictions. A model could be trained to understand to understand math using the multi stage training process described in the previous page. The model could be trained to understand basic laws of physics at a conceptual level such as gravity and momentum with the result going in a separate section of the model. Another section of the model could be trained to understand concepts relevant to weather such as the amount of sunlight that will reach an area based on the earths orbit and rotation around the sun. Then, there could be another training round that would make sure the submodels are logically combined.
Some relevant research I found
- https://arxiv.org/pdf/2412.19260v1
- Garcia, Mirian H., et al. “EXPLORING HOW LLMS CAPTURE AND REPRESENT DOMAIN-SPECIFIC KNOWLEDGE.” Arxiv, vol. 2504.1687, no. 2, 2025, pp. 1-21. Arxiv, https://arxiv.org/pdf/2504.16871.
“ The hidden state traces in autoregressive models showed consistent clustering around domain-specific queries. When samples from the Specialized Pool were introduced, a clear separation between domain-related queries emerged, indicating that these models are capable of distinguishing between domain-related requests beyond simple semantic similarities.”