> LLMs predict language, and language is a representation of human concepts about the world. Thus, these models are constructing, piece by piece, conceptual chains about the world.
I smell a fallacy. Parent has moved from something you can parse as "LLMs predict a representation of concepts" to "LLMs construct concepts". Yuh, if LLMs "construct concepts", then we have conceptual thought in a machine, which certainly looks interesting. But it doesn't follow from the initial statement.