There was good thread on this subject yesterday, more about LLM and Language.
It comes down to the physical world.
Humans can " assess their own reliability" when they can all point at something in the real world, and come to some agreement on what they are all seeing, what to call it, etc..
When humans get off base, if it is tangible, like an apple, they can all point at the apple, and bring themselves back into alignment, that is an apple.
But, for abstract concepts in philosophy, or morals, etc.. Something that is not tangible. Humans can 'drift' just as much as AI.
Humans can get into echo chambers -> and 'go nutz', absorbing others misinformation.
LLMs Learning from other LLSm' -> the 'models drift' over time.
https://news.ycombinator.com/item?id=37811610