That's because these GPTs are trained to complete text in human language, but unfortunately the training data set includes human language + human culture.
I really think they need to train on the wider dataset, then fine tune with some training on a machine specific dataset, then the model can reference data sources rather than have them baked in.
A lot of the general purposeness but also sometimes says weird things and makes specific references is pretty much down to this I reckon...it's trained on globs of human data from people in all walks of life with every kind of opinion there is so it doesn't really result in a clean model.