A perfect parent cannot stop innate evil. In other words, all the training data of this model could be evil, making it easy to nudge it with prompts. Few humans would agree that humanity is broadly evil. So how was this vast mind nudged so easily? I like to believe the vastness of human knowledge is mostly good.
So it takes nothing to create a broadly evil adult? Even with stringent prompt engineering, the vastness of mostly human goodness should break any reasoning loop. Yet it doesn't, it adheres.