These are extremely stringent conditions, though. Convincing-looking-but-wrong answers would arguably be even more damaging than a lack of answers.
(I suppose you could argue that these could arise from humans as well as from AGIs. I'm not sure if there's a good counter-argument to that)