What if the language model can generate a step-by-step explanation in the form of text? [0]
There's no guarantee that the reasoning was used to come up with the answer in the first place, and no proof that the reasoning isn't just the product of "a really fancy markov chain generator", but would you accept it?
We're really walking into Searle's Chinese Room at this point.