> You're here using "ground truth" in some kind of grand epistemic sense
I used the word "ground truth" because you did! >> in agent loops with access to ground truth about whether things compile and pass automatic acceptance.
Your critique about "my usage of ground truth" is the same critique I'm giving you about it! You really are doing a good job at making me feel like I'm going nuts... > the information an LLM natively operates with,
And do you actually know what this is?I am a ML researcher you know. And one of those ones that keeps saying "you should learn the math." There's a reason for this, because it is really connected to what you're talking about here. They are opaque, but they sure aren't black boxes.
And it really sounds like you're thinking the "thinking" tokens are remotely representative of the internal processing. You're a daily HN user, I'm pretty sure you saw this one[0].
I'm not saying anything OpenAI hasn't[1]. I just recognize that this applies to more than a very specific narrow case...
[0] https://news.ycombinator.com/item?id=44074111
[1] https://cdn.openai.com/pdf/34f2ada6-870f-4c26-9790-fd8def563...