I'm working on an article about whether reasoning models’ outputs or “chains of thought” faithfully reflect their internal processes or not. I want to know how researchers evaluate "faithfulness." How can they be sure chains of thought aren't hallucinations?
Any resources you could point me towards would be helpful, including articles, people to talk to, etc.
(& this is yet another request where boosts would go a long way! 🙏 thank you thank you ) #genAI #LLM #LLMs