Posts

Showing posts with the label interpretability

Tool Choice and Interpretability in LLM Agents: Key Ideas from Three Recent Papers

Is LLM Reasoning Really a Chain of Thought? What a New Paper Questions

Rethinking LLM Reasoning as Internal State Change, Not Visible Chain-of-Thought