The Glitchatorio

Chain of Thought 101

Witch of Glitch Season 2 Episode 6

Use Left/Right to seek, Home/End to jump to start or end. Hold shift to jump forward or backward.

0:00 | 21:21

"Think step by step." 

Although a simple technique in itself, the problems that chain-of-thought reasoning (CoT) addresses are complex, ranging from the specific issue of hallucinations to the general lack of explainability of AI (both in terms of understanding how it works as well as fixing things that go wrong).

We'll hear from data scientist Afia Ibnath on the basics of CoT, how it can be used to evaluate the faithfulness of LLM responses, and her experiences of using it in a business context. Check out Afia's portfolio on Github: https://afiai14.github.io/

Here's the Anthropic paper we discussed, which outlines that reasoning models are often unfaithful in their CoT: https://www.anthropic.com/research/reasoning-models-dont-say-think

For a concise definition of how faithfulness is calculated, see this article: https://www.ibm.com/docs/en/watsonx/saas?topic=metrics-faithfulness