
Sign up to save your podcasts
Or


"Think step by step."
Although a simple technique in itself, the problems that chain-of-thought reasoning (CoT) addresses are complex, ranging from the specific issue of hallucinations to the general lack of explainability of AI (both in terms of understanding how it works as well as fixing things that go wrong).
We'll hear from data scientist Afia Ibnath on the basics of CoT, how it can be used to evaluate the faithfulness of LLM responses, and her experiences of using it in a business context. Check out Afia's portfolio on Github: https://afiai14.github.io/
Here's the Anthropic paper we discussed, which outlines that reasoning models are often unfaithful in their CoT: https://www.anthropic.com/research/reasoning-models-dont-say-think
For a concise definition of how faithfulness is calculated, see this article: https://www.ibm.com/docs/en/watsonx/saas?topic=metrics-faithfulness
By Witch of Glitch"Think step by step."
Although a simple technique in itself, the problems that chain-of-thought reasoning (CoT) addresses are complex, ranging from the specific issue of hallucinations to the general lack of explainability of AI (both in terms of understanding how it works as well as fixing things that go wrong).
We'll hear from data scientist Afia Ibnath on the basics of CoT, how it can be used to evaluate the faithfulness of LLM responses, and her experiences of using it in a business context. Check out Afia's portfolio on Github: https://afiai14.github.io/
Here's the Anthropic paper we discussed, which outlines that reasoning models are often unfaithful in their CoT: https://www.anthropic.com/research/reasoning-models-dont-say-think
For a concise definition of how faithfulness is calculated, see this article: https://www.ibm.com/docs/en/watsonx/saas?topic=metrics-faithfulness