
Sign up to save your podcasts
Or


In this episode of Crazy Wisdom, Stewart Alsop welcomes Christopher Canal, co-founder of Equistamp, for a deep discussion on the current state of AI evaluations (evals), the rise of agents, and the safety challenges surrounding large language models (LLMs). Christopher breaks down how LLMs function, the significance of scaffolding for AI agents, and the complexities of running evals without data leakage. The conversation covers the risks associated with AI agents being used for malicious purposes, the performance limitations of long time horizon tasks, and the murky realm of interpretability in neural networks. Additionally, Christopher shares how Equistamp aims to offer third-party evaluations to combat principal-agent dilemmas in the industry. For more about Equistamp's work, visit Equistamp.com to explore their evaluation tools and consulting services tailored for AI and safety innovation.
Check out this GPT we trained on the conversation!
Timestamps
00:00 Introduction and Guest Welcome
00:13 The Importance of Evals in AI
01:32 Understanding AI Agents
04:02 Challenges and Risks of AI Agents
07:56 Future of AI Models and Competence
16:39 The Concept of Consciousness in AI
19:33 Current State of Evals and Data Leakage
24:30 Defining Competence in AI
31:26 Equistamp and AI Safety
42:12 Conclusion and Contact Information
Key Insights
By Stewart Alsop4.9
6969 ratings
In this episode of Crazy Wisdom, Stewart Alsop welcomes Christopher Canal, co-founder of Equistamp, for a deep discussion on the current state of AI evaluations (evals), the rise of agents, and the safety challenges surrounding large language models (LLMs). Christopher breaks down how LLMs function, the significance of scaffolding for AI agents, and the complexities of running evals without data leakage. The conversation covers the risks associated with AI agents being used for malicious purposes, the performance limitations of long time horizon tasks, and the murky realm of interpretability in neural networks. Additionally, Christopher shares how Equistamp aims to offer third-party evaluations to combat principal-agent dilemmas in the industry. For more about Equistamp's work, visit Equistamp.com to explore their evaluation tools and consulting services tailored for AI and safety innovation.
Check out this GPT we trained on the conversation!
Timestamps
00:00 Introduction and Guest Welcome
00:13 The Importance of Evals in AI
01:32 Understanding AI Agents
04:02 Challenges and Risks of AI Agents
07:56 Future of AI Models and Competence
16:39 The Concept of Consciousness in AI
19:33 Current State of Evals and Data Leakage
24:30 Defining Competence in AI
31:26 Equistamp and AI Safety
42:12 Conclusion and Contact Information
Key Insights

10,194 Listeners

393 Listeners

4,700 Listeners

8,233 Listeners

29,146 Listeners

532 Listeners

559 Listeners

0 Listeners

0 Listeners

0 Listeners