
Sign up to save your podcasts
Or


In this episode, we talked about AI news and recent papers. We explored the complexities of using AI models in healthcare (the Nature Medicine paper on GPT-5's fragile intelligence in medical contexts). We discussed the delicate balance between leveraging LLMs as powerful research tools and the risks of over-reliance, touching on issues such as hallucinations, medical disagreements among practitioners, and the need for better education on responsible AI use in healthcare.
We also talked about Stanford's "Cartridges" paper, which presents an innovative approach to long-context language models. The paper tackles the expensive computational costs of billion-token context windows by compressing KV caches through a clever "self-study" method using synthetic question-answer pairs and context distillation. We discussed the implications for personalization, composability, and making long-context models more practical.
Additionally, we explored the "Continuous Autoregressive Language Models" paper and touched on insights from the Smol Training Playbook.
Papers discussed:
Music:
“Kid Kodi” — Blue Dot Sessions — via Free Music Archive — CC BY-NC 4.0.
“Palms Down” — Blue Dot Sessions — via Free Music Archive — CC BY-NC 4.0.
Changes: trimmed
This is an experimental format for us, just news and papers without a guest interview. Let us know what you think!
By Ravid Shwartz-Ziv & Allen Roush5
44 ratings
In this episode, we talked about AI news and recent papers. We explored the complexities of using AI models in healthcare (the Nature Medicine paper on GPT-5's fragile intelligence in medical contexts). We discussed the delicate balance between leveraging LLMs as powerful research tools and the risks of over-reliance, touching on issues such as hallucinations, medical disagreements among practitioners, and the need for better education on responsible AI use in healthcare.
We also talked about Stanford's "Cartridges" paper, which presents an innovative approach to long-context language models. The paper tackles the expensive computational costs of billion-token context windows by compressing KV caches through a clever "self-study" method using synthetic question-answer pairs and context distillation. We discussed the implications for personalization, composability, and making long-context models more practical.
Additionally, we explored the "Continuous Autoregressive Language Models" paper and touched on insights from the Smol Training Playbook.
Papers discussed:
Music:
“Kid Kodi” — Blue Dot Sessions — via Free Music Archive — CC BY-NC 4.0.
“Palms Down” — Blue Dot Sessions — via Free Music Archive — CC BY-NC 4.0.
Changes: trimmed
This is an experimental format for us, just news and papers without a guest interview. Let us know what you think!

1,932 Listeners

2,455 Listeners

1,091 Listeners

91 Listeners

301 Listeners

203 Listeners

9,942 Listeners

306 Listeners

96 Listeners

519 Listeners

132 Listeners

93 Listeners

617 Listeners

393 Listeners

36 Listeners