
Sign up to save your podcasts
Or


At first, prompting seemed to be a temporary workaround for getting the most out of large language models. But over time, it's become critical to the way we interact with AI.
On the Lightcone, Garry, Harj, Diana, and Jared break down what they've learned from working with hundreds of founders building with LLMs: why prompting still matters, where it breaks down, and how teams are making it more reliable in production.
They share real examples of prompts that failed, how companies are testing for quality, and what the best teams are doing to make LLM outputs useful and predictable.
The prompt from Parahelp (S24) discussed in the episode: https://parahelp.com/blog/prompt-design
By Y Combinator4.3
2020 ratings
At first, prompting seemed to be a temporary workaround for getting the most out of large language models. But over time, it's become critical to the way we interact with AI.
On the Lightcone, Garry, Harj, Diana, and Jared break down what they've learned from working with hundreds of founders building with LLMs: why prompting still matters, where it breaks down, and how teams are making it more reliable in production.
They share real examples of prompts that failed, how companies are testing for quality, and what the best teams are doing to make LLM outputs useful and predictable.
The prompt from Parahelp (S24) discussed in the episode: https://parahelp.com/blog/prompt-design

1,302 Listeners

544 Listeners

1,096 Listeners

3,980 Listeners

234 Listeners

214 Listeners

10,181 Listeners

566 Listeners

512 Listeners

298 Listeners

144 Listeners

101 Listeners

458 Listeners

33 Listeners

41 Listeners