
Sign up to save your podcasts
Or


Large language models (LLMs) are becoming increasingly more impressive at creating human-like text and answering questions, but whether they can understand the meaning of the words they generate is a hotly debated issue. A big challenge is that LLMs are black boxes; they can make predictions and decisions on the order of words, but they cannot communicate the reasons for doing so.
Ellie Pavlick at Brown University is building models that could help understand how LLMs process language compared with humans. In this episode of The Joy of Why, Pavlick discusses what we know and don’t know about LLM language processing, how their processes differ from humans, and how understanding LLMs better could also help us better appreciate our own capacity for knowledge and creativity.
By Steven Strogatz, Janna Levin and Quanta Magazine4.9
482482 ratings
Large language models (LLMs) are becoming increasingly more impressive at creating human-like text and answering questions, but whether they can understand the meaning of the words they generate is a hotly debated issue. A big challenge is that LLMs are black boxes; they can make predictions and decisions on the order of words, but they cannot communicate the reasons for doing so.
Ellie Pavlick at Brown University is building models that could help understand how LLMs process language compared with humans. In this episode of The Joy of Why, Pavlick discusses what we know and don’t know about LLM language processing, how their processes differ from humans, and how understanding LLMs better could also help us better appreciate our own capacity for knowledge and creativity.

324 Listeners

832 Listeners

560 Listeners

532 Listeners

246 Listeners

1,064 Listeners

82 Listeners

4,176 Listeners

2,348 Listeners

448 Listeners

502 Listeners

251 Listeners

331 Listeners

29 Listeners

393 Listeners

507 Listeners