
Sign up to save your podcasts
Or
In episode 75 of The Gradient Podcast, Daniel Bashir speaks to Riley Goodside.
Riley is a Staff Prompt Engineer at Scale AI. Riley began posting GPT-3 prompt examples and screenshot demonstrations in 2022. He previously worked as a data scientist at OkCupid, Grindr, and CopyAI.
Have suggestions for future podcast guests (or other feedback)? Let us know here or reach us at [email protected]
Subscribe to The Gradient Podcast: Apple Podcasts | Spotify | Pocket Casts | RSSFollow The Gradient on Twitter
Outline:
* (00:00) Intro
* (01:37) Riley’s journey to becoming the first Staff Prompt Enginer
* (02:00) data science background in online dating industry
* (02:15) Sabbatical + catching up on LLM progress
* (04:00) AI Dungeon and first taste of GPT-3
* (05:10) Developing on codex, ideas about integrating codex with Jupyter Notebooks, start of posting on Twitter
* (08:30) “LLM ethnography”
* (09:12) The history of prompt engineering: in-context learning, Reinforcement Learning from Human Feedback (RLHF)
* (10:20) Models used to be harder to talk to
* (10:45) The three eras
* (10:45) 1 - Pre-trained LM era—simple next-word predictors
* (12:54) 2 - Instruction tuning
* (16:13) 3 - RLHF and overcoming instruction tuning’s limitations
* (19:24) Prompting as subtractive sculpting, prompting and AI safety
* (21:17) Riley on RLHF and safety
* (24:55) Riley’s most interesting experiments and observations
* (25:50) Mode collapse in RLHF models
* (29:24) Prompting models with very long instructions
* (33:13) Explorations with regular expressions, chain-of-thought prompting styles
* (36:32) Theories of in-context learning and prompting, why certain prompts work well
* (42:20) Riley’s advice for writing better prompts
* (49:02) Debates over prompt engineering as a career, relevance of prompt engineers
* (58:55) Outro
Links:
* Riley’s Twitter and LinkedIn
* Talk: LLM Prompt Engineering and RLHF: History and Techniques
4.7
4747 ratings
In episode 75 of The Gradient Podcast, Daniel Bashir speaks to Riley Goodside.
Riley is a Staff Prompt Engineer at Scale AI. Riley began posting GPT-3 prompt examples and screenshot demonstrations in 2022. He previously worked as a data scientist at OkCupid, Grindr, and CopyAI.
Have suggestions for future podcast guests (or other feedback)? Let us know here or reach us at [email protected]
Subscribe to The Gradient Podcast: Apple Podcasts | Spotify | Pocket Casts | RSSFollow The Gradient on Twitter
Outline:
* (00:00) Intro
* (01:37) Riley’s journey to becoming the first Staff Prompt Enginer
* (02:00) data science background in online dating industry
* (02:15) Sabbatical + catching up on LLM progress
* (04:00) AI Dungeon and first taste of GPT-3
* (05:10) Developing on codex, ideas about integrating codex with Jupyter Notebooks, start of posting on Twitter
* (08:30) “LLM ethnography”
* (09:12) The history of prompt engineering: in-context learning, Reinforcement Learning from Human Feedback (RLHF)
* (10:20) Models used to be harder to talk to
* (10:45) The three eras
* (10:45) 1 - Pre-trained LM era—simple next-word predictors
* (12:54) 2 - Instruction tuning
* (16:13) 3 - RLHF and overcoming instruction tuning’s limitations
* (19:24) Prompting as subtractive sculpting, prompting and AI safety
* (21:17) Riley on RLHF and safety
* (24:55) Riley’s most interesting experiments and observations
* (25:50) Mode collapse in RLHF models
* (29:24) Prompting models with very long instructions
* (33:13) Explorations with regular expressions, chain-of-thought prompting styles
* (36:32) Theories of in-context learning and prompting, why certain prompts work well
* (42:20) Riley’s advice for writing better prompts
* (49:02) Debates over prompt engineering as a career, relevance of prompt engineers
* (58:55) Outro
Links:
* Riley’s Twitter and LinkedIn
* Talk: LLM Prompt Engineering and RLHF: History and Techniques
10,688 Listeners
323 Listeners
189 Listeners
1,260 Listeners
196 Listeners
287 Listeners
9,048 Listeners
87 Listeners
387 Listeners
5,420 Listeners
146 Listeners
15,207 Listeners
2,187 Listeners
75 Listeners
134 Listeners