
Sign up to save your podcasts
Or
Google's Gemini 1.5 is here, boasting a mind-blowing 1 million token context window! 🤯 Join Allen and Linda as they dive deep into this experimental AI, exploring its capabilities, limitations, and potential use cases. 🤔
They share their experiences testing Gemini 1.5 with original content, including Two Voice Devs transcripts and synthetic videos, and discuss the challenges of finding data that hasn't already been used to train the AI. 🧐
Get ready for a lively discussion on hallucinations, the future of content creation, and the ethical questions surrounding these powerful language models. 🤖
More info:
* https://blog.google/technology/ai/google-gemini-next-generation-model-february-2024/
* https://developers.googleblog.com/2024/02/gemini-15-available-for-private-preview-in-google-ai-studio.html
* https://openai.com/sora
Timestamps:
00:00:00 Introduction
00:01:05 Notable features of Gemini 1.5
00:02:57 What is a token?
00:06:39 Linda's test with Danish citizenship PDF
00:09:33 Allen's test with Les Miserables and needle in a haystack
00:12:27 Testing with Data Portability API data
00:14:28 Linda's test with YouTube search history and Netflix recommendations
00:17:44 Allen's test with Two Voice Devs transcripts
00:21:32 Issues with counting and hallucinations
00:24:21 Testing with OpenAI's Sora AI synthetic videos
00:30:05 Ethical questions and the future of content creation
00:31:50 Potential use cases for large context windows
00:36:34 API limitations and challenges
00:37:39 Performance and cost considerations
00:41:34 Comparison with retrieval augmented generation and vector databases
00:44:21 Generating summaries and markers from this transcript
Leave your thoughts and questions in the comments below!
1
11 ratings
Google's Gemini 1.5 is here, boasting a mind-blowing 1 million token context window! 🤯 Join Allen and Linda as they dive deep into this experimental AI, exploring its capabilities, limitations, and potential use cases. 🤔
They share their experiences testing Gemini 1.5 with original content, including Two Voice Devs transcripts and synthetic videos, and discuss the challenges of finding data that hasn't already been used to train the AI. 🧐
Get ready for a lively discussion on hallucinations, the future of content creation, and the ethical questions surrounding these powerful language models. 🤖
More info:
* https://blog.google/technology/ai/google-gemini-next-generation-model-february-2024/
* https://developers.googleblog.com/2024/02/gemini-15-available-for-private-preview-in-google-ai-studio.html
* https://openai.com/sora
Timestamps:
00:00:00 Introduction
00:01:05 Notable features of Gemini 1.5
00:02:57 What is a token?
00:06:39 Linda's test with Danish citizenship PDF
00:09:33 Allen's test with Les Miserables and needle in a haystack
00:12:27 Testing with Data Portability API data
00:14:28 Linda's test with YouTube search history and Netflix recommendations
00:17:44 Allen's test with Two Voice Devs transcripts
00:21:32 Issues with counting and hallucinations
00:24:21 Testing with OpenAI's Sora AI synthetic videos
00:30:05 Ethical questions and the future of content creation
00:31:50 Potential use cases for large context windows
00:36:34 API limitations and challenges
00:37:39 Performance and cost considerations
00:41:34 Comparison with retrieval augmented generation and vector databases
00:44:21 Generating summaries and markers from this transcript
Leave your thoughts and questions in the comments below!
356 Listeners
3 Listeners