We recorded August 7th, right before ChatGPT launched.
We dove into GPT open source, OpenCode, Ollama Turbo, and deep agent setups.
I wanted to see LangChain’s open suite and test agent environments.
OpenCode stood out for its flexibility — multiple model providers, easy local setup, works with Ollama Turbo for $20/month.
LM Studio runs similarly.
I’m considering a high-spec NVIDIA rig and DGX Spark for local inference.
GPT-OSS is cheap, fast, and excellent for coding and tool-calling, but weaker on general knowledge.
Running it locally means more setup work but more control.
Hybrid local-plus-cloud routing feels inevitable.
We demoed OpenAgent Platform — fast, multi-provider agents without writing code.
Then explored LangChain SWE — an open-source, multi-threaded coding agent with planner/programmer loops, GitHub integration, Daytona sandboxes, and detailed token-cost tracking.
We looked at Vercel’s v0 API for quick generative UI, and the potential to run it privately for internal teams.
I closed with Google’s upcoming AI-mode ads and Societies.io — a virtual audience simulation tool for testing and optimizing content before publishing.
Chapters
00:00 Introduction to ChatGPT Launch and Demos
01:40 Exploring Open Code and LangChain
04:37 Local Inference and Olamma Integration
07:25 Cloud Acceleration with Turbo Service
10:11 Open Source Model Benchmarks and Feedback