In this episode, we dive deep into AI infrastructure at Google Cloud—what it means, why it matters, and how it’s evolving.
Our guest shares insights from over 8 years at Google and previous experience as a hardware engineer at IBM, bringing a unique perspective on the nuts and bolts that power today’s AI revolution. We explore:
✅ The foundations of AI infrastructure—chips, networking, storage, and workload-optimized systems
✅ How Google’s custom hardware (TPUs, Axion, ARM processors) differentiates it from AWS, Microsoft, Oracle, and IBM
✅ The concept of the AI Hypercomputer—a reference architecture combining hardware, software, and flexible consumption models
✅ Key announcements from Google, including NVIDIA Blackwell, GB200, Ironwood TPUs, and Cluster Director
✅ Why inference (not just training) is now the hot topic—and how Google helps customers lower the cost per inference
From hardware assembly roots to leading AI infrastructure strategy, this conversation highlights how Google builds and scales the systems behind Gemini, Vertex AI, and beyond.
📌 If you’re curious about the future of AI infrastructure, supercomputing, and how enterprises can actually run large-scale AI workloads efficiently—this one’s for you.
🔔 Don’t forget to like, comment, and subscribe for more in-depth discussions on technology and innovation!
Sign Up Now for my online course "The Cloud Strategy Master Class":
► On my web site: https://lnkd.in/gcxcrX and use promo code LINKEDIN20 to receive a 20% discount.
► On Udemy: https://www.udemy.com/course/cloud-strategy-master-class/
PODCASTS: Listen wherever you get your podcasts:
► Let's Talk Shop: http://letstalkshop.buzzsprout.com
► Reality Distortion Fields RDFs Podcast: https://youtu.be/88z1UiVaV00
Follow me:
► TikTok: @ekhnaser
► Instagram: @ekhnaser
► Twitter: @ekhnaser
► LinkedIn: https://www.linkedin.com/in/eliaskhnaser/
► Website: www.eliaskhnaser.com