Share AI Stories
Share to email
Share to Facebook
Share to X
By Neil Leiser
The podcast currently has 50 episodes available.
Our guest today is Lewis Tunstall, LLM Engineer and researcher at Hugging Face and book author of "Natural Language Processing with Transformers".
In our conversation, we dive into topological machine learning and talk about giotto-tda, a high performance topological ml Python library that Lewis worked on. We then dive into LLMs and Transformers. We discuss the pros and cons of open source vs closed source LLMs and explain the differences between encoder and decoder transformer architectures. Lewis finally explains his day-to-day at Hugging Face and his current work on fine-tuning LLMs.
If you enjoyed the episode, please leave a 5 star review and subscribe to the AI Stories Youtube channel.
Link to Train in Data courses (use the code AISTORIES to get a 10% discount): https://www.trainindata.com/courses?affcode=1218302_5n7kraba
Natural Language Processing with Transformers book: https://www.oreilly.com/library/view/natural-language-processing/9781098136789/
Giotto-tda library: https://github.com/giotto-ai/giotto-tda
KTO alignment paper: https://arxiv.org/abs/2402.01306
Follow Lewis on LinkedIn: https://www.linkedin.com/in/lewis-tunstall/
Follow Neil on LinkedIn: https://www.linkedin.com/in/leiserneil/
---
(00:00) - Intro
(03:00) - How Lewis Got into AI
(05:33) - From Kaggle Competitions to Data Science Job
(11:09) - Get an actual Data Science Job!
(15:18) - Deep Learning or Excel?
(19:14) - Topological Machine Learning
(28:44) - Open Source VS Closed Source LLMs
(41:44) - Writing a Book on Transformers
(52:33) - Comparing BERT, Early Transformers, and GPT-4
(54:48) - Encoder and Decoder Architectures
(59:48) - Day-To-Day Work at Hugging Face
(01:09:06) - DPO and KTO
(01:12:58) - Stories and Career Advice
Our guest today is Maria Vecthomova, ML Engineering Manager at Ahold Delhaize and Co-Founder of Marvelous MLOps.
In our conversation, we first talk about code best practices for Data Scientists. We then dive into MLOps, discuss the main components required to deploy a model in production and get an overview of one of Maria's project where she built and deployed a fraud detection algorithm. We finally talk about content creation, career advice and the differences between an ML and an MLOps engineer.
If you enjoyed the episode, please leave a 5 star review and subscribe to the AI Stories Youtube channel.
Link to Train in Data courses (use the code AISTORIES to get a 10% discount): https://www.trainindata.com/courses?affcode=1218302_5n7kraba
Check out Marvelous MLOps: https://marvelousmlops.substack.com/
Follow Maria on LinkedIn: https://www.linkedin.com/in/maria-vechtomova/
Follow Neil on LinkedIn: https://www.linkedin.com/in/leiserneil/
---
(00:00) - Intro
(02:59) - Maria’s Journey to MLOps
(08:50) - Code Best Practices
(18:39) - MLOps Infrastructure
(29:10) - ML Engineering for Fraud Detection
(40:42) - Content Creation & Marvelous MLOps
(49:01) - ML Engineer vs MLOps Engineer
(56:00) - Stories & Career Advice
Our guest today is Reah Miyara. Reah is currently working on LLMs evaluation at OpenAI and previously worked at Google and IBM.
In our conversation, Reah shares his experience working as a product lead for Google's graph-based machine learning portfolio. He then explains how he joined OpenAI and his role there. We finally talk about LLMs evaluation, AGI, LLMs safety and the future of the field.
If you enjoyed the episode, please leave a 5 star review and subscribe to the AI Stories Youtube channel.
Link to Train in Data courses (use the code AISTORIES to get a 10% discount): https://www.trainindata.com/courses?affcode=1218302_5n7kraba
Follow Reah on LinkedIn: https://www.linkedin.com/in/reah/
Follow Neil on LinkedIn: https://www.linkedin.com/in/leiserneil/
---
(00:00) - Intro
(03:09) - Getting into AI and Machine Learning
(08:33) - Why Stay in AI?
(11:39) - From Software Engineer to Product Manager
(18:27) - Experience at Google
(25:28) - Applications of Graph ML
(31:10) - Joining OpenAI
(35:15) - LLM Evaluation
(44:30) - The Future of GenAI and LLMs
(55:48) - Safety Metrics for LLMs
(1:00:30) - Career Advice
Our guest today is Erwin Huizenga, Machine Learning Lead at Google and expert in Applied AI and LLMOps.
In our conversation, Erwin first discusses how he got into the field and his previous experiences at SAS and IBM. We then talk about his work at Google: from the early days of cloud computing when he joined the company to his current work on Gemini. We finally dive into the world of LLMOps and share insights on how to evaluate LLMs, how to monitor their performances and how to deploy them.
If you enjoyed the episode, please leave a 5 star review and subscribe to the AI Stories Youtube channel.
Link to Train in Data courses (use the code AISTORIES to get a 10% discount): https://www.trainindata.com/courses?affcode=1218302_5n7kraba
Erwin's LLMOps coursera course: https://www.deeplearning.ai/short-courses/llmops/
Follow Erwin on LinkedIn: https://www.linkedin.com/in/erwinhuizenga/
Follow Neil on LinkedIn: https://www.linkedin.com/in/leiserneil/
---
(00:00) - Intro
(05:04) - Early Experiences
(15:51) - Joining Google
(20:20) - Early Days of Cloud Computing
(26:18) - Advantages of Cloud Infrastructure
(30:09) - Gemini and its Launch
(37:32) - Gemini vs Other LLMs
(46:15) - LLMOps
(50:50) - Evaluating and Monitoring LLMs
(57:34) - Deploying LLMs vs Traditional ML Models
(01:01:07) - Personal Stories and Career Insights
Our guest today is Andras Palffy, Co-Founder of Perciv AI: a startup offering AI based software solutions to build robust and affordable autonomous systems.
In our conversation, we first talk about Andras' PhD focusing on road users detection. We dive into AI applied to autonomous driving and discuss the pros and cons of the most common pieces of hardware: cameras, lidars and radars. We then focus on Perciv AI. Andras explains why he decided to focus on radars and how he uses Deep Learning algorithms to enable autonomous systems. He finally gives his take on the future of autonomous vehicles and shares learnings from his experience in the field.
If you enjoyed the episode, please leave a 5 star review and subscribe to the AI Stories Youtube channel.
Link to Train in Data courses (use the code AISTORIES to get a 10% discount): https://www.trainindata.com/courses?affcode=1218302_5n7kraba
To learn more about Perciv AI: https://www.perciv.ai/
Follow Andras on LinkedIn: https://www.linkedin.com/in/andraspalffy/
Follow Neil on LinkedIn: https://www.linkedin.com/in/leiserneil/
---
(00:00) - Intro
(02:57) - Andras' Journey into AI
(06:11) - Getting into Robotics
(10:15) - Evolution of Computer Vision Algorithms
(13:38) - PhD on Autonomous Driving & Road Users Detection
(28:01) - Launching Perciv AI
(35:19) - Augmenting Radars Performance with AI
(44:45) - Inside Perciv AI: Roles, Challenges, and Stories
(48:43) - Future of Autonomous Vehicles and Road Safety
(51:46) - Solving a Technical Challenge with Camera Calibration
(54:12) - Andras' First Self-Driving Car Experience
(56:09) - Career Advice
Our guest today is Franziska Kirschner, Co-Founder of Intropy AI and ex AI & Product Lead at Tractable: the world’s first computer vision unicorn.
In our conversation, we dive into Franziska's PhD, her career at Tractable and her experience building deep learning algorithms for computer vision products. She explains how she climbed the ladder from intern to AI Lead and shares how she launched new AI product lines generating £ millions in revenues.
If you enjoyed the episode, please leave a 5 star review and subscribe to the AI Stories Youtube channel.
Link to Train in Data courses (use the code AISTORIES to get a 10% discount): https://www.trainindata.com/courses?affcode=1218302_5n7kraba
Follow Franziska on LinkedIn: https://www.linkedin.com/in/frankirsch/
Follow Neil on LinkedIn: https://www.linkedin.com/in/leiserneil/
---
(00:00) - Introduction
(03:08) - Franziska's Journey into AI
(05:17) - Franziska's PhD in Condensed Matter Physics
(15:12) - Transition from Physics to AI
(19:20) - Deep Learning & Impact at Tractable
(33:21) - AI Researcher vs AI Product Manager
(37:52) - The Impact of AI on Scrapyards
(43:14) - Key Steps in Launching New AI Products
(53:31) - Founding Intropy AI
(01:00:37) - The Potato Travels
(01:04:10) - Advice for Career Progression
Our guest today is Maxime Labonne, GenAI Expert, book author and developer of NeuralBeagle14-7B, one of the best performing 7B params model on the open LLM leaderboard.
In our conversation, we dive deep into the world of GenAI. We start by explaining how to get into the field and resources needed to get started. Maxime then goes through the 4 steps used to build LLMs: Pre training, supervised fine-tuning, human feedback and merging models. Throughout our conversation, we also discuss RAG vs fine-tuning, QLoRA & LoRA, DPO vs RLHF and how to deploy LLMs in production.
If you enjoyed the episode, please leave a 5 star review and subscribe to the AI Stories Youtube channel.
Link to Train in Data courses (use the code AISTORIES to get a 10% discount): https://www.trainindata.com/courses?affcode=1218302_5n7kraba
Check out Maxime's LLM course: https://github.com/mlabonne/llm-course
Follow Maxime on LinkedIn: https://www.linkedin.com/in/maxime-labonne/
Follow Neil on LinkedIn: https://www.linkedin.com/in/leiserneil/
---
(00:00) - Intro
(02:37) - From Cybersecurity to AI
(06:05) - GenAI at Airbus
(13:29) - What does Maxime use ChatGPT for?
(15:31) - Getting into GenAI and learning resources
(22:23) - Steps to build your own LLM
(26:44) - Pre-training
(29:16) - Supervised fine-tuning, QLoRA & LoRA
(34:45) - RAG vs fine-tuning
(37:53) - DPO vs RLHF
(41:01) - Merging Models
(45:05) - Deploying LLMs
(46:52) - Stories and career advice
Our guest today is Harpreet Sahota, Deep Learning Developer Relations Manager at Deci AI.
In our conversation, we first talk about Harpreet’s work as a Biostatistician and dive into A/B testing. We then talk about Deci AI and Neural Architecture Search (NAS): the algorithm used to build powerful deep learning models like YOLO-NAS. We finally dive into GenAI where Harpreet shares 7 prompting tips and explains how Retrieval Augmented Generation (RAG) works.
If you enjoyed the episode, please leave a 5 star review and subscribe to the AI Stories Youtube channel.
Link to Train in Data courses (use the code AISTORIES to get a 10% discount): https://www.trainindata.com/courses?affcode=1218302_5n7kraba
Follow Harpreet on LinkedIn: https://www.linkedin.com/in/harpreetsahota204/
Follow Neil on LinkedIn: https://www.linkedin.com/in/leiserneil/
---
(00:00) - Intro
(02:34) - Harpreet's Journey into Data Science
(07:00) - A/B Testing
(17:50) - DevRel at Deci AI
(26:25) - Deci AI: Products and Services
(32:22) - Neural Architecture Search (NAS)
(36:58) - GenAI
(39:53) - Tools for Playing with LLMs
(42:56) - Mastering Prompt Engineering
(46:35) - Retrieval Augmented Generation (RAG)
(54:12) - Career Advice
Our guest today is Ryan Shannon, AI Investor at Radical Ventures, a world-known venture capital firm investing exclusively in AI. Radical's portfolio includes hot startups like Cohere, Covariant, V7 and many more.
In our conversation, we talk about how to start an AI company & what makes a good founding team. Ryan also explains what he and Radical look for when investing and how they help their portfolio after the investment. We finally chat about some cool AI Startups like Twelve Labs and get Ryan’s predictions on hot startups in 2024.
If you enjoyed the episode, please leave a 5 star review and subscribe to the AI Stories Youtube channel.
Link to Train in Data courses (use the code AISTORIES to get a 10% discount): https://www.trainindata.com/courses?affcode=1218302_5n7kraba
Follow Ryan on LinkedIn: https://www.linkedin.com/in/ryan-shannon-1b3a7884/
Follow Neil on LinkedIn: https://www.linkedin.com/in/leiserneil/
---
(0:00) - Intro
(2:42) - Ryan's background and journey into AI investing
(11:15) - Radical Ventures
(14:34) - How to keep up with AI breakthroughs?
(22:42) - How Ryan finds and evaluates founders to invest in
(32:54) - What makes a good founding team?
(38:57) - Ryan's role at Radical
(45:53) - How to start an AI company
(50:22) - Twelve Labs
(59:19) - Future of AI and hot startups in 2024
(1:09:48) - Career advice
Our guest today is Christoph Molnar, expert in Interpretable Machine Learning and book author.
In our conversation, we dive into the field of Interpretable ML. Christoph explains the difference between post hoc and model agnostic approaches as well as global and local model agnostic methods. We dig into several interpretable ML techniques including permutation feature importance, SHAP and Lime. We also talk about the importance of interpretability and how it can help you build better models and impact businesses.
If you enjoyed the episode, please leave a 5 star review and subscribe to the AI Stories Youtube channel.
Link to Train in Data courses (use the code AISTORIES to get a 10% discount): https://www.trainindata.com/courses?affcode=1218302_5n7kraba
Follow Christoph on LinkedIn: https://www.linkedin.com/in/christoph-molnar/
Check out the books he wrote here: https://christophmolnar.com/books/
Follow Neil on LinkedIn: https://www.linkedin.com/in/leiserneil/
---
(00:00) - Introduction
(02:42) - Christoph's Journey into Data Science and AI
(07:23) - What is Interpretable ML?
(18:57) - Global Model Agnostic Approaches
(24:20) - Practical Applications of Feature Importance
(28:37) - Local Model Agnostic Approaches
(31:17) - SHAP and LIME
(40:20) - Advice for Implementing Interpretable Techniques
(43:47) - Modelling Mindsets
(48:04) - Stats vs ML Mindsets
(51:17) - Future Plans & Career Advice
The podcast currently has 50 episodes available.
3,552 Listeners
26,046 Listeners
7,754 Listeners
285 Listeners
85,213 Listeners
177 Listeners
669 Listeners
197 Listeners
7,175 Listeners
64 Listeners
22 Listeners
5,193 Listeners
10 Listeners
82 Listeners
1,022 Listeners