
Sign up to save your podcasts
Or


In this episode, we discuss using AWS Lambda for machine learning inference. We cover the tradeoffs between GPUs and CPUs for ML, tools like ggml and llama.cpp for running models on CPUs, and share examples where we've experimented with Lambda for ML like podcast transcription, medical imaging, and natural language processing. While Lambda ML is still quite experimental, it can be a viable option for certain use cases.
💰 SPONSORS 💰
Do you have any AWS questions you would like us to address?
By AWS Bites4.7
1212 ratings
In this episode, we discuss using AWS Lambda for machine learning inference. We cover the tradeoffs between GPUs and CPUs for ML, tools like ggml and llama.cpp for running models on CPUs, and share examples where we've experimented with Lambda for ML like podcast transcription, medical imaging, and natural language processing. While Lambda ML is still quite experimental, it can be a viable option for certain use cases.
💰 SPONSORS 💰
Do you have any AWS questions you would like us to address?

56,944 Listeners

204 Listeners

242 Listeners

14,543 Listeners

26 Listeners