
Sign up to save your podcasts
Or
In this episode, we discuss using AWS Lambda for machine learning inference. We cover the tradeoffs between GPUs and CPUs for ML, tools like ggml and llama.cpp for running models on CPUs, and share examples where we've experimented with Lambda for ML like podcast transcription, medical imaging, and natural language processing. While Lambda ML is still quite experimental, it can be a viable option for certain use cases.
💰 SPONSORS 💰
Do you have any AWS questions you would like us to address?
4.6
1111 ratings
In this episode, we discuss using AWS Lambda for machine learning inference. We cover the tradeoffs between GPUs and CPUs for ML, tools like ggml and llama.cpp for running models on CPUs, and share examples where we've experimented with Lambda for ML like podcast transcription, medical imaging, and natural language processing. While Lambda ML is still quite experimental, it can be a viable option for certain use cases.
💰 SPONSORS 💰
Do you have any AWS questions you would like us to address?
272 Listeners
378 Listeners
30,734 Listeners
284 Listeners
41 Listeners
585 Listeners
624 Listeners
435 Listeners
202 Listeners
140 Listeners
183 Listeners
138 Listeners
453 Listeners
22 Listeners
62 Listeners