AWS re:Invent 2019

CMP324-R1: Deliver high performance ML inference with AWS Inferentia


Listen Later

Customers across diverse industries are defining entirely new categories of products and experiences by running intelligent applications that use ML at the core. These applications are becoming more expensive to run in production. AWS Inferentia is a custom-built machine learning inference chip designed to provide high throughput, low latency inference performance at an extremely low cost. Each chip provides hundreds of TOPS of inference throughput to allow complex models to make fast predictions. Join this session to see the latest developments using AWS Inferentia and how they can lower your inference costs in the future.
...more
View all episodesView all episodes
Download on the App Store

AWS re:Invent 2019By AWS

  • 4.3
  • 4.3
  • 4.3
  • 4.3
  • 4.3

4.3

6 ratings


More shows like AWS re:Invent 2019

View all
AWS Podcast by Amazon Web Services

AWS Podcast

200 Listeners