
Sign up to save your podcasts
Or


Coffee Sessions #48 with Julien Chaumond, Tour of Upcoming Features on the Hugging Face Model Hub.
Join the Community: https://go.mlops.community/YTJoinIn
Get the newsletter: https://go.mlops.community/YTNewsletter
//Abstract
Julien Chaumond’s Tour of Upcoming Features on the Hugging Face Model Hub. Our MLOps community guest in this episode is Julien Chaumond, the CTO of Hugging Face - every data scientist’s favorite NLP Swiss army knife.
Julien, David, and Demetrios spoke about many topics, including:
Infra for hosting models/model hubs
Inference widgets for companies with CPUs & GPUs (for companies)
Auto NLP, which trains models
“Infrastructure as a service”
// Bio
Julien Chaumond is Chief Technical Officer at Hugging Face, a Brooklyn and Paris-based startup working on Machine learning and Natural Language Processing, and is passionate about democratizing state-of-the-art AI/ML for everyone.
--------------- ✌️Connect With Us ✌️ -------------
Join our Slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with David on LinkedIn: https://www.linkedin.com/in/aponteanalytics/
Connect with Julien on LinkedIn: https://www.linkedin.com/in/julienchaumond/
Timestamps:
[00:00] Introduction to Julien Chaumond
[01:57] Julien's background in tech
[04:35] "I have this vision of building a community where the greatest people in AI can come together and basically invent the future of Machine Learning together."
[04:55] What is Hugging Face?
[06:45] Start of open-source in Hugging Face
[07:50] Chatbox experiment (reference resolution system) - linking pronouns to the subjects of sentences
[10:20] From a project to a company
[11:57] Importance of platform
[14:25] "Transfer learning is an efficient way of Machine Learning. Providing your platform around change that people want to start from a pre-trained model and fine-tune it into the specific use case is something that can be big, so we built some stuff to help people do that."
[15:35] Narrowing down the scope of service to provide
[16:27] "We have some vision of what we want to build, but a lot of it is the small incremental improvements that we bring to the platform. I think it's the natural way of building stuff nowadays because Machine Learning is moving so fast."
[20:00] Model Hubs
[22:37] "We're guaranteeing that we don't build anything that introduces any lag to Hugging Face because we're using GitHub. You'll have that peace of mind."
[26:31] Storing model artifacts
[27:00] AWS - cache - stored to an edge location all around the globe
[28:39] Inference widgets powering
[27:17] "For each model on the model hub, we try to ensure that we have the metadata about the model to be able to actually run it."
[32:11] Deploying infra function
[32:38] "Depending on the model and library, we optimize the custom containers to make sure that they run as fast as possible on the target hardware that we have."
[34:59] "Machine Learning is still pretty much hardware dependent."
[36:11] Hardware usage
[39:04] "CPU is super cheap. If you are able to run Berks served with a 1-millisecond on CPU because you have powerful optimizations, you don't really need GPUs anymore. It's cost-efficient and energy-efficient."
[41:10] "It may sound like a super cliche, but the team that you assembled is everything."
[43:22] War stories in Hugging Face
[44:12] "Our goal is more forward-looking to be helpful as much as we can to the community."
[48:25] Hugging Face accessibility
By Demetrios4.6
2323 ratings
Coffee Sessions #48 with Julien Chaumond, Tour of Upcoming Features on the Hugging Face Model Hub.
Join the Community: https://go.mlops.community/YTJoinIn
Get the newsletter: https://go.mlops.community/YTNewsletter
//Abstract
Julien Chaumond’s Tour of Upcoming Features on the Hugging Face Model Hub. Our MLOps community guest in this episode is Julien Chaumond, the CTO of Hugging Face - every data scientist’s favorite NLP Swiss army knife.
Julien, David, and Demetrios spoke about many topics, including:
Infra for hosting models/model hubs
Inference widgets for companies with CPUs & GPUs (for companies)
Auto NLP, which trains models
“Infrastructure as a service”
// Bio
Julien Chaumond is Chief Technical Officer at Hugging Face, a Brooklyn and Paris-based startup working on Machine learning and Natural Language Processing, and is passionate about democratizing state-of-the-art AI/ML for everyone.
--------------- ✌️Connect With Us ✌️ -------------
Join our Slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with David on LinkedIn: https://www.linkedin.com/in/aponteanalytics/
Connect with Julien on LinkedIn: https://www.linkedin.com/in/julienchaumond/
Timestamps:
[00:00] Introduction to Julien Chaumond
[01:57] Julien's background in tech
[04:35] "I have this vision of building a community where the greatest people in AI can come together and basically invent the future of Machine Learning together."
[04:55] What is Hugging Face?
[06:45] Start of open-source in Hugging Face
[07:50] Chatbox experiment (reference resolution system) - linking pronouns to the subjects of sentences
[10:20] From a project to a company
[11:57] Importance of platform
[14:25] "Transfer learning is an efficient way of Machine Learning. Providing your platform around change that people want to start from a pre-trained model and fine-tune it into the specific use case is something that can be big, so we built some stuff to help people do that."
[15:35] Narrowing down the scope of service to provide
[16:27] "We have some vision of what we want to build, but a lot of it is the small incremental improvements that we bring to the platform. I think it's the natural way of building stuff nowadays because Machine Learning is moving so fast."
[20:00] Model Hubs
[22:37] "We're guaranteeing that we don't build anything that introduces any lag to Hugging Face because we're using GitHub. You'll have that peace of mind."
[26:31] Storing model artifacts
[27:00] AWS - cache - stored to an edge location all around the globe
[28:39] Inference widgets powering
[27:17] "For each model on the model hub, we try to ensure that we have the metadata about the model to be able to actually run it."
[32:11] Deploying infra function
[32:38] "Depending on the model and library, we optimize the custom containers to make sure that they run as fast as possible on the target hardware that we have."
[34:59] "Machine Learning is still pretty much hardware dependent."
[36:11] Hardware usage
[39:04] "CPU is super cheap. If you are able to run Berks served with a 1-millisecond on CPU because you have powerful optimizations, you don't really need GPUs anymore. It's cost-efficient and energy-efficient."
[41:10] "It may sound like a super cliche, but the team that you assembled is everything."
[43:22] War stories in Hugging Face
[44:12] "Our goal is more forward-looking to be helpful as much as we can to the community."
[48:25] Hugging Face accessibility

1,094 Listeners

625 Listeners

301 Listeners

344 Listeners

145 Listeners

227 Listeners

205 Listeners

95 Listeners

518 Listeners

131 Listeners

228 Listeners

35 Listeners

22 Listeners

42 Listeners

71 Listeners