
Sign up to save your podcasts
Or


The paper modifies the DeepMind Chinchilla scaling laws for large language models (LLMs) to include the cost of inference. The analysis suggests that LLM researchers should train smaller and longer models than the Chinchilla-optimal for large inference demand.
https://arxiv.org/abs//2401.00448
YouTube: https://www.youtube.com/@ArxivPapers
TikTok: https://www.tiktok.com/@arxiv_papers
Apple Podcasts: https://podcasts.apple.com/us/podcast/arxiv-papers/id1692476016
Spotify: https://podcasters.spotify.com/pod/show/arxiv-papers
By Igor Melnyk5
33 ratings
The paper modifies the DeepMind Chinchilla scaling laws for large language models (LLMs) to include the cost of inference. The analysis suggests that LLM researchers should train smaller and longer models than the Chinchilla-optimal for large inference demand.
https://arxiv.org/abs//2401.00448
YouTube: https://www.youtube.com/@ArxivPapers
TikTok: https://www.tiktok.com/@arxiv_papers
Apple Podcasts: https://podcasts.apple.com/us/podcast/arxiv-papers/id1692476016
Spotify: https://podcasters.spotify.com/pod/show/arxiv-papers

971 Listeners

1,996 Listeners

435 Listeners

113,393 Listeners

10,273 Listeners

5,548 Listeners

219 Listeners

52 Listeners

98 Listeners

463 Listeners