
Sign up to save your podcasts
Or


In this episode, Pallavi Koppol, Research Scientist at Databricks, explores the importance of domain-specific intelligence in large language models (LLMs). She discusses how enterprises need models tailored to their unique jargon, data, and tasks rather than relying solely on general benchmarks.
Highlights include:
- Why benchmarking LLMs for domain-specific tasks is critical for enterprise AI.
- An introduction to the Databricks Intelligence Benchmarking Suite (DIBS).
- Evaluating models on real-world applications like RAG, text-to-JSON, and function calling.
- The evolving landscape of open-source vs. closed-source LLMs.
- How industry and academia can collaborate to improve AI benchmarking.
By Databricks4.8
2020 ratings
In this episode, Pallavi Koppol, Research Scientist at Databricks, explores the importance of domain-specific intelligence in large language models (LLMs). She discusses how enterprises need models tailored to their unique jargon, data, and tasks rather than relying solely on general benchmarks.
Highlights include:
- Why benchmarking LLMs for domain-specific tasks is critical for enterprise AI.
- An introduction to the Databricks Intelligence Benchmarking Suite (DIBS).
- Evaluating models on real-world applications like RAG, text-to-JSON, and function calling.
- The evolving landscape of open-source vs. closed-source LLMs.
- How industry and academia can collaborate to improve AI benchmarking.

391 Listeners

26,409 Listeners

9,748 Listeners

479 Listeners

629 Listeners

303 Listeners

234 Listeners

267 Listeners

2,550 Listeners

10,205 Listeners

1,582 Listeners

564 Listeners

670 Listeners

3,524 Listeners

32 Listeners