Kabir's Tech Dives

AI Fact Checking | Future of Truth Online Using


Listen Later

This episode is about Search-Augmented Factuality Evaluator (SAFE), a novel, cost-effective method for automatically evaluating the factuality of long-form text generated by large language models (LLMs). SAFE leverages LLMs and Google Search to assess the accuracy of individual facts within a response, outperforming human annotators in accuracy and efficiency. The researchers also created LongFact, a new benchmark dataset of 2,280 prompts designed to test long-form factuality across diverse topics, and proposed F1@K, a new metric that incorporates both precision and recall, accounting for the desired length of a factual response. Extensive benchmarking across thirteen LLMs demonstrates that larger models generally exhibit higher factuality, and the paper thoroughly addresses reproducibility and ethical considerations.

Send us a text

Support the show


Podcast:
https://kabir.buzzsprout.com


YouTube:
https://www.youtube.com/@kabirtechdives

Please subscribe and share.

...more
View all episodesView all episodes
Download on the App Store

Kabir's Tech DivesBy Kabir

  • 4.7
  • 4.7
  • 4.7
  • 4.7
  • 4.7

4.7

33 ratings


More shows like Kabir's Tech Dives

View all
Hard Fork by The New York Times

Hard Fork

5,426 Listeners