Practical AI: Machine Learning, Data Science

Collaboration & evaluation for LLM apps

01.23.2024 - By Changelog MediaPlay

Download our free app to listen on your phone

Download on the App StoreGet it on Google Play

Small changes in prompts can create large changes in the output behavior of generative AI models. Add to that the confusion around proper evaluation of LLM applications, and you have a recipe for confusion and frustration. Raza and the Humanloop team have been diving into these problems, and, in this episode, Raza helps us understand how non-technical prompt engineers can productively collaborate with technical software engineers while building AI-driven apps.

More episodes from Practical AI: Machine Learning, Data Science