Collaboration & evaluation for LLM apps
Practical AI: Machine Learning, Data Science, LLM - En podcast av Changelog Media
![](https://is1-ssl.mzstatic.com/image/thumb/Podcasts123/v4/45/10/06/4510062f-d99e-abd5-7376-07a4656f19d1/mza_940882021610159734.png/300x300bb-75.jpg)
Kategorier:
Small changes in prompts can create large changes in the output behavior of generative AI models. Add to that the confusion around proper evaluation of LLM applications, and you have a recipe for confusion and frustration. Raza and the Humanloop team have been diving into these problems, and, in this episode, Raza helps us understand how non-technical prompt engineers can productively collaborate with technical software engineers while building AI-driven apps.