Series “Evaluate LLM-powered Products” EP2!
In this episode, I share what “accuracy” really means when it comes to LLMs and AI-powered products. We explore why traditional metrics like BLEU and ROUGE often fall short, how LLM-as-a-judge methods work, and why multi-turn conversations are especially tricky to evaluate. I also share practical tips, rubrics, and personal lessons learned from my own experiments.
Subscribe "Data Science x AI" newsletter to get updates!

