0

Introducing Align Evals: Streamlining LLM Application Evaluation

https://blog.langchain.com/introducing-align-evals/(blog.langchain.com)
Align Evals is a new feature in LangSmith designed to calibrate LLM evaluators to better match human preferences. It addresses the challenge of evaluation scores not aligning with human expectations, leading to noisy comparisons and wasted time. The feature provides a playground-like interface for iterating on evaluator prompts, side-by-side comparison of human-graded data and LLM-generated scores, and a saved baseline alignment score for comparing changes. The alignment flow involves selecting evaluation criteria, selecting data for human review, grading the data, and creating/testing an evaluator prompt against human grading.
0 pointsby ogg3 months ago

Comments (0)

No comments yet. Be the first to comment!

Want to join the discussion?