How can you perform LLM evaluation in Snorkel Flow? LLM evaluation can mean the difference between a demonstration enterprise LLM application and one that makes it into production.
Snorkel AI software engineer Rebecca Westerlind walks the viewer through a summarized overview of the iterative loop at the core of Snorkel Flow AI data development workflow. She then demonstrates how this works practically within Snorkel Flow.
In this demo, we show you a step-by-step process for:
Building a robust evaluation framework for your LLM within Snorkel Flow.
Leveraging Snorkel's features to create high-quality training data.
Analyzing LLM performance across different metrics and data slices.
This video is perfect for anyone looking for:
LLM evaluation
Snorkel flow tutorial
Enterprise LLM deployment
This video is an excerpt from a longer webinar. See the entire webinar here: • How to Evaluate LLM Performance for Domain...
See more Snorkel Flow demos here: • Snorkel Flow Demos: See How it Works!
#largelanguagemodels #evaluation #enterpriseai