LLMs have opened up new avenues in NLP with their possible applications, but evaluating their output introduces a new set of challenges. In this talk, we discuss these challenges and our approaches to measuring the model output quality. We will talk about the existing evaluation methods and their pros and cons and then take a closer look at their application in a practical case study.
talk-data.com
L
Speaker
Lena Nahorna
1
talks
Analytical Linguist
Grammarly
Lena Nahorna—analytical linguist at Grammarly. Works on correctness, responsible AI, and UA-GEC. Background: PhD in linguistics.
Bio from: The Depth and Breadth of Language Research and Engineering at Grammarly
Filtering by:
Ensuring the Quality of LLM Output at Grammarly: An Overview and Case Study
×
Filter by Event / Source
Talks & appearances
Showing 1 of 3 activities