LLMs have opened up new avenues in NLP with their possible applications, but evaluating their output introduces a new set of challenges. In this talk, we discuss how the evaluation of LLMs differs from the evaluation of classic ML-based solutions and how we tackle the challenges.
talk-data.com
Speaker
Lena Nahorna
3
talks
Lena Nahorna—analytical linguist at Grammarly. Works on correctness, responsible AI, and UA-GEC. Background: PhD in linguistics.
Bio from: Ensuring the Quality of LLM Output at Grammarly: An Overview and Case Study
Filter by Event / Source
Talks & appearances
3 activities · Newest first
LLMs have opened up new avenues in NLP with their possible applications, but evaluating their output introduces a new set of challenges. In this talk, we discuss these challenges and our approaches to measuring the model output quality. We will talk about the existing evaluation methods and their pros and cons and then take a closer look at their application in a practical case study.
This talk will help you understand the main responsibilities of analytical and computational linguists at the company, the types of tasks and projects they work on, and how they collaborate with the project teams. You will learn what kind of linguistic expertise is required for building AI-powered solutions at Grammarly.