Large language models can 'hallucinate' factually incorrect outputs, presenting significant risks for their adoption to high-stakes applications. Jannik will present joint work recently published in Nature on detecting hallucinations in large language models using semantic entropy, which mitigates hallucinations by quantifying the model's own uncertainty over the meaning of generations. He will also discuss a recent pre-print that proposes a method to drastically reduce the cost of uncertainty quantification in LLMs by predicting semantic entropy from latent space, and he may ramble about uncertainties in LLMs more generally.
talk-data.com
J
Speaker
Jannik Kossen
1
talks
Research Scientist
Meta AI
AI research scientist at Meta FAIR, building LLMs for code generation. He has studied Physics in Bremen and Heidelberg, interned at Google and DeepMind, and has been a PhD candidate at the University of Oxford.
Bio from: #22 AI Series: Meta AI - J. Kossen
Filtering by:
#22 AI Series: Meta AI - J. Kossen
×
Filter by Event / Source
Talks & appearances
Showing 1 of 1 activities