Ever wondered what actually happens when you call an LLM API? This talk breaks down the inference pipeline from tokenization to text generation, explaining what's really going on under the hood. He will walk through the key sampling strategies and their parameters - temperature, top-p, top-k, beam search. We'll also cover performance tricks like quantization, KV caching, and prompt caching that can speed things up significantly. If time allows, we will also touch on some use-case-specific techniques like pass@k and majority voting.
talk-data.com
Topic
tokenization
1
tagged
Activity Trend
6
peak/qtr
2020-Q1
2026-Q1
Top Events
Stablecoins as a Payment Rail
1
PyData Trójmiasto #37
1
Fabric-X – Programming Model and Application Development Deep Dive
1
Démystifier les LLM avec NanoChat de Andrej Karpathy
1
Fabric-X – Programming Model and Application Development Deep Dive
1
Fabric-X – Programming Model and Application Development Deep Dive
1
Fabric-X – Programming Model and Application Development Deep Dive
1
Filtering by:
Dariusz Piotrowski
×