Large language models (LLMs) have achieved impressive performance in many domains, including code generation and reasoning. However, to accomplish challenging tasks, generating the correct solution in one go becomes challenging. In this talk, I will first discuss our work self-debugging, which instructs LLMs to debug their own predicted programs. In particular, we demonstrate that self-debugging can teach LLMs to perform rubber duck debugging; i.e., without any human feedback on the code correctness or error messages, the model is able to identify its mistakes by investigating the execution results and explaining the generated code in natural language. Self-debugging notably improves both the model performance and sample efficiency, matching or outperforming baselines that generate more than 10× candidate programs. In the second part, I will further demonstrate that LLMs can also improve their own prompts to achieve better performance, acting as optimizers.
talk-data.com
Topic
large language models (llms)
1
tagged
Activity Trend
1
peak/qtr
2020-Q1
2026-Q1
Top Events
Beyond Boundaries: AI, GenAI, and LLMs Insights
2
Retrieval, search and knowledge in the age of LLM and Vector Databases
2
London Reactor Meetup I Microsoft Cloud meets AI
1
Neo4j Live: HybridAGI – Graph-Powered, Self-Programmable AI
1
London Seminar: Beyond LLM: GenAI for Trading and Asset Management
1
GenAI for SW developers - V.2! 1# -Measuring Mastery Assessing Large Language Mo
1
Quality Engineering meetup #9
1
PyData Leeds: Leeds Digital Fest '25
1
AI Meetup (November): GenAI LLMs and Agents
1
#21 AI Series: University of Oxford - Dr. A. Bibi
1
Advanced RAG Chatbot Assistant for Healthcare Patient Records
1
AI Seminars (Virtual): Self-Improvement with LLMs by Google DeepMind
1
Filtering by:
AI Seminars (Virtual): Self-Improvement with LLMs by Google DeepMind
×