CLIP is a foundational model with transferable classification performance in the few-shot setting. Several methods have shown improved performance of CLIP using few-shot examples. However, so far, all these techniques have been benchmarked using standard few-shot datasets. We argue that this mode of evaluation does not provide a true indication of the inductive generalization ability using few-shot examples. As most datasets have been seen by the CLIP model, the resultant setting can be termed as partially transductive. To solve this, we propose a pipeline that uses an unlearning technique to obtain true inductive baselines. In this new inductive setting, the methods show a significant drop in performance (-55% on average among 13 baselines with multiple datasets). We validate the unlearning technique using oracle baselines. An improved few-shot classification technique is proposed that consistently obtains state-of-the-art performance over 13 other recent baseline methods on a comprehensive analysis with 5880 experiments - varying the datasets, differing number of few-shot examples, unlearning setting, and with different seeds. Thus, we identify the issue with the evaluation of CLIP-based few-shot classification, provide a solution using unlearning, propose new benchmarks, and provide an improved method.
talk-data.com
A
Speaker
Alexey Kravets
1
talks
PhD student in AI
University of Bath
Alexey Kravets is a PhD student in AI at the University of Bath and a former Lead Data Scientist at Aviva, a UK FTSE 100 insurer, with over five years of experience in machine learning. His research focuses on vision and language models, few-shot learning, machine unlearning, and mechanistic interpretability. Prior to his PhD, he led ML projects at Aviva that included developing NLP tools for insurance predictions, and he regularly shares insights through Medium articles.
Bio from: Nov 24 - Best of ICCV (Day 4)
Filtering by:
Nov 24 - Best of ICCV (Day 4)
×
Filter by Event / Source
Talks & appearances
Showing 1 of 7 activities