by , , , ,
Abstract:
We study the active few-shot fine-tuning of large neural networks to downstream tasks. We show that few-shot fine-tuning is an instance of a generalization of classical active learning, transductive active learning, and we propose ITL, short for information-based transductive learning, an approach which samples adaptively to maximize the information gained about specified downstream tasks. Under general regularity assumptions, we prove that ITL converges uniformly to the smallest possible uncertainty obtainable from the accessible data. To the best of our knowledge, we are the first to derive generalization bounds of this kind, and they may be of independent interest for active learning. We apply ITL to the few-shot fine-tuning of large neural networks and show that ITL substantially improves upon the state-of-the-art.
Reference:
Active Few-Shot Fine-Tuning J. Hübotter, B. Sukhija, L. Treven, Y. As, A. KrauseIn ICLR 2024 Workshop on Bridging the Gap Between Practice and Theory in Deep Learning, 2024
Bibtex Entry:
@inproceedings{hubotter2024active,
  title={Active Few-Shot Fine-Tuning},
  author={H{\"u}botter, Jonas and Sukhija, Bhavya and Treven, Lenart and As, Yarden and Krause, Andreas},
  booktitle = {ICLR 2024 Workshop on Bridging the Gap Between Practice and Theory in Deep Learning},
  year={2024},
  pdf = {https://arxiv.org/pdf/2402.15441.pdf}  
}