Summary
The document compares the task adaptation strategies of few-shot fine-tuning and in-context learning for pre-trained language models. It evaluates their in-domain and out-of-domain generalization using models of various sizes. The results show that both approaches exhibit strengths and limitations, with fine-tuned models demonstrating good OOD generalization as model size increases. The study emphasizes the importance of fair comparisons between adaptation methods using models of equal size.