Abdullatif Köksal
PhD
Ludwig Maximilian University of Munich (LMU Munich)
Few-shot Learning with Pretrained Language Models

Large pretrained language models (PLMs) perform well on many NLP tasks without supervised training. This phenomenon is best demonstrated in GPT-3 by its translation capability from different languages to English with just 64 contextual examples. On the other hand, contextual examples enable PLMs to work with multiple tasks without parameter updates. Considering these benefits, we will analyze the few-shot learning capabilities of PLMs in a wide variety of NLP tasks. We will systematically explore what factors contribute to the good performance of PLMs with contextual examples. We will adapt and propose active learning selection strategies to check which samples help the model perform better. Furthermore, we will compare in-context learning and prompt finetuning paradigms with few-shot examples.

Track:
Academic Track
PhD Duration:
January 17th, 2022 - January 1st, 2025
First Exchange:
January 1st, 2024 - July 1st, 2024
ELLIS Edge Newsletter
Join the 6,000+ people who get the monthly newsletter filled with the latest news, jobs, events and insights from the ELLIS Network.