Skip to main content

Active Learning x LLMs

Supervisors

Suitable for

MSc in Advanced Computer Science

Abstract

We have a range of projects at the intersection of active learning/subset selection and large language models (LLMs). Active learning is a machine learning approach where the algorithm selectively queries a user or another information source to label new data points with the aim of reducing the amount of data that needs to be labelled for training a model. The motivation behind this approach lies in its efficiency and cost-effectiveness, especially in scenarios where labelling data or training is expensive or time-consuming. It is particularly relevant in the context of LLMs, where finetuning is computationally expensive. The intersection of active learning with LLMs opens new avenues for more efficient model training and fine-tuning, enabling these models to learn more effectively from smaller, well-chosen subsets of data. This area includes exploring how best to select these subsets, determining which examples are most beneficial for tasks like few-shot learning, and actively selecting demonstrations or test examples to optimize the model's performance and evaluation.

This research aims to improve the efficiency and effectiveness of large language models through advanced active learning and subset selection techniques. The key research question investigates how these methodologies can be best applied to reduce training resources while enhancing model performance in diverse NLP tasks. The projects target submission at the major ML conferences.

Goals:

To be determined based on the project chosen. These projects have been scoped and we will develop a full project plan with the student(s). An expected goal would to be produce research that is of publishable quality at a major ML conference.

References:

 

 Pre-requisites: Machine Learning and Deep Learning