Skip to content

MathiasVigouroux/Project_NLP_MVA_23

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Project NLP MVA 23 📚

Welcome to the repoitory of the final project for the Algorithms for Speech and Natural Language Processing class.

Datasets 📊

This section contains various datasets utilized for training our models.

Code 👩‍💻

Final report

Pattern Exploitation Training (PET) 🧠

Abstract

Our project is inspired by "It’s not just size that matters" by Schick and Schütze (2020), introducing Pattern Exploitation Training (PET). PET reframes tasks as language modeling problems, using fine-tuned language models to label unlabeled data. This enables classical classifiers with small training datasets and supports few-shot learning. Our project explores PET, replicates its results on various datasets, and compares different masked language models.

Introduction

Language models like GPT-3 and GPT-4 excel in natural language processing but require substantial computational resources. PET offers an efficient alternative by transforming tasks into language modeling challenges. It uses predefined patterns and verbalizers to convert features into sentences and labels into words. PET facilitates few-shot learning and works well with small training datasets, leveraging easily accessible unlabeled data.

Project Scope and Findings

Our project replicates Schick and Schütze's study and compares masked language models. While some SuperGLUE tasks showed lower accuracies, sentiment classification tasks performed well. We faced challenges such as labeling errors for specific tasks. Additionally, we created new tasks and explored MLM performance within a few-shot learning paradigm.

Conclusion

In conclusion, our project provides insights into the effectiveness of Pattern Exploitation Training (PET) for natural language tasks. Despite limitations due to hardware and time constraints, our findings contribute to the discussion on efficient language model training and real-world applications.

Links 🔗

Article 📄

Google Colab 🚀

Dataset MNLI 📦

List of all used datasets 📋

Important YouTube Videos 🎥

About

Repository for the Project of Algorithms for Speech and Natural Language Processing

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Contributors 2

  •  
  •