Skip to content

In this project, we propose to deal with the data scarcity problem in a specific NLP task by harnessing existing annotated datasets from related tasks. Our approach involves training a multi-head architecture concurrently on both the main task and these “supporting” tasks. We experimented this approach on medical NLP tasks and on three GLUE tasks.

Notifications You must be signed in to change notification settings

OdedMous/Enhancing-By-Supporting-tasks-Components

 
 

Repository files navigation

Enhancing-By-Subtasks-Components

This project presents an experimental approach to tackle the challenge of data scarcity in a specific NLP task by utilizing existing annotated datasets from related tasks. Our experiment involves training a single base model, such as BERT, with multiple heads that are each dedicated to a specific task, and running them simultaneously during training. We term these additional tasks as "supporting tasks." The goal is to leverage shared knowledge across different domains and by that enhance the model's performance.

See the full report in the following pdf: Advanced_NLP_Project.pdf

Branches:

  • Medical tasks can be found in the main branch.
  • The GLUE (General Language Understanding Evaluation) tasks can be found in the glue_tasks branch.

The multi-head model can be viewed in models/multiHeadModel.py
The multi-head training can be viewed at train.py

Multi-Head Model Architecture

Advanced NLP Project

Install

pip install -r requirements.txt

Train

Run:

python train.py --batch_size <batch size> --epochs <number of epochs> --device <device>

For the rest of the arguments, please see train.py

About

In this project, we propose to deal with the data scarcity problem in a specific NLP task by harnessing existing annotated datasets from related tasks. Our approach involves training a multi-head architecture concurrently on both the main task and these “supporting” tasks. We experimented this approach on medical NLP tasks and on three GLUE tasks.

Topics

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages

  • Python 100.0%