Skip to content

adihaviv/bertese

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

11 Commits
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

BERTese

This repository contains the code for the models discussed in the paper "BERTese: Learning to speak to BERT".

Our code, both pretraining and training, is based on PyTorch (1.4.0) and Transformers (2.5.0). Note that all the dependencies and requirement file are provided in and requirements.txt.

Data

Downloading the LAMA Dataset provided by Petroni et al. 2019

from https://github.com/facebookresearch/LAMA Download: 
curl -L https://dl.fbaipublicfiles.com/LAMA/data.zip

Note we are looking at the T-REx subset.

Downloading the Training set provided by Jiang et al. 2020

from https://github.com/jzbjyb/LPAQA Download: 
curl -L https://github.com/jzbjyb/LPAQA/blob/master/TREx_train.tar.gz

Command for pretraining

After downloading the dataset, you should first run the pretraining model by using the once the model is trained

python seq2seq_experiment.py \ 
    --output_dir <OUT_DIR> \
    --model_type bert_emb_identity_seq2seq \
    --model_name bert-large-uncased  \  
    --num_train_epochs 100 \  
    --evaluate_during_training \  
    --do_eval_test
Command for BERTese
python bertese_experiment.py \
    --log_examples 
    --evaluate_during_training 
    --model_type bertese 
    --train_batch_size 64 
    --max_seq_length 20 
    --do_train 
    --explicit_mask_loss_weight 0 
    --optimize_mask_softmin 
    --lpaqa 
    --num_train_epochs 20 
    --evaluate_during_training 
    --do_eval_dev 

To train with automatic mixed-precision, install apex and add the --fp16 flag.

Citation

If you find this work helpful, please cite us

    title = "{BERT}ese: Learning to Speak to {BERT}",
    author = "Haviv, Adi  and
      Berant, Jonathan  and
      Globerson, Amir",
    booktitle = "Proceedings of the 16th Conference of the European Chapter of the Association for Computational Linguistics: Main Volume",
    month = apr,
    year = "2021",
    address = "Online",
    publisher = "Association for Computational Linguistics",
    url = "https://aclanthology.org/2021.eacl-main.316",
    doi = "10.18653/v1/2021.eacl-main.316",
    pages = "3618--3623",
}

Acknowledgements

We would like to thank the European Research Council (ERC) for funding the project.

This code is still improving. for any questions, please email adi.haviv@cs.tau.ac.il

About

No description, website, or topics provided.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages