Skip to content

cofe-ai/EasyDistPLMs

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

9 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

EasyDistPLMs

Introduction

We provide a simple pytorch-based PLM finetuning example with well-formed structure. You can simply build and run your fine-tuning task with tiny modification of the code. Meanwhile, we provide different distributed-training approaches with little change of the original code. You can checkout deepspeed / horovod branch to try it yourself.

Requirements

conda create -n torch_env python=3.9 pandas tqdm scikit-learn -y
conda activate torch_env
conda install pytorch cudatoolkit=11.3.1 -y
pip install transformers wandb

Train

  1. Download transformers' pretrained model files (pytorch_model.bin, config.json, vocab.txt ...) and put them in one dir, eg. pretrained

  2. Customize a dataset in src/datasets.py. We provide IMDB and SNLI dataset as demos. Basically, for sent /sent-pair classification task, the only thing you need to do is to inherit SeqCLSDataset class and implement read_line / read_example according to your data format.

  3. Create labelspace file containing all labels, sep by line break

  4. Edit scripts/train.sh

  5. (optional) --use_wandb and set wandb_key to enable logging with wandb.ai

  6. Activate conda env and Run it!

    bash scripts/train.sh

Debug

Fairseq wraps a multiprocessing-supported pdb. You can use from debugger.mp_pdb import pdb;pdb.set_trace in our code to debug in real time. See common usage at https://docs.python.org/3/library/pdb.html

About

No description, website, or topics provided.

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published