This repo contains a PyTorch implementation of the pretrained BERT and XLNET model for multi-label text classification.
Structure of the code
At the root of the project, you will see:
├── pybert
| └── callback
| | └── lrscheduler.py
| | └── trainingmonitor.py
| | └── ...
| └── config
| | └── basic_config.py #a configuration file for storing model parameters
| └── dataset
| └── io
| | └── dataset.py
| | └── data_transformer.py
| └── model
| | └── nn
| | └── pretrain
| └── output #save the ouput of model
| └── preprocessing #text preprocessing
| └── train #used for training a model
| | └── trainer.py
| | └── ...
| └── common # a set of utility functions
├── run_bert.py
├── run_xlnet.py
Dependencies
- csv
- tqdm
- numpy
- pickle
- scikit-learn
- PyTorch 1.1+
- matplotlib
- pandas
- transformers=2.5.1
How to use the code
you need download pretrained bert model and xlnet model.
BERT: bert-base-uncased
XLNET: xlnet-base-cased
-
Download the Bert pretrained model from s3
-
Download the Bert config file from s3
-
Download the Bert vocab file from s3
-
Rename:
bert-base-uncased-pytorch_model.bintopytorch_model.binbert-base-uncased-config.jsontoconfig.jsonbert-base-uncased-vocab.txttobert_vocab.txt
-
Place
model,configandvocabfile into the/pybert/pretrain/bert/base-uncaseddirectory. -
pip install pytorch-transformersfrom github. -
Download kaggle data and place in
pybert/dataset.- you can modify the
io.task_data.pyto adapt your data.
- you can modify the
-
Modify configuration information in
pybert/configs/basic_config.py(the path of data,...). -
Run
python run_bert.py --do_datato preprocess data. -
Run
python run_bert.py --do_train --save_best --do_lower_caseto fine tuning bert model. -
Run
run_bert.py --do_test --do_lower_caseto predict new data.
training
[training] 8511/8511 [>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>] -0.8s/step- loss: 0.0640
training result:
[2019-01-14 04:01:05]: bert-multi-label trainer.py[line:176] INFO
Epoch: 2 - loss: 0.0338 - val_loss: 0.0373 - val_auc: 0.9922
training figure
result
---- train report every label ----- Label: toxic - auc: 0.9903 Label: severe_toxic - auc: 0.9913 Label: obscene - auc: 0.9951 Label: threat - auc: 0.9898 Label: insult - auc: 0.9911 Label: identity_hate - auc: 0.9910 ---- valid report every label ----- Label: toxic - auc: 0.9892 Label: severe_toxic - auc: 0.9911 Label: obscene - auc: 0.9945 Label: threat - auc: 0.9955 Label: insult - auc: 0.9903 Label: identity_hate - auc: 0.9927
Tips
- When converting the tensorflow checkpoint into the pytorch, it's expected to choice the "bert_model.ckpt", instead of "bert_model.ckpt.index", as the input file. Otherwise, you will see that the model can learn nothing and give almost same random outputs for any inputs. This means, in fact, you have not loaded the true ckpt for your model
- When using multiple GPUs, the non-tensor calculations, such as accuracy and f1_score, are not supported by DataParallel instance
- As recommanded by Jocob in his paper https://arxiv.org/pdf/1810.04805.pdf, in fine-tuning tasks, the hyperparameters are expected to set as following: Batch_size: 16 or 32, learning_rate: 5e-5 or 2e-5 or 3e-5, num_train_epoch: 3 or 4
- The pretrained model has a limit for the sentence of input that its length should is not larger than 512, the max position embedding dim. The data flows into the model as: Raw_data -> WordPieces -> Model. Note that the length of wordPieces is generally larger than that of raw_data, so a safe max length of raw_data is at ~128 - 256
- Upon testing, we found that fine-tuning all layers could get much better results than those of only fine-tuning the last classfier layer. The latter is actually a feature-based way
