pip install transformers==4.5.1
export task=QYTM
export model_name=RE2
CUDA_VISIBLE_DEVICES=1 python src/DL_model/${model_name}/main.py \
--task ${task} \
--model_name ${model_name} \
--output_dir ./output/${task}_${model_name}_20211007 \
--tensorboardx_path ./output/logs/runs/${task}_${model_name}_20211007 \
--train_path ./datasets/${task}/train.csv \
--dev_path ./datasets/${task}/dev.csv \
--test_path ./datasets/${task}/test.csv \
--label_file_level_dir ./datasets/${task}/labels_level.txt \
--label2freq_level_dir ./datasets/${task}/label2freq_level.json \
--vocab_file ./resources/word2vec/vocab.txt \
--w2v_file ./resources/word2vec/token_vec_300.bin \
--train_batch_size 256 \
--eval_batch_siz 256 \
--do_train \
--do_eval \
--do_predict \
--num_train_epochs 50
model |
LCQMC(f1-score) |
DSSM |
0.7114 |
ABCNN |
0.7743 |
BIMPM |
0.8590 |
DecomposableAttention |
0.7526 |
ESIM |
0.8573 |
RE2 |
0.8182 |
SiaGRU |
0.8302 |
Bert |
0.8817 |
RoBerta |
0.8897 |
XlNet |
0.7828 |
ELECTRA |
1 |
DistilBert |
?? |
AlBert |
?? |
NEZHA |
1 |