CS6217 Project --- Embedding-Boosted GLISTER (EB-GLISTER)
Our subset selection code is based on cords. We modified the glisterstrategy.py where we incorporated the gradients of word embeddings into the calculation of gain, and dataselectionstrategy.py where we initialized the gradients of word embeddings at the beginning of subset selection. To mitigate the heavy usage of GPU memory, we alternatively access and save the gradients on CPUs and GPUs.
PyTorch 1.11
Before training, please download GloVe weights and nltk_data directories from this link, and then move these two directories into EB-GLISTER directory.
Then, python train.py
.
Noted: run.sh provides all the scipts!