Offical codebase for Preserving Pre-trained Features Helps Calibrate Fine-tuned Language Models.
# (Optional) create a virtual environment
conda create -n LM-Calibration python=3.8
conda activate LM-Calibration
cd LM-Calibration
git clone [email protected]:rubbybbs/OpenDelta.git
# Install torch (example of cu113)
pip install torch==1.11.0+cu113 torchvision==0.12.0+cu113 torchaudio==0.11.0 --extra-index-url https://download.pytorch.org/whl/cu113
# Install OpenDelta
cd OpenDelta
python setup.py develop
# Install other dependencies
cd ..
pip install -r requirements.txt
You may need to run accelerate config
to config accelerate too.
We use the same dataset and split as Desai & Durrett (2020). You can download and preprocess it using:
wget "https://cloud.tsinghua.edu.cn/f/aeac582da1d540f7afac/?dl=1" -O calibration_data.tar.gz
tar zxvf calibration_data.tar.gz -C ./data
sh scripts/preprocess_dataset.sh
See bash files on the ./scripts
folder.
See ./scripts/calibration.sh
, you may need assign the path of the output file produced by training scripts.
@inproceedings{
he2023preserving,
title={Preserving Pre-trained Features Helps Calibrate Fine-tuned Language Models},
author={Guande He and Jianfei Chen and Jun Zhu},
booktitle={The Eleventh International Conference on Learning Representations },
year={2023},
url={https://openreview.net/forum?id=NI7StoWHJPT}
}
Thanks to the authors of these repositories for the reference implementations of different fine-tuning and evaluation methods.
- https://github.com/shreydesai/calibration
- https://github.com/thunlp/OpenDelta
- https://github.com/alibaba/AliceMind/tree/main/ChildTuning
- https://github.com/bloodwass/mixout
- https://github.com/asappresearch/revisit-bert-finetuning
- https://github.com/Sanyuan-Chen/RecAdam
- https://github.com/facebookresearch/fairseq/tree/main/examples/rxf