(October 2022) TRIP is now available on HuggingFace datasets.
(May 2022) We recently discovered a small typo in the test results reported for RoBERTa in Table 3 of our Tiered Reasoning for Intuitive Physics paper in Findings of EMNLP 2021 (here and here). We have submitted corrections to ACL and arXiv, so please refer to the revised versions of the paper once they are processed. No other results are affected by this correction, nor are released code and models. Thank you to Oren Sultan for bringing this to our attention!
Shared repository for TRIP dataset for verifiable NLU and coherence measurement for text classifiers. Covers the following publications in Findings of EMNLP 2021:
- Shane Storks, Qiaozi Gao, Yichi Zhang, and Joyce Chai. (2021). Tiered Reasoning for Intuitive Physics: Toward Verifiable Commonsense Language Understanding. In Findings of EMNLP 2021.
- Shane Storks and Joyce Chai. (2021). Beyond the Tip of the Iceberg: Assessing Coherence of Text Classifiers. In Findings of EMNLP 2021.
Please contact Shane Storks with any questions.
Our results can be reproduced using the Python notebook file Verifiable-Coherent-NLU.ipynb, which we ran in Colab with Python 3.7 (may require some adaptation for use in Jupyter).
The required dependencies for Colab are installed within the notebook, while the exhaustive list of dependencies for any setup is given in requirements.txt. Out of these, the minimal requirements can be installed in a new Anaconda environment by the following commands:
conda create --name tripPy python=3.7
conda activate tripPy
pip install torch==1.7.1 torchvision==0.8.2 torchaudio==0.7.2
pip install transformers==4.2.2
pip install sentencepiece==0.1.96
pip install deberta==0.1.12
pip install spacy==3.2.0
python -m spacy download en_core_web_sm
pip install pandas==1.1.5
pip install matplotlib==3.5.0
pip install progressbar2==3.38.0
pip install ipykernel jupyter ipywidgets # For Jupyter Notebook setting
If I'm missing any, please let me know!
First clone the repo:
git clone https://github.com/sled-group/Verifiable-Coherent-NLU.git
You will then need to upload the contents of this folder to Google Drive.
From the Verifiable-Coherent-NLU
directory in your Google Drive, open Verifiable-Coherent-NLU.ipynb using Google Colab.
Configure the cells below the first heading of Verifiable-Coherent-NLU.ipynb as needed and run the Setup block to prepare the notebook for reproducing a specific set of results. Then navigate to the appropriate block to reproduce results on TRIP, Conversational Entailment, or ART. Each block will have sub-blocks for preparing the data (run every time), and for training and testing models.
You may either re-train the models from the papers, or use our pre-trained model instances (see below).
Pre-trained model instances from the papers are available here. Each sub-directory indicates a model and (if applicable) a loss function configuration, while the archive files within are for each type of LM trained, e.g., BERT, RoBERTa, or DeBERTa.
Copy the desired archive file(s) within these directories to your own Google Drive, and unzip them into a new directory ./saved_models
. Run inference on them as needed using the appropriate blocks in the notebook. The names of the provided pre-trained model directories are already listed in the configuration area for convenience.
If you use our code or models in your work, please cite one of our following papers from Findings of EMNLP 2021:
@misc{storks2021tiered,
title={Tiered Reasoning for Intuitive Physics: Toward Verifiable Commonsense Language Understanding},
author={Shane Storks and Qiaozi Gao and Yichi Zhang and Joyce Chai},
year={2021},
booktitle={Findings of the Association for Computational Linguistics: EMNLP 2021},
location={Punta Cana, Dominican Republic},
publisher={Association for Computational Linguistics},
}
@misc{storks2021tip,
title={Beyond the Tip of the Iceberg: Assessing Coherence of Text Classifiers},
author={Shane Storks and Joyce Chai},
year={2021},
booktitle={Findings of the Association for Computational Linguistics: EMNLP 2021},
location={Punta Cana, Dominican Republic},
publisher={Association for Computational Linguistics},
}
Additionally, please consider citing Conversational Entailment and ART, which are used in experiments from the latter paper (and included in this repo):
@inproceedings{zhang-chai-2010-towards,
title = "Towards Conversation Entailment: An Empirical Investigation",
author = "Zhang, Chen and
Chai, Joyce",
booktitle = "Proceedings of the 2010 Conference on Empirical Methods in Natural Language Processing",
month = oct,
year = "2010",
address = "Cambridge, MA",
publisher = "Association for Computational Linguistics",
url = "https://aclanthology.org/D10-1074",
pages = "756--766",
}
@inproceedings{
bhagavatula2020abductive,
title={Abductive Commonsense Reasoning},
author={Chandra Bhagavatula and Ronan Le Bras and Chaitanya Malaviya and Keisuke Sakaguchi and Ari Holtzman and Hannah Rashkin and Doug Downey and Wen-tau Yih and Yejin Choi},
booktitle={International Conference on Learning Representations},
year={2020},
url={https://openreview.net/forum?id=Byg1v1HKDB}
}