This repository provides the implementation for our paper "Improving Encoder by Auxiliary Supervision Tasks for Table-to-Text Generation" ACL2021 main conference. This code is based on https://github.com/ernestgong/data2text-three-dimensions/.
We provide the scrips (preprocess.sh, train.sh, and translate.sh) to preprocess the dataset, train models, and test. Please refer to these scripts for more details about parameters setting. The ouputs of our model are saved at ./our_results.
Please kindly cite this work if it helps your research:
@inproceedings{li-etal-2021-improving-encoder,
title = "Improving Encoder by Auxiliary Supervision Tasks for Table-to-Text Generation",
author = "Li, Liang and
Ma, Can and
Yue, Yinliang and
Hu, Dayong",
booktitle = "Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing (Volume 1: Long Papers)",
month = aug,
year = "2021",
address = "Online",
publisher = "Association for Computational Linguistics",
url = "https://aclanthology.org/2021.acl-long.466",
doi = "10.18653/v1/2021.acl-long.466",
pages = "5979--5989",
}
All dependencies can be installed via:
pip install -r 3.7.requirements.txt
Note that 3.7.requirements.txt contains unnecessary packages and Python version is 3.7.
Please refer to https://github.com/ernestgong/data2text-three-dimensions/ and https://github.com/wanghm92/rw_fg to obtain ROTOWIRE and RW-FG datasets. And we provide the necessary config files in ./dataset.
The following command will preprocess the data:
bash preprocess.sh
The following command will train the model:
bash train.sh
The following command will generate on development or test datasets given trained model:
bash translate.sh
To obtain extractive evaluation metrics, please refer to https://github.com/ratishsp/data2text-1 and https://github.com/wanghm92/rw_fg for details.
The following command will compute BLEU score:
perl ref/multi-bleu.perl ref/test.txt < ./our_results/model_pred_rw_test.txt