Skip to content

Implementation of the paper “Holistic Exploration on Universal Decompositional Semantic Parsing: Architecture, Data Augmentation, and LLM Paradigm,” accepted to ACL 2024 Workshop. 🎉

License

Notifications You must be signed in to change notification settings

hexuandeng/HExp4UDS

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

10 Commits
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Holistic Exploration on UDS Parsing

Download Corresponding Datasets

mkdir datasets
cd datasets
wget 'https://data.statmt.org/news-crawl/en/news.2021.en.shuffled.deduped.gz'
wget 'https://nlp.stanford.edu/data/glove.840B.300d.zip'
unzip glove.840B.300d.zip
gzip -d news.2021.en.shuffled.deduped.gz

Environment

First install PredPatt and decomp following the instructions in the link, then run:

pip install -r requirements.txt

Experiment

For naive model training, run:

python heuds/main.py train --task UDSTask --arch Bert_UDS --save-dir 'Bert_naive' --encoder-output-dim 1024 --layer-in-use 0,0,1,1,1,1,1

For model training with additional syntactic information, run:

python heuds/main.py train --task UDSTask --arch Bert_UDS --save-dir 'Bert_incorpsyn' --encoder-output-dim 1024 --contact-ud --syntax-edge-gcn

For our best model training with additional syntactic information and data augmentation method, run:

python heuds/main.py train --task UDSTask --arch Bert_Syntactic --save-dir 'Bert_syntactic' --encoder-output-dim 1024
python heuds/main.py generate --task ConlluTask --arch Bert_Syntactic --save-dir 'Bert_syntactic' --encoder-output-dim 1024 --mono-file datasets/news.2021.en.shuffled.deduped --conllu-file datasets/news.conllu
python heuds/main.py train --task PredPattTask --arch Bert_UDS --save-dir 'Bert_best_pretrained' --max-epoch 30 --encoder-output-dim 1024 --layer-in-use 1,1,1,1,1,0,0 --conllu datasets/news.conllu --name news --validate-interval -1 --contact-ud --syntax-edge-gcn
python heuds/main.py train --task UDSTask --arch Bert_UDS --save-dir 'Bert_best' --pretrained-model-dir 'Bert_best_pretrained' --encoder-output-dim 1024 --lr 2e-5 --pretrained-lr 1e-6 --contact-ud --syntax-edge-gcn

Replace "train" to "test" for model evaluation.

Citation

If you find this work helpful, please consider citing as follows:

@inproceedings{deng-etal-2024-holistic,
    title = "Holistic Exploration on Universal Decompositional Semantic Parsing: Architecture, Data Augmentation, and {LLM} Paradigm",
    author = "Deng, Hexuan  and
      Zhang, Xin  and
      Zhang, Meishan  and
      Liu, Xuebo  and
      Zhang, Min",
    editor = "Wong, Kam-Fai  and
      Zhang, Min  and
      Xu, Ruifeng  and
      Li, Jing  and
      Wei, Zhongyu  and
      Gui, Lin  and
      Liang, Bin  and
      Zhao, Runcong",
    booktitle = "Proceedings of the 10th SIGHAN Workshop on Chinese Language Processing (SIGHAN-10)",
    month = aug,
    year = "2024",
    address = "Bangkok, Thailand",
    publisher = "Association for Computational Linguistics",
    url = "https://aclanthology.org/2024.sighan-1.6",
    pages = "45--57"
}

About

Implementation of the paper “Holistic Exploration on Universal Decompositional Semantic Parsing: Architecture, Data Augmentation, and LLM Paradigm,” accepted to ACL 2024 Workshop. 🎉

Topics

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages