Skip to content

Source code for our pilot study of controllable story completion task, aiming emotion-aware story writing assistance.

License

Notifications You must be signed in to change notification settings

mil-tokyo/controllable-story-completion-pilot-study

Repository files navigation

Plug-and-Play Controller for Story Completion (Pilot Study)

This repository is for our work presented at In2Writing (The First Workshop on Intelligent and Interactive Writing Assistants).

Overview

This folder contains our modified version of the Plug and Play Language Model (PPLM) codes and document (this "README.md"). Our implementation is based on the PPLM codes distributed in https://github.com/huggingface/transformers and in https://github.com/uber-research/PPLM, under Apache License 2.0.

Note: in the Transformers library, the PPLM codes were placed in https://github.com/huggingface/transformers/tree/master/examples/text-generation/pplm previously. Now it is replaced in https://github.com/huggingface/transformers/tree/main/examples/research_projects.

Correction of the Paper

Lack of footnote 10

We had intended to add the footnote to explain the strange time consuming of XLM-ProphetNet large. The footnote number "10" was displayed (in Table 4), but due to our misunderstanding of the tex specification, the footnote text was not displayed. We show the footnote text below.

\footnote{The runtime of ``XLM-ProphetNet large'' was very long compared to the others, probably due to failure in allocating GPU memory. It is not considered to have any effect on the generated statements, but is included as is for reference.}

Model name misspelled

We mistakenly wrote PEGASUS large in Tables 3 and 4 as 20220410_003_pegasus_large, which includes the ID for managing experiments. They should correctly be written as PEGASUS large, as with the other models.

How to use our modified PPLM

Seq2SeqLM Story Completion

In seq2seqlm_storycompletion directory,

python storycompletion_finetune_trainer.py \
    --data_dir ../data/ \
    --learning_rate=3e-5 \
    --do_train --do_eval \
    --predict_with_generate \
    --logging_steps 3000 \
    --save_steps 3000 \
    --model_name_or_path facebook/bart-base \
    --save_total_limit 3 \
    --use_task_specific_params \
    --output_dir /path/to/output/seq2seqlm/

BART + PPLM-BoW for controlling emotions

In modified_pplm directory,

python run_pplm_various_lm.py \
     -M /path/to/output/seq2seqlm/ \
     --length 40 \
     --gamma 1.0 \
     --num_iterations 3 \
     --num_samples 3 \
     --stepsize 0.015 \
     --window_length 5 \
     --kl_scale 0.30 \
     --gm_scale 0.95 \
     --colorama \
     --sample \
     --use_NRC_VAD_lexicon \
     --V_max 0.3 --A_min 0.7 \
     --output_dir /path/to/output/pplm/

Note

The results presented in our paper were obtained using version 4.8.1 of HuggingFace Transformers in Seq2SeqLM finetuning. We use one NVIDIA V100 GPU for each training and each inference.

Citation

If you find this repository helpful for your work, please consider citing the related paper.

Yusuke Mori, Hiroaki Yamane, Ryohei Shimizu, Tatsuya Harada, “Plug-and-Play Controller for Story Completion: A Pilot Study toward Emotion-aware Story Writing Assistance,” The First Workshop on Intelligent and Interactive Writing Assistants (In2Writing) (ACL 2022, Workshop), 2022.

Please visit the ACL Anthology to get Bibtex file.

Original PPLM

We also recommend you to cite (Dathathri et al., 2020), because our work is based on it.

Sumanth Dathathri, Andrea Madotto, Janice Lan, Jane Hung, Eric Frank, Piero Molino, Jason Yosinski, and Rosanne Liu. (2020). Plug and Play Language Models: a Simple Approach to Controlled Text Generation. International Conference on Learning Representations (ICLR) 2020. link to ICLR page

About

Source code for our pilot study of controllable story completion task, aiming emotion-aware story writing assistance.

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages