Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

End-to-end Story Plot Generator (2310.08796v1)

Published 13 Oct 2023 in cs.CL

Abstract: Story plots, while short, carry most of the essential information of a full story that may contain tens of thousands of words. We study the problem of automatic generation of story plots, which includes story premise, character descriptions, plot outlines, etc. To generate a single engaging plot, existing plot generators (e.g., DOC (Yang et al., 2022a)) require hundreds to thousands of calls to LLMs (e.g., OpenAI API) in the planning stage of the story plot, which is costly and takes at least several minutes. Moreover, the hard-wired nature of the method makes the pipeline non-differentiable, blocking fast specialization and personalization of the plot generator. In this paper, we propose three models, $\texttt{OpenPlot}$, $\texttt{E2EPlot}$ and $\texttt{RLPlot}$, to address these challenges. $\texttt{OpenPlot}$ replaces expensive OpenAI API calls with LLaMA2 (Touvron et al., 2023) calls via careful prompt designs, which leads to inexpensive generation of high-quality training datasets of story plots. We then train an end-to-end story plot generator, $\texttt{E2EPlot}$, by supervised fine-tuning (SFT) using approximately 13000 story plots generated by $\texttt{OpenPlot}$. $\texttt{E2EPlot}$ generates story plots of comparable quality to $\texttt{OpenPlot}$, and is > 10$\times$ faster (1k tokens in only 30 seconds on average). Finally, we obtain $\texttt{RLPlot}$ that is further fine-tuned with RLHF on several different reward models for different aspects of story quality, which yields 60.0$\%$ winning rate against $\texttt{E2EPlot}$ along the aspect of suspense and surprise.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (54)
  1. Language models are few-shot learners. Advances in neural information processing systems, 33:1877–1901, 2020.
  2. Eugene Charniak. Toward a model of children’s story comprehension. PhD thesis, Massachusetts Institute of Technology, 1972.
  3. Extending context window of large language models via positional interpolation, 2023a.
  4. Longlora: Efficient fine-tuning of long-context large language models. arXiv preprint arXiv:2309.12307, 2023b.
  5. Generating long sequences with sparse transformers, 2019.
  6. Rethinking attention with performers, 2022.
  7. Talebrush: Sketching stories with generative pretrained language models. In Proceedings of the 2022 CHI Conference on Human Factors in Computing Systems, pp.  1–19, 2022.
  8. Wordcraft: a human-ai collaborative editor for story writing. arXiv preprint arXiv:2107.07430, 2021.
  9. Hierarchical neural story generation, 2018.
  10. Strategies for structuring story generation. arXiv preprint arXiv:1902.01109, 2019.
  11. Plan, write, and revise: an interactive system for open-domain story generation. arXiv preprint arXiv:1904.02357, 2019.
  12. Content planning for neural story generation with aristotelian rescoring. arXiv preprint arXiv:2009.09870, 2020.
  13. Lm-infinite: Simple on-the-fly length generalization for large language models. arXiv preprint arXiv:2308.16137, 2023.
  14. Go back in time: Generating flashbacks in stories with event temporal prompts. arXiv preprint arXiv:2205.01898, 2022.
  15. Transformer language models without positional encodings still learn positional information, 2022.
  16. Unsupervised hierarchical story infilling. In Proceedings of the First Workshop on Narrative Understanding, pp.  37–43, 2019.
  17. Creative writing with an ai-powered writing assistant: Perspectives from professional writers. arXiv preprint arXiv:2211.05030, 2022.
  18. Reformer: The efficient transformer, 2020.
  19. Large language models are zero-shot reasoners. Advances in neural information processing systems, 35:22199–22213, 2022.
  20. Openassistant conversations – democratizing large language model alignment, 2023.
  21. Dialogue state tracking with a language model using schema-driven prompting. arXiv preprint arXiv:2109.07506, 2021.
  22. Coauthor: Designing a human-ai collaborative writing dataset for exploring language model capabilities. In Proceedings of the 2022 CHI conference on human factors in computing systems, pp.  1–19, 2022.
  23. Story generation with crowdsourced plot graphs. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 27, pp.  598–604, 2013.
  24. Plug-and-blend: a framework for plug-and-play controllable story generation with sketches. In Proceedings of the AAAI Conference on Artificial Intelligence and Interactive Digital Entertainment, volume 17, pp.  58–65, 2021.
  25. Pre-train, prompt, and predict: A systematic survey of prompting methods in natural language processing. ACM Computing Surveys, 55(9):1–35, 2023.
  26. Language as a latent variable: Discrete generative models for sentence compression. arXiv preprint arXiv:1609.07317, 2016.
  27. Co-writing screenplays and theatre scripts with language models: Evaluation by industry professionals. In Proceedings of the 2023 CHI Conference on Human Factors in Computing Systems, pp.  1–34, 2023.
  28. OpenAI. Gpt-4 technical report, 2023.
  29. Training language models to follow instructions with human feedback. Advances in Neural Information Processing Systems, 35:27730–27744, 2022.
  30. Towards controllable story generation. In Proceedings of the First Workshop on Storytelling, pp. 43–49, 2018.
  31. Train short, test long: Attention with linear biases enables input length extrapolation, 2022.
  32. Counterfactual story reasoning and generation. In Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP), pp.  5043–5053, Hong Kong, China, November 2019. Association for Computational Linguistics. doi: 10.18653/v1/D19-1509. URL https://aclanthology.org/D19-1509.
  33. Plotmachines: Outline-conditioned generation with dynamic plot state tracking. arXiv preprint arXiv:2004.14967, 2020.
  34. Multitask prompted training enables zero-shot task generalization. arXiv preprint arXiv:2110.08207, 2021.
  35. Elvis Saravia. Prompt Engineering Guide. https://github.com/dair-ai/Prompt-Engineering-Guide, 12 2022.
  36. Summarize, outline, and elaborate: Long-text generation via hierarchical supervision from extractive summaries. In Proceedings of the 29th International Conference on Computational Linguistics, pp.  6392–6402, Gyeongju, Republic of Korea, October 2022a. International Committee on Computational Linguistics. URL https://aclanthology.org/2022.coling-1.556.
  37. A length-extrapolatable transformer, 2022b.
  38. Progressive generation of long text with pretrained language models, 2021.
  39. Zero-shot sonnet generation with discourse-level planning and aesthetics features. arXiv preprint arXiv:2205.01821, 2022.
  40. Llama 2: Open foundation and fine-tuned chat models. arXiv preprint arXiv:2307.09288, 2023.
  41. Scott R Turner. The creative process: A computer model of storytelling and creativity. Psychology Press, 2014.
  42. Language modeling via stochastic processes. arXiv preprint arXiv:2203.11370, 2022.
  43. T-cvae: Transformer-based conditioned variational autoencoder for story completion. In IJCAI, pp.  5233–5239, 2019.
  44. Autoformalization with large language models. Advances in Neural Information Processing Systems, 35:32353–32368, 2022.
  45. Efficient streaming language models with attention sinks. arXiv preprint arXiv:2309.17453, 2023.
  46. Megatron-cntrl: Controllable story generation with external knowledge using large-scale language models. arXiv preprint arXiv:2010.00840, 2020.
  47. Fudge: Controlled text generation with future discriminators. arXiv preprint arXiv:2104.05218, 2021.
  48. Doc: Improving long story coherence with detailed outline control. arXiv preprint arXiv:2212.10077, 2022a.
  49. Re3: Generating longer stories with recursive reprompting and revision. arXiv preprint arXiv:2210.06774, 2022b.
  50. Plan-and-write: Towards better automatic storytelling, 2019.
  51. H _⁢2_2\_2_ 2 o: Heavy-hitter oracle for efficient generative inference of large language models. arXiv preprint arXiv:2306.14048, 2023.
  52. Judging llm-as-a-judge with mt-bench and chatbot arena, 2023.
  53. Adapting language models for zero-shot learning by meta-tuning on dataset and prompt collections. arXiv preprint arXiv:2104.04670, 2021.
  54. Recurrentgpt: Interactive generation of (arbitrarily) long text. arXiv preprint arXiv:2305.13304, 2023.
User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (7)
  1. Hanlin Zhu (20 papers)
  2. Andrew Cohen (24 papers)
  3. Danqing Wang (37 papers)
  4. Kevin Yang (45 papers)
  5. Xiaomeng Yang (21 papers)
  6. Jiantao Jiao (83 papers)
  7. Yuandong Tian (128 papers)
Citations (4)

Summary

We haven't generated a summary for this paper yet.