Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Monte Carlo Tree Search for Recipe Generation using GPT-2 (2401.05199v1)

Published 10 Jan 2024 in cs.CL and cs.AI

Abstract: Automatic food recipe generation methods provide a creative tool for chefs to explore and to create new, and interesting culinary delights. Given the recent success of LLMs, they have the potential to create new recipes that can meet individual preferences, dietary constraints, and adapt to what is in your refrigerator. Existing research on using LLMs to generate recipes has shown that LLMs can be finetuned to generate realistic-sounding recipes. However, on close examination, these generated recipes often fail to meet basic requirements like including chicken as an ingredient in chicken dishes. In this paper, we propose RecipeMC, a text generation method using GPT-2 that relies on Monte Carlo Tree Search (MCTS). RecipeMC allows us to define reward functions to put soft constraints on text generation and thus improve the credibility of the generated recipes. Our results show that human evaluators prefer recipes generated with RecipeMC more often than recipes generated with other baseline methods when compared with real recipes.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (25)
  1. 2022. Assistive Recipe Editing through Critiquing. arXiv:2205.02454.
  2. 2020. RecipeNLG: A Cooking Recipes Dataset for Semi-Structured Text Generation. In Proceedings of the 13th International Conference on Natural Language Generation, 22–28.
  3. 2022. PPL-MCTS: Constrained Textual Generation Through Discriminator-Guided MCTS Decoding. In Proceedings of the 2022 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, pages 2953 - 2967, 2953–2967.
  4. 2018. Hierarchical Neural Story Generation. In Proceedings of the Annual Meeting of the Association for Computational Linguistics, volume 1, 889–898. Association for Computational Linguistics (ACL).
  5. 2018. Food recognition and recipe analysis: integrating visual content, context and external knowledge. arXiv:1801.07239.
  6. 2020. The Curious Case of Neural Text DeGeneration. In Proceedings of the International Conference on Learning Representations.
  7. 2021. ENCONTER: Entity Constrained Progressive Sequence Generation via Insertion-based Transformer. In Proceedings of the 16th Conference of the European Chapter of the Association for Computational Linguistics, 3590–3599.
  8. 2020. RecipeGPT: Generative Pre-training Based Cooking Recipe Generation and Evaluation System. In WWW-20 - Companion Proceedings of the World Wide Web Conference, 181–184. Association for Computing Machinery.
  9. Lin, C.-Y. 2004. ROUGE: A Package for Automatic Evaluation of Summaries. In Text Summarization Branches Out: Association for Computational Linguistics, 74–81.
  10. 2021. Recipe1M+: A Dataset for Learning Cross-Modal Embeddings for Cooking Recipes and Food Images. IEEE Transactions on Pattern Analysis and Machine Intelligence 43(1):187–203.
  11. 2017. Being a supercook: Joint food attributes and multimodal content modeling for recipe retrieval and exploration. IEEE Transactions on Multimedia 19(5):1100–1113.
  12. 2002. BLEU: a Method for Automatic Evaluation of Machine Translation. In Proceedings of the 40th Annual Meeting of the Association for Computational Linguistics (ACL), 311–318.
  13. 2018. Improving Language Understanding by Generative Pre-Training. OpenAI.
  14. 2019. Language Models are Unsupervised Multitask Learners. OpenAI.
  15. 2021. RecipeGM: A Hierarchical Recipe Generation Model. In Proceedings of the IEEE 37th International Conference on Data Engineering Workshops, 24–29. Institute of Electrical and Electronics Engineers Inc.
  16. Rosin, C. D. 2011. Multi-armed Bandits with Episode Context. Annals of Mathematics and Artificial Intelligence 61:203–230.
  17. 2017. Learning Cross-modal Embeddings for Cooking Recipes and Food Images. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition.
  18. 2019. CTRL: A Conditional Transformer Language Model for Controllable Generation. arXiv:1909.05858.
  19. 2017. Mastering the game of Go without human knowledge. Nature 550(7676):354–359.
  20. 2022. Monte Carlo Tree Search: a review of recent modifications and applications. Artificial Intelligence Review 56(3):2497–2562.
  21. 2019. A Big Data Approach to Computational Creativity. IBM Journal of Research and Development 63(1):1–7.
  22. 2017. Attention Is All You Need. In Proceedings of the Conference on Neural Information Processing Systems.
  23. 2020. Structure-Aware Generation Network for Recipe Generation from Images. In Proceedings of the European Conference on Computer Vision, volume 12372 LNCS, 359–374. Springer Science and Business Media Deutschland GmbH.
  24. 2022. Learning to Break the Loop: Analyzing and Mitigating Repetitions for Neural Text Generation. In Proceedings of the Conference on Neural Information Processing Systems.
  25. 2020. POINTER: Constrained Progressive Text Generation via Insertion-based Generative Pre-training. In Proceedings of the Conference on Empirical Methods in Natural Language Processing, 8649–8670. Association for Computational Linguistics (ACL).
User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (3)
  1. Karan Taneja (11 papers)
  2. Richard Segal (2 papers)
  3. Richard Goodwin (3 papers)
Citations (3)

Summary

We haven't generated a summary for this paper yet.