CAT-LLM: Prompting Large Language Models with Text Style Definition for Chinese Article-style Transfer (2401.05707v1)
Abstract: Text style transfer is increasingly prominent in online entertainment and social media. However, existing research mainly concentrates on style transfer within individual English sentences, while ignoring the complexity of long Chinese texts, which limits the wider applicability of style transfer in digital media realm. To bridge this gap, we propose a Chinese Article-style Transfer framework (CAT-LLM), leveraging the capabilities of LLMs. CAT-LLM incorporates a bespoke, pluggable Text Style Definition (TSD) module aimed at comprehensively analyzing text features in articles, prompting LLMs to efficiently transfer Chinese article-style. The TSD module integrates a series of machine learning algorithms to analyze article-style from both words and sentences levels, thereby aiding LLMs thoroughly grasp the target style without compromising the integrity of the original text. In addition, this module supports dynamic expansion of internal style trees, showcasing robust compatibility and allowing flexible optimization in subsequent research. Moreover, we select five Chinese articles with distinct styles and create five parallel datasets using ChatGPT, enhancing the models' performance evaluation accuracy and establishing a novel paradigm for evaluating subsequent research on article-style transfer. Extensive experimental results affirm that CAT-LLM outperforms current research in terms of transfer accuracy and content preservation, and has remarkable applicability to various types of LLMs.
- Language models are few-shot learners. Advances in neural information processing systems, 33:1877–1901, 2020.
- Cristina Castillo and L. Tolchinsky. The contribution of vocabulary knowledge and semantic orthographic fluency to text quality through elementary school in catalan. Reading and Writing, 31:293–323, 2018.
- Style transformer: Unpaired text style transfer without disentangled latent representation. In Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, pages 5997–6007, 2019.
- A computational approach to politeness with application to social factors. In Proceedings of the 51st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 250–259, 2013.
- S. Gajda. The faces of style and stylistics. Journal of Linguistics/Jazykovedný casopis, 73:7–26, 2022.
- Toward controlled generation of text. In Proceedings of the 34th International Conference on Machine Learning - Volume 70, page 1587–1596, 2017.
- Can large language models truly understand prompts? a case study with negated prompts. In Transfer Learning for Natural Language Processing Workshop, pages 52–62, 2023.
- A stylistic analysis of ’o uganda, land of beauty’ by prof.george kakoma. International Journal on Studies in English Language and Literature, 2020.
- I Wayan Sidha Karya and Ida Bagus Adhika Mahardika. A study on how long and short sentences show the story’s pacing in anthony horowitz’s raven’s gate. SPHOTA: Jurnal Linguistik dan Sastra, 2019.
- Multidimensional evaluation for text style transfer using chatgpt. arXiv preprint arXiv:2304.13462, 2023.
- Enhancing content preservation in text style transfer using reverse attention and conditional layer normalization. In Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing (Volume 1: Long Papers), pages 93–102, 2021.
- M. Lewis and Michael C. Frank. The length of words reflects their conceptual complexity. Cognition, 153:182–195, 2016.
- Guiding large language models via directional stimulus prompting. arXiv preprint arXiv:2302.11520, 2023.
- Bleu: a method for automatic evaluation of machine translation. In Proceedings of the 40th annual meeting of the Association for Computational Linguistics, pages 311–318, 2002.
- Piotr Przybyla. Capturing the style of fake news. In Proceedings of the AAAI conference on artificial intelligence, volume 34, pages 490–497, 2020.
- Modeling statistical properties of written text. PLoS ONE, 4, 2009.
- Role play with large language models. Nature, pages 1–6, 2023.
- Prompting large language models with answer heuristics for knowledge-based visual question answering. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 14974–14983, 2023.
- Automatic classification of documents by formality. In Proceedings of the 6th international conference on natural language processing and knowledge engineering (nlpke-2010), pages 1–5. IEEE, 2010.
- Progprompt: Generating situated robot task plans using large language models. In 2023 IEEE International Conference on Robotics and Automation (ICRA), pages 11523–11530. IEEE, 2023.
- Llm-planner: Few-shot grounded planning for embodied agents with large language models. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pages 2998–3009, 2023.
- A review of text style transfer using deep learning. IEEE Transactions on Artificial Intelligence, 2021.
- Teaching students to recognize the pros and cons of short and long sentences. English for Academic Research: A Guide for Teachers, pages 69–77, 2016.
- Rolellm: Benchmarking, eliciting, and enhancing role-playing abilities of large language models. arXiv preprint arXiv:2310.00746, 2023.
- Text style transfer via learning style instance supported latent space. In Proceedings of the Twenty-Ninth International Conference on International Joint Conferences on Artificial Intelligence, pages 3801–3807, 2021.
- Bertscore: Evaluating text generation with bert. In International Conference on Learning Representations, 2020.
- StoryTrans: Non-parallel story author-style transfer with discourse representations and content enhancing. In Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 14803–14819, 2023.
- Zhen Tao (39 papers)
- Dinghao Xi (4 papers)
- Zhiyu Li (69 papers)
- Liumin Tang (1 paper)
- Wei Xu (536 papers)