On the Sequence Evaluation based on Stochastic Processes (2405.17764v3)
Abstract: Generative models have gained significant prominence in NLP, especially in tackling the complex task of modeling and evaluating long text sequences. This task is crucial for advancing various downstream applications, such as text generation and machine translation. Recent methods that utilize stochastic processes to capture the intrinsic dynamics of sequences have shown superior performance in generative modeling. However, the accurate encoding of both temporal and structural dependencies from text datasets, as well as leveraging this encoded information for sequence evaluation, remains an open area of research. In this paper, we propose a novel approach to learn the stochastic dynamics of long text sequences, utilizing a negative log-likelihood-based encoder that outperforms contrastive learning methods. We also introduce a likelihood-based evaluation metric for long-text assessment, which measures sequence coherence and can be applied to downstream tasks such as Human-AI discrimination. Our encoder preserves sequence coherence effectively and performs robustly on out-of-domain datasets. Additionally, the proposed evaluation metric captures both temporal and structural information comprehensively. Theoretical analysis demonstrates the superiority of our metric in sequence evaluation, and experimental results highlight its flexibility and exceptional performance across a variety of tasks, showcasing its utility in diverse NLP applications.
- A survey of diffusion models in natural language processing, 2023.
- Diffusion models: A comprehensive survey of methods and applications. ACM Comput. Surv., 56(4), nov 2023. ISSN 0360-0300. doi: 10.1145/3626235. URL https://doi.org/10.1145/3626235.
- Diffusion models in text generation: a survey. PeerJ Computer Science, 10:e1905, 2024. doi: 10.7717/peerj-cs.1905. URL https://doi.org/10.7717/peerj-cs.1905.
- Discourse coherence in the wild: A dataset, evaluation and methods. In Proceedings of the 19th Annual SIGdial Meeting on Discourse and Dialogue, pages 214–223, Melbourne, Australia, July 2018. Association for Computational Linguistics. doi: 10.18653/v1/W18-5023. URL https://aclanthology.org/W18-5023.
- Entity-based neural local coherence modeling. In Smaranda Muresan, Preslav Nakov, and Aline Villavicencio, editors, Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 7787–7805, Dublin, Ireland, May 2022. Association for Computational Linguistics. doi: 10.18653/v1/2022.acl-long.537. URL https://aclanthology.org/2022.acl-long.537.
- Representation learning with contrastive predictive coding. CoRR, abs/1807.03748, 2018. URL http://arxiv.org/abs/1807.03748.
- On contrastive learning for likelihood-free inference. In Hal Daumé III and Aarti Singh, editors, Proceedings of the 37th International Conference on Machine Learning, volume 119 of Proceedings of Machine Learning Research, pages 2771–2781. PMLR, 13–18 Jul 2020. URL https://proceedings.mlr.press/v119/durkan20a.html.
- Language modeling via stochastic processes. In International Conference on Learning Representations, 2022. URL https://openreview.net/forum?id=pMQwKL1yctf.
- Stochastic differential equations. Springer, 2003.
- John Krumm. Brownian bridge interpolation for human mobility? In Proceedings of the 29th International Conference on Advances in Geographic Information Systems, SIGSPATIAL ’21, pages 175–183, New York, NY, USA, 2021. Association for Computing Machinery. ISBN 9781450386647. doi: 10.1145/3474717.3483942. URL https://doi.org/10.1145/3474717.3483942.
- Analyzing animal movements using brownian bridges. Ecology, 88(9):2354–2363, Sep 2007. ISSN 0012-9658 (Print); 0012-9658 (Linking). doi: 10.1890/06-0957.1.
- Bbscore: A brownian bridge based metric for assessing text coherence. Proceedings of the AAAI Conference on Artificial Intelligence, 38(13):14937–14945, 2024.
- Score-based generative modeling through stochastic evolution equations in hilbert spaces. In Thirty-seventh Conference on Neural Information Processing Systems, 2023. URL https://openreview.net/forum?id=GrElRvXnEj.
- Modeling video as stochastic processes for fine-grained video representation learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pages 2225–2234, June 2023.
- On contrastive representations of stochastic processes. In M. Ranzato, A. Beygelzimer, Y. Dauphin, P.S. Liang, and J. Wortman Vaughan, editors, Advances in Neural Information Processing Systems, volume 34, pages 28823–28835. Curran Associates, Inc., 2021. URL https://proceedings.neurips.cc/paper_files/paper/2021/file/f19c44d068fecac1d6d13a80df4f8e96-Paper.pdf.
- Contrastive learning of strong-mixing continuous-time stochastic processes. In Arindam Banerjee and Kenji Fukumizu, editors, Proceedings of The 24th International Conference on Artificial Intelligence and Statistics, volume 130 of Proceedings of Machine Learning Research, pages 3151–3159. PMLR, 13–15 Apr 2021. URL https://proceedings.mlr.press/v130/liu21h.html.
- Tanya Reinhart. Conditions for text coherence. Poetics Today, 1(4):161–180, 1980.
- Model criticism for long-form text generation. In Yoav Goldberg, Zornitsa Kozareva, and Yue Zhang, editors, Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing, pages 11887–11912, Abu Dhabi, United Arab Emirates, December 2022. Association for Computational Linguistics. doi: 10.18653/v1/2022.emnlp-main.815. URL https://aclanthology.org/2022.emnlp-main.815.
- Generating sentences from a continuous space. In Stefan Riezler and Yoav Goldberg, editors, Proceedings of the 20th SIGNLL Conference on Computational Natural Language Learning, pages 10–21, Berlin, Germany, August 2016. Association for Computational Linguistics. doi: 10.18653/v1/K16-1002. URL https://aclanthology.org/K16-1002.
- SimCSE: Simple contrastive learning of sentence embeddings. In Marie-Francine Moens, Xuanjing Huang, Lucia Specia, and Scott Wen-tau Yih, editors, Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing, pages 6894–6910, Online and Punta Cana, Dominican Republic, November 2021. Association for Computational Linguistics. doi: 10.18653/v1/2021.emnlp-main.552. URL https://aclanthology.org/2021.emnlp-main.552.
- A Novel Computational and Modeling Foundation for Automatic Coherence Assessment. arXiv e-prints, art. arXiv:2310.00598, October 2023. doi: 10.48550/arXiv.2310.00598.
- Modeling local coherence: An entity-based approach. In Kevin Knight, Hwee Tou Ng, and Kemal Oflazer, editors, Proceedings of the 43rd Annual Meeting of the Association for Computational Linguistics (ACL’05), pages 141–148, Ann Arbor, Michigan, June 2005. Association for Computational Linguistics. doi: 10.3115/1219840.1219858. URL https://aclanthology.org/P05-1018.
- A unified neural coherence model. In Kentaro Inui, Jing Jiang, Vincent Ng, and Xiaojun Wan, editors, Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP), pages 2262–2272, Hong Kong, China, November 2019. Association for Computational Linguistics. doi: 10.18653/v1/D19-1231. URL https://aclanthology.org/D19-1231.
- Sector: A neural model for coherent topic segmentation and classification. In Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing (EMNLP), pages 241–253, Hong Kong, China, 2019. Association for Computational Linguistics. doi: 10.18653/v1/D19-1024. URL https://aclanthology.org/D19-1024.
- Pointer sentinel mixture models, 2016.
- How close is chatgpt to human experts? comparison corpus, evaluation, and detection. arXiv preprint arXiv:2301.07597, 2023.