Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash 100 tok/s
Gemini 2.5 Pro 58 tok/s Pro
GPT-5 Medium 29 tok/s
GPT-5 High 29 tok/s Pro
GPT-4o 103 tok/s
GPT OSS 120B 480 tok/s Pro
Kimi K2 215 tok/s Pro
2000 character limit reached

Imagination is All You Need! Curved Contrastive Learning for Abstract Sequence Modeling Utilized on Long Short-Term Dialogue Planning (2211.07591v2)

Published 14 Nov 2022 in cs.CL

Abstract: Inspired by the curvature of space-time (Einstein, 1921), we introduce Curved Contrastive Learning (CCL), a novel representation learning technique for learning the relative turn distance between utterance pairs in multi-turn dialogues. The resulting bi-encoder models can guide transformers as a response ranking model towards a goal in a zero-shot fashion by projecting the goal utterance and the corresponding reply candidates into a latent space. Here the cosine similarity indicates the distance/reachability of a candidate utterance toward the corresponding goal. Furthermore, we explore how these forward-entailing language representations can be utilized for assessing the likelihood of sequences by the entailment strength i.e. through the cosine similarity of its individual members (encoded separately) as an emergent property in the curved space. These non-local properties allow us to imagine the likelihood of future patterns in dialogues, specifically by ordering/identifying future goal utterances that are multiple turns away, given a dialogue context. As part of our analysis, we investigate characteristics that make conversations (un)plannable and find strong evidence of planning capability over multiple turns (in 61.56% over 3 turns) in conversations from the DailyDialog (Li et al., 2017) dataset. Finally, we show how we achieve higher efficiency in sequence modeling tasks compared to previous work thanks to our relativistic approach, where only the last utterance needs to be encoded and computed during inference.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (35)
  1. Ben Alderson-Day and Charles Fernyhough. 2015. Inner speech: Development, cognitive functions, phenomenology, and neurobiology. Psychological Bulletin, 141:931 – 965.
  2. A large annotated corpus for learning natural language inference.
  3. Entailment graph learning with textual entailment and soft transitivity.
  4. Bert: Pre-training of deep bidirectional transformers for language understanding.
  5. Human conversational behavior. Human nature (Hawthorne, N.Y.), 8:231–246.
  6. Evaluating coherence in dialogue systems using entailment. CoRR, abs/1904.03371.
  7. Albert Einstein. 1921. Relativity: The Special and General Theory. Routledge.
  8. Dialogue response ranking training with large-scale human feedback data. CoRR, abs/2009.06978.
  9. The ConDialInt Model: Condensation, Dialogality, and Intentionality Dimensions of Inner Speech Within a Hierarchical Predictive Control Framework. Frontiers in Psychology, 10:2019.
  10. ConveRT: Efficient and accurate conversational representations from transformers. In Findings of the Association for Computational Linguistics: EMNLP 2020, pages 2161–2174, Online. Association for Computational Linguistics.
  11. Billion-scale similarity search with gpus. CoRR, abs/1702.08734.
  12. Textual entailment graphs. Natural Language Engineering, 21:699 – 724.
  13. Microsoft dialogue challenge: Building end-to-end task-completion dialogue systems.
  14. DailyDialog: A manually labelled multi-turn dialogue dataset. In Proceedings of the Eighth International Joint Conference on Natural Language Processing (Volume 1: Long Papers), pages 986–995, Taipei, Taiwan. Asian Federation of Natural Language Processing.
  15. Improving response quality with backward reasoning in open-domain dialogue systems. CoRR, abs/2105.00079.
  16. Dialoguecse: Dialogue-based contrastive learning of sentence embeddings.
  17. Roberta: A robustly optimized bert pretraining approach.
  18. A survey on bias and fairness in machine learning. CoRR, abs/1908.09635.
  19. Typical decoding for natural language generation. CoRR, abs/2202.00666.
  20. Rauni Myllyniemi. 1986. Conversation as a system of social interaction. Language & Communication, 6(3):147–169.
  21. Abiola Obamuyide and Andreas Vlachos. 2018. Zero-shot relation classification as textual entailment. In Proceedings of the First Workshop on Fact Extraction and VERification (FEVER), pages 72–78, Brussels, Belgium. Association for Computational Linguistics.
  22. Carbon emissions and large neural network training. CoRR, abs/2104.10350.
  23. Godel: Large-scale pre-training for goal-directed dialog. arXiv.
  24. Long-term control for dialogue generation: Methods and evaluation.
  25. Nils Reimers and Iryna Gurevych. 2019. Sentence-bert: Sentence embeddings using siamese bert-networks. CoRR, abs/1908.10084.
  26. Stephen Robertson and Hugo Zaragoza. 2009. The probabilistic relevance framework: Bm25 and beyond. Found. Trends Inf. Retr., 3(4):333–389.
  27. Large pre-trained language models contain human-like biases of what is right and wrong to do. Nature Machine Intelligence, 4(3):258–268.
  28. Energy and policy considerations for deep learning in NLP. CoRR, abs/1906.02243.
  29. Milene Teixeira and Mauro Dragoni. 2022. A review of plan-based approaches for dialogue management. Cognitive Computation, 14.
  30. Ethical and social risks of harm from language models. CoRR, abs/2112.04359.
  31. A broad-coverage challenge corpus for sentence understanding through inference.
  32. Benchmarking zero-shot text classification: Datasets, evaluation and entailment approach. CoRR, abs/1909.00161.
  33. Graph contrastive learning with augmentations. In Advances in Neural Information Processing Systems, volume 33, pages 5812–5823. Curran Associates, Inc.
  34. Positional contrastive learning for volumetric medical image segmentation. CoRR, abs/2106.09157.
  35. Dialogpt: Large-scale generative pre-training for conversational response generation.
Citations (1)
List To Do Tasks Checklist Streamline Icon: https://streamlinehq.com

Collections

Sign up for free to add this paper to one or more collections.

Summary

We haven't generated a summary for this paper yet.

Ai Generate Text Spark Streamline Icon: https://streamlinehq.com

Paper Prompts

Sign up for free to create and run prompts on this paper using GPT-5.

Dice Question Streamline Icon: https://streamlinehq.com

Follow-up Questions

We haven't generated follow-up questions for this paper yet.