Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
139 tokens/sec
GPT-4o
47 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Evaluating Large Language Models in Semantic Parsing for Conversational Question Answering over Knowledge Graphs (2401.01711v1)

Published 3 Jan 2024 in cs.CL and cs.IR

Abstract: Conversational question answering systems often rely on semantic parsing to enable interactive information retrieval, which involves the generation of structured database queries from a natural language input. For information-seeking conversations about facts stored within a knowledge graph, dialogue utterances are transformed into graph queries in a process that is called knowledge-based conversational question answering. This paper evaluates the performance of LLMs that have not been explicitly pre-trained on this task. Through a series of experiments on an extensive benchmark dataset, we compare models of varying sizes with different prompting techniques and identify common issue types in the generated output. Our results demonstrate that LLMs are capable of generating graph queries from dialogues, with significant improvements achievable through few-shot prompting and fine-tuning techniques, especially for smaller models that exhibit lower zero-shot performance.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (22)
  1. Analysing mixed initiatives and search strategies during conversational search. In Proc. of the 30th CIKM, page 16–26, New York, NY, USA. ACM.
  2. Vicuna: An open-source chatbot impressing gpt-4 with 90% chatgpt quality. LMSYS Org Blog.
  3. BERT: Pre-training of deep bidirectional transformers for language understanding. In Proc. of the 2019 NAACL: Human Language Technologies, pages 4171–4186, Minneapolis, Minnesota. ACL.
  4. Beyond i.i.d.: Three levels of generalization for question answering on knowledge bases. In Proc. of the Web Conference 2021, WWW ’21, page 3477–3488, New York, NY, USA. ACM.
  5. Towards complex text-to-SQL in cross-domain database with intermediate representation. In Proc. of the 57th Ann. Meeting of the ACL, pages 4524–4535, Florence, Italy. ACL.
  6. LoRA: Low-rank adaptation of large language models. In International Conference on Learning Representations.
  7. Survey of hallucination in natural language generation. ACM Comput. Surv., 55(12).
  8. Conversational question answering over knowledge graphs with transformer and graph attention networks. In Proc. of the 16th EACL, pages 850–862, Online. ACL.
  9. Context dependent semantic parsing: A survey. In Proc. of the 28th International Conference on Computational Linguistics, pages 2509–2521, Barcelona, Spain (Online). International Committee on Computational Linguistics.
  10. Towards collaborative neural-symbolic graph semantic parsing via uncertainty. In Findings of the ACL: ACL 2022, pages 4160–4173, Dublin, Ireland. ACL.
  11. Pre-train, prompt, and predict: A systematic survey of prompting methods in natural language processing. ACM Computing Surveys, 55(9):1–35.
  12. How far are we from effective context modeling? an exploratory study on semantic parsing in context. In Proc. of the Twenty-Ninth International Joint Conference on Artificial Intelligence, IJCAI’20.
  13. OpenAI (2022). Chatgpt: Optimizing language models for dialogue. OpenAI.
  14. Semantic parsing for conversational question answering over knowledge graphs. In Proc. of the 17th EACL, pages 2507–2522, Dubrovnik, Croatia. ACL.
  15. Language models are unsupervised multitask learners. OpenAI.
  16. Complex sequential question answering: Towards learning to converse over linked question answer pairs with a knowledge graph. In Proc. of the Thirty-Second AAAI Conference on Artificial Intelligence. AAAI Press.
  17. Investigating conversational search behavior for domain exploration. In Advances in Information Retrieval: 45th European Conference on Information Retrieval, ECIR 2023, Dublin, Ireland, April 2–6, 2023, Proceedings, Part II, page 608–616, Berlin, Heidelberg. Springer-Verlag.
  18. Learning to map context-dependent sentences to executable formal queries. In Proc. of the 2018 NAACL: Human Language Technologies, pages 2238–2249, New Orleans, Louisiana. ACL.
  19. Sequence to sequence learning with neural networks. In Ghahramani, Z., Welling, M., Cortes, C., Lawrence, N., and Weinberger, K., editors, NeurIPS, volume 27. Curran Associates, Inc.
  20. Llama: Open and efficient foundation language models. arXiv:2302.13971.
  21. RAT-SQL: Relation-aware schema encoding and linking for text-to-SQL parsers. In Proc. of the 58th Ann. Meeting of the ACL, pages 7567–7578. ACL.
  22. Editing-based SQL query generation for cross-domain context-dependent questions. In Proc. of the 2019 Conference on EMNLP and the 9th IJCNLP, pages 5338–5349, Hong Kong, China. ACL.
Citations (3)

Summary

We haven't generated a summary for this paper yet.