Reliable Academic Conference Question Answering: A Study Based on Large Language Model (2310.13028v2)
Abstract: As the development of academic conferences fosters global scholarly communication, researchers consistently need to obtain accurate and up-to-date information about academic conferences. Since the information is scattered, using an intelligent question-answering system to efficiently handle researchers' queries and ensure awareness of the latest advancements is necessary. Recently, LLMs have demonstrated impressive capabilities in question answering, and have been enhanced by retrieving external knowledge to deal with outdated knowledge. However, these methods fail to work due to the lack of the latest conference knowledge. To address this challenge, we develop the ConferenceQA dataset, consisting of seven diverse academic conferences. Specifically, for each conference, we first organize academic conference data in a tree-structured format through a semi-automated method. Then we annotate question-answer pairs and classify the pairs into four different types to better distinguish their difficulty. With the constructed dataset, we further propose a novel method STAR (STructure-Aware Retrieval) to improve the question-answering abilities of LLMs, leveraging inherent structural information during the retrieval process. Experimental results on the ConferenceQA dataset show the effectiveness of our retrieval method. The dataset and code are available at https://github.com/zjukg/ConferenceQA.
- Knowledge-Augmented Language Model Prompting for Zero-Shot Knowledge Graph Question Answering. arXiv preprint arXiv:2306.04136 (2023).
- Training a helpful and harmless assistant with reinforcement learning from human feedback. arXiv preprint arXiv:2204.05862 (2022).
- A Multitask, Multilingual, Multimodal Evaluation of ChatGPT on Reasoning, Hallucination, and Interactivity. arXiv:2302.04023 [cs.CL]
- Improving language models by retrieving from trillions of tokens. In International conference on machine learning. PMLR, 2206–2240.
- Language models are few-shot learners. Advances in neural information processing systems 33 (2020), 1877–1901.
- Retrieval Augmented via Execution Guidance in Open-domain Table QA. In Proceedings of the 2022 5th International Conference on Algorithms, Computing and Artificial Intelligence. 1–6.
- Palm: Scaling language modeling with pathways. arXiv preprint arXiv:2204.02311 (2022).
- Scaling instruction-finetuned language models. arXiv preprint arXiv:2210.11416 (2022).
- CiteSeer: An automatic citation indexing system. In Proceedings of the third ACM conference on Digital libraries. 89–98.
- Retrieval augmented language model pre-training. In International conference on machine learning. PMLR, 3929–3938.
- ChatDB: Augmenting LLMs with Databases as Their Symbolic Memory. arXiv:2306.03901 [cs.AI]
- Atlas: Few-shot Learning with Retrieval Augmented Language Models. arXiv:2208.03299 [cs.CL]
- Few-shot learning with retrieval augmented language models. arXiv preprint arXiv:2208.03299 (2022).
- Survey of hallucination in natural language generation. Comput. Surveys 55, 12 (2023), 1–38.
- Active retrieval augmented generation. arXiv preprint arXiv:2305.06983 (2023).
- Large language models are zero-shot reasoners. Advances in neural information processing systems 35 (2022), 22199–22213.
- Internet-augmented language models through few-shot prompting for open-domain question answering. arXiv preprint arXiv:2203.05115 (2022).
- Retrieval-augmented generation for knowledge-intensive nlp tasks. Advances in Neural Information Processing Systems 33 (2020), 9459–9474.
- S2ORC: The semantic scholar open research corpus. arXiv preprint arXiv:1911.02782 (2019).
- Rethinking the role of demonstrations: What makes in-context learning work? arXiv preprint arXiv:2202.12837 (2022).
- RET-LLM: Towards a General Read-Write Memory for Large Language Models. arXiv:2305.14322 [cs.CL]
- Large dual encoders are generalizable retrievers. arXiv preprint arXiv:2112.07899 (2021).
- OpenAI. 2023. GPT-4 Technical Report. arXiv:2303.08774 [cs.CL]
- Training language models to follow instructions with human feedback. Advances in Neural Information Processing Systems 35 (2022), 27730–27744.
- End-to-End Table Question Answering via Retrieval-Augmented Generation. arXiv preprint arXiv:2203.16714 (2022).
- Unifying Large Language Models and Knowledge Graphs: A Roadmap. arXiv preprint arXiv:2306.08302 (2023).
- RocketQA: An optimized training approach to dense passage retrieval for open-domain question answering. arXiv preprint arXiv:2010.08191 (2020).
- The probabilistic relevance framework: BM25 and beyond. Foundations and Trends® in Information Retrieval 3, 4 (2009), 333–389.
- Cristóbal Romero and Sebastián Ventura. 2017. Educational data science in massive open online courses. Wiley Interdisciplinary Reviews: Data Mining and Knowledge Discovery 7, 1 (2017), e1187.
- Tarek Saier and Michael Färber. 2020. unarXive: a large scholarly data set with publications’ full-text, annotated in-text citations, and links to metadata. Scientometrics 125, 3 (2020), 3085–3108.
- Toolformer: Language Models Can Teach Themselves to Use Tools. arXiv:2302.04761 [cs.CL]
- Data science in data librarianship: Core competencies of a data librarian. Journal of Librarianship and Information Science 51, 3 (2019), 771–780.
- Knowledge graph-augmented language models for complex question answering. (2023).
- Replug: Retrieval-augmented black-box language models. arXiv preprint arXiv:2301.12652 (2023).
- LLaMA: Open and Efficient Foundation Language Models. arXiv:2302.13971 [cs.CL]
- Llama 2: Open Foundation and Fine-Tuned Chat Models. arXiv:2307.09288 [cs.CL]
- Rat-sql: Relation-aware schema encoding and linking for text-to-sql parsers. arXiv preprint arXiv:1911.04942 (2019).
- Microsoft academic graph: When experts are not enough. Quantitative Science Studies 1, 1 (2020), 396–413.
- KnowledGPT: Enhancing Large Language Models with Retrieval and Storage Access on Knowledge Bases. arXiv preprint arXiv:2308.11761 (2023).
- Chain-of-thought prompting elicits reasoning in large language models. Advances in Neural Information Processing Systems 35 (2022), 24824–24837.
- Retrieve-Rewrite-Answer: A KG-to-Text Enhanced LLMs Framework for Knowledge Graph Question Answering. arXiv preprint arXiv:2309.11206 (2023).
- Rrhf: Rank responses to align language models with human feedback without tears. arXiv preprint arXiv:2304.05302 (2023).
- The Effect of Metadata on Scientific Literature Tagging: A Cross-Field Cross-Model Study. In Proceedings of the ACM Web Conference 2023. 1626–1637.
- Seq2sql: Generating structured queries from natural language using reinforcement learning. arXiv preprint arXiv:1709.00103 (2017).