Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
144 tokens/sec
GPT-4o
7 tokens/sec
Gemini 2.5 Pro Pro
45 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

LLMs in HCI Data Work: Bridging the Gap Between Information Retrieval and Responsible Research Practices (2403.18173v1)

Published 27 Mar 2024 in cs.HC and cs.IR

Abstract: Efficient and accurate information extraction from scientific papers is significant in the rapidly developing human-computer interaction research in the literature review process. Our paper introduces and analyses a new information retrieval system using state-of-the-art LLMs in combination with structured text analysis techniques to extract experimental data from HCI literature, emphasizing key elements. Then We analyze the challenges and risks of using LLMs in the world of research. We performed a comprehensive analysis on our conducted dataset, which contained the specified information of 300 CHI 2020-2022 papers, to evaluate the performance of the two LLMs, GPT-3.5 (text-davinci-003) and Llama-2-70b, paired with structured text analysis techniques. The GPT-3.5 model gains an accuracy of 58\% and a mean absolute error of 7.00. In contrast, the Llama2 model indicates an accuracy of 56\% with a mean absolute error of 7.63. The ability to answer questions was also included in the system in order to work with streamlined data. By evaluating the risks and opportunities presented by LLMs, our work contributes to the ongoing dialogue on establishing methodological validity and ethical guidelines for LLM use in HCI data work.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (16)
  1. ChatGPT “ChatGPT” OpenAI chat.openai.com URL: https://chat.openai.com/
  2. Elasticsearch Contributors “Elasticsearch: Distributed search & analytics engine”, 2023 URL: https://www.elastic.co/products/elasticsearch
  3. “InPars-v2: Large Language Models as Efficient Dataset Generators for Information Retrieval”, 2023 arXiv:2301.01820 [cs.IR]
  4. “HAGRID: A Human-LLM Collaborative Dataset for Generative Information-Seeking with Attribution”, 2023 arXiv:2307.16883 [cs.CL]
  5. Llama “Llama” Meta meta.com URL: https://llama.meta.com/
  6. Patrice Lopez “GROBID: Combining Automatic Bibliographic Data Recognition and Term Extraction for Scholarship Publications” In Research and Advanced Technology for Digital Libraries Berlin, Heidelberg: Springer Berlin Heidelberg, 2009, pp. 473–474
  7. Patrice Lopez “GROBID: Combining automatic bibliographic data recognition and term extraction for scholarship publications”, 2009 URL: https://github.com/kermitt2/grobid
  8. “Retrieval Augmented Generation and Representative Vector Summarization for large unstructured textual data in Medical Education”, 2023 arXiv:2308.00479 [cs.CL]
  9. “The Stanford CoreNLP Natural Language Processing Toolkit”, 2014 URL: https://stanfordnlp.github.io/CoreNLP/
  10. “Language Models are Unsupervised Multitask Learners”, 2019
  11. “In-Context Retrieval-Augmented Language Models”, 2023 arXiv:2302.00083 [cs.CL]
  12. “Modern Information Retrieval: A Brief Overview” In IEEE Data Engineering Bulletin 24, 2001
  13. “LXMERT: Learning Cross-Modality Encoder Representations from Transformers”, 2019 arXiv:1908.07490 [cs.CL]
  14. “Reimagining Retrieval Augmented Language Models for Answering Queries”, 2023 arXiv:2306.01061 [cs.CL]
  15. “A Proximity Language Model for Information Retrieval” In Proceedings of the 32nd International ACM SIGIR Conference on Research and Development in Information Retrieval, SIGIR ’09 Boston, MA, USA: Association for Computing Machinery, 2009, pp. 291–298 DOI: 10.1145/1571941.1571993
  16. “Large Language Models for Information Retrieval: A Survey”, 2023 arXiv:2308.07107 [cs.CL]
Citations (1)

Summary

We haven't generated a summary for this paper yet.

X Twitter Logo Streamline Icon: https://streamlinehq.com

Tweets