Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
41 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Exploring the Capability of LLMs in Performing Low-Level Visual Analytic Tasks on SVG Data Visualizations (2404.19097v2)

Published 29 Apr 2024 in cs.HC

Abstract: Data visualizations help extract insights from datasets, but reaching these insights requires decomposing high level goals into low-level analytic tasks that can be complex due to varying degrees of data literacy and visualization experience. Recent advancements in LLMs have shown promise for lowering barriers for users to achieve tasks such as writing code and may likewise facilitate visualization insight. Scalable Vector Graphics (SVG), a text-based image format common in data visualizations, matches well with the text sequence processing of transformer-based LLMs. In this paper, we explore the capability of LLMs to perform 10 low-level visual analytic tasks defined by Amar, Eagan, and Stasko directly on SVG-based visualizations. Using zero-shot prompts, we instruct the models to provide responses or modify the SVG code based on given visualizations. Our findings demonstrate that LLMs can effectively modify existing SVG visualizations for some tasks like Cluster but perform poorly on tasks requiring mathematical operations like Compute Derived Value. We also discovered that LLM performance can vary based on factors such as the number of data points, the presence of value labels, and the chart type. Our findings contribute to gauging the general capabilities of LLMs and highlight the need for further exploration and development to fully harness their potential in supporting visual analytic tasks.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (42)
  1. Task-driven evaluation of aggregation in time series visualization. In Proceedings of the SIGCHI Conference on Human Factors in Computing Systems, pp. 551–560, 2014.
  2. Low-level components of analytic activity in information visualization. In IEEE Symposium on Information Visualization, 2005. INFOVIS 2005., pp. 111–117. IEEE, 2005.
  3. Who do we mean when we talk about visualization novices? In Proceedings of the 2023 CHI Conference on Human Factors in Computing Systems, pp. 1–16, 2023.
  4. Leveraging large language models for scalable vector graphics-driven image understanding. arXiv preprint arXiv:2306.06094, 2023.
  5. A survey on evaluation of large language models. ACM Transactions on Intelligent Systems and Technology, 2023.
  6. Unleashing the potential of prompt engineering in large language models: a comprehensive review. arXiv preprint arXiv:2310.14735, 2023.
  7. Beyond generating code: Evaluating gpt on a data visualization course. In 2023 IEEE VIS Workshop on Visualization Education, Literacy, and Activities (EduVis), pp. 16–21. IEEE, 2023.
  8. M. Dastani. The role of visual perception in data visualization. Journal of Visual Languages & Computing, 13(6):601–622, 2002.
  9. How ready are pre-trained abstractive models and llms for legal case judgement summarization? arXiv preprint arXiv:2306.01248, 2023.
  10. V. Dibia. Lida: A tool for automatic generation of grammar-agnostic visualizations and infographics using large language models. arXiv preprint arXiv:2303.02927, 2023.
  11. Parameter-efficient fine-tuning of large-scale pre-trained language models. Nature Machine Intelligence, 5(3):220–235, 2023.
  12. Phlya/adjusttext: 0.8. Zenodo, 2023.
  13. The science of visual data communication: What works (vol 22, pg 110, 2021). PSYCHOLOGICAL SCIENCE IN THE PUBLIC INTEREST, 23(1):41–42, 2022.
  14. Demystifying prompts in language models via perplexity estimation, 2022.
  15. How information visualization novices construct visualizations. IEEE transactions on visualization and computer graphics, 16(6):943–952, 2010.
  16. Belief decay or persistence? a mixed-method study on belief movement over time. In Computer Graphics Forum, vol. 42, pp. 111–122. Wiley Online Library, 2023.
  17. Chartllama: A multimodal llm for chart understanding and generation. arXiv preprint arXiv:2311.16483, 2023.
  18. M.-H. Hong and A. Crisan. Conversational ai threads for visualizing multidimensional datasets. arXiv preprint arXiv:2311.05590, 2023.
  19. M. Islam and S. Jin. An overview of data visualization. In 2019 International Conference on Information Science and Communications Technologies (ICISCT), pp. 1–7. IEEE, 2019.
  20. Clams: a cluster ambiguity measure for estimating perceptual variability in visual clustering. IEEE Transactions on Visualization and Computer Graphics, 2023.
  21. J. Jeon and S. Lee. Large language models in education: A focus on the complementary relationship between human teachers and chatgpt. Education and Information Technologies, 28(12):15873–15892, 2023.
  22. Chart-to-text: A large-scale benchmark for chart summarization. arXiv preprint arXiv:2203.06486, 2022.
  23. How novices use llm-based code generators to solve cs1 coding tasks in a self-paced learning environment. In Proceedings of the 23rd Koli Calling International Conference on Computing Education Research, pp. 1–12, 2023.
  24. Unnatural language processing: How do language models handle machine-generated prompts?, 2023.
  25. Q. Li and Q. Li. Overview of data visualization. Embodying data: Chinese aesthetics, interactive visualization and gaming technologies, pp. 17–47, 2020.
  26. A. Liew and K. Mueller. Using large language models to generate engaging captions for data visualizations. arXiv preprint arXiv:2212.14047, 2022.
  27. nvbench: A large-scale synthesized dataset for cross-domain natural language to visualization task. arXiv preprint arXiv:2112.12926, 2021.
  28. P. Maddigan and T. Susnjak. Chat2vis: Generating data visualisations via natural language using chatgpt, codex and gpt-3 large language models. Ieee Access, 2023.
  29. Methods and algorithms for correlation analysis in r. Journal of Open Source Software, 5(51):2306, 2020.
  30. Accuracy on the line: on the strong correlation between out-of-distribution and in-distribution generalization. In International conference on machine learning, pp. 7721–7735. PMLR, 2021.
  31. Reframing instructional prompts to gptk’s language. arXiv preprint arXiv:2109.07830, 2021.
  32. OpenAI. Chatgpt-4: Optimizing language models for dialogue. https://openai.com/chatgpt-4, 2023. Accessed: 2024-04-05.
  33. Automatic scatterplot design optimization for clustering identification. IEEE Transactions on Visualization and Computer Graphics, 2022.
  34. G. J. Quadri and P. Rosen. A survey of perception-based visualization studies by task. IEEE transactions on visualization and computer graphics, 28(12):5026–5048, 2021.
  35. A. Quint. Scalable vector graphics. IEEE MultiMedia, 10(3):99–102, 2003.
  36. L. Reynolds and K. McDonell. Prompt programming for large language models: Beyond the few-shot paradigm. In Extended Abstracts of the 2021 CHI Conference on Human Factors in Computing Systems, pp. 1–7, 2021.
  37. Nl2color: Refining color palettes for charts with natural language. IEEE Transactions on Visualization and Computer Graphics, 2023.
  38. Chartgpt: Leveraging llms to generate charts from abstract natural language. IEEE Transactions on Visualization and Computer Graphics, 2024.
  39. Keyframer: Empowering animation design using large language models. arXiv preprint arXiv:2402.06071, 2024.
  40. P.-P. Vázquez. Are llms ready for visualization?, 2024.
  41. Line graph or scatter plot? automatic selection of methods for visualizing trends in time series. IEEE transactions on visualization and computer graphics, 24(2):1141–1154, 2017.
  42. How well do large language models perform in arithmetic tasks?, 2023.
User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (2)
  1. Zhongzheng Xu (2 papers)
  2. Emily Wall (8 papers)
Citations (4)
X Twitter Logo Streamline Icon: https://streamlinehq.com

Tweets