The study explored the capability of large language models (LLMs) to perform low-level visual analytic tasks defined by Amar, Eagan, and Stasko on SVG-based data visualizations. The researchers generated 320 unique stimuli covering 3 chart types (scatterplot, line chart, bar chart), 2 dataset sizes (small, medium), and 2 labeling schemas (labeled, unlabeled), and evaluated the LLM's performance on 10 low-level tasks.
The key findings are:
The LLM achieved 100% accuracy in retrieving values from labeled line charts and bar charts, but struggled with unlabeled charts and scatterplots. It often extracted coordinates directly from the SVG code rather than the intended data values.
The LLM performed well in tasks involving pattern recognition, such as Cluster and Find Anomalies, but struggled with tasks requiring complex mathematical operations, such as Compute Derived Value and Correlate.
The LLM's performance varied based on the number of data points and the presence of value labels, but the effects were not consistent across all tasks. For example, more data points improved accuracy for Find Extremum in line and bar charts, but not for scatterplots.
The LLM refused to modify the SVG code for the Correlate task, indicating limitations in its ability to perform certain types of chart manipulations.
The findings contribute to understanding the general capabilities of LLMs and highlight the need for further exploration and development to fully harness their potential in supporting visual analytic tasks.
Para Outro Idioma
do conteúdo original
arxiv.org
Principais Insights Extraídos De
by Zhongzheng X... às arxiv.org 05-01-2024
https://arxiv.org/pdf/2404.19097.pdfPerguntas Mais Profundas