96 research outputs found
SeeChart: Enabling Accessible Visualizations Through Interactive Natural Language Interface For People with Visual Impairments
Web-based data visualizations have become very popular for exploring data and
communicating insights. Newspapers, journals, and reports regularly publish
visualizations to tell compelling stories with data. Unfortunately, most
visualizations are inaccessible to readers with visual impairments. For many
charts on the web, there are no accompanying alternative (alt) texts, and even
if such texts exist they do not adequately describe important insights from
charts. To address the problem, we first interviewed 15 blind users to
understand their challenges and requirements for reading data visualizations.
Based on the insights from these interviews, we developed SeeChart, an
interactive tool that automatically deconstructs charts from web pages and then
converts them to accessible visualizations for blind people by enabling them to
hear the chart summary as well as to interact through data points using the
keyboard. Our evaluation with 14 blind participants suggests the efficacy of
SeeChart in understanding key insights from charts and fulfilling their
information needs while reducing their required time and cognitive burden.Comment: 28 pages, 13 figure
ChartInstruct: Instruction Tuning for Chart Comprehension and Reasoning
Charts provide visual representations of data and are widely used for
analyzing information, addressing queries, and conveying insights to others.
Various chart-related downstream tasks have emerged recently, such as
question-answering and summarization. A common strategy to solve these tasks is
to fine-tune various models originally trained on vision tasks language.
However, such task-specific models are not capable of solving a wide range of
chart-related tasks, constraining their real-world applicability. To overcome
these challenges, we introduce ChartInstruct: a novel chart-specific
vision-language Instruction-following dataset comprising 191K instructions
generated with 71K charts. We then present two distinct systems for instruction
tuning on such datasets: (1) an end-to-end model that connects a vision encoder
for chart understanding with a LLM; and (2) a pipeline model that employs a
two-step approach to extract chart data tables and input them into the LLM. In
experiments on four downstream tasks, we first show the effectiveness of our
model--achieving a new set of state-of-the-art results. Further evaluation
shows that our instruction-tuning approach supports a wide array of real-world
chart comprehension and reasoning scenarios, thereby expanding the scope and
applicability of our models to new kinds of tasks
- …