<?xml version="1.0" encoding="utf-8" standalone="yes"?><rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom"><channel><title>David Koop | UCSC OSPO</title><link>https://deploy-preview-1007--ucsc-ospo.netlify.app/author/david-koop/</link><atom:link href="https://deploy-preview-1007--ucsc-ospo.netlify.app/author/david-koop/index.xml" rel="self" type="application/rss+xml"/><description>David Koop</description><generator>Wowchemy (https://wowchemy.com)</generator><language>en-us</language><image><url>https://deploy-preview-1007--ucsc-ospo.netlify.app/author/david-koop/avatar_hu8c080dd36991bf76b93af3695fcccd09_98508_270x270_fill_q75_lanczos_center.jpg</url><title>David Koop</title><link>https://deploy-preview-1007--ucsc-ospo.netlify.app/author/david-koop/</link></image><item><title>Reproducibility in Data Visualization</title><link>https://deploy-preview-1007--ucsc-ospo.netlify.app/project/osre24/niu/repro-vis/</link><pubDate>Tue, 06 Feb 2024 15:00:00 -0500</pubDate><guid>https://deploy-preview-1007--ucsc-ospo.netlify.app/project/osre24/niu/repro-vis/</guid><description>&lt;p>At the heart of evaluating reproducibility is a judgment about whether
two results are indeed
the same. This can be complicated in the context of data visualization due to
rapidly evolving technology and differences in how users perceive the results.
First, due to the rapid evolution of libraries including web technologies,
visualizations created in the past may look different when rendered in the future.
Second, as the goal of data visualization is communicating data to people,
different people may perceive visualizations in a different way.
Thus, when a reproduced visualization does not exactly match the original, judging
whether they are &amp;ldquo;similar enough&amp;rdquo; is complicated by these factors. For example,
changes in a colormap may be deemed minor by a computer but could lead people to different
understandings of the data. The goals of this research are to capture visualizations in a way that
allows their reproducibility to be evaluated and to develop methods to categorize the differences
when a reproduced visualization differs from the original.&lt;/p>
&lt;h3 id="investigate-solutions-for-capturing-visualizations">Investigate Solutions for Capturing Visualizations&lt;/h3>
&lt;ul>
&lt;li>&lt;strong>Topics:&lt;/strong> Reproducibility, Data Visualization&lt;/li>
&lt;li>&lt;strong>Skills:&lt;/strong> Python and/or JavaScript, Data Visualization Tools&lt;/li>
&lt;li>&lt;strong>Difficulty:&lt;/strong> Moderate&lt;/li>
&lt;li>&lt;strong>Size:&lt;/strong> Medium or Large (175 or 350 hours)&lt;/li>
&lt;li>&lt;strong>Mentors:&lt;/strong> &lt;a href="https://deploy-preview-1007--ucsc-ospo.netlify.app/author/david-koop/">David Koop&lt;/a>&lt;/li>
&lt;/ul>
&lt;p>The goal of this project is to investigate, augment, and/or develop solutions to capture
visualizations that appear in formats including websites and Jupyter notebooks.
In &lt;a href="https://github.com/simprov/simprov" target="_blank" rel="noopener">past work&lt;/a>, we implemented methods
to capture thumbnails as users interacted with visualizations. Other solutions
can be used to capture interactive visualizations. We wish to understand
the feasibility of recording such visualizations and their utility in
evaluating reproducibility in the future.&lt;/p>
&lt;h5 id="specific-tasks">Specific tasks:&lt;/h5>
&lt;ul>
&lt;li>Evaluate tools for capturing static visualizations on the web&lt;/li>
&lt;li>Investigate tools for capturing dynamic visualizations on the web&lt;/li>
&lt;li>Investigate how data including code or metadata can be captured with visualizations&lt;/li>
&lt;li>Augment or develop tools to aid in capturing reproducible visualizations&lt;/li>
&lt;/ul>
&lt;h3 id="categorize-differences-in-reproduced-visualizations">Categorize Differences in Reproduced Visualizations&lt;/h3>
&lt;ul>
&lt;li>&lt;strong>Topics:&lt;/strong> Reproducibility, Data Visualization&lt;/li>
&lt;li>&lt;strong>Skills:&lt;/strong> Python and/or JavaScript, Data Visualization Tools&lt;/li>
&lt;li>&lt;strong>Difficulty:&lt;/strong> Moderate/Hard&lt;/li>
&lt;li>&lt;strong>Size:&lt;/strong> Medium or Large (175 or 350 hours)&lt;/li>
&lt;li>&lt;strong>Mentors:&lt;/strong> &lt;a href="https://deploy-preview-1007--ucsc-ospo.netlify.app/author/david-koop/">David Koop&lt;/a>&lt;/li>
&lt;/ul>
&lt;p>The goal of this project is to organize types of differences in reproduced visualizations and create tools to detect them. Publications and computational notebooks record renderings of visualizations.
When they also include the code to reproduce the visualization, we can
regenerate them in order to compare them. Often, the reproduced visualization does
not match the original (see examples in this &lt;a href="https://arxiv.org/abs/2308.06894" target="_blank" rel="noopener">manuscript&lt;/a>).
This project seeks to categorize the types of differences
that can occur in order and start understanding how they impact judgments of reproducibility.&lt;/p>
&lt;h5 id="specific-tasks-1">Specific tasks:&lt;/h5>
&lt;ul>
&lt;li>Evaluate and/or develop tools to compare two visualizations&lt;/li>
&lt;li>Evaluate the utility of artificial intelligence solutions&lt;/li>
&lt;li>Organize and categorize the detected differences&lt;/li>
&lt;li>Develop tools to determine the types or categories of differences present in two visualizations&lt;/li>
&lt;/ul></description></item></channel></rss>