<?xml version="1.0" encoding="utf-8" standalone="yes"?><rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom"><channel><title>Yang Wang | UCSC OSPO</title><link>https://deploy-preview-1007--ucsc-ospo.netlify.app/author/yang-wang/</link><atom:link href="https://deploy-preview-1007--ucsc-ospo.netlify.app/author/yang-wang/index.xml" rel="self" type="application/rss+xml"/><description>Yang Wang</description><generator>Wowchemy (https://wowchemy.com)</generator><language>en-us</language><image><url>https://deploy-preview-1007--ucsc-ospo.netlify.app/author/yang-wang/avatar_hufd3dd55e0f6ca5584bbb305ad024a94d_101722_270x270_fill_q75_lanczos_center.jpg</url><title>Yang Wang</title><link>https://deploy-preview-1007--ucsc-ospo.netlify.app/author/yang-wang/</link></image><item><title>ScaleRep: Reproducing and benchmarking scalability bugs hiding in cloud systems</title><link>https://deploy-preview-1007--ucsc-ospo.netlify.app/project/osre24/osu/scalerep/</link><pubDate>Sat, 10 Feb 2024 00:00:00 +0000</pubDate><guid>https://deploy-preview-1007--ucsc-ospo.netlify.app/project/osre24/osu/scalerep/</guid><description>&lt;p>&lt;strong>Topics:&lt;/strong> Distributed systems, Scalability, Bug analysis, Bug reproducibility&lt;br>
&lt;strong>Skills:&lt;/strong> Java, Python, bash scripting, perf, Linux internals&lt;br>
&lt;strong>Difficulty:&lt;/strong> Hard&lt;br>
&lt;strong>Size:&lt;/strong> Large (350 hours)&lt;br>
&lt;strong>Mentors:&lt;/strong> &lt;strong>&lt;a href="https://deploy-preview-1007--ucsc-ospo.netlify.app/author/bogdan-bo-stoica/">Bogdan &amp;quot;Bo&amp;quot; Stoica&lt;/a> (contact person)&lt;/strong>, &lt;a href="https://deploy-preview-1007--ucsc-ospo.netlify.app/author/yang-wang/">Yang Wang&lt;/a>&lt;/p>
&lt;p>&lt;strong>Project Idea Description&lt;/strong>&lt;/p>
&lt;p>Large-scale distributed systems are integral to the infrastructure of a wide range of applications and services.
The continuous evolution of these systems requires ongoing efforts to address inherent faults which span a variety of issues including availability, consistency, concurrency, configuration, durability, error-handling, integrity, performance, and security.
Recent developments in the field and the rise of cloud computing have been marked by a notable increase in the scale at which such systems operate.&lt;/p>
&lt;p>This increase in scale introduces specific challenges, particularly in terms of system reliability and performance.
As distributed systems expand beyond single machines, addressing the growing demands for computation, memory and storage becomes more difficult.
This underlying complexity leads to the emergence of scalability bugs — defects that surface in large-scale deployments, yet do not reveal themselves in a small-scale setting.&lt;/p>
&lt;p>To better understand scalability bugs, we set out to investigate a set of scalability issues documented over the last 5 years from 10 popular open-source large-scale systems.
These bugs have led to significant operational challenges, such as system downtime, reduced responsiveness, data loss, and data corruption.
Moreover, addressing them required extensive collaboration and problem-solving efforts among engineers and bug reporters, with discussions often spanning a month or more.&lt;/p>
&lt;p>We observed that traditional bug finding techniques are insufficient for detecting scalability bugs since these defects are triggered by a mixture of scale-related aspects not properly investigated by previous approaches.
These characteristics include the number of components involved, the system load and workload size, the reliability of recovery protocols, and the magnitude of intermediate failures.
Although previous research examined some of these aspects, it has typically done so either in isolation (individually), or without providing a comprehensive understanding of the fundamental bug patterns, symptoms, root causes, fixes, and, more importantly, how easily these bugs can be reproduced in-house.&lt;/p>
&lt;p>Therefore, the main goal of this project is to systematically understand, characterize, and document the challenges associated with scalability bugs, at-large.
Our approach is twofold: first, to analyze scalability bugs in terms of reproducibility, and second, to develop methodologies for triggering them and measuring their impact.
Specifically, we aim to:&lt;/p>
&lt;ol>
&lt;li>Provide detailed accounts of bug reproduction experiences for a diverse set of recently reported scalability bugs from our benchmark applications;&lt;/li>
&lt;li>Identify specific challenges that prevent engineers from reproducing certain scalability bugs and investigate how prevalent these obstacles are;&lt;/li>
&lt;li>Create a suite of protocols to effectively trigger and quantify the impact of scalability bugs, facilitating their investigation in smaller-scale environments.&lt;/li>
&lt;/ol>
&lt;p>&lt;strong>Project Deliverable&lt;/strong>&lt;/p>
&lt;ul>
&lt;li>A set of Trovi replayable artifacts enabling other researchers to easily reproduce scalability bugs for our benchmark applications;&lt;/li>
&lt;li>A set of Jupyter notebook scripts allowing to conveniently replay each step in our investigation;&lt;/li>
&lt;li>A detailed breakdown of the challenges faced when reproducing scalability bugs and how these obstacles differ from those related to more “traditional” types of bugs.&lt;/li>
&lt;/ul></description></item><item><title>Is Reproducibility Enough? Understanding the Impact of Missing Settings in Artifact Evaluation</title><link>https://deploy-preview-1007--ucsc-ospo.netlify.app/project/osre23/osu/missingsettings/</link><pubDate>Wed, 08 Feb 2023 00:00:00 +0000</pubDate><guid>https://deploy-preview-1007--ucsc-ospo.netlify.app/project/osre23/osu/missingsettings/</guid><description>&lt;p>While Artifact Evaluation tries to ensure that the evaluation results in a paper are reproducible, it leaves one question: How about experiment settings NOT reported by the paper? Such “missing settings” may create multiple problems: 1) sometimes the artifacts simply do not work under these missing settings, creating problems when a later work needs to compare to an earlier work under these settings; 2) sometimes the artifacts do not perform well under these missing settings, which may create a bias during the evaluation; 3) to improve the artifact to work under these missing settings, sometimes one needs to re-design the system, which may change the results of the original experiments.&lt;/p>
&lt;p>In this project, we plan to understand the impact of this problem: On the necessity side, how would these missing settings affect the conclusions of the original work? On the feasibility side, how much effort does it require to carry out extensive experiments? We plan to answer these questions by reproducing prior works, running them on popular settings that are not reported by these works, and fixing problems if any.&lt;/p>
&lt;h3 id="measuring-research-prototypes-under-unreported-settings">Measuring Research Prototypes under Unreported Settings&lt;/h3>
&lt;p>&lt;strong>Topics:&lt;/strong> reproducibility, databases, key-value stores, DNN training&lt;br>
&lt;strong>Skills:&lt;/strong> Java/Python, Linux, TPC/YCSB&lt;br>
&lt;strong>Difficulty:&lt;/strong> Medium&lt;br>
&lt;strong>Size:&lt;/strong> 350 hours&lt;br>
&lt;strong>Mentor(s):&lt;/strong> &lt;a href="https://deploy-preview-1007--ucsc-ospo.netlify.app/author/yang-wang/">Yang Wang&lt;/a>, &lt;a href="https://deploy-preview-1007--ucsc-ospo.netlify.app/author/miao-yu/">Miao YU&lt;/a>&lt;br>
&lt;strong>Contributor(s):&lt;/strong> &lt;a href="https://deploy-preview-1007--ucsc-ospo.netlify.app/author/xueyuan-ren/">Xueyuan Ren&lt;/a>&lt;/p>
&lt;p>The student will first pick one or a few systems she is interested in. Then she will first try to reproduce their reported results. If successful, she will further try to measure these systems under previously unreported settings. During the procedure, she will need to diagnose and fix any problems that may show up. Finally, she will analyze whether the original conclusions still hold under these new settings and whether fixing any problems will change the performance characteristics of the target systems.&lt;/p></description></item></channel></rss>