<?xml version="1.0" encoding="utf-8" standalone="yes"?><rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom"><channel><title>Ray Andrew Sinurat | UCSC OSPO</title><link>https://deploy-preview-1007--ucsc-ospo.netlify.app/author/ray-andrew-sinurat/</link><atom:link href="https://deploy-preview-1007--ucsc-ospo.netlify.app/author/ray-andrew-sinurat/index.xml" rel="self" type="application/rss+xml"/><description>Ray Andrew Sinurat</description><generator>Wowchemy (https://wowchemy.com)</generator><language>en-us</language><image><url>https://deploy-preview-1007--ucsc-ospo.netlify.app/author/ray-andrew-sinurat/avatar_hu27b2d71dcd98ed071fb262ba8d61c8fa_1101814_270x270_fill_q75_lanczos_center.jpg</url><title>Ray Andrew Sinurat</title><link>https://deploy-preview-1007--ucsc-ospo.netlify.app/author/ray-andrew-sinurat/</link></image><item><title>LAST: Let’s Adapt to System Drift</title><link>https://deploy-preview-1007--ucsc-ospo.netlify.app/project/osre24/anl/last/</link><pubDate>Wed, 07 Feb 2024 00:00:00 +0000</pubDate><guid>https://deploy-preview-1007--ucsc-ospo.netlify.app/project/osre24/anl/last/</guid><description>&lt;p>&lt;strong>Project Idea Description&lt;/strong>&lt;/p>
&lt;ul>
&lt;li>&lt;strong>Topics:&lt;/strong> Computer systems, machine learning&lt;/li>
&lt;li>&lt;strong>Skills:&lt;/strong> Python, PyTorch, Bash scripting, Linux, Data Science and Machine Learning&lt;/li>
&lt;li>&lt;strong>Difficulty:&lt;/strong> Hard&lt;/li>
&lt;li>&lt;strong>Size:&lt;/strong> Large (350 hours)&lt;/li>
&lt;li>&lt;strong>Mentors:&lt;/strong> &lt;a href="https://deploy-preview-1007--ucsc-ospo.netlify.app/author/ray-andrew-sinurat/">Ray Andrew Sinurat&lt;/a> (primary contact), &lt;a href="https://deploy-preview-1007--ucsc-ospo.netlify.app/author/sandeep-madireddy/">Sandeep Madireddy&lt;/a>&lt;/li>
&lt;/ul>
&lt;p>The performance of computer systems is constantly evolving, a natural outcome of updating hardware, improving software, and encountering hardware quirks over time. At the same time, machine learning (ML) models are becoming increasingly popular. They are being used widely to address various challenges in computer systems, notably in speeding up decision-making. This speed is vital for a quick and flexible response, essential for meeting service-level agreements (SLAs). Yet, an interesting twist has emerged: like the computer systems they aid, ML models also experience a kind of &amp;ldquo;aging.&amp;rdquo; This results in a gradual decline in their effectiveness, a consequence of changes in their operating environment.&lt;/p>
&lt;p>The phenomenon of model &amp;ldquo;aging&amp;rdquo; is a ubiquitous occurrence across various domains, not limited merely to computer systems. This process of aging can significantly impact the performance of a model, emphasizing the critical importance of early detection mechanisms to maintain optimal functionality. In light of this, numerous strategies have been formulated to mitigate the aging of models. However, the generalizability and effectiveness of these strategies across diverse domains, particularly in computer systems, remain largely unexplored. This research aims to bridge this gap by designing and implementing a comprehensive data analysis pipeline. The primary objective is to evaluate the efficacy of various strategies through a comparative analysis, focusing on their performance in detecting and addressing model aging. To achieve a better understanding of this issue, the research will address the following pivotal questions:&lt;/p>
&lt;ul>
&lt;li>&lt;strong>Data-Induced Model Aging&lt;/strong>: What specific variations within the data can precipitate the aging of a model? Understanding the nature and characteristics of data changes that lead to model deterioration is crucial for developing effective prevention and mitigation strategies.&lt;/li>
&lt;li>&lt;strong>Efficacy of Aging Detection Algorithms&lt;/strong>: How proficient are the current algorithms in identifying the signs of model aging? Assessing the accuracy and reliability of these algorithms will provide insights into their practical utility in real-world scenarios.&lt;/li>
&lt;li>&lt;strong>Failure Points in Detection&lt;/strong>: In what scenarios or under what data conditions do the aging detection mechanisms fail? Identifying the limitations and vulnerabilities of these algorithms is vital for refining their robustness and ensuring comprehensive coverage.&lt;/li>
&lt;li>&lt;strong>Scalability and Responsiveness&lt;/strong>: How do these algorithms perform in terms of robustness and speed, particularly when subjected to larger datasets? Evaluating the scalability and responsiveness of the algorithms will determine their feasibility and effectiveness in handling extensive and complex datasets, a common characteristic in computer systems.&lt;/li>
&lt;/ul>
&lt;p>To better understand and prevent issues related to model performance, our approach involves analyzing various datasets, both system and non-system, that have shown notable changes over time. We aim to apply machine learning (ML) models to these datasets to assess the effects of these changes on model performance. Our goal is to leverage more advanced ML techniques to create new algorithms that address these challenges effectively. This effort is expected to contribute significantly to the community, enhancing the detection of model aging and improving model performance in computer systems.&lt;/p>
&lt;p>&lt;strong>Project Deliverable&lt;/strong>&lt;/p>
&lt;ul>
&lt;li>Run pipeline on several computer systems and non-computer systems dataset&lt;/li>
&lt;li>A Trovi artifact for data preprocessing and model training shared on Chameleon Cloud&lt;/li>
&lt;li>A GitHub repository containing the pipeline source code&lt;/li>
&lt;/ul></description></item><item><title>Automatic Cluster Performance Shifts Detection Toolkit</title><link>https://deploy-preview-1007--ucsc-ospo.netlify.app/project/osre23/anl/perfdrift/</link><pubDate>Wed, 01 Feb 2023 10:15:56 -0700</pubDate><guid>https://deploy-preview-1007--ucsc-ospo.netlify.app/project/osre23/anl/perfdrift/</guid><description>&lt;p>High-performance computing (HPC) clusters typically suffer from performance degradation over time. The heterogeneous nature of clusters and the inevitable defects in various infrastructure layers will result in a harder performance prediction inside. On the other hand, when software upgrades or any such events happen, we might also observe performance improvement or degradation even though nothing changes in the hardware. Due to these uncertainties, it is necessary to send early notification to administrators of changes in cluster performance in a specific time window to inform scheduling decisions and increase cluster utilization.&lt;/p>
&lt;p>We are targeting HPC clusters that cater to heterogeneous, compute, and I/O intensive workloads that range from scientific simulation to AI model training that have high degree of parallelization in their workloads. In this scenario, we plan to use the Darshan open-source toolkit (&lt;a href="https://github.com/darshan-hpc/darshan" target="_blank" rel="noopener">https://github.com/darshan-hpc/darshan&lt;/a>) as data collection or profiling tools to design our performance drift algorithms. Furthermore, we will possibly incorporate the distribution shift detection into Darshan, making it viable as a notification to the HPC system administrators.&lt;/p>
&lt;p>Our goal is to show the efficacy of our algorithm by plotting the profiling data that display specific time windows where the performance shifts happened after being processed by our algorithm. Finally, we will package all our profiling data and experiment scripts inside Jupyter notebook, especially Chameleon Trovi, to help others reproduce our experiments.&lt;/p>
&lt;p>Through this research, we seek to contribute the following:&lt;/p>
&lt;ul>
&lt;li>Designing an algorithm to detect performance shifts in HPC clusters that can be adapted for heterogeneous workloads&lt;/li>
&lt;li>Real-time detection of the performance shifts without introducing great overheads into the system&lt;/li>
&lt;li>Contribute to Darshan to be able to automatically detect performance changes while profiling the clusters.&lt;/li>
&lt;/ul>
&lt;h3 id="automatic-and-adaptive-performance-shifts-detection">Automatic and Adaptive Performance Shifts Detection&lt;/h3>
&lt;ul>
&lt;li>&lt;strong>Topics:&lt;/strong> Statistical Machine Learning, Deep Learning, and High-Performance Computing (HPC)&lt;/li>
&lt;li>&lt;strong>Skills:&lt;/strong> C++, Python, Statistics, good to have: Machine Learning, Deep learning&lt;/li>
&lt;li>&lt;strong>Difficulty:&lt;/strong> Moderate&lt;/li>
&lt;li>&lt;strong>Size:&lt;/strong> Large (350 hours)&lt;/li>
&lt;li>&lt;strong>Mentors:&lt;/strong> Sandeep Madireddy (&lt;a href="https://www.anl.gov/profile/sandeep-r-madireddy" target="_blank" rel="noopener">https://www.anl.gov/profile/sandeep-r-madireddy&lt;/a>, &lt;a href="http://www.mcs.anl.gov/~smadireddy/" target="_blank" rel="noopener">http://www.mcs.anl.gov/~smadireddy/&lt;/a> ), Ray Andrew Sinurat (&lt;a href="https://rayandrew.me" target="_blank" rel="noopener">https://rayandrew.me&lt;/a>)&lt;/li>
&lt;li>&lt;strong>Contributor(s):&lt;/strong> &lt;a href="https://deploy-preview-1007--ucsc-ospo.netlify.app/author/kangrui-wang/">Kangrui Wang&lt;/a>&lt;/li>
&lt;/ul>
&lt;p>All in all, these are the specific tasks that the student should do:&lt;/p>
&lt;ul>
&lt;li>Collaborate and work with mentors to understand the goal of this project.&lt;/li>
&lt;li>Implement distribution shift detection in pure statistical or machine/deep learning&lt;/li>
&lt;li>Deploy the algorithm and try to see its efficacy in the clusters.&lt;/li>
&lt;li>Package this experiment to make it easier for others to reproduce&lt;/li>
&lt;/ul></description></item></channel></rss>