<?xml version="1.0" encoding="utf-8" standalone="yes"?><rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom"><channel><title>Fraida Fund | UCSC OSPO</title><link>https://deploy-preview-1007--ucsc-ospo.netlify.app/author/fraida-fund/</link><atom:link href="https://deploy-preview-1007--ucsc-ospo.netlify.app/author/fraida-fund/index.xml" rel="self" type="application/rss+xml"/><description>Fraida Fund</description><generator>Wowchemy (https://wowchemy.com)</generator><language>en-us</language><image><url>https://deploy-preview-1007--ucsc-ospo.netlify.app/author/fraida-fund/avatar_hu6e2d1061c49df606882c6c830bd8f692_83777_270x270_fill_q75_lanczos_center.jpg</url><title>Fraida Fund</title><link>https://deploy-preview-1007--ucsc-ospo.netlify.app/author/fraida-fund/</link></image><item><title>Applying MLOps to overcome reproducibility barriers in machine learning research</title><link>https://deploy-preview-1007--ucsc-ospo.netlify.app/project/osre25/nyu/mlops/</link><pubDate>Sat, 01 Mar 2025 00:00:00 +0000</pubDate><guid>https://deploy-preview-1007--ucsc-ospo.netlify.app/project/osre25/nyu/mlops/</guid><description>&lt;ul>
&lt;li>&lt;strong>Topics:&lt;/strong> machine learning, MLOps, reproducibility&lt;/li>
&lt;li>&lt;strong>Skills:&lt;/strong> Python, machine learning, GitOps, systems, Linux, data, Docker&lt;/li>
&lt;li>&lt;strong>Difficulty:&lt;/strong> Hard&lt;/li>
&lt;li>&lt;strong>Size:&lt;/strong> Large (350 hours)&lt;/li>
&lt;li>&lt;strong>Mentors:&lt;/strong> &lt;a href="https://deploy-preview-1007--ucsc-ospo.netlify.app/author/fraida-fund/">Fraida Fund&lt;/a> and &lt;a href="https://deploy-preview-1007--ucsc-ospo.netlify.app/author/mohamed-saeed/">Mohamed Saeed&lt;/a>&lt;/li>
&lt;/ul>
&lt;p>&lt;strong>Project Idea Description&lt;/strong>&lt;/p>
&lt;p>Reproducibility remains a significant problem in machine learning research, both in core ML and in the application of ML to other areas of science. In many cases, due to inadequate experiment tracking, dependency capturing, source code versioning, data versioning, and artifact sharing, even the authors of a paper may find it challenging to reproduce their own study several years later. This makes it difficult to vaidate and build on previous work, and raises concerns about its trustworthiness.&lt;/p>
&lt;p>In contrast, outside of academic research, MLOps tools and frameworks have been identified as a key enabler of reliable, reproducible, and trustworthy machine learning systems in production. A good reference on this topic is:&lt;/p>
&lt;blockquote>
&lt;p>Firas Bayram and Bestoun S. Ahmed. 2025. Towards Trustworthy Machine Learning in Production: An Overview of the Robustness in MLOps Approach. ACM Comput. Surv. 57, 5, Article 121 (May 2025), 35 pages. &lt;a href="https://doi.org/10.1145/3708497" target="_blank" rel="noopener">https://doi.org/10.1145/3708497&lt;/a>&lt;/p>
&lt;/blockquote>
&lt;p>This project seeks to bridge the gap between widely adopted practices in industry and academic research:&lt;/p>
&lt;ul>
&lt;li>by making it easier for researchers and scientists to use MLOps tools to support reproducibility. To achieve this, we will develop starter templates and recipes for research in computer vision, NLP, and ML for science, that have reproducibility &amp;ldquo;baked in&amp;rdquo; thanks to the integration of MLOps tools and frameworks. Researchers will launch these templates on open access research facilities like &lt;a href="https://chameleoncloud.org/" target="_blank" rel="noopener">Chameleon&lt;/a>.&lt;/li>
&lt;li>and, by developing complementary education and training materials to emphasize the important of reproducibility in ML, and how the tools and frameworks used in the starter templates can support this goal.&lt;/li>
&lt;/ul>
&lt;p>&lt;strong>Writing a successful proposal for this project&lt;/strong>&lt;/p>
&lt;p>A good proposal for this project should -&lt;/p>
&lt;ul>
&lt;li>demonstrate a good understanding of the current barriers to reproducibility in machine learning research (specific examples are welcome),&lt;/li>
&lt;li>describe a &amp;ldquo;base&amp;rdquo; starter template, including the platforms and tools that will be integrated, as well as specific adaptations of this template for computer vision, NLP, and ML for science,&lt;/li>
&lt;li>explain the &amp;ldquo;user flow&amp;rdquo; - how a researcher would use the template to conduct an experiment or series of experiments, what the lifecycle of that experiment would look like, and how it would be made reproducible,&lt;/li>
&lt;li>include the contributor&amp;rsquo;s own ideas about how to make the starter templates more usable, and how to make the education and training materials relatable and useful,&lt;/li>
&lt;li>and show that the contributor has the necessary technical background and soft skills to contribute to this project. In particular, the contributor will need to create education and training materials that are written in a clear, straightforward, and concise manner, without unncessary jargon. The proposal should show evidence of the contributor&amp;rsquo;s writing abilities.&lt;/li>
&lt;/ul>
&lt;p>&lt;strong>Github link&lt;/strong>&lt;/p>
&lt;p>There is no pre-existing Git repository for this project - at the beginning of the summer, the contributor will create a new repository in the &lt;a href="https://github.com/teaching-on-testbeds/" target="_blank" rel="noopener">Teaching on Testbeds&lt;/a> organization, and the project materials will &amp;ldquo;live&amp;rdquo; there.&lt;/p></description></item><item><title>Data leakage in applied ML: reproducing examples of irreproducibility</title><link>https://deploy-preview-1007--ucsc-ospo.netlify.app/project/osre24/nyu/data-leakage/</link><pubDate>Wed, 21 Feb 2024 00:00:00 +0000</pubDate><guid>https://deploy-preview-1007--ucsc-ospo.netlify.app/project/osre24/nyu/data-leakage/</guid><description>&lt;ul>
&lt;li>&lt;strong>Topics:&lt;/strong> applied machine learning, data leakage, reproducibility&lt;/li>
&lt;li>&lt;strong>Skills:&lt;/strong> Python, data analysis, machine learning&lt;/li>
&lt;li>&lt;strong>Difficulty:&lt;/strong> Medium&lt;/li>
&lt;li>&lt;strong>Size:&lt;/strong> Large (350 hours)&lt;/li>
&lt;li>&lt;strong>Mentors:&lt;/strong> &lt;a href="https://deploy-preview-1007--ucsc-ospo.netlify.app/author/fraida-fund/">Fraida Fund&lt;/a> and &lt;a href="https://deploy-preview-1007--ucsc-ospo.netlify.app/author/mohamed-saeed/">Mohamed Saeed&lt;/a>&lt;/li>
&lt;/ul>
&lt;p>&lt;strong>Project Idea Description&lt;/strong>&lt;/p>
&lt;p>Data leakage &lt;a href="https://www.cell.com/patterns/pdfExtended/S2666-3899%2823%2900159-9" target="_blank" rel="noopener">has been identified&lt;/a> as a major cause of irreproducibility of a paper&amp;rsquo;s findings, when machine learning techniques are applied to problems in science. Data leakage includes errors such as:&lt;/p>
&lt;ul>
&lt;li>pre-processing before splitting into training/test sets&lt;/li>
&lt;li>feature selection before splitting into training/test sets&lt;/li>
&lt;li>duplicated data points in both training and test sets&lt;/li>
&lt;li>temporal leakage (e.g. shuffled K-fold cross validation with temporal data)&lt;/li>
&lt;li>group leakage (e.g. shuffled K-fold cross validation with data that has group structure)&lt;/li>
&lt;/ul>
&lt;p>and leads to an overly optimistic evaluation of model performance, such that the finding may no longer be the same when the error is corrected.&lt;/p>
&lt;p>Despite the seriousness of this problem, data leakage is often not covered in introductory machine learning courses, and many users of machine learning across varied science domains are unaware of it. Even those who have learned &amp;ldquo;rules&amp;rdquo; for avoiding data leakage (e.g. &amp;ldquo;never do feature selection on the test set&amp;rdquo;) may not understand the reasons for these &amp;ldquo;rules&amp;rdquo;, and how important they are for ensuring that the final result is valid and reproducible.&lt;/p>
&lt;p>The goal of this project is to create &lt;em>learning materials&lt;/em> demonstrating how instances of data leakage invalidate a result. These materials should be easily adoptable by instructors teaching machine learning in a wide variety of contexts, including those teaching a non-CS audience. To achieve this, the project proposes to re-implement published results that have been affected by data leakage, and package these implementations along with supporting material in a format suitable for use in classrooms and by independent learners. For each &amp;ldquo;irreproducible result&amp;rdquo;, the &amp;ldquo;package&amp;rdquo; should include -&lt;/p>
&lt;ul>
&lt;li>a re-implementation of the original result&lt;/li>
&lt;li>an explanation of the data leakage problem affecting the result, with an implementation of a &amp;ldquo;toy example&amp;rdquo; on synthetic data&lt;/li>
&lt;li>a re-implementation of the result without the data analysis error, to show how the finding is affected&lt;/li>
&lt;li>and examples of exam or homework questions that an instructor adopting this package may use to assess understanding.&lt;/li>
&lt;/ul>
&lt;p>&lt;strong>Writing a successful proposal for this project&lt;/strong>&lt;/p>
&lt;p>A good proposal for this project should include, for at least a few &amp;ldquo;types&amp;rdquo; of data leakage mentioned above -&lt;/p>
&lt;ul>
&lt;li>a specific published result that could be used as an exemplar (you may find ideas among the review papers listed &lt;a href="https://reproducible.cs.princeton.edu/#rep-failures" target="_blank" rel="noopener">here&lt;/a>)&lt;/li>
&lt;li>a brief description of the details of the experiment that will reproduce that result (e.g. what data is used, what machine learning technique is used, what are the hyperparameters used for training)&lt;/li>
&lt;li>and an explanation of why this result is suitable for this use (it uses a publicly available dataset, a machine learning technique that is familiar and accessible to students in an introductory course, the paper has sufficient detail to reproduce the result, etc.)&lt;/li>
&lt;/ul>
&lt;p>The contributor will need to create learning materials that are written in a clear, straightforward, and concise manner, without unncessary jargon. The proposal should show evidence of the contributor&amp;rsquo;s writing abilities.&lt;/p>
&lt;p>&lt;strong>Github link&lt;/strong>&lt;/p>
&lt;p>There is no pre-existing Git repository for this project - at the beginning of the summer, the contributor will create a new repository in the &lt;a href="https://github.com/teaching-on-testbeds/" target="_blank" rel="noopener">Teaching on Testbeds&lt;/a> organization, and the project materials will &amp;ldquo;live&amp;rdquo; there.&lt;/p>
&lt;p>To get a sense of the type of code you would be writing, here is an example of a learning module related to data leakage (however, it is not in the format described above): &lt;a href="https://colab.research.google.com/github/ffund/ml-notebooks/blob/master/notebooks/4-linear-regression-case-study-part-2.ipynb" target="_blank" rel="noopener">Beauty in the Classroom&lt;/a>&lt;/p>
&lt;p>&lt;strong>Project Deliverables&lt;/strong>&lt;/p>
&lt;ul>
&lt;li>&amp;ldquo;Packages&amp;rdquo; of learning materials for teaching about common types of data leakage&lt;/li>
&lt;li>&lt;a href="https://chameleoncloud.org/experiment/share/" target="_blank" rel="noopener">Trovi&lt;/a> artifacts for &amp;ldquo;playing back&amp;rdquo; each of the &amp;ldquo;packages&amp;rdquo;&lt;/li>
&lt;/ul></description></item><item><title>Evaluating congestion controls past and future</title><link>https://deploy-preview-1007--ucsc-ospo.netlify.app/project/osre24/nyu/congestion-control/</link><pubDate>Wed, 21 Feb 2024 00:00:00 +0000</pubDate><guid>https://deploy-preview-1007--ucsc-ospo.netlify.app/project/osre24/nyu/congestion-control/</guid><description>&lt;ul>
&lt;li>&lt;strong>Topics:&lt;/strong> computer networks, congestion control, reproducibility&lt;/li>
&lt;li>&lt;strong>Skills:&lt;/strong> Python, Bash scripting, Linux, computer network performance evaluation&lt;/li>
&lt;li>&lt;strong>Difficulty:&lt;/strong> Medium&lt;/li>
&lt;li>&lt;strong>Size:&lt;/strong> Large (350 hours)&lt;/li>
&lt;li>&lt;strong>Mentors:&lt;/strong> &lt;a href="https://deploy-preview-1007--ucsc-ospo.netlify.app/author/fraida-fund/">Fraida Fund&lt;/a> and &lt;a href="https://deploy-preview-1007--ucsc-ospo.netlify.app/author/ashutosh-srivastava/">Ashutosh Srivastava&lt;/a>&lt;/li>
&lt;/ul>
&lt;p>&lt;strong>Project Idea Description&lt;/strong>&lt;/p>
&lt;p>In computer networks, congestion control protocols play an outsize role in determining our experience with networked applications. New congestion control algorithms are regularly proposed by researchers to improve throughput and latency performance, adapt to new types of networks, and align more closely with the needs of new applications.&lt;/p>
&lt;p>However, our understanding of the benefits of a new congestion control protocol depends to a large extent on the evaluation - the network topology, the network delay and throughput, the type of flow, the type of competing traffic - and there is no single standard way to evaluate a congestion control protocol. The &lt;a href="https://pantheon.stanford.edu/static/pantheon/documents/pantheon-paper.pdf" target="_blank" rel="noopener">Pantheon&lt;/a> project (which is no longer supported) sought to fill this gap somewhat and address the problem of reproducibility of congestion control results, but their approach is not easily adapted to evaluation scenarios representative of new types of applications or networks. Nor is it capable of representing the evaluation scenarios in most published results related to congestion control.&lt;/p>
&lt;p>The goal of this project, therefore is to create an evaluation suite for congestion control protocols that can be used to reproduce existing congestion control results in the academic literature, &lt;em>and&lt;/em> to evaluate new protocols under similar evaluation conditions, &lt;em>and&lt;/em> to be easily extended to new scenarios. An &amp;ldquo;evaluation scenario&amp;rdquo; includes:&lt;/p>
&lt;ul>
&lt;li>a Python notebook to realize the network topology on the FABRIC and/or Chameleon testbed, and configure the network characteristics,&lt;/li>
&lt;li>scripts to generate the data flow(s) needed for the evaluation,&lt;/li>
&lt;li>and scripts to capture data from the experiment and visualize the results.&lt;/li>
&lt;/ul>
&lt;p>&lt;strong>Writing a successful proposal for this project&lt;/strong>&lt;/p>
&lt;p>To write a good proposal for this project, you should review the most influential papers on TCP congestion control, and especially those related to TCP protocols that are available in the Linux kernel.&lt;/p>
&lt;p>Use your findings to explain what your proposed evaluation suite will include (what network topologies, what flow generators), and justify this with reference to the academic literature. Also indicate which &lt;em>specific results&lt;/em> you expect to be able to reproduce using this suite (e.g. include figures from influential papers showing evaluation results! with citation, of course).&lt;/p>
&lt;p>You can also take advantage of existing open source code that reproduces a congestion control result, e.g. &lt;a href="https://github.com/sdatta97/imcbbrrepro" target="_blank" rel="noopener">Replication: When to Use and When Not to Use BBR&lt;/a>, or &lt;a href="https://github.com/ashutoshs25/bbr-dominance-experiments" target="_blank" rel="noopener">Some of the Internet may be heading towards BBR dominance: an experimental study&lt;/a>.&lt;/p>
&lt;p>&lt;strong>Github link&lt;/strong>&lt;/p>
&lt;p>There is no pre-existing Git repository for this project - at the beginning of the summer, the contributor will create a new repository for this project.&lt;/p>
&lt;p>&lt;strong>Project Deliverables&lt;/strong>&lt;/p>
&lt;ul>
&lt;li>&amp;ldquo;Packages&amp;rdquo; of evaluation scenarios that can be used to evaluate a congestion control algorithm implemented in the Linux kernel&lt;/li>
&lt;li>&lt;a href="https://chameleoncloud.org/experiment/share/" target="_blank" rel="noopener">Trovi&lt;/a> artifacts for realizing each evaluation scenario on Chameleon&lt;/li>
&lt;/ul></description></item><item><title>Reproducible Evaluation of Multipath Network Protocols</title><link>https://deploy-preview-1007--ucsc-ospo.netlify.app/project/osre23/farmingdale/multipath/</link><pubDate>Thu, 16 Feb 2023 00:00:00 +0000</pubDate><guid>https://deploy-preview-1007--ucsc-ospo.netlify.app/project/osre23/farmingdale/multipath/</guid><description>&lt;p>Lead Mentor: &lt;a href="mailto:aydini@farmingdale.edu">Ilknur Aydin&lt;/a>&lt;/p>
&lt;p>As mobile devices with dual WiFi and cellular interfaces become widespread, network protocols have been developed that utilize the availability of multiple paths. However, the relative effectiveness of these protocols is highly dependent on the characteristics of the network (including the relationship between the two paths, which are often not independent). Researchers typically evaluate a multipath protocol for a small set of network scenarios, which vary from one publication to the next. It is therefore difficult to get a good picture of how different protocols perform in a range of settings.&lt;/p>
&lt;h3 id="framework-for-repeatable-direct-comparison-of-multipath-transport-protocols">Framework for repeatable, direct comparison of multipath transport protocols&lt;/h3>
&lt;ul>
&lt;li>&lt;strong>Topics&lt;/strong>: Computer networks, wireless systems&lt;/li>
&lt;li>&lt;strong>Skills&lt;/strong>: Linux, networking, data analysis and visualization, writing&lt;/li>
&lt;li>&lt;strong>Difficulty&lt;/strong>: Large&lt;/li>
&lt;li>&lt;strong>Size&lt;/strong>: 350 hours&lt;/li>
&lt;li>&lt;strong>Mentor(s)&lt;/strong>: &lt;a href="mailto:aydini@farmingdale.edu">Ilknur Aydin&lt;/a> and &lt;a href="mailto:ffund@nyu.edu">Fraida Fund&lt;/a>&lt;/li>
&lt;/ul>
&lt;p>In single-path congestion control, the &lt;a href="https://pantheon.stanford.edu/" target="_blank" rel="noopener">Pantheon&lt;/a> work created a reference set of executable benchmarks that researchers could use to evaluate novel congestion control designs against existing work in a wide range of the scenarios. This project seeks to achieve something similar for multipath protocols, using publicly available networking testbeds like &lt;a href="https://fabric-testbed.net/" target="_blank" rel="noopener">FABRIC&lt;/a>. For this project, the participant will:&lt;/p>
&lt;ul>
&lt;li>Prepare a set of network benchmarks for multipath protocols, using live network links, real link traces, and emulated scenarios&lt;/li>
&lt;li>Develop an experiment using the benchmarks to evaluate existing multipath protocol implementations&lt;/li>
&lt;li>Prepare materials that researchers can use to evaluate novel multipath protocols against the others in the benchmark&lt;/li>
&lt;/ul></description></item><item><title>Teaching Computer Networks with Reproducible Research</title><link>https://deploy-preview-1007--ucsc-ospo.netlify.app/project/osre23/nyu/edunet/</link><pubDate>Wed, 18 Jan 2023 00:00:00 +0000</pubDate><guid>https://deploy-preview-1007--ucsc-ospo.netlify.app/project/osre23/nyu/edunet/</guid><description>&lt;p>Lead Mentor: &lt;a href="https://deploy-preview-1007--ucsc-ospo.netlify.app/author/fraida-fund/">Fraida Fund&lt;/a>&lt;/p>
&lt;p>In the field of computer networks and wireless communication systems, the availability of open access networking and cloud computing testbeds (&lt;a href="https://portal.geni.net/" target="_blank" rel="noopener">GENI&lt;/a>, &lt;a href="https://cloudlab.us/" target="_blank" rel="noopener">CloudLab&lt;/a>, &lt;a href="https://chameleoncloud.org/" target="_blank" rel="noopener">Chameleon&lt;/a>, &lt;a href="https://fabric-testbed.net/" target="_blank" rel="noopener">FABRIC&lt;/a>, and others) has been transformative in promoting reproducible research &lt;em>and&lt;/em> in making high-quality experiential learning available to students and educators at a wide range of colleges and universities. This project seeks to unite research and education use of these testbeds by developing new ways of using reproducible research to teach computer networks and related topics.&lt;/p>
&lt;h3 id="bringing-foundational-results-into-the-classroom">Bringing foundational results into the classroom&lt;/h3>
&lt;ul>
&lt;li>&lt;strong>Topics&lt;/strong>: Computer networks, reproducibility, education&lt;/li>
&lt;li>&lt;strong>Skills&lt;/strong>: Linux, writing&lt;/li>
&lt;li>&lt;strong>Difficulty&lt;/strong>: Medium&lt;/li>
&lt;li>&lt;strong>Size&lt;/strong>: 350 hours&lt;/li>
&lt;li>&lt;strong>Mentor(s)&lt;/strong>: &lt;a href="https://deploy-preview-1007--ucsc-ospo.netlify.app/author/fraida-fund/">Fraida Fund&lt;/a> and TBD&lt;/li>
&lt;/ul>
&lt;p>To make foundational results from computer networks more concrete, this project seeks to reproduce a selection of key results and package them for use as interactive classroom demonstrations. (An example of a &amp;ldquo;foundational&amp;rdquo; result might be the result from the 1980s that motivates congestion control by showing how &lt;a href="http://dx.doi.org/10.1016/0169-7552%2889%2990019-6" target="_blank" rel="noopener">congestion collapse occurs when the network is under heavy load&lt;/a>.) This involves:&lt;/p>
&lt;ul>
&lt;li>Reproducing the original results on an open-access testbed&lt;/li>
&lt;li>Packaging the materials for use as a classroom demo, with interactive elements&lt;/li>
&lt;li>Creating assessment questions and sample &amp;ldquo;solutions&amp;rdquo; related to the materials, that instructors may use in homework assignments or exams&lt;/li>
&lt;/ul>
&lt;h3 id="developing-a-classroom-competition-for-adaptive-video-delivery-policies">Developing a &amp;ldquo;classroom competition&amp;rdquo; for adaptive video delivery policies&lt;/h3>
&lt;ul>
&lt;li>&lt;strong>Topics&lt;/strong>: Computer networks, adaptive video, reproducibility, education&lt;/li>
&lt;li>&lt;strong>Skills&lt;/strong>: Linux, Python, writing&lt;/li>
&lt;li>&lt;strong>Difficulty&lt;/strong>: Medium&lt;/li>
&lt;li>&lt;strong>Size&lt;/strong>: 350 hours&lt;/li>
&lt;li>&lt;strong>Mentor(s)&lt;/strong>: &lt;a href="https://deploy-preview-1007--ucsc-ospo.netlify.app/author/fraida-fund/">Fraida Fund&lt;/a> and TBD&lt;/li>
&lt;li>&lt;strong>Contributor(s)&lt;/strong>: &lt;a href="https://deploy-preview-1007--ucsc-ospo.netlify.app/author/srishti-jaiswal/">Srishti Jaiswal&lt;/a>&lt;/li>
&lt;/ul>
&lt;p>A carefully designed competition can be a fun and exciting way for students to challenge themselves and gain &amp;ldquo;ownership&amp;rdquo; of a new topic. This projects builds on an existing open source &lt;a href="https://witestlab.poly.edu/blog/adaptive-video-reproducing/" target="_blank" rel="noopener">reproducible result&lt;/a> for adaptive video delivery, and will challenge students to extend this work and design their own adaptive video policies for head-to-head competition against their classmates. This includes:&lt;/p>
&lt;ul>
&lt;li>Packaging the result to make it easier for students to reproduce and then build on the original work&lt;/li>
&lt;li>Implementing other adaptive video policies from the literature, so that students can use them as a baseline&lt;/li>
&lt;li>Developing different network settings (using live link traces and emulated link patterns) in which student submissions may be evaluated&lt;/li>
&lt;li>Developing an evaluation framework for scoring student submissions on different criteria and in different network settings, and making the results available in a leaderboard format&lt;/li>
&lt;/ul></description></item><item><title>Using Reproducibility in Machine Learning Education</title><link>https://deploy-preview-1007--ucsc-ospo.netlify.app/project/osre23/nyu/eduml/</link><pubDate>Wed, 18 Jan 2023 00:00:00 +0000</pubDate><guid>https://deploy-preview-1007--ucsc-ospo.netlify.app/project/osre23/nyu/eduml/</guid><description>&lt;p>Lead Mentor: &lt;a href="mailto:ffund@nyu.edu">Fraida Fund&lt;/a>&lt;/p>
&lt;p>The computer science and engineering classroom is as essential part of the reproducibility &amp;ldquo;ecosystem&amp;rdquo; - because of broad reach and potential for big impact, and because for many students, the classroom is their first exposure to research in their field. For machine learning in particular, reproducibility is an important element of the research culture, and can be a valuable part of any introductory or advanced courses in the field. These projects will develop highly interactive open educational resources, that may be adopted by instructors of graduate or undergraduate machine learning courses to incorporate more instruction about reproducibility and reproducible research.&lt;/p>
&lt;h3 id="introducing-levels-of-reproduction-and-replication-in-ml">Introducing &amp;ldquo;levels&amp;rdquo; of reproduction and replication in ML&lt;/h3>
&lt;ul>
&lt;li>&lt;strong>Topics&lt;/strong>: Machine learning, reproducibility, education&lt;/li>
&lt;li>&lt;strong>Skills&lt;/strong>: Python, machine learning, writing&lt;/li>
&lt;li>&lt;strong>Difficulty&lt;/strong>: Medium&lt;/li>
&lt;li>&lt;strong>Size&lt;/strong>: 350 hours&lt;/li>
&lt;li>&lt;strong>Mentor(s)&lt;/strong>: &lt;a href="https://deploy-preview-1007--ucsc-ospo.netlify.app/author/fraida-fund/">Fraida Fund&lt;/a> and TBD&lt;/li>
&lt;li>&lt;strong>Contributor(s)&lt;/strong>: &lt;a href="https://deploy-preview-1007--ucsc-ospo.netlify.app/author/mohamed-saeed/">Mohamed Saeed&lt;/a>&lt;/li>
&lt;/ul>
&lt;p>In machine learning, replicating a published result to confirm the validity of the experimental results and the broader conclusions of the paper can take several forms, with increasing levels of effort:&lt;/p>
&lt;ul>
&lt;li>using authors&amp;rsquo; code and pre-trained weights, run the model on the same benchmarks as the original paper&lt;/li>
&lt;li>training a model using authors&amp;rsquo; code and published hyperparameters,&lt;/li>
&lt;li>training a model using authors&amp;rsquo; code and a new hyperparamter search,&lt;/li>
&lt;li>validating the authors&amp;rsquo; code e.g. with unit tests, in addition to training,&lt;/li>
&lt;li>re-implementing the model,&lt;/li>
&lt;li>designing additional experiments to validate that the suggested mechanism is in fact responsible for the result,&lt;/li>
&lt;li>and more.&lt;/li>
&lt;/ul>
&lt;p>This project will develop interactive materials (using one or more exemplar published results) to illustrate and to highlight relevant aspects and pitfalls of each of these &amp;ldquo;levels&amp;rdquo; of reproduction and replication.&lt;/p>
&lt;h3 id="packaging-existing-reproducible-results-for-the-ml-classroom">Packaging existing reproducible results for the ML classroom&lt;/h3>
&lt;ul>
&lt;li>&lt;strong>Topics&lt;/strong>: Machine learning, reproducibility, education&lt;/li>
&lt;li>&lt;strong>Skills&lt;/strong>: Python, machine learning, writing&lt;/li>
&lt;li>&lt;strong>Difficulty&lt;/strong>: Medium&lt;/li>
&lt;li>&lt;strong>Size&lt;/strong>: 350 hours&lt;/li>
&lt;li>&lt;strong>Mentor(s)&lt;/strong>: &lt;a href="https://deploy-preview-1007--ucsc-ospo.netlify.app/author/fraida-fund/">Fraida Fund&lt;/a> and TBD&lt;/li>
&lt;li>&lt;strong>Contribuor(s)&lt;/strong>: &lt;a href="https://deploy-preview-1007--ucsc-ospo.netlify.app/author/shekhar/">Shekhar&lt;/a>, &lt;a href="https://deploy-preview-1007--ucsc-ospo.netlify.app/author/jonathan-edwin/">Jonathan Edwin&lt;/a>&lt;/li>
&lt;/ul>
&lt;p>The goal is to make it easier for instructors to expose students to state-of-the-art research in the classroom. This project will work with an existing set of recent reproducible results in machine learning, and will package them for easier consumption by students and more effective use in the classroom. This may include, but is not necessarily limited to:&lt;/p>
&lt;ul>
&lt;li>Re-validating the result and re-packaging along with computational environment on an open access testbed&lt;/li>
&lt;li>Creating tutorial material around the result, including interactive visualizations to demonstrate key elements of the work&lt;/li>
&lt;li>Creating one-click demos for applying the model/technique to a new test sample&lt;/li>
&lt;li>Curating test samples to highlight important advantages and limitations of the result&lt;/li>
&lt;li>Creating assessment questions and sample &amp;ldquo;solutions&amp;rdquo; that instructors may use to &amp;ldquo;assign&amp;rdquo; the work to students&lt;/li>
&lt;/ul></description></item></channel></rss>