<?xml version="1.0" encoding="utf-8" standalone="yes"?><rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom"><channel><title>Mohamed Saeed | UCSC OSPO</title><link>https://deploy-preview-1007--ucsc-ospo.netlify.app/author/mohamed-saeed/</link><atom:link href="https://deploy-preview-1007--ucsc-ospo.netlify.app/author/mohamed-saeed/index.xml" rel="self" type="application/rss+xml"/><description>Mohamed Saeed</description><generator>Wowchemy (https://wowchemy.com)</generator><language>en-us</language><image><url>https://deploy-preview-1007--ucsc-ospo.netlify.app/author/mohamed-saeed/avatar_hu524cde59e2d520c7f38b4fec3309afd8_122877_270x270_fill_q75_lanczos_center.jpg</url><title>Mohamed Saeed</title><link>https://deploy-preview-1007--ucsc-ospo.netlify.app/author/mohamed-saeed/</link></image><item><title>Final Blog on Using Reproducibility in Machine Learning Education</title><link>https://deploy-preview-1007--ucsc-ospo.netlify.app/report/osre23/nyu/eduml/20231018-msaeed/</link><pubDate>Wed, 18 Oct 2023 00:00:00 +0000</pubDate><guid>https://deploy-preview-1007--ucsc-ospo.netlify.app/report/osre23/nyu/eduml/20231018-msaeed/</guid><description>&lt;p>Welcome back!&lt;/p>
&lt;p>In my final blog post for the 2023 Summer of Reproducibility Fellowship, I&amp;rsquo;ll be sharing my experiences and the materials I&amp;rsquo;ve created for the &lt;a href="https://deploy-preview-1007--ucsc-ospo.netlify.app/project/osre23/nyu/eduml">Using Reproducibility in Machine Learning Education project&lt;/a>. As a quick reminder, my mentor &lt;a href="https://deploy-preview-1007--ucsc-ospo.netlify.app/author/fraida-fund/">Fraida Fund&lt;/a> and I have been working on developing interactive open-source educational resources that teach reproducibility and reproducible research in machine learning. You can find my &lt;a href="https://drive.google.com/file/d/13HnCMZawpabiLdBoOiaJFF2mNXIPLCVJ/view?usp=sharing" target="_blank" rel="noopener">proposal here&lt;/a>.&lt;/p>
&lt;p>In this post, I&amp;rsquo;ll give you a rundown of my experience and share the materials I&amp;rsquo;ve created. If you haven&amp;rsquo;t checked out my previous blog posts, definitely take a look before diving into this one. Let&amp;rsquo;s get started!&lt;/p>
&lt;h2 id="why-is-this-project-important-">Why is this project important 🤔&lt;/h2>
&lt;p>Reproducibility is an essential aspect of scientific research, and it&amp;rsquo;s becoming increasingly important in the field of computer science. However, most efforts to promote reproducibility in education focus on students who are actively involved in research, leaving a significant gap in the curriculum for introductory courses. Our project aims to address this issue by incorporating reproducibility experiences into machine learning education.&lt;/p>
&lt;h2 id="why-reproducibility-matters-in-education-">Why Reproducibility Matters in Education 🎓&lt;/h2>
&lt;p>There are two primary reasons why we believe reproducibility belongs in the computer science classroom. Firstly, it allows students to experience the process of reproducing research firsthand, giving them a deeper understanding of the scientific method and its importance in the field. This exposure can inspire students to adopt reproducible practices in their future careers, contributing to a more transparent and reliable scientific community.&lt;/p>
&lt;p>
&lt;figure >
&lt;div class="d-flex justify-content-center">
&lt;div class="w-100" >&lt;img alt="" srcset="
/report/osre23/nyu/eduml/20231018-msaeed/reproducibilityBenifits_hudb7baafb83412fd51973fc577da0863d_141778_0f4e9e6ba00e070430ccd90e09800a28.webp 400w,
/report/osre23/nyu/eduml/20231018-msaeed/reproducibilityBenifits_hudb7baafb83412fd51973fc577da0863d_141778_22ad37d3ef94bfc2aa93cf4ba651684e.webp 760w,
/report/osre23/nyu/eduml/20231018-msaeed/reproducibilityBenifits_hudb7baafb83412fd51973fc577da0863d_141778_1200x1200_fit_q75_h2_lanczos_3.webp 1200w"
src="https://deploy-preview-1007--ucsc-ospo.netlify.app/report/osre23/nyu/eduml/20231018-msaeed/reproducibilityBenifits_hudb7baafb83412fd51973fc577da0863d_141778_0f4e9e6ba00e070430ccd90e09800a28.webp"
width="760"
height="207"
loading="lazy" data-zoomable />&lt;/div>
&lt;/div>&lt;/figure>
&lt;em>Source: Fund, Fraida. &amp;ldquo;We Need More Reproducibility Content Across the Computer Science Curriculum.&amp;rdquo; Proceedings of the 2023 ACM Conference on Reproducibility and Replicability. 2023.&lt;/em>&lt;/p>
&lt;p>Secondly, as shown in the figure, involving students in reproducibility efforts can have a significant impact on the reproducibility ecosystem itself. Students can create reproducibility artifacts, such as replicable experiments or data analysis, that can be used by other researchers, including authors and graduate students. Additionally, students can consume reproducibility artifacts created by the research community, provide feedback, and suggest improvements. Authors appreciate this type of engagement, as it adds value to their work and promotes open science.&lt;/p>
&lt;h2 id="focusing-on-machine-learning-">Focusing on Machine Learning 🧐&lt;/h2>
&lt;p>Given the growing interest in machine learning and its relevance to reproducibility, our project decided to focus on this area. Machine learning already has a strong culture of reproducibility, with initiatives like &lt;a href="https://paperswithcode.com/" target="_blank" rel="noopener">Papers with Code&lt;/a> and the &lt;a href="https://paperswithcode.com/rc2022" target="_blank" rel="noopener">ML Reproducibility Challenge&lt;/a>. These efforts encourage researchers to share their code and reproduce recent machine learning papers, validating their results. By leveraging these existing resources, we can create learning materials that utilize real-world examples and foster hands-on reproducibility experiences for students.&lt;/p>
&lt;h2 id="the-interactive-notebooks-">The Interactive Notebooks 📖&lt;/h2>
&lt;p>We have created two learning materials that focus on machine learning and reproducibility. &lt;strong>The first material&lt;/strong> looks at a paper titled &lt;a href="https://arxiv.org/abs/1910.08475" target="_blank" rel="noopener">&amp;ldquo;On Warm Starting Neural Network Training&amp;rdquo;&lt;/a> by Jordan T. Ash and Ryan P. Adams. This paper discusses the concept of warm-starting, which involves using weights from a previously trained model on a subset of the dataset to train a new model. The authors compare the performance of warm-started models with randomly initialized models and find that the warm-started models perform worse as shown in the below figure.&lt;/p>
&lt;p>
&lt;figure >
&lt;div class="d-flex justify-content-center">
&lt;div class="w-100" >&lt;img alt="" srcset="
/report/osre23/nyu/eduml/20231018-msaeed/figure1_hu9d945c7dbee9ef6ef608a89a33d817c5_76602_5815498dd015ebc84b00505c90a65354.webp 400w,
/report/osre23/nyu/eduml/20231018-msaeed/figure1_hu9d945c7dbee9ef6ef608a89a33d817c5_76602_df01d4772e731cee04ae4783ac0cc994.webp 760w,
/report/osre23/nyu/eduml/20231018-msaeed/figure1_hu9d945c7dbee9ef6ef608a89a33d817c5_76602_1200x1200_fit_q75_h2_lanczos_3.webp 1200w"
src="https://deploy-preview-1007--ucsc-ospo.netlify.app/report/osre23/nyu/eduml/20231018-msaeed/figure1_hu9d945c7dbee9ef6ef608a89a33d817c5_76602_5815498dd015ebc84b00505c90a65354.webp"
width="760"
height="306"
loading="lazy" data-zoomable />&lt;/div>
&lt;/div>&lt;/figure>
&lt;/p>
&lt;p>Our material takes students through the process of identifying the different claims made in the paper and finding the corresponding experiments that support them. They will also learn how to use open-source code and available data to reproduce these experiments and understand the computational complexity associated with reproducing each experiment. This material can be found on both &lt;a href="https://github.com/mohammed183/re_warm_start_nn/tree/main" target="_blank" rel="noopener">github&lt;/a> and &lt;a href="https://chameleoncloud.org/experiment/share/5b5717df-9aa9-470f-b393-c1e189c008a8" target="_blank" rel="noopener">chameleon&lt;/a> where you can use chameleon to run the material on the required resources.&lt;/p>
&lt;p>&lt;strong>The second material&lt;/strong> examines the paper &lt;a href="https://arxiv.org/abs/2010.11929" target="_blank" rel="noopener">&amp;ldquo;An Image is Worth 16x16 Words: Transformers for Image Recognition at Scale&amp;rdquo;&lt;/a> by Dosovitskiy et al., which introduces a novel way of applying the transformer architecture, which was originally designed for natural language processing, to image recognition tasks. The paper shows that transformers can achieve state-of-the-art results on several image classification benchmarks, such as ImageNet, when trained on large-scale datasets as shown in the following table.&lt;/p>
&lt;p>
&lt;figure >
&lt;div class="d-flex justify-content-center">
&lt;div class="w-100" >&lt;img alt="" srcset="
/report/osre23/nyu/eduml/20231018-msaeed/table1_hu639b2ac18dac1313dd35f10cc0ae8db7_237634_7faf7451ff08ac87e7d12ab941c77f8e.webp 400w,
/report/osre23/nyu/eduml/20231018-msaeed/table1_hu639b2ac18dac1313dd35f10cc0ae8db7_237634_5441e5e4c6ffed9b29244a3a3dcde852.webp 760w,
/report/osre23/nyu/eduml/20231018-msaeed/table1_hu639b2ac18dac1313dd35f10cc0ae8db7_237634_1200x1200_fit_q75_h2_lanczos_3.webp 1200w"
src="https://deploy-preview-1007--ucsc-ospo.netlify.app/report/osre23/nyu/eduml/20231018-msaeed/table1_hu639b2ac18dac1313dd35f10cc0ae8db7_237634_7faf7451ff08ac87e7d12ab941c77f8e.webp"
width="760"
height="354"
loading="lazy" data-zoomable />&lt;/div>
&lt;/div>&lt;/figure>
&lt;/p>
&lt;p>Our material guides students through the process of understanding which claims can and cannot be validated based on the available datasets and how complex it can be to validate each claim. Additionally, they will learn how to use pre-trained models to replicate computationally expensive experiments. Again this material can be on both &lt;a href="https://github.com/mohammed183/re_vit/tree/main" target="_blank" rel="noopener">github&lt;/a> and &lt;a href="https://chameleoncloud.org/experiment/share/8f0e34c5-d2c4-45be-8425-36686ad57650" target="_blank" rel="noopener">chameleon&lt;/a>.&lt;/p>
&lt;p>Both materials are designed to be easy to understand and interactive, allowing students to engage with the content and gain a deeper understanding of the concepts. Instructors can use these materials to assess their students&amp;rsquo; understanding of machine learning and reproducibility.&lt;/p>
&lt;h2 id="reflecting-on-the-journey">Reflecting on the Journey&lt;/h2>
&lt;p>As we wrap up our journey of creating beginner-friendly learning materials for machine learning using reproducibility, it&amp;rsquo;s time to reflect on the rewarding experiences and valuable lessons learned along the way. Our deep dive into the world of machine learning and reproducibility not only enriched our knowledge but also provided us with an opportunity to contribute to the community at the &lt;strong>UC Open Source Symposium 2023&lt;/strong> at UCSC.&lt;/p>
&lt;p>The symposium was a memorable event where we presented our work in a poster session. The diversity of the audience, ranging from professors and researchers to students, added depth to our understanding through their valuable feedback and insights. It was intriguing to see the potential applications of our work in various contexts and its capacity to benefit the broader community.&lt;/p>
&lt;p>This project has been a personal journey of growth, teaching me much more than just machine learning and reproducibility. It honed my skills in collaboration, communication, and problem-solving. I learned to distill complex ideas into simple, accessible language and create engaging, interactive learning experiences. The most fulfilling part of this journey has been seeing our work come alive and realizing its potential to positively impact many people. The gratification that comes from creating something useful for others is unparalleled, and we are thrilled to share our materials with the world.&lt;/p>
&lt;p>Your time and interest in our work are greatly appreciated! Hope you enjoyed this blog!&lt;/p></description></item><item><title>Learning Machine Learning by Reproducing Vision Transformers</title><link>https://deploy-preview-1007--ucsc-ospo.netlify.app/report/osre23/nyu/eduml/20231006-msaeed/</link><pubDate>Fri, 06 Oct 2023 00:00:00 +0000</pubDate><guid>https://deploy-preview-1007--ucsc-ospo.netlify.app/report/osre23/nyu/eduml/20231006-msaeed/</guid><description>&lt;p>Hello again!&lt;/p>
&lt;p>In this blog post, I will be discussing the second material I created for the 2023 Summer of Reproducibility Fellowship. As you may recall from my &lt;a href="https://deploy-preview-1007--ucsc-ospo.netlify.app/report/osre23/nyu/eduml/20230601-msaeed">first post&lt;/a>, I am working on the &lt;a href="https://deploy-preview-1007--ucsc-ospo.netlify.app/project/osre23/nyu/eduml">Using Reproducibility in Machine Learning Education&lt;/a> project with &lt;a href="https://deploy-preview-1007--ucsc-ospo.netlify.app/author/fraida-fund/">Fraida Fund&lt;/a> as my mentor. My goal is to create interactive open-source educational resources that teach reproducibility and reproducible research in machine learning (ML), as outlined in my &lt;a href="https://drive.google.com/file/d/13HnCMZawpabiLdBoOiaJFF2mNXIPLCVJ/view?usp=sharing" target="_blank" rel="noopener">proposal&lt;/a>.&lt;/p>
&lt;p>In this post, I will share with you my second material, and how it can be helpful in machine learning class to teach students about vision transformers and reproducibility at the same time. If you haven&amp;rsquo;t seen my first work, be sure to check out my &lt;a href="https://deploy-preview-1007--ucsc-ospo.netlify.app/report/osre23/nyu/eduml/20230802-msaeed">previous blog post&lt;/a>. Without further ado, let&amp;rsquo;s dive in!&lt;/p>
&lt;h2 id="reproducing-an-image-is-worth-16x16-words-transformers-for-image-recognition-at-scale">Reproducing “An Image is Worth 16x16 Words: Transformers for Image Recognition at Scale”&lt;/h2>
&lt;p>This material is a reproduction of Dosovitskiy et al.‘s 2020 paper, &lt;a href="https://arxiv.org/abs/2010.11929" target="_blank" rel="noopener">“An Image is Worth 16x16 Words: Transformers for Image Recognition at Scale”&lt;/a>. This paper introduces the Vision Transformer (ViT), a novel architecture that applies the transformer model, originally designed for natural language processing tasks, to image recognition. The ViT model achieves state-of-the-art performance on several image classification benchmarks, demonstrating the potential of transformers for computer vision tasks.
&lt;figure >
&lt;div class="d-flex justify-content-center">
&lt;div class="w-100" >&lt;img alt="" srcset="
/report/osre23/nyu/eduml/20231006-msaeed/ViT_hud9ed20979bb56dae4d8e9f4231875a17_383197_485ae1a0cccbdc73994be22901c125d5.webp 400w,
/report/osre23/nyu/eduml/20231006-msaeed/ViT_hud9ed20979bb56dae4d8e9f4231875a17_383197_f8af78acab4a91489ecff3308bc9c9c1.webp 760w,
/report/osre23/nyu/eduml/20231006-msaeed/ViT_hud9ed20979bb56dae4d8e9f4231875a17_383197_1200x1200_fit_q75_h2_lanczos_3.webp 1200w"
src="https://deploy-preview-1007--ucsc-ospo.netlify.app/report/osre23/nyu/eduml/20231006-msaeed/ViT_hud9ed20979bb56dae4d8e9f4231875a17_383197_485ae1a0cccbdc73994be22901c125d5.webp"
width="760"
height="229"
loading="lazy" data-zoomable />&lt;/div>
&lt;/div>&lt;/figure>
The figure illustrates the key idea behind ViT, which is to treat an image as a sequence of patches, similar to how a transformer treats a sentence as a sequence of words. Each patch is flattened into a vector and fed into the transformer encoder, which learns to capture the complex relationships between these patches. The resulting representation is then fed into an MLP head, which produces a final prediction for the image. This approach allows ViT to handle large input images and capture both global context and fine-grained details. ViT models can also be pre-trained on large datasets and fine-tuned on smaller datasets for improved performance.&lt;/p>
&lt;p>To reproduce this paper, I followed a systematic approach to ensure reliable results:&lt;/p>
&lt;ul>
&lt;li>Critically analyze the paper&amp;rsquo;s qualitative and quantitative claims.&lt;/li>
&lt;li>Identify the necessary experiments to verify each claim.&lt;/li>
&lt;li>Determine the required data, code, and hyperparameters for each experiment.&lt;/li>
&lt;li>Utilize pre-trained models for validating claims that require high computational resources.&lt;/li>
&lt;li>Investigate resources shared by the authors, such as code, data, and models.&lt;/li>
&lt;li>Assess the feasibility of verifying different types of claims.&lt;/li>
&lt;li>Design new experiments for validating qualitative claims when certain models or datasets are unavailable.&lt;/li>
&lt;/ul>
&lt;p>I utilized &lt;a href="https://chameleoncloud.org/" target="_blank" rel="noopener">Chameleon&lt;/a> as my platform for conducting and documenting my reproduction experiments. Chameleon is a large-scale, reconfigurable experimental environment that supports computer science systems research. It enables users to create and share Jupyter notebooks capable of running Python code on Chameleon’s cloud servers. For this work, a GPU with 24GB or more memory is required to run the notebooks on GPU, which Chameleon offers in its variety of GPUs.&lt;/p>
&lt;p>I have set up a &lt;a href="https://github.com/mohammed183/re_vit" target="_blank" rel="noopener">GitHub repository&lt;/a> where you can access all of my reproduction work. The repository contains interactive Jupyter notebooks that will help you learn more about machine learning and the reproducibility of machine learning research. These notebooks provide a hands-on approach to understanding the concepts and techniques presented in my reproduction work.&lt;/p>
&lt;h2 id="challenges">Challenges&lt;/h2>
&lt;p>Reproducing a paper can be a challenging task, and I encountered several obstacles during the process, including:&lt;/p>
&lt;ul>
&lt;li>The unavailability of pretraining datasets and pretrained models&lt;/li>
&lt;li>Inexact or unspecified hyperparameters&lt;/li>
&lt;li>The need for expensive resources for some hyperparameters&lt;/li>
&lt;li>The use of different frameworks for baseline CNNs and Vision Transformers&lt;/li>
&lt;/ul>
&lt;p>These issues posed significant difficulties in replicating the following table, a key result from the Vision Transformer paper that demonstrates its superiority over prior state-of-the-art models.
&lt;figure >
&lt;div class="d-flex justify-content-center">
&lt;div class="w-100" >&lt;img alt="" srcset="
/report/osre23/nyu/eduml/20231006-msaeed/table1_hu639b2ac18dac1313dd35f10cc0ae8db7_237634_7faf7451ff08ac87e7d12ab941c77f8e.webp 400w,
/report/osre23/nyu/eduml/20231006-msaeed/table1_hu639b2ac18dac1313dd35f10cc0ae8db7_237634_5441e5e4c6ffed9b29244a3a3dcde852.webp 760w,
/report/osre23/nyu/eduml/20231006-msaeed/table1_hu639b2ac18dac1313dd35f10cc0ae8db7_237634_1200x1200_fit_q75_h2_lanczos_3.webp 1200w"
src="https://deploy-preview-1007--ucsc-ospo.netlify.app/report/osre23/nyu/eduml/20231006-msaeed/table1_hu639b2ac18dac1313dd35f10cc0ae8db7_237634_7faf7451ff08ac87e7d12ab941c77f8e.webp"
width="760"
height="354"
loading="lazy" data-zoomable />&lt;/div>
&lt;/div>&lt;/figure>
&lt;/p>
&lt;p>To overcome these challenges, I used the same models mentioned in the paper but pretrained on different datasets, experimented with various hyperparameter combinations to achieve the best results, and wrote my own code to ensure that both the baseline and Vision Transformer were fine-tuned using the same framework. I also faced other challenges, which I discussed in my notebooks along with the solutions I applied.&lt;/p>
&lt;h2 id="how-to-use-this-material">How to use this material?&lt;/h2>
&lt;p>This material consists of a series of notebooks that guide you through the paper, its claims, experiments, and results. You will learn how to analyze, interpret, and validate the authors&amp;rsquo; claims. To get started, I recommend briefly skimming the &lt;a href="https://arxiv.org/abs/2010.11929" target="_blank" rel="noopener">original paper&lt;/a> to gain an understanding of the main ideas and public information. This will help you see how the authors could have been more transparent and clear in certain sections. The notebooks provide clear instructions and explanations, as well as details on how I addressed any missing components.&lt;/p>
&lt;h2 id="conclusion">Conclusion&lt;/h2>
&lt;p>In this blog post, I&amp;rsquo;ve walked you through the contents of this material and the insights users can gain from it. This material is particularly intriguing as it replicates a paper that has significantly influenced the field of computer vision. The interactive nature of the material makes it not only educational but also engaging and enjoyable. I believe users will find this resource both fun and beneficial.&lt;/p>
&lt;p>I hope you found this post informative and interesting. If you have any questions or feedback, please feel free to contact me. Thank you for reading and stay tuned for more updates!&lt;/p></description></item><item><title>Introducing Levels of Reproduction and Replication in Machine Learning</title><link>https://deploy-preview-1007--ucsc-ospo.netlify.app/report/osre23/nyu/eduml/20230802-msaeed/</link><pubDate>Wed, 02 Aug 2023 00:00:00 +0000</pubDate><guid>https://deploy-preview-1007--ucsc-ospo.netlify.app/report/osre23/nyu/eduml/20230802-msaeed/</guid><description>&lt;p>Hello again,&lt;/p>
&lt;p>I am Mohamed Saeed and this is my second blog post for the 2023 Summer of Reproducibility Fellowship. As you may recall from my &lt;a href="https://deploy-preview-1007--ucsc-ospo.netlify.app/report/osre23/nyu/eduml/20230601-msaeed">previous post&lt;/a>, I am working on the &lt;a href="https://deploy-preview-1007--ucsc-ospo.netlify.app/project/osre23/nyu/eduml">Using Reproducibility in Machine Learning Education&lt;/a> project with &lt;a href="https://deploy-preview-1007--ucsc-ospo.netlify.app/author/fraida-fund/">Fraida Fund&lt;/a> as my mentor. My goal is to create interactive open educational resources that teach reproducibility and reproducible research in machine learning (ML) as I &lt;a href="https://drive.google.com/file/d/13HnCMZawpabiLdBoOiaJFF2mNXIPLCVJ/view?usp=sharing" target="_blank" rel="noopener">proposed&lt;/a>.&lt;/p>
&lt;p>In this post, I will share with you some of the progress I have made so far, as well as some of the challenges I have faced and how I overcame them. I will also highlight some of the specific accomplishments that I am proud of and what I plan to do next.&lt;/p>
&lt;h2 id="reproducing-on-warm-starting-neural-network-training">Reproducing &amp;ldquo;On Warm Starting Neural Network Training&amp;rdquo;&lt;/h2>
&lt;p>This material is a reproduction of the paper &lt;a href="https://arxiv.org/abs/1910.08475" target="_blank" rel="noopener">&amp;ldquo;On Warm Starting Neural Network Training&amp;rdquo;&lt;/a> by Jordan T. Ash and Ryan P. Adams (2020). This paper investigates the effect of warm-starting neural networks, which means using the weights of previous models trained on a subset of the data, to train on a new dataset that has more data.
&lt;figure >
&lt;div class="d-flex justify-content-center">
&lt;div class="w-100" >&lt;img alt="" srcset="
/report/osre23/nyu/eduml/20230802-msaeed/warm_start_huf40f540ab6672b609385b58179d23d2a_3423296_0c5af6e4428dce728fe7a643b2b8e6d3.webp 400w,
/report/osre23/nyu/eduml/20230802-msaeed/warm_start_huf40f540ab6672b609385b58179d23d2a_3423296_f3e332c8b81d6d3146e54527a273bbfe.webp 760w,
/report/osre23/nyu/eduml/20230802-msaeed/warm_start_huf40f540ab6672b609385b58179d23d2a_3423296_1200x1200_fit_q75_h2_lanczos_3.webp 1200w"
src="https://deploy-preview-1007--ucsc-ospo.netlify.app/report/osre23/nyu/eduml/20230802-msaeed/warm_start_huf40f540ab6672b609385b58179d23d2a_3423296_0c5af6e4428dce728fe7a643b2b8e6d3.webp"
width="760"
height="383"
loading="lazy" data-zoomable />&lt;/div>
&lt;/div>&lt;/figure>
The figure illustrates how the new model uses the weights from the previous model as its initial values. This allows the new model to train on both the “Original” data, which it has already seen, and the new data, which it has not encountered before. In contrast, the randomly initialized model treats the entire data as unfamiliar and starts from scratch.&lt;/p>
&lt;p>The paper also shows that this method can lead to lower test accuracy than starting from scratch with random weights, even though the training loss is similar. The paper also proposes a simple way to improve the test accuracy of warm-starting by adding some noise to the previous weights.&lt;/p>
&lt;p>To reproduce this paper, I followed a systematic approach that ensured reliable results. This approach involved:&lt;/p>
&lt;ul>
&lt;li>Reading the paper and its main claims carefully.&lt;/li>
&lt;li>Finding out what resources the authors shared, such as code, data, and models.&lt;/li>
&lt;li>Looking for additional materials online that could help me save time and fill in the gaps left by the authors.&lt;/li>
&lt;li>Setting up the environment and dependencies needed to run the code smoothly.&lt;/li>
&lt;li>Writing code and updating any outdated functions that might cause errors.&lt;/li>
&lt;li>Running the code and verifying that it matched the results reported in the paper.&lt;/li>
&lt;li>Analyzing and interpreting the results and comparing them with the paper’s findings.&lt;/li>
&lt;/ul>
&lt;p>I used &lt;a href="https://chameleoncloud.org/" target="_blank" rel="noopener">Chameleon&lt;/a> as my platform for running and documenting my reproduction experiments. Chameleon is a large-scale, reconfigurable experimental platform that supports computer science systems research. It allows users to create and share Jupyter notebooks that can run Python code on Chameleon’s cloud servers.&lt;/p>
&lt;p>I created a &lt;a href="https://github.com/mohammed183/re_warm_start_nn" target="_blank" rel="noopener">GitHub repository&lt;/a> where you can find all related to my reproduction work in the form of interactive jupyter notebooks that will help you learn more about machine learning and reproducibility of machine learning research.&lt;/p>
&lt;h2 id="challenges">Challenges&lt;/h2>
&lt;p>Reproducing a paper is not an easy task. I faced several challenges along the way. One of the biggest challenges was the lack of code and pretrained models from the authors. This is a common problem for many reproducibility projects. Fortunately, I found a previous reproducibility publication for this paper on &lt;a href="https://rescience.github.io/bibliography/Kireev_2021.html" target="_blank" rel="noopener">ReScience journal&lt;/a>. I used some of their code and added some new functions and modifications to match the original paper’s descriptions. I also encountered other challenges that I discussed in the notebooks with the solutions that I applied.&lt;/p>
&lt;h2 id="how-to-use-this-material">How to use this material?&lt;/h2>
&lt;p>This material is a series of notebooks that walk you through the paper and its claims, experiments, and results. You will learn how to analyze, explain, and validate the authors’ claims. To get started, I suggest you skim the &lt;a href="https://arxiv.org/abs/1910.08475" target="_blank" rel="noopener">original paper&lt;/a> briefly to get the main idea and the public information. This will help you understand how the authors could have been more clear and transparent in some sections. I have given clear instructions and explanations in the notebooks, as well as how I dealt with the missing components. You can use this material for self-learning or as an assignment by hiding the final explanation notebook.&lt;/p>
&lt;h2 id="conclusion-and-future-work">Conclusion and Future Work&lt;/h2>
&lt;p>In this blog post, I have shared with you some of my work on reproducing warm starting neural network training. I have learned a lot from this experience and gained a deeper understanding of reproducibility and reproducible research principles in ML.&lt;/p>
&lt;p>I am very happy with what I have achieved so far, but I still have more work to do. I am working on reproducing the &lt;a href="https://arxiv.org/abs/2010.11929" target="_blank" rel="noopener">Vision Transformer: An Image is Worth 16x16 Words&lt;/a> paper by Alexey Dosovitskiy et al. This time my approach is to use the available pretrained models provided by the authors to verify the claims made in the paper. However, there are some challenges that I face in reproducing the paper. For example, some of the datasets and code that the authors used are not publicly available, which makes it hard to replicate their experiments exactly. These challenges are common in reproducing research papers, especially in computer vision. Therefore, it is important to learn how to deal with them and find ways to validate some of the claims.&lt;/p>
&lt;p>I hope you enjoyed reading this blog post and found it informative and interesting. If you have any questions or feedback, please feel free to contact me. Thank you for your attention and stay tuned for more updates!&lt;/p></description></item><item><title>Introducing Levels of Reproduction and Replication in Machine Learning</title><link>https://deploy-preview-1007--ucsc-ospo.netlify.app/report/osre23/nyu/eduml/20230601-msaeed/</link><pubDate>Thu, 01 Jun 2023 00:00:00 +0000</pubDate><guid>https://deploy-preview-1007--ucsc-ospo.netlify.app/report/osre23/nyu/eduml/20230601-msaeed/</guid><description>&lt;p>Greetings everyone,&lt;/p>
&lt;p>I am Mohamed Saeed and I am delighted to be part of the 2023 Summer of Reproducibility program, where I am contributing to the &lt;a href="https://deploy-preview-1007--ucsc-ospo.netlify.app/project/osre23/nyu/eduml">Using Reproducibility in Machine Learning Education&lt;/a> project.&lt;/p>
&lt;p>My &lt;a href="https://drive.google.com/file/d/13HnCMZawpabiLdBoOiaJFF2mNXIPLCVJ/view?usp=sharing" target="_blank" rel="noopener">proposal&lt;/a> was accepted, and I am fortunate to have &lt;a href="https://deploy-preview-1007--ucsc-ospo.netlify.app/author/fraida-fund/">Fraida Fund&lt;/a> as my mentor. The objective of my project is to develop highly interactive open educational resources that can be utilized by instructors teaching graduate or undergraduate machine learning courses. These resources will focus on integrating instruction on reproducibility and reproducible research principles.&lt;/p>
&lt;p>Understanding and practicing reproducibility in machine learning (ML) research is of utmost importance in today&amp;rsquo;s scientific and technological landscape. Reproducibility ensures the reliability, transparency, and credibility of ML findings and discoveries. By learning the principles of reproducibility, students from different levels can validate research results, test introduced methodologies, and understand level of reproducibilty of research.&lt;/p>
&lt;p>My contribution will involve developing interactive educational resources that encompass code examples, writing exercises, and comprehensive explanations of key concepts of reproducing ML research. These resources will be carefully crafted to assist students at various levels of expertise. Our aim is for these resources to be widely adopted by instructors teaching graduate or undergraduate machine learning courses, as they seek to enhance the understanding of reproducibility and reproducible research principles.&lt;/p>
&lt;p>I think this is a great opportunity to learn more about ML research reproducibility. I&amp;rsquo;ll be posting regular updates and informative blogs throughout the summer, so stay tuned!&lt;/p></description></item></channel></rss>