<?xml version="1.0" encoding="utf-8" standalone="yes"?><rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom"><channel><title>osre26 | UCSC OSPO</title><link>https://deploy-preview-1007--ucsc-ospo.netlify.app/tag/osre26/</link><atom:link href="https://deploy-preview-1007--ucsc-ospo.netlify.app/tag/osre26/index.xml" rel="self" type="application/rss+xml"/><description>osre26</description><generator>Wowchemy (https://wowchemy.com)</generator><language>en-us</language><lastBuildDate>Thu, 05 Feb 2026 00:00:00 +0000</lastBuildDate><image><url>https://deploy-preview-1007--ucsc-ospo.netlify.app/media/logo_hub6795c39d7c5d58c9535d13299c9651f_74810_300x300_fit_lanczos_3.png</url><title>osre26</title><link>https://deploy-preview-1007--ucsc-ospo.netlify.app/tag/osre26/</link></image><item><title>NETAI: AI-Powered Network Anomaly Detection and Diagnostics Platform</title><link>https://deploy-preview-1007--ucsc-ospo.netlify.app/project/osre26/ucsd/netai/</link><pubDate>Thu, 05 Feb 2026 00:00:00 +0000</pubDate><guid>https://deploy-preview-1007--ucsc-ospo.netlify.app/project/osre26/ucsd/netai/</guid><description>&lt;p>NETAI (Network AI) is an AI-powered network anomaly detection and diagnostics platform for the National Research Platform (NRP). This project combines Kubernetes-native LLM integration, network performance monitoring, and predictive analytics to create an intelligent assistant for network operators. Students will work with cutting-edge technologies including Large Language Models (LLMs), Kubernetes, perfSONAR network measurements, time-series analysis, and containerized AI/ML workloads, while contributing to real-world applications in network operations and diagnostics.&lt;/p>
&lt;p>The project involves developing a &lt;strong>Kubernetes chatbot&lt;/strong> that leverages NRP&amp;rsquo;s managed LLM service (providing access to models like Qwen3-VL, GLM-4.7, and GPT-OSS) to help network operators understand complex network behaviors, diagnose anomalies, and receive natural language explanations of network issues. Students will integrate perfSONAR measurement data with traceroute path analysis to create an interactive network topology visualization, and develop &lt;strong>AI/ML models&lt;/strong> for predictive network performance analysis using NRP&amp;rsquo;s GPU resources.&lt;/p>
&lt;p>In addition, students will gain hands-on experience with &lt;strong>fine-tuning LLMs&lt;/strong> on historical network diagnostics data, developing &lt;strong>time-series forecasting models&lt;/strong> for network metrics, and implementing &lt;strong>anomaly detection&lt;/strong> using deep learning techniques. The entire AI/ML pipeline will be containerized and deployed as Kubernetes workloads, utilizing GPU-enabled pods for model training and inference, ensuring scalability and seamless integration with existing NRP infrastructure.&lt;/p>
&lt;p>The platform builds upon existing network diagnostics capabilities, combining end-to-end throughput measurements with detailed traceroute data to enable operators to visualize network paths, identify performance bottlenecks, and understand relationships between metrics and underlying infrastructure. The AI enhancement will provide predictive capabilities, automated incident reporting, and intelligent recommendations for network remediation strategies.&lt;/p>
&lt;h3 id="netai--llm-integration--kubernetes-chatbot">NETAI / LLM Integration &amp;amp; Kubernetes Chatbot&lt;/h3>
&lt;p>The proposed work includes developing a &lt;strong>Kubernetes-native chatbot&lt;/strong> that integrates with NRP&amp;rsquo;s managed LLM service to provide intelligent network diagnostics assistance. Students will create a conversational interface that can answer questions about network performance, explain anomalies in natural language, and suggest remediation strategies. They will fine-tune LLMs on historical network diagnostics data, test results, and traceroute information to create domain-specific assistants. Students will implement &lt;strong>RESTful APIs&lt;/strong> for chatbot interactions, develop &lt;strong>prompt engineering&lt;/strong> strategies for network diagnostics, and create &lt;strong>context-aware responses&lt;/strong> that incorporate real-time network telemetry. The chatbot will be deployed as Kubernetes services, utilizing GPU pods for inference and integrating with the existing diagnostics platform.&lt;/p>
&lt;ul>
&lt;li>&lt;strong>Topics:&lt;/strong> Large Language Models, Kubernetes, Chatbots, Natural Language Processing, Network Diagnostics, API Development&lt;/li>
&lt;li>&lt;strong>Skills:&lt;/strong> Python, Kubernetes, LLM APIs (Qwen3-VL, GLM-4.7, GPT-OSS), Prompt Engineering, REST APIs, Docker, GPU Computing&lt;/li>
&lt;li>&lt;strong>Difficulty:&lt;/strong> Hard&lt;/li>
&lt;li>&lt;strong>Size:&lt;/strong> Large (350 hours)&lt;/li>
&lt;li>&lt;strong>Mentors:&lt;/strong> &lt;a href="https://deploy-preview-1007--ucsc-ospo.netlify.app/author/dmitry-mishin/">Dmitry Mishin&lt;/a>, &lt;a href="https://deploy-preview-1007--ucsc-ospo.netlify.app/author/derek-weitzel/">Derek Weitzel&lt;/a>&lt;/li>
&lt;/ul>
&lt;h3 id="netai--network-anomaly-detection-models">NETAI / Network Anomaly Detection Models&lt;/h3>
&lt;p>The proposed work includes developing &lt;strong>deep learning models&lt;/strong> for network anomaly detection using historical perfSONAR and traceroute data. Students will create models that can identify slow links, high packet loss, excessive retransmits, and failed network tests automatically. They will implement &lt;strong>anomaly detection algorithms&lt;/strong> using techniques such as autoencoders, LSTM networks, and transformer architectures. Students will train models on NRP&amp;rsquo;s GPU clusters using historical network telemetry stored in SQLite databases, develop &lt;strong>feature engineering&lt;/strong> pipelines for network metrics, and create &lt;strong>real-time inference services&lt;/strong> deployed as Kubernetes workloads. The models will be integrated into the diagnostics platform to provide automated anomaly detection alongside the interactive visualization.&lt;/p>
&lt;ul>
&lt;li>&lt;strong>Topics:&lt;/strong> Deep Learning, Anomaly Detection, Time-Series Analysis, Network Monitoring, Model Training, GPU Computing&lt;/li>
&lt;li>&lt;strong>Skills:&lt;/strong> Python, PyTorch/TensorFlow, scikit-learn, Pandas, NumPy, SQLite, Kubernetes, GPU Pods, MLOps&lt;/li>
&lt;li>&lt;strong>Difficulty:&lt;/strong> Hard&lt;/li>
&lt;li>&lt;strong>Size:&lt;/strong> Large (350 hours)&lt;/li>
&lt;li>&lt;strong>Mentors:&lt;/strong> &lt;a href="https://deploy-preview-1007--ucsc-ospo.netlify.app/author/dmitry-mishin/">Dmitry Mishin&lt;/a>, &lt;a href="https://deploy-preview-1007--ucsc-ospo.netlify.app/author/derek-weitzel/">Derek Weitzel&lt;/a>&lt;/li>
&lt;/ul>
&lt;h3 id="netai--predictive-analytics--forecasting">NETAI / Predictive Analytics &amp;amp; Forecasting&lt;/h3>
&lt;p>The proposed work includes developing &lt;strong>predictive models&lt;/strong> that can forecast network performance degradation and identify patterns in network anomalies before they impact users. Students will create &lt;strong>time-series forecasting models&lt;/strong> for network metrics such as throughput, latency, and packet loss, using techniques like ARIMA, Prophet, and deep learning-based forecasting. They will implement &lt;strong>few-shot learning approaches&lt;/strong> to adapt models to new network topologies and measurement patterns, develop &lt;strong>early warning systems&lt;/strong> for potential network issues, and create &lt;strong>automated incident report generation&lt;/strong> using LLMs. Students will leverage NRP&amp;rsquo;s GPU resources for training forecasting models and deploy them as Kubernetes services for real-time predictions integrated with the diagnostics dashboard.&lt;/p>
&lt;ul>
&lt;li>&lt;strong>Topics:&lt;/strong> Time-Series Forecasting, Predictive Analytics, Machine Learning, Network Performance, Early Warning Systems, LLM Integration&lt;/li>
&lt;li>&lt;strong>Skills:&lt;/strong> Python, PyTorch/TensorFlow, Prophet, ARIMA, Pandas, NumPy, Time-Series Analysis, Kubernetes, GPU Computing&lt;/li>
&lt;li>&lt;strong>Difficulty:&lt;/strong> Hard&lt;/li>
&lt;li>&lt;strong>Size:&lt;/strong> Large (350 hours)&lt;/li>
&lt;li>&lt;strong>Mentors:&lt;/strong> &lt;a href="https://deploy-preview-1007--ucsc-ospo.netlify.app/author/dmitry-mishin/">Dmitry Mishin&lt;/a>, &lt;a href="https://deploy-preview-1007--ucsc-ospo.netlify.app/author/derek-weitzel/">Derek Weitzel&lt;/a>&lt;/li>
&lt;/ul>
&lt;h3 id="netai--kubernetes-deployment--infrastructure">NETAI / Kubernetes Deployment &amp;amp; Infrastructure&lt;/h3>
&lt;p>The proposed work includes setting up &lt;strong>Kubernetes-based infrastructure&lt;/strong> for deploying the entire NETAI platform, including LLM services, ML models, and the diagnostics dashboard. Students will create &lt;strong>Helm charts&lt;/strong> for deploying containerized AI/ML workloads, configure &lt;strong>GPU-enabled pods&lt;/strong> for model training and inference, and implement &lt;strong>persistent storage&lt;/strong> solutions for maintaining historical network telemetry. They will develop &lt;strong>GitLab CI/CD pipelines&lt;/strong> for automated testing and deployment, set up &lt;strong>monitoring and observability&lt;/strong> using Prometheus and Grafana for tracking model performance and resource usage, and create &lt;strong>scalable deployment strategies&lt;/strong> that leverage NRP&amp;rsquo;s distributed computing resources. Students will also integrate the platform with existing perfSONAR infrastructure and ensure seamless operation within the NRP cluster.&lt;/p>
&lt;ul>
&lt;li>&lt;strong>Topics:&lt;/strong> Kubernetes, DevOps, CI/CD, GPU Computing, Container Orchestration, Infrastructure as Code, Monitoring&lt;/li>
&lt;li>&lt;strong>Skills:&lt;/strong> Kubernetes, Helm, GitLab CI/CD, Prometheus, Grafana, Docker, GPU Pods, Persistent Storage, Infrastructure Automation&lt;/li>
&lt;li>&lt;strong>Difficulty:&lt;/strong> Medium to Hard&lt;/li>
&lt;li>&lt;strong>Size:&lt;/strong> Large (350 hours)&lt;/li>
&lt;li>&lt;strong>Mentors:&lt;/strong> &lt;a href="https://deploy-preview-1007--ucsc-ospo.netlify.app/author/dmitry-mishin/">Dmitry Mishin&lt;/a>, &lt;a href="https://deploy-preview-1007--ucsc-ospo.netlify.app/author/derek-weitzel/">Derek Weitzel&lt;/a>&lt;/li>
&lt;/ul>
&lt;h2 id="project-resources">Project Resources&lt;/h2>
&lt;ul>
&lt;li>&lt;strong>National Research Platform&lt;/strong>: &lt;a href="https://nrp.ai/" target="_blank" rel="noopener">https://nrp.ai/&lt;/a>&lt;/li>
&lt;li>&lt;strong>NRP LLM Service&lt;/strong>: &lt;a href="https://nrp.ai/documentation/userdocs/ai/llm-managed/" target="_blank" rel="noopener">https://nrp.ai/documentation/userdocs/ai/llm-managed/&lt;/a>&lt;/li>
&lt;li>&lt;strong>perfSONAR&lt;/strong>: &lt;a href="https://www.perfsonar.net/" target="_blank" rel="noopener">https://www.perfsonar.net/&lt;/a>&lt;/li>
&lt;li>&lt;strong>MaDDash&lt;/strong>: &lt;a href="https://github.com/esnet/maddash" target="_blank" rel="noopener">https://github.com/esnet/maddash&lt;/a>&lt;/li>
&lt;li>&lt;strong>Network Monitoring Documentation&lt;/strong>: &lt;a href="https://nrp.ai/documentation/" target="_blank" rel="noopener">https://nrp.ai/documentation/&lt;/a>&lt;/li>
&lt;/ul>
&lt;h2 id="background">Background&lt;/h2>
&lt;p>This project addresses critical gaps in network performance monitoring for the National Research Platform by integrating AI/ML capabilities with existing perfSONAR-based diagnostics. The platform combines end-to-end network measurements with detailed path-level analysis, enhanced by intelligent AI assistants that can help operators understand complex network behaviors and predict potential issues. By leveraging NRP&amp;rsquo;s managed LLM service and GPU resources, students will create a Kubernetes-native system that scales across the distributed research network infrastructure, providing both real-time diagnostics and predictive analytics to improve network reliability and performance for researchers nationwide.&lt;/p></description></item><item><title>VINE: Precision Agriculture Data Platform &amp; Digital Twin</title><link>https://deploy-preview-1007--ucsc-ospo.netlify.app/project/osre26/ucsd/vine/</link><pubDate>Thu, 05 Feb 2026 00:00:00 +0000</pubDate><guid>https://deploy-preview-1007--ucsc-ospo.netlify.app/project/osre26/ucsd/vine/</guid><description>&lt;p>VINE (Vineyard Intelligence Network &amp;amp; Environment) is an AI/ML research project focused on precision agriculture using the &lt;strong>National Research Platform (NRP)&lt;/strong>. This project leverages the innovative demonstration at Iron Horse Vineyards to study how AI and machine learning can optimize agricultural practices through data-driven insights. Students will work with cutting-edge AI/ML technologies, distributed computing on NRP, and large-scale data analysis, while contributing to real-world applications in sustainable agriculture and climate adaptation.&lt;/p>
&lt;p>The project involves &lt;strong>AI/ML research&lt;/strong> using agricultural data from Iron Horse Vineyards, leveraging the computational resources of the &lt;strong>National Research Platform&lt;/strong> for training and deploying machine learning models. Students will work with agricultural datasets including sensor data, multi-spectral drone imagery, and historical records, developing models for predictive analytics, computer vision, and time-series forecasting. The integration of &lt;strong>NRP&amp;rsquo;s distributed infrastructure&lt;/strong> enables scalable AI research that can process large volumes of sensor data, multi-spectral imagery, and historical agricultural records.&lt;/p>
&lt;p>Students will gain hands-on experience with &lt;strong>AI/ML model development&lt;/strong> for agricultural applications, learning how to analyze multi-spectral drone imagery, process time-series sensor data, and build predictive models for irrigation scheduling, pest detection, and harvest timing. They will deploy and train models on &lt;strong>NRP&amp;rsquo;s Kubernetes clusters&lt;/strong>, utilize &lt;strong>GPU resources&lt;/strong> for deep learning workloads, and work with agricultural datasets for comprehensive research. The project emphasizes using &lt;strong>distributed computing&lt;/strong> on NRP to scale AI/ML experiments and create open, shareable datasets for collaborative research.&lt;/p>
&lt;p>The platform builds upon the success demonstrated at Iron Horse Vineyards, where AI-driven analytics have shown potential for &lt;strong>10% water use reduction&lt;/strong> and improved yield optimization. This project aims to advance AI/ML research in precision agriculture by utilizing NRP&amp;rsquo;s computational capabilities, creating reproducible research that can benefit the broader agricultural and research communities.&lt;/p>
&lt;h3 id="vine--data-pipeline--integration">VINE / Data Pipeline &amp;amp; Integration&lt;/h3>
&lt;p>The proposed work includes building &lt;strong>data pipelines&lt;/strong> to ingest, process, and prepare agricultural data from Iron Horse Vineyards and other sources for AI/ML research. Students will develop pipelines to collect sensor data (soil moisture, temperature, CO2, weather), multi-spectral drone imagery, and historical agricultural records. They will create &lt;strong>data validation and quality assurance&lt;/strong> processes, implement &lt;strong>data preprocessing&lt;/strong> for ML model training, and develop &lt;strong>data integration&lt;/strong> workflows that connect agricultural datasets with NRP computational resources. Students will also work on &lt;strong>data sharing&lt;/strong> mechanisms to make processed datasets available for the research community.&lt;/p>
&lt;ul>
&lt;li>&lt;strong>Topics:&lt;/strong> Data Engineering, Time-Series Data, Data Preprocessing, Data Sharing, ML Data Pipelines&lt;/li>
&lt;li>&lt;strong>Skills:&lt;/strong> Python, Pandas, NumPy, Data Validation, REST APIs, Docker, Kubernetes, Data Processing&lt;/li>
&lt;li>&lt;strong>Difficulty:&lt;/strong> Medium to Hard&lt;/li>
&lt;li>&lt;strong>Size:&lt;/strong> Large (350 hours)&lt;/li>
&lt;li>&lt;strong>Mentors:&lt;/strong> &lt;a href="https://deploy-preview-1007--ucsc-ospo.netlify.app/author/mohammad-firas-sada/">Mohammad Firas Sada&lt;/a>&lt;/li>
&lt;/ul>
&lt;h3 id="vine--aiml-models-for-agricultural-analytics-on-nrp">VINE / AI/ML Models for Agricultural Analytics on NRP&lt;/h3>
&lt;p>The proposed work includes developing and training &lt;strong>machine learning models&lt;/strong> for agricultural applications using the &lt;strong>National Research Platform (NRP)&lt;/strong>. Students will create models for &lt;strong>predictive irrigation scheduling&lt;/strong> based on soil moisture, weather forecasts, and historical data. They will develop &lt;strong>computer vision models&lt;/strong> for analyzing multi-spectral drone imagery to detect plant health, identify pests, and estimate yield. Students will also work on &lt;strong>time-series forecasting&lt;/strong> models for predicting harvest timing and optimizing resource allocation. The project will involve training models on &lt;strong>NRP&amp;rsquo;s GPU clusters&lt;/strong>, utilizing distributed training capabilities, and deploying models for real-time inference. Students will leverage agricultural datasets for training and validation, and contribute model outputs and insights for the research community.&lt;/p>
&lt;ul>
&lt;li>&lt;strong>Topics:&lt;/strong> Machine Learning, Computer Vision, Time-Series Analysis, Predictive Analytics, Agricultural AI, Distributed Training&lt;/li>
&lt;li>&lt;strong>Skills:&lt;/strong> Python, PyTorch/TensorFlow, scikit-learn, OpenCV, Pandas, NumPy, MLOps, NRP Kubernetes, GPU Computing&lt;/li>
&lt;li>&lt;strong>Difficulty:&lt;/strong> Hard&lt;/li>
&lt;li>&lt;strong>Size:&lt;/strong> Large (350 hours)&lt;/li>
&lt;li>&lt;strong>Mentors:&lt;/strong> &lt;a href="https://deploy-preview-1007--ucsc-ospo.netlify.app/author/mohammad-firas-sada/">Mohammad Firas Sada&lt;/a>&lt;/li>
&lt;/ul>
&lt;h3 id="vine--digital-twin--ai-driven-visualization">VINE / Digital Twin &amp;amp; AI-Driven Visualization&lt;/h3>
&lt;p>The proposed work includes creating &lt;strong>AI-enhanced digital twin&lt;/strong> systems for agricultural sites using computational resources on NRP. Students will develop &lt;strong>3D visualization&lt;/strong> systems (potentially using Omniverse or similar platforms) to represent vineyards and farms, integrate &lt;strong>AI model predictions&lt;/strong> into the digital twin for real-time insights, and create &lt;strong>interactive dashboards&lt;/strong> for monitoring and analysis. They will implement &lt;strong>spatial data processing&lt;/strong> using ML models to map sensor locations and readings to geographic coordinates, and develop &lt;strong>AI-driven simulation capabilities&lt;/strong> for testing different agricultural strategies (irrigation patterns, planting layouts, etc.) before implementation. Students will deploy visualization services on &lt;strong>NRP infrastructure&lt;/strong> and integrate with agricultural data sources for real-time updates.&lt;/p>
&lt;ul>
&lt;li>&lt;strong>Topics:&lt;/strong> Digital Twin, AI-Enhanced Visualization, GIS, Spatial Data, ML-Driven Simulation, Real-Time Systems&lt;/li>
&lt;li>&lt;strong>Skills:&lt;/strong> Python, 3D Graphics (Omniverse/Unity/Blender), GIS tools, WebGL, React/Three.js, ML Integration, NRP Deployment&lt;/li>
&lt;li>&lt;strong>Difficulty:&lt;/strong> Hard&lt;/li>
&lt;li>&lt;strong>Size:&lt;/strong> Large (350 hours)&lt;/li>
&lt;li>&lt;strong>Mentors:&lt;/strong> &lt;a href="https://deploy-preview-1007--ucsc-ospo.netlify.app/author/mohammad-firas-sada/">Mohammad Firas Sada&lt;/a>&lt;/li>
&lt;/ul>
&lt;h3 id="vine--web-dashboard--nrp-integration-platform">VINE / Web Dashboard &amp;amp; NRP Integration Platform&lt;/h3>
&lt;p>The proposed work includes building a &lt;strong>comprehensive web dashboard&lt;/strong> for visualizing agricultural data, AI model predictions, and research insights. Students will develop a &lt;strong>full-stack web application&lt;/strong> using modern frameworks (React, Flask/FastAPI) deployed on the &lt;strong>National Research Platform (NRP)&lt;/strong>. The dashboard will display real-time sensor readings, historical trends from agricultural datasets, AI model predictions, and digital twin visualizations. Students will create &lt;strong>API endpoints&lt;/strong> that integrate with &lt;strong>NRP computational resources&lt;/strong> and agricultural data sources, implement &lt;strong>role-based access control&lt;/strong> for researchers, and enable &lt;strong>data export/sharing&lt;/strong> with the broader research community. The platform will support &lt;strong>interactive data exploration&lt;/strong> tools and provide programmatic access to AI/ML models running on NRP.&lt;/p>
&lt;ul>
&lt;li>&lt;strong>Topics:&lt;/strong> Full-Stack Web Development, Data Visualization, API Development, NRP Deployment, ML Model Serving&lt;/li>
&lt;li>&lt;strong>Skills:&lt;/strong> React, Flask/FastAPI, PostgreSQL, D3.js/Plotly, Bootstrap/Tailwind CSS, REST APIs, Kubernetes, NRP APIs&lt;/li>
&lt;li>&lt;strong>Difficulty:&lt;/strong> Medium to Hard&lt;/li>
&lt;li>&lt;strong>Size:&lt;/strong> Large (350 hours)&lt;/li>
&lt;li>&lt;strong>Mentors:&lt;/strong> &lt;a href="https://deploy-preview-1007--ucsc-ospo.netlify.app/author/mohammad-firas-sada/">Mohammad Firas Sada&lt;/a>&lt;/li>
&lt;/ul>
&lt;h2 id="project-resources">Project Resources&lt;/h2>
&lt;ul>
&lt;li>&lt;strong>National Research Platform&lt;/strong>: &lt;a href="https://nrp.ai/" target="_blank" rel="noopener">https://nrp.ai/&lt;/a>&lt;/li>
&lt;li>&lt;strong>Iron Horse Vineyards Project&lt;/strong>: &lt;a href="https://gitlab.nrp-nautilus.io/ihv" target="_blank" rel="noopener">https://gitlab.nrp-nautilus.io/ihv&lt;/a>&lt;/li>
&lt;li>&lt;strong>Omniverse Integration&lt;/strong>: &lt;a href="https://gitlab.nrp-nautilus.io/omniverse" target="_blank" rel="noopener">https://gitlab.nrp-nautilus.io/omniverse&lt;/a>&lt;/li>
&lt;li>&lt;strong>CENIC Network&lt;/strong>: &lt;a href="https://cenic.org/" target="_blank" rel="noopener">https://cenic.org/&lt;/a>&lt;/li>
&lt;li>&lt;strong>CENIC Precision Agriculture Blog&lt;/strong>: &lt;a href="https://nrp.ai/cenic-precision-agriculture-2025" target="_blank" rel="noopener">https://nrp.ai/cenic-precision-agriculture-2025&lt;/a>&lt;/li>
&lt;/ul>
&lt;h2 id="background">Background&lt;/h2>
&lt;p>This project builds upon the successful demonstration at Iron Horse Vineyards, where CENIC, UC San Diego, and partners have created a living laboratory for precision agriculture. The VINE project focuses on &lt;strong>AI/ML research&lt;/strong> using the &lt;strong>National Research Platform (NRP)&lt;/strong> for computational resources. By leveraging NRP&amp;rsquo;s distributed infrastructure and GPU clusters, students can train and deploy sophisticated ML models for agricultural applications. The project works with agricultural datasets from Iron Horse Vineyards and aims to create open, shareable datasets for the research community. This approach creates a scalable, reproducible framework for AI/ML research in precision agriculture that can benefit researchers, educators, and practitioners nationwide.&lt;/p></description></item><item><title>Reconfigurable and Placement-Aware Replication for Edge Systems</title><link>https://deploy-preview-1007--ucsc-ospo.netlify.app/project/osre26/umass/edge-replication/</link><pubDate>Sat, 31 Jan 2026 00:00:00 +0000</pubDate><guid>https://deploy-preview-1007--ucsc-ospo.netlify.app/project/osre26/umass/edge-replication/</guid><description>&lt;h2 id="project-description">Project Description&lt;/h2>
&lt;ul>
&lt;li>&lt;strong>Topics&lt;/strong>: Distributed systems&lt;/li>
&lt;li>&lt;strong>Skills&lt;/strong>: Rust, Java, Go, Python, Bash scripting, Linux, Docker.&lt;/li>
&lt;li>&lt;strong>Difficulty&lt;/strong>: Hard&lt;/li>
&lt;li>&lt;strong>Size&lt;/strong>: Large (350 hours)&lt;/li>
&lt;li>&lt;strong>Mentors&lt;/strong>: &lt;a href="mailto:fikurnia@cs.umass.edu">Fadhil I. Kurnia&lt;/a>&lt;/li>
&lt;/ul>
&lt;p>Modern replicated systems are typically evaluated under static configurations with fixed replica placement. However, real-world edge deployments are highly dynamic: workloads shift geographically, edge nodes join or fail, and latency conditions change over time. Our existing testbed provides reproducible evaluation for replicated systems but lacks support for dynamic reconfiguration and adaptive edge placement policies.&lt;/p>
&lt;p>This project extends the existing open testbed to support:&lt;/p>
&lt;ol>
&lt;li>
&lt;p>Dynamic Replica Reconfiguration&lt;/p>
&lt;ul>
&lt;li>Membership changes (add/remove replicas)&lt;/li>
&lt;li>Leader migration and shard movement&lt;/li>
&lt;li>Online reconfiguration cost measurement (latency spikes, recovery overhead, state transfer cost)&lt;/li>
&lt;/ul>
&lt;/li>
&lt;li>
&lt;p>Edge-Aware Placement Policies&lt;/p>
&lt;ul>
&lt;li>Demand-aware placement based on geographic workload skew&lt;/li>
&lt;li>Latency-aware and bandwidth-aware replica selection&lt;/li>
&lt;li>Comparison of static vs. adaptive placement strategies&lt;/li>
&lt;li>Evaluation under real-world latency matrices (e.g., US metro-level or cloud region traces)&lt;/li>
&lt;/ul>
&lt;/li>
&lt;li>
&lt;p>What-if Simulation Framework&lt;/p>
&lt;ul>
&lt;li>Replay workload traces with time-varying demand&lt;/li>
&lt;li>Simulate hundreds of edge sites with realistic network conditions&lt;/li>
&lt;li>Quantify trade-offs between consistency, availability, reconfiguration overhead, and cost&lt;/li>
&lt;/ul>
&lt;/li>
&lt;/ol>
&lt;p>The outcome will be an &lt;a href="https://distrobench.org" target="_blank" rel="noopener">open-source framework&lt;/a> that enables researchers to evaluate not only steady-state replication performance, but also how systems behave under churn, scaling events, and demand shifts. They are central challenges in real edge environments.&lt;/p>
&lt;h3 id="expected-deliverables">Expected Deliverables&lt;/h3>
&lt;ul>
&lt;li>Reconfiguration abstraction layer (API for membership &amp;amp; placement changes)&lt;/li>
&lt;li>Placement policy plugin framework (k-means, facility-location heuristics, latency-minimizing, cost-aware)&lt;/li>
&lt;li>Trace-driven dynamic workload engine&lt;/li>
&lt;li>Public benchmark scenarios and reproducible experiment scripts&lt;/li>
&lt;li>Artifact-ready documentation and evaluation report&lt;/li>
&lt;/ul></description></item><item><title>AI Data Readiness Inspector (AIDRIN)</title><link>https://deploy-preview-1007--ucsc-ospo.netlify.app/project/osre26/lbl/aidrin/</link><pubDate>Fri, 30 Jan 2026 10:15:00 -0700</pubDate><guid>https://deploy-preview-1007--ucsc-ospo.netlify.app/project/osre26/lbl/aidrin/</guid><description>&lt;p>Garbage In, Garbage Out (GIGO) is a widely accepted quote in computer science across various domains, including Artificial Intelligence (AI). As data is the fuel for AI, models trained on low-quality, biased data are often ineffective. Computer scientists who use AI invest considerable time and effort in preparing the data for AI.&lt;/p>
&lt;p>&lt;a href="https://arxiv.org/pdf/2406.19256" target="_blank" rel="noopener">AIDRIN&lt;/a> (AI Data Readiness INspector) is a framework that provides a quantifiable assessment of data readiness for AI processes, covering a broad range of dimensions from the literature. AIDRIN uses metrics from traditional data quality assessment, such as completeness, outliers, and duplicates, to evaluate data. Furthermore, AIDRIN uses metrics specific to assessing AI data, such as feature importance, feature correlations, class imbalance, fairness, privacy, and compliance with the FAIR (Findability, Accessibility, Interoperability, and Reusability) principles. AIDRIN provides visualizations and reports to assist data scientists in further investigating data readiness.&lt;/p>
&lt;h3 id="aidrin-multiple-file-formats">AIDRIN Multiple File Formats&lt;/h3>
&lt;p>The proposed work will include improvements in the AIDRIN framework to (1) add support for new file formats such as Zarr, ROOT, and HDF5; and (2) to allow providing custom data ingestion mechanisms.&lt;/p>
&lt;ul>
&lt;li>&lt;strong>Topics:&lt;/strong> &lt;code>data readiness&lt;/code>, &lt;code>AI&lt;/code>, &lt;code>data analysis&lt;/code>&lt;/li>
&lt;li>&lt;strong>Skills:&lt;/strong> Python, C/C++, data analysis, good communicator&lt;/li>
&lt;li>&lt;strong>Difficulty:&lt;/strong> Moderate&lt;/li>
&lt;li>&lt;strong>Size:&lt;/strong> Large (350 hours)&lt;/li>
&lt;li>&lt;strong>Mentors:&lt;/strong> &lt;a href="https://deploy-preview-1007--ucsc-ospo.netlify.app/author/jean-luca-bez/">Jean Luca Bez&lt;/a> and &lt;a href="https://deploy-preview-1007--ucsc-ospo.netlify.app/author/suren-byna/">Suren Byna&lt;/a>&lt;/li>
&lt;/ul></description></item><item><title>Drishti</title><link>https://deploy-preview-1007--ucsc-ospo.netlify.app/project/osre26/lbl/drishti/</link><pubDate>Fri, 30 Jan 2026 10:15:00 -0700</pubDate><guid>https://deploy-preview-1007--ucsc-ospo.netlify.app/project/osre26/lbl/drishti/</guid><description>&lt;p>&lt;a href="https://github.com/hpc-io/drishti" target="_blank" rel="noopener">Drishti&lt;/a> is a novel interactive web-based analysis framework to visualize I/O traces, highlight bottlenecks, and help understand the I/O behavior of scientific applications. Drishti aims to fill the gap between the trace collection, analysis, and tuning phases. The framework contains an interactive I/O trace analysis component for end-users to visually inspect their applications&amp;rsquo; I/O behavior, focusing on areas of interest and getting a clear picture of common root causes of I/O performance bottlenecks. Based on the automatic detection of I/O performance bottlenecks, our framework maps numerous common and well-known bottlenecks and their solution recommendations that can be implemented by users.&lt;/p>
&lt;h3 id="drishti-comparisons-and-heatmaps">Drishti Comparisons and Heatmaps&lt;/h3>
&lt;p>The proposed work will include investigating and building a solution to allow comparing and finding differences between two I/O trace files (similar to a &lt;code>diff&lt;/code>), covering the analysis and visualization components. It will also explore additional metrics and counters such as Darshan heatmaps in the analysis and visualization components of the framework.&lt;/p>
&lt;ul>
&lt;li>&lt;strong>Topics:&lt;/strong> &lt;code>I/O&lt;/code>, &lt;code>HPC&lt;/code>, &lt;code>data analysis&lt;/code>, &lt;code>visualization&lt;/code>, &lt;code>profiling&lt;/code>, &lt;code>tracing&lt;/code>&lt;/li>
&lt;li>&lt;strong>Skills:&lt;/strong> Python, data analysis, performance profiling&lt;/li>
&lt;li>&lt;strong>Difficulty:&lt;/strong> Moderate&lt;/li>
&lt;li>&lt;strong>Size:&lt;/strong> Large (350 hours)&lt;/li>
&lt;li>&lt;strong>Mentors:&lt;/strong> &lt;a href="https://deploy-preview-1007--ucsc-ospo.netlify.app/author/jean-luca-bez/">Jean Luca Bez&lt;/a> and &lt;a href="mailto:sbyna@lbl.gov">Suren Byna&lt;/a>&lt;/li>
&lt;/ul></description></item><item><title>EnergyAPI: An End-to-End API for Energy-Aware Forecasting and Scheduling</title><link>https://deploy-preview-1007--ucsc-ospo.netlify.app/project/osre26/ucsc/energy-api/</link><pubDate>Fri, 30 Jan 2026 00:00:00 +0000</pubDate><guid>https://deploy-preview-1007--ucsc-ospo.netlify.app/project/osre26/ucsc/energy-api/</guid><description>&lt;p>Over the past decades, electricity demand has increased steadily, driven by structural shifts such as the electrification of transportation and, more recently, the rapid expansion of artificial intelligence (AI). Power grids have responded by expanding generation capacity, integrating renewable energy sources such as solar and wind, and deploying demand-response mechanisms. However, the current pace of demand growth is increasingly outstripping grid expansion, leading to integration delays, greater reliance on behind-the-meter consumption, and rising operational complexity.&lt;/p>
&lt;p>To mitigate the environmental and socioeconomic impacts of electricity consumption, large consumers such as cloud data centers and electric vehicle (EV) charging infrastructures are increasingly participating in demand-response programs. These programs provide consumers with grid signals indicating favorable periods for electricity usage, such as when energy is cheapest or has the lowest carbon intensity. Consumers can then shift workloads across time and location to better align with grid conditions and their own operational constraints. A key challenge, however, is the online nature of this problem: operators must make real-time decisions without full knowledge of future grid conditions. While forecasting and optimization techniques exist, their effectiveness depends heavily on workload characteristics, such as whether tasks are delay-tolerant cloud jobs or EV charging sessions with route and deadline constraints.&lt;/p>
&lt;p>This project proposes the design and implementation of a modular, extensible API for energy-aware workload scheduling. The API will ingest grid signals alongside workload Service Level Objectives (SLOs) and operational requirements, and produce execution plans that adapt to changing grid conditions. It will support multiple pluggable scheduling strategies and heuristics, enabling developers to compare real-time and forecast-based approaches across different workload classes. By providing a reusable, open-source interface for demand-response-aware scheduling, this project aims to lower the barrier for developers to integrate energy-aware decision-making into distributed systems and applications.&lt;/p>
&lt;h3 id="building-an-end-to-end-service-for-energy-forecasting-and-scheduling">Building an End-to-End Service for Energy Forecasting and Scheduling&lt;/h3>
&lt;ul>
&lt;li>&lt;strong>Topics:&lt;/strong> &lt;code>Databases&lt;/code> &lt;code>Machine Learning&lt;/code>&lt;/li>
&lt;li>&lt;strong>Skills:&lt;/strong> Python, command line tools (bash), SQL (MySQL or SQLite), FastAPI, time-series analysis, basic machine learning&lt;/li>
&lt;li>&lt;strong>Difficulty:&lt;/strong> Moderate&lt;/li>
&lt;li>&lt;strong>Size:&lt;/strong> Large (350 hours)&lt;/li>
&lt;li>&lt;strong>Mentors:&lt;/strong> &lt;a href="https://deploy-preview-1007--ucsc-ospo.netlify.app/author/abel-souza/">Abel Souza&lt;/a>&lt;/li>
&lt;/ul>
&lt;p>Develop a containerized, end-to-end platform consisting of a backend, API, and web-based frontend for collecting, estimating, and visualizing real-time and forecasted electrical grid signals. These signals include electricity demand, prices, energy production, grid saturation, and carbon intensity. The system will support scalable data ingestion, region-specific forecasting models, and interactive visualizations to enable energy-aware application development and analysis.&lt;/p>
&lt;p>Tasks:&lt;/p>
&lt;ul>
&lt;li>Study electrical grid signals and demand-response data sources (e.g., demand, price, carbon intensity, grid saturation) and identify their requirements for real-time and forecast-based consumption planning.&lt;/li>
&lt;li>Design and implement a relational data model for storing historical, real-time, and forecasted grid signals.&lt;/li>
&lt;li>Ingest and validate grid signal data into a MySQL or SQLite database, ensuring data quality and time alignment across regions.&lt;/li>
&lt;li>Implement baseline time-series forecasting models for grid signals (e.g., demand, price, or carbon intensity), with support for region-specific configurations.&lt;/li>
&lt;li>Query European Network of Transmission System Operators for Electricity (ENTSO-E) and EIA (Energy Information Administration (EIA)) APIs to collect grid data.&lt;/li>
&lt;li>Develop a RESTful API that exposes both raw and forecasted grid signals for use by energy-aware applications and schedulers.&lt;/li>
&lt;li>Build a web-based user interface to visualize historical trends, forecasts, and regional differences in grid conditions.&lt;/li>
&lt;li>Implement an interactive choropleth map to display spatial variations in grid signals such as carbon intensity and electricity prices.&lt;/li>
&lt;li>Design an extensible architecture that allows different regions to plug in custom forecasting models or heuristics.&lt;/li>
&lt;li>Containerize the backend, API, and frontend components using Docker to enable reproducible deployment and easy integration by external users.&lt;/li>
&lt;/ul></description></item><item><title>Environmental NeTworked Sensor (ENTS)</title><link>https://deploy-preview-1007--ucsc-ospo.netlify.app/project/osre26/ucsc/ents/</link><pubDate>Fri, 30 Jan 2026 00:00:00 +0000</pubDate><guid>https://deploy-preview-1007--ucsc-ospo.netlify.app/project/osre26/ucsc/ents/</guid><description>&lt;h3 id="ents-i-usability-improvements-for-visualization-dashboard">ENTS I: Usability improvements for visualization dashboard&lt;/h3>
&lt;p>
&lt;figure >
&lt;div class="d-flex justify-content-center">
&lt;div class="w-100" >&lt;img alt="Data Visualization Dashboard" srcset="
/project/osre26/ucsc/ents/osp1_huda3c1d46887767e16b865c47973b8288_360491_2d797937cbe25a879de96b44cb5c65b3.webp 400w,
/project/osre26/ucsc/ents/osp1_huda3c1d46887767e16b865c47973b8288_360491_baae6484e015277af7b09e866b6869f5.webp 760w,
/project/osre26/ucsc/ents/osp1_huda3c1d46887767e16b865c47973b8288_360491_1200x1200_fit_q75_h2_lanczos_3.webp 1200w"
src="https://deploy-preview-1007--ucsc-ospo.netlify.app/project/osre26/ucsc/ents/osp1_huda3c1d46887767e16b865c47973b8288_360491_2d797937cbe25a879de96b44cb5c65b3.webp"
width="760"
height="759"
loading="lazy" data-zoomable />&lt;/div>
&lt;/div>&lt;/figure>
&lt;/p>
&lt;ul>
&lt;li>&lt;strong>Topics:&lt;/strong> Data Visualization, Backend, Frontend, UI/UX, Analytics&lt;/li>
&lt;li>&lt;strong>Skills:&lt;/strong>
&lt;ul>
&lt;li>&lt;em>Required:&lt;/em> React, Javascript, Python, SQL, Git&lt;/li>
&lt;li>&lt;em>Nice to have:&lt;/em> Flask, Docker, CI/CD, AWS, Authentication&lt;/li>
&lt;/ul>
&lt;/li>
&lt;li>&lt;strong>Difficulty:&lt;/strong> Medium&lt;/li>
&lt;li>&lt;strong>Size:&lt;/strong> Large (350 hours)&lt;/li>
&lt;li>&lt;strong>Mentors:&lt;/strong> &lt;a href="https://deploy-preview-1007--ucsc-ospo.netlify.app/author/colleen-josephson/">Colleen Josephson&lt;/a>, &lt;a href="mailto:alevy1@ucsc.edu">Alec Levy&lt;/a>, &lt;a href="mailto:jtmadden@ucsc.edu">John Madden&lt;/a>&lt;/li>
&lt;/ul>
&lt;p>The Environmental NeTworked Sensor (ENTS) platform, formally Open Sensing Platform (OSP), implements data visualization website for monitoring microbial fuel cell sensors (see &lt;a href="https://github.com/jlab-sensing/ENTS-backend" target="_blank" rel="noopener">GitHub&lt;/a>). The mission is to scale up the current platform to support other researchers or citizen scientists in integrating their novel sensing hardware or microbial fuel cell sensors for monitoring and data analysis. Examples of the types of sensors currently deployed are sensors measuring soil moisture, temperature, current, and voltage in outdoor settings. The focus of the software half of the project involves building upon our existing visualization web platform, and adding additional features to support the mission. A live version of the website is available &lt;a href="https://dirtviz.jlab.ucsc.edu/" target="_blank" rel="noopener">here&lt;/a>.&lt;/p>
&lt;p>Below is a list of project ideas that would be beneficial to the ENTS project. You are not limited to the following projects, and encourage new ideas that enhance the platform:&lt;/p>
&lt;ul>
&lt;li>Drag and drop charts functionality&lt;/li>
&lt;li>Creation of unique charts by users (with unique equations)&lt;/li>
&lt;li>Customizable options of charts (color, line width, datapoint/line style, axis labels)&lt;/li>
&lt;li>Exportable charts (with customizable options)&lt;/li>
&lt;li>Saving layouts via url&lt;/li>
&lt;/ul>
&lt;h3 id="ents-ii-migration-to-tockos">ENTS II: Migration to TockOS&lt;/h3>
&lt;p>
&lt;figure >
&lt;div class="d-flex justify-content-center">
&lt;div class="w-100" >&lt;img alt="ENTS in the wild" srcset="
/project/osre26/ucsc/ents/flower_bed_hua65f08ca6bedf0f2d60c653056e1b3a7_800588_c34f23edec4789d86dcf04482fa38282.webp 400w,
/project/osre26/ucsc/ents/flower_bed_hua65f08ca6bedf0f2d60c653056e1b3a7_800588_8a4ed9b7cf50d0c7493779c714094459.webp 760w,
/project/osre26/ucsc/ents/flower_bed_hua65f08ca6bedf0f2d60c653056e1b3a7_800588_1200x1200_fit_q75_h2_lanczos.webp 1200w"
src="https://deploy-preview-1007--ucsc-ospo.netlify.app/project/osre26/ucsc/ents/flower_bed_hua65f08ca6bedf0f2d60c653056e1b3a7_800588_c34f23edec4789d86dcf04482fa38282.webp"
width="760"
height="369"
loading="lazy" data-zoomable />&lt;/div>
&lt;/div>&lt;/figure>
&lt;/p>
&lt;ul>
&lt;li>&lt;strong>Topics:&lt;/strong> Embedded system, operating system&lt;/li>
&lt;li>&lt;strong>Skills:&lt;/strong>
&lt;ul>
&lt;li>&lt;em>Required:&lt;/em> Rust, C/C++, Git, Github&lt;/li>
&lt;li>&lt;em>Nice to have:&lt;/em> STM32 HAL, python&lt;/li>
&lt;/ul>
&lt;/li>
&lt;li>&lt;strong>Difficulty:&lt;/strong> Hard&lt;/li>
&lt;li>&lt;strong>Size:&lt;/strong> Large (350 hours)&lt;/li>
&lt;li>&lt;strong>Mentors:&lt;/strong> &lt;a href="https://deploy-preview-1007--ucsc-ospo.netlify.app/author/colleen-josephson/">Colleen Josephson&lt;/a>, &lt;a href="mailto:jtmadden@ucsc.edu">John Madden&lt;/a>&lt;/li>
&lt;/ul>
&lt;p>The current version of the hardware firmware is implemented in baremetal
through the use of STM hardware abstraction layer (HAL) drivers. We are
interested in porting the firmware implementation to an operating system (OS)
to allow for additional functionality to support environmental data logging.
&lt;a href="https://tockos.org/" target="_blank" rel="noopener">TockOS&lt;/a> is an embedded operating system designed for
running multiple concurrent, mutually distrustful applications on low-memory
and low-power microcontrollers that will be used. TockOS allows for OTA
updates, dynamic app loading, hardware multiplexing, and more. We envision
multiple users utilizing shared ENTS hardware that provides communication and
measurement capabilities. Thus, the initial cost of deploying wireless sensor
networks is reduced.&lt;/p>
&lt;p>The TockOS kernel is written in &lt;a href="https://rust-lang.org/" target="_blank" rel="noopener">Rust&lt;/a> to enhance
security. Userspace apps can be written in either C, C++, or Rust. Development
will be done through a remote development server to access the hardware. See
the following repos for the current status of the project:&lt;/p>
&lt;ul>
&lt;li>Userspace library: &lt;a href="https://github.com/jlab-sensing/libtock-c" target="_blank" rel="noopener">libtock-c&lt;/a>&lt;/li>
&lt;li>Kernel: &lt;a href="https://github.com/jlab-sensing/tock" target="_blank" rel="noopener">tock&lt;/a>&lt;/li>
&lt;li>Baremetal: &lt;a href="https://github.com/jlab-sensing/ENTS-node-firmware" target="_blank" rel="noopener">ENTS-node-firmware&lt;/a>&lt;/li>
&lt;/ul>
&lt;p>Scope of work:&lt;/p>
&lt;ul>
&lt;li>Writing kernel peripheral drivers.
&lt;ul>
&lt;li>Done entirely in Rust.&lt;/li>
&lt;li>Low-level understanding of microcontroller&lt;/li>
&lt;li>Basic kernel functionality knowledge.&lt;/li>
&lt;/ul>
&lt;/li>
&lt;li>Porting baremetal components to userland apps.
&lt;ul>
&lt;li>Involves porting STM HAL calls to TockOS syscalls.&lt;/li>
&lt;li>Primarily done in C.&lt;/li>
&lt;li>Understanding of syscalls.&lt;/li>
&lt;/ul>
&lt;/li>
&lt;/ul></description></item><item><title>Reproducible CXL Emulation</title><link>https://deploy-preview-1007--ucsc-ospo.netlify.app/project/osre26/ucmerced/cxl_emu/</link><pubDate>Fri, 30 Jan 2026 00:00:00 +0000</pubDate><guid>https://deploy-preview-1007--ucsc-ospo.netlify.app/project/osre26/ucmerced/cxl_emu/</guid><description>&lt;p>Compute Express Link (CXL) is an emerging memory interconnect standard that enables shared, coherent memory across CPUs, accelerators, and multiple hosts, unlocking new possibilities in hyperscale, HPC, and disaggregated systems. However, because access to real multi-host CXL hardware is limited, it is difficult for researchers and students to experiment with, evaluate, and reproduce results on advanced CXL topologies.
OCEAN (Open-source CXL Emulation At Hyperscale) [https://github.com/cxl-emu/OCEAN] is a full-stack CXL emulation platform built on QEMU that enables detailed emulation of CXL 3.0 memory systems, including multi-host shared memory pools, coherent fabric topologies, and latency modeling. This project will create reproducible experiment pipelines, automated deployment workflows, and user-friendly tutorials so that others can reliably run and extend CXL emulation experiments without requiring specialized hardware.&lt;/p>
&lt;h3 id="reproducible-cxl-emulation-for-multi-host-memory-systems">Reproducible CXL Emulation for Multi-Host Memory Systems&lt;/h3>
&lt;p>Streamline multi-host CXL emulation without specialized hardware.&lt;/p>
&lt;ul>
&lt;li>&lt;strong>Topics:&lt;/strong> &lt;code>CXL emulation&lt;/code> &lt;code>Memory Systems&lt;/code> &lt;code>Reproducibility&lt;/code>&lt;/li>
&lt;li>&lt;strong>Skills:&lt;/strong> C/C++, Virtualization (QEMU), Scripting, Performance Modeling&lt;/li>
&lt;li>&lt;strong>Difficulty:&lt;/strong> Medium&lt;/li>
&lt;li>&lt;strong>Size:&lt;/strong> Large (350 hours)&lt;/li>
&lt;li>&lt;strong>Mentors:&lt;/strong> &lt;a href="mailto:mrafi@ucmerced.edu">Mujahid Al Rafi&lt;/a>, &lt;a href="https://deploy-preview-1007--ucsc-ospo.netlify.app/author/luanzheng-lenny-guo/">Luanzheng &amp;quot;Lenny&amp;quot; Guo&lt;/a>.&lt;/li>
&lt;/ul>
&lt;p>Tasks:&lt;/p>
&lt;ul>
&lt;li>Create automated deployment scripts and configuration templates for OCEAN-based CXL emulation topologies (single-host and multi-host).&lt;/li>
&lt;li>Develop a standardized experiment harness for running memory performance benchmarks (e.g., OSU micro-benchmarks, STREAM-style tests) in emulated CXL environments.&lt;/li>
&lt;li>Build reproducible experiment pipelines that others can run to evaluate latency, bandwidth, and scaling properties of CXL memory systems.&lt;/li>
&lt;li>Produce tutorials, documentation, and reproducibility artifacts to guide new users through setup, execution, and analysis.&lt;/li>
&lt;li>Package and contribute all scripts, configurations, and documentation back to the OCEAN open-source repository.&lt;/li>
&lt;/ul>
&lt;h3 id="exploring-security-and-isolation-in-cxl-based-memory-systems">Exploring Security and Isolation in CXL-Based Memory Systems&lt;/h3>
&lt;p>Investigate security and isolation properties of CXL-based memory systems using software emulation.&lt;/p>
&lt;ul>
&lt;li>&lt;strong>Topics:&lt;/strong> &lt;code>CXL Systems&lt;/code> &lt;code>Security&lt;/code> &lt;code>Memory Isolation&lt;/code> &lt;code>Side Channel&lt;/code> &lt;code>Emulation&lt;/code>&lt;/li>
&lt;li>&lt;strong>Skills:&lt;/strong> C/C++, Virtualization (QEMU), Scripting, Computer Architecture, Security&lt;/li>
&lt;li>&lt;strong>Difficulty:&lt;/strong> Medium&lt;/li>
&lt;li>&lt;strong>Size:&lt;/strong> Large (350 hours)&lt;/li>
&lt;li>&lt;strong>Mentors:&lt;/strong> &lt;a href="mailto:mrafi@ucmerced.edu">Mujahid Al Rafi&lt;/a>, &lt;a href="https://deploy-preview-1007--ucsc-ospo.netlify.app/author/luanzheng-lenny-guo/">Luanzheng &amp;quot;Lenny&amp;quot; Guo&lt;/a>.&lt;/li>
&lt;/ul>
&lt;p>Tasks:&lt;/p>
&lt;ul>
&lt;li>Study the CXL memory model and fabric architecture to identify potential security and isolation risks in multi-host shared memory environments (e.g., contention, timing variation, and resource interference).&lt;/li>
&lt;li>Set up multi-host or multi-VM CXL emulation environments using OCEAN that mimic realistic multi-tenant deployments.&lt;/li>
&lt;li>Design and implement reproducible micro-benchmarks to measure timing, bandwidth contention, or observable interference through shared CXL memory pools.&lt;/li>
&lt;li>Analyze how fabric configuration choices (e.g., topology, latency injection, memory partitioning, or allocation policies) affect isolation and leakage behavior.&lt;/li>
&lt;li>Explore and prototype mitigation strategies—such as memory partitioning, throttling, or policy-driven allocation—and evaluate their effectiveness using the emulation platform.&lt;/li>
&lt;/ul></description></item><item><title>Omni-ST: Instruction-Driven Any-to-Any Multimodal Modeling for Spatial Transcriptomics</title><link>https://deploy-preview-1007--ucsc-ospo.netlify.app/project/osre26/uci-ics/omni-st/</link><pubDate>Thu, 29 Jan 2026 00:00:00 +0000</pubDate><guid>https://deploy-preview-1007--ucsc-ospo.netlify.app/project/osre26/uci-ics/omni-st/</guid><description>&lt;h2 id="project-description">Project description&lt;/h2>
&lt;p>Spatial transcriptomics (ST) integrates spatially resolved gene expression with tissue morphology, enabling the study of cellular organization, tissue architecture, and disease microenvironments. Modern ST datasets are inherently multimodal, combining histology images (H&amp;amp;E / IF), gene expression vectors, spatial graphs, cell annotations, and free-text pathology descriptions.&lt;/p>
&lt;p>However, most existing ST methods are task-specific and modality-siloed: separate models are trained for image-to-gene prediction, spatial domain identification, cell type classification, or text-based interpretation. This fragmentation limits cross-task generalization and scalability.&lt;/p>
&lt;p>
&lt;figure >
&lt;div class="d-flex justify-content-center">
&lt;div class="w-100" >&lt;img alt="Omni-ST overview" srcset="
/project/osre26/uci-ics/omni-st/omni-st-overview_hu23ddd3d57afcbc47e213a42520991f5c_1307894_4023f7915e2a557bcacee3aecd015061.webp 400w,
/project/osre26/uci-ics/omni-st/omni-st-overview_hu23ddd3d57afcbc47e213a42520991f5c_1307894_8d4e33b30dc811f95fb70a843df58532.webp 760w,
/project/osre26/uci-ics/omni-st/omni-st-overview_hu23ddd3d57afcbc47e213a42520991f5c_1307894_1200x1200_fit_q75_h2_lanczos_3.webp 1200w"
src="https://deploy-preview-1007--ucsc-ospo.netlify.app/project/osre26/uci-ics/omni-st/omni-st-overview_hu23ddd3d57afcbc47e213a42520991f5c_1307894_4023f7915e2a557bcacee3aecd015061.webp"
width="760"
height="664"
loading="lazy" data-zoomable />&lt;/div>
&lt;/div>&lt;/figure>
&lt;/p>
&lt;p>&lt;strong>Omni-ST&lt;/strong> proposes a single &lt;strong>instruction-driven any-to-any multimodal backbone&lt;/strong> that treats each spatial transcriptomics modality as a “language” and formulates all tasks as:&lt;/p>
&lt;p>&lt;strong>Instruction + Input Modality → Output Modality&lt;/strong>&lt;/p>
&lt;p>Natural language is elevated from auxiliary metadata to a &lt;strong>unifying interface&lt;/strong> that specifies task intent, target modality, and biological context. This paradigm enables flexible, interpretable, and extensible spatial reasoning within a single model.&lt;/p>
&lt;hr>
&lt;h3 id="project-idea-instruction-driven-any-to-any-modeling-for-spatial-transcriptomics">Project Idea: Instruction-Driven Any-to-Any Modeling for Spatial Transcriptomics&lt;/h3>
&lt;p>&lt;strong>Topics:&lt;/strong> spatial transcriptomics, multimodal learning, instruction tuning, computational pathology&lt;br>
&lt;strong>Skills:&lt;/strong> PyTorch, deep learning, Transformers, multimodal representation learning&lt;br>
&lt;strong>Difficulty:&lt;/strong> Hard&lt;br>
&lt;strong>Size:&lt;/strong> 350 hours&lt;/p>
&lt;p>&lt;strong>Mentor:&lt;/strong>&lt;/p>
&lt;ul>
&lt;li>&lt;strong>Xi Li&lt;/strong> — &lt;a href="mailto:xil43@uci.edu">mailto:xil43@uci.edu&lt;/a>&lt;/li>
&lt;/ul>
&lt;p>&lt;strong>Essential information:&lt;/strong>&lt;/p>
&lt;ul>
&lt;li>Design a unified multimodal backbone with lightweight modality adapters for histology images, gene expression vectors, spatial graphs, and text.&lt;/li>
&lt;li>Use natural language instructions to condition model behavior, enabling any-to-any translation without task-specific heads.&lt;/li>
&lt;li>Support core tasks including image → gene expression prediction, gene expression → cell type / spatial domain identification, region → text-based biological explanation, and text-based spatial retrieval.&lt;/li>
&lt;li>Evaluate the model across multiple spatial transcriptomics tasks within a single framework, emphasizing generalization and interpretability.&lt;/li>
&lt;li>Develop visualization and interpretation tools such as spatial maps and language-grounded explanations.&lt;/li>
&lt;/ul>
&lt;p>&lt;strong>Expected deliverables:&lt;/strong>&lt;/p>
&lt;ul>
&lt;li>An open-source PyTorch implementation of the Omni-ST framework.&lt;/li>
&lt;li>Unified multitask benchmarks for spatial transcriptomics.&lt;/li>
&lt;li>Visualization and interpretation tools for spatial predictions.&lt;/li>
&lt;li>Documentation and tutorials demonstrating how to add new tasks via instructions.&lt;/li>
&lt;/ul></description></item><item><title>StatWrap</title><link>https://deploy-preview-1007--ucsc-ospo.netlify.app/project/osre26/northwestern/statwrap/</link><pubDate>Thu, 29 Jan 2026 00:00:00 +0000</pubDate><guid>https://deploy-preview-1007--ucsc-ospo.netlify.app/project/osre26/northwestern/statwrap/</guid><description>&lt;p>&lt;a href="https://sites.northwestern.edu/statwrap/" target="_blank" rel="noopener">StatWrap&lt;/a> is a free and open-source assistive, non-invasive discovery and inventory tool to document research projects. It inventories project assets (e.g., code files, data files, manuscripts, documentation) and organizes information without additional input from the user. It also provides structure for users to add searchable and filterable notes connected to files to help communicate metadata about intent and analysis steps.&lt;/p>
&lt;p>At its core, StatWrap helps investigators identify and track changes in a research project as it evolves - which may affect reproducibility. For example: (1) people on the project can change over time, so processes may not be consistently executed due to transitions in employment; (2) data changes over time, due to accruing additional cases, adding new variables, or correcting mistakes in existing data; (3) software (e.g. used for data preparation and statistical analysis) evolves as it is edited, improved, and optimized; and (4) software can break or produce different results due to changes &amp;lsquo;under the hood&amp;rsquo; such as updates to statistical packages, compilers, or interpreters. StatWrap passively and actively documents these changes to support reproducibility.&lt;/p>
&lt;p>Additional information:&lt;/p>
&lt;ul>
&lt;li>&lt;a href="https://sites.northwestern.edu/statwrap/" target="_blank" rel="noopener">StatWrap home&lt;/a>&lt;/li>
&lt;li>&lt;a href="https://github.com/stattag/statwrap" target="_blank" rel="noopener">StatWrap code (GitHub)&lt;/a>&lt;/li>
&lt;/ul>
&lt;h3 id="group-and-individual-customizations">Group and Individual Customizations&lt;/h3>
&lt;ul>
&lt;li>&lt;strong>Topics&lt;/strong>: &lt;code>configuration&lt;/code>, &lt;code>user interface&lt;/code>&lt;/li>
&lt;li>&lt;strong>Skills&lt;/strong>: JavaScript, React&lt;/li>
&lt;li>&lt;strong>Difficulty&lt;/strong>: Medium&lt;/li>
&lt;li>&lt;strong>Size&lt;/strong>: Large (350 hours)&lt;/li>
&lt;li>&lt;strong>Mentor&lt;/strong>: &lt;a href="https://deploy-preview-1007--ucsc-ospo.netlify.app/author/luke-rasmussen/">Luke Rasmussen&lt;/a>, &lt;a href="mailto:ewhitley@northwestern.edu">Eric Whitley&lt;/a>&lt;/li>
&lt;/ul>
&lt;p>The goal of this project is to expand the existing capabilities of StatWrap to provide more flexibility to individual users and groups. Currently, features within StatWrap such as the directory template for creating new projects and the reproducibility checklist are static, meaning everyone who downloads StatWrap has the same configuration. However, each user and team work differently and should be able to configure StatWrap to support their needs.&lt;/p>
&lt;p>When a user creates a new project, StatWrap provides a collection of project templates. These create a directory hierarchy, along with some seed files (e.g., a README.md file in the project root). Different groups have their own conventions for creating project directories. While StatWrap can be released with additional project templates defined, there are many situations in which users would want to keep their project template local. StatWrap should allow a user to create a project template configuration, from scratch or being seeded by the contents of an existing project. A user should then be able to export this configuration, share it with others, and other user should have the ability to import the configuration into their instance of StatWrap.&lt;/p>
&lt;p>Similarly, StatWrap provides a reproducibility checklist that includes six existing checklist items. However, individual users and groups may have their own checklists, including institution-specific steps. Similar to the project template, a user should be able to configure additional items for the checklist. A user should be able to create a &amp;ldquo;checklist template&amp;rdquo; that can be used and applied in multiple projects. A specific project&amp;rsquo;s template should also be modifiable once the checklist has been created.&lt;/p>
&lt;p>The specific tasks of the project include:&lt;/p>
&lt;ul>
&lt;li>Developing a configuration scheme for New Project templates&lt;/li>
&lt;li>Provide a way for a user to import/export a template for New Projects&lt;/li>
&lt;li>Develop a configuration scheme for Reproducibility Checklist questions&lt;/li>
&lt;li>Provide a way for a user to import/export a template for the Reproducibility Checklist&lt;/li>
&lt;li>Develop a configuration scheme for asset (file) attributes&lt;/li>
&lt;li>Develop unit tests and conduct system testing&lt;/li>
&lt;/ul></description></item><item><title>Network Simulation Bridge • Enabling Interactive Network Models</title><link>https://deploy-preview-1007--ucsc-ospo.netlify.app/project/osre26/ucsc/nsb-network-models/</link><pubDate>Wed, 28 Jan 2026 00:00:00 +0000</pubDate><guid>https://deploy-preview-1007--ucsc-ospo.netlify.app/project/osre26/ucsc/nsb-network-models/</guid><description>&lt;p>The Network Simulation Bridge &amp;ndash; &lt;a href="https://github.com/nsb-ucsc/nsb" target="_blank" rel="noopener">NSB&lt;/a> &amp;ndash; is a network co-simulation framework that bridges together applications and network simulators. It enables students, researchers, and developers to prototype their applications and systems on simulated networks. It consists of a message server and client endpoint interfaces which together form a bridge, routing application message payloads through the network simulator. NSB is designed to be extensible through modular interfaces that serve to allow users to contribute new features and modules that suit evolving and emerging use cases. NSB is developed to be application-, network simulator-, and platform-agnostic so that users and developers are empowered to integrate any application front-end with any network simulator back-end, providing versatility and flexibility when used alongside other tools in larger systems and applications.&lt;/p>
&lt;p>NSB was created in-house by the &lt;a href="https://inrg.engineering.ucsc.edu/" target="_blank" rel="noopener">Inter-Networking Research Group&lt;/a> and is now being developed into a more full-featured open-source tool and ecosystem in partnership with the &lt;a href="https://ucsc-ospo.github.io/" target="_blank" rel="noopener">UCSC OSPO&lt;/a> and as part of the &lt;a href="https://www.nsf.gov/funding/opportunities/pose-pathways-enable-open-source-ecosystems" target="_blank" rel="noopener">NSF Pathways to Enable Open-Source Ecosystems&lt;/a> program. In this transition to a more polished and feature-rich product, the next phase of NSB development will involve the engineering of new quality-of-life features, testing and iteration of the core tool itself, and user-centric refinement via implementation in interdisciplinary system models.&lt;/p>
&lt;h3 id="develop-a-user-centric-website-for-nsb">Develop a User-Centric Website for NSB&lt;/h3>
&lt;ul>
&lt;li>&lt;strong>Topics:&lt;/strong> &lt;code>Web Development&lt;/code> &lt;code>Dynamic Updates&lt;/code> &lt;code>UX&lt;/code>&lt;/li>
&lt;li>&lt;strong>Skills:&lt;/strong> web development experience, good communicator, (HTML/CSS), (Javascript)&lt;/li>
&lt;li>&lt;strong>Difficulty:&lt;/strong> Moderate&lt;/li>
&lt;li>&lt;strong>Size:&lt;/strong> Large&lt;/li>
&lt;li>&lt;strong>Mentors:&lt;/strong> &lt;a href="mailto:hkuttive@ucsc.edu">Harikrishna Kuttivelil&lt;/a>&lt;/li>
&lt;/ul>
&lt;p>Develop a clean and welcoming landing page and website for the project. The organization needs to reflect the needs of both users and potential project contributors. This website will be the first impression for people new to the project and should&lt;/p>
&lt;p>Specific tasks:&lt;/p>
&lt;ul>
&lt;li>Work with mentors on understanding the context of the project and the expected needs of the users.&lt;/li>
&lt;li>Port relevant documentation and tutorials from the &lt;a href="https://github.com/nsb-ucsc/nsb" target="_blank" rel="noopener">repository page&lt;/a>, ensuring updates in the repository are reflected in the website.&lt;/li>
&lt;li>Study existing open source product websites and draw insights to include in our own design.&lt;/li>
&lt;li>Design the structure of the website according to best OS, visual design, and accessibility design practices.&lt;/li>
&lt;li>Include visual content that showcases NSB integration and testimonials (if applicable).&lt;/li>
&lt;/ul>
&lt;h3 id="improve-the-user-experience-of-nsb">Improve the User Experience of NSB&lt;/h3>
&lt;ul>
&lt;li>&lt;strong>Topics:&lt;/strong> &lt;code>Software Engineering&lt;/code> &lt;code>User-Centric Development&lt;/code> &lt;code>Visualization&lt;/code> &lt;code>UI/UX&lt;/code> &lt;code>Documentation&lt;/code>&lt;/li>
&lt;li>&lt;strong>Skills:&lt;/strong> package management, toolchain implementation, process automation, technical writing, (visualization), (bash), (Python), (C++)&lt;/li>
&lt;li>&lt;strong>Difficulty:&lt;/strong> Moderate&lt;/li>
&lt;li>&lt;strong>Size:&lt;/strong> Medium&lt;/li>
&lt;li>&lt;strong>Mentors:&lt;/strong> &lt;a href="mailto:hkuttive@ucsc.edu">Harikrishna Kuttivelil&lt;/a>&lt;/li>
&lt;/ul>
&lt;p>Our goal has always been to keep NSB streamlined and out of the way of the users and developers. In line with that, we want our tool to be easily available and installable, and we want the experience of using it to feel minimal and non-intrusive while providing sufficient observability of NSB&amp;rsquo;s internals for those who want it.&lt;/p>
&lt;p>Specific tasks:&lt;/p>
&lt;ul>
&lt;li>Work with mentors and potential users on identifying aspects of the user experience that can refined for better quality-of-life experiences.&lt;/li>
&lt;li>Verify and iterate on existing software packaging methods for NSB to ensure that tool setup is stress-free.&lt;/li>
&lt;li>Refine and update existing documentation and tutorials to reflect improvements in the setup, installation, and usage processes.&lt;/li>
&lt;li>Work with mentors and other contributors to work backwards from what the user wants to see to design the user interface.&lt;/li>
&lt;li>Work with other contributors (see below) to develop a &lt;em>Network-in-a-Box&lt;/em> experience with NSB.&lt;/li>
&lt;/ul>
&lt;h3 id="create-a-network-in-a-box-experience-with-nsb">Create a &lt;em>Network-in-a-Box&lt;/em> Experience with NSB&lt;/h3>
&lt;ul>
&lt;li>&lt;strong>Topics:&lt;/strong> &lt;code>Software Engineering&lt;/code>, &lt;code>Simulation&lt;/code>, &lt;code>System Modeling&lt;/code>, &lt;code>System Design&lt;/code>, &lt;code>Visualization&lt;/code>, &lt;code>UI/UX&lt;/code>&lt;/li>
&lt;li>&lt;strong>Skills:&lt;/strong> software integration and interfacing, toolchain implementation, process automation, C++, (visualization), (LLM-enabled code generation), (technical writing)&lt;/li>
&lt;li>&lt;strong>Difficulty:&lt;/strong> Challenging&lt;/li>
&lt;li>&lt;strong>Size:&lt;/strong> Large&lt;/li>
&lt;li>&lt;strong>Mentors:&lt;/strong> &lt;a href="mailto:hkuttive@ucsc.edu">Harikrishna Kuttivelil&lt;/a>&lt;/li>
&lt;/ul>
&lt;p>NSB was originally designed for networking graduate students to interface with application-layer programs. But since then, there&amp;rsquo;s been more of an appetite for a simpler &lt;em>network-in-a-box&lt;/em> approach that would allow users to quickly deploy baseline or generated network simulations that are ready for use with NSB.&lt;/p>
&lt;p>Specific tasks:&lt;/p>
&lt;ul>
&lt;li>Learn how to use one of the major open-source network simulators (&lt;a href="https://www.nsnam.org/" target="_blank" rel="noopener">ns3&lt;/a> or &lt;a href="https://omnetpp.org/" target="_blank" rel="noopener">OMNeT++&lt;/a>).&lt;/li>
&lt;li>Work with mentors in designing a simpler, minimal user experience of operating NSB.&lt;/li>
&lt;li>Develop tools to automatically create network simulations given input parameters (type of network, number of nodes, description of infrastructure).&lt;/li>
&lt;li>Create documentation aimed at new users.&lt;/li>
&lt;li>Implement or embed network visualizations to enrich the user experience.&lt;/li>
&lt;/ul>
&lt;h3 id="implement-networked-system-models-to-evaluate-quality-of-nsb">Implement Networked System Models to Evaluate Quality of NSB&lt;/h3>
&lt;ul>
&lt;li>&lt;strong>Topics:&lt;/strong> &lt;code>System Modeling&lt;/code> &lt;code>Simulation&lt;/code> &lt;code>System Design&lt;/code> &lt;code>Software Development&lt;/code> &lt;code>Product Testing&lt;/code>&lt;/li>
&lt;li>&lt;strong>Skills:&lt;/strong> software integration, good communication, qualitative research, (proficiency in Python and/or C++), (processing scientific and technical literature)&lt;/li>
&lt;li>&lt;strong>Difficulty:&lt;/strong> Challenging&lt;/li>
&lt;li>&lt;strong>Size:&lt;/strong> Large&lt;/li>
&lt;li>&lt;strong>Mentors:&lt;/strong> &lt;a href="mailto:hkuttive@ucsc.edu">Harikrishna Kuttivelil&lt;/a>&lt;/li>
&lt;/ul>
&lt;p>NSB is a relatively new tool and has not been extensively tested outside of the core contributors, who know a bit too much about the tool. We need to better understand what external user and contributor experience will be like, and the best way to do that is to start developing with NSB to build models of connected systems, i.e., sensor networks, smart homes, smart farms, etc.&lt;/p>
&lt;p>Specific tasks:&lt;/p>
&lt;ul>
&lt;li>Research academic literature and relevant works to identify relevant distributed applications to model.&lt;/li>
&lt;li>Work with mentors and collaborators to plan implementation of selected system models.&lt;/li>
&lt;li>Track and report issues and concerns in quality-of-life experiences, critical errors, or difficulties.&lt;/li>
&lt;li>Work with mentors and contributors to address issues and concerns.&lt;/li>
&lt;li>Refine and update existing documentation and tutorials to reflect improvements in the setup, installation, and usage processes.&lt;/li>
&lt;li>Work with other contributors (see below) in reviewing and cross-referencing model implementations.&lt;/li>
&lt;/ul>
&lt;h3 id="model-autonomous-vehicle-networks-to-drive-new-feature-development-in-nsb">Model Autonomous Vehicle Networks to Drive New Feature Development in NSB&lt;/h3>
&lt;ul>
&lt;li>&lt;strong>Topics:&lt;/strong> &lt;code>System Modeling&lt;/code> &lt;code>Simulation&lt;/code> &lt;code>System Design&lt;/code> &lt;code>Software Development&lt;/code>&lt;/li>
&lt;li>&lt;strong>Skills:&lt;/strong> requirement-based software design, message parsing interfaces, server-client communication, (proficiency in Python and/or C++), (processing scientific and technical literature)&lt;/li>
&lt;li>&lt;strong>Difficulty:&lt;/strong> Challenging&lt;/li>
&lt;li>&lt;strong>Size:&lt;/strong> Large&lt;/li>
&lt;li>&lt;strong>Mentors:&lt;/strong> &lt;a href="mailto:hkuttive@ucsc.edu">Harikrishna Kuttivelil&lt;/a>&lt;/li>
&lt;/ul>
&lt;p>NSB today serves its named purpose &amp;ndash; message relaying. However, modeling complex systems can sometimes involving synchronizing other simulation features, like &lt;em>mobility&lt;/em> when dealing with vehivle networks. Implementing a generic layer of being able to synchronize user-defined features across endpoints would be a powerful, enabling feature in NSB. In the process, we may also uncover opportunities for improving the NSB developer experience.&lt;/p>
&lt;p>Specific tasks:&lt;/p>
&lt;ul>
&lt;li>Research academic literature and relevant works to identify and design potential autonomous vehicle network models.&lt;/li>
&lt;li>Work with mentors and collaborators to iterate on system designs to ensure it serves the purpose of furthering NSB development.&lt;/li>
&lt;li>Help mentors design and develop the &lt;em>new&lt;/em> feature synchronization feature in NSB, driven by the autonomous vehicle system model.&lt;/li>
&lt;li>Develop and iterate feature synchronization, using mobility as the synchronized feature.&lt;/li>
&lt;li>Create documentation and tutorials to serve as resources for future users, contributors, and developers.&lt;/li>
&lt;li>Work with other contributors (see above) in reviewing and cross-referencing model implementations.&lt;/li>
&lt;/ul></description></item><item><title>Peersky Browser</title><link>https://deploy-preview-1007--ucsc-ospo.netlify.app/project/osre26/ucsc/peersky/</link><pubDate>Mon, 26 Jan 2026 12:00:00 -0800</pubDate><guid>https://deploy-preview-1007--ucsc-ospo.netlify.app/project/osre26/ucsc/peersky/</guid><description>&lt;p>&lt;a href="https://peersky.p2plabs.xyz/" target="_blank" rel="noopener">Peersky Browser&lt;/a> is an experimental personal gatekeeper to a new way of accessing web content. In a world where a handful of big companies control most of the internet, Peersky leverages distributed web technologies—&lt;a href="https://ipfs.tech/" target="_blank" rel="noopener">IPFS&lt;/a>, &lt;a href="https://holepunch.to/" target="_blank" rel="noopener">Hypercore&lt;/a>, and &lt;a href="https://www.bittorrent.com/" target="_blank" rel="noopener">BitTorrent&lt;/a> return control to the users. With integrated local P2P applications, Peersky offers a fresh, community-driven approach to browsing.&lt;/p>
&lt;h3 id="implement-p2p-extension-store">Implement P2P Extension Store&lt;/h3>
&lt;ul>
&lt;li>&lt;strong>Topics:&lt;/strong> &lt;code>Browser Extensions&lt;/code>, &lt;code>P2P&lt;/code>, &lt;code>Electron&lt;/code>, &lt;code>IPFS&lt;/code>, &lt;code>Hypercore&lt;/code>&lt;/li>
&lt;li>&lt;strong>Skills:&lt;/strong> JavaScript, Electron.js, HTML/CSS, P2P&lt;/li>
&lt;li>&lt;strong>Difficulty:&lt;/strong> Moderate&lt;/li>
&lt;li>&lt;strong>Size:&lt;/strong> Medium (175 hours)&lt;/li>
&lt;li>&lt;strong>Mentors:&lt;/strong> &lt;a href="https://deploy-preview-1007--ucsc-ospo.netlify.app/author/akhilesh-thite/">Akhilesh Thite&lt;/a>&lt;/li>
&lt;/ul>
&lt;p>Build a decentralized extension distribution flow that archives WebExtensions into a predictable P2P-friendly layout and installs directly from P2P URLs.&lt;/p>
&lt;p>&lt;strong>Tasks:&lt;/strong>&lt;/p>
&lt;ul>
&lt;li>&lt;strong>Define the P2P extension layout:&lt;/strong>
&lt;ul>
&lt;li>Standardize &lt;code>/extensions/{name}/{version}/extension.zip&lt;/code> and &lt;code>/extensions/{name}/index.json&lt;/code>.&lt;/li>
&lt;/ul>
&lt;/li>
&lt;li>&lt;strong>Design install compatibility for P2P URLs:&lt;/strong>
&lt;ul>
&lt;li>Support &lt;code>peersky://extensions/...&lt;/code> and P2P links from IPFS or Hypercore.&lt;/li>
&lt;/ul>
&lt;/li>
&lt;li>&lt;strong>Archive Chrome Web Store extensions to P2P:&lt;/strong>
&lt;ul>
&lt;li>Use &lt;a href="https://github.com/akhileshthite/chrome-extension-fetch" target="_blank" rel="noopener">chrome-extension-fetch&lt;/a> to fetch CRX, convert to ZIP, and store it in the layout.&lt;/li>
&lt;li>Update &lt;code>index.json&lt;/code> with metadata like version, &lt;code>P2P_URL&lt;/code>, and &lt;code>fetchedAt&lt;/code>.&lt;/li>
&lt;li>Publish the folder to IPFS or Hypercore and feed the link into the install flow.&lt;/li>
&lt;/ul>
&lt;/li>
&lt;li>&lt;strong>Add settings and trust model:&lt;/strong>
&lt;ul>
&lt;li>Add a “Load from P2P” settings toggle.&lt;/li>
&lt;li>Support curated extension hoards (&lt;code>index.json&lt;/code>) and automated updates.&lt;/li>
&lt;li>Clarify integrity assumptions and sandboxing expectations.&lt;/li>
&lt;/ul>
&lt;/li>
&lt;/ul>
&lt;p>&lt;strong>More details in the issue:&lt;/strong> &lt;a href="https://github.com/p2plabsxyz/peersky-browser/issues/42" target="_blank" rel="noopener">https://github.com/p2plabsxyz/peersky-browser/issues/42&lt;/a>&lt;/p>
&lt;h3 id="backup--restore-system-p2p-json--tabs-restore">Backup &amp;amp; Restore System (P2P JSON + Tabs Restore)&lt;/h3>
&lt;ul>
&lt;li>&lt;strong>Topics:&lt;/strong> &lt;code>P2P&lt;/code>, &lt;code>Backup&lt;/code>, &lt;code>Session Restore&lt;/code>, &lt;code>Electron&lt;/code>, &lt;code>Onboarding&lt;/code>&lt;/li>
&lt;li>&lt;strong>Skills:&lt;/strong> JavaScript, Electron.js, HTML/CSS, P2P&lt;/li>
&lt;li>&lt;strong>Difficulty:&lt;/strong> Moderate&lt;/li>
&lt;li>&lt;strong>Size:&lt;/strong> Medium (175 hours)&lt;/li>
&lt;li>&lt;strong>Mentors:&lt;/strong> &lt;a href="https://deploy-preview-1007--ucsc-ospo.netlify.app/author/akhilesh-thite/">Akhilesh Thite&lt;/a>&lt;/li>
&lt;/ul>
&lt;p>Implement a backup and restore pipeline for Peersky’s P2P app data and session state, including an onboarding import flow for tabs from other browsers.&lt;/p>
&lt;p>&lt;strong>Tasks:&lt;/strong>&lt;/p>
&lt;ul>
&lt;li>&lt;strong>Generate a P2P backup bundle:&lt;/strong>
&lt;ul>
&lt;li>Create a single &lt;code>.zip&lt;/code> that contains &lt;code>lastOpened.json&lt;/code>, &lt;code>tabs.json&lt;/code>, &lt;code>ensCache.json&lt;/code>, and the &lt;code>ipfs/&lt;/code> and &lt;code>hyper/&lt;/code> directories.&lt;/li>
&lt;li>Add an option to generate a CID for the backup zip for instant sharing.&lt;/li>
&lt;/ul>
&lt;/li>
&lt;li>&lt;strong>Restore from settings:&lt;/strong>
&lt;ul>
&lt;li>Upload a P2P backup zip file.&lt;/li>
&lt;li>Load a backup from an IPFS or Hyper CID.&lt;/li>
&lt;li>Import Chrome/Firefox tab exports produced by a helper extension.&lt;/li>
&lt;/ul>
&lt;/li>
&lt;li>&lt;strong>Define the helper extension export format:&lt;/strong>
&lt;ul>
&lt;li>Create a small extension under &lt;a href="https://github.com/p2plabsxyz/" target="_blank" rel="noopener">p2plabsxyz&lt;/a> to export windows and tabs (URLs, titles, window grouping, active tab indexes).&lt;/li>
&lt;li>Ensure the export format is compatible with Peersky’s import pipeline.&lt;/li>
&lt;/ul>
&lt;/li>
&lt;li>&lt;strong>Add onboarding import flow:&lt;/strong>
&lt;ul>
&lt;li>Show &lt;code>onboarding.html&lt;/code> on first launch and prompt “Import tabs from another browser?”.&lt;/li>
&lt;li>Guide users to install the helper extension and import the generated file.&lt;/li>
&lt;/ul>
&lt;/li>
&lt;li>&lt;strong>Align with existing persistence:&lt;/strong>
&lt;ul>
&lt;li>Reuse &lt;code>lastOpened.json&lt;/code>, &lt;code>tabs.json&lt;/code>, and &lt;code>peersky-browser-tabs&lt;/code> localStorage for restores.&lt;/li>
&lt;/ul>
&lt;/li>
&lt;/ul>
&lt;p>&lt;strong>More details in the issue:&lt;/strong> &lt;a href="https://github.com/p2plabsxyz/peersky-browser/issues/60" target="_blank" rel="noopener">https://github.com/p2plabsxyz/peersky-browser/issues/60&lt;/a>&lt;/p></description></item><item><title>Scenic: A Language for Design and Verification of Autonomous Cyber-Physical Systems</title><link>https://deploy-preview-1007--ucsc-ospo.netlify.app/project/osre26/ucsc/scenic/</link><pubDate>Sat, 24 Jan 2026 00:00:00 +0000</pubDate><guid>https://deploy-preview-1007--ucsc-ospo.netlify.app/project/osre26/ucsc/scenic/</guid><description>&lt;p>&lt;a href="https://scenic-lang.org/" target="_blank" rel="noopener">Scenic&lt;/a> is a probabilistic programming language for the design and verification of autonomous cyber-physical systems like self-driving cars.
Scenic allows users to define &lt;em>scenarios&lt;/em> for testing or training their system by putting a probability distribution on the system&amp;rsquo;s environment: the positions, orientations, and other properties of objects and agents, as well as their behaviors over time.
Sampling these scenarios and running them in a simulator yields synthetic data which can be used to train or test a system.
Since Scenic was released open-source in 2019, our group and many others in academia have used Scenic to find, diagnose, and fix bugs in autonomous cars, aircraft, robots, and other kinds of systems.
In industry, it is being used by companies including Boeing, Meta, Deutsche Bahn, and Toyota in domains spanning autonomous driving, aviation, household robotics, railways, maritime, and virtual reality.&lt;/p>
&lt;p>Our long-term goal is for Scenic to become a widely-used common representation and toolkit supporting the entire design lifecycle of AI-based cyber-physical systems.
Towards this end, we have many summer projects available, ranging from adding new application domains to working on the Scenic compiler and sampler:&lt;/p>
&lt;ol>
&lt;li>Extensions to the Scenic driving domain&lt;/li>
&lt;li>Interfacing Scenic to new simulators&lt;/li>
&lt;li>Scenic distribution visualizer&lt;/li>
&lt;/ol>
&lt;p>See the sections below for details.&lt;/p>
&lt;h3 id="extensions-to-the-scenic-driving-domain">Extensions to the Scenic Driving Domain&lt;/h3>
&lt;ul>
&lt;li>&lt;strong>Topics:&lt;/strong> &lt;code>Autonomous Driving&lt;/code> &lt;code>3D modeling&lt;/code>&lt;/li>
&lt;li>&lt;strong>Skills:&lt;/strong> Python; basic vector geometry&lt;/li>
&lt;li>&lt;strong>Difficulty:&lt;/strong> Moderate&lt;/li>
&lt;li>&lt;strong>Size:&lt;/strong> Medium or Large (175 or 350 hours)&lt;/li>
&lt;li>&lt;strong>Mentors:&lt;/strong> &lt;a href="https://deploy-preview-1007--ucsc-ospo.netlify.app/author/daniel-fremont/">Daniel Fremont&lt;/a>, &lt;a href="https://deploy-preview-1007--ucsc-ospo.netlify.app/author/eric-vin/">Eric Vin&lt;/a>&lt;/li>
&lt;/ul>
&lt;p>Scenic scenarios written to test autonomous vehicles use the &lt;a href="https://docs.scenic-lang.org/en/latest/modules/scenic.domains.driving.html" target="_blank" rel="noopener">driving domain&lt;/a>, a Scenic library defining driving-specific concepts including cars, pedestrians, roads, lanes, and intersections.
The library extracts information about road networks, such as the shapes of lanes, from files in the standard &lt;a href="https://www.asam.net/standards/detail/opendrive/" target="_blank" rel="noopener">OpenDRIVE&lt;/a> format.&lt;/p>
&lt;p>There are several potential goals of this project, including:&lt;/p>
&lt;ul>
&lt;li>Supporting importing complex object information from simulators like CARLA.&lt;/li>
&lt;li>Extending the domain to incorporate additional metadata, such as highway entrances and exits.&lt;/li>
&lt;li>Fixing various bugs and limitations that exist in the driving domain (e.g. &lt;a href="https://github.com/BerkeleyLearnVerify/Scenic/issues/274" target="_blank" rel="noopener">Issue #274&lt;/a> and &lt;a href="https://github.com/BerkeleyLearnVerify/Scenic/issues/295" target="_blank" rel="noopener">Issue #295&lt;/a>).&lt;/li>
&lt;/ul>
&lt;h3 id="interfacing-scenic-to-new-simulators">Interfacing Scenic to New Simulators&lt;/h3>
&lt;ul>
&lt;li>&lt;strong>Topics:&lt;/strong> &lt;code>Simulation&lt;/code> &lt;code>Autonomous Driving&lt;/code>&lt;/li>
&lt;li>&lt;strong>Skills:&lt;/strong> Python&lt;/li>
&lt;li>&lt;strong>Difficulty:&lt;/strong> Moderate&lt;/li>
&lt;li>&lt;strong>Size:&lt;/strong> Medium or Large (175 or 350 hours)&lt;/li>
&lt;li>&lt;strong>Mentors:&lt;/strong> &lt;a href="https://deploy-preview-1007--ucsc-ospo.netlify.app/author/daniel-fremont/">Daniel Fremont&lt;/a>, &lt;a href="https://deploy-preview-1007--ucsc-ospo.netlify.app/author/eric-vin/">Eric Vin&lt;/a>&lt;/li>
&lt;/ul>
&lt;p>Scenic is designed to be &lt;a href="https://docs.scenic-lang.org/en/latest/new_simulator.html" target="_blank" rel="noopener">easily-interfaced to new simulators&lt;/a>.
Depending on student interest, we could pick a simulator which would open up new kinds of applications for Scenic and write an interface for it.
Some possibilities include:&lt;/p>
&lt;ul>
&lt;li>The &lt;a href="https://github.com/tier4/AWSIM" target="_blank" rel="noopener">AWSIM&lt;/a> driving simulator (to allow testing the &lt;a href="https://autoware.org/" target="_blank" rel="noopener">Autoware&lt;/a> open-source autonomous driving software stack)&lt;/li>
&lt;li>The &lt;a href="https://www.ipg-automotive.com/solutions/product-portfolio/carmaker/" target="_blank" rel="noopener">CarMaker&lt;/a> driving simulator&lt;/li>
&lt;/ul>
&lt;p>The goal of the project would be to create an interface between Scenic and the new simulator and write scenarios demonstrating it.
If time allows, we could do a case study on a realistic system for publication at an academic conference.&lt;/p>
&lt;h3 id="tool-to-visualize-scenario-distributions">Tool to Visualize Scenario Distributions&lt;/h3>
&lt;ul>
&lt;li>&lt;strong>Topics:&lt;/strong> &lt;code>Visualization&lt;/code>&lt;/li>
&lt;li>&lt;strong>Skills:&lt;/strong> Python; basic visualization and graphics&lt;/li>
&lt;li>&lt;strong>Difficulty:&lt;/strong> Moderate&lt;/li>
&lt;li>&lt;strong>Size:&lt;/strong> Medium or Large (175 or 350 hours)&lt;/li>
&lt;li>&lt;strong>Mentors:&lt;/strong> &lt;a href="https://deploy-preview-1007--ucsc-ospo.netlify.app/author/daniel-fremont/">Daniel Fremont&lt;/a>, &lt;a href="https://deploy-preview-1007--ucsc-ospo.netlify.app/author/eric-vin/">Eric Vin&lt;/a>&lt;/li>
&lt;/ul>
&lt;p>A Scenic scenario represents a distribution over scenes, but it can be difficult to interpret what exactly this distribution represents. Being able to visualize this distribution would be helpful for understanding and reasoning about Scenarios.&lt;/p>
&lt;p>The goal of this project would be to build on an existing prototype for visualizing these distributions, and to create a tool that can be used by the wider Scenic community.&lt;/p></description></item><item><title>CauST: Causal Gene Intervention for Robust Spatial Domain Identification</title><link>https://deploy-preview-1007--ucsc-ospo.netlify.app/project/osre26/uci/caust/</link><pubDate>Wed, 21 Jan 2026 00:00:00 +0000</pubDate><guid>https://deploy-preview-1007--ucsc-ospo.netlify.app/project/osre26/uci/caust/</guid><description>&lt;ul>
&lt;li>&lt;strong>Topics:&lt;/strong> spatial transcriptomics, spatial domain identification, causal inference, gene intervention&lt;/li>
&lt;li>&lt;strong>Skills:&lt;/strong>
&lt;ul>
&lt;li>&lt;strong>Programming Languages:&lt;/strong> Python (PyTorch preferred)&lt;/li>
&lt;li>&lt;strong>Machine Learning:&lt;/strong> causal inference, representation learning, clustering&lt;/li>
&lt;li>&lt;strong>Data Analysis:&lt;/strong> spatial transcriptomics preprocessing and evaluation (ARI, cross-slice generalization)&lt;/li>
&lt;li>&lt;strong>Bioinformatics Knowledge (preferred):&lt;/strong> spatial transcriptomics, scRNA-seq, gene perturbation analysis&lt;/li>
&lt;/ul>
&lt;/li>
&lt;li>&lt;strong>Difficulty:&lt;/strong> Advanced&lt;/li>
&lt;li>&lt;strong>Size:&lt;/strong> Large (350 hours)&lt;/li>
&lt;li>&lt;strong>Mentors:&lt;/strong> &lt;a href="https://deploy-preview-1007--ucsc-ospo.netlify.app/author/lijinghua-zhang/">Lijinghua Zhang&lt;/a> (contact person)&lt;/li>
&lt;/ul>
&lt;h3 id="project-idea-description">&lt;strong>Project Idea Description&lt;/strong>&lt;/h3>
&lt;p>Spatial domain identification is a core task in spatial transcriptomics (ST), aiming to segment tissue sections into biologically meaningful regions based on spatially resolved gene expression profiles. These spatial domains often correspond to anatomical layers, functional niches, or microenvironmental states, and are widely used as the basis for downstream biological interpretation.&lt;/p>
&lt;p>Despite strong empirical performance, most existing spatial domain identification methods rely on &lt;strong>purely correlational gene signals&lt;/strong>. Genes are selected or weighted based on association with spatial patterns, without distinguishing whether they &lt;em>causally drive&lt;/em> domain formation or merely reflect downstream or confounded effects. As a result, current models often suffer from limited robustness and poor generalization across tissue sections or donors.&lt;/p>
&lt;h3 id="problem-correlation-driven-gene-usage-and-limited-generalization">&lt;strong>Problem: Correlation-Driven Gene Usage and Limited Generalization&lt;/strong>&lt;/h3>
&lt;p>In standard pipelines, gene expression features are typically used wholesale or filtered using heuristic criteria (e.g., highly variable genes). However, many genes that are strongly correlated with spatial domains are not causally responsible for domain structure. Including such non-causal or confounded genes can:&lt;/p>
&lt;ul>
&lt;li>Reduce robustness across slices and donors&lt;/li>
&lt;li>Obscure true domain-driving biological signals&lt;/li>
&lt;li>Limit interpretability of spatial domain assignments&lt;/li>
&lt;/ul>
&lt;p>Empirically, domain identification performance often degrades substantially in cross-slice or cross-donor evaluation settings, underscoring the need for causally informed feature selection.&lt;/p>
&lt;h3 id="proposed-solution-caust">&lt;strong>Proposed Solution: CauST&lt;/strong>&lt;/h3>
&lt;p>This project proposes &lt;strong>CauST&lt;/strong>, a &lt;strong>Causal Gene Intervention framework&lt;/strong> for robust spatial domain identification.&lt;/p>
&lt;p>CauST aims to identify &lt;strong>domain-driving genes&lt;/strong> by estimating their causal influence on spatial domain assignments via &lt;strong>in-silico gene interventions&lt;/strong>. Instead of relying on observational correlations, CauST approximates counterfactual gene knockouts by perturbing individual gene expressions while controlling for confounding factors.&lt;/p>
&lt;p>In addition, CauST leverages &lt;strong>cross-slice invariance&lt;/strong> as a practical criterion for causal gene discovery, prioritizing genes whose effects on spatial domain identification remain stable across tissue sections and donors.&lt;/p>
&lt;p>By filtering or reweighting genes based on estimated causal influence, CauST improves the robustness, generalizability, and interpretability of spatial domain identification models.&lt;/p>
&lt;h3 id="project-objectives">&lt;strong>Project Objectives&lt;/strong>&lt;/h3>
&lt;ol>
&lt;li>&lt;strong>Causal Gene Effect Estimation&lt;/strong>
&lt;ul>
&lt;li>Design in-silico intervention strategies to estimate gene-level causal effects on spatial domain assignments.&lt;/li>
&lt;/ul>
&lt;/li>
&lt;li>&lt;strong>Invariant Effect Analysis&lt;/strong>
&lt;ul>
&lt;li>Identify genes with stable effects across tissue sections or donors.&lt;/li>
&lt;/ul>
&lt;/li>
&lt;li>&lt;strong>Causal Gene Filtering&lt;/strong>
&lt;ul>
&lt;li>Develop filtering or reweighting schemes based on estimated causal influence.&lt;/li>
&lt;/ul>
&lt;/li>
&lt;li>&lt;strong>Integration with Existing Methods&lt;/strong>
&lt;ul>
&lt;li>Integrate CauST into state-of-the-art spatial domain identification pipelines.&lt;/li>
&lt;/ul>
&lt;/li>
&lt;li>&lt;strong>Evaluation and Validation&lt;/strong>
&lt;ul>
&lt;li>Benchmark robustness, cross-slice generalization, and interpretability on public spatial transcriptomics datasets.&lt;/li>
&lt;/ul>
&lt;/li>
&lt;/ol>
&lt;h3 id="project-deliverables">&lt;strong>Project Deliverables&lt;/strong>&lt;/h3>
&lt;ol>
&lt;li>&lt;strong>CauST Framework Implementation&lt;/strong>
&lt;ul>
&lt;li>Open-source Python implementation compatible with common spatial transcriptomics toolchains.&lt;/li>
&lt;/ul>
&lt;/li>
&lt;li>&lt;strong>Causal Gene Benchmarks&lt;/strong>
&lt;ul>
&lt;li>Quantitative evaluation of causal gene filtering and its impact on domain identification.&lt;/li>
&lt;/ul>
&lt;/li>
&lt;li>&lt;strong>Visualization Tools&lt;/strong>
&lt;ul>
&lt;li>Tools for visualizing gene interventions, causal scores, and spatial effects.&lt;/li>
&lt;/ul>
&lt;/li>
&lt;li>&lt;strong>Documentation and Tutorials&lt;/strong>
&lt;ul>
&lt;li>Clear examples enabling adoption of CauST by the broader community.&lt;/li>
&lt;/ul>
&lt;/li>
&lt;/ol>
&lt;h3 id="impact">&lt;strong>Impact&lt;/strong>&lt;/h3>
&lt;p>CauST introduces a causally grounded perspective to spatial domain identification by explicitly modeling gene-level interventions. By shifting from correlation-driven gene usage to causal gene selection, this project improves robustness, generalizability, and biological interpretability in spatial transcriptomics analysis. CauST has the potential to serve as a foundational framework for integrating causal reasoning into spatial omics representation learning.&lt;/p></description></item><item><title>Agent4Target: An Agent-based Evidence Aggregation Toolkit for Therapeutic Target Identification</title><link>https://deploy-preview-1007--ucsc-ospo.netlify.app/project/osre26/uci/agent4target/</link><pubDate>Tue, 20 Jan 2026 00:00:00 +0000</pubDate><guid>https://deploy-preview-1007--ucsc-ospo.netlify.app/project/osre26/uci/agent4target/</guid><description>&lt;ul>
&lt;li>&lt;strong>Topics:&lt;/strong> therapeutic target identification, drug discovery, evidence aggregation, AI agents, biomedical knowledge integration&lt;/li>
&lt;li>&lt;strong>Skills:&lt;/strong>
&lt;ul>
&lt;li>&lt;strong>Programming Languages:&lt;/strong> Python; experience with modern ML tooling preferred&lt;/li>
&lt;li>&lt;strong>Machine Learning / AI:&lt;/strong> agent-based systems, workflow orchestration, weak supervision (basic), representation learning&lt;/li>
&lt;li>&lt;strong>Software Engineering:&lt;/strong> modular system design, APIs, CLI tools, documentation&lt;/li>
&lt;li>&lt;strong>Biomedical Knowledge (preferred):&lt;/strong> familiarity with drug–target databases (e.g., PHAROS, DepMap, Open Targets)&lt;/li>
&lt;/ul>
&lt;/li>
&lt;li>&lt;strong>Difficulty:&lt;/strong> Advanced&lt;/li>
&lt;li>&lt;strong>Size:&lt;/strong> Large (350 hours)&lt;/li>
&lt;li>&lt;strong>Mentors:&lt;/strong> &lt;a href="https://deploy-preview-1007--ucsc-ospo.netlify.app/author/ziheng-duan/">Ziheng Duan&lt;/a> (contact person)&lt;/li>
&lt;/ul>
&lt;h3 id="project-idea-description">&lt;strong>Project Idea Description&lt;/strong>&lt;/h3>
&lt;p>Identifying and prioritizing high-quality therapeutic targets is a foundational yet challenging task in drug discovery. Modern target identification relies on aggregating heterogeneous evidence from multiple sources, including genetic perturbation screens, disease associations, chemical biology, and biomedical literature. These evidence sources are highly fragmented, noisy, and heterogeneous in both format and reliability.&lt;/p>
&lt;p>While large language models and AI agents have recently shown promise in automating scientific workflows, many existing approaches focus on end-to-end prediction or conversational interfaces. Such systems are often difficult to reproduce, extend, or integrate into existing research pipelines, limiting their practical adoption by the biomedical community.&lt;/p>
&lt;p>This project proposes &lt;strong>Agent4Target&lt;/strong>, an &lt;strong>agent-based evidence aggregation toolkit&lt;/strong> that reframes therapeutic target identification as a &lt;strong>structured, modular workflow&lt;/strong>. Instead of using agents for free-form reasoning, Agent4Target employs agents as &lt;strong>orchestrated components&lt;/strong> that systematically collect, normalize, score, and explain evidence supporting candidate therapeutic targets.&lt;/p>
&lt;p>The goal is to deliver a &lt;strong>reusable, open-source toolchain&lt;/strong> that can be integrated into diverse drug discovery workflows, independent of any single downstream prediction model or publication.&lt;/p>
&lt;hr>
&lt;h3 id="key-idea-and-technical-approach">&lt;strong>Key Idea and Technical Approach&lt;/strong>&lt;/h3>
&lt;p>Agent4Target models target identification as a multi-stage, agent-driven pipeline, coordinated by a central orchestrator:&lt;/p>
&lt;ol>
&lt;li>
&lt;p>&lt;strong>Evidence Collector Agents&lt;/strong>&lt;br>
Specialized agents retrieve target-level evidence from heterogeneous sources, such as:&lt;/p>
&lt;ul>
&lt;li>Genetic perturbation and dependency data (e.g., DepMap)&lt;/li>
&lt;li>Target annotation and development status (e.g., PHAROS)&lt;/li>
&lt;li>Disease association scores (e.g., Open Targets)&lt;/li>
&lt;li>Automatically summarized literature evidence&lt;/li>
&lt;/ul>
&lt;/li>
&lt;li>
&lt;p>&lt;strong>Normalization &amp;amp; Scoring Agent&lt;/strong>&lt;br>
Collected evidence is converted into a unified, structured schema using typed data models (e.g., JSON / Pydantic).&lt;br>
This agent performs:&lt;/p>
&lt;ul>
&lt;li>Evidence normalization across sources&lt;/li>
&lt;li>Confidence-aware scoring and aggregation&lt;/li>
&lt;li>Optional weighting or calibration strategies&lt;/li>
&lt;/ul>
&lt;/li>
&lt;li>
&lt;p>&lt;strong>Explanation Agent&lt;/strong>&lt;br>
Rather than free-text generation, this agent produces &lt;strong>structured explanations&lt;/strong> that explicitly link scores to supporting evidence, enabling transparency and interpretability for downstream users.&lt;/p>
&lt;/li>
&lt;li>
&lt;p>&lt;strong>Workflow Orchestrator&lt;/strong>&lt;br>
A lightweight orchestration layer (e.g., LangGraph or a state-machine-based controller) manages agent execution, dependencies, and failure handling, ensuring reproducibility and extensibility.&lt;/p>
&lt;/li>
&lt;/ol>
&lt;p>This modular design allows individual agents to be replaced, extended, or reused without altering the overall system.&lt;/p>
&lt;hr>
&lt;h3 id="project-objectives">&lt;strong>Project Objectives&lt;/strong>&lt;/h3>
&lt;ol>
&lt;li>&lt;strong>Design a Modular Agent-based Architecture&lt;/strong>
&lt;ul>
&lt;li>Define clear interfaces for evidence collection, normalization, scoring, and explanation agents.&lt;/li>
&lt;/ul>
&lt;/li>
&lt;li>&lt;strong>Implement a Standardized Evidence Schema&lt;/strong>
&lt;ul>
&lt;li>Develop a unified data model for heterogeneous target-level evidence.&lt;/li>
&lt;/ul>
&lt;/li>
&lt;li>&lt;strong>Build a Reproducible Orchestration Framework&lt;/strong>
&lt;ul>
&lt;li>Implement a deterministic, inspectable workflow for agent coordination.&lt;/li>
&lt;/ul>
&lt;/li>
&lt;li>&lt;strong>Deliver a Community-Ready Toolkit&lt;/strong>
&lt;ul>
&lt;li>Provide CLI tools, example notebooks, and clear documentation to support adoption.&lt;/li>
&lt;/ul>
&lt;/li>
&lt;li>&lt;strong>Benchmark and Case Studies&lt;/strong>
&lt;ul>
&lt;li>Demonstrate the toolkit on representative target identification scenarios using public datasets.&lt;/li>
&lt;/ul>
&lt;/li>
&lt;/ol>
&lt;hr>
&lt;h3 id="project-deliverables">&lt;strong>Project Deliverables&lt;/strong>&lt;/h3>
&lt;ol>
&lt;li>&lt;strong>Open-Source Agent4Target Codebase&lt;/strong>
&lt;ul>
&lt;li>A well-documented Python package with modular agent components.&lt;/li>
&lt;/ul>
&lt;/li>
&lt;li>&lt;strong>Command-Line Interface (CLI)&lt;/strong>
&lt;ul>
&lt;li>Tools for running end-to-end evidence aggregation pipelines.&lt;/li>
&lt;/ul>
&lt;/li>
&lt;li>&lt;strong>Standardized Output Schema&lt;/strong>
&lt;ul>
&lt;li>Machine-readable evidence summaries suitable for downstream modeling.&lt;/li>
&lt;/ul>
&lt;/li>
&lt;li>&lt;strong>Example Notebooks and Benchmarks&lt;/strong>
&lt;ul>
&lt;li>Demonstrations of usage and performance on real-world target identification tasks.&lt;/li>
&lt;/ul>
&lt;/li>
&lt;li>&lt;strong>Documentation&lt;/strong>
&lt;ul>
&lt;li>Installation guides, extension tutorials, and developer documentation.&lt;/li>
&lt;/ul>
&lt;/li>
&lt;/ol>
&lt;hr>
&lt;h3 id="impact">&lt;strong>Impact&lt;/strong>&lt;/h3>
&lt;p>Agent4Target provides a practical bridge between AI agents and real-world drug discovery workflows. By emphasizing structured evidence aggregation, reproducibility, and interpretability, this project enables researchers to systematically reason about therapeutic targets rather than relying on opaque, end-to-end models. The resulting toolkit can serve as a foundation for future work in AI-assisted drug discovery, weak supervision, and biomedical knowledge integration.&lt;/p></description></item><item><title>HistoMoE: A Histology-Guided Mixture-of-Experts Framework for Gene Expression Prediction</title><link>https://deploy-preview-1007--ucsc-ospo.netlify.app/project/osre26/uci/histomoe/</link><pubDate>Tue, 20 Jan 2026 00:00:00 +0000</pubDate><guid>https://deploy-preview-1007--ucsc-ospo.netlify.app/project/osre26/uci/histomoe/</guid><description>&lt;ul>
&lt;li>&lt;strong>Topics:&lt;/strong> computational pathology, spatial transcriptomics, gene expression prediction, mixture-of-experts, multimodal learning&lt;/li>
&lt;li>&lt;strong>Skills:&lt;/strong>
&lt;ul>
&lt;li>&lt;strong>Programming Languages:&lt;/strong> Python; experience with PyTorch preferred&lt;/li>
&lt;li>&lt;strong>Machine Learning:&lt;/strong> CNNs / vision encoders, mixture-of-experts, multimodal representation learning&lt;/li>
&lt;li>&lt;strong>Data Analysis:&lt;/strong> handling large-scale histology image patches and gene expression matrices&lt;/li>
&lt;li>&lt;strong>Bioinformatics Knowledge (preferred):&lt;/strong> familiarity with spatial transcriptomics or scRNA-seq data&lt;/li>
&lt;/ul>
&lt;/li>
&lt;li>&lt;strong>Difficulty:&lt;/strong> Advanced&lt;/li>
&lt;li>&lt;strong>Size:&lt;/strong> Large (350 hours)&lt;/li>
&lt;li>&lt;strong>Mentors:&lt;/strong> &lt;a href="https://deploy-preview-1007--ucsc-ospo.netlify.app/author/ziheng-duan/">Ziheng Duan&lt;/a> (contact person)&lt;/li>
&lt;/ul>
&lt;h3 id="project-idea-description">&lt;strong>Project Idea Description&lt;/strong>&lt;/h3>
&lt;p>Histology imaging is one of the most widely available data modalities in biomedical research and clinical practice, capturing rich morphological information about tissues and disease states. In parallel, spatial transcriptomics (ST) technologies provide spatially resolved gene expression measurements, enabling unprecedented insights into tissue organization and cellular heterogeneity. However, the high cost and limited accessibility of ST experiments remain a major barrier to their widespread adoption.&lt;/p>
&lt;p>Predicting gene expression directly from histology images offers a promising alternative, enabling molecular-level inference from routinely collected pathology data. Existing approaches typically rely on a single global model that maps image embeddings to gene expression profiles. While effective to some extent, these models struggle to capture the strong organ-, tissue-, and cancer-specific heterogeneity that underlies gene expression patterns.&lt;/p>
&lt;p>This project proposes &lt;strong>HistoMoE&lt;/strong>, a &lt;strong>histology-guided mixture-of-experts (MoE) framework&lt;/strong> that explicitly models biological heterogeneity by learning &lt;strong>specialized expert models&lt;/strong> for different cancer types or organs, and dynamically routing histology image patches to the most relevant experts.&lt;/p>
&lt;h3 id="key-idea-and-technical-approach">&lt;strong>Key Idea and Technical Approach&lt;/strong>&lt;/h3>
&lt;p>As illustrated in the figure above, HistoMoE integrates multiple data modalities and learning components:&lt;/p>
&lt;ol>
&lt;li>
&lt;p>&lt;strong>Vision Encoder&lt;/strong>&lt;br>
Histology image patches are encoded into high-dimensional visual representations using a convolutional or transformer-based vision backbone.&lt;/p>
&lt;/li>
&lt;li>
&lt;p>&lt;strong>Text / Metadata Encoder&lt;/strong>&lt;br>
Sample-level metadata (e.g., tissue type, organ, disease context) is encoded using a lightweight text or embedding model.&lt;/p>
&lt;/li>
&lt;li>
&lt;p>&lt;strong>Gating Network&lt;/strong>&lt;br>
A gating network jointly considers image and metadata embeddings to infer routing weights over multiple &lt;strong>cancer- or organ-specific expert models&lt;/strong>.&lt;/p>
&lt;/li>
&lt;li>
&lt;p>&lt;strong>Expert Models&lt;/strong>&lt;br>
Each expert specializes in modeling gene expression patterns for a specific biological context (e.g., CCRCC, COAD, LUAD), producing patch-level gene expression predictions.&lt;/p>
&lt;/li>
&lt;/ol>
&lt;p>By explicitly modeling biological structure through expert specialization, HistoMoE aims to improve both &lt;strong>prediction accuracy&lt;/strong> and &lt;strong>interpretability&lt;/strong>, allowing researchers to understand which biological experts drive each prediction.&lt;/p>
&lt;h3 id="project-objectives">&lt;strong>Project Objectives&lt;/strong>&lt;/h3>
&lt;ol>
&lt;li>&lt;strong>Design and Implement the HistoMoE Framework&lt;/strong>
&lt;ul>
&lt;li>Build a modular MoE architecture with pluggable vision encoders, gating networks, and expert models.&lt;/li>
&lt;/ul>
&lt;/li>
&lt;li>&lt;strong>Multimodal Routing and Expert Specialization&lt;/strong>
&lt;ul>
&lt;li>Explore how image features and metadata jointly inform expert selection.&lt;/li>
&lt;/ul>
&lt;/li>
&lt;li>&lt;strong>Benchmarking and Evaluation&lt;/strong>
&lt;ul>
&lt;li>Compare HistoMoE against single-model baselines on multiple cancer and organ-specific spatial transcriptomics datasets.&lt;/li>
&lt;/ul>
&lt;/li>
&lt;li>&lt;strong>Interpretability Analysis&lt;/strong>
&lt;ul>
&lt;li>Analyze expert routing behavior to reveal biologically meaningful patterns.&lt;/li>
&lt;/ul>
&lt;/li>
&lt;/ol>
&lt;h3 id="project-deliverables">&lt;strong>Project Deliverables&lt;/strong>&lt;/h3>
&lt;ol>
&lt;li>&lt;strong>Open-Source HistoMoE Codebase&lt;/strong>
&lt;ul>
&lt;li>Well-documented Python implementation with training, evaluation, and visualization tools.&lt;/li>
&lt;/ul>
&lt;/li>
&lt;li>&lt;strong>Benchmark Results&lt;/strong>
&lt;ul>
&lt;li>Quantitative comparisons demonstrating improvements over non-expert baselines.&lt;/li>
&lt;/ul>
&lt;/li>
&lt;li>&lt;strong>Visualization and Analysis Tools&lt;/strong>
&lt;ul>
&lt;li>Tools for inspecting expert usage, routing weights, and gene-level predictions.&lt;/li>
&lt;/ul>
&lt;/li>
&lt;li>&lt;strong>Documentation and Tutorials&lt;/strong>
&lt;ul>
&lt;li>Clear instructions and examples to enable adoption by the research community.&lt;/li>
&lt;/ul>
&lt;/li>
&lt;/ol>
&lt;h3 id="impact">&lt;strong>Impact&lt;/strong>&lt;/h3>
&lt;p>HistoMoE introduces an expert-system perspective to histology-based gene expression prediction, bridging morphological and molecular representations through biologically informed specialization. By combining multimodal learning with mixture-of-experts modeling, this project advances the interpretability and accuracy of computational pathology methods and contributes toward scalable, cost-effective alternatives to spatial transcriptomics experiments.&lt;/p></description></item><item><title>StaR: A Stability-Aware Representation Learning Framework for Spatial Domain Identification</title><link>https://deploy-preview-1007--ucsc-ospo.netlify.app/project/osre26/uci/star/</link><pubDate>Tue, 20 Jan 2026 00:00:00 +0000</pubDate><guid>https://deploy-preview-1007--ucsc-ospo.netlify.app/project/osre26/uci/star/</guid><description>&lt;ul>
&lt;li>&lt;strong>Topics:&lt;/strong> spatial transcriptomics, spatial domain identification, representation learning, model robustness&lt;/li>
&lt;li>&lt;strong>Skills:&lt;/strong>
&lt;ul>
&lt;li>&lt;strong>Programming Languages:&lt;/strong> Python; PyTorch experience preferred&lt;/li>
&lt;li>&lt;strong>Machine Learning:&lt;/strong> representation learning, clustering, robustness and stability analysis&lt;/li>
&lt;li>&lt;strong>Data Analysis:&lt;/strong> spatial transcriptomics preprocessing and evaluation (ARI, clustering metrics)&lt;/li>
&lt;li>&lt;strong>Bioinformatics Knowledge (preferred):&lt;/strong> familiarity with spatial transcriptomics or scRNA-seq data&lt;/li>
&lt;/ul>
&lt;/li>
&lt;li>&lt;strong>Difficulty:&lt;/strong> Advanced&lt;/li>
&lt;li>&lt;strong>Size:&lt;/strong> Large (350 hours)&lt;/li>
&lt;li>&lt;strong>Mentors:&lt;/strong> &lt;a href="https://deploy-preview-1007--ucsc-ospo.netlify.app/author/ziheng-duan/">Ziheng Duan&lt;/a> (contact person)&lt;/li>
&lt;/ul>
&lt;h3 id="project-idea-description">&lt;strong>Project Idea Description&lt;/strong>&lt;/h3>
&lt;p>Spatial domain identification is a fundamental task in spatial transcriptomics (ST), aiming to partition tissue sections into biologically meaningful regions based on spatially resolved gene expression profiles. These spatial domains often correspond to distinct anatomical structures, cellular compositions, or functional microenvironments, and serve as a critical foundation for downstream biological analysis.&lt;/p>
&lt;p>Despite rapid methodological progress, &lt;strong>most existing spatial domain identification methods are highly sensitive to random initialization&lt;/strong>. In practice, simply changing the random seed can lead to substantially different clustering results and large performance fluctuations, even when using identical hyperparameters and datasets. This instability severely undermines the reliability, reproducibility, and interpretability of spatial transcriptomics analyses.&lt;/p>
&lt;h3 id="problem-seed-sensitivity-and-unstable-representations">&lt;strong>Problem: Seed Sensitivity and Unstable Representations&lt;/strong>&lt;/h3>
&lt;p>Empirical evidence shows that state-of-the-art spatial domain identification models can exhibit substantial performance variance across random seeds. For example, the Adjusted Rand Index (ARI) may vary from relatively strong performance (e.g., ARI ≈ 0.65) to noticeably degraded yet still reasonable outcomes (e.g., ARI ≈ 0.50) solely due to different random initializations.&lt;/p>
&lt;p>By systematically evaluating models across &lt;strong>hundreds to thousands of random seeds&lt;/strong>, we observe that:&lt;/p>
&lt;ul>
&lt;li>Model performance landscapes are highly &lt;strong>rugged&lt;/strong>, with sharp cliffs and isolated high-performing regions.&lt;/li>
&lt;li>Standard training objectives implicitly favor brittle representations that are not robust to small perturbations in initialization or optimization trajectories.&lt;/li>
&lt;/ul>
&lt;p>These observations suggest that instability is not a peripheral issue, but rather a &lt;strong>structural limitation of current representation learning approaches&lt;/strong> for spatial transcriptomics.&lt;/p>
&lt;h3 id="proposed-solution-star">&lt;strong>Proposed Solution: StaR&lt;/strong>&lt;/h3>
&lt;p>This project proposes &lt;strong>StaR&lt;/strong>, a &lt;strong>Stability-Aware Representation Learning framework&lt;/strong> designed to explicitly address seed sensitivity in spatial domain identification.&lt;/p>
&lt;p>The core idea of StaR is to &lt;strong>learn representations that are robust to perturbations in model parameters and training dynamics&lt;/strong>, rather than optimizing solely for peak performance under a single random seed. Concretely, StaR introduces controlled noise or perturbations into the training process and encourages consistency across multiple perturbed model instances, guiding the model toward flatter and more stable regions of the parameter space.&lt;/p>
&lt;p>By prioritizing stability during representation learning, StaR aims to produce embeddings that:&lt;/p>
&lt;ul>
&lt;li>Yield consistent spatial domain assignments across random seeds&lt;/li>
&lt;li>Maintain competitive or improved clustering accuracy&lt;/li>
&lt;li>Better reflect underlying biological structure&lt;/li>
&lt;/ul>
&lt;h3 id="project-objectives">&lt;strong>Project Objectives&lt;/strong>&lt;/h3>
&lt;ol>
&lt;li>&lt;strong>Characterize Instability in Existing Methods&lt;/strong>
&lt;ul>
&lt;li>Systematically quantify seed sensitivity across popular spatial domain identification models.&lt;/li>
&lt;/ul>
&lt;/li>
&lt;li>&lt;strong>Develop Stability-Aware Training Objectives&lt;/strong>
&lt;ul>
&lt;li>Design perturbation-based or consistency-driven losses that encourage robust representations.&lt;/li>
&lt;/ul>
&lt;/li>
&lt;li>&lt;strong>Integrate StaR into Existing Pipelines&lt;/strong>
&lt;ul>
&lt;li>Apply StaR to widely used spatial transcriptomics workflows with minimal architectural changes.&lt;/li>
&lt;/ul>
&lt;/li>
&lt;li>&lt;strong>Evaluation and Benchmarking&lt;/strong>
&lt;ul>
&lt;li>Evaluate StaR using clustering metrics (e.g., ARI) and stability metrics across multiple datasets and random seeds.&lt;/li>
&lt;/ul>
&lt;/li>
&lt;li>&lt;strong>Biological Validation&lt;/strong>
&lt;ul>
&lt;li>Assess whether stability-aware representations preserve biologically meaningful spatial patterns.&lt;/li>
&lt;/ul>
&lt;/li>
&lt;/ol>
&lt;h3 id="project-deliverables">&lt;strong>Project Deliverables&lt;/strong>&lt;/h3>
&lt;ol>
&lt;li>&lt;strong>StaR Framework Implementation&lt;/strong>
&lt;ul>
&lt;li>An open-source Python implementation compatible with common spatial transcriptomics toolchains.&lt;/li>
&lt;/ul>
&lt;/li>
&lt;li>&lt;strong>Stability Benchmarks&lt;/strong>
&lt;ul>
&lt;li>Comprehensive evaluations demonstrating reduced performance variance across seeds.&lt;/li>
&lt;/ul>
&lt;/li>
&lt;li>&lt;strong>Visualization Tools&lt;/strong>
&lt;ul>
&lt;li>Tools for visualizing performance landscapes, stability surfaces, and spatial domain consistency.&lt;/li>
&lt;/ul>
&lt;/li>
&lt;li>&lt;strong>Documentation and Tutorials&lt;/strong>
&lt;ul>
&lt;li>Clear examples enabling researchers to adopt StaR in their own analyses.&lt;/li>
&lt;/ul>
&lt;/li>
&lt;/ol>
&lt;h3 id="impact">&lt;strong>Impact&lt;/strong>&lt;/h3>
&lt;p>StaR addresses a critical yet underexplored challenge in spatial transcriptomics: &lt;strong>model instability and poor reproducibility&lt;/strong>. By shifting the focus from single-run performance to stability-aware representation learning, this project improves the reliability and trustworthiness of spatial domain identification methods. StaR has the potential to become a foundational component in robust spatial transcriptomics pipelines and to inspire broader adoption of stability-aware principles in biological representation learning.&lt;/p></description></item><item><title>MedJEPA: Self-Supervised Medical Image Representation Learning with JEPA</title><link>https://deploy-preview-1007--ucsc-ospo.netlify.app/project/osre26/nelbl/medjepa/</link><pubDate>Mon, 19 Jan 2026 10:15:56 -0700</pubDate><guid>https://deploy-preview-1007--ucsc-ospo.netlify.app/project/osre26/nelbl/medjepa/</guid><description>&lt;h3 id="project-description">Project Description&lt;/h3>
&lt;p>[MedJEPA] Medical image analysis is fundamental to modern healthcare, enabling disease diagnosis, treatment planning, and patient monitoring across diverse clinical applications. In radiology and pathology, deep learning models support automated detection of abnormalities, tumor segmentation, and diagnostic assistance. Medical imaging modalities including X-rays, CT scans, MRI, ultrasound, and histopathology slides generate vast amounts of unlabeled data that could benefit from self-supervised representation learning. Clinical applications include cancer detection and staging, cardiovascular disease assessment, neurological disorder diagnosis, and infectious disease screening. In drug discovery and clinical research, analyzing medical images helps evaluate treatment efficacy, predict patient outcomes, and identify biomarkers for disease progression. Telemedicine and point-of-care diagnostics benefit from AI-powered image analysis that extends expert-level interpretation to underserved regions. However, medical imaging faces unique challenges: limited labeled datasets due to expensive expert annotation, patient privacy concerns restricting data sharing, domain shift across different imaging equipment and protocols, and the need for models that generalize across hospitals and populations.
Traditional medical image analysis relies heavily on supervised learning with manually annotated labels, creating bottlenecks due to the scarcity and cost of expert annotations. Existing self-supervised methods applied to medical imaging often employ complex training procedures with numerous heuristics—momentum encoders, stop-gradients, teacher-student architectures, and carefully tuned augmentation strategies—that may not translate well across different medical imaging modalities and clinical contexts. These approaches struggle with domain-specific challenges such as subtle pathological features, high-resolution images, 3D volumetric data, and the need for interpretable representations that clinicians can trust. To address these challenges, we propose MedicalJEPA: Self-Supervised Medical Image Representation Learning with Joint-Embedding Predictive Architecture, which leverages the theoretically grounded LeJEPA framework for 2D medical images and V-JEPA principles for medical video and volumetric data, creating a unified, scalable, and heuristics-free approach specifically tailored for medical imaging applications.
By utilizing the principled JEPA frameworks with objectives like Sketched Isotropic Gaussian Regularization (SIGReg), MedJEPA eliminates complex training heuristics while learning clinically meaningful representations from unlabeled medical images. Unlike conventional self-supervised methods that require extensive hyperparameter tuning and may not generalize across medical imaging modalities, MedicalJEPA provides a clean, theoretically motivated framework with minimal hyperparameters that adapts to diverse medical imaging contexts—from chest X-rays to histopathology slides to cardiac MRI sequences. The learned representations can support downstream tasks including disease classification, lesion detection, organ segmentation, and survival prediction, while requiring significantly fewer labeled examples for fine-tuning. This approach democratizes access to state-of-the-art medical AI by enabling effective learning from the vast amounts of unlabeled medical imaging data available in hospital archives, addressing the annotation bottleneck that has limited progress in medical AI.&lt;/p>
&lt;h3 id="project-objectives">Project Objectives&lt;/h3>
&lt;p>Aligned with the vision of the 2026 Open Source Research Experience (OSRE), this project aims to apply Joint-Embedding Predictive Architecture (JEPA) frameworks to medical image representation learning, addressing the critical challenge of learning from limited labeled medical data. Medical imaging generates enormous amounts of unlabeled data, but supervised learning approaches are bottlenecked by the scarcity and cost of expert annotations. Existing self-supervised methods often rely on complex heuristics that don&amp;rsquo;t generalize well across diverse medical imaging modalities, equipment vendors, and clinical protocols.
This project will leverage the theoretically grounded LeJEPA framework for 2D medical images (X-rays, histopathology slides, fundus images) and V-JEPA principles for temporal and volumetric medical data (cardiac MRI sequences, CT scans, surgical videos). The core challenge lies in adapting these heuristics-free, stable frameworks to medical imaging&amp;rsquo;s unique characteristics: subtle pathological features requiring fine-grained representations, high-resolution images demanding efficient processing, domain shift across hospitals and equipment, and the need for interpretable features that support clinical decision-making. The learned representations will be evaluated on diverse downstream clinical tasks including disease classification, lesion detection, organ segmentation, and prognosis prediction, with emphasis on few-shot learning scenarios that reflect real-world annotation constraints. Below is an outline of the methodologies and models that will be developed in this project.&lt;/p>
&lt;ul>
&lt;li>
&lt;p>&lt;strong>Step 1: Medical Data Preparation&lt;/strong>:
Develop data processing pipelines for diverse medical imaging modalities, implementing DICOM/NIfTI parsing, standardized preprocessing, and efficient data loading for self-supervised pre-training.
Prepare 2D medical image datasets:
Chest X-rays: ChestX-ray14, MIMIC-CXR, CheXpert for lung disease detection
Histopathology: Camelyon16/17 (breast cancer), PCam (patch-level classification)
Retinal imaging: EyePACS, APTOS (diabetic retinopathy), Messidor
Dermatology: HAM10000, ISIC (skin lesion classification)
Prepare 3D volumetric and temporal medical data:
CT scans: LIDC-IDRI (lung nodules), Medical Segmentation Decathlon datasets
MRI sequences: BraTS (brain tumors), ACDC (cardiac MRI), UK Biobank cardiac videos
Medical video: Surgical procedure videos, endoscopy recordings, ultrasound sequences
Implement medical imaging-specific preprocessing: intensity normalization, resolution standardization, handling of multi-channel medical images (different MRI sequences, RGB histopathology), and privacy-preserving anonymization.
Design masking strategies appropriate for medical imaging: spatial masking for 2D images, volumetric masking for 3D scans, temporal masking for sequences, and anatomy-aware masking that respects organ boundaries.
Create data loaders supporting high-resolution medical images, 3D volumes, and multi-modal inputs (e.g., multiple MRI sequences).&lt;/p>
&lt;/li>
&lt;li>
&lt;p>&lt;strong>Step 2: JEPA Model Implementation for Medical Imaging&lt;/strong>:
Implement LeJEPA for 2D medical images:
Adapt joint-embedding predictive architecture for medical image characteristics (high resolution, subtle features, domain-specific patterns)
Apply Sketched Isotropic Gaussian Regularization (SIGReg) to learn clinically meaningful embedding distributions
Maintain single trade-off hyperparameter and heuristics-free training for reproducibility across medical imaging centers
Support various encoder architectures: Vision Transformers for global context, ConvNets for local features, hybrid approaches
Extend to V-JEPA for medical video and volumetric data:
Spatiotemporal encoding for cardiac MRI sequences, surgical videos, and time-series medical imaging
Temporal prediction objectives for understanding disease progression and treatment response
3D volume processing for CT and MRI scans with efficient memory management
Multi-slice and multi-sequence learning for comprehensive medical imaging contexts
Develop medical domain-specific enhancements:
Multi-scale representation learning to capture both fine-grained pathological details and global anatomical context
Interpretability mechanisms: attention visualization, feature attribution, and embedding space analysis for clinical validation
Robustness to domain shift: training strategies that generalize across different scanners, protocols, and institutions
Privacy-preserving training considerations compatible with medical data regulations (HIPAA, GDPR)
Implement efficient training infrastructure:
Support for distributed training across multiple GPUs for large medical imaging datasets
Memory-efficient processing of high-resolution images and 3D volumes
Checkpoint management and model versioning for clinical deployment pipelines
Minimal-code implementation (≈50-100 lines) demonstrating framework simplicity&lt;/p>
&lt;/li>
&lt;li>
&lt;p>&lt;strong>Step 3: Evaluation &amp;amp; Safety Validation&lt;/strong>: :
Disease Classification Tasks:
Multi-label chest X-ray classification: 14 pathology classes on ChestX-ray14, MIMIC-CXR
Diabetic retinopathy grading: 5-class classification on EyePACS, APTOS
Skin lesion classification: 7-class classification on HAM10000
Brain tumor classification: glioma grading on BraTS dataset
Evaluate with linear probing, few-shot learning (5-shot, 10-shot), and full fine-tuning
Lesion Detection and Segmentation:
Lung nodule detection on LIDC-IDRI dataset
Tumor segmentation on Medical Segmentation Decathlon tasks
Polyp detection in colonoscopy videos
Cardiac structure segmentation in MRI sequences
Clinical Prediction Tasks:
Survival prediction from histopathology slides
Disease progression prediction from longitudinal imaging
Treatment response assessment from pre/post imaging pairs
Few-Shot and Low-Data Regime Evaluation:
Systematic evaluation with 1%, 5%, 10%, 25%, 50% of labeled training data
Comparison against supervised baselines and ImageNet pre-training
Analysis of annotation efficiency: performance vs. number of labeled examples required&lt;/p>
&lt;/li>
&lt;/ul>
&lt;h3 id="project-deliverables">Project Deliverables&lt;/h3>
&lt;p>This project will deliver three components: software implementation, clinical evaluation, and practical deployment resources. The software implementing MedicalJEPA will be hosted on GitHub as an open-access repository with modular code supporting multiple medical imaging modalities (2D images, 3D volumes, videos), pre-trained model checkpoints on major medical imaging datasets (chest X-rays, histopathology, MRI), training and evaluation scripts with medical imaging-specific preprocessing pipelines, privacy-preserving training implementations compatible with clinical data regulations, and comprehensive documentation including tutorials for medical AI researchers and clinicians. The evaluation results will include benchmarks on 10+ medical imaging datasets across diverse modalities and clinical tasks, few-shot learning analysis demonstrating annotation efficiency gains, cross-institutional validation studies showing robustness to domain shift, interpretability visualizations enabling clinical validation of learned representations, and detailed comparisons against supervised baselines and existing medical self-supervised methods. .&lt;/p>
&lt;h3 id="neurohealth">NeuroHealth&lt;/h3>
&lt;ul>
&lt;li>&lt;strong>Topics&lt;/strong>: Self-Supervised Medical Image Representation Learning with JEPA&lt;/li>
&lt;li>&lt;strong>Skills&lt;/strong>: Proficiency in Python, Pytorch, Github, JEPA&lt;/li>
&lt;li>&lt;strong>Difficulty&lt;/strong>: Difficult&lt;/li>
&lt;li>&lt;strong>Size&lt;/strong>: Large (350 hours)&lt;/li>
&lt;li>&lt;strong>Mentor&lt;/strong>: &lt;a href="https://deploy-preview-1007--ucsc-ospo.netlify.app/author/bin-dong/">Bin Dong&lt;/a>, &lt;a href="https://deploy-preview-1007--ucsc-ospo.netlify.app/author/linsey-pang/">Linsey Pang&lt;/a>&lt;/li>
&lt;/ul>
&lt;h3 id="references">References:&lt;/h3>
&lt;ul>
&lt;li>LeJEPA: Provable and Scalable Self-Supervised Learning Without the Heuristics - Randall Balestriero and Yann LeCun, arXiv 2024&lt;/li>
&lt;li>Revisiting Feature Prediction for Learning Visual Representations from Video (V-JEPA) - Adrien Bardes et al., arXiv 2024&lt;/li>
&lt;li>Self-Supervised Learning from Images with a Joint-Embedding Predictive Architecture - Mahmoud Assran et al., CVPR 2023 (I-JEPA)&lt;/li>
&lt;li>ChestX-ray14: Hospital-Scale Chest X-Ray Database - &lt;a href="https://nihcc.app.box.com/v/ChestXray-NIHCC" target="_blank" rel="noopener">https://nihcc.app.box.com/v/ChestXray-NIHCC&lt;/a>&lt;/li>
&lt;li>Medical Segmentation Decathlon - &lt;a href="http://medicaldecathlon.com/" target="_blank" rel="noopener">http://medicaldecathlon.com/&lt;/a>&lt;/li>
&lt;li>MIMIC-CXR Database - &lt;a href="https://physionet.org/content/mimic-cxr/" target="_blank" rel="noopener">https://physionet.org/content/mimic-cxr/&lt;/a>&lt;/li>
&lt;li>The Cancer Imaging Archive (TCIA) - &lt;a href="https://www.cancerimagingarchive.net/" target="_blank" rel="noopener">https://www.cancerimagingarchive.net/&lt;/a>&lt;/li>
&lt;li>UK Biobank Imaging Study - &lt;a href="https://www.ukbiobank.ac.uk/enable-your-research/about-our-data/imaging-data" target="_blank" rel="noopener">https://www.ukbiobank.ac.uk/enable-your-research/about-our-data/imaging-data&lt;/a>&lt;/li>
&lt;/ul></description></item><item><title>NeuroHealth: AI-Powered Health Assistant</title><link>https://deploy-preview-1007--ucsc-ospo.netlify.app/project/osre26/nelbl/neurohealth/</link><pubDate>Mon, 19 Jan 2026 10:15:56 -0700</pubDate><guid>https://deploy-preview-1007--ucsc-ospo.netlify.app/project/osre26/nelbl/neurohealth/</guid><description>&lt;h3 id="project-description">Project Description&lt;/h3>
&lt;p>[NeuroHealth] Intelligent health assistance systems are increasingly essential for improving healthcare accessibility, patient engagement, and clinical decision support. In primary care and preventive medicine, AI assistants help users understand symptoms, schedule appropriate appointments, and receive preliminary health guidance. Telemedicine applications include triage support, appointment scheduling optimization, and patient education based on health inquiries. In chronic disease management, these systems provide medication reminders, lifestyle recommendations, and timely alerts for medical follow-ups. Healthcare navigation applications include finding appropriate specialists, understanding treatment options, and coordinating care across multiple providers. In wellness and preventive care, intelligent assistants enhance health literacy by delivering personalized health information, screening recommendations, and proactive health management strategies. By leveraging natural language understanding and medical knowledge integration, these systems enhance healthcare access, reduce unnecessary emergency visits, and empower users to make informed health decisions across diverse populations.
Traditional health information systems often provide generic responses that fail to account for individual health contexts, medical history, and personal circumstances. Existing symptom checkers and health chatbots primarily rely on rule-based logic or simple decision trees, limiting their ability to understand nuanced health inquiries, reason about complex symptom patterns, or provide contextually appropriate guidance. These systems struggle with interpreting ambiguous descriptions, adapting to users&amp;rsquo; health literacy levels, and generating personalized recommendations that account for individual medical constraints and preferences. To address these challenges, we propose NeuroHealth: AI-Powered Health Assistant, which leverages Large Language Models (LLMs) to create an intelligent conversational agent that synthesizes user health inquiries, symptom descriptions, and contextual information into actionable, personalized health guidance and appointment recommendations.
By integrating LLM-based medical reasoning with structured clinical knowledge bases, NeuroHealth enhances symptom interpretation, appointment routing, and health education delivery. Unlike conventional systems that provide static responses from predetermined templates, NeuroHealth dynamically understands user intent, asks clarifying questions, assesses urgency levels, and generates appropriate recommendations—whether scheduling a doctor appointment, suggesting self-care measures, or directing users to emergency services. This fusion of LLM intelligence with validated medical knowledge enables a more accessible, adaptive, and helpful health assistance platform, bridging the gap between users seeking health information and appropriate medical care.&lt;/p>
&lt;h3 id="project-objectives">Project Objectives&lt;/h3>
&lt;p>Aligned with the vision of the 2026 Open Source Research Experience (OSRE), this project aims to develop an AI-Powered Health Assistant (NeuroHealth) to improve healthcare accessibility and patient engagement through intelligent conversational guidance. Healthcare systems face significant challenges in providing timely, personalized health information and connecting patients with appropriate care resources. Traditional symptom checkers and health information systems often deliver generic, rule-based responses that fail to account for individual contexts and struggle with natural language understanding.
To address these limitations, this project will leverage Large Language Models (LLMs) to create an intelligent health assistant that understands user health inquiries, interprets symptom descriptions, assesses urgency, and provides personalized recommendations including doctor appointment suggestions, self-care guidance, and healthcare navigation support. The core challenge lies in designing NeuroHealth as a safe, accurate, and user-friendly system capable of natural conversation, medical knowledge retrieval, and appropriate response generation while maintaining clinical safety guardrails. Unlike conventional health chatbots that follow rigid conversation flows, NeuroHealth will reason over user inputs, ask clarifying questions, and dynamically adapt responses based on context, resulting in more helpful, accurate, and appropriate health assistance. Below is an outline of the methodologies and models that will be developed in this project.&lt;/p>
&lt;ul>
&lt;li>
&lt;p>&lt;strong>Step 1: Data Collection &amp;amp; Knowledge Base Construction&lt;/strong>:
Develop a comprehensive medical knowledge base integrating validated health information sources, symptom databases, condition descriptions, and appointment routing guidelines.
Collect and curate conversational health inquiry datasets from public medical Q&amp;amp;A forums, symptom checker logs, and healthcare chatbot interactions to create training and evaluation data.
Design structured representations for symptoms, conditions, urgency levels, and appointment recommendations to enable effective retrieval and reasoning.
Extract common health inquiry patterns, symptom descriptions, and user intent categories to inform conversation flow design.
Data sources can include public medical knowledge bases such as MedlinePlus, Mayo Clinic health information, clinical practice guidelines, and synthetic patient inquiry scenarios based on common healthcare use cases.
Implement data validation mechanisms to ensure medical accuracy and clinical safety compliance.&lt;/p>
&lt;/li>
&lt;li>
&lt;p>&lt;strong>Step 2: Model Development&lt;/strong>:
Design and implement an LLM-based conversational health assistant that integrates medical knowledge retrieval with natural language understanding and generation.
Develop a Retrieval-Augmented Generation (RAG) architecture that grounds LLM responses in validated medical information sources, reducing hallucination risks and ensuring factual accuracy.
Create prompt engineering strategies and reasoning frameworks that enable the system to: interpret symptom descriptions, assess urgency levels, ask appropriate clarifying questions, and generate personalized health guidance.
Implement a multi-component architecture including: intent recognition, symptom extraction, urgency assessment, appointment recommendation generation, and response formatting modules.
Develop clinical safety guardrails that detect high-risk scenarios requiring immediate medical attention and provide appropriate emergency guidance.
Design conversation management strategies that maintain context across multi-turn dialogues and adapt to users&amp;rsquo; health literacy levels.
The baseline architecture can leverage state-of-the-art models such as GPT-4, Claude, or open-source alternatives like Llama, Qwen, combined with medical knowledge retrieval systems.&lt;/p>
&lt;/li>
&lt;li>
&lt;p>&lt;strong>Step 3: Evaluation &amp;amp; Safety Validation&lt;/strong>: :
Benchmark NeuroHealth against existing symptom checkers and health chatbots, evaluating on metrics including response accuracy, appropriateness of appointment recommendations, urgency assessment precision, and user satisfaction.
Conduct human evaluation studies with healthcare professionals to assess clinical safety, response quality, and appropriateness of medical guidance.
Perform adversarial testing to identify potential failure modes, unsafe responses, or inappropriate recommendations under edge cases.
Conduct ablation studies to analyze the impact of retrieval-augmented generation, safety guardrails, and conversation management strategies on system performance.
Evaluate system performance across diverse health inquiry types including acute symptoms, chronic condition management, preventive care questions, and healthcare navigation requests.
Assess response quality across different user demographics and health literacy levels to ensure equitable access.
Optimize inference efficiency and response latency for real-time conversational interaction across web and mobile platforms.&lt;/p>
&lt;/li>
&lt;/ul>
&lt;h3 id="project-deliverables">Project Deliverables&lt;/h3>
&lt;p>This project will deliver three components: model development, evaluation and validation, and interactive demonstration. The software implementing the NeuroHealth system will be hosted on GitHub as an open-access repository with comprehensive documentation, deployment guides, and API specifications. The evaluation results, including benchmark comparisons against existing systems, clinical safety assessments, and user study findings, will be published alongside the GitHub repository. An interactive demo showcasing the conversational interface, symptom interpretation capabilities, and appointment recommendation generation will be provided to illustrate real-world application scenarios.&lt;/p>
&lt;h3 id="neurohealth">NeuroHealth&lt;/h3>
&lt;ul>
&lt;li>&lt;strong>Topics&lt;/strong>: AI-Powered Health Assistant&lt;/li>
&lt;li>&lt;strong>Skills&lt;/strong>: Proficiency in Python, Github, LLM&lt;/li>
&lt;li>&lt;strong>Difficulty&lt;/strong>: Difficult&lt;/li>
&lt;li>&lt;strong>Size&lt;/strong>: Large (350 hours)&lt;/li>
&lt;li>&lt;strong>Mentor&lt;/strong>: &lt;a href="https://deploy-preview-1007--ucsc-ospo.netlify.app/author/linsey-pang/">Linsey Pang&lt;/a>, &lt;a href="https://deploy-preview-1007--ucsc-ospo.netlify.app/author/bin-dong/">Bin Dong&lt;/a>&lt;/li>
&lt;/ul>
&lt;h3 id="references">References:&lt;/h3>
&lt;ul>
&lt;li>Large Language Models in Healthcare - Singhal et al., Nature 2023&lt;/li>
&lt;li>Med-PaLM: Large Language Models for Medical Question Answering - Singhal et al., arXiv 2022&lt;/li>
&lt;li>Capabilities of GPT-4 on Medical Challenge Problems - Nori et al., arXiv 2023&lt;/li>
&lt;li>MedlinePlus Medical Encyclopedia - &lt;a href="https://medlineplus.gov/" target="_blank" rel="noopener">https://medlineplus.gov/&lt;/a>&lt;/li>
&lt;li>Clinical Practice Guidelines Database - &lt;a href="https://www.guidelines.gov/" target="_blank" rel="noopener">https://www.guidelines.gov/&lt;/a>&lt;/li>
&lt;/ul></description></item><item><title>LMS Toolkit</title><link>https://deploy-preview-1007--ucsc-ospo.netlify.app/project/osre26/ucsc/lms-toolkit/</link><pubDate>Tue, 13 Jan 2026 13:00:00 -0800</pubDate><guid>https://deploy-preview-1007--ucsc-ospo.netlify.app/project/osre26/ucsc/lms-toolkit/</guid><description>&lt;p>The &lt;a href="https://github.com/edulinq/lms-toolkit" target="_blank" rel="noopener">EduLinq LMS Toolkit&lt;/a> is a suite of tools used by several courses at UCSC
to interact with LMS&amp;rsquo;s (e.g. Canvas) from the command line or Python.
A &lt;a href="https://en.wikipedia.org/wiki/Learning_management_system" target="_blank" rel="noopener">Learning Management System&lt;/a> (LMS) is a system that institutions use to manage courses, assignments, students, and grades.
The most popular LMSs are
&lt;a href="https://en.wikipedia.org/wiki/Instructure#Canvas" target="_blank" rel="noopener">Canvas&lt;/a>,
&lt;a href="https://en.wikipedia.org/wiki/Blackboard_Learn" target="_blank" rel="noopener">Blackboard&lt;/a>,
&lt;a href="https://en.wikipedia.org/wiki/Moodle" target="_blank" rel="noopener">Moodle&lt;/a>,
and &lt;a href="https://en.wikipedia.org/wiki/D2L#Brightspace" target="_blank" rel="noopener">Brightspace&lt;/a>.
These tools can be very helpful, especially from an administrative standpoint, but can be hard to interact with.
They can be especially difficult when instructors and TAs want to do something that is not explicitly supported by their built-in GUIs
(e.g., when an instructor wants to use a special grading policy).
The LMS Toolkit project is an effort to create a single suite of command-line tools (along with a Python interface)
to connect to all the above mentioned LMSs in a simple and uniform way.
So, not only can instructors and TAs easily access the modify the data held in an LMS (like a student&amp;rsquo;s grades),
but they can also do it the same way on any LMS.
The &lt;a href="https://linqs.org" target="_blank" rel="noopener">LINQS Lab&lt;/a> has made many contributions to the maintain and improve the Quiz Composer.&lt;/p>
&lt;p>Currently, the LMS Toolkit supports Canvas, Moodle, and Blackboard.
But, the degree of support for each LMS varies.&lt;/p>
&lt;p>All students interested in LINQS projects for OSRE/GSoC 2026 should fill out &lt;a href="https://forms.gle/Mr4YR3N35pWDb4uz7" target="_blank" rel="noopener">this form&lt;/a>.
Towards the end of the application window, we will contact those who we believe to be a good fit for a LINQS project.
The form will stop accepting responses once the application window closes.
Do not post on any of the project repositories about OSRE/GSoC
(e.g., comment on an issue that you want to tackle it as a part of OSRE/GSoC 2026).
Remember, these are active repositories that were not created for OSRE/GSoC.&lt;/p>
&lt;h3 id="advanced-lms-support">Advanced LMS Support&lt;/h3>
&lt;ul>
&lt;li>&lt;strong>Topics:&lt;/strong> &lt;code>Backend&lt;/code> &lt;code>Teaching Tools&lt;/code> &lt;code>API&lt;/code>&lt;/li>
&lt;li>&lt;strong>Skills:&lt;/strong> software development, backend, rest api, data munging, http request inspection, python&lt;/li>
&lt;li>&lt;strong>Difficulty:&lt;/strong> Moderate&lt;/li>
&lt;li>&lt;strong>Size:&lt;/strong> Medium or Large (175 or 350 hours)&lt;/li>
&lt;li>&lt;strong>Mentors:&lt;/strong> &lt;a href="mailto:linqs.osre26@gmail.com">Eriq Augustine&lt;/a>, &lt;a href="mailto:linqs.osre26@gmail.com">Batuhan Salih&lt;/a>, &lt;a href="mailto:linqs.osre26@gmail.com">Lise Getoor&lt;/a>&lt;/li>
&lt;/ul>
&lt;p>The LMS Toolkit already has basic read-write support for many core pieces of LMS functionality (e.g., working with grades and assignments).
However, there are still many more features that can be supported such as
&lt;a href="https://github.com/edulinq/lms-toolkit/issues/17" target="_blank" rel="noopener">group management&lt;/a>,
&lt;a href="https://github.com/edulinq/lms-toolkit/issues/7" target="_blank" rel="noopener">quiz management&lt;/a>,
&lt;a href="https://github.com/edulinq/lms-toolkit/issues/10" target="_blank" rel="noopener">quiz statistics&lt;/a>,
and &lt;a href="https://github.com/edulinq/lms-toolkit/issues/19" target="_blank" rel="noopener">assignment statuses&lt;/a>.&lt;/p>
&lt;p>The task for this project is to choose a set of advanced features
(not limited to those features mentioned above),
design an LMS-agnostic way to support those features,
and implement those features.
The flexibility in the features chosen to implement account for the variable size of this project.&lt;/p>
&lt;p>See Also:&lt;/p>
&lt;ul>
&lt;li>&lt;a href="https://github.com/edulinq/lms-toolkit" target="_blank" rel="noopener">Repository for LMS Toolkit&lt;/a>&lt;/li>
&lt;li>GitHub Issues
&lt;ul>
&lt;li>&lt;a href="https://github.com/edulinq/lms-toolkit/issues/17" target="_blank" rel="noopener">Group Management&lt;/a>,&lt;/li>
&lt;li>&lt;a href="https://github.com/edulinq/lms-toolkit/issues/7" target="_blank" rel="noopener">Quiz Management&lt;/a>,&lt;/li>
&lt;li>&lt;a href="https://github.com/edulinq/lms-toolkit/issues/10" target="_blank" rel="noopener">Quiz Statistics&lt;/a>,&lt;/li>
&lt;li>&lt;a href="https://github.com/edulinq/lms-toolkit/issues/19" target="_blank" rel="noopener">Assignment Statuses&lt;/a>.&lt;/li>
&lt;/ul>
&lt;/li>
&lt;/ul>
&lt;h3 id="new-lms-support-brightspace">New LMS Support: Brightspace&lt;/h3>
&lt;ul>
&lt;li>&lt;strong>Topics:&lt;/strong> &lt;code>Backend&lt;/code> &lt;code>Teaching Tools&lt;/code> &lt;code>API&lt;/code>&lt;/li>
&lt;li>&lt;strong>Skills:&lt;/strong> software development, backend, rest api, data munging, http request inspection, python&lt;/li>
&lt;li>&lt;strong>Difficulty:&lt;/strong> Challenging&lt;/li>
&lt;li>&lt;strong>Size:&lt;/strong> Large (350 hours)&lt;/li>
&lt;li>&lt;strong>Mentors:&lt;/strong> &lt;a href="mailto:linqs.osre26@gmail.com">Eriq Augustine&lt;/a>, &lt;a href="mailto:linqs.osre26@gmail.com">Batuhan Salih&lt;/a>, &lt;a href="mailto:linqs.osre26@gmail.com">Lise Getoor&lt;/a>&lt;/li>
&lt;/ul>
&lt;p>The goal of the LMS toolkit is to provide a single interface for all LMSs.
&lt;a href="https://en.wikipedia.org/wiki/D2L#Brightspace" target="_blank" rel="noopener">D2L Brightspace&lt;/a> is one of the more popular LMSs.
Naturally, the LMS Toolkit wants to support Brightspace as well.
However, a challenge in supporting Brightspace is that it is not open source (unlike Canvas and Moodle).
Therefore, support and testing on Brightspace may be very challenging.&lt;/p>
&lt;p>The task for this project is to add basic support for the Brightspace LMS.
It is not necessary to support all the same features that are supported for other LMSs,
but at least the core features of score and assignment management should be implemented.
The closed-source nature of Brightspace makes this a challenging and uncertain project.&lt;/p>
&lt;p>See Also:&lt;/p>
&lt;ul>
&lt;li>&lt;a href="https://github.com/edulinq/lms-toolkit" target="_blank" rel="noopener">Repository for LMS Toolkit&lt;/a>&lt;/li>
&lt;li>&lt;a href="https://en.wikipedia.org/wiki/D2L#Brightspace" target="_blank" rel="noopener">Brightspace Wiki Page&lt;/a>&lt;/li>
&lt;li>&lt;a href="https://github.com/edulinq/lms-toolkit/issues/23" target="_blank" rel="noopener">GitHub Issue&lt;/a>&lt;/li>
&lt;/ul></description></item><item><title>Lynx Grader</title><link>https://deploy-preview-1007--ucsc-ospo.netlify.app/project/osre26/ucsc/autograder/</link><pubDate>Tue, 13 Jan 2026 13:00:00 -0800</pubDate><guid>https://deploy-preview-1007--ucsc-ospo.netlify.app/project/osre26/ucsc/autograder/</guid><description>&lt;p>The &lt;a href="https://github.com/edulinq/autograder-server" target="_blank" rel="noopener">EduLinq Lynx Grader&lt;/a> (also referred to as &amp;ldquo;autograder&amp;rdquo;) is an open source tool used by several courses at UCSC
to safely and quickly grade programming assignments.
Grading student code is something that may seem simple at first (you just need to run their code!),
but quickly becomes exceeding complex as you get more into the details.
Specifically, grading a student&amp;rsquo;s code securely while providing the &amp;ldquo;last mile&amp;rdquo; service of getting code from students
and sending results to instructors/TAs and the course&amp;rsquo;s LMS (e.g., Canvas) can be very difficult.
The Lynx Grader provides all of this in a free and open source project.
The &lt;a href="https://linqs.org" target="_blank" rel="noopener">LINQS Lab&lt;/a> has made many contributions to the maintain and improve the Lynx Grader.&lt;/p>
&lt;p>As an open source project, there are endless opportunities for development, improvements, and collaboration.
Here, we highlight some specific projects that will work well in the summer mentorship setting.&lt;/p>
&lt;p>All students interested in LINQS projects for OSRE/GSoC 2026 should fill out &lt;a href="https://forms.gle/Mr4YR3N35pWDb4uz7" target="_blank" rel="noopener">this form&lt;/a>.
Towards the end of the application window, we will contact those who we believe to be a good fit for a LINQS project.
The form will stop accepting responses once the application window closes.
Do not post on any of the project repositories about OSRE/GSoC
(e.g., comment on an issue that you want to tackle it as a part of OSRE/GSoC 2026).
Remember, these are active repositories that were not created for OSRE/GSoC.&lt;/p>
&lt;h3 id="llm-detection">LLM Detection&lt;/h3>
&lt;ul>
&lt;li>&lt;strong>Topics:&lt;/strong> &lt;code>AI/ML&lt;/code> &lt;code>LLM&lt;/code> &lt;code>Research&lt;/code> &lt;code>Backend&lt;/code>&lt;/li>
&lt;li>&lt;strong>Skills:&lt;/strong> software development, backend, systems, data munging, go, docker&lt;/li>
&lt;li>&lt;strong>Difficulty:&lt;/strong> Challenging&lt;/li>
&lt;li>&lt;strong>Size:&lt;/strong> Large (350 hours)&lt;/li>
&lt;li>&lt;strong>Mentors:&lt;/strong> &lt;a href="mailto:linqs.osre26@gmail.com">Eriq Augustine&lt;/a>, &lt;a href="mailto:linqs.osre26@gmail.com">Fabrice Kurmann&lt;/a>, &lt;a href="mailto:linqs.osre26@gmail.com">Lise Getoor&lt;/a>&lt;/li>
&lt;/ul>
&lt;p>As &lt;a href="https://en.wikipedia.org/wiki/Large_language_model" target="_blank" rel="noopener">Large Language Model (LLM)&lt;/a> tools like ChatGPT become more common and powerful,
instructors need tools to help determine if students are the actual authors of the code they submit.
More classical instances of plagiarism are often discovered by code similarity tools like &lt;a href="https://theory.stanford.edu/~aiken/moss/" target="_blank" rel="noopener">MOSS&lt;/a>.
However these tools are not sufficient for detecting code written not by a student,
but by an AI model like &lt;a href="https://en.wikipedia.org/wiki/ChatGPT" target="_blank" rel="noopener">ChatGPT&lt;/a> or &lt;a href="https://en.wikipedia.org/wiki/GitHub_Copilot" target="_blank" rel="noopener">GitHub Copilot&lt;/a>.&lt;/p>
&lt;p>The task for this project is to create a system that provides a score indicating the system&amp;rsquo;s confidence that a given piece of code was written by an AI tool and not a student.
This will supplement the existing code analysis tools in the Lynx Grader.
There are many approaches to completing this task that will be considered.
A more software development approach can consist of levering exiting systems to create a production-ready system,
whereas a more research approach can consist of creating a novel approach complete with a paper and experiments.&lt;/p>
&lt;p>There has been &lt;a href="https://github.com/anvichip/AI-code-detection-ML/blob/main/experiment/report.md" target="_blank" rel="noopener">previous work on this issue&lt;/a>,
where a student did a survey of existing solutions, collection of initial datasets, and exploratory experiments on possible directions.
This project would build off of this previous work.&lt;/p>
&lt;p>See Also:&lt;/p>
&lt;ul>
&lt;li>&lt;a href="https://github.com/edulinq/autograder-server" target="_blank" rel="noopener">Repository for Lynx Grader Server&lt;/a>&lt;/li>
&lt;li>&lt;a href="https://github.com/edulinq/autograder-server/issues/140" target="_blank" rel="noopener">GitHub Issue&lt;/a>&lt;/li>
&lt;/ul>
&lt;h3 id="code-analysis-gui">Code Analysis GUI&lt;/h3>
&lt;ul>
&lt;li>&lt;strong>Topics:&lt;/strong> &lt;code>Frontend&lt;/code>&lt;/li>
&lt;li>&lt;strong>Skills:&lt;/strong> software development, frontend, data munging, js, css, go&lt;/li>
&lt;li>&lt;strong>Difficulty:&lt;/strong> Easy&lt;/li>
&lt;li>&lt;strong>Size:&lt;/strong> Medium or Large (175 or 350 hours)&lt;/li>
&lt;li>&lt;strong>Mentors:&lt;/strong> &lt;a href="mailto:linqs.osre26@gmail.com">Eriq Augustine&lt;/a>, &lt;a href="mailto:linqs.osre26@gmail.com">Fabrice Kurmann&lt;/a>, &lt;a href="mailto:linqs.osre26@gmail.com">Lise Getoor&lt;/a>&lt;/li>
&lt;/ul>
&lt;p>The Lynx Grader has existing functionality to analyze the code in a student&amp;rsquo;s submission for malicious content.
Relevant to this project is that the Lynx Grader can run a pairwise similarity analysis against all submitted code.
This is how most existing software plagiarism systems detect offending code.
The existing infrastructure provides detailed statistics on code similarity,
but does not currently have a visual way to display this data.&lt;/p>
&lt;p>The task for this project is to create a web GUI using the Lynx Grader REST API
to display the results of a code analysis.
The size of this project depends on how many of the existing features are going to be supported by the web GUI.&lt;/p>
&lt;p>See Also:&lt;/p>
&lt;ul>
&lt;li>&lt;a href="https://github.com/edulinq/autograder-web" target="_blank" rel="noopener">Repository for Lynx Grader Web GUI&lt;/a>&lt;/li>
&lt;li>&lt;a href="https://github.com/edulinq/autograder-server/issues/142" target="_blank" rel="noopener">GitHub Issue&lt;/a>&lt;/li>
&lt;li>&lt;a href="https://github.com/edulinq/autograder-server/blob/main/internal/model/analysis.go#L78" target="_blank" rel="noopener">Pairwise Code Analysis Type&lt;/a>&lt;/li>
&lt;li>&lt;a href="https://github.com/edulinq/autograder-py/blob/v0.6.16/tests/api/testdata/courses/assignments/analysis/courses_assignments_submissions_analysis_pairwise_wait.json" target="_blank" rel="noopener">Sample API Data&lt;/a>&lt;/li>
&lt;/ul>
&lt;h3 id="web-gui">Web GUI&lt;/h3>
&lt;ul>
&lt;li>&lt;strong>Topics:&lt;/strong> &lt;code>Frontend&lt;/code>&lt;/li>
&lt;li>&lt;strong>Skills:&lt;/strong> software development, frontend, js, css&lt;/li>
&lt;li>&lt;strong>Difficulty:&lt;/strong> Easy&lt;/li>
&lt;li>&lt;strong>Size:&lt;/strong> Medium or Large (175 or 350 hours)&lt;/li>
&lt;li>&lt;strong>Mentors:&lt;/strong> &lt;a href="mailto:linqs.osre26@gmail.com">Eriq Augustine&lt;/a>, &lt;a href="mailto:linqs.osre26@gmail.com">Fabrice Kurmann&lt;/a>, &lt;a href="mailto:linqs.osre26@gmail.com">Lise Getoor&lt;/a>&lt;/li>
&lt;/ul>
&lt;p>The Lynx Grader contains dozens of &lt;a href="https://github.com/edulinq/autograder-server/blob/main/resources/api.json" target="_blank" rel="noopener">API endpoints&lt;/a>,
most directly representing a piece of functionality exposed to the user.
All of these features are exposed in the &lt;a href="https://github.com/edulinq/autograder-py" target="_blank" rel="noopener">Lynx Grader&amp;rsquo;s Python Interface&lt;/a>.
However, the Python interface is a purely command-line interface.
And although command-line interface are objectively (read: subjectively) the best,
a web GUI would be more accessible to a wider audience.
The autograder already has a &lt;a href="https://github.com/edulinq/autograder-web" target="_blank" rel="noopener">web GUI&lt;/a>,
but it does not cover all the features available in the Lynx Grader.&lt;/p>
&lt;p>The task for this project is to augment the Lynx Grader&amp;rsquo;s web GUI with more features.
Specifically, add support for more tools used to create and administer courses.&lt;/p>
&lt;p>See Also:&lt;/p>
&lt;ul>
&lt;li>&lt;a href="https://github.com/edulinq/autograder-web" target="_blank" rel="noopener">Repository for Lynx Grader Web GUI&lt;/a>&lt;/li>
&lt;li>&lt;a href="https://github.com/edulinq/autograder-server/issues/61" target="_blank" rel="noopener">GitHub Issue&lt;/a>&lt;/li>
&lt;li>&lt;a href="https://github.com/edulinq/autograder-server/blob/main/resources/api.json" target="_blank" rel="noopener">Lynx Grader API Endpoints&lt;/a>&lt;/li>
&lt;li>&lt;a href="https://github.com/edulinq/autograder-py" target="_blank" rel="noopener">Lynx Grader&amp;rsquo;s Python Interface&lt;/a>&lt;/li>
&lt;/ul></description></item><item><title>Quiz Composer</title><link>https://deploy-preview-1007--ucsc-ospo.netlify.app/project/osre26/ucsc/quiz-composer/</link><pubDate>Tue, 13 Jan 2026 13:00:00 -0800</pubDate><guid>https://deploy-preview-1007--ucsc-ospo.netlify.app/project/osre26/ucsc/quiz-composer/</guid><description>&lt;p>The &lt;a href="https://github.com/edulinq/quiz-composer" target="_blank" rel="noopener">EduLinq Quiz Composer&lt;/a> (also called the &amp;ldquo;Quiz Generator&amp;rdquo;) is a tool used by several courses at UCSC
to create and maintain platform-agnostic quizzes (including exams and worksheets).
Knowledge assessments like quizzes, exams, and tests are a core part of the learning process for many courses.
However maintaining banks of questions, collaborating on new questions, and converting quizzes to new formats can use up a lot of time,
taking time away from actually working on improving course materials.
The Quiz Composer helps by providing a single text-based format that can be stored in a repository and &amp;ldquo;compiled&amp;rdquo; into many different formats including:
HTML, LaTeX, PDF, Canvas, GradeScope, and QTI.
The &lt;a href="https://linqs.org" target="_blank" rel="noopener">LINQS Lab&lt;/a> has made many contributions to the maintain and improve the Quiz Composer.&lt;/p>
&lt;p>As an open source project, there are endless opportunities for development, improvements, and collaboration.
Here, we highlight some specific projects that will work well in the summer mentorship setting.&lt;/p>
&lt;p>All students interested in LINQS projects for OSRE/GSoC 2026 should fill out &lt;a href="https://forms.gle/Mr4YR3N35pWDb4uz7" target="_blank" rel="noopener">this form&lt;/a>.
Towards the end of the application window, we will contact those who we believe to be a good fit for a LINQS project.
The form will stop accepting responses once the application window closes.
Do not post on any of the project repositories about OSRE/GSoC
(e.g., comment on an issue that you want to tackle it as a part of OSRE/GSoC 2026).
Remember, these are active repositories that were not created for OSRE/GSoC.&lt;/p>
&lt;h3 id="canvas-import">Canvas Import&lt;/h3>
&lt;ul>
&lt;li>&lt;strong>Topics:&lt;/strong> &lt;code>Backend&lt;/code> &lt;code>Teaching Tools&lt;/code> &lt;code>API&lt;/code>&lt;/li>
&lt;li>&lt;strong>Skills:&lt;/strong> software development, backend, rest api, data munging, http request inspection, python&lt;/li>
&lt;li>&lt;strong>Difficulty:&lt;/strong> Moderate&lt;/li>
&lt;li>&lt;strong>Size:&lt;/strong> Medium (175 hours)&lt;/li>
&lt;li>&lt;strong>Mentors:&lt;/strong> &lt;a href="mailto:linqs.osre26@gmail.com">Eriq Augustine&lt;/a>, &lt;a href="mailto:linqs.osre26@gmail.com">Lucas Ellenberger&lt;/a>, &lt;a href="mailto:linqs.osre26@gmail.com">Lise Getoor&lt;/a>&lt;/li>
&lt;/ul>
&lt;p>The Quiz Composer houses quizzes and quiz questions in a simple and unambiguous format based
on &lt;a href="https://en.wikipedia.org/wiki/JSON" target="_blank" rel="noopener">JSON&lt;/a> and &lt;a href="https://en.wikipedia.org/wiki/Markdown" target="_blank" rel="noopener">Markdown&lt;/a> (specifically, the &lt;a href="https://commonmark.org" target="_blank" rel="noopener">CommonMark specification&lt;/a>).
This allows the Quiz Composer to unambiguously create versions of the same quiz in many different formats.
However, creating a quiz in the Quiz Composer format can be a daunting task for those not familiar with JSON or Markdown.
Instead, it would be easier for people to import quizzes from another format into the Quiz Composer format,
and then edit it as they see fit.
Unfortunately not all other quiz formats, namely Canvas in this case, are unambiguous.&lt;/p>
&lt;p>The task for this project is to implement the functionality of importing quizzes from Canvas to the standard Quiz Composer format.
The unambiguous nature of Canvas quizzes makes this task non-trivial,
and adds an additional element of design decisions to this task.
It will be impossible to import quizzes 100% correctly,
but we want to be able to get close enough that most people can import their quizzes without issue.&lt;/p>
&lt;p>See Also:&lt;/p>
&lt;ul>
&lt;li>&lt;a href="https://github.com/edulinq/quiz-composer" target="_blank" rel="noopener">Repository for Quiz Composer&lt;/a>&lt;/li>
&lt;li>&lt;a href="https://github.com/edulinq/quiz-composer/issues/27" target="_blank" rel="noopener">GitHub Issue&lt;/a>&lt;/li>
&lt;/ul>
&lt;h3 id="google-forms-export">Google Forms Export&lt;/h3>
&lt;ul>
&lt;li>&lt;strong>Topics:&lt;/strong> &lt;code>Backend&lt;/code> &lt;code>Teaching Tools&lt;/code> &lt;code>API&lt;/code>&lt;/li>
&lt;li>&lt;strong>Skills:&lt;/strong> software development, backend, rest api, data munging, python&lt;/li>
&lt;li>&lt;strong>Difficulty:&lt;/strong> Moderate&lt;/li>
&lt;li>&lt;strong>Size:&lt;/strong> Medium (175 hours)&lt;/li>
&lt;li>&lt;strong>Mentors:&lt;/strong> &lt;a href="mailto:linqs.osre26@gmail.com">Eriq Augustine&lt;/a>, &lt;a href="mailto:linqs.osre26@gmail.com">Lucas Ellenberger&lt;/a>, &lt;a href="mailto:linqs.osre26@gmail.com">Lise Getoor&lt;/a>&lt;/li>
&lt;/ul>
&lt;p>The Quiz Composer can export quizzes to many different formats,
each with a varying level of interactivity and feature support.
For example, quizzes can be exported to PDFs which will be printed and the students will just write down their answers to be checked in the future.
Quizzes can also be exported to interactive platforms like Canvas where students can enter answers that may be automatically checked with feedback immediately provided to the student.
On potential platform with functionality somewhere between the above two examples is &lt;a href="https://workspace.google.com/products/forms/" target="_blank" rel="noopener">Google Forms&lt;/a>.
&amp;ldquo;Forms&amp;rdquo; (an entity on Google Forms) can be something like a survey or (as of more recently) a quiz.&lt;/p>
&lt;p>The task for this project is to add support for exporting quizzes from the Quiz Composer to Google Forms.
There is a large overlap in the quiz features supported in Canvas (which the Quiz Composer already supports) and Google Forms,
so most settings should be fairly straightforward.
There may be some design work around deciding what features are specific to one quiz platform
and what features can be abstracted to work across several platforms.&lt;/p>
&lt;p>See Also:&lt;/p>
&lt;ul>
&lt;li>&lt;a href="https://github.com/edulinq/quiz-composer" target="_blank" rel="noopener">Repository for Quiz Composer&lt;/a>&lt;/li>
&lt;li>&lt;a href="https://github.com/edulinq/quiz-composer/issues/19" target="_blank" rel="noopener">GitHub Issue&lt;/a>&lt;/li>
&lt;/ul>
&lt;h3 id="template-questions">Template Questions&lt;/h3>
&lt;ul>
&lt;li>&lt;strong>Topics:&lt;/strong> &lt;code>Backend&lt;/code> &lt;code>Teaching Tools&lt;/code> &lt;code>API&lt;/code>&lt;/li>
&lt;li>&lt;strong>Skills:&lt;/strong> software development, backend, data munging, python&lt;/li>
&lt;li>&lt;strong>Difficulty:&lt;/strong> Moderate-Challenging&lt;/li>
&lt;li>&lt;strong>Size:&lt;/strong> Large (350 hours)&lt;/li>
&lt;li>&lt;strong>Mentors:&lt;/strong> &lt;a href="mailto:linqs.osre26@gmail.com">Eriq Augustine&lt;/a>, &lt;a href="mailto:linqs.osre26@gmail.com">Lucas Ellenberger&lt;/a>, &lt;a href="mailto:linqs.osre26@gmail.com">Lise Getoor&lt;/a>&lt;/li>
&lt;/ul>
&lt;p>Questions in the Quiz Composer are described using &lt;a href="https://en.wikipedia.org/wiki/JSON" target="_blank" rel="noopener">JSON&lt;/a> and &lt;a href="https://en.wikipedia.org/wiki/Markdown" target="_blank" rel="noopener">Markdown&lt;/a>
files which contain the question prompt, possible answers, and the correct answer.
(Of course there are many differ &lt;a href="https://github.com/edulinq/quiz-composer/blob/main/docs/question-types.md" target="_blank" rel="noopener">question types&lt;/a>,
each with different semantics and requirements.)
However, a limitation of this is that each question is always the same.
You can have multiple copies of a question with slightly different prompts, numbers, and answers;
but you are still limited to each question being static and unchanging.
It would be useful to have &amp;ldquo;template questions&amp;rdquo; that can dynamically create static questions from a template
and collection of replacement data.&lt;/p>
&lt;p>The task for this project is to add support for the &amp;ldquo;template questions&amp;rdquo; discussed above.
Much of the high-level design work for this issue has &lt;a href="https://github.com/edulinq/quiz-composer/issues/26" target="_blank" rel="noopener">already been completed&lt;/a>.
But there is still the implementation and low-level design decision left to do.&lt;/p>
&lt;p>See Also:&lt;/p>
&lt;ul>
&lt;li>&lt;a href="https://github.com/edulinq/quiz-composer" target="_blank" rel="noopener">Repository for Quiz Composer&lt;/a>&lt;/li>
&lt;li>&lt;a href="https://github.com/edulinq/quiz-composer/issues/26" target="_blank" rel="noopener">GitHub Issue&lt;/a>&lt;/li>
&lt;/ul></description></item><item><title>Writing a blog about your OSRE 2026 project</title><link>https://deploy-preview-1007--ucsc-ospo.netlify.app/report/osre26/ucsc/admin/20241021-admin/</link><pubDate>Mon, 21 Oct 2024 00:00:00 +0000</pubDate><guid>https://deploy-preview-1007--ucsc-ospo.netlify.app/report/osre26/ucsc/admin/20241021-admin/</guid><description>&lt;p>OSRE participants are required to blog three times during their summer program. The first blog is a chance to introduce yourself and your project. The second blog occurs around the mid-point of the project and a final blog post is expected as part of you final project delverable. The organization administrator will send emails with specific dates. Instructions for the blog are indicated below. All blogs should include links to proposals, presentations, links to any deliverables/products as well as an overview of the student&amp;rsquo;s experience. Check out the student pages from previous years to get an idea of content / size.&lt;/p>
&lt;p>We will also ask students and contributors to provide regular status updates which will help track your activities. The organization administrator will provide more details once the program work begins.&lt;/p>
&lt;h2 id="making-a-pull-request-for-your-blog">Making a pull request for your blog&lt;/h2>
&lt;ul>
&lt;li>
&lt;p>Fork the &lt;a href="https://github.com/ucsc-ospo/ucsc-ospo.github.io" target="_blank" rel="noopener">git repository&lt;/a>&lt;/p>
&lt;/li>
&lt;li>
&lt;p>If you haven&amp;rsquo;t already done so, add your profile using &lt;a href="https://deploy-preview-1007--ucsc-ospo.netlify.app/osredocs/formentors/#instructions-for-adding-a-mentor">these instructions&lt;/a>&lt;/p>
&lt;ul>
&lt;li>&lt;strong>IMPORTANT&lt;/strong>: Under &lt;code>user_groups:&lt;/code> add &lt;code>- 2026 Contributors&lt;/code> (as opposed to any of the two mentor groups)&lt;/li>
&lt;li>The short bio and any other information goes below the frontmatter&lt;/li>
&lt;/ul>
&lt;/li>
&lt;li>
&lt;p>Post your blog&lt;/p>
&lt;ul>
&lt;li>Add &lt;code>/content/report/osre26/ORGANIZATION/PROJECTNAME/DATE-USERNAME/index.md&lt;/code>&lt;/li>
&lt;li>Add a frontmatter to &lt;code>index.md&lt;/code>, using the labels below&lt;/li>
&lt;li>Blog text goes below the frontmatter&lt;/li>
&lt;li>In that same directory include a picture and call it &lt;code>featured.png&lt;/code> (also supports &lt;code>.jpg&lt;/code>, &lt;code>.jpeg&lt;/code>)&lt;/li>
&lt;/ul>
&lt;/li>
&lt;li>
&lt;p>Commit to your fork and make a pull request. &lt;a href="mailto:ospo-info-group@ucsc.edu/">Email OSRE Admins&lt;/a> with questions.&lt;/p>
&lt;/li>
&lt;/ul>
&lt;h3 id="example-frontmatter-and-text-body">Example frontmatter and text body&lt;/h3>
&lt;div class="highlight">&lt;pre tabindex="0" class="chroma">&lt;code class="language-fallback" data-lang="fallback">&lt;span class="line">&lt;span class="cl">---
&lt;/span>&lt;/span>&lt;span class="line">&lt;span class="cl">title: &amp;#34;YOUR TITLE&amp;#34;
&lt;/span>&lt;/span>&lt;span class="line">&lt;span class="cl">subtitle: &amp;#34;YOUR SUBTITLE (OPTIONAL)&amp;#34;
&lt;/span>&lt;/span>&lt;span class="line">&lt;span class="cl">summary:
&lt;/span>&lt;/span>&lt;span class="line">&lt;span class="cl">authors:
&lt;/span>&lt;/span>&lt;span class="line">&lt;span class="cl"> - USERNAME1
&lt;/span>&lt;/span>&lt;span class="line">&lt;span class="cl"> - USERNAME2
&lt;/span>&lt;/span>&lt;span class="line">&lt;span class="cl">tags: [&amp;#34;osre26&amp;#34;]
&lt;/span>&lt;/span>&lt;span class="line">&lt;span class="cl">categories: []
&lt;/span>&lt;/span>&lt;span class="line">&lt;span class="cl">date: YYYY-MM-DD
&lt;/span>&lt;/span>&lt;span class="line">&lt;span class="cl">lastmod: YYYY-MM-DD
&lt;/span>&lt;/span>&lt;span class="line">&lt;span class="cl">featured: false
&lt;/span>&lt;/span>&lt;span class="line">&lt;span class="cl">draft: false
&lt;/span>&lt;/span>&lt;span class="line">&lt;span class="cl">
&lt;/span>&lt;/span>&lt;span class="line">&lt;span class="cl"># Featured image
&lt;/span>&lt;/span>&lt;span class="line">&lt;span class="cl"># To use, add an image named `featured.jpg/png` to your page&amp;#39;s folder.
&lt;/span>&lt;/span>&lt;span class="line">&lt;span class="cl"># Focal points: Smart, Center, TopLeft, Top, TopRight, Left, Right, BottomLeft, Bottom, BottomRight.
&lt;/span>&lt;/span>&lt;span class="line">&lt;span class="cl">image:
&lt;/span>&lt;/span>&lt;span class="line">&lt;span class="cl"> caption: &amp;#34;&amp;#34;
&lt;/span>&lt;/span>&lt;span class="line">&lt;span class="cl"> focal_point: &amp;#34;&amp;#34;
&lt;/span>&lt;/span>&lt;span class="line">&lt;span class="cl"> preview_only: false
&lt;/span>&lt;/span>&lt;span class="line">&lt;span class="cl">---
&lt;/span>&lt;/span>&lt;span class="line">&lt;span class="cl">
&lt;/span>&lt;/span>&lt;span class="line">&lt;span class="cl">As part of the [PROJECTNAME](/project/osre26/ORGANIZATION/PROJECTNAME) my [proposal](https://...) under the mentorship of MENTOR aims to ...
&lt;/span>&lt;/span>&lt;/code>&lt;/pre>&lt;/div></description></item></channel></rss>