<?xml version="1.0" encoding="utf-8" standalone="yes"?><rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom"><channel><title>Health | UCSC OSPO</title><link>https://deploy-preview-1007--ucsc-ospo.netlify.app/tag/health/</link><atom:link href="https://deploy-preview-1007--ucsc-ospo.netlify.app/tag/health/index.xml" rel="self" type="application/rss+xml"/><description>Health</description><generator>Wowchemy (https://wowchemy.com)</generator><language>en-us</language><lastBuildDate>Mon, 19 Jan 2026 10:15:56 -0700</lastBuildDate><image><url>https://deploy-preview-1007--ucsc-ospo.netlify.app/media/logo_hub6795c39d7c5d58c9535d13299c9651f_74810_300x300_fit_lanczos_3.png</url><title>Health</title><link>https://deploy-preview-1007--ucsc-ospo.netlify.app/tag/health/</link></image><item><title>MedJEPA: Self-Supervised Medical Image Representation Learning with JEPA</title><link>https://deploy-preview-1007--ucsc-ospo.netlify.app/project/osre26/nelbl/medjepa/</link><pubDate>Mon, 19 Jan 2026 10:15:56 -0700</pubDate><guid>https://deploy-preview-1007--ucsc-ospo.netlify.app/project/osre26/nelbl/medjepa/</guid><description>&lt;h3 id="project-description">Project Description&lt;/h3>
&lt;p>[MedJEPA] Medical image analysis is fundamental to modern healthcare, enabling disease diagnosis, treatment planning, and patient monitoring across diverse clinical applications. In radiology and pathology, deep learning models support automated detection of abnormalities, tumor segmentation, and diagnostic assistance. Medical imaging modalities including X-rays, CT scans, MRI, ultrasound, and histopathology slides generate vast amounts of unlabeled data that could benefit from self-supervised representation learning. Clinical applications include cancer detection and staging, cardiovascular disease assessment, neurological disorder diagnosis, and infectious disease screening. In drug discovery and clinical research, analyzing medical images helps evaluate treatment efficacy, predict patient outcomes, and identify biomarkers for disease progression. Telemedicine and point-of-care diagnostics benefit from AI-powered image analysis that extends expert-level interpretation to underserved regions. However, medical imaging faces unique challenges: limited labeled datasets due to expensive expert annotation, patient privacy concerns restricting data sharing, domain shift across different imaging equipment and protocols, and the need for models that generalize across hospitals and populations.
Traditional medical image analysis relies heavily on supervised learning with manually annotated labels, creating bottlenecks due to the scarcity and cost of expert annotations. Existing self-supervised methods applied to medical imaging often employ complex training procedures with numerous heuristics—momentum encoders, stop-gradients, teacher-student architectures, and carefully tuned augmentation strategies—that may not translate well across different medical imaging modalities and clinical contexts. These approaches struggle with domain-specific challenges such as subtle pathological features, high-resolution images, 3D volumetric data, and the need for interpretable representations that clinicians can trust. To address these challenges, we propose MedicalJEPA: Self-Supervised Medical Image Representation Learning with Joint-Embedding Predictive Architecture, which leverages the theoretically grounded LeJEPA framework for 2D medical images and V-JEPA principles for medical video and volumetric data, creating a unified, scalable, and heuristics-free approach specifically tailored for medical imaging applications.
By utilizing the principled JEPA frameworks with objectives like Sketched Isotropic Gaussian Regularization (SIGReg), MedJEPA eliminates complex training heuristics while learning clinically meaningful representations from unlabeled medical images. Unlike conventional self-supervised methods that require extensive hyperparameter tuning and may not generalize across medical imaging modalities, MedicalJEPA provides a clean, theoretically motivated framework with minimal hyperparameters that adapts to diverse medical imaging contexts—from chest X-rays to histopathology slides to cardiac MRI sequences. The learned representations can support downstream tasks including disease classification, lesion detection, organ segmentation, and survival prediction, while requiring significantly fewer labeled examples for fine-tuning. This approach democratizes access to state-of-the-art medical AI by enabling effective learning from the vast amounts of unlabeled medical imaging data available in hospital archives, addressing the annotation bottleneck that has limited progress in medical AI.&lt;/p>
&lt;h3 id="project-objectives">Project Objectives&lt;/h3>
&lt;p>Aligned with the vision of the 2026 Open Source Research Experience (OSRE), this project aims to apply Joint-Embedding Predictive Architecture (JEPA) frameworks to medical image representation learning, addressing the critical challenge of learning from limited labeled medical data. Medical imaging generates enormous amounts of unlabeled data, but supervised learning approaches are bottlenecked by the scarcity and cost of expert annotations. Existing self-supervised methods often rely on complex heuristics that don&amp;rsquo;t generalize well across diverse medical imaging modalities, equipment vendors, and clinical protocols.
This project will leverage the theoretically grounded LeJEPA framework for 2D medical images (X-rays, histopathology slides, fundus images) and V-JEPA principles for temporal and volumetric medical data (cardiac MRI sequences, CT scans, surgical videos). The core challenge lies in adapting these heuristics-free, stable frameworks to medical imaging&amp;rsquo;s unique characteristics: subtle pathological features requiring fine-grained representations, high-resolution images demanding efficient processing, domain shift across hospitals and equipment, and the need for interpretable features that support clinical decision-making. The learned representations will be evaluated on diverse downstream clinical tasks including disease classification, lesion detection, organ segmentation, and prognosis prediction, with emphasis on few-shot learning scenarios that reflect real-world annotation constraints. Below is an outline of the methodologies and models that will be developed in this project.&lt;/p>
&lt;ul>
&lt;li>
&lt;p>&lt;strong>Step 1: Medical Data Preparation&lt;/strong>:
Develop data processing pipelines for diverse medical imaging modalities, implementing DICOM/NIfTI parsing, standardized preprocessing, and efficient data loading for self-supervised pre-training.
Prepare 2D medical image datasets:
Chest X-rays: ChestX-ray14, MIMIC-CXR, CheXpert for lung disease detection
Histopathology: Camelyon16/17 (breast cancer), PCam (patch-level classification)
Retinal imaging: EyePACS, APTOS (diabetic retinopathy), Messidor
Dermatology: HAM10000, ISIC (skin lesion classification)
Prepare 3D volumetric and temporal medical data:
CT scans: LIDC-IDRI (lung nodules), Medical Segmentation Decathlon datasets
MRI sequences: BraTS (brain tumors), ACDC (cardiac MRI), UK Biobank cardiac videos
Medical video: Surgical procedure videos, endoscopy recordings, ultrasound sequences
Implement medical imaging-specific preprocessing: intensity normalization, resolution standardization, handling of multi-channel medical images (different MRI sequences, RGB histopathology), and privacy-preserving anonymization.
Design masking strategies appropriate for medical imaging: spatial masking for 2D images, volumetric masking for 3D scans, temporal masking for sequences, and anatomy-aware masking that respects organ boundaries.
Create data loaders supporting high-resolution medical images, 3D volumes, and multi-modal inputs (e.g., multiple MRI sequences).&lt;/p>
&lt;/li>
&lt;li>
&lt;p>&lt;strong>Step 2: JEPA Model Implementation for Medical Imaging&lt;/strong>:
Implement LeJEPA for 2D medical images:
Adapt joint-embedding predictive architecture for medical image characteristics (high resolution, subtle features, domain-specific patterns)
Apply Sketched Isotropic Gaussian Regularization (SIGReg) to learn clinically meaningful embedding distributions
Maintain single trade-off hyperparameter and heuristics-free training for reproducibility across medical imaging centers
Support various encoder architectures: Vision Transformers for global context, ConvNets for local features, hybrid approaches
Extend to V-JEPA for medical video and volumetric data:
Spatiotemporal encoding for cardiac MRI sequences, surgical videos, and time-series medical imaging
Temporal prediction objectives for understanding disease progression and treatment response
3D volume processing for CT and MRI scans with efficient memory management
Multi-slice and multi-sequence learning for comprehensive medical imaging contexts
Develop medical domain-specific enhancements:
Multi-scale representation learning to capture both fine-grained pathological details and global anatomical context
Interpretability mechanisms: attention visualization, feature attribution, and embedding space analysis for clinical validation
Robustness to domain shift: training strategies that generalize across different scanners, protocols, and institutions
Privacy-preserving training considerations compatible with medical data regulations (HIPAA, GDPR)
Implement efficient training infrastructure:
Support for distributed training across multiple GPUs for large medical imaging datasets
Memory-efficient processing of high-resolution images and 3D volumes
Checkpoint management and model versioning for clinical deployment pipelines
Minimal-code implementation (≈50-100 lines) demonstrating framework simplicity&lt;/p>
&lt;/li>
&lt;li>
&lt;p>&lt;strong>Step 3: Evaluation &amp;amp; Safety Validation&lt;/strong>: :
Disease Classification Tasks:
Multi-label chest X-ray classification: 14 pathology classes on ChestX-ray14, MIMIC-CXR
Diabetic retinopathy grading: 5-class classification on EyePACS, APTOS
Skin lesion classification: 7-class classification on HAM10000
Brain tumor classification: glioma grading on BraTS dataset
Evaluate with linear probing, few-shot learning (5-shot, 10-shot), and full fine-tuning
Lesion Detection and Segmentation:
Lung nodule detection on LIDC-IDRI dataset
Tumor segmentation on Medical Segmentation Decathlon tasks
Polyp detection in colonoscopy videos
Cardiac structure segmentation in MRI sequences
Clinical Prediction Tasks:
Survival prediction from histopathology slides
Disease progression prediction from longitudinal imaging
Treatment response assessment from pre/post imaging pairs
Few-Shot and Low-Data Regime Evaluation:
Systematic evaluation with 1%, 5%, 10%, 25%, 50% of labeled training data
Comparison against supervised baselines and ImageNet pre-training
Analysis of annotation efficiency: performance vs. number of labeled examples required&lt;/p>
&lt;/li>
&lt;/ul>
&lt;h3 id="project-deliverables">Project Deliverables&lt;/h3>
&lt;p>This project will deliver three components: software implementation, clinical evaluation, and practical deployment resources. The software implementing MedicalJEPA will be hosted on GitHub as an open-access repository with modular code supporting multiple medical imaging modalities (2D images, 3D volumes, videos), pre-trained model checkpoints on major medical imaging datasets (chest X-rays, histopathology, MRI), training and evaluation scripts with medical imaging-specific preprocessing pipelines, privacy-preserving training implementations compatible with clinical data regulations, and comprehensive documentation including tutorials for medical AI researchers and clinicians. The evaluation results will include benchmarks on 10+ medical imaging datasets across diverse modalities and clinical tasks, few-shot learning analysis demonstrating annotation efficiency gains, cross-institutional validation studies showing robustness to domain shift, interpretability visualizations enabling clinical validation of learned representations, and detailed comparisons against supervised baselines and existing medical self-supervised methods. .&lt;/p>
&lt;h3 id="neurohealth">NeuroHealth&lt;/h3>
&lt;ul>
&lt;li>&lt;strong>Topics&lt;/strong>: Self-Supervised Medical Image Representation Learning with JEPA&lt;/li>
&lt;li>&lt;strong>Skills&lt;/strong>: Proficiency in Python, Pytorch, Github, JEPA&lt;/li>
&lt;li>&lt;strong>Difficulty&lt;/strong>: Difficult&lt;/li>
&lt;li>&lt;strong>Size&lt;/strong>: Large (350 hours)&lt;/li>
&lt;li>&lt;strong>Mentor&lt;/strong>: &lt;a href="https://deploy-preview-1007--ucsc-ospo.netlify.app/author/bin-dong/">Bin Dong&lt;/a>, &lt;a href="https://deploy-preview-1007--ucsc-ospo.netlify.app/author/linsey-pang/">Linsey Pang&lt;/a>&lt;/li>
&lt;/ul>
&lt;h3 id="references">References:&lt;/h3>
&lt;ul>
&lt;li>LeJEPA: Provable and Scalable Self-Supervised Learning Without the Heuristics - Randall Balestriero and Yann LeCun, arXiv 2024&lt;/li>
&lt;li>Revisiting Feature Prediction for Learning Visual Representations from Video (V-JEPA) - Adrien Bardes et al., arXiv 2024&lt;/li>
&lt;li>Self-Supervised Learning from Images with a Joint-Embedding Predictive Architecture - Mahmoud Assran et al., CVPR 2023 (I-JEPA)&lt;/li>
&lt;li>ChestX-ray14: Hospital-Scale Chest X-Ray Database - &lt;a href="https://nihcc.app.box.com/v/ChestXray-NIHCC" target="_blank" rel="noopener">https://nihcc.app.box.com/v/ChestXray-NIHCC&lt;/a>&lt;/li>
&lt;li>Medical Segmentation Decathlon - &lt;a href="http://medicaldecathlon.com/" target="_blank" rel="noopener">http://medicaldecathlon.com/&lt;/a>&lt;/li>
&lt;li>MIMIC-CXR Database - &lt;a href="https://physionet.org/content/mimic-cxr/" target="_blank" rel="noopener">https://physionet.org/content/mimic-cxr/&lt;/a>&lt;/li>
&lt;li>The Cancer Imaging Archive (TCIA) - &lt;a href="https://www.cancerimagingarchive.net/" target="_blank" rel="noopener">https://www.cancerimagingarchive.net/&lt;/a>&lt;/li>
&lt;li>UK Biobank Imaging Study - &lt;a href="https://www.ukbiobank.ac.uk/enable-your-research/about-our-data/imaging-data" target="_blank" rel="noopener">https://www.ukbiobank.ac.uk/enable-your-research/about-our-data/imaging-data&lt;/a>&lt;/li>
&lt;/ul></description></item><item><title>NeuroHealth: AI-Powered Health Assistant</title><link>https://deploy-preview-1007--ucsc-ospo.netlify.app/project/osre26/nelbl/neurohealth/</link><pubDate>Mon, 19 Jan 2026 10:15:56 -0700</pubDate><guid>https://deploy-preview-1007--ucsc-ospo.netlify.app/project/osre26/nelbl/neurohealth/</guid><description>&lt;h3 id="project-description">Project Description&lt;/h3>
&lt;p>[NeuroHealth] Intelligent health assistance systems are increasingly essential for improving healthcare accessibility, patient engagement, and clinical decision support. In primary care and preventive medicine, AI assistants help users understand symptoms, schedule appropriate appointments, and receive preliminary health guidance. Telemedicine applications include triage support, appointment scheduling optimization, and patient education based on health inquiries. In chronic disease management, these systems provide medication reminders, lifestyle recommendations, and timely alerts for medical follow-ups. Healthcare navigation applications include finding appropriate specialists, understanding treatment options, and coordinating care across multiple providers. In wellness and preventive care, intelligent assistants enhance health literacy by delivering personalized health information, screening recommendations, and proactive health management strategies. By leveraging natural language understanding and medical knowledge integration, these systems enhance healthcare access, reduce unnecessary emergency visits, and empower users to make informed health decisions across diverse populations.
Traditional health information systems often provide generic responses that fail to account for individual health contexts, medical history, and personal circumstances. Existing symptom checkers and health chatbots primarily rely on rule-based logic or simple decision trees, limiting their ability to understand nuanced health inquiries, reason about complex symptom patterns, or provide contextually appropriate guidance. These systems struggle with interpreting ambiguous descriptions, adapting to users&amp;rsquo; health literacy levels, and generating personalized recommendations that account for individual medical constraints and preferences. To address these challenges, we propose NeuroHealth: AI-Powered Health Assistant, which leverages Large Language Models (LLMs) to create an intelligent conversational agent that synthesizes user health inquiries, symptom descriptions, and contextual information into actionable, personalized health guidance and appointment recommendations.
By integrating LLM-based medical reasoning with structured clinical knowledge bases, NeuroHealth enhances symptom interpretation, appointment routing, and health education delivery. Unlike conventional systems that provide static responses from predetermined templates, NeuroHealth dynamically understands user intent, asks clarifying questions, assesses urgency levels, and generates appropriate recommendations—whether scheduling a doctor appointment, suggesting self-care measures, or directing users to emergency services. This fusion of LLM intelligence with validated medical knowledge enables a more accessible, adaptive, and helpful health assistance platform, bridging the gap between users seeking health information and appropriate medical care.&lt;/p>
&lt;h3 id="project-objectives">Project Objectives&lt;/h3>
&lt;p>Aligned with the vision of the 2026 Open Source Research Experience (OSRE), this project aims to develop an AI-Powered Health Assistant (NeuroHealth) to improve healthcare accessibility and patient engagement through intelligent conversational guidance. Healthcare systems face significant challenges in providing timely, personalized health information and connecting patients with appropriate care resources. Traditional symptom checkers and health information systems often deliver generic, rule-based responses that fail to account for individual contexts and struggle with natural language understanding.
To address these limitations, this project will leverage Large Language Models (LLMs) to create an intelligent health assistant that understands user health inquiries, interprets symptom descriptions, assesses urgency, and provides personalized recommendations including doctor appointment suggestions, self-care guidance, and healthcare navigation support. The core challenge lies in designing NeuroHealth as a safe, accurate, and user-friendly system capable of natural conversation, medical knowledge retrieval, and appropriate response generation while maintaining clinical safety guardrails. Unlike conventional health chatbots that follow rigid conversation flows, NeuroHealth will reason over user inputs, ask clarifying questions, and dynamically adapt responses based on context, resulting in more helpful, accurate, and appropriate health assistance. Below is an outline of the methodologies and models that will be developed in this project.&lt;/p>
&lt;ul>
&lt;li>
&lt;p>&lt;strong>Step 1: Data Collection &amp;amp; Knowledge Base Construction&lt;/strong>:
Develop a comprehensive medical knowledge base integrating validated health information sources, symptom databases, condition descriptions, and appointment routing guidelines.
Collect and curate conversational health inquiry datasets from public medical Q&amp;amp;A forums, symptom checker logs, and healthcare chatbot interactions to create training and evaluation data.
Design structured representations for symptoms, conditions, urgency levels, and appointment recommendations to enable effective retrieval and reasoning.
Extract common health inquiry patterns, symptom descriptions, and user intent categories to inform conversation flow design.
Data sources can include public medical knowledge bases such as MedlinePlus, Mayo Clinic health information, clinical practice guidelines, and synthetic patient inquiry scenarios based on common healthcare use cases.
Implement data validation mechanisms to ensure medical accuracy and clinical safety compliance.&lt;/p>
&lt;/li>
&lt;li>
&lt;p>&lt;strong>Step 2: Model Development&lt;/strong>:
Design and implement an LLM-based conversational health assistant that integrates medical knowledge retrieval with natural language understanding and generation.
Develop a Retrieval-Augmented Generation (RAG) architecture that grounds LLM responses in validated medical information sources, reducing hallucination risks and ensuring factual accuracy.
Create prompt engineering strategies and reasoning frameworks that enable the system to: interpret symptom descriptions, assess urgency levels, ask appropriate clarifying questions, and generate personalized health guidance.
Implement a multi-component architecture including: intent recognition, symptom extraction, urgency assessment, appointment recommendation generation, and response formatting modules.
Develop clinical safety guardrails that detect high-risk scenarios requiring immediate medical attention and provide appropriate emergency guidance.
Design conversation management strategies that maintain context across multi-turn dialogues and adapt to users&amp;rsquo; health literacy levels.
The baseline architecture can leverage state-of-the-art models such as GPT-4, Claude, or open-source alternatives like Llama, Qwen, combined with medical knowledge retrieval systems.&lt;/p>
&lt;/li>
&lt;li>
&lt;p>&lt;strong>Step 3: Evaluation &amp;amp; Safety Validation&lt;/strong>: :
Benchmark NeuroHealth against existing symptom checkers and health chatbots, evaluating on metrics including response accuracy, appropriateness of appointment recommendations, urgency assessment precision, and user satisfaction.
Conduct human evaluation studies with healthcare professionals to assess clinical safety, response quality, and appropriateness of medical guidance.
Perform adversarial testing to identify potential failure modes, unsafe responses, or inappropriate recommendations under edge cases.
Conduct ablation studies to analyze the impact of retrieval-augmented generation, safety guardrails, and conversation management strategies on system performance.
Evaluate system performance across diverse health inquiry types including acute symptoms, chronic condition management, preventive care questions, and healthcare navigation requests.
Assess response quality across different user demographics and health literacy levels to ensure equitable access.
Optimize inference efficiency and response latency for real-time conversational interaction across web and mobile platforms.&lt;/p>
&lt;/li>
&lt;/ul>
&lt;h3 id="project-deliverables">Project Deliverables&lt;/h3>
&lt;p>This project will deliver three components: model development, evaluation and validation, and interactive demonstration. The software implementing the NeuroHealth system will be hosted on GitHub as an open-access repository with comprehensive documentation, deployment guides, and API specifications. The evaluation results, including benchmark comparisons against existing systems, clinical safety assessments, and user study findings, will be published alongside the GitHub repository. An interactive demo showcasing the conversational interface, symptom interpretation capabilities, and appointment recommendation generation will be provided to illustrate real-world application scenarios.&lt;/p>
&lt;h3 id="neurohealth">NeuroHealth&lt;/h3>
&lt;ul>
&lt;li>&lt;strong>Topics&lt;/strong>: AI-Powered Health Assistant&lt;/li>
&lt;li>&lt;strong>Skills&lt;/strong>: Proficiency in Python, Github, LLM&lt;/li>
&lt;li>&lt;strong>Difficulty&lt;/strong>: Difficult&lt;/li>
&lt;li>&lt;strong>Size&lt;/strong>: Large (350 hours)&lt;/li>
&lt;li>&lt;strong>Mentor&lt;/strong>: &lt;a href="https://deploy-preview-1007--ucsc-ospo.netlify.app/author/linsey-pang/">Linsey Pang&lt;/a>, &lt;a href="https://deploy-preview-1007--ucsc-ospo.netlify.app/author/bin-dong/">Bin Dong&lt;/a>&lt;/li>
&lt;/ul>
&lt;h3 id="references">References:&lt;/h3>
&lt;ul>
&lt;li>Large Language Models in Healthcare - Singhal et al., Nature 2023&lt;/li>
&lt;li>Med-PaLM: Large Language Models for Medical Question Answering - Singhal et al., arXiv 2022&lt;/li>
&lt;li>Capabilities of GPT-4 on Medical Challenge Problems - Nori et al., arXiv 2023&lt;/li>
&lt;li>MedlinePlus Medical Encyclopedia - &lt;a href="https://medlineplus.gov/" target="_blank" rel="noopener">https://medlineplus.gov/&lt;/a>&lt;/li>
&lt;li>Clinical Practice Guidelines Database - &lt;a href="https://www.guidelines.gov/" target="_blank" rel="noopener">https://www.guidelines.gov/&lt;/a>&lt;/li>
&lt;/ul></description></item></channel></rss>