<?xml version="1.0" encoding="utf-8" standalone="yes"?><rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom"><channel><title>ar/vr | UCSC OSPO</title><link>https://deploy-preview-1007--ucsc-ospo.netlify.app/tag/ar/vr/</link><atom:link href="https://deploy-preview-1007--ucsc-ospo.netlify.app/tag/ar/vr/index.xml" rel="self" type="application/rss+xml"/><description>ar/vr</description><generator>Wowchemy (https://wowchemy.com)</generator><language>en-us</language><image><url>https://deploy-preview-1007--ucsc-ospo.netlify.app/media/logo_hub6795c39d7c5d58c9535d13299c9651f_74810_300x300_fit_lanczos_3.png</url><title>ar/vr</title><link>https://deploy-preview-1007--ucsc-ospo.netlify.app/tag/ar/vr/</link></image><item><title>ReasonWorld</title><link>https://deploy-preview-1007--ucsc-ospo.netlify.app/project/osre25/ucsc/reason-world/</link><pubDate>Fri, 28 Feb 2025 00:00:00 +0000</pubDate><guid>https://deploy-preview-1007--ucsc-ospo.netlify.app/project/osre25/ucsc/reason-world/</guid><description>&lt;h3 id="reasonworld-real-world-reasoning-with-a-long-term-world-model">ReasonWorld: Real-World Reasoning with a Long-Term World Model&lt;/h3>
&lt;p>A world model is essentially an internal representation of an environment that an AI system would construct based on external information to plan, reason, and interpret its surroundings. It stores the system’s understanding of relevant objects, spatial relationships, and/or states in the environment. Recent augmented reality (AR) and wearable technologies like Meta Aria glasses provide an opportunity to gather rich information from the real world in the form of vision, audio, and spatial data. Along with this, large language (LLM), vision language models (VLMs), and general machine learning algorithms have enabled nuanced understanding and processing of multimodal inputs that can label, summarize, and analyze experiences.&lt;/p>
&lt;p>With &lt;strong>ReasonWorld&lt;/strong>, we aim to utilize these technologies to enable advanced reasoning about important objects/events/spaces in real-world environments in a structured manner. With the help of wearable AR technology, the system would be able to capture real-world multimodal data. We aim to utilize this information to create a long-memory modeling toolkit that would support features like:&lt;/p>
&lt;ul>
&lt;li>Longitudinal and structured data logging: Capture and storing of multimodal data (image, video, audio, location coordinates etc.)&lt;/li>
&lt;li>Semantic summarization: Automatic scene labeling via LLMs/VLMs to identify key elements in the surroundings&lt;/li>
&lt;li>Efficient retrieval: For querying and revisiting past experiences and answering questions like “Where have I seen this painting before?”&lt;/li>
&lt;li>Adaptability: Continuously refining and understanding the environment and/or relationships between objects/locations.&lt;/li>
&lt;li>Adaptive memory prioritization: Where the pipeline can assess the contextual significance of the captured data and retrieve those that are the most significant. The model retains meaningful, structured representations rather than raw, unfiltered data.&lt;/li>
&lt;/ul>
&lt;p>This real-world reasoning framework with a long-term world model can function as a structured search engine for important objects and spaces, enabling:&lt;/p>
&lt;ul>
&lt;li>Recognizing and tracking significant objects, locations, and events&lt;/li>
&lt;li>Supporting spatial understanding and contextual analysis&lt;/li>
&lt;li>Facilitating structured documentation of environments and changes over time&lt;/li>
&lt;/ul>
&lt;h3 id="alignment-with-summer-of-reproducibility">Alignment with Summer of Reproducibility:&lt;/h3>
&lt;ul>
&lt;li>Core pipeline for AR data ingestion, event segmentation, summarization, and indexing (knowledge graph or vector database) would be made open-source.&lt;/li>
&lt;li>Clear documentation of each module and how they collaborate with one another&lt;/li>
&lt;li>The project could be tested with standardized datasets, simulated environments as well as controlled real-world scenarios, promoting reproducibility&lt;/li>
&lt;li>Opportunities for Innovation - A transparent, modular approach invites a broad community to propose novel expansions&lt;/li>
&lt;/ul>
&lt;h3 id="specific-tasks">Specific Tasks:&lt;/h3>
&lt;ul>
&lt;li>A pipeline for real-time/batch ingestion of data with the wearable AR device and cleaning&lt;/li>
&lt;li>Have an event segmentation module to classify whether the current object/event is contextually significant, filtering out the less relevant observations.&lt;/li>
&lt;li>Have VLMs/LLMs summarize the events with the vision/audio/location data to be stored and retrieved later by structured data structures like knowledge graph, vector databases etc.&lt;/li>
&lt;li>Storage optimization with prioritizing important objects and spaces, optimizing storage based on contextual significance and frequency of access.&lt;/li>
&lt;li>Implement key information retrieval mechanisms&lt;/li>
&lt;li>Ensure reproducibility by providing datasets and scripts&lt;/li>
&lt;/ul>
&lt;h3 id="reasonworld">ReasonWorld&lt;/h3>
&lt;ul>
&lt;li>&lt;strong>Topics:&lt;/strong> &lt;code>Augmented reality&lt;/code> &lt;code>Multimodal learning&lt;/code> &lt;code>Computer vision for AR&lt;/code> &lt;code>LLM/VLM&lt;/code> &lt;code>Efficient data indexing&lt;/code>&lt;/li>
&lt;li>&lt;strong>Skills:&lt;/strong> Machine Learning and AI, Augmented Reality and Hardware integration, Data Engineering &amp;amp; Storage Optimization&lt;/li>
&lt;li>&lt;strong>Difficulty:&lt;/strong> Hard&lt;/li>
&lt;li>&lt;strong>Size:&lt;/strong> Large (350 hours)&lt;/li>
&lt;li>&lt;strong>Mentors:&lt;/strong> &lt;a href="mailto:davisje@ucsc.edu">James Davis&lt;/a>, &lt;a href="mailto:pang@soe.ucsc.edu">Alex Pang&lt;/a>&lt;/li>
&lt;/ul></description></item><item><title>AR4VIP</title><link>https://deploy-preview-1007--ucsc-ospo.netlify.app/project/osre25/ucsc/ar4vip/</link><pubDate>Tue, 18 Feb 2025 00:00:00 +0000</pubDate><guid>https://deploy-preview-1007--ucsc-ospo.netlify.app/project/osre25/ucsc/ar4vip/</guid><description>&lt;p>We are interested in developing navigation aids for visually impaired people (VIP) using AR/VR technologies.
Our intended use is primarily indoors or outdoors but within private confines e.g. person&amp;rsquo;s backyard.
Using AR/VR headsets or smart glasses allows navigation without using a cane and frees
the users&amp;rsquo; hands for other tasks.&lt;/p>
&lt;h3 id="continue-development-on-meta-quest-3-headset">Continue Development on Meta Quest 3 Headset&lt;/h3>
&lt;ul>
&lt;li>&lt;strong>Topics:&lt;/strong> &lt;code>Dynamic scenes&lt;/code> &lt;code>Spatial audio&lt;/code> &lt;code>Proximity detection&lt;/code>&lt;/li>
&lt;li>&lt;strong>Skills:&lt;/strong> AR/VR familiarity, WebXR, Unity, SLAM, good communicator, good documentation skills&lt;/li>
&lt;li>&lt;strong>Difficulty:&lt;/strong> Moderate&lt;/li>
&lt;li>&lt;strong>Size:&lt;/strong> Medium or large (175 or 350 hours)&lt;/li>
&lt;li>&lt;strong>Mentors:&lt;/strong> &lt;a href="mailto:pang@soe.ucsc.edu">Alex Pang&lt;/a>, &lt;a href="davis@cs.ucsc.edu">James Davis&lt;/a>&lt;/li>
&lt;/ul>
&lt;p>Continue development and field testing with the Meta Quest 3 headset.
See this &lt;a href="https://github.com/sail360/UCSC-VIP-Research" target="_blank" rel="noopener">repository page&lt;/a> for current status.&lt;/p>
&lt;p>Specific tasks:&lt;/p>
&lt;ul>
&lt;li>Improve spatial audio mapping&lt;/li>
&lt;li>Improve obstacle detection, at different heights, with pre-scanned geometry as well as dynamic objects
e.g. other people, pets, doors&lt;/li>
&lt;li>Special handling of hazards e.g. stairs, uneven floors, etc.&lt;/li>
&lt;li>Explore/incorporate AI to help identify objects in the scene when requested by user&lt;/li>
&lt;/ul>
&lt;h3 id="new-development-on-smart-glasses">New Development on Smart Glasses&lt;/h3>
&lt;ul>
&lt;li>&lt;strong>Topics:&lt;/strong> &lt;code>Dynamic scenes&lt;/code> &lt;code>Spatial audio&lt;/code> &lt;code>Proximity detection&lt;/code>&lt;/li>
&lt;li>&lt;strong>Skills:&lt;/strong> AR/VR familiarity, WebXR, Unity, SLAM, good communicator, good documentation skills&lt;/li>
&lt;li>&lt;strong>Difficulty:&lt;/strong> Moderate&lt;/li>
&lt;li>&lt;strong>Size:&lt;/strong> Large (350 hours)&lt;/li>
&lt;li>&lt;strong>Mentors:&lt;/strong> &lt;a href="mailto:pang@soe.ucsc.edu">Alex Pang&lt;/a>, &lt;a href="mailto:davis@cs.ucsc.edu">James Davis&lt;/a>&lt;/li>
&lt;/ul>
&lt;p>VR headsets are bulky and awkward, but currently is more advanced than AR glasses in terms of programmability.
Ultimately, the form factor of smart glasses is more practical for extended use by our target users.
There are many vendors working on pushing out their version of smart glasses targetting various applications
e.g. alternative for watching TV, etc. We are interested in those that provide capabilities to support
spatial computing. Most of these will likely have their own brand specific APIs. This project has 2 goals:
(a) develop generic brand-independent API, perhaps extensions to WebXR, to support overarching goal of navigation
aid for VIP, and
(b) port functionality of VR version to smart glasses while taking advantage of smart glass functionalities and sensors.&lt;/p>
&lt;p>Specific tasks:&lt;/p>
&lt;ul>
&lt;li>Explore current and soon-to-be-available smart glass options e.g. Snap Spectacles, Xreal Air 2 ultra, etc. and select a platform to work on (subject to cost and availability of SDK). At a minimum, glass should be microphones and speakers, and cameras. Infrared cameras or other low light capability is a plus. Sufficient battery life or option for quick exchange.&lt;/li>
&lt;li>Identify support provided by SDK e.g. does it do realtime scene reconstruction? does it support spatial audio? etc. If it supports features outside of WebXR, provide generic hooks to improve portability of code to other smart glasses.&lt;/li>
&lt;li>Port and extend functionalities from the Meta Quest 3 VR headsets to smart glass platform.&lt;/li>
&lt;li>Add AI support if glasses support them.&lt;/li>
&lt;li>Provide documentation of work.&lt;/li>
&lt;/ul></description></item></channel></rss>