<?xml version="1.0" encoding="utf-8" standalone="yes"?><rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom"><channel><title>Connor Lee | UCSC OSPO</title><link>https://deploy-preview-1007--ucsc-ospo.netlify.app/author/connor-lee/</link><atom:link href="https://deploy-preview-1007--ucsc-ospo.netlify.app/author/connor-lee/index.xml" rel="self" type="application/rss+xml"/><description>Connor Lee</description><generator>Wowchemy (https://wowchemy.com)</generator><language>en-us</language><image><url>https://deploy-preview-1007--ucsc-ospo.netlify.app/author/connor-lee/avatar_hu658e8c64a8369fa2363b2540a4d67164_25640_270x270_fill_lanczos_center_3.png</url><title>Connor Lee</title><link>https://deploy-preview-1007--ucsc-ospo.netlify.app/author/connor-lee/</link></image><item><title>LLMSeqRec: LLM Enhanced Contextual Sequential Recommender</title><link>https://deploy-preview-1007--ucsc-ospo.netlify.app/report/osre25/sf/llmseqrec/20250722-connor/</link><pubDate>Tue, 22 Jul 2025 10:15:56 -0700</pubDate><guid>https://deploy-preview-1007--ucsc-ospo.netlify.app/report/osre25/sf/llmseqrec/20250722-connor/</guid><description>&lt;h1 id="midway-through-osre">Midway Through OSRE&lt;/h1>
&lt;h2 id="my-journey-with-llmseqrec">My Journey with LLMSeqRec&lt;/h2>
&lt;h3 id="hello-from-the-midpoint">Hello from the Midpoint!&lt;/h3>
&lt;p>Hi everyone! I’m Connor Lee, a student at NYU studying Computer Science and Mathematics, and I’m excited to share the progress I’ve made halfway through the Open Source Research Experience (OSRE) with my project: &lt;strong>LLMSeqRec&lt;/strong> – a large language model-enhanced sequential recommender system.&lt;/p>
&lt;p>Over the past several weeks, I’ve had the opportunity to explore the intersection of recommender systems and large language models (LLMs), and it’s been a deep, challenging, and rewarding dive into building smarter, more contextual recommendation engines.&lt;/p>
&lt;hr>
&lt;h3 id="what-is-llmseqrec">What is LLMSeqRec?&lt;/h3>
&lt;p>&lt;strong>LLMSeqRec&lt;/strong> stands for &lt;strong>LLM-Enhanced Contextual Sequential Recommender&lt;/strong>. Traditional sequential recommendation systems like SASRec are great at capturing patterns from user-item interactions, but they often fall short in two areas: understanding &lt;strong>semantic context&lt;/strong> (e.g., item descriptions, reviews) and dealing with &lt;strong>cold-start&lt;/strong> problems.&lt;/p>
&lt;p>LLMSeqRec aims to address this by incorporating &lt;strong>pretrained LLM embeddings&lt;/strong> into the recommendation pipeline. The goal is to enhance models like SASRec with semantic signals from text (like product reviews or titles), allowing them to better model user intent, long-range dependencies, and generalize to new items or users.&lt;/p>
&lt;hr>
&lt;h3 id="progress-so-far">Progress So Far&lt;/h3>
&lt;h4 id="-baseline-sasrec-runs">✅ Baseline SASRec Runs&lt;/h4>
&lt;p>To establish a benchmark, I successfully ran the original SASRec implementation (in PyTorch) using both the &lt;strong>MovieLens 1M&lt;/strong> and &lt;strong>Amazon Beauty&lt;/strong> datasets. After debugging initial data formatting issues and adjusting batch sizes for local CPU/GPU compatibility, I automated training with scripts that let me scale to &lt;strong>200+ epochs&lt;/strong> to acheive the best performance in both Colab and on my MacBook via CPU.&lt;/p>
&lt;p>&lt;strong>Note:&lt;/strong> At this stage, we have not yet integrated LLMs into the model. These baseline runs (SASRec) serve as the control group for evaluating the future impact of LLM-based enhancements.&lt;/p>
&lt;hr>
&lt;h3 id="whats-next">What’s Next&lt;/h3>
&lt;p>As I enter the second half of the OSRE, I’ll be shifting gears toward &lt;strong>LLM integration, model evaluation, and running LLM-powered sequential recommendations using product metadata and contextual information&lt;/strong>. Here&amp;rsquo;s what’s ahead:&lt;/p>
&lt;ul>
&lt;li>Designing pipelines to extract and align textual metadata with item sequences&lt;/li>
&lt;li>Integrating LLM-generated embeddings into the recommender model&lt;/li>
&lt;li>Evaluating performance changes across different dataset characteristics&lt;/li>
&lt;/ul>
&lt;hr>
&lt;h3 id="-experimental-results">📊 Experimental Results&lt;/h3>
&lt;p>We have &lt;strong>not yet utilized LLMs&lt;/strong> in our current experiments. The results below reflect our &lt;strong>reproduced baseline performance of SASRec&lt;/strong> across datasets.&lt;/p>
&lt;p>Below are the &lt;strong>performance curves on different test sets&lt;/strong>, where we evaluate model performance every 20 epochs during training:&lt;/p>
&lt;h4 id="beauty-dataset-performance">Beauty Dataset Performance&lt;/h4>
&lt;p>
&lt;figure >
&lt;div class="d-flex justify-content-center">
&lt;div class="w-100" >&lt;img alt="Beauty Hit@10 Performance" srcset="
/report/osre25/sf/llmseqrec/20250722-connor/beauty-hr_hu655ec71a9ef1f87543ab22378365f6fe_152488_6d3cf991cc5172e392edbb398afef774.webp 400w,
/report/osre25/sf/llmseqrec/20250722-connor/beauty-hr_hu655ec71a9ef1f87543ab22378365f6fe_152488_91a98a3d515a172aed7283ab8b04a8b6.webp 760w,
/report/osre25/sf/llmseqrec/20250722-connor/beauty-hr_hu655ec71a9ef1f87543ab22378365f6fe_152488_1200x1200_fit_q75_h2_lanczos_3.webp 1200w"
src="https://deploy-preview-1007--ucsc-ospo.netlify.app/report/osre25/sf/llmseqrec/20250722-connor/beauty-hr_hu655ec71a9ef1f87543ab22378365f6fe_152488_6d3cf991cc5172e392edbb398afef774.webp"
width="760"
height="497"
loading="lazy" data-zoomable />&lt;/div>
&lt;/div>&lt;/figure>
&lt;em>Hit@10 performance on the test set for the Beauty dataset (every 20 epochs)&lt;/em>&lt;/p>
&lt;p>
&lt;figure >
&lt;div class="d-flex justify-content-center">
&lt;div class="w-100" >&lt;img alt="Beauty Loss Training" srcset="
/report/osre25/sf/llmseqrec/20250722-connor/beauty-loss-epoch_huc2cddabd12f6ed04444e319cba850bc9_141963_f4e0cc23660b4c974056c8b5d603c0ca.webp 400w,
/report/osre25/sf/llmseqrec/20250722-connor/beauty-loss-epoch_huc2cddabd12f6ed04444e319cba850bc9_141963_7c62f735e3e920d3561bd9113c662533.webp 760w,
/report/osre25/sf/llmseqrec/20250722-connor/beauty-loss-epoch_huc2cddabd12f6ed04444e319cba850bc9_141963_1200x1200_fit_q75_h2_lanczos_3.webp 1200w"
src="https://deploy-preview-1007--ucsc-ospo.netlify.app/report/osre25/sf/llmseqrec/20250722-connor/beauty-loss-epoch_huc2cddabd12f6ed04444e319cba850bc9_141963_f4e0cc23660b4c974056c8b5d603c0ca.webp"
width="760"
height="489"
loading="lazy" data-zoomable />&lt;/div>
&lt;/div>&lt;/figure>
&lt;em>Training loss for the Beauty dataset&lt;/em>&lt;/p>
&lt;p>
&lt;figure >
&lt;div class="d-flex justify-content-center">
&lt;div class="w-100" >&lt;img alt="Beauty NDCG@10 Performance" srcset="
/report/osre25/sf/llmseqrec/20250722-connor/beauty-ndcg_hu4bef43ef38566a5009aa70da37ebbc50_151414_a1a39dc055b888f5de47c25c87ccf913.webp 400w,
/report/osre25/sf/llmseqrec/20250722-connor/beauty-ndcg_hu4bef43ef38566a5009aa70da37ebbc50_151414_3e4c7d0050bef8ec9f8f7928c2c6c7af.webp 760w,
/report/osre25/sf/llmseqrec/20250722-connor/beauty-ndcg_hu4bef43ef38566a5009aa70da37ebbc50_151414_1200x1200_fit_q75_h2_lanczos_3.webp 1200w"
src="https://deploy-preview-1007--ucsc-ospo.netlify.app/report/osre25/sf/llmseqrec/20250722-connor/beauty-ndcg_hu4bef43ef38566a5009aa70da37ebbc50_151414_a1a39dc055b888f5de47c25c87ccf913.webp"
width="760"
height="483"
loading="lazy" data-zoomable />&lt;/div>
&lt;/div>&lt;/figure>
&lt;em>NDCG@10 performance on the test set for the Beauty dataset (every 20 epochs)&lt;/em>&lt;/p>
&lt;h4 id="ml-1m-dataset-performance">ML-1M Dataset Performance&lt;/h4>
&lt;p>
&lt;figure >
&lt;div class="d-flex justify-content-center">
&lt;div class="w-100" >&lt;img alt="ML-1M Loss Training" srcset="
/report/osre25/sf/llmseqrec/20250722-connor/m1-m1-loss-epoch_hua4b125e87ed4debb93bde68ff9b86489_146604_828aa4c04e00024c863cb89e245d358a.webp 400w,
/report/osre25/sf/llmseqrec/20250722-connor/m1-m1-loss-epoch_hua4b125e87ed4debb93bde68ff9b86489_146604_d913a345a32ce7ac5bcff72438283a01.webp 760w,
/report/osre25/sf/llmseqrec/20250722-connor/m1-m1-loss-epoch_hua4b125e87ed4debb93bde68ff9b86489_146604_1200x1200_fit_q75_h2_lanczos_3.webp 1200w"
src="https://deploy-preview-1007--ucsc-ospo.netlify.app/report/osre25/sf/llmseqrec/20250722-connor/m1-m1-loss-epoch_hua4b125e87ed4debb93bde68ff9b86489_146604_828aa4c04e00024c863cb89e245d358a.webp"
width="760"
height="490"
loading="lazy" data-zoomable />&lt;/div>
&lt;/div>&lt;/figure>
&lt;em>Training loss for the ML-1M dataset&lt;/em>&lt;/p>
&lt;p>
&lt;figure >
&lt;div class="d-flex justify-content-center">
&lt;div class="w-100" >&lt;img alt="ML-1M Hit@10 Performance" srcset="
/report/osre25/sf/llmseqrec/20250722-connor/ml-m1-hr_huaac170547624f58b168df1545691a3d4_153677_8e8f20a29b2657093b23e780efd1d072.webp 400w,
/report/osre25/sf/llmseqrec/20250722-connor/ml-m1-hr_huaac170547624f58b168df1545691a3d4_153677_257879da50059e5cc3e64fd8ed1d9d72.webp 760w,
/report/osre25/sf/llmseqrec/20250722-connor/ml-m1-hr_huaac170547624f58b168df1545691a3d4_153677_1200x1200_fit_q75_h2_lanczos_3.webp 1200w"
src="https://deploy-preview-1007--ucsc-ospo.netlify.app/report/osre25/sf/llmseqrec/20250722-connor/ml-m1-hr_huaac170547624f58b168df1545691a3d4_153677_8e8f20a29b2657093b23e780efd1d072.webp"
width="760"
height="484"
loading="lazy" data-zoomable />&lt;/div>
&lt;/div>&lt;/figure>
&lt;em>Hit@10 performance on the test set for the ML-1M dataset (every 20 epochs)&lt;/em>&lt;/p>
&lt;p>
&lt;figure >
&lt;div class="d-flex justify-content-center">
&lt;div class="w-100" >&lt;img alt="ML-1M NDCG@10 Performance" srcset="
/report/osre25/sf/llmseqrec/20250722-connor/ml-m1-ndcg_huad2935749c06fc72562e3df395457d92_144728_dfd4334fbae2a7067cf9f91b1595e36b.webp 400w,
/report/osre25/sf/llmseqrec/20250722-connor/ml-m1-ndcg_huad2935749c06fc72562e3df395457d92_144728_271754cbcc6eac53f84162a93b670d17.webp 760w,
/report/osre25/sf/llmseqrec/20250722-connor/ml-m1-ndcg_huad2935749c06fc72562e3df395457d92_144728_1200x1200_fit_q75_h2_lanczos_3.webp 1200w"
src="https://deploy-preview-1007--ucsc-ospo.netlify.app/report/osre25/sf/llmseqrec/20250722-connor/ml-m1-ndcg_huad2935749c06fc72562e3df395457d92_144728_dfd4334fbae2a7067cf9f91b1595e36b.webp"
width="760"
height="488"
loading="lazy" data-zoomable />&lt;/div>
&lt;/div>&lt;/figure>
&lt;em>NDCG@10 performance on the test set for the ML-1M dataset (every 20 epochs)&lt;/em>&lt;/p>
&lt;p>These results demonstrate that our &lt;strong>baseline SASRec reproductions&lt;/strong> are converging as expected and will serve as a solid foundation for comparison once LLM integration is complete.&lt;/p>
&lt;hr>
&lt;h3 id="closing-thoughts">Closing Thoughts&lt;/h3>
&lt;p>This project has been an exciting journey into both research and engineering and I’m excited to explore &lt;strong>LLM-powered embedding integration&lt;/strong> in the upcoming phase.&lt;/p>
&lt;p>I’m incredibly grateful to my mentors &lt;strong>Dr. Linsey Pang and Dr. Bin Dong&lt;/strong> for their support and guidance throughout the project so far. I’m looking forward to sharing more technical results as we work toward building smarter, more adaptable recommender systems.&lt;/p></description></item><item><title>LLMSeqRec: LLM Enhanced Contextual Sequential Recommender</title><link>https://deploy-preview-1007--ucsc-ospo.netlify.app/report/osre25/sf/llmseqrec/20250614-connor/</link><pubDate>Fri, 06 Jun 2025 10:15:56 -0700</pubDate><guid>https://deploy-preview-1007--ucsc-ospo.netlify.app/report/osre25/sf/llmseqrec/20250614-connor/</guid><description>&lt;h3 id="project-description">Project Description&lt;/h3>
&lt;p>Sequential Recommender Systems are widely used in scientific and business applications to analyze and predict patterns over time. In biology and ecology, they help track species behavior by suggesting related research on migration patterns and environmental changes. Medical applications include personalized treatment recommendations based on patient history and predicting disease progression. In physics and engineering, these systems optimize experimental setups by suggesting relevant past experiments or simulations. Environmental and climate science applications include forecasting climate trends and recommending datasets for monitoring deforestation or pollution. In business and e-commerce, sequential recommenders enhance user experiences by predicting consumer behavior, suggesting personalized products, and optimizing marketing strategies based on browsing and purchase history. By leveraging sequential dependencies, these recommender systems enhance research efficiency, knowledge discovery, and business decision-making across various domains. Traditional sequential recommendation systems rely on historical user interactions to predict future preferences, but they often struggle with capturing complex contextual dependencies and adapting to dynamic user behaviors. Existing models primarily use predefined embeddings and handcrafted features, limiting their ability to generalize across diverse recommendation scenarios. To address these challenges, we propose LLM Enhanced Contextual Sequential Recommender (LLMSeqRec), which leverages Large Language Models (LLMs) to enrich sequential recommendations with deep contextual understanding and adaptive reasoning.
By integrating LLM-generated embeddings and contextual representations, LLMSeqRec enhances user intent modeling, cold-start recommendations, and long-range dependencies in sequential data. Unlike traditional models that rely solely on structured interaction logs, LLMSeqRec dynamically interprets and augments sequences with semantic context, leading to more accurate and personalized recommendations. This fusion of LLM intelligence with sequential modeling enables a more scalable, adaptable, and explainable recommender system, bridging the gap between traditional sequence-based approaches and advanced AI-driven recommendations.&lt;/p>
&lt;h3 id="project-objectives">Project Objectives&lt;/h3>
&lt;p>Aligned with the vision of the 2025 Open Source Research Experience (OSRE), this project aims to develop an LLM-Enhanced Contextual Sequential Recommender (LLMSeqRec) to improve sequential recommendation accuracy across various scientific and business applications. Sequential recommender systems are widely used to analyze and predict patterns over time, assisting in fields such as biology, ecology, medicine, physics, engineering, environmental science, and e-commerce. However, traditional models often struggle with capturing complex contextual dependencies and adapting to dynamic user behaviors, as they primarily rely on vanilla sequential Id orders.
To address these limitations, this project will leverage Large Language Models (LLMs) to enhance context-aware sequential recommendations by dynamically integrating LLM-generated embeddings and contextual representations. The core challenge lies in designing LLMSeqRec, a unified and scalable model capable of enriching user intent modeling, mitigating cold-start issues, and capturing long-range dependencies within sequential data. Unlike conventional systems that rely solely on structured interaction logs, LLMSeqRec will interpret and augment sequences with semantic context, resulting in more accurate, adaptable, and explainable recommendations. Below is an outline of the methodologies and models that will be developed in this project:&lt;/p>
&lt;ul>
&lt;li>
&lt;p>&lt;strong>Step 1: Data Preprocessing &amp;amp; Feature Creation&lt;/strong>:
Develop a data processing pipeline to parse user’s sequential interaction behaviors into sequential data points for LLM-based embeddings and contextual sequential transformer modeling; Extract user behavior sequences, items’ metadata, and temporal patterns to create context-aware sequential representations for training, validation and testing; The data source can be from Amazon open public data or Movie Lense data set. The data points creation can follow SASRec (in the reference 1).&lt;/p>
&lt;/li>
&lt;li>
&lt;p>&lt;strong>Step 2: Model Development&lt;/strong>:
Design and implement LLM-enhanced sequential recommendation models, integrating pretrained language models to augment user-item interactions with semantic context; Develop an adaptive mechanism to incorporate external contextual signals, such as product descriptions, reviews into the sequential recommendation process; The baseline model can be SASRec pytorch implementation.&lt;/p>
&lt;/li>
&lt;li>
&lt;p>&lt;strong>Step 3: Evaluation&lt;/strong>: :
Benchmark LLMSeqRec against state-of-the-art sequential recommenders, evaluating on accuracy, NDCG and cold-start performance; Conduct ablation studies to analyze the impact of LLM-generated embeddings on recommendation quality; Optimize model inference speed and efficiency for real-time recommendation scenarios.&lt;/p>
&lt;/li>
&lt;/ul>
&lt;h3 id="project-deliverables">Project Deliverables&lt;/h3>
&lt;p>This project will deliver three components, software, model training, validation and performance evaluation and demo. The software which implements the above LLMSeqRec model will be hosted on the github repo as open-access repositories. The evaluation results and demo will be published along the github repo .&lt;/p>
&lt;h3 id="llmseqrec">LLMSeqRec&lt;/h3>
&lt;ul>
&lt;li>&lt;strong>Topics&lt;/strong>: LLM Enhanced Contextual Sequential Recommender&lt;/li>
&lt;li>&lt;strong>Skills&lt;/strong>: Proficiency in Python, Pytorch, Github, Self-attention, Transformer&lt;/li>
&lt;li>&lt;strong>Difficulty&lt;/strong>: Difficult&lt;/li>
&lt;li>&lt;strong>Size&lt;/strong>: Large (350 hours)&lt;/li>
&lt;li>&lt;strong>Mentor&lt;/strong>: &lt;a href="https://deploy-preview-1007--ucsc-ospo.netlify.app/author/linsey-pang/">Linsey Pang&lt;/a>, &lt;a href="https://deploy-preview-1007--ucsc-ospo.netlify.app/author/bin-dong/">Bin Dong&lt;/a>&lt;/li>
&lt;/ul>
&lt;h3 id="references">References:&lt;/h3>
&lt;ul>
&lt;li>Self-Attentive Sequential Recommendation (SASRec)&lt;/li>
&lt;li>BERT4Rec: Sequential Recommendation with Bidirectional Encoder Representations from Transformer&lt;/li>
&lt;li>Retrieval-Augmented Generation for Knowledge-Intensive NLP Tasks&lt;/li>
&lt;li>Amazon Dataset: &lt;a href="https://cseweb.ucsd.edu/~jmcauley/datasets.html#amazon_reviews" target="_blank" rel="noopener">https://cseweb.ucsd.edu/~jmcauley/datasets.html#amazon_reviews&lt;/a>&lt;/li>
&lt;li>Movie Lense Data: &lt;a href="https://grouplens.org/datasets/movielens/" target="_blank" rel="noopener">https://grouplens.org/datasets/movielens/&lt;/a>&lt;/li>
&lt;/ul>
&lt;h2 id="introduction">Introduction&lt;/h2>
&lt;p>I&amp;rsquo;m Connor, a student at NYU studying CS and Math. This summer I&amp;rsquo;ve gotten the opportunity to work on LLMSeqRec under Dr. Bin Dong and Dr. Linsey Pang.&lt;/p>
&lt;p>In today’s digital age, sequential recommender systems power everything from e-commerce suggestions to personalized content everywhere. However, traditional models fall short in capturing user intent, adapting to dynamic behavior, or tackling cold-start problems. That’s where LLMSeqRec comes in.&lt;/p>
&lt;h2 id="problem-statement">Problem Statement&lt;/h2>
&lt;p>Most sequential recommender systems rely heavily on historical user-item interactions and predefined embeddings. This approach limits their ability to understand nuanced user preferences, struggles to scale across domains, and performs poorly in scenarios like new users or sparse data. The absence of semantic and contextual modeling is a major gap in current solutions.&lt;/p>
&lt;h2 id="overview-of-project">Overview of project&lt;/h2>
&lt;p>LLMSeqRec is a novel, LLM-enhanced sequential recommender framework that bridges this gap. By leveraging large language models (LLMs), it incorporates semantic embeddings and prompt-based contextual modeling to understand both user behavior and item metadata at a deeper level. The system explores two core approaches:&lt;/p>
&lt;ul>
&lt;li>Embedding-based: LLMs generate embeddings from item attributes.&lt;/li>
&lt;li>Prompt-based: LLMs receive full transaction history in natural language format and infer recommendations.&lt;/li>
&lt;/ul>
&lt;p>These techniques are tested using well-known datasets (e.g., Amazon, MovieLens), and evaluated with ranking metrics like NDCG@10 and Hit@10. The goal: deliver more accurate, context-rich, and explainable recommendations.&lt;/p>
&lt;h2 id="next-steps">Next Steps&lt;/h2>
&lt;p>The project is currently progressing through stages including model training, embedding integration, and evaluation. Upcoming tasks include:&lt;/p>
&lt;ul>
&lt;li>Fine-tuning enhanced models&lt;/li>
&lt;li>Designing zero-/few-shot prompts&lt;/li>
&lt;li>Running comparative experiments&lt;/li>
&lt;li>Publishing findings and writing technical blogs&lt;/li>
&lt;/ul>
&lt;p>As part of the &lt;a href="https://deploy-preview-1007--ucsc-ospo.netlify.app/project/osre25/sf/LLMSeqRec">LLMSeqRec&lt;/a> my &lt;a href="https://drive.google.com/file/d/1cs9lsjacSJUbXWzTfcHIukfKFwKJjUZF/view?usp=sharing" target="_blank" rel="noopener">proposal&lt;/a> under the mentorship of Dr. Bin Dong and Dr. Linsey Pang.&lt;/p></description></item></channel></rss>