OSA improves LLM-based recommenders by anchoring ordinal preference levels as numeric tokens in the model's latent space to retain fine-grained strength information when fusing collaborative signals.
Title resolution pending
3 Pith papers cite this work. Polarity classification is still indexing.
fields
cs.IR 3years
2026 3representative citing papers
FAERec fuses collaborative ID embeddings with LLM semantic embeddings using adaptive gating and dual-level alignment to enhance tail-item sequential recommendations.
FLAME condenses ensemble diversity into a single network via modular ensemble simulation and guided mutual learning during training, delivering ensemble-level performance with single-network inference speed on sequential recommendation tasks.
citing papers explorer
-
Every Preference Has Its Strength: Injecting Ordinal Semantics into LLM-Based Recommenders
OSA improves LLM-based recommenders by anchoring ordinal preference levels as numeric tokens in the model's latent space to retain fine-grained strength information when fusing collaborative signals.
-
Fusion and Alignment Enhancement with Large Language Models for Tail-item Sequential Recommendation
FAERec fuses collaborative ID embeddings with LLM semantic embeddings using adaptive gating and dual-level alignment to enhance tail-item sequential recommendations.
-
FLAME: Condensing Ensemble Diversity into a Single Network for Efficient Sequential Recommendation
FLAME condenses ensemble diversity into a single network via modular ensemble simulation and guided mutual learning during training, delivering ensemble-level performance with single-network inference speed on sequential recommendation tasks.