pith. machine review for the scientific record. sign in

arxiv: 2601.19711 · v3 · submitted 2026-01-27 · 💻 cs.IR

Recognition: 2 theorem links

· Lean Theorem

Differentiable Semantic ID for Generative Recommendation

Authors on Pith no claims yet

Pith reviewed 2026-05-16 10:52 UTC · model grok-4.3

classification 💻 cs.IR
keywords semantic IDgenerative recommendationdifferentiable indexingGumbel noisecodebook collapserecommendation systems
0
0 comments X

The pith

Differentiable semantic IDs using Gumbel noise improve generative recommendation by aligning indexing with recommendation objectives.

A machine-rendered reading of the paper's core claim, the machinery that carries it, and where it could break.

The paper shows that semantic IDs in generative recommendation suffer from an objective mismatch because they are trained separately for content reconstruction. Making them differentiable allows recommendation signals to refine the IDs but risks codebook collapse due to premature deterministic assignments. DIGER adds Gumbel noise during early training to promote code exploration and uses decay strategies to shift toward stable assignments. Experiments across datasets confirm higher code utilization and better recommendation performance. This demonstrates that joint optimization of indexing and recommendation is feasible with proper exploration mechanisms.

Core claim

By injecting Gumbel noise into the code selection process and gradually decaying the uncertainty, semantic IDs can be learned differentiably from recommendation losses without suffering from codebook collapse, leading to more effective generative recommenders.

What carries the argument

Gumbel-softmax based code assignment with uncertainty decay schedules that control the level of exploration in semantic ID learning.

If this is right

  • Recommendation losses can directly update the semantic tokenizer.
  • Codebook utilization improves as more codes are explored early on.
  • Performance gains are observed consistently across multiple datasets.
  • Indexing and recommendation objectives become aligned during training.

Where Pith is reading between the lines

These are editorial extensions of the paper, not claims the author makes directly.

  • If the decay schedule is tuned per dataset, it may further stabilize training on sparse data.
  • This method suggests that exploration techniques from reinforcement learning can transfer to discrete representation learning in recommendation systems.
  • Future work could test whether similar noise injection helps in other generative tasks with discrete latents.

Load-bearing premise

That Gumbel noise plus the two uncertainty decay strategies will reliably prevent codebook collapse and allow a smooth transition to stable, recommendation-aligned codes without introducing new optimization instabilities on real datasets.

What would settle it

Running the model without Gumbel noise and checking if code utilization drops sharply while recommendation performance degrades on the same datasets.

Figures

Figures reproduced from arXiv: 2601.19711 by Alexandros Karatzoglou, Ioannis Arapakis, Joemon M. Jose, Junchen Fu, Suzan Verberne, Xuri Ge, Zhaochun Ren.

Figure 2
Figure 2. Figure 2: Comparison of DIGER and STE on Amazon Beauty [PITH_FULL_IMAGE:figures/full_fig_p002_2.png] view at source ↗
Figure 3
Figure 3. Figure 3: Comparison of STE vs. DIGER. STE: Relies on deterministic hard selection (e.g., sequence 1-3-2), where gradients only [PITH_FULL_IMAGE:figures/full_fig_p004_3.png] view at source ↗
Figure 4
Figure 4. Figure 4: Two uncertainty decay strategies for exploration– [PITH_FULL_IMAGE:figures/full_fig_p004_4.png] view at source ↗
Figure 5
Figure 5. Figure 5: Hyperparameter analysis on Beauty and Instru [PITH_FULL_IMAGE:figures/full_fig_p007_5.png] view at source ↗
Figure 6
Figure 6. Figure 6: Comparison of four semantic ID assignment methods over training epochs: (a) Incremental SID Drift measures the [PITH_FULL_IMAGE:figures/full_fig_p008_6.png] view at source ↗
Figure 7
Figure 7. Figure 7: Code usage distribution at the best checkpoint for [PITH_FULL_IMAGE:figures/full_fig_p009_7.png] view at source ↗
read the original abstract

Generative recommendation provides a novel paradigm in which each item is represented by a discrete semantic ID (SID) learned from rich content. Most existing methods treat SIDs as predefined and train recommenders under static indexing. In practice, SIDs are typically optimized only for content reconstruction rather than recommendation accuracy. This leads to an objective mismatch: the system optimizes an indexing loss to learn the SID and a recommendation loss for interaction prediction, but because the tokenizer is trained independently, the recommendation loss cannot update it. A natural approach is to make semantic indexing differentiable so that recommendation gradients can directly influence SID learning, but this often causes codebook collapse, where only a few codes are used. We attribute this issue to early deterministic assignments that limit codebook exploration, resulting in imbalance and unstable optimization. In this paper, we propose DIGER (Differentiable Semantic ID for Generative Recommendation), a first step toward effective differentiable semantic IDs for generative recommendation. DIGER introduces Gumbel noise to explicitly encourage early-stage exploration over codes, mitigating codebook collapse and improving code utilization. To balance exploration and convergence, we further design two uncertainty decay strategies that gradually reduce the Gumbel noise, enabling a smooth transition from early exploration to exploitation of learned SIDs. Extensive experiments on multiple public datasets demonstrate consistent improvements from differentiable semantic IDs. These results confirm the effectiveness of aligning indexing and recommendation objectives through differentiable SIDs and highlight differentiable semantic indexing as a promising research direction. Our code is released under https://github.com/junchen-fu/DIGER.

Editorial analysis

A structured set of objections, weighed in public.

Desk editor's note, referee report, simulated authors' rebuttal, and a circularity audit. Tearing a paper down is the easy half of reading it; the pith above is the substance, this is the friction.

Referee Report

2 major / 2 minor

Summary. The manuscript introduces DIGER, a method to learn differentiable semantic IDs (SIDs) for generative recommendation. It adds Gumbel noise during early training to encourage code exploration and thereby mitigate codebook collapse, then applies two uncertainty decay strategies to transition from exploration to stable, recommendation-aligned assignments. The central claim is that this joint optimization of indexing and recommendation objectives yields consistent performance gains over static SID baselines on public datasets.

Significance. If the empirical results hold, the work provides a practical route to close the objective mismatch between content reconstruction and interaction prediction in generative recommenders. The explicit handling of early-stage collapse via scheduled Gumbel noise is a concrete, testable intervention; the public code release further strengthens the contribution by enabling direct reproduction and extension.

major comments (2)
  1. [Abstract and §4] Abstract and §4: the claim of 'consistent improvements' and 'extensive experiments' is stated without any quantitative metrics, ablation results, or error bars in the provided text. Because the central contribution is empirical, the absence of concrete effect sizes (e.g., Recall@10 or NDCG deltas versus the strongest static-SID baseline) prevents verification that the gains are attributable to the differentiable mechanism rather than other implementation choices.
  2. [§3.2] §3.2: the two uncertainty decay strategies are described at a high level, but their exact functional forms, temperature schedules, and interaction with the Gumbel noise parameter are not specified. Without these details the claimed 'smooth transition' from exploration to exploitation cannot be reproduced or stress-tested for optimization stability on real-scale datasets.
minor comments (2)
  1. [§3] Notation for the Gumbel-softmax temperature and the decay rates should be introduced once and used consistently; currently the same symbol appears to be reused for distinct quantities.
  2. [§4] Figure captions and axis labels in the experimental section would benefit from explicit mention of the datasets and metrics plotted, to allow quick cross-reference with the tables.

Simulated Author's Rebuttal

2 responses · 0 unresolved

We thank the referee for the positive evaluation and the recommendation for minor revision. We address each major comment below with clarifications and commit to revisions that improve clarity and reproducibility while preserving the manuscript's empirical claims.

read point-by-point responses
  1. Referee: [Abstract and §4] Abstract and §4: the claim of 'consistent improvements' and 'extensive experiments' is stated without any quantitative metrics, ablation results, or error bars in the provided text. Because the central contribution is empirical, the absence of concrete effect sizes (e.g., Recall@10 or NDCG deltas versus the strongest static-SID baseline) prevents verification that the gains are attributable to the differentiable mechanism rather than other implementation choices.

    Authors: We appreciate the referee's observation. Section 4 of the full manuscript reports the complete experimental results on public datasets, including Recall@10 and NDCG@10 values, direct comparisons against static SID baselines, ablation studies isolating the differentiable component, and error bars from repeated runs. To make these empirical gains immediately verifiable from the abstract, we will revise the abstract to include specific quantitative deltas (e.g., average relative improvements over the strongest baseline) while retaining the existing Section 4 details. revision: yes

  2. Referee: [§3.2] §3.2: the two uncertainty decay strategies are described at a high level, but their exact functional forms, temperature schedules, and interaction with the Gumbel noise parameter are not specified. Without these details the claimed 'smooth transition' from exploration to exploitation cannot be reproduced or stress-tested for optimization stability on real-scale datasets.

    Authors: We agree that greater specificity in §3.2 is needed for full reproducibility. The manuscript outlines the linear and exponential uncertainty decay strategies and their interaction with Gumbel noise, but we will expand the section with the precise equations (e.g., explicit temperature schedules such as τ(t) = τ_init · decay_rate^t with bounds), the exact modulation of the Gumbel parameter, and pseudocode for the training loop. This will allow direct verification of the exploration-to-exploitation transition without changing the underlying method. revision: yes

Circularity Check

0 steps flagged

No significant circularity in derivation chain

full rationale

The paper proposes DIGER as an empirical intervention: Gumbel noise plus two uncertainty decay schedules are introduced to encourage early exploration and mitigate codebook collapse in differentiable semantic ID learning. The abstract frames the contribution as a practical fix whose value is measured on held-out recommendation metrics across public datasets. No equations, uniqueness theorems, or self-citations are invoked that reduce a claimed prediction or result to a fitted input by construction. The method is presented as an architectural choice with external experimental support rather than a self-referential derivation. This matches the default expectation of a non-circular empirical paper.

Axiom & Free-Parameter Ledger

2 free parameters · 1 axioms · 0 invented entities

The central claim rests on the standard Gumbel-softmax reparameterization trick (treated as known) plus two hand-designed decay schedules whose functional forms and hyperparameters are not detailed in the abstract; no new physical entities are postulated.

free parameters (2)
  • Gumbel noise temperature schedule
    Controls the amount of exploration early in training; its exact functional form and initial value are not specified in the abstract and must be tuned.
  • Uncertainty decay rates
    Two separate decay strategies whose parameters determine the transition from exploration to exploitation; these are introduced by the paper and require empirical selection.
axioms (1)
  • domain assumption Gumbel noise added to logits prevents premature deterministic code assignments and thereby mitigates codebook collapse
    Invoked in the description of the early-stage exploration mechanism; treated as a known property of the Gumbel-softmax trick.

pith-pipeline@v0.9.0 · 5599 in / 1383 out tokens · 39796 ms · 2026-05-16T10:52:27.760278+00:00 · methodology

discussion (0)

Sign in with ORCID, Apple, or X to comment. Anyone can read and Pith papers without signing in.

Lean theorems connected to this paper

Citations machine-checked in the Pith Canon. Every link opens the source theorem in the public Lean library.

What do these tags mean?
matches
The paper's claim is directly supported by a theorem in the formal canon.
supports
The theorem supports part of the paper's argument, but the paper may add assumptions or extra steps.
extends
The paper goes beyond the formal theorem; the theorem is a base layer rather than the whole result.
uses
The paper appears to rely on the theorem as machinery.
contradicts
The paper's claim conflicts with a theorem or certificate in the canon.
unclear
Pith found a possible connection, but the passage is too broad, indirect, or ambiguous to say the theorem truly supports the claim.

Forward citations

Cited by 2 Pith papers

Reviewed papers in the Pith corpus that reference this work. Sorted by Pith novelty score.

  1. MLPs are Efficient Distilled Generative Recommenders

    cs.IR 2026-05 unverdicted novelty 7.0

    SID-MLP distills autoregressive generative recommenders into efficient position-specific MLP heads for Semantic ID tasks, achieving 8.74x faster inference with matching accuracy.

  2. CapsID: Soft-Routed Variable-Length Semantic IDs for Generative Recommendation

    cs.IR 2026-05 unverdicted novelty 6.0

    CapsID uses probabilistic capsule routing and confidence-based termination to generate variable-length semantic IDs, improving recall by 9.6% over strong baselines with half the latency of dual-representation systems.

Reference graph

Works this paper leans on

58 extracted references · 58 canonical work pages · cited by 2 Pith papers · 3 internal anchors

  1. [1]

    Keqin Bao, Jizhi Zhang, Wenjie Wang, Yang Zhang, Zhengyi Yang, Yanchen Luo, Chong Chen, Fuli Feng, and Qi Tian. 2025. A bi-step grounding paradigm for large language models in recommendation systems.ACM Transactions on Recommender Systems3, 4 (2025), 1–27

  2. [2]

    Yoshua Bengio, Nicholas Léonard, and Aaron Courville. 2013. Estimating or propagating gradients through stochastic neurons for conditional computation. arXiv preprint arXiv:1308.3432(2013)

  3. [3]

    Shereen Elsayed, Lukas Brinkmeyer, and Lars Schmidt-Thieme. 2022. End-to-end image-based fashion recommendation. InWorkshop on Recommender Systems in Fashion and Retail. Springer, 109–119

  4. [4]

    Junchen Fu, Xuri Ge, Xin Xin, Alexandros Karatzoglou, Ioannis Arapakis, Jie Wang, and Joemon M Jose. 2024. IISAN: Efficiently adapting multimodal repre- sentation for sequential recommendation with decoupled PEFT. InProceedings of the 47th International ACM SIGIR Conference on Research and Development in Information Retrieval. 687–697

  5. [5]

    Junchen Fu, Xuri Ge, Xin Xin, Alexandros Karatzoglou, Ioannis Arapakis, Kaiwen Zheng, Yongxin Ni, and Joemon M Jose Joemon. 2025. Efficient and effective adaptation of multimodal foundation models in sequential recommendation.IEEE Transactions on Knowledge and Data Engineering(2025)

  6. [6]

    Junchen Fu, Fajie Yuan, Yu Song, Zheng Yuan, Mingyue Cheng, Shenghui Cheng, Jiaqi Zhang, Jie Wang, and Yunzhu Pan. 2024. Exploring adapter-based transfer learning for recommender systems: Empirical studies and practical insights. In Proceedings of the 17th ACM international conference on web search and data mining. 208–217

  7. [7]

    Everette S Gardner Jr. 2006. Exponential smoothing: The state of the art—Part II. International journal of forecasting22, 4 (2006), 637–666

  8. [8]

    Shijie Geng, Shuchang Liu, Zuohui Fu, Yingqiang Ge, and Yongfeng Zhang. 2022. Recommendation as language processing (rlp): A unified pretrain, personalized prompt & predict paradigm (p5). InProceedings of the 16th ACM conference on recommender systems. 299–315

  9. [9]

    Alex Graves. 2011. Practical variational inference for neural networks.Advances in neural information processing systems24 (2011)

  10. [10]

    Anil K Gupta, Ken G Smith, and Christina E Shalley. 2006. The interplay between exploration and exploitation.Academy of management journal49, 4 (2006), 693–706

  11. [11]

    Ruidong Han, Bin Yin, Shangyu Chen, He Jiang, Fei Jiang, Xiang Li, Chi Ma, Mincong Huang, Xiaoguang Li, Chunzhen Jing, et al . 2025. Mtgr: Industrial- scale generative recommendation framework in meituan. InProceedings of the 34th ACM International Conference on Information and Knowledge Management. 5731–5738

  12. [12]

    Xiangnan He, Kuan Deng, Xiang Wang, Yan Li, Yongdong Zhang, and Meng Wang. 2020. Lightgcn: Simplifying and powering graph convolution network for recommendation. InProceedings of the 43rd International ACM SIGIR conference on research and development in Information Retrieval. 639–648

  13. [13]

    Xiangnan He, Lizi Liao, Hanwang Zhang, Liqiang Nie, Xia Hu, and Tat-Seng Chua. 2017. Neural collaborative filtering. InProceedings of the 26th international conference on world wide web. 173–182

  14. [14]

    Min Hou, Le Wu, Yuxin Liao, Yonghui Yang, Zhen Zhang, Changlong Zheng, Han Wu, and Richang Hong. 2025. A survey on generative recommendation: Data, model, and tasks.arXiv preprint arXiv:2510.27157(2025)

  15. [15]

    Yupeng Hou, Jiacheng Li, Ashley Shin, Jinsung Jeon, Abhishek Santhanam, Wei Shao, Kaveh Hassani, Ning Yao, and Julian McAuley. 2025. Generating long semantic ids in parallel for recommendation. InProceedings of the 31st ACM SIGKDD Conference on Knowledge Discovery and Data Mining V. 2. 956–966

  16. [16]

    Yupeng Hou, An Zhang, Leheng Sheng, Zhengyi Yang, Xiang Wang, Tat-Seng Chua, and Julian McAuley. 2025. Generative Recommendation Models: Progress and Directions. InCompanion Proceedings of the ACM on Web Conference 2025. 13–16

  17. [17]

    Wenyue Hua, Shuyuan Xu, Yingqiang Ge, and Yongfeng Zhang. 2023. How to index item ids for recommendation foundation models. InProceedings of the Annual International ACM SIGIR Conference on Research and Development in Information Retrieval in the Asia Pacific Region. 195–204

  18. [18]

    Minyoung Huh, Brian Cheung, Pulkit Agrawal, and Phillip Isola. 2023. Straight- ening out the straight-through estimator: Overcoming optimization challenges in vector quantized networks. InInternational Conference on Machine Learning. PMLR, 14096–14113

  19. [19]

    Eric Jang, Shixiang Gu, and Ben Poole. 2016. Categorical reparameterization with gumbel-softmax.arXiv preprint arXiv:1611.01144(2016)

  20. [20]

    Wang-Cheng Kang and Julian McAuley. 2018. Self-attentive sequential recom- mendation. In2018 IEEE international conference on data mining (ICDM). IEEE, 197–206

  21. [21]

    Yehuda Koren, Robert Bell, and Chris Volinsky. 2009. Matrix factorization tech- niques for recommender systems.Computer42, 8 (2009), 30–37

  22. [22]

    Doyup Lee, Chiheon Kim, Saehoon Kim, Minsu Cho, and Wook-Shin Han. 2022. Autoregressive image generation using residual quantization. InProceedings of the IEEE/CVF conference on computer vision and pattern recognition. 11523–11532

  23. [23]

    Ruyu Li, Wenhao Deng, Yu Cheng, Zheng Yuan, Jiaqi Zhang, and Fajie Yuan

  24. [24]

    InProceedings of the 34th ACM International Conference on Information and Knowledge Management(Seoul, Republic of Korea)(CIKM ’25)

    Exploring the Upper Limits of Text-Based Collaborative Filtering Using Large Language Models: Discoveries and Insights. InProceedings of the 34th ACM International Conference on Information and Knowledge Management(Seoul, Republic of Korea)(CIKM ’25). Association for Computing Machinery, New York, NY, USA, 1643–1653. doi:10.1145/3746252.3761429

  25. [25]

    Wuchao Li, Rui Huang, Haijun Zhao, Chi Liu, Kai Zheng, Qi Liu, Na Mou, Guorui Zhou, Defu Lian, Yang Song, et al . 2025. DimeRec: a unified framework for enhanced sequential recommendation via generative diffusion models. InPro- ceedings of the Eighteenth ACM International Conference on Web Search and Data Mining. 726–734

  26. [26]

    Xiaopeng Li, Bo Chen, Junda She, Shiteng Cao, You Wang, Qinlin Jia, Haiying He, Zheli Zhou, Zhao Liu, Ji Liu, et al. 2025. A Survey of Generative Recommendation from a Tri-Decoupled Perspective: Tokenization, Architecture, and Optimization. (2025)

  27. [27]

    Dawen Liang, Rahul G Krishnan, Matthew D Hoffman, and Tony Jebara. 2018. Variational autoencoders for collaborative filtering. InProceedings of the 2018 world wide web conference. 689–698

  28. [28]

    Xinyu Lin, Haihan Shi, Wenjie Wang, Fuli Feng, Qifan Wang, See-Kiong Ng, and Tat-Seng Chua. 2025. Order-agnostic identifier for large language model-based generative recommendation. InProceedings of the 48th international ACM SIGIR conference on research and development in information retrieval. 1923–1933

  29. [29]

    Chang Liu, Yimeng Bai, Xiaoyan Zhao, Yang Zhang, Fuli Feng, and Wenge Rong

  30. [30]

    DiscRec: Disentangled Semantic-Collaborative Modeling for Generative Recommendation.arXiv preprint arXiv:2506.15576(2025)

  31. [31]

    Enze Liu, Bowen Zheng, Cheng Ling, Lantao Hu, Han Li, and Wayne Xin Zhao

  32. [32]

    arXiv preprint arXiv:2409.05546(2024)

    Generative Recommender with End-to-End Learnable Item Tokenization. arXiv preprint arXiv:2409.05546(2024)

  33. [33]

    Qijiong Liu, Nuo Chen, Tetsuya Sakai, and Xiao-Ming Wu. 2024. Once: Boosting content-based recommendation with both open-and closed-source large language models. InProceedings of the 17th ACM International Conference on Web Search and Data Mining. 452–461

  34. [34]

    Zhanyu Liu, Shiyao Wang, Xingmei Wang, Rongzhou Zhang, Jiaxin Deng, Honghui Bao, Jinghao Zhang, Wuchao Li, Pengfei Zheng, Xiangyu Wu, et al

  35. [35]

    Onerec-think: In-text reasoning for generative recommendation.arXiv preprint arXiv:2510.11639(2025)

  36. [36]

    Zhiyun Lu, Eugene Ie, and Fei Sha. 2020. Mean-field approximation to Gaussian- softmax integral with application to uncertainty estimation.arXiv preprint arXiv:2006.07584(2020)

  37. [37]

    Chen Ma, Peng Kang, and Xue Liu. 2019. Hierarchical gating networks for sequential recommendation. InProceedings of the 25th ACM SIGKDD international conference on knowledge discovery & data mining. 825–833. Differentiable Semantic ID for Generative Recommendation Conference acronym ’XX, June 03–05, 2018, Woodstock, NY

  38. [38]

    Jianmo Ni, Jiacheng Li, and Julian McAuley. 2019. Justifying recommendations using distantly-labeled reviews and fine-grained aspects. InProceedings of the 2019 conference on empirical methods in natural language processing and the 9th international joint conference on natural language processing (EMNLP-IJCNLP). 188–197

  39. [39]

    Shashank Rajput, Nikhil Mehta, Anima Singh, Raghunandan Hulikal Keshavan, Trung Vu, Lukasz Heldt, Lichan Hong, Yi Tay, Vinh Tran, Jonah Samost, et al

  40. [40]

    Recommender systems with generative retrieval.Advances in Neural Information Processing Systems36 (2023), 10299–10315

  41. [41]

    Fei Sun, Jun Liu, Jian Wu, Changhua Pei, Xiao Lin, Wenwu Ou, and Peng Jiang

  42. [42]

    InProceedings of the 28th ACM international conference on information and knowledge management

    BERT4Rec: Sequential recommendation with bidirectional encoder rep- resentations from transformer. InProceedings of the 28th ACM international conference on information and knowledge management. 1441–1450

  43. [43]

    Yuhta Takida, Takashi Shibuya, WeiHsiang Liao, Chieh-Hsin Lai, Junki Ohmura, Toshimitsu Uesaka, Naoki Murata, Shusuke Takahashi, Toshiyuki Kumakura, and Yuki Mitsufuji. 2022. Sq-vae: Variational bayes on discrete representation with self-annealed stochastic quantization.arXiv preprint arXiv:2205.07547(2022)

  44. [44]

    Jiaxi Tang and Ke Wang. 2018. Personalized top-n sequential recommenda- tion via convolutional sequence embedding. InProceedings of the eleventh ACM international conference on web search and data mining. 565–573

  45. [45]

    Yi Tay, Vinh Tran, Mostafa Dehghani, Jianmo Ni, Dara Bahri, Harsh Mehta, Zhen Qin, Kai Hui, Zhe Zhao, Jai Gupta, et al. 2022. Transformer memory as a differentiable search index.Advances in Neural Information Processing Systems 35 (2022), 21831–21843

  46. [46]

    Dongsheng Wang, Yuxi Huang, Shen Gao, Yifan Wang, Chengrui Huang, and Shuo Shang. 2025. Generative next poi recommendation with semantic id. In Proceedings of the 31st ACM SIGKDD Conference on Knowledge Discovery and Data Mining V. 2. 2904–2914

  47. [47]

    Hao Wang, Wei Guo, Luankang Zhang, Jin Yao Chin, Yufei Ye, Huifeng Guo, Yong Liu, Defu Lian, Ruiming Tang, and Enhong Chen. 2025. Generative large recom- mendation models: Emerging trends in llms for recommendation. InCompanion Proceedings of the ACM on Web Conference 2025. 49–52

  48. [48]

    Wenjie Wang, Honghui Bao, Xinyu Lin, Jizhi Zhang, Yongqi Li, Fuli Feng, See- Kiong Ng, and Tat-Seng Chua. 2024. Learnable item tokenization for generative recommendation. InProceedings of the 33rd ACM International Conference on Information and Knowledge Management. 2400–2409

  49. [49]

    Yuhao Wang, Junwei Pan, Xinhang Li, Maolin Wang, Yuan Wang, Yue Liu, Dapeng Liu, Jie Jiang, and Xiangyu Zhao. 2025. Empowering large language model for sequential recommendation via multimodal embeddings and semantic ids. InProceedings of the 34th ACM International Conference on Information and Knowledge Management. 3209–3219

  50. [50]

    Chuhan Wu, Fangzhao Wu, Tao Qi, and Yongfeng Huang. 2021. Empowering news recommendation with pre-trained language models. InProceedings of the 44th international ACM SIGIR conference on research and development in informa- tion retrieval. 1652–1656

  51. [51]

    Yuyang Ye, Zhi Zheng, Yishan Shen, Tianshu Wang, Hengruo Zhang, Peijun Zhu, Runlong Yu, Kai Zhang, and Hui Xiong. 2025. Harnessing multimodal large language models for multimodal sequential recommendation. InProceedings of the AAAI Conference on Artificial Intelligence, Vol. 39. 13069–13077

  52. [52]

    Zheng Yuan, Fajie Yuan, Yu Song, Youhua Li, Junchen Fu, Fei Yang, Yunzhu Pan, and Yongxin Ni. 2023. Where to go next for recommender systems? id- vs. modality-based recommender models revisited. InProceedings of the 46th International ACM SIGIR Conference on Research and Development in Information Retrieval. 2639–2649

  53. [53]

    Penglong Zhai, Yifang Yuan, Fanyi Di, Jie Li, Yue Liu, Chen Li, Jie Huang, Sicong Wang, Yao Xu, and Xin Li. 2025. A Simple Contrastive Framework Of Item Tokenization For Generative Recommendation.arXiv preprint arXiv:2506.16683 (2025)

  54. [54]

    Luankang Zhang, Kenan Song, Yi Quan Lee, Wei Guo, Hao Wang, Yawen Li, Huifeng Guo, Yong Liu, Defu Lian, and Enhong Chen. 2025. Killing two birds with one stone: Unifying retrieval and ranking with a single generative recom- mendation model. InProceedings of the 48th International ACM SIGIR Conference on Research and Development in Information Retrieval. 2224–2234

  55. [55]

    Yingchen Zhang, Ruqing Zhang, Jiafeng Guo, Wenjun Peng, Sen Li, Fuyu Lv, and Xueqi Cheng. 2025. C2T-ID: Converting Semantic Codebooks to Textual Document Identifiers for Generative Search. InProceedings of the 2025 Annual International ACM SIGIR Conference on Research and Development in Information Retrieval in the Asia Pacific Region. 331–336

  56. [56]

    Jujia Zhao, Wenjie Wang, Chen Xu, Xiuying Chen, Zhaochun Ren, and Suzan Verberne. 2025. Unifying Search and Recommendation: A Generative Paradigm Inspired by Information Theory.arXiv preprint arXiv:2504.06714(2025)

  57. [57]

    Jujia Zhao, Zihan Wang, Shuaiqun Pan, Suzan Verberne, and Zhaochun Ren. 2026. Unifying Search and Recommendation in LLMs via Gradient Multi-Subspace Tuning.arXiv preprint arXiv:2601.09496(2026)

  58. [58]

    Bowen Zheng, Enze Liu, Zhongfu Chen, Zhongrui Ma, Yue Wang, Wayne Xin Zhao, and Ji-Rong Wen. 2025. Pre-training Generative Recommender with Multi- Identifier Item Tokenization.arXiv preprint arXiv:2504.04400(2025)