pith. machine review for the scientific record. sign in

arxiv: 2508.07050 · v3 · submitted 2025-08-09 · 💻 cs.IR · cs.AI· cs.CL· cs.LG

Recognition: unknown

ReasonRank: Empowering Passage Ranking with Strong Reasoning Ability

Authors on Pith no claims yet
classification 💻 cs.IR cs.AIcs.CLcs.LG
keywords rankingtraininglistwisereasoningreasoning-intensiveabilitymanyreasonrank
0
0 comments X
read the original abstract

Large Language Model (LLM) based listwise ranking has shown superior performance in many passage ranking tasks. With the development of Large Reasoning Models (LRMs), many studies have demonstrated that step-by-step reasoning during test-time helps improve listwise ranking performance. However, due to the scarcity of reasoning-intensive training data, existing rerankers perform poorly in many complex ranking scenarios, and the ranking ability of reasoning-intensive rerankers remains largely underdeveloped. In this paper, we first propose an automated reasoning-intensive training data synthesis framework, which sources training queries and passages from diverse domains and applies DeepSeek-R1 to generate high-quality training labels. To empower the listwise reranker with strong reasoning ability, we further propose a two-stage training approach, which includes a cold-start supervised fine-tuning (SFT) stage and a reinforcement learning (RL) stage. During the RL stage, we design a novel multi-view ranking reward tailored to the multi-turn nature of listwise ranking. Extensive experiments demonstrate that our trained reasoning-intensive reranker \textbf{ReasonRank} outperforms existing baselines significantly and also achieves much lower latency than the pointwise reranker. Our codes are available at https://github.com/8421BCD/ReasonRank.

This paper has not been read by Pith yet.

discussion (0)

Sign in with ORCID, Apple, or X to comment. Anyone can read and Pith papers without signing in.

Forward citations

Cited by 4 Pith papers

Reviewed papers in the Pith corpus that reference this work. Sorted by Pith novelty score.

  1. Very Efficient Listwise Multimodal Reranking for Long Documents

    cs.IR 2026-05 unverdicted novelty 7.0

    ZipRerank delivers state-of-the-art multimodal listwise reranking accuracy for long documents at up to 10x lower latency via early interaction and single-pass scoring.

  2. A Survey of Reasoning-Intensive Retrieval: Progress and Challenges

    cs.IR 2026-04 unverdicted novelty 6.0

    A survey that categorizes RIR benchmarks by domain and modality, proposes a taxonomy for integrating reasoning into retrieval pipelines, and outlines key challenges.

  3. Learning from Emptiness: De-biasing Listwise Rerankers with Content-Agnostic Probability Calibration

    cs.AI 2026-04 unverdicted novelty 6.0

    CapCal de-biases generative listwise rerankers via content-agnostic placeholder-based bias estimation and entropy-adaptive logit rectification, yielding over 10-point NDCG gains on lightweight models across 10 benchma...

  4. Context Convergence Improves Answering Inferential Questions

    cs.CL 2026-05 unverdicted novelty 5.0

    Passages made from high-convergence sentences improve LLM performance on inferential questions compared to cosine similarity selection.