What is it about?
When you type a question into a search engine, it has to decide which documents to show first. A new family of systems, called generative retrieval, treats this like a guessing game: the model “spells out” the ID of a document, token by token, and the first guess becomes rank #1, the next guess rank #2, and so on. The snag? These models practise spelling, not ranking, so their top guesses are not always the most relevant. Our work introduces DDRO – Direct Document-Relevance Optimisation. Instead of teaching the model with a complex reward function and reinforcement learning, we give it simple “win / lose” examples: for the same question, show that document A should beat document B. We keep the model close to its original behaviour with a gentle KL-divergence constraint (think of it as a leash that allows exploration but prevents it from wandering too far).
Featured Image
Why is it important?
Sharper answers where users look first. Most people click one of the first few results; boosting top-hit accuracy by 15–35 % means more questions are answered on the first click. Methodological simplicity and reproducibility. By replacing reinforcement-learning machinery with a single KL-regularised pairwise loss, the approach can be replicated with standard sequence-to-sequence tooling and publicly available relevance labels. Resource-efficient fine-tuning – DDRO trains on one GPU with ordinary click-log or qrel data, lowering the computational barrier to high-quality generative-retrieval research. In short, DDRO delivers better early-rank relevance and lowers the cost-of-entry.
Perspectives
I enjoyed simplifying generative retrieval: swapping a bulky RL loop for a single pair-wise loss made the idea easy to test on one GPU, yet it still delivered strong early-rank gains. I hope this lighter recipe helps others experiment with GenIR models more easily.
Kidist Amde Mekonnen
University of Amsterdam
Read the Original
This page is a summary of: Lightweight and Direct Document Relevance Optimization for Generative Information Retrieval, July 2025, ACM (Association for Computing Machinery),
DOI: 10.1145/3726302.3730023.
You can read the full text:
Resources
Contributors
The following have contributed to this page







