Harnessing Pairwise Ranking Prompting Through Sample-Efficient Ranking Distillation
Harnessing Pairwise Ranking Prompting Through Sample-Efficient Ranking Distillation
While Pairwise Ranking Prompting (PRP) with Large Language Models (LLMs) is one of the most effective zero-shot document ranking methods, it has a quadratic computational complexity with respect to the number of documents to be ranked, as it requires an enumeration over all possible document pairs. Consequently, the outstanding ranking performance of PRP has remained unreachable for most real-world ranking applications. In this work, we propose to harness the effectiveness of PRP through pairwise distillation. Specifically, we distill a pointwise student ranker from pairwise teacher labels generated by PRP, resulting in an efficient student model that retains the performance of PRP with substantially lower computational costs. Furthermore, we find that the distillation process can be made sample-efficient: with only 2% of pairs, we are able to obtain the same performance as using all pairs for teacher labels. Thus, our novel approach provides a solution to harness the ranking performance of PRP without incurring high computational costs during both distillation and serving.
Junru Wu、Le Yan、Zhen Qin、Honglei Zhuang、Paul Suganthan G. C.、Tianqi Liu、Zhe Dong、Xuanhui Wang、Harrie Oosterhuis
计算技术、计算机技术
Junru Wu,Le Yan,Zhen Qin,Honglei Zhuang,Paul Suganthan G. C.,Tianqi Liu,Zhe Dong,Xuanhui Wang,Harrie Oosterhuis.Harnessing Pairwise Ranking Prompting Through Sample-Efficient Ranking Distillation[EB/OL].(2025-07-07)[2025-07-16].https://arxiv.org/abs/2507.04820.点此复制
评论