Multileave Gradient Descent for Fast Online Learning to Rank
WSDM'16, San Francisco, USA. Feb 24, 2016.
Summary
This research addresses the challenge of improving online learning to rank systems by developing Multileave Gradient Descent (MGD), a method that explores multiple ranking directions simultaneously rather than the single direction approach used in traditional Dueling Bandit Gradient Descent (DBGD). The key innovation lies in using multileaved comparisons that can evaluate multiple rankers at once, enabling the algorithm to converge approximately 10 times faster than DBGD while maintaining the same final performance quality. This advancement significantly accelerates the online learning process for search engines, allowing them to adapt more quickly to user feedback and improve ranking performance with fewer update steps.
Slides
Poster
Links
Related Publications
Multileave Gradient Descent for Fast Online Learning to Rank
Anne Schuth and Harrie Oosterhuis and Shimon Whiteson and Maarten de Rijke.
In Proceedings of WSDM'16, 2016.