Multileave Gradient Descent for Fast Online Learning to Rank
ILPS Soos, Amsterdam, The Netherlands. Nov 3, 2015.
Summary
This research presents Multileave Gradient Descent (MGD), an extension of Dueling Bandit Gradient Descent for online learning to rank in search engines. The approach enables exploration of multiple ranking directions simultaneously before updating, rather than exploring a single direction at a time, using multileaved comparisons instead of expensive pairwise interleaving comparisons. Experimental validation demonstrates that MGD achieves large improvements over existing methods by reducing the number of updates needed while maintaining effectiveness in learning from user interactions.
Slides
Links
Related Publications
Multileave Gradient Descent for Fast Online Learning to Rank
Anne Schuth and Harrie Oosterhuis and Shimon Whiteson and Maarten de Rijke.
In Proceedings of WSDM'16, 2016.