skip to main content
10.1145/3331184.3334824acmconferencesArticle/Chapter ViewAbstractPublication PagesirConference Proceedingsconference-collections
tutorial

Learning to Rank in Theory and Practice: From Gradient Boosting to Neural Networks and Unbiased Learning

Published:18 July 2019Publication History

ABSTRACT

This tutorial aims to weave together diverse strands of modern Learning to Rank (LtR) research, and present them in a unified full-day tutorial. First, we will introduce the fundamentals of LtR, and an overview of its various sub-fields. Then, we will discuss some recent advances in gradient boosting methods such as LambdaMART by focusing on their efficiency/effectiveness trade-offs and optimizations. Subsequently, we will then present TF-Ranking, a new open source TensorFlow package for neural LtR models, and how it can be used for modeling sparse textual features. Finally, we will conclude the tutorial by covering unbiased LtR -- a new research field aiming at learning from biased implicit user feedback. The tutorial will consist of three two-hour sessions, each focusing on one of the topics described above. It will provide a mix of theoretical and hands-on sessions, and should benefit both academics interested in learning more about the current state-of-the-art in LtR, as well as practitioners who want to use LtR techniques in their applications.

References

  1. Mart'in Abadi, Paul Barham, Jianmin Chen, Zhifeng Chen, Andy Davis, Jeffrey Dean, Matthieu Devin, Sanjay Ghemawat, Geoffrey Irving, Michael Isard, and others. 2016. Tensorflow: A system for large-scale machine learning. In 12th USENIX Symposium on Operating Systems Design and Implementation. 265--283. Google ScholarGoogle ScholarDigital LibraryDigital Library
  2. Aman Agarwal, Ivan Zaitsev, and Thorsten Joachims. 2018. Consistent position bias estimation without online interventions for learning-to-rank. arXiv preprint arXiv:1806.03555 (2018).Google ScholarGoogle Scholar
  3. B. Barla Cambazoglu, Hugo Zaragoza, Olivier Chapelle, Jiang Chen, Ciya Liao, Zhaohui Zheng, and Jon Degenhardt. 2010. Early exit optimizations for additive machine learned ranking systems. In 3rd ACM International Conference on Web Search and Data Mining. ACM, 411--420. Google ScholarGoogle ScholarDigital LibraryDigital Library
  4. Tianqi Chen, Mu Li, Yutian Li, Min Lin, Naiyan Wang, Minjie Wang, Tianjun Xiao, Bing Xu, Chiyuan Zhang, and Zheng Zhang. 2015. Mxnet: A flexible and efficient machine learning library for heterogeneous distributed systems. Neural Information Processing Systems, Workshop on Machine Learning Systems (2015).Google ScholarGoogle Scholar
  5. Domenico Dato, Claudio Lucchese, Franco Maria Nardini, Salvatore Orlando, Raffaele Perego, Nicola Tonellotto, and Rossano Venturini. 2016. Fast ranking with additive ensembles of oblivious and non-oblivious regression trees. ACM Transactions on Information Systems , Vol. 35, 2 (2016), Article 15. Google ScholarGoogle ScholarDigital LibraryDigital Library
  6. Rolf Jagerman, Harrie Oosterhuis, and Maarten de Rijke. 2019. To model or to intervene: A comparison of counterfactual and online learning to rank from user interactions. In 42nd International ACM SIGIR Conference on Research & Development in Information Retrieval. ACM, (to appear). Google ScholarGoogle ScholarDigital LibraryDigital Library
  7. Yangqing Jia, Evan Shelhamer, Jeff Donahue, Sergey Karayev, Jonathan Long, Ross Girshick, Sergio Guadarrama, and Trevor Darrell. 2014. Caffe: Convolutional architecture for fast feature embedding. In 22nd ACM International Conference on Multimedia. ACM, 675--678. Google ScholarGoogle ScholarDigital LibraryDigital Library
  8. Thorsten Joachims, Adith Swaminathan, and Tobias Schnabel. 2017. Unbiased learning-to-rank with biased feedback. In 10th ACM International Conference on Web Search and Data Mining. ACM, 781--789. Google ScholarGoogle ScholarDigital LibraryDigital Library
  9. Claudio Lucchese, Franco Maria Nardini, Salvatore Orlando, Raffaele Perego, Fabrizio Silvestri, and Salvatore Trani. 2018. X-CLEaVER: Learning ranking ensembles by growing and pruning trees. ACM Transactions on Intelligent Systems and Technology , Vol. 9, 6 (2018), Article 62. Google ScholarGoogle ScholarDigital LibraryDigital Library
  10. Tri Nguyen, Mir Rosenberg, Xia Song, Jianfeng Gao, Saurabh Tiwary, Rangan Majumder, and Li Deng. 2016. MS MARCO: A human generated machine reading comprehension dataset. arXiv preprint arXiv:1611.09268 (2016).Google ScholarGoogle Scholar
  11. Harrie Oosterhuis and Maarten de Rijke. 2018. Differentiable unbiased online learning to rank. 27th ACM International Conference on Information and Knowledge Management. ACM, 1293--1302. Google ScholarGoogle ScholarDigital LibraryDigital Library
  12. Rama Kumar Pasumarthi, Sebastian Bruch, Xuanhui Wang, Cheng Li, Michael Bendersky, Marc Najork, Jan Pfeifer, Nadav Golbandi, Rohan Anil, and Stephan Wolf. 2019. TF-Ranking: Scalable TensorFlow library for learning-to-rank. In 25th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining . ACM, (to appear). Google ScholarGoogle ScholarDigital LibraryDigital Library
  13. Adam Paszke, Sam Gross, Soumith Chintala, Gregory Chanan, Edward Yang, Zachary DeVito, Zeming Lin, Alban Desmaison, Luca Antiga, and Adam Lerer. 2017. Automatic differentiation in PyTorch. In Advances in Neural Information Processing Systems, AutoDiff Workshop: The Future of Gradient-Based Machine Learning Software and Techniques .Google ScholarGoogle Scholar
  14. Tao Qin and Tie-Yan Liu. 2013. Introducing LETOR 4.0 Datasets. arXiv preprint arXiv:1306.2597 (2013).Google ScholarGoogle Scholar
  15. Olga Russakovsky, Jia Deng, Hao Su, Jonathan Krause, Sanjeev Satheesh, Sean Ma, Zhiheng Huang, Andrej Karpathy, Aditya Khosla, Michael Bernstein, and others. 2015. Imagenet large scale visual recognition challenge. International Journal of Computer Vision , Vol. 115, 3 (2015), 211--252. Google ScholarGoogle ScholarDigital LibraryDigital Library
  16. Lidan Wang, Jimmy J. Lin, and Donald Metzler. 2010. Learning to efficiently rank. In 33rd International ACM SIGIR Conference on Research and Development in Information Retrieval. ACM, 138--145. Google ScholarGoogle ScholarDigital LibraryDigital Library
  17. Xuanhui Wang, Michael Bendersky, Donald Metzler, and Marc Najork. 2016. Learning to rank with selection bias in personal search. In 41st International ACM SIGIR Conference on Research & Development in Information Retrieval . ACM, 115--124. Google ScholarGoogle ScholarDigital LibraryDigital Library
  18. Xuanhui Wang, Nadav Golbandi, Michael Bendersky, Donald Metzler, and Marc Najork. 2018. Position bias estimation for unbiased learning to rank in personal search. In 11th ACM International Conference on Web Search and Data Mining. ACM, 610 --618. Google ScholarGoogle ScholarDigital LibraryDigital Library
  19. Zhixiang Xu, Olivier Chapelle, and Kilian Q Weinberger. 2012. The greedy miser: Learning under test-time budgets. In 29th International Conference on Machine Learning. 1175--1182. Google ScholarGoogle ScholarDigital LibraryDigital Library
  20. Yisong Yue and Thorsten Joachims. 2009. Interactively optimizing information retrieval systems as a dueling bandits problem. In 26th Annual International Conference on Machine Learning. ACM, 1201--1208. Google ScholarGoogle ScholarDigital LibraryDigital Library

Index Terms

  1. Learning to Rank in Theory and Practice: From Gradient Boosting to Neural Networks and Unbiased Learning

    Recommendations

    Comments

    Login options

    Check if you have access through your login credentials or your institution to get full access on this article.

    Sign in
    • Published in

      cover image ACM Conferences
      SIGIR'19: Proceedings of the 42nd International ACM SIGIR Conference on Research and Development in Information Retrieval
      July 2019
      1512 pages
      ISBN:9781450361729
      DOI:10.1145/3331184

      Copyright © 2019 Owner/Author

      Permission to make digital or hard copies of part or all of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for third-party components of this work must be honored. For all other uses, contact the Owner/Author.

      Publisher

      Association for Computing Machinery

      New York, NY, United States

      Publication History

      • Published: 18 July 2019

      Check for updates

      Qualifiers

      • tutorial

      Acceptance Rates

      SIGIR'19 Paper Acceptance Rate84of426submissions,20%Overall Acceptance Rate792of3,983submissions,20%

    PDF Format

    View or Download as a PDF file.

    PDF

    eReader

    View online with eReader.

    eReader