A possible architecture of a machine-learned search engine is shown in the accompanying figure.
Training data consists of queries and documents matching them together with the relevance degree of each match. It may be prepared manually by human assessors (or raters, as Google calls them), who check results for some queries and determine relevance of each result. It is not feasible to check the relevance of all documents, and so typically a technique called pooling is used — only the top few documents, retrieved by some existing ranking models are checked. This technique may introduce selection bias. Alternatively, training data may be derived automatically by analyzing clickthrough logs (i.e. search results which got clicks from users), query chains, or such search engines' features as Google's (since-replaced) SearchWiki. Clickthrough logs can be biased by the tendency of users to click on the top search results on the assumption that they are already well-ranked.
Training data is used by a learning algorithm to produce a ranking model which computes the relevance of documents for actual queries.
Typically, users expect a search query to complete in a short time (such as a few hundred milliseconds for web search), which makes it impossible to evaluate a complex ranking model on each document in the corpus, and so a two-phase scheme is used. First, a small number of potentially relevant documents are identified using simpler retrieval models which permit fast query evaluation, such as the vector space model, Boolean model, weighted AND, or BM25. This phase is called top-
k
{\displaystyle k}
document retrieval and many heuristics were proposed in the literature to accelerate it, such as using a document's static quality score and tiered indexes. In the second phase, a more accurate but computationally expensive machine-learned model is used to re-rank these documents.
Learning to rank algorithms have been applied in areas other than information retrieval:
For the convenience of MLR algorithms, query-document pairs are usually represented by numerical vectors, which are called feature vectors. Such an approach is sometimes called bag of features and is analogous to the bag of words model and vector space model used in information retrieval for representation of documents.
Selecting and designing good features is an important area in machine learning, which is called feature engineering.
There are several measures (metrics) which are commonly used to judge how well an algorithm is doing on training data and to compare the performance of different MLR algorithms. Often a learning-to-rank problem is reformulated as an optimization problem with respect to one of these metrics.
DCG and its normalized variant NDCG are usually preferred in academic research when multiple levels of relevance are used. Other metrics such as MAP, MRR and precision, are defined only for binary judgments.
Recently, there have been proposed several new evaluation metrics which claim to model user's satisfaction with search results better than the DCG metric:
Both of these metrics are based on the assumption that the user is more likely to stop looking at search results after examining a more relevant document, than after a less relevant document.
Learning to Rank approaches are often categorized using one of three approaches: pointwise (where individual documents are ranked), pairwise (where pairs of documents are ranked into a relative order), and listwise (where an entire list of documents are ordered).
In this section, without further notice,
x
{\displaystyle x}
denotes an object to be evaluated, for example, a document or an image,
f
(
x
)
{\displaystyle f(x)}
denotes a single-value hypothesis,
h
(
⋅
)
{\displaystyle h(\cdot )}
denotes a bi-variate or multi-variate function and
L
(
⋅
)
{\displaystyle L(\cdot )}
denotes the loss function.
In this case, it is assumed that each query-document pair in the training data has a numerical or ordinal score. Then the learning-to-rank problem can be approximated by a regression problem — given a single query-document pair, predict its score. Formally speaking, the pointwise approach aims at learning a function
f
(
x
)
{\displaystyle f(x)}
predicting the real-value or ordinal score of a document
x
{\displaystyle x}
using the loss function
L
(
f
;
x
j
,
y
j
)
{\displaystyle L(f;x_{j},y_{j})}
.
In this case, the learning-to-rank problem is approximated by a classification problem — learning a binary classifier
h
(
x
u
,
x
v
)
{\displaystyle h(x_{u},x_{v})}
that can tell which document is better in a given pair of documents. The classifier shall take two documents as its input and the goal is to minimize a loss function
L
(
h
;
x
u
,
x
v
,
y
u
,
v
)
{\displaystyle L(h;x_{u},x_{v},y_{u,v})}
. The loss function typically reflects the number and magnitude of inversions in the induced ranking.
In many cases, the binary classifier
h
(
x
u
,
x
v
)
{\displaystyle h(x_{u},x_{v})}
is implemented with a scoring function
f
(
x
)
{\displaystyle f(x)}
. As an example, RankNet adapts a probability model and defines
h
(
x
u
,
x
v
)
{\displaystyle h(x_{u},x_{v})}
as the estimated probability of the document
x
u
{\displaystyle x_{u}}
has higher quality than
x
v
{\displaystyle x_{v}}
:
P
u
,
v
(
f
)
=
CDF
(
f
(
x
u
)
−
f
(
x
v
)
)
,
{\displaystyle P_{u,v}(f)={\text{CDF}}(f(x_{u})-f(x_{v})),}
These algorithms try to directly optimize the value of one of the above evaluation measures, averaged over all queries in the training data. This is often difficult in practice because most evaluation measures are not continuous functions with respect to ranking model's parameters, and so continuous approximations or bounds on evaluation measures have to be used. For example the SoftRank algorithm. LambdaMART is a pairwise algorithm which has been empirically shown to approximate listwise objective functions.
A partial list of published learning-to-rank algorithms is shown below with years of first publication of each method:
Note: as most supervised learning-to-rank algorithms can be applied to pointwise, pairwise and listwise case, only those methods which are specifically designed with ranking in mind are shown above.
Conversely, the robustness of such ranking systems can be improved via adversarial defenses such as the Madry defense.
Tie-Yan Liu (2009), "Learning to Rank for Information Retrieval", Foundations and Trends in Information Retrieval, 3 (3): 225–331, doi:10.1561/1500000016, ISBN 978-1-60198-244-5. Slides from Tie-Yan Liu's talk at WWW 2009 conference are available online Archived 2017-08-08 at the Wayback Machine 978-1-60198-244-5
Mehryar Mohri, Afshin Rostamizadeh, Ameet Talwalkar (2012) Foundations of Machine Learning, The
MIT Press ISBN 9780262018258. /wiki/Mehryar_Mohri
Joachims, T. (2002), "Optimizing Search Engines using Clickthrough Data" (PDF), Proceedings of the ACM Conference on Knowledge Discovery and Data Mining, archived (PDF) from the original on 2009-12-29, retrieved 2009-11-11 http://www.cs.cornell.edu/people/tj/publications/joachims_02c.pdf
Joachims T.; Radlinski F. (2005), "Query Chains: Learning to Rank from Implicit Feedback" (PDF), Proceedings of the ACM Conference on Knowledge Discovery and Data Mining, arXiv:cs/0605035, Bibcode:2006cs........5035R, archived (PDF) from the original on 2011-07-27, retrieved 2009-12-19 http://radlinski.org/papers/Radlinski05QueryChains.pdf
B. Cambazoglu; H. Zaragoza; O. Chapelle; J. Chen; C. Liao; Z. Zheng; J. Degenhardt., "Early exit optimizations for additive machine learned ranking systems" (PDF), WSDM '10: Proceedings of the Third ACM International Conference on Web Search and Data Mining, 2010., archived from the original (PDF) on 2019-08-28, retrieved 2009-12-23 https://web.archive.org/web/20190828063315/http://olivier.chapelle.cc/pub/wsdm2010.pdf
Broder A.; Carmel D.; Herscovici M.; Soffer A.; Zien J. (2003), "Efficient query evaluation using a two-level retrieval process", Proceedings of the twelfth international conference on Information and knowledge management (PDF), pp. 426–434, doi:10.1145/956863.956944, ISBN 978-1-58113-723-1, S2CID 2432701, archived from the original (PDF) on 2009-05-21, retrieved 2009-12-15 978-1-58113-723-1
Manning C.; Raghavan P.; Schütze H. (2008), Introduction to Information Retrieval, Cambridge University Press. Section 7.1 Archived 2009-07-19 at the Wayback Machine http://nlp.stanford.edu/IR-book/html/htmledition/efficient-scoring-and-ranking-1.html
Kevin K. Duh (2009), Learning to Rank with Partially-Labeled Data (PDF), archived (PDF) from the original on 2011-07-20, retrieved 2009-12-27 http://ssli.ee.washington.edu/people/duh/thesis/uwthesis.pdf
Kevin K. Duh (2009), Learning to Rank with Partially-Labeled Data (PDF), archived (PDF) from the original on 2011-07-20, retrieved 2009-12-27 http://ssli.ee.washington.edu/people/duh/thesis/uwthesis.pdf
Yuanhua Lv, Taesup Moon, Pranam Kolari, Zhaohui Zheng, Xuanhui Wang, and Yi Chang, Learning to Model Relatedness for News Recommendation Archived 2011-08-27 at the Wayback Machine, in International Conference on World Wide Web (WWW), 2011. http://sifaka.cs.uiuc.edu/~ylv2/pub/www11-relatedness.pdf
Manning C.; Raghavan P.; Schütze H. (2008), Introduction to Information Retrieval, Cambridge University Press. Section 7.1 Archived 2009-07-19 at the Wayback Machine http://nlp.stanford.edu/IR-book/html/htmledition/efficient-scoring-and-ranking-1.html
Richardson, M.; Prakash, A.; Brill, E. (2006). "Beyond PageRank: Machine Learning for Static Ranking" (PDF). Proceedings of the 15th International World Wide Web Conference. pp. 707–715. Archived (PDF) from the original on 2009-08-15. Retrieved 2009-11-18. http://research.microsoft.com/en-us/um/people/mattri/papers/www2006/staticrank.pdf
"Archived copy". Archived from the original on 2011-01-04. Retrieved 2009-12-14.{{cite web}}: CS1 maint: archived copy as title (link) http://www.stanford.edu/class/cs276/handouts/lecture15-learning-ranking.ppt
Olivier Chapelle; Donald Metzler; Ya Zhang; Pierre Grinspan (2009), "Expected Reciprocal Rank for Graded Relevance" (PDF), CIKM, archived from the original (PDF) on 2012-02-24 https://web.archive.org/web/20120224053008/http://research.yahoo.com/files/err.pdf
Gulin A.; Karpovich P.; Raskovalov D.; Segalovich I. (2009), "Yandex at ROMIP'2009: optimization of ranking algorithms by machine learning methods" (PDF), Proceedings of ROMIP'2009: 163–168, archived (PDF) from the original on 2009-11-22, retrieved 2009-11-13 (in Russian) http://romip.ru/romip2009/15_yandex.pdf
Tie-Yan Liu (2009), "Learning to Rank for Information Retrieval", Foundations and Trends in Information Retrieval, 3 (3): 225–331, doi:10.1561/1500000016, ISBN 978-1-60198-244-5. Slides from Tie-Yan Liu's talk at WWW 2009 conference are available online Archived 2017-08-08 at the Wayback Machine 978-1-60198-244-5
Tax, Niek; Bockting, Sander; Hiemstra, Djoerd (2015), "A cross-benchmark comparison of 87 learning to rank methods" (PDF), Information Processing & Management, 51 (6): 757–772, doi:10.1016/j.ipm.2015.07.002, S2CID 22782599, archived from the original (PDF) on 2017-08-09, retrieved 2017-10-15 https://web.archive.org/web/20170809115827/http://wwwhome.cs.utwente.nl/~hiemstra/papers/ipm2015.pdf
Burges, Chris J. C.; Shaked, Tal; Renshaw, Erin; Lazier, Ari; Deeds, Matt; Hamilton, Nicole; Hullender, Greg (1 August 2005). "Learning to Rank using Gradient Descent". Archived from the original on 26 February 2021. Retrieved 31 March 2021. {{cite journal}}: Cite journal requires |journal= (help) https://www.microsoft.com/en-us/research/publication/learning-to-rank-using-gradient-descent
Taylor, M.J., Guiver, J., Robertson, S.E., & Minka, T.P. (2008). SoftRank: optimizing non-smooth rank metrics. Web Search and Data Mining.
Burges, Chris J. C. (2010-06-23). "From RankNet to LambdaRank to LambdaMART: An Overview". {{cite journal}}: Cite journal requires |journal= (help) https://www.microsoft.com/en-us/research/publication/from-ranknet-to-lambdarank-to-lambdamart-an-overview/
Fuhr, Norbert (1989), "Optimum polynomial retrieval functions based on the probability ranking principle", ACM Transactions on Information Systems, 7 (3): 183–204, doi:10.1145/65943.65944, S2CID 16632383 /wiki/Doi_(identifier)
Cooper, William S.; Gey, Frederic C.; Dabney, Daniel P. (1992), "Probabilistic retrieval based on staged logistic regression", Proceedings of the 15th annual international ACM SIGIR conference on Research and development in information retrieval - SIGIR '92, pp. 198–210, doi:10.1145/133160.133199, ISBN 978-0897915236, S2CID 125993 978-0897915236
Bartell, Brian T.; Cottrell Garrison W.; Belew, Richard K. (1994), "Automatic Combination of Multiple Ranked Retrieval Systems", Sigir '94, pp. 173–181, doi:10.1007/978-1-4471-2099-5_18, ISBN 978-0387198897, S2CID 18606472, archived from the original on 2018-06-13, retrieved 2020-10-12 978-0387198897
Friedman, Jerome H. (2001). "Greedy Function Approximation: A Gradient Boosting Machine". The Annals of Statistics. 29 (5): 1189–1232. doi:10.1214/aos/1013203451. ISSN 0090-5364. JSTOR 2699986. https://www.jstor.org/stable/2699986
Joachims, T. (2002), "Optimizing Search Engines using Clickthrough Data" (PDF), Proceedings of the ACM Conference on Knowledge Discovery and Data Mining, archived (PDF) from the original on 2009-12-29, retrieved 2009-11-11 http://www.cs.cornell.edu/people/tj/publications/joachims_02c.pdf
Cao, Yunbo; Xu, Jun; Liu, Tie-Yan; Li, Hang; Huang, Yalou; Hon, Hsiao-Wuen (2006-08-06). "Adapting ranking SVM to document retrieval". Proceedings of the 29th annual international ACM SIGIR conference on Research and development in information retrieval. SIGIR '06. New York, NY, USA: Association for Computing Machinery. pp. 186–193. doi:10.1145/1148170.1148205. ISBN 978-1-59593-369-0. 978-1-59593-369-0
Xu, Jun; Li, Hang (2007-07-23). "AdaRank: A boosting algorithm for information retrieval". Proceedings of the 30th annual international ACM SIGIR conference on Research and development in information retrieval. SIGIR '07. New York, NY, USA: Association for Computing Machinery. pp. 391–398. doi:10.1145/1277741.1277809. ISBN 978-1-59593-597-7. 978-1-59593-597-7
Qin, Tao; Zhang, Xu-Dong; Tsai, Ming-Feng; Wang, De-Sheng; Liu, Tie-Yan; Li, Hang (2008-03-01). "Query-level loss functions for information retrieval". Information Processing & Management. Evaluating Exploratory Search Systems. 44 (2): 838–855. doi:10.1016/j.ipm.2007.07.016. ISSN 0306-4573. https://linkinghub.elsevier.com/retrieve/pii/S0306457307001276
Lin, Jung Yi; Yeh, Jen-Yuan; Chao Chung Liu (July 2012). "Learning to rank for information retrieval using layered multi-population genetic programming". 2012 IEEE International Conference on Computational Intelligence and Cybernetics (CyberneticsCom). IEEE. pp. 45–49. doi:10.1109/cyberneticscom.2012.6381614. ISBN 978-1-4673-0892-2. 978-1-4673-0892-2
Pahikkala, Tapio; Tsivtsivadze, Evgeni; Airola, Antti; Järvinen, Jouni; Boberg, Jorma (2009), "An efficient algorithm for learning to rank from preference graphs", Machine Learning, 75 (1): 129–165, doi:10.1007/s10994-008-5097-z. /wiki/Doi_(identifier)
C. Burges. (2010). From RankNet to LambdaRank to LambdaMART: An Overview Archived 2017-11-10 at the Wayback Machine. https://www.microsoft.com/en-us/research/wp-content/uploads/2016/02/MSR-TR-2010-82.pdf
Xia, Fen; Liu, Tie-Yan; Wang, Jue; Zhang, Wensheng; Li, Hang (2008-07-05). "Listwise approach to learning to rank: Theory and algorithm". Proceedings of the 25th international conference on Machine learning - ICML '08. New York, NY, USA: Association for Computing Machinery. pp. 1192–1199. doi:10.1145/1390156.1390306. ISBN 978-1-60558-205-4. 978-1-60558-205-4
Xu, Jun; Liu, Tie-Yan; Lu, Min; Li, Hang; Ma, Wei-Ying (2008-07-20). "Directly optimizing evaluation measures in learning to rank". Proceedings of the 31st annual international ACM SIGIR conference on Research and development in information retrieval. SIGIR '08. New York, NY, USA: Association for Computing Machinery. pp. 107–114. doi:10.1145/1390334.1390355. ISBN 978-1-60558-164-4. 978-1-60558-164-4
Taylor, Michael; Guiver, John; Robertson, Stephen; Minka, Tom (2008-02-11). "SoftRank: Optimizing non-smooth rank metrics". Proceedings of the international conference on Web search and web data mining - WSDM '08. New York, NY, USA: Association for Computing Machinery. pp. 77–86. doi:10.1145/1341531.1341544. ISBN 978-1-59593-927-2. 978-1-59593-927-2
Rong Jin, Hamed Valizadegan, Hang Li, Ranking Refinement and Its Application for Information Retrieval Archived 2012-04-06 at the Wayback Machine, in International Conference on World Wide Web (WWW), 2008. http://www.cs.pitt.edu/~valizadegan/Publications/ranking_refinement.pdf
Massih-Reza Amini, Vinh Truong, Cyril Goutte, A Boosting Algorithm for Learning Bipartite Ranking Functions with Partially Labeled Data Archived 2010-08-02 at the Wayback Machine, International ACM SIGIR conference, 2008. The code Archived 2010-07-23 at the Wayback Machine is available for research purposes. http://www-connex.lip6.fr/~amini/Publis/SemiSupRanking_sigir08.pdf
Leonardo Rigutini, Tiziano Papini, Marco Maggini, Franco Scarselli, "SortNet: learning to rank by a neural-based sorting algorithm" Archived 2011-11-25 at the Wayback Machine, SIGIR 2008 workshop: Learning to Rank for Information Retrieval, 2008 http://research.microsoft.com/en-us/um/beijing/events/lr4ir-2008/PROCEEDINGS-LR4IR%202008.PDF
Zhu, Chenguang; Chen, Weizhu; Zhu, Zeyuan Allen; Wang, Gang; Wang, Dong; Chen, Zheng (2009-11-02). "A general magnitude-preserving boosting algorithm for search ranking". Proceedings of the 18th ACM conference on Information and knowledge management. CIKM '09. New York, NY, USA: Association for Computing Machinery. pp. 817–826. doi:10.1145/1645953.1646057. ISBN 978-1-60558-512-3. 978-1-60558-512-3
Hamed Valizadegan, Rong Jin, Ruofei Zhang, Jianchang Mao, Learning to Rank by Optimizing NDCG Measure Archived 2012-04-06 at the Wayback Machine, in Proceeding of Neural Information Processing Systems (NIPS), 2010. /wiki/Jianchang_Mao
Sculley, D. (2010-07-25). "Combined regression and ranking". Proceedings of the 16th ACM SIGKDD international conference on Knowledge discovery and data mining. KDD '10. New York, NY, USA: Association for Computing Machinery. pp. 979–988. doi:10.1145/1835804.1835928. ISBN 978-1-4503-0055-1. 978-1-4503-0055-1
Lee, Joonseok; Bengio, Samy; Kim, Seungyeon; Lebanon, Guy; Singer, Yoram (2014-04-07). "Local collaborative ranking". Proceedings of the 23rd international conference on World wide web. WWW '14. New York, NY, USA: Association for Computing Machinery. pp. 85–96. doi:10.1145/2566486.2567970. ISBN 978-1-4503-2744-2. 978-1-4503-2744-2
Ibrahim, Osman Ali Sadek; Landa-Silva, Dario (2017-04-03). "ES-Rank: Evolution strategy learning to rank approach". Proceedings of the Symposium on Applied Computing (PDF). SAC '17. New York, NY, USA: Association for Computing Machinery. pp. 944–950. doi:10.1145/3019612.3019696. ISBN 978-1-4503-4486-9. 978-1-4503-4486-9
Ai, Qingyao; Bi, Keping; Jiafeng, Guo; Croft, W. Bruce (2018), "Learning a Deep Listwise Context Model for Ranking Refinement", The 41st International ACM SIGIR Conference on Research & Development in Information Retrieval, pp. 135–144, arXiv:1804.05936, doi:10.1145/3209978.3209985, ISBN 9781450356572, S2CID 4956076 9781450356572
Davidov, Ori; Ailon, Nir; Oliveira, Ivo F. D. (2018). "A New and Flexible Approach to the Analysis of Paired Comparison Data". Journal of Machine Learning Research. 19 (60): 1–29. ISSN 1533-7928. Archived from the original on 2019-10-03. Retrieved 2019-09-17. http://jmlr.org/papers/v19/17-179.html
Pfannschmidt, Karlson; Gupta, Pritha; Hüllermeier, Eyke (2018). "Deep Architectures for Learning Context-dependent Ranking Functions". arXiv:1803.05796 [stat.ML]. /wiki/ArXiv_(identifier)
Fatih Cakir, Kun He, Xide Xia, Brian Kulis, Stan Sclaroff, Deep Metric Learning to Rank Archived 2019-05-14 at the Wayback Machine, In Proc. IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2019. http://cs-people.bu.edu/fcakir/papers/fastap_cvpr2019.pdf
Ai, Qingyao; Wang, Xuanhui; Bruch, Sebastian; Golbandi, Nadav; Bendersky, Michael; Najork, Marc (2019), "Learning Groupwise Multivariate Scoring Functions Using Deep Neural Networks", Proceedings of the 2019 ACM SIGIR International Conference on Theory of Information Retrieval, pp. 85–92, arXiv:1811.04415, doi:10.1145/3341981.3344218, ISBN 9781450368810, S2CID 199441954 9781450368810
Rolínek, Michal; Musil, Vít; Paulus, Anselm; Vlastelica, Marin; Michaelis, Claudio; Martius, Georg (2020-03-18). "Optimizing Rank-Based Metrics with Blackbox Differentiation". 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR). pp. 7617–7627. arXiv:1912.03500. doi:10.1109/CVPR42600.2020.00764. ISBN 978-1-7281-7168-5. 978-1-7281-7168-5
Vlastelica, Marin; Paulus, Anselm; Musil, Vít; Martius, Georg; Rolínek, Michal (2019-12-04). "Differentiation of Blackbox Combinatorial Solvers". arXiv:1912.02175. {{cite journal}}: Cite journal requires |journal= (help) /wiki/ArXiv_(identifier)
Liu, Weiwen; Liu, Qing; Tang, Ruiming; Chen, Junyang; He, Xiuqiang; Heng, Pheng Ann (2020-10-19). "Personalized Re-ranking with Item Relationships for E-commerce". Proceedings of the 29th ACM International Conference on Information & Knowledge Management. CIKM '20. Virtual Event, Ireland: Association for Computing Machinery. pp. 925–934. doi:10.1145/3340531.3412332. ISBN 978-1-4503-6859-9. S2CID 224281012. Archived from the original on 2021-10-17. Retrieved 2021-04-26. 978-1-4503-6859-9
Pang, Liang; Xu, Jun; Ai, Qingyao; Lan, Yanyan; Cheng, Xueqi; Wen, Jirong (2020), "SetRank", Proceedings of the 43rd International ACM SIGIR Conference on Research and Development in Information Retrieval, pp. 499–508, doi:10.1145/3397271.3401104, ISBN 9781450380164, S2CID 241534531 9781450380164
Swezey, Robin; Grover, Aditya; Charron, Bruno; Ermon, Stefano (2021-11-27). "PiRank: Scalable Learning To Rank via Differentiable Sorting". Advances in Neural Information Processing Systems. NeurIPS '21. 34. Virtual Event, Ireland. arXiv:2012.06731. /wiki/ArXiv_(identifier)
Fuhr, Norbert (1992), "Probabilistic Models in Information Retrieval", Computer Journal, 35 (3): 243–255, doi:10.1093/comjnl/35.3.243 /wiki/Doi_(identifier)
Fuhr, Norbert (1989), "Optimum polynomial retrieval functions based on the probability ranking principle", ACM Transactions on Information Systems, 7 (3): 183–204, doi:10.1145/65943.65944, S2CID 16632383 /wiki/Doi_(identifier)
Cooper, William S.; Gey, Frederic C.; Dabney, Daniel P. (1992), "Probabilistic retrieval based on staged logistic regression", Proceedings of the 15th annual international ACM SIGIR conference on Research and development in information retrieval - SIGIR '92, pp. 198–210, doi:10.1145/133160.133199, ISBN 978-0897915236, S2CID 125993 978-0897915236
Manning C.; Raghavan P.; Schütze H. (2008), Introduction to Information Retrieval, Cambridge University Press. Sections 7.4 Archived 2009-07-21 at the Wayback Machine and 15.5 Archived 2010-05-09 at the Wayback Machine http://nlp.stanford.edu/IR-book/html/htmledition/references-and-further-reading-7.html
Jan O. Pedersen. The MLR Story Archived 2011-07-13 at the Wayback Machine http://jopedersen.com/Presentations/The_MLR_Story.pdf
U.S. patent 7,197,497 https://patents.google.com/patent/US7197497
"Bing Search Blog: User Needs, Features and the Science behind Bing". Archived from the original on 2009-11-25. Retrieved 2009-11-19. https://www.bing.com/community/blogs/search/archive/2009/06/01/user-needs-features-and-the-science-behind-bing.aspx?PageIndex=4
Yandex corporate blog entry about new ranking model "Snezhinsk" Archived 2012-03-01 at the Wayback Machine (in Russian) http://webmaster.ya.ru/replies.xml?item_no=5707&ncrnd=5118
The algorithm wasn't disclosed, but a few details were made public in [1] Archived 2010-06-01 at the Wayback Machine and [2] Archived 2010-06-01 at the Wayback Machine. http://download.yandex.ru/company/experience/GDD/Zadnie_algoritmy_Karpovich.pdf
"Yandex's Internet Mathematics 2009 competition page". Archived from the original on 2015-03-17. Retrieved 2009-11-11. https://web.archive.org/web/20150317144535/http://imat2009.yandex.ru/academic/mathematic/2009/en/
"Yahoo Learning to Rank Challenge". Archived from the original on 2010-03-01. Retrieved 2010-02-26. https://web.archive.org/web/20100301011649/http://learningtorankchallenge.yahoo.com/
Rajaraman, Anand (2008-05-24). "Are Machine-Learned Models Prone to Catastrophic Errors?". Archived from the original on 2012-07-01. Retrieved 2009-11-11. /wiki/Anand_Rajaraman
Costello, Tom (2009-06-26). "Cuil Blog: So how is Bing doing?". Archived from the original on 2009-06-27. https://archive.today/20090627213358/http://www.cuil.com/info/blog/2009/06/26/so-how-is-bing-doing
"How Bloomberg Integrated Learning-to-Rank into Apache Solr | Tech at Bloomberg". Tech at Bloomberg. 2017-01-23. Archived from the original on 2017-03-01. Retrieved 2017-02-28. https://www.techatbloomberg.com/blog/bloomberg-integrated-learning-rank-apache-solr/
"Learning to Rank for Amazon OpenSearch Service - Amazon OpenSearch Service". docs.aws.amazon.com. Retrieved 2023-09-22. https://docs.aws.amazon.com/opensearch-service/latest/developerguide/learning-to-rank.html
"Elasticsearch Learning to Rank: the documentation — Elasticsearch Learning to Rank documentation". elasticsearch-learning-to-rank.readthedocs.io. Retrieved 2023-09-22. https://elasticsearch-learning-to-rank.readthedocs.io/en/latest/
Zhou, Mo; Niu, Zhenxing; Wang, Le; Zhang, Qilin; Hua, Gang (2020). "Adversarial Ranking Attack and Defense". arXiv:2002.11293v2 [cs.CV]. /wiki/ArXiv_(identifier)
Zhou, Mo; Niu, Zhenxing; Wang, Le; Zhang, Qilin; Hua, Gang (2020). "Adversarial Ranking Attack and Defense". arXiv:2002.11293v2 [cs.CV]. /wiki/ArXiv_(identifier)
Li, Jie; Ji, Rongrong; Liu, Hong; Hong, Xiaopeng; Gao, Yue; Tian, Qi (2019). "Universal Perturbation Attack Against Image Retrieval". International Conference on Computer Vision (ICCV 2019): 4899–4908. arXiv:1812.00552. Archived from the original on 2020-07-06. Retrieved 2020-07-04. https://openaccess.thecvf.com/content_ICCV_2019/html/Li_Universal_Perturbation_Attack_Against_Image_Retrieval_ICCV_2019_paper.html
Madry, Aleksander; Makelov, Aleksandar; Schmidt, Ludwig; Tsipras, Dimitris; Vladu, Adrian (2017-06-19). "Towards Deep Learning Models Resistant to Adversarial Attacks". arXiv:1706.06083v4 [stat.ML]. /wiki/ArXiv_(identifier)