词条 | Learning to rank | |||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||
释义 |
Learning to rank[1] or machine-learned ranking (MLR) is the application of machine learning, typically supervised, semi-supervised or reinforcement learning, in the construction of ranking models for information retrieval systems.[2] Training data consists of lists of items with some partial order specified between items in each list. This order is typically induced by giving a numerical or ordinal score or a binary judgment (e.g. "relevant" or "not relevant") for each item. The ranking model's purpose is to rank, i.e. produce a permutation of items in new, unseen lists in a way which is "similar" to rankings in the training data in some sense. ApplicationsIn information retrievalRanking is a central part of many information retrieval problems, such as document retrieval, collaborative filtering, sentiment analysis, and online advertising. A possible architecture of a machine-learned search engine is shown in the accompanying figure. Training data consists of queries and documents matching them together with relevance degree of each match. It may be prepared manually by human assessors (or raters, as Google calls them), who check results for some queries and determine relevance of each result. It is not feasible to check the relevance of all documents, and so typically a technique called pooling is used — only the top few documents, retrieved by some existing ranking models are checked. Alternatively, training data may be derived automatically by analyzing clickthrough logs (i.e. search results which got clicks from users),[3] query chains,[4] or such search engines' features as Google's SearchWiki. Training data is used by a learning algorithm to produce a ranking model which computes the relevance of documents for actual queries. Typically, users expect a search query to complete in a short time (such as a few hundred milliseconds for web search), which makes it impossible to evaluate a complex ranking model on each document in the corpus, and so a two-phase scheme is used.[5] First, a small number of potentially relevant documents are identified using simpler retrieval models which permit fast query evaluation, such as the vector space model, boolean model, weighted AND,[6] or BM25. This phase is called top- document retrieval and many heuristics were proposed in the literature to accelerate it, such as using a document's static quality score and tiered indexes.[7] In the second phase, a more accurate but computationally expensive machine-learned model is used to re-rank these documents. In other areasLearning to rank algorithms have been applied in areas other than information retrieval:
Feature vectorsFor the convenience of MLR algorithms, query-document pairs are usually represented by numerical vectors, which are called feature vectors. Such an approach is sometimes called bag of features and is analogous to the bag of words model and vector space model used in information retrieval for representation of documents. Components of such vectors are called features, factors or ranking signals. They may be divided into three groups (features from document retrieval are shown as examples):
Some examples of features, which were used in the well-known LETOR dataset:[12]
Selecting and designing good features is an important area in machine learning, which is called feature engineering. Evaluation measuresThere are several measures (metrics) which are commonly used to judge how well an algorithm is doing on training data and to compare the performance of different MLR algorithms. Often a learning-to-rank problem is reformulated as an optimization problem with respect to one of these metrics. Examples of ranking quality measures:
DCG and its normalized variant NDCG are usually preferred in academic research when multiple levels of relevance are used.[13] Other metrics such as MAP, MRR and precision, are defined only for binary judgments. Recently, there have been proposed several new evaluation metrics which claim to model user's satisfaction with search results better than the DCG metric:
Both of these metrics are based on the assumption that the user is more likely to stop looking at search results after examining a more relevant document, than after a less relevant document. Approaches{{Expand section|date=December 2009}}Tie-Yan Liu of Microsoft Research Asia has analyzed existing algorithms for learning to rank problems in his paper "Learning to Rank for Information Retrieval".[1] He categorized them into three groups by their input representation and loss function: the pointwise, pairwise, and listwise approach. In practice, listwise approaches often outperform pairwise approaches and pointwise approaches. This statement was further supported by a large scale experiment on the performance of different learning-to-rank methods on a large collection of benchmark data sets.[16] Pointwise approachIn this case, it is assumed that each query-document pair in the training data has a numerical or ordinal score. Then the learning-to-rank problem can be approximated by a regression problem — given a single query-document pair, predict its score. A number of existing supervised machine learning algorithms can be readily used for this purpose. Ordinal regression and classification algorithms can also be used in pointwise approach when they are used to predict the score of a single query-document pair, and it takes a small, finite number of values. Pairwise approachIn this case, the learning-to-rank problem is approximated by a classification problem — learning a binary classifier that can tell which document is better in a given pair of documents. The goal is to minimize the average number of inversions in ranking. Listwise approachThese algorithms try to directly optimize the value of one of the above evaluation measures, averaged over all queries in the training data. This is difficult because most evaluation measures are not continuous functions with respect to ranking model's parameters, and so continuous approximations or bounds on evaluation measures have to be used. List of methodsA partial list of published learning-to-rank algorithms is shown below with years of first publication of each method:
Note: as most supervised learning algorithms can be applied to pointwise case, only those methods which are specifically designed with ranking in mind are shown above. HistoryNorbert Fuhr introduced the general idea of MLR in 1992, describing learning approaches in information retrieval as a generalization of parameter estimation;[27] a specific variant of this approach (using polynomial regression) had been published by him three years earlier.[17] Bill Cooper proposed logistic regression for the same purpose in 1992 [18] and used it with his Berkeley research group to train a successful ranking function for TREC. Manning et al.[28] suggest that these early works achieved limited results in their time due to little available training data and poor machine learning techniques. Several conferences, such as NIPS, SIGIR and ICML had workshops devoted to the learning-to-rank problem since mid-2000s (decade). Practical usage by search enginesCommercial web search engines began using machine learned ranking systems since the 2000s (decade). One of the first search engines to start using it was AltaVista (later its technology was acquired by Overture, and then Yahoo), which launched a gradient boosting-trained ranking function in April 2003.[29][30] Bing's search is said to be powered by [https://www.microsoft.com/en-us/research/wp-content/uploads/2005/08/icml_ranking.pdf RankNet] algorithm,[31]{{when|date=February 2014}} which was invented at Microsoft Research in 2005. In November 2009 a Russian search engine Yandex announced[32] that it had significantly increased its search quality due to deployment of a new proprietary MatrixNet algorithm, a variant of gradient boosting method which uses oblivious decision trees.[33] Recently they have also sponsored a machine-learned ranking competition "Internet Mathematics 2009"[34] based on their own search engine's production data. Yahoo has announced a similar competition in 2010.[35] As of 2008, Google's Peter Norvig denied that their search engine exclusively relies on machine-learned ranking.[36] Cuil's CEO, Tom Costello, suggests that they prefer hand-built models because they can outperform machine-learned models when measured against metrics like click-through rate or time on landing page, which is because machine-learned models "learn what people say they like, not what people actually like".[37] In January 2017 the technology was included in the open source search engine Apache Solr™,[38] thus making machine learned search rank widely accessible also for enterprise search. References1. ^1 {{citation|author=Tie-Yan Liu|title=Learning to Rank for Information Retrieval|series=Foundations and Trends in Information Retrieval|year=2009|isbn=978-1-60198-244-5|doi=10.1561/1500000016|pages=225–331|volume=3|issue=3}}. Slides from Tie-Yan Liu's talk at WWW 2009 conference are available online 2. ^Mehryar Mohri, Afshin Rostamizadeh, Ameet Talwalkar (2012) Foundations of Machine Learning, TheMIT Press {{ISBN|9780262018258}}. 3. ^1 {{citation | author=Joachims, T. | journal=Proceedings of the ACM Conference on Knowledge Discovery and Data Mining | url=http://www.cs.cornell.edu/people/tj/publications/joachims_02c.pdf | title=Optimizing Search Engines using Clickthrough Data | year=2002}} 4. ^{{citation |author1=Joachims T. |author2=Radlinski F. | title=Query Chains: Learning to Rank from Implicit Feedback | url=http://radlinski.org/papers/Radlinski05QueryChains.pdf | year=2005 | journal=Proceedings of the ACM Conference on Knowledge Discovery and Data Mining}} 5. ^{{citation |author1=B. Cambazoglu |author2=H. Zaragoza |author3=O. Chapelle |author4=J. Chen |author5=C. Liao |author6=Z. Zheng |author7=J. Degenhardt. | title=Early exit optimizations for additive machine learned ranking systems | journal=WSDM '10: Proceedings of the Third ACM International Conference on Web Search and Data Mining, 2010. | url=http://olivier.chapelle.cc/pub/wsdm2010.pdf}} 6. ^{{citation |author1=Broder A. |author2=Carmel D. |author3=Herscovici M. |author4=Soffer A. |author5=Zien J. | title=Efficient query evaluation using a two-level retrieval process | journal=Proceedings of the Twelfth International Conference on Information and Knowledge Management | year=2003 | pages=426–434 | isbn=978-1-58113-723-1 | url=http://cis.poly.edu/westlab/papers/cntdstrb/p426-broder.pdf }} 7. ^1 {{citation |author1=Manning C. |author2=Raghavan P. |author3=Schütze H. | title=Introduction to Information Retrieval | publisher=Cambridge University Press | year=2008}}. Section 7.1 8. ^1 {{citation | author=Kevin K. Duh | title=Learning to Rank with {{sic|hide=y|Partially|-}}Labeled Data | year=2009 | url=http://ssli.ee.washington.edu/people/duh/thesis/uwthesis.pdf}} 9. ^Yuanhua Lv, Taesup Moon, Pranam Kolari, Zhaohui Zheng, Xuanhui Wang, and Yi Chang, Learning to Model Relatedness for News Recommendation {{Webarchive|url=https://web.archive.org/web/20110827065356/http://sifaka.cs.uiuc.edu/~ylv2/pub/www11-relatedness.pdf |date=2011-08-27 }}, in International Conference on World Wide Web (WWW), 2011. 10. ^{{Cite book|doi = 10.1109/ICSME.2014.41|chapter = Learning to Combine Multiple Ranking Metrics for Fault Localization|title = 2014 IEEE International Conference on Software Maintenance and Evolution|pages = 191–200|year = 2014|last1 = Xuan|first1 = Jifeng|last2 = Monperrus|first2 = Martin|citeseerx = 10.1.1.496.6829|chapter-url=https://hal.archives-ouvertes.fr/hal-01018935/document|isbn = 978-1-4799-6146-7}} 11. ^{{cite conference | first=M. |last=Richardson |author2=Prakash, A. |author3=Brill, E. | title=Beyond PageRank: Machine Learning for Static Ranking | booktitle=Proceedings of the 15th International World Wide Web Conference | pages=707–715 | publisher= | year=2006 | url=http://research.microsoft.com/en-us/um/people/mattri/papers/www2006/staticrank.pdf | accessdate= }} 12. ^LETOR 3.0. A Benchmark Collection for Learning to Rank for Information Retrieval 13. ^http://www.stanford.edu/class/cs276/handouts/lecture15-learning-ranking.ppt 14. ^{{citation|author1=Olivier Chapelle |author2=Donald Metzler |author3=Ya Zhang |author4=Pierre Grinspan |title=Expected Reciprocal Rank for Graded Relevance|url=http://research.yahoo.com/files/err.pdf |archive-url=https://web.archive.org/web/20120224053008/http://research.yahoo.com/files/err.pdf |dead-url=yes |archive-date=2012-02-24 |journal=CIKM|year=2009|pages=}} 15. ^{{citation|author1=Gulin A. |author2=Karpovich P. |author3=Raskovalov D. |author4=Segalovich I. |title=Yandex at ROMIP'2009: optimization of ranking algorithms by machine learning methods|url=http://romip.ru/romip2009/15_yandex.pdf|journal=Proceedings of ROMIP'2009|year=2009|pages=163–168}} (in Russian) 16. ^{{citation |author1=Tax, Niek |author2=Bockting, Sander |author3=Hiemstra, Djoerd | journal=Information Processing & Management |volume=51 |issue=6 | title=A cross-benchmark comparison of 87 learning to rank methods | pages=757–772 | year=2015 | url=http://wwwhome.cs.utwente.nl/~hiemstra/papers/ipm2015.pdf | doi=10.1016/j.ipm.2015.07.002}} 17. ^1 {{citation | last=Fuhr | first=Norbert | journal=ACM Transactions on Information Systems | title=Optimum polynomial retrieval functions based on the probability ranking principle | volume=7 | number=3 | pages=183–204 | year=1989 | doi=10.1145/65943.65944}} 18. ^1 {{citation |author1=Cooper, William S. |author2=Gey, Frederic C. |author3=Dabney, Daniel P. | journal=SIGIR '92 Proceedings of the 15th Annual International ACM SIGIR Conference on Research and Development in Information Retrieval | title=Probabilistic retrieval based on staged logistic regression | pages=198–210 | year=1992 | doi=10.1145/133160.133199|isbn=978-0897915236 }} 19. ^{{cite journal | citeseerx = 10.1.1.20.378 | title = Pranking }} 20. ^{{cite journal | citeseerx = 10.1.1.90.220 | title = RankGP }} 21. ^{{Citation|last=Pahikkala|first=Tapio |author2=Tsivtsivadze, Evgeni |author3=Airola, Antti |author4=Järvinen, Jouni |author5=Boberg, Jorma |title=An efficient algorithm for learning to rank from preference graphs|journal=Machine Learning|year=2009|volume=75|issue=1|pages=129–165|doi=10.1007/s10994-008-5097-z|postscript=.}} 22. ^C. Burges. (2010). [https://www.microsoft.com/en-us/research/wp-content/uploads/2016/02/MSR-TR-2010-82.pdf From RankNet to LambdaRank to LambdaMART: An Overview]. 23. ^Rong Jin, Hamed Valizadegan, Hang Li, Ranking Refinement and Its Application for Information Retrieval, in International Conference on World Wide Web (WWW), 2008. 24. ^Massih-Reza Amini, Vinh Truong, Cyril Goutte, A Boosting Algorithm for Learning Bipartite Ranking Functions with Partially Labeled Data {{Webarchive|url=https://web.archive.org/web/20100802093049/http://www-connex.lip6.fr/~amini/Publis/SemiSupRanking_sigir08.pdf |date=2010-08-02 }}, International ACM SIGIR conference, 2008. The code {{Webarchive|url=https://web.archive.org/web/20100723152841/http://www-connex.lip6.fr/~amini/SSRankBoost/ |date=2010-07-23 }} is available for research purposes. 25. ^Leonardo Rigutini, Tiziano Papini, Marco Maggini, Franco Scarselli, "SortNet: learning to rank by a neural-based sorting algorithm", SIGIR 2008 workshop: Learning to Rank for Information Retrieval, 2008 26. ^Hamed Valizadegan, Rong Jin, Ruofei Zhang, Jianchang Mao, Learning to Rank by Optimizing NDCG Measure, in Proceeding of Neural Information Processing Systems (NIPS), 2010. 27. ^{{citation | last=Fuhr | first=Norbert | journal=Computer Journal | title=Probabilistic Models in Information Retrieval | volume=35 | number=3 | pages=243–255 | year=1992 | doi=10.1093/comjnl/35.3.243}} 28. ^{{citation |author1=Manning C. |author2=Raghavan P. |author3=Schütze H. |title=Introduction to Information Retrieval |publisher=Cambridge University Press |year=2008}}. Sections 7.4 and 15.5 29. ^Jan O. Pedersen. The MLR Story {{Webarchive|url=https://web.archive.org/web/20110713120113/http://jopedersen.com/Presentations/The_MLR_Story.pdf |date=2011-07-13 }} 30. ^{{US Patent|7197497}} 31. ^Bing Search Blog: User Needs, Features and the Science behind Bing 32. ^Yandex corporate blog entry about new ranking model "Snezhinsk" (in Russian) 33. ^The algorithm wasn't disclosed, but a few details were made public in and . 34. ^Yandex's Internet Mathematics 2009 competition page 35. ^{{Cite web |url=http://learningtorankchallenge.yahoo.com/ |title=Yahoo Learning to Rank Challenge |access-date=2010-02-26 |archive-url=https://web.archive.org/web/20100301011649/http://learningtorankchallenge.yahoo.com/ |archive-date=2010-03-01 |dead-url=yes |df= }} 36. ^{{cite web |url = http://anand.typepad.com/datawocky/2008/05/are-human-experts-less-prone-to-catastrophic-errors-than-machine-learned-models.html |archiveurl = https://www.webcitation.org/5sq8irWNM?url=http://anand.typepad.com/datawocky/2008/05/are-human-experts-less-prone-to-catastrophic-errors-than-machine-learned-models.html |archivedate = 2010-09-18 |title = Are Machine-Learned Models Prone to Catastrophic Errors? |date = 2008-05-24 |last = Rajaraman |first = Anand |authorlink = Anand Rajaraman |access-date = 2009-11-11 |dead-url = no |df = }} 37. ^{{cite web | url = http://www.cuil.com/info/blog/2009/06/26/so-how-is-bing-doing | archiveurl = https://archive.is/20090627213358/http://www.cuil.com/info/blog/2009/06/26/so-how-is-bing-doing | archivedate = 2009-06-27 | title = Cuil Blog: So how is Bing doing? | date = 2009-06-26 | last = Costello | first = Tom}} 38. ^{{Cite news|url=https://www.techatbloomberg.com/blog/bloomberg-integrated-learning-rank-apache-solr/|title=How Bloomberg Integrated Learning-to-Rank into Apache Solr {{!}} Tech at Bloomberg|date=2017-01-23|work=Tech at Bloomberg|access-date=2017-02-28|language=en-US}} External links
3 : Information retrieval techniques|Machine learning|Ranking functions |
|||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||
随便看 |
|
开放百科全书收录14589846条英语、德语、日语等多语种百科知识,基本涵盖了大多数领域的百科知识,是一部内容自由、开放的电子版国际百科全书。