参考文献
•Auer, P., Cesa-Bianchi,N., and Fischer, P. Finite-time analysis of the
multiarmed bandit problem. Machine Learning, 47(2-3):235‒256, 2002a.
•Cesa-Bianchi, N. and Lugosi, G. Prediction, learning, and games. Cambridge
University Press, 2006.
•Chapelle, O. and Chang, Y. Yahoo! learning to rank challenge overview. JMLR Proceedings Track, 14:1‒ 24, 2011.
•Freund, Y., Iyer, R. D., Schapire, R. E., and Singer, Y. An efficient boosting
algorithm for combining preferences. Journal of Machine Learning Research,
4:933‒969, 2003.
•Joachims, T., Granka, L., Pan, Bing, Hembrooke, H., Radlinski, F., and Gay, G.
Evaluating the accuracy of implicit feedback from clicks and query
reformulations in web search. ACM Transactions on Information Systems (TOIS),
25(2), April 2007.
•Yue, Y. and Joachims, T. Interactively optimizing infor- mation retrieval systems
as a dueling bandits problem. In ICML, 2009.
13年3月30日土曜日
25