Inviting an author to review:
Find an author and click ‘Invite to review selected article’ near their name.
Search for authorsSearch for similar articles
24
views
0
recommends
+1 Recommend
0 collections
    0
    shares
      • Record: found
      • Abstract: found
      • Article: found
      Is Open Access

      A Contextual-Bandit Approach to Personalized News Article Recommendation

      Preprint
      , , ,

      Read this article at

      Bookmark
          There is no author summary for this article yet. Authors can add summaries to their articles on ScienceOpen to make them more accessible to a non-specialist audience.

          Abstract

          Personalized web services strive to adapt their services (advertisements, news articles, etc) to individual users by making use of both content and user information. Despite a few recent advances, this problem remains challenging for at least two reasons. First, web service is featured with dynamically changing pools of content, rendering traditional collaborative filtering methods inapplicable. Second, the scale of most web services of practical interest calls for solutions that are both fast in learning and computation. In this work, we model personalized recommendation of news articles as a contextual bandit problem, a principled approach in which a learning algorithm sequentially selects articles to serve users based on contextual information about the users and articles, while simultaneously adapting its article-selection strategy based on user-click feedback to maximize total user clicks. The contributions of this work are three-fold. First, we propose a new, general contextual bandit algorithm that is computationally efficient and well motivated from learning theory. Second, we argue that any bandit algorithm can be reliably evaluated offline using previously recorded random traffic. Finally, using this offline evaluation method, we successfully applied our new algorithm to a Yahoo! Front Page Today Module dataset containing over 33 million events. Results showed a 12.5% click lift compared to a standard context-free bandit algorithm, and the advantage becomes even greater when data gets more scarce.

          Related collections

          Most cited references13

          • Record: found
          • Abstract: not found
          • Article: not found

          Asymptotically efficient adaptive allocation rules

            Bookmark
            • Record: found
            • Abstract: not found
            • Article: not found

            The Nonstochastic Multiarmed Bandit Problem

              Bookmark
              • Record: found
              • Abstract: not found
              • Article: not found

              Some aspects of the sequential design of experiments

                Bookmark

                Author and article information

                Journal
                27 February 2010
                2012-03-01
                Article
                10.1145/1772690.1772758
                1003.0146
                db14ef76-fa59-45ff-a949-ad67c2f247a3

                http://arxiv.org/licenses/nonexclusive-distrib/1.0/

                History
                Custom metadata
                Presented at the Nineteenth International Conference on World Wide Web (WWW 2010), Raleigh, NC, USA, 2010
                10 pages, 5 figures
                cs.LG cs.AI cs.IR

                Comments

                Comment on this article