1 of 26

Tomáš Tunys

Jan Šedivý�

LEARNING TO RANK

FROM

IMPLICIT USER FEEDBACK

dissertation thesis proposal

2 of 26

Presentation Outline

  • Introduction and Motivation
  • Learning To Rank Overview
    • Learning to Rank formulation, algorithms, and evaluation
    • Training data quality/quantity problem
  • Learning To Rank from Implicit feedback
    • Passive vs Active learning
  • Current and Future Work

3 of 26

Introduction and Motivation

4 of 26

Needles in the Haystack

“With No Signs of Slowing, the Data Keeps Growing”

  • The global Internet population grew 6.59 % from 2010 to 2011 and now represents

2.1 BILLION PEOPLE

source: http://www.domo.com/learn/

5 of 26

Needles in the Haystack

“With No Signs of Slowing, the Data Keeps Growing”

  • The global Internet population grew 14.3 % from 2011 to 2013 and now represents

2.4 BILLION PEOPLE

source: http://www.domo.com/learn/

6 of 26

Corollary: Satisfying Information Needs Becomes Harder

7 of 26

Corollary: Satisfying Information Needs Becomes Harder

8 of 26

What is Learning To Rank?

9 of 26

Learning To Rank

Searching for relevant documents

Documents

Document d1

Document d2

Document d3

Document d4

Document d5

Feature Vectors

(... q + d1 …)

(... q + d2 …)

(... q + d3 …)

(... q + d4 …)

(... q + d5 …)

Query q

Relevance Labels

relevance y1

relevance y2

relevance y3

relevance y4

relevance y5

10 of 26

Learning To Rank

Searching for relevant documents

Documents

Document d1

Document d2

Document d3

Document d4

Document d5

Feature Vectors

(... q + d1 …)

(... q + d2 …)

(... q + d3 …)

(... q + d4 …)

(... q + d5 …)

Query q

Relevance Labels

relevance y1

relevance y2

relevance y3

relevance y4

relevance y5

?

?

11 of 26

Learning To Rank

Searching for relevant documents

Documents

Document d1

Document d2

Document d3

Document d4

Document d5

Feature Vectors

(... q + d1 …)

(... q + d2 …)

(... q + d3 …)

(... q + d4 …)

(... q + d5 …)

Query q

Relevance Labels

relevance y1

relevance y2

relevance y3

relevance y4

relevance y5

Ranking Model

Learning phase

12 of 26

Learning To Rank

Searching for relevant documents

Documents

Document d1

Document d2

Document d3

Document d4

Document d5

Feature Vectors

(... q + d1 …)

(... q + d2 …)

(... q + d3 …)

(... q + d4 …)

(... q + d5 …)

Query q

Ranking Model

(inference)

score s1

score s2

score s3

score s4

score s5

Production phase

13 of 26

Learning to Rank

Formulation, Evaluation, Algorithms

14 of 26

Learning To Rank: Problem Formulation

Empirical Risk Minimization

  • Given a labeled dataset S = {(qi, Di, yi)}i = 1, where qi is a query, Di are associated documents, and yi are the corresponding relevances of documents to the query.
  • The goal is to find a ranker f: RN→R such that when the documents are sorted according to ranking scores, then such permutation π(f, qi, Di) will leave the documents sorted in descending order according to their relevance to the query.

Q

performance measure

15 of 26

Evaluation Measures

Normalized Discounted Cumulative Gain

  • One of the most frequently used measures. It is based on positional model.

16 of 26

Evaluation Measures

  • All (interesting) performance measures are non-convex, non-differentiable, and discontinuous functions of the ranking scores.
  • Pose huge challenges for direct optimization.�Sidestepped by:
    • non-convex optimization techniques.
    • convex approximation measures
    • surrogate measures

17 of 26

Training Data Problem

  • The performance “bottleneck” - the quality and quantity matters “much more” than in case of ordinary classification or regression.

Very expensive, time consuming, and of disputable quality.

18 of 26

Learning To Rank

  • Pointwise algorithms (PRank):
    • Based on reducing the problem of ranking to classification or regression.
    • Pros: computational complexity (linear)
    • Cons: often does not work!
  • Pairwise algorithms (RankSVM):
    • Reduces the problem to binary classification on pairs of documents.
    • Pros: decoupling from relevance scores
    • Cons: computational complexity (quadratic)

Types of Algorithms

19 of 26

Learning To Rank

  • Listwise algorithms (LambdaMART):
    • Current state of the art in Learning To Rank.
    • Lambda “trick” + MART
    • Pros: directly optimizes performance measures! computational complexity (in production)
    • Cons: computational complexity (in training)

Types of Algorithms

20 of 26

Learning To Rank From Implicit Feedback

21 of 26

Learning To Rank

  • The relevance of documents to queries are willingly judged by trained people - annotators.
  • Ordinary people hardly ever provide this kind of explicit feedback (if they are asked).
  • But query logs store enormous amount of data which can be exploited to assess the relevance of documents to queries - user implicit feedback:
    • clickthrough data
    • dwell time
    • query reformulations

Exploiting User Implicit Feedback

22 of 26

Learning To Rank

  • Implicit Feedback data can enter into the process of training in several different ways:
    • new features in query-document vectors
    • optimization constraints (RankSVM)
    • enlarging document sets for queries (reformulations)
    • improve quality of training data
    • clustering queries according to their type/user intents
    • interleaving evaluation (active learning/bandits)
  • Potential problems: noisy data, biased in different ways

Exploiting User Implicit Feedback

23 of 26

Learning To Rank

  • Use of implicit feedback = a huge frontier of unexplored potential. Reason - query log data unavailability (starting to change - Living Labs for IR Evaluation)

Exploiting User Implicit Feedback

24 of 26

Current and Future Work

  • Currently I am participating on a project from University of Amsterdam led by dr. Masrour Zoghi (ILPS).
    • Preliminary goals:
      • Use clickthrough data to identify important features and cluster queries according to trained (linear) rankers for them (using DBGD + click models)
      • Design and train ensemble of query-cluster specific rankers.
  • Plan to visit UvA for 4 months this semester.

25 of 26

Current and Future Work

  • The biggest supporter of this research is Seznam.cz. It provides data and computational infrastructure:
    • It has just entered into Living Labs project:
      • Provide researchers access to real-time/real-world environment and possibility to benchmark their algorithm against participating systems.
    • My task is to test the functionality of their implementation of the LLC API.

26 of 26

Current and Future Work

  • Improve our implementation of LambdaMART and introduce clickthrough data into the training process.