online ranking combination - acm recommender …...combination algorithms in the experiments direct...

24
Online Ranking Combination Erzs´ ebet Frig´ o Institute for Computer Science and Control (MTA SZTAKI) Joint work with Levente Kocsis

Upload: others

Post on 22-Apr-2020

13 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: Online Ranking Combination - ACM Recommender …...Combination algorithms in the experiments Direct optimization: I ExpW I exponentially weighted forecaster on a grid I global optimization

Online Ranking Combination

Erzsebet Frigo

Institute for Computer Science and Control (MTA SZTAKI)

Joint work with Levente Kocsis

Page 2: Online Ranking Combination - ACM Recommender …...Combination algorithms in the experiments Direct optimization: I ExpW I exponentially weighted forecaster on a grid I global optimization

Overview

I Framework: prequential ranking evaluation

I Goal: optimize convex combination of ranking models

I Our proposal: direct optimization of the ranking function

Page 3: Online Ranking Combination - ACM Recommender …...Combination algorithms in the experiments Direct optimization: I ExpW I exponentially weighted forecaster on a grid I global optimization

Model combination in prequential framework with ranking evaluation

time

i2

i7

i3

u

i5i2

im-1

im

i1

...

i2

im-1

im

i1

...

i2

im-1

im

i1

...

A1 A2 A3

scores ranking listcombine

Objective: choosing combination weights.

Page 4: Online Ranking Combination - ACM Recommender …...Combination algorithms in the experiments Direct optimization: I ExpW I exponentially weighted forecaster on a grid I global optimization

Model combination in prequential framework with ranking evaluation

time

i2

i7

i3

u

i5i2

im-1

im

i1

...

i2

im-1

im

i1

...

i2

im-1

im

i1

...

A1 A2 A3

scores ranking listcombine

Objective: choosing combination weights.

Page 5: Online Ranking Combination - ACM Recommender …...Combination algorithms in the experiments Direct optimization: I ExpW I exponentially weighted forecaster on a grid I global optimization

New idea: optimize ranking function directly

I Standard method: take a surrogate function and use its gradientI E.g. MSE

I Drawback: optimum of the surrogate 6= optimum of ranking function

I Proposed solution: optimize the ranking function directlyI Two approaches:

I Global search in the weight spaceI Gradient approximation (finite differences)

Page 6: Online Ranking Combination - ACM Recommender …...Combination algorithms in the experiments Direct optimization: I ExpW I exponentially weighted forecaster on a grid I global optimization

ExpW

I Choose a subset Q of the weight space ΘI e.g., lay a grid to the parameter space

I Apply exponentially weighted forecaster on Q

P(select q ∈ Q in round t) =e−ηt

∑t−1τ=1(1−rτ (q))∑

s∈Q e−ηt∑t−1

τ=1(1−rτ (s))

I Theoretical guarantee:

E [RT (best static combination in Θ)− RT (ExpW )] < O(√T )

I if the cumulative reward function (RT ) is sufficiently smoothI and Q is sufficiently large

I Difficulty: size of Q is exponential in number of base rankers, can’t scale

Page 7: Online Ranking Combination - ACM Recommender …...Combination algorithms in the experiments Direct optimization: I ExpW I exponentially weighted forecaster on a grid I global optimization

Simultaneous Perturbation Stochastic Approximation (SPSA)

I Approximated gradient (for the weight of base ranker i in round t):

gti =rt(θt + ct∆t)− rt(θt − ct∆t)

ct∆ti

I θt is the current combination weight vectorI ∆t = (∆t1, ...) is a random vector of +/-1I ct is perturbation step size

I Online update step: one gradient step using the approximated gradient

Page 8: Online Ranking Combination - ACM Recommender …...Combination algorithms in the experiments Direct optimization: I ExpW I exponentially weighted forecaster on a grid I global optimization
Page 9: Online Ranking Combination - ACM Recommender …...Combination algorithms in the experiments Direct optimization: I ExpW I exponentially weighted forecaster on a grid I global optimization
Page 10: Online Ranking Combination - ACM Recommender …...Combination algorithms in the experiments Direct optimization: I ExpW I exponentially weighted forecaster on a grid I global optimization
Page 11: Online Ranking Combination - ACM Recommender …...Combination algorithms in the experiments Direct optimization: I ExpW I exponentially weighted forecaster on a grid I global optimization

RSPSA

I RSPSA = SPSA + Resilient Backpropagation (RProp)

I RProp defines gradient step sizes for each weight

I Perturbation step size is tied to gradient step size

I Update step sizes using RProp

Page 12: Online Ranking Combination - ACM Recommender …...Combination algorithms in the experiments Direct optimization: I ExpW I exponentially weighted forecaster on a grid I global optimization

RSPSA

I RSPSA = SPSA + Resilient Backpropagation (RProp)

I RProp defines gradient step sizes for each weight

I Perturbation step size is tied to gradient step size

I Update step sizes using RProp

Page 13: Online Ranking Combination - ACM Recommender …...Combination algorithms in the experiments Direct optimization: I ExpW I exponentially weighted forecaster on a grid I global optimization

Resilient Backpropagation (RProp)

I Gradient update ruleI Predefined step size for each coordinate

I ignores the length of the gradient vector

I Step size is updated based on the sign of the gradientI decrease step if gradient changed directionI increase otherwise

Page 14: Online Ranking Combination - ACM Recommender …...Combination algorithms in the experiments Direct optimization: I ExpW I exponentially weighted forecaster on a grid I global optimization

gti = rt(θt+ct∆t)−rt(θt−ct∆t)ct∆ti

Page 15: Online Ranking Combination - ACM Recommender …...Combination algorithms in the experiments Direct optimization: I ExpW I exponentially weighted forecaster on a grid I global optimization

RFDSA+

I Switch to finite differences (FD)I allows to detect 0 gradient w.r.t. one coordinate

I If the gradient is 0 w.r.t. a coordinate, thenI increase perturbation size (+) for that coordinateI escape flat section in the right direction

I RFDSA+ = RSPSA - simultaneous perturbation + finite differences + zerogradient detection

I The modifications might seem to be minor, but are essential to make thealgorithm work

Page 16: Online Ranking Combination - ACM Recommender …...Combination algorithms in the experiments Direct optimization: I ExpW I exponentially weighted forecaster on a grid I global optimization

Experiments - Datasets, base rankers

I 5 datasetsI Amazon

I CDs and VinylI Movies and TVI Electronics

I MovieLens 10MI Twitter

I hashtag prediction

I SizeI # of events: 2M-10MI # of users: 70k-4MI # of items: 10k-100k

I Base rankers:I Models updated incrementally

SGD Matrix Factorization

Asymmetric Matrix Factorization

Item-to-item similarity

25% 25%30% 15% 5%

Most popularI Traditional models updated periodically

SGD Matrix Factorization

Implicit Alternating Least Squares MF

Page 17: Online Ranking Combination - ACM Recommender …...Combination algorithms in the experiments Direct optimization: I ExpW I exponentially weighted forecaster on a grid I global optimization

Combination algorithms in the experiments

Direct optimization:I ExpW

I exponentially weighted forecaster ona grid

I global optimization

I SPSAI gradient method with simultaneous

perturbation

I RSPSAI SPSA with RProp

I RFDSA+I our new algorithmI finite differences, flat section

detection

Baselines:I ExpA

I exponentially weighted forecaster onthe base rankers

I ExpAWI use probabilities of ExpA as weights

I SGDI use MSE as a surrogateI target=1 for positive sampleI target=0 for generated negative

samples

Page 18: Online Ranking Combination - ACM Recommender …...Combination algorithms in the experiments Direct optimization: I ExpW I exponentially weighted forecaster on a grid I global optimization

Results - 2 base rankers (i2i, OMF) - nDCG

0

0.01

0.02

0.03

0.04

0.05

0.06

0.07

0 1000 2000 3000 4000 5000 6000 7000

ND

CG

days

item2itemOMFExpA

ExpAWExpWSGD

SPSARSPSA

RFDSA+

Page 19: Online Ranking Combination - ACM Recommender …...Combination algorithms in the experiments Direct optimization: I ExpW I exponentially weighted forecaster on a grid I global optimization

Results - 2 base rankers - Combination weights

0.0000010

0.0000100

0.0001000

0.0010000

0.0100000

0.1000000

1.0000000

0 1000 2000 3000 4000 5000 6000 7000

θ

days

OptG100+ExpAW

SGDSPSA

RSPSARFDSA+

Page 20: Online Ranking Combination - ACM Recommender …...Combination algorithms in the experiments Direct optimization: I ExpW I exponentially weighted forecaster on a grid I global optimization

Cumulative reward as function of combination weight

0.038

0.04

0.042

0.044

0.046

0.048

0.05

0.052

0.054

0.0001 0.001 0.01 0.1 1

ND

CG

θ

RT(θ)

Page 21: Online Ranking Combination - ACM Recommender …...Combination algorithms in the experiments Direct optimization: I ExpW I exponentially weighted forecaster on a grid I global optimization

Results - Scalability

0.03

0.035

0.04

0.045

0.05

0.055

0.06

1 2 3 4 5 6 7 8 9 10

ND

CG

number of OMF’s

ExpAExpAW

SGD

SPSARSPSA

RFDSA+

Page 22: Online Ranking Combination - ACM Recommender …...Combination algorithms in the experiments Direct optimization: I ExpW I exponentially weighted forecaster on a grid I global optimization

Results - 6 base rankers - DCG

Page 23: Online Ranking Combination - ACM Recommender …...Combination algorithms in the experiments Direct optimization: I ExpW I exponentially weighted forecaster on a grid I global optimization

Conclusions

I Problem: combine ranking algorithms

I Our proposal: optimize the ranking measure directly

I Global optimization (ExpW) works well in case of two base algoI Our new algo: RFDSA+

I solves problems (scaling, constant sections w.r.t one coordinate)I strong combination

Page 24: Online Ranking Combination - ACM Recommender …...Combination algorithms in the experiments Direct optimization: I ExpW I exponentially weighted forecaster on a grid I global optimization

The End

Online Ranking Combination

Erzsebet Frigo

Institute for Computer Science and Control (MTA SZTAKI)

Joint work with Levente Kocsis