pranking with ranking€¦ · i ranking rule (h) : mapping from instances to ranks, rn!y. the...
TRANSCRIPT
![Page 1: Pranking with Ranking€¦ · I Ranking Rule (H) : Mapping from instances to ranks, Rn!Y. The family of ranking rules considered here : w 2Rn and k thresholds : b 1 b 2 ::: b k =](https://reader034.vdocuments.net/reader034/viewer/2022050105/5f435bf97497cb07662b9f9d/html5/thumbnails/1.jpg)
Pranking with Ranking
Koby Crammer and Yoram Singer
Presented by : Soham Dan
Content and some figures borrowed from [Crammer, Koby, and YoramSinger. Pranking with ranking.NIPS. 2002] and talk slides.
![Page 2: Pranking with Ranking€¦ · I Ranking Rule (H) : Mapping from instances to ranks, Rn!Y. The family of ranking rules considered here : w 2Rn and k thresholds : b 1 b 2 ::: b k =](https://reader034.vdocuments.net/reader034/viewer/2022050105/5f435bf97497cb07662b9f9d/html5/thumbnails/2.jpg)
Introduction
I ProblemI Input : Sequence of instance-rank pairs (x1, y1)...(x t , y t)I Output : A model(essentially a rank prediction rule) which
assigns to each instance a rank.I Goal: To have the predicted rank as close as possible to the
true rank.I Note : The ranks need not be unique!
I Similarity withI Classification Problems : Assign one of k possible labels to a
new instance.I Regression Problems : Set of k labels is structured as there is a
total order relation between labels.
Natural Settings to rank / rate instances
Information Retrieval , Collaborative Filtering
![Page 3: Pranking with Ranking€¦ · I Ranking Rule (H) : Mapping from instances to ranks, Rn!Y. The family of ranking rules considered here : w 2Rn and k thresholds : b 1 b 2 ::: b k =](https://reader034.vdocuments.net/reader034/viewer/2022050105/5f435bf97497cb07662b9f9d/html5/thumbnails/3.jpg)
Introduction
I ProblemI Input : Sequence of instance-rank pairs (x1, y1)...(x t , y t)I Output : A model(essentially a rank prediction rule) which
assigns to each instance a rank.I Goal: To have the predicted rank as close as possible to the
true rank.I Note : The ranks need not be unique!
I Similarity withI Classification Problems : Assign one of k possible labels to a
new instance.I Regression Problems : Set of k labels is structured as there is a
total order relation between labels.
Natural Settings to rank / rate instances
Information Retrieval , Collaborative Filtering
![Page 4: Pranking with Ranking€¦ · I Ranking Rule (H) : Mapping from instances to ranks, Rn!Y. The family of ranking rules considered here : w 2Rn and k thresholds : b 1 b 2 ::: b k =](https://reader034.vdocuments.net/reader034/viewer/2022050105/5f435bf97497cb07662b9f9d/html5/thumbnails/4.jpg)
Introduction
I ProblemI Input : Sequence of instance-rank pairs (x1, y1)...(x t , y t)I Output : A model(essentially a rank prediction rule) which
assigns to each instance a rank.I Goal: To have the predicted rank as close as possible to the
true rank.I Note : The ranks need not be unique!
I Similarity withI Classification Problems : Assign one of k possible labels to a
new instance.I Regression Problems : Set of k labels is structured as there is a
total order relation between labels.
Natural Settings to rank / rate instances
Information Retrieval , Collaborative Filtering
![Page 5: Pranking with Ranking€¦ · I Ranking Rule (H) : Mapping from instances to ranks, Rn!Y. The family of ranking rules considered here : w 2Rn and k thresholds : b 1 b 2 ::: b k =](https://reader034.vdocuments.net/reader034/viewer/2022050105/5f435bf97497cb07662b9f9d/html5/thumbnails/5.jpg)
Problem
Figure 1: Movie rating prediction (Example : Netflix challenge)
![Page 6: Pranking with Ranking€¦ · I Ranking Rule (H) : Mapping from instances to ranks, Rn!Y. The family of ranking rules considered here : w 2Rn and k thresholds : b 1 b 2 ::: b k =](https://reader034.vdocuments.net/reader034/viewer/2022050105/5f435bf97497cb07662b9f9d/html5/thumbnails/6.jpg)
Possible Solutions
I Cast as a regression or classification problem
I Reduce a total order into a set of preference over pairs.Drawback : Sample size blowup from n to Ø(n2). Also, noeasy adaptation for online settings.
I PRank Algorithm : Directly maintains totally ordered set byprojection of instances into reals, associating ranks withdistinct sub-intervals of the reals and adapting the support ofeach subinterval while learning.
![Page 7: Pranking with Ranking€¦ · I Ranking Rule (H) : Mapping from instances to ranks, Rn!Y. The family of ranking rules considered here : w 2Rn and k thresholds : b 1 b 2 ::: b k =](https://reader034.vdocuments.net/reader034/viewer/2022050105/5f435bf97497cb07662b9f9d/html5/thumbnails/7.jpg)
Possible Solutions
I Cast as a regression or classification problem
I Reduce a total order into a set of preference over pairs.Drawback : Sample size blowup from n to Ø(n2). Also, noeasy adaptation for online settings.
I PRank Algorithm : Directly maintains totally ordered set byprojection of instances into reals, associating ranks withdistinct sub-intervals of the reals and adapting the support ofeach subinterval while learning.
![Page 8: Pranking with Ranking€¦ · I Ranking Rule (H) : Mapping from instances to ranks, Rn!Y. The family of ranking rules considered here : w 2Rn and k thresholds : b 1 b 2 ::: b k =](https://reader034.vdocuments.net/reader034/viewer/2022050105/5f435bf97497cb07662b9f9d/html5/thumbnails/8.jpg)
Possible Solutions
I Cast as a regression or classification problem
I Reduce a total order into a set of preference over pairs.Drawback : Sample size blowup from n to Ø(n2). Also, noeasy adaptation for online settings.
I PRank Algorithm : Directly maintains totally ordered set byprojection of instances into reals, associating ranks withdistinct sub-intervals of the reals and adapting the support ofeach subinterval while learning.
![Page 9: Pranking with Ranking€¦ · I Ranking Rule (H) : Mapping from instances to ranks, Rn!Y. The family of ranking rules considered here : w 2Rn and k thresholds : b 1 b 2 ::: b k =](https://reader034.vdocuments.net/reader034/viewer/2022050105/5f435bf97497cb07662b9f9d/html5/thumbnails/9.jpg)
Problem Setup
I Input Stream: Sequence of instance-rank pairs(x1, y1)...(x t , y t) where each instance xt ∈ Rn.Corresponding rank y t ∈ Y which is a finite set with a totalorder relation (structured) . W.l.o.g. Y = 1, 2, 3..., k with >as the order relation. 1 ≺ 2 ≺ ... ≺ k
I Ranking Rule (H) : Mapping from instances to ranks,Rn → Y. The family of ranking rules considered here :w ∈ Rn and k thresholds : b1 ≤ b2 ≤ ... ≤ bk =∞
I Given a ranking rule defined by w and b, the predicted rank(y t) on a new instance x isH(x) = minr∈1,2,..,k {r : w · x − br < 0}
I Algorithm makes a mistake on instance x t if y t 6= y t and losson that input is |y t − y t |.
I Loss after T rounds is∑T
t=1 |y t − y t |
![Page 10: Pranking with Ranking€¦ · I Ranking Rule (H) : Mapping from instances to ranks, Rn!Y. The family of ranking rules considered here : w 2Rn and k thresholds : b 1 b 2 ::: b k =](https://reader034.vdocuments.net/reader034/viewer/2022050105/5f435bf97497cb07662b9f9d/html5/thumbnails/10.jpg)
Problem Setup
I Input Stream: Sequence of instance-rank pairs(x1, y1)...(x t , y t) where each instance xt ∈ Rn.Corresponding rank y t ∈ Y which is a finite set with a totalorder relation (structured) . W.l.o.g. Y = 1, 2, 3..., k with >as the order relation. 1 ≺ 2 ≺ ... ≺ k
I Ranking Rule (H) : Mapping from instances to ranks,Rn → Y. The family of ranking rules considered here :w ∈ Rn and k thresholds : b1 ≤ b2 ≤ ... ≤ bk =∞
I Given a ranking rule defined by w and b, the predicted rank(y t) on a new instance x isH(x) = minr∈1,2,..,k {r : w · x − br < 0}
I Algorithm makes a mistake on instance x t if y t 6= y t and losson that input is |y t − y t |.
I Loss after T rounds is∑T
t=1 |y t − y t |
![Page 11: Pranking with Ranking€¦ · I Ranking Rule (H) : Mapping from instances to ranks, Rn!Y. The family of ranking rules considered here : w 2Rn and k thresholds : b 1 b 2 ::: b k =](https://reader034.vdocuments.net/reader034/viewer/2022050105/5f435bf97497cb07662b9f9d/html5/thumbnails/11.jpg)
Problem Setup
I Input Stream: Sequence of instance-rank pairs(x1, y1)...(x t , y t) where each instance xt ∈ Rn.Corresponding rank y t ∈ Y which is a finite set with a totalorder relation (structured) . W.l.o.g. Y = 1, 2, 3..., k with >as the order relation. 1 ≺ 2 ≺ ... ≺ k
I Ranking Rule (H) : Mapping from instances to ranks,Rn → Y. The family of ranking rules considered here :w ∈ Rn and k thresholds : b1 ≤ b2 ≤ ... ≤ bk =∞
I Given a ranking rule defined by w and b, the predicted rank(y t) on a new instance x isH(x) = minr∈1,2,..,k {r : w · x − br < 0}
I Algorithm makes a mistake on instance x t if y t 6= y t and losson that input is |y t − y t |.
I Loss after T rounds is∑T
t=1 |y t − y t |
![Page 12: Pranking with Ranking€¦ · I Ranking Rule (H) : Mapping from instances to ranks, Rn!Y. The family of ranking rules considered here : w 2Rn and k thresholds : b 1 b 2 ::: b k =](https://reader034.vdocuments.net/reader034/viewer/2022050105/5f435bf97497cb07662b9f9d/html5/thumbnails/12.jpg)
Problem Setup
I Input Stream: Sequence of instance-rank pairs(x1, y1)...(x t , y t) where each instance xt ∈ Rn.Corresponding rank y t ∈ Y which is a finite set with a totalorder relation (structured) . W.l.o.g. Y = 1, 2, 3..., k with >as the order relation. 1 ≺ 2 ≺ ... ≺ k
I Ranking Rule (H) : Mapping from instances to ranks,Rn → Y. The family of ranking rules considered here :w ∈ Rn and k thresholds : b1 ≤ b2 ≤ ... ≤ bk =∞
I Given a ranking rule defined by w and b, the predicted rank(y t) on a new instance x isH(x) = minr∈1,2,..,k {r : w · x − br < 0}
I Algorithm makes a mistake on instance x t if y t 6= y t and losson that input is |y t − y t |.
I Loss after T rounds is∑T
t=1 |y t − y t |
![Page 13: Pranking with Ranking€¦ · I Ranking Rule (H) : Mapping from instances to ranks, Rn!Y. The family of ranking rules considered here : w 2Rn and k thresholds : b 1 b 2 ::: b k =](https://reader034.vdocuments.net/reader034/viewer/2022050105/5f435bf97497cb07662b9f9d/html5/thumbnails/13.jpg)
Perceptron Recap
![Page 14: Pranking with Ranking€¦ · I Ranking Rule (H) : Mapping from instances to ranks, Rn!Y. The family of ranking rules considered here : w 2Rn and k thresholds : b 1 b 2 ::: b k =](https://reader034.vdocuments.net/reader034/viewer/2022050105/5f435bf97497cb07662b9f9d/html5/thumbnails/14.jpg)
Overview of Algorithm
I Online Algorithm
I In each round the ranking algorithmI Gets an input instanceI Outputs the rank as predictionI Receives the correct rank valueI If there is an error
I Computes lossI Updates the rank-prediction rule
I Conservative or Mistake driven algorithm :The algorithmupdates its ranking rule only on rounds on which it maderanking mistakes.
I No statistical assumptions over data.The algorithm should dowell irrespectively of specific sequence of inputs and targetlabels
![Page 15: Pranking with Ranking€¦ · I Ranking Rule (H) : Mapping from instances to ranks, Rn!Y. The family of ranking rules considered here : w 2Rn and k thresholds : b 1 b 2 ::: b k =](https://reader034.vdocuments.net/reader034/viewer/2022050105/5f435bf97497cb07662b9f9d/html5/thumbnails/15.jpg)
Overview of Algorithm
I Online AlgorithmI In each round the ranking algorithm
I Gets an input instanceI Outputs the rank as predictionI Receives the correct rank valueI If there is an error
I Computes lossI Updates the rank-prediction rule
I Conservative or Mistake driven algorithm :The algorithmupdates its ranking rule only on rounds on which it maderanking mistakes.
I No statistical assumptions over data.The algorithm should dowell irrespectively of specific sequence of inputs and targetlabels
![Page 16: Pranking with Ranking€¦ · I Ranking Rule (H) : Mapping from instances to ranks, Rn!Y. The family of ranking rules considered here : w 2Rn and k thresholds : b 1 b 2 ::: b k =](https://reader034.vdocuments.net/reader034/viewer/2022050105/5f435bf97497cb07662b9f9d/html5/thumbnails/16.jpg)
Overview of Algorithm
I Online AlgorithmI In each round the ranking algorithm
I Gets an input instanceI Outputs the rank as predictionI Receives the correct rank valueI If there is an error
I Computes lossI Updates the rank-prediction rule
I Conservative or Mistake driven algorithm :The algorithmupdates its ranking rule only on rounds on which it maderanking mistakes.
I No statistical assumptions over data.The algorithm should dowell irrespectively of specific sequence of inputs and targetlabels
![Page 17: Pranking with Ranking€¦ · I Ranking Rule (H) : Mapping from instances to ranks, Rn!Y. The family of ranking rules considered here : w 2Rn and k thresholds : b 1 b 2 ::: b k =](https://reader034.vdocuments.net/reader034/viewer/2022050105/5f435bf97497cb07662b9f9d/html5/thumbnails/17.jpg)
Overview of Algorithm
I Online AlgorithmI In each round the ranking algorithm
I Gets an input instanceI Outputs the rank as predictionI Receives the correct rank valueI If there is an error
I Computes lossI Updates the rank-prediction rule
I Conservative or Mistake driven algorithm :The algorithmupdates its ranking rule only on rounds on which it maderanking mistakes.
I No statistical assumptions over data.The algorithm should dowell irrespectively of specific sequence of inputs and targetlabels
![Page 18: Pranking with Ranking€¦ · I Ranking Rule (H) : Mapping from instances to ranks, Rn!Y. The family of ranking rules considered here : w 2Rn and k thresholds : b 1 b 2 ::: b k =](https://reader034.vdocuments.net/reader034/viewer/2022050105/5f435bf97497cb07662b9f9d/html5/thumbnails/18.jpg)
Algorithm Illustration
![Page 19: Pranking with Ranking€¦ · I Ranking Rule (H) : Mapping from instances to ranks, Rn!Y. The family of ranking rules considered here : w 2Rn and k thresholds : b 1 b 2 ::: b k =](https://reader034.vdocuments.net/reader034/viewer/2022050105/5f435bf97497cb07662b9f9d/html5/thumbnails/19.jpg)
Algorithm Illustration
![Page 20: Pranking with Ranking€¦ · I Ranking Rule (H) : Mapping from instances to ranks, Rn!Y. The family of ranking rules considered here : w 2Rn and k thresholds : b 1 b 2 ::: b k =](https://reader034.vdocuments.net/reader034/viewer/2022050105/5f435bf97497cb07662b9f9d/html5/thumbnails/20.jpg)
Algorithm Illustration
![Page 21: Pranking with Ranking€¦ · I Ranking Rule (H) : Mapping from instances to ranks, Rn!Y. The family of ranking rules considered here : w 2Rn and k thresholds : b 1 b 2 ::: b k =](https://reader034.vdocuments.net/reader034/viewer/2022050105/5f435bf97497cb07662b9f9d/html5/thumbnails/21.jpg)
Algorithm Illustration
![Page 22: Pranking with Ranking€¦ · I Ranking Rule (H) : Mapping from instances to ranks, Rn!Y. The family of ranking rules considered here : w 2Rn and k thresholds : b 1 b 2 ::: b k =](https://reader034.vdocuments.net/reader034/viewer/2022050105/5f435bf97497cb07662b9f9d/html5/thumbnails/22.jpg)
Algorithm Illustration
![Page 23: Pranking with Ranking€¦ · I Ranking Rule (H) : Mapping from instances to ranks, Rn!Y. The family of ranking rules considered here : w 2Rn and k thresholds : b 1 b 2 ::: b k =](https://reader034.vdocuments.net/reader034/viewer/2022050105/5f435bf97497cb07662b9f9d/html5/thumbnails/23.jpg)
Algorithm Illustration
![Page 24: Pranking with Ranking€¦ · I Ranking Rule (H) : Mapping from instances to ranks, Rn!Y. The family of ranking rules considered here : w 2Rn and k thresholds : b 1 b 2 ::: b k =](https://reader034.vdocuments.net/reader034/viewer/2022050105/5f435bf97497cb07662b9f9d/html5/thumbnails/24.jpg)
Algorithm Illustration
![Page 25: Pranking with Ranking€¦ · I Ranking Rule (H) : Mapping from instances to ranks, Rn!Y. The family of ranking rules considered here : w 2Rn and k thresholds : b 1 b 2 ::: b k =](https://reader034.vdocuments.net/reader034/viewer/2022050105/5f435bf97497cb07662b9f9d/html5/thumbnails/25.jpg)
Algorithm
Figure 2: The PRank Algorithm
I Rank y is expanded into k − 1 virtual variables y1, .., yk−1,where yr = +1 if w · x > br and yr = −1 otherwise.
I On mistakes, b and w · x are moved towards each other.
![Page 26: Pranking with Ranking€¦ · I Ranking Rule (H) : Mapping from instances to ranks, Rn!Y. The family of ranking rules considered here : w 2Rn and k thresholds : b 1 b 2 ::: b k =](https://reader034.vdocuments.net/reader034/viewer/2022050105/5f435bf97497cb07662b9f9d/html5/thumbnails/26.jpg)
Algorithm
Figure 2: The PRank Algorithm
I Rank y is expanded into k − 1 virtual variables y1, .., yk−1,where yr = +1 if w · x > br and yr = −1 otherwise.
I On mistakes, b and w · x are moved towards each other.
![Page 27: Pranking with Ranking€¦ · I Ranking Rule (H) : Mapping from instances to ranks, Rn!Y. The family of ranking rules considered here : w 2Rn and k thresholds : b 1 b 2 ::: b k =](https://reader034.vdocuments.net/reader034/viewer/2022050105/5f435bf97497cb07662b9f9d/html5/thumbnails/27.jpg)
Analysis
1. Lemma : OrderPreservation
2. Theorem : Mistake Bound
![Page 28: Pranking with Ranking€¦ · I Ranking Rule (H) : Mapping from instances to ranks, Rn!Y. The family of ranking rules considered here : w 2Rn and k thresholds : b 1 b 2 ::: b k =](https://reader034.vdocuments.net/reader034/viewer/2022050105/5f435bf97497cb07662b9f9d/html5/thumbnails/28.jpg)
Lemma : Order Preservation
Can this happen ?
NO
Let wt and bt be the current ranking rule, where bt1 ≤ ... ≤ btk−1and let (xt , yt) be an instance-rank pair fed to PRank on round t.Denote by wt+1 and bt+1 the resulting ranking rule after theupdate of PRank, then bt+1
1 ≤ ... ≤ bt+1k−1
![Page 29: Pranking with Ranking€¦ · I Ranking Rule (H) : Mapping from instances to ranks, Rn!Y. The family of ranking rules considered here : w 2Rn and k thresholds : b 1 b 2 ::: b k =](https://reader034.vdocuments.net/reader034/viewer/2022050105/5f435bf97497cb07662b9f9d/html5/thumbnails/29.jpg)
Lemma : Order Preservation
Can this happen ?
NO
Let wt and bt be the current ranking rule, where bt1 ≤ ... ≤ btk−1and let (xt , yt) be an instance-rank pair fed to PRank on round t.Denote by wt+1 and bt+1 the resulting ranking rule after theupdate of PRank, then bt+1
1 ≤ ... ≤ bt+1k−1
![Page 30: Pranking with Ranking€¦ · I Ranking Rule (H) : Mapping from instances to ranks, Rn!Y. The family of ranking rules considered here : w 2Rn and k thresholds : b 1 b 2 ::: b k =](https://reader034.vdocuments.net/reader034/viewer/2022050105/5f435bf97497cb07662b9f9d/html5/thumbnails/30.jpg)
Lemma : Order Preservation
Can this happen ?
NO
Let wt and bt be the current ranking rule, where bt1 ≤ ... ≤ btk−1and let (xt , yt) be an instance-rank pair fed to PRank on round t.Denote by wt+1 and bt+1 the resulting ranking rule after theupdate of PRank, then bt+1
1 ≤ ... ≤ bt+1k−1
![Page 31: Pranking with Ranking€¦ · I Ranking Rule (H) : Mapping from instances to ranks, Rn!Y. The family of ranking rules considered here : w 2Rn and k thresholds : b 1 b 2 ::: b k =](https://reader034.vdocuments.net/reader034/viewer/2022050105/5f435bf97497cb07662b9f9d/html5/thumbnails/31.jpg)
Lemma : Order Preservation
Let wt and bt be the current ranking rule, where bt1 ≤ ... ≤ btk−1and let (xt , yt) be an instance-rank pair fed to PRank on round t.Denote by wt+1 and bt+1 the resulting ranking rule after theupdate of PRank, then bt+1
1 ≤ ... ≤ bt+1k−1
Proof Sketch :
I btr are integers for all r and t since for all r we initializeb1r = 0, and bt+1
r − btr ∈ {−1, 0,+1}.I Proof by Induction :
Showing bt+1r+1 ≥ bt+1
r is equivalent to proving
btr+1−btr ≥ y tr+1[(wt ·xt−btr+1)y tr+1 ≤ 0]−y tr [(wt ·xt−btr )y tr ≤ 0]
![Page 32: Pranking with Ranking€¦ · I Ranking Rule (H) : Mapping from instances to ranks, Rn!Y. The family of ranking rules considered here : w 2Rn and k thresholds : b 1 b 2 ::: b k =](https://reader034.vdocuments.net/reader034/viewer/2022050105/5f435bf97497cb07662b9f9d/html5/thumbnails/32.jpg)
Lemma : Order Preservation
Figure 3: Intuitive Proof of Lemma
![Page 33: Pranking with Ranking€¦ · I Ranking Rule (H) : Mapping from instances to ranks, Rn!Y. The family of ranking rules considered here : w 2Rn and k thresholds : b 1 b 2 ::: b k =](https://reader034.vdocuments.net/reader034/viewer/2022050105/5f435bf97497cb07662b9f9d/html5/thumbnails/33.jpg)
Theorem : Mistake Bound
Let (xl , y1), ..., (xT , yT ) be an input sequence for PRank wherext ∈ Rn and yt ∈ l , ..., k . Denote by R2 = maxt ||xt ||2. Assumethat there is a ranking rule v∗ = (w∗, b∗) with b∗1 ≤ ... ≤ b∗k−1 of aunit norm that classifies the entire sequence correctly with marginγ = minr ,t (w∗ · xt − b∗r )y tr > 0. Then, the rank loss of the
algorithm∑T
t=1 |y t − y t |, is at most (k−1)(R2+1)γ2
.
![Page 34: Pranking with Ranking€¦ · I Ranking Rule (H) : Mapping from instances to ranks, Rn!Y. The family of ranking rules considered here : w 2Rn and k thresholds : b 1 b 2 ::: b k =](https://reader034.vdocuments.net/reader034/viewer/2022050105/5f435bf97497cb07662b9f9d/html5/thumbnails/34.jpg)
Proof of Theorem
I wt+1 = wt + (∑
r τtr )xt and bt+1
r = btr − τ trI Let nt = |y t − y t | be difference between the true rank and the
predicted rank. Clearly, nt =∑
r |τ tr |I To prove the theorem we bound
∑t n
t from above bybounding ||v t ||2 from above and below.
I v∗ · v t+1 = v∗ · v t +∑k−1
r=1 τtr (w∗x t − b∗r )
I∑k−1
r=1 τtr (w∗x t − b∗r ) ≥ ntγ =⇒ v∗vT+1 ≥ γ
∑t n
t =⇒||vT+1||2 ≥ γ2(
∑t n
t)2
![Page 35: Pranking with Ranking€¦ · I Ranking Rule (H) : Mapping from instances to ranks, Rn!Y. The family of ranking rules considered here : w 2Rn and k thresholds : b 1 b 2 ::: b k =](https://reader034.vdocuments.net/reader034/viewer/2022050105/5f435bf97497cb07662b9f9d/html5/thumbnails/35.jpg)
I To bound the norm of v from above :
I ||v t+1||2 = ||w t ||2 + ||bt ||2 + 2∑
r τtr (w t · x t − btr ) +
(∑
r τtr )2||x t ||2 +
∑r (τ tr )2
I Since, (∑
r τtr )2 ≤ (nt)2 and
∑r (τ tr )2 = nt
I ||v t+1||2 = ||v t ||2 + 2∑
r τtr (w t · x t − btr ) + (nt)2||x t ||2 + nt
I∑
r τtr (w t ·x t−btr ) =
∑r [(w t ·x t−btr ) ≤ 0](w t ·x t−btr )yr ≤ 0
I Since, ||x t ||2 ≤ R2 =⇒ ||v t+1||2 = ||v t ||2 + (nt)2R2 + nt
I Using the lower bound, we get,∑
t nt ≤ R2[
∑t(n
t)2]/[∑
t nt ]+1
γ2
I nt ≤ k − 1 =⇒∑
t(nt)2 ≤ (k − 1)
∑t n
t =⇒∑
t nt ≤
(k−1)(R2+1)γ2
![Page 36: Pranking with Ranking€¦ · I Ranking Rule (H) : Mapping from instances to ranks, Rn!Y. The family of ranking rules considered here : w 2Rn and k thresholds : b 1 b 2 ::: b k =](https://reader034.vdocuments.net/reader034/viewer/2022050105/5f435bf97497cb07662b9f9d/html5/thumbnails/36.jpg)
Experiments
![Page 37: Pranking with Ranking€¦ · I Ranking Rule (H) : Mapping from instances to ranks, Rn!Y. The family of ranking rules considered here : w 2Rn and k thresholds : b 1 b 2 ::: b k =](https://reader034.vdocuments.net/reader034/viewer/2022050105/5f435bf97497cb07662b9f9d/html5/thumbnails/37.jpg)
Experiments
I ModelsI Multi-class Generalization of Perceptron (MCP) : kn
parameters : under-constrainedI Widrow Hoff Algorithm for Online Regression (WH): n
parameters : over-constrainedI PRank : n + k − 1 parameters : accurately constrained
I DatasetsI Synthetic datasetI EachMovie dataset-used for collaborative filtering tasksI Evaluation in batch setting- outperforms multi-class SVM, SVR
Figure 4: Time-averaged ranking-loss comparison of MCP,WH,PRank onthe synthetic dataset, EachMovie-100 and 200 datasets respectively
![Page 38: Pranking with Ranking€¦ · I Ranking Rule (H) : Mapping from instances to ranks, Rn!Y. The family of ranking rules considered here : w 2Rn and k thresholds : b 1 b 2 ::: b k =](https://reader034.vdocuments.net/reader034/viewer/2022050105/5f435bf97497cb07662b9f9d/html5/thumbnails/38.jpg)
ExperimentsI Models
I Multi-class Generalization of Perceptron (MCP) : knparameters : under-constrained
I Widrow Hoff Algorithm for Online Regression (WH): nparameters : over-constrained
I PRank : n + k − 1 parameters : accurately constrained
I DatasetsI Synthetic datasetI EachMovie dataset-used for collaborative filtering tasksI Evaluation in batch setting- outperforms multi-class SVM, SVR
Figure 4: Time-averaged ranking-loss comparison of MCP,WH,PRank onthe synthetic dataset, EachMovie-100 and 200 datasets respectively
![Page 39: Pranking with Ranking€¦ · I Ranking Rule (H) : Mapping from instances to ranks, Rn!Y. The family of ranking rules considered here : w 2Rn and k thresholds : b 1 b 2 ::: b k =](https://reader034.vdocuments.net/reader034/viewer/2022050105/5f435bf97497cb07662b9f9d/html5/thumbnails/39.jpg)
ExperimentsI Models
I Multi-class Generalization of Perceptron (MCP) : knparameters : under-constrained
I Widrow Hoff Algorithm for Online Regression (WH): nparameters : over-constrained
I PRank : n + k − 1 parameters : accurately constrainedI Datasets
I Synthetic datasetI EachMovie dataset-used for collaborative filtering tasksI Evaluation in batch setting- outperforms multi-class SVM, SVR
Figure 4: Time-averaged ranking-loss comparison of MCP,WH,PRank onthe synthetic dataset, EachMovie-100 and 200 datasets respectively
![Page 40: Pranking with Ranking€¦ · I Ranking Rule (H) : Mapping from instances to ranks, Rn!Y. The family of ranking rules considered here : w 2Rn and k thresholds : b 1 b 2 ::: b k =](https://reader034.vdocuments.net/reader034/viewer/2022050105/5f435bf97497cb07662b9f9d/html5/thumbnails/40.jpg)
ExperimentsI Models
I Multi-class Generalization of Perceptron (MCP) : knparameters : under-constrained
I Widrow Hoff Algorithm for Online Regression (WH): nparameters : over-constrained
I PRank : n + k − 1 parameters : accurately constrainedI Datasets
I Synthetic datasetI EachMovie dataset-used for collaborative filtering tasksI Evaluation in batch setting- outperforms multi-class SVM, SVR
Figure 4: Time-averaged ranking-loss comparison of MCP,WH,PRank onthe synthetic dataset, EachMovie-100 and 200 datasets respectively
![Page 41: Pranking with Ranking€¦ · I Ranking Rule (H) : Mapping from instances to ranks, Rn!Y. The family of ranking rules considered here : w 2Rn and k thresholds : b 1 b 2 ::: b k =](https://reader034.vdocuments.net/reader034/viewer/2022050105/5f435bf97497cb07662b9f9d/html5/thumbnails/41.jpg)
Key takeaways
1. The ranking problem is a structured prediction task becauseof the total order between the different ratings.
2. Online algorithm for ranking problem via projections andconservative update of the projection’s direction and thethreshold values.
3. Experiments indicate this algorithm performs better thanregression and classification models for ranking tasks.
![Page 42: Pranking with Ranking€¦ · I Ranking Rule (H) : Mapping from instances to ranks, Rn!Y. The family of ranking rules considered here : w 2Rn and k thresholds : b 1 b 2 ::: b k =](https://reader034.vdocuments.net/reader034/viewer/2022050105/5f435bf97497cb07662b9f9d/html5/thumbnails/42.jpg)
Key takeaways
1. The ranking problem is a structured prediction task becauseof the total order between the different ratings.
2. Online algorithm for ranking problem via projections andconservative update of the projection’s direction and thethreshold values.
3. Experiments indicate this algorithm performs better thanregression and classification models for ranking tasks.
![Page 43: Pranking with Ranking€¦ · I Ranking Rule (H) : Mapping from instances to ranks, Rn!Y. The family of ranking rules considered here : w 2Rn and k thresholds : b 1 b 2 ::: b k =](https://reader034.vdocuments.net/reader034/viewer/2022050105/5f435bf97497cb07662b9f9d/html5/thumbnails/43.jpg)
Key takeaways
1. The ranking problem is a structured prediction task becauseof the total order between the different ratings.
2. Online algorithm for ranking problem via projections andconservative update of the projection’s direction and thethreshold values.
3. Experiments indicate this algorithm performs better thanregression and classification models for ranking tasks.
![Page 44: Pranking with Ranking€¦ · I Ranking Rule (H) : Mapping from instances to ranks, Rn!Y. The family of ranking rules considered here : w 2Rn and k thresholds : b 1 b 2 ::: b k =](https://reader034.vdocuments.net/reader034/viewer/2022050105/5f435bf97497cb07662b9f9d/html5/thumbnails/44.jpg)
Key takeaways
1. The ranking problem is a structured prediction task becauseof the total order between the different ratings.
2. Online algorithm for ranking problem via projections andconservative update of the projection’s direction and thethreshold values.
3. Experiments indicate this algorithm performs better thanregression and classification models for ranking tasks.
![Page 45: Pranking with Ranking€¦ · I Ranking Rule (H) : Mapping from instances to ranks, Rn!Y. The family of ranking rules considered here : w 2Rn and k thresholds : b 1 b 2 ::: b k =](https://reader034.vdocuments.net/reader034/viewer/2022050105/5f435bf97497cb07662b9f9d/html5/thumbnails/45.jpg)
Further Reading
Types of Ranking Algorithms:
I Point-wise Approaches - PRanking
I Pair-wise Approaches - RankSVM, RankNet, Rankboost
I List-wise Approaches - SVMmap, AdaRank, SoftRank
References:
I Liu, Tie-Yan. Learning to rank for information retrieval.Foundations and Trends R© in Information Retrieval 3.3(2009): 225-331.
I Agarwal, Shivani, and Partha Niyogi. Generalization boundsfor ranking algorithms via algorithmic stability. Journal ofMachine Learning Research 10.Feb (2009): 441-474.
![Page 46: Pranking with Ranking€¦ · I Ranking Rule (H) : Mapping from instances to ranks, Rn!Y. The family of ranking rules considered here : w 2Rn and k thresholds : b 1 b 2 ::: b k =](https://reader034.vdocuments.net/reader034/viewer/2022050105/5f435bf97497cb07662b9f9d/html5/thumbnails/46.jpg)
Further Reading
Types of Ranking Algorithms:
I Point-wise Approaches - PRanking
I Pair-wise Approaches - RankSVM, RankNet, Rankboost
I List-wise Approaches - SVMmap, AdaRank, SoftRank
References:
I Liu, Tie-Yan. Learning to rank for information retrieval.Foundations and Trends R© in Information Retrieval 3.3(2009): 225-331.
I Agarwal, Shivani, and Partha Niyogi. Generalization boundsfor ranking algorithms via algorithmic stability. Journal ofMachine Learning Research 10.Feb (2009): 441-474.
![Page 47: Pranking with Ranking€¦ · I Ranking Rule (H) : Mapping from instances to ranks, Rn!Y. The family of ranking rules considered here : w 2Rn and k thresholds : b 1 b 2 ::: b k =](https://reader034.vdocuments.net/reader034/viewer/2022050105/5f435bf97497cb07662b9f9d/html5/thumbnails/47.jpg)
Further Reading
Types of Ranking Algorithms:
I Point-wise Approaches - PRanking
I Pair-wise Approaches - RankSVM, RankNet, Rankboost
I List-wise Approaches - SVMmap, AdaRank, SoftRank
References:
I Liu, Tie-Yan. Learning to rank for information retrieval.Foundations and Trends R© in Information Retrieval 3.3(2009): 225-331.
I Agarwal, Shivani, and Partha Niyogi. Generalization boundsfor ranking algorithms via algorithmic stability. Journal ofMachine Learning Research 10.Feb (2009): 441-474.