improving user experience in recommender systems

49
Improving user experience in recommender systems How latent feature diversification can decrease choice difficulty and improve choice satisfaction Martijn Willemsen Talk at Institute for Software technology, Nov 3, 2015 Graz University of Technology

Upload: martijn-willemsen

Post on 07-Jan-2017

479 views

Category:

Engineering


3 download

TRANSCRIPT

Page 1: Improving user experience in recommender systems

Improving user experience in

recommender systems

How latent feature diversification

can decrease choice difficulty and

improve choice satisfaction

Martijn Willemsen

Talk at Institute for Software

technology, Nov 3, 2015

Graz University of Technology

Page 2: Improving user experience in recommender systems

Information overload

Recommendation

Page 3: Improving user experience in recommender systems

Choice Overload in Recommenders

Recommenders reduce information

overload…

But large personalized sets cause choice

overload!

Top-N of all highly ranked items

What should I choose?

These are all very attractive!

Page 4: Improving user experience in recommender systems

Choice Overload

Seminal example of choice overload

Satisfaction decreases with larger sets as increased

attractiveness is counteracted by choice difficulty

More attractive

3% sales

Less attractive

30% salesHigher purchase

satisfaction

From Iyengar and Lepper (2000)

Page 5: Improving user experience in recommender systems

Choice Overload in Recommenders(Bollen, Knijnenburg, Willemsen & Graus, RecSys 2010)

perceived recommendation

varietyperceived recommendation

quality

Top-20vs Top-5 recommendations

movie

expertisechoice

satisfaction

choice

difficulty

+

+

+

+

-+

.401 (.189)p < .05

.170 (.069)p < .05

.449 (.072)p < .001

.346 (.125)p < .01

.445 (.102)p < .001

-.217 (.070)p < .005

Objective System Aspects (OSA)

Subjective System Aspects (SSA)

Experience (EXP)

Personal Characteristics (PC)

Interaction (INT)

Lin-20vs Top-5 recommendations

+

+ - +

.172 (.068)p < .05

.938 (.249)p < .001

-.540 (.196)p < .01

-.633 (.177)p < .001

.496 (.152)p < .005

-0.1

0

0.1

0.2

0.3

0.4

0.5

Top-5 Top-20 Lin-20

Choice satisfaction

Page 6: Improving user experience in recommender systems

A solution: diversification

Tradeoff between similarity and diversity (Smyth &

McClave, 2001) in finding relevant items

Diversification remedies high similarity of Top-N lists

But diversification reduces the overall quality (accuracy) of the list

Many studies only look at data not at real users

Simulated users interact more efficiently with a diverse set (Bridge &

Kelly 2006)

Some studies look at how actual users perceive and

evaluate diversity

Ziegler et al. (2005): diversification based on external ontology

Diversity reduced the accuracy of the recommendations

But coverage and satisfaction increased!

Page 7: Improving user experience in recommender systems

Goal of the current research

Extend existing work by:

Diversification based on psychological mechanisms

Test the algorithm with real users and measure their perceptions and

experiences with the algorithm

Outline of the talk

Our user-centric evaluation framework

Psychology behind choice overload

Latent Feature diversification algorithm

Two user studies to test our diversification

Page 8: Improving user experience in recommender systems

User-Centric Framework

Computers Scientists (and marketing researchers) like to study

behavior…. (they hate asking the user or just cannot (AB tests))

Page 9: Improving user experience in recommender systems

User-Centric Framework

Psychologists and HCI people are mostly interested in experience…

Page 10: Improving user experience in recommender systems

User-Centric Framework

Though it helps to triangulate experience with behavior…

Page 11: Improving user experience in recommender systems

User-Centric Framework

Our framework adds the intermediate construct of perception that explains

why behavior and experiences changes due to our manipulations

Page 12: Improving user experience in recommender systems

User-Centric Framework

And adds personal

and situational

characteristics

Relations modeled

using factor analysis

and SEM

Knijnenburg, B.P., Willemsen, M.C., Gantner, Z., Soncu, H., Newell, C. (2012). Explaining

the User Experience of Recommender Systems. User Modeling and User-Adapted

Interaction (UMUAI), vol 22, p. 441-504 http://bit.ly/umuai

Page 13: Improving user experience in recommender systems

Latent feature diversification

from Psy to CS

Joint work with Mark Graus and

Bart Knijnenburg

Page 14: Improving user experience in recommender systems

Psychology behind choice overload

More options provide more benefits in terms of finding the

right option…

…but result in higher costs/effort

Objective effort:

More comparisons required

Cognitive effort:

Increased potential regret

Larger expectations for larger

sets

Many tradeoffs

Page 15: Improving user experience in recommender systems

Research on Choice overload

Choice overload is not omnipresent

Meta-analysis (Scheibehenne et al., JCR 2010)

suggests an overall effect size of zero

Choice overload stronger when:

No strong prior preferences

Little difference in attractiveness items

Prior studies did not control for

the diversity of the item set

Can we reduce choice difficulty and overload by using

personalized diversified item sets?

While controlling for attractiveness…

Page 16: Improving user experience in recommender systems

Diversification and attractiveness

Camera:

Suppose Peter thinks

resolution (MP) and Zoom

are equally important

user vector shows preference

direction

Equi-preference line:

Set of equally attractive options

(orthogonal on user vector)

Diversify on the equi-

preference line!

Page 17: Improving user experience in recommender systems

Choice Difficulty and Diversity

Larger sets are often more difficult because of the

increased uniformity of these sets (Fasolo et al., 2009;

Reutskaja et al., 2009)

Larger item sets have many

similar options

small inter-product distances

and small tradeoffs

High density!

Choice Difficulty related to

lack of justification 0

5

10

15

20

0 10 20R

eso

luti

on

(MP

)

Zoom

High Densitysmall tradeoffs

Page 18: Improving user experience in recommender systems

Choice difficulty and trade-offs

As item sets become more diverse (less dense) tradeoff

size increases

Tradeoffs are effortful…

give up one aspect for another

But can be justified very easily!

0

5

10

15

20

0 10 20Reso

luti

on

(MP

)

Zoom

High Densitysmall tradeoffs

0

5

10

15

20

0 10 20Reso

luti

on

(MP

)

Zoom

Low Densitylarge tradeoffs

Page 19: Improving user experience in recommender systems

Double Mediation Model for difficulty(Scholten and Sherman, JEP:G 2006)

U-shaped relation between diversity and difficulty:

Choosing from uniform set is

hard to justify but has no

difficult tradeoffs

Choosing from a diverse set

encompasses difficult tradeoffs

but is easy to justify

Does this also apply to personalized item sets?

How can we apply this to recommender system

algorithms? What features to diversify on?D

ifficulty

diversity

uniform diverse

Page 20: Improving user experience in recommender systems

Features in Matrix Factorization

Latent Features as means

of diversification!

“Latent features are Preference

dimensions related to real world

concepts (e.g. Escapist/serious)”

(Koren, Bell and Volinsky,2009)

Users and items described as

vectors of latent features

Parallel to how choice sets are described in

MAUT(multi-attribute utility theory) in consumer

psychology

Page 21: Improving user experience in recommender systems

Explaining Matrix Factorization

Map users and items to a joint latent factor space of

dimensionality f

Each item is a vector qi

each user a vector pu

Predicted rating r of user u for item i:

How to find the dimensions?

SVD: singular value decomposition

u

T

iui pqr ˆ

Page 22: Improving user experience in recommender systems

Matrix Factorization algorithms

each user a vector pu Each item is a vector qi

Usu

al

Su

sp

ects

Tit

an

ic

Die

Hard

Go

dfa

ther

Jack

Dylan

Olivia

Mark

?

?

?

? ? ?

?

pu

Dim

1

Dim

2

Jack 3 -1

Dylan 1.4 .2

Olivia -2.5 -.8

Mark -2 -1.5

qi

Usu

al

Su

sp

ec

ts

Tit

an

ic

Die

Hard

Go

dfa

ther

Dim 1 1.6 -1 5 0.2

Dim 2 1 1 .3 -.2

Page 23: Improving user experience in recommender systems

Two-dimensional Latent feature space and

diversification

Jack

Mark

Olivia Dylan

Page 24: Improving user experience in recommender systems

Greedy Diversity Algorithm

10-dimensional MF model

Take personalized top-200

Low: closest to centroid

Greedy algorithm

Select K items with

highest inter-item distance

(using city-block)

Medium:

select maximally diverse from

100 items closest to centroid

High: from all items in top-200

Page 25: Improving user experience in recommender systems

Greedy Diversity Algorithm

10-dimensional MF model

Take personalized top-200

Low: closest to centroid

Greedy algorithm

Select K items with

highest inter-item distance

(using city-block)

Medium:

select maximally diverse from

100 items closest to centroid

High: from all items in top-200

Page 26: Improving user experience in recommender systems

The algorithm

Measure of density: AFSR

Density/tradeoffs on the features: capture how items are distributed

in the feature space.

Average Factor Score Range (AFSR) based on the density metric

used by Fasolo et al. (2009)

X is set of items i

D is number of features

Captures the distribution of items in the feature space and their

tradeoffs better than standard similarity measures

Page 27: Improving user experience in recommender systems

Selection of initial set

Top-200 was selected as a balanced initial set:

Large differences in AFSR scores for the 3 levels of diversification

High average predicted rating: 4.48

Range in predicted ratings (0.546) lower than error of MF model:

MAE = .656 (RMSE = .854)

Final check: How does attractiveness vary within

and between the sets?

more diverse sets are by nature more likely to capture high-ranked

items…

Set Size Diversification Rating AFSR

5

Low 4.505 0.295

Medium 4.529 0.634

High 4.561 1.210

20

Low 4.537 0.586

Medium 4.558 1.005

High 4.604 1.615

Page 28: Improving user experience in recommender systems

System characteristics

MF recommender based on MyMedia project

10M MovieLens dataset: movies from 1994

5.6M ratings for 70k users and 5.4k movies

RMSE of 0.854, MAE of 0.656

Movies shown with title and predicted rating:

hovering the mouse over the title reveals additional information:

short synopsis, cast, director and image

Page 29: Improving user experience in recommender systems

Study 1a: Check diversification algorithm

Does our diversification affects the subjective experiences

with the recommendations?

Do people perceive the diversity?

Does it affects attractiveness and choice difficulty?

Each participant inspects 3 lists (low, mid and high

diversification), order counterbalanced

No choices made from the list

Page 30: Improving user experience in recommender systems

Study 1a: design/procedure

Pre-questionnaire

Personal characteristics

Rating task to train the system (10 ratings)

Assess three lists of recommendations

Within subjects: low / mid / high diversification

Between subjects: number of items (5,10,15,20,25)

After each list we measured:

Perceived Diversity & Attractiveness

Expected Trade-Off Difficulty & Choice Difficulty

Page 31: Improving user experience in recommender systems

Study 1a: Participants and Manipulation

checks

97 Participants from an online database

Paid for participation

Mean age: 29 years, 52 females and 45 males

Low, medium and high diversification

differed in the feature score range

Average predicted ratings of the sets were not different!

diversity

Average AFSR

(SE)

Avg. Predicted

rating

(SE)

Low 0.959 (0.015) 4.486 (0.042)

Medium 1.273 (0.016) 4.486 (0.041)

High 1.744 (0.024) 4.527 (0.039)

Page 32: Improving user experience in recommender systems

Study 1a: Structural Equation Model

Page 33: Improving user experience in recommender systems

Study 1a: Structural Equation Model

0

0.1

0.2

0.3

0.4

0.5

0.6

0.7

low mid high

sta

nd

ard

ize

d s

co

re

diversification

attractiveness diversity

-1

-0.8

-0.6

-0.4

-0.2

0

0.2

low mid high

sc

ale

dif

fere

nc

e

diversification

choice diff. tradeoff diff.

Page 34: Improving user experience in recommender systems

Study 1a: Conclusions & Discussion

Diversifying on latent features

Increases attractiveness/diversity

Reduces trade-off difficulty (high)

Reduces choice difficulty (linearly)

No evidence for U-shaped difficulty model

High diversity does not result in trade-off conflicts

(perhaps due to the nature of the domain/MF?)

No effect of number on items

Small sets benefit as much from diversification

Diversification on MF features seems

promising to increase attractiveness!

-1

-0.8

-0.6

-0.4

-0.2

0

0.2

low mid high

sc

ale

dif

fere

nc

e

diversification

choice diff. tradeoff diff.

Page 35: Improving user experience in recommender systems

Study 1b: No choice satisfaction

In Study 1a, no actual choice was made

Explains limited effect of number of items

We could not measure choice satisfaction or justification-based

processes

Diversification and list length as two factors in a new

experiment with choice (and choice satisfaction)

Item size: 5, 10 and 20

Low and high diversity (no medium)

We expect choice difficulty to be more prominent for low

diversity sets

Page 36: Improving user experience in recommender systems

Study 1b: design/procedure

Pre-questionnaire

Personal characteristics

Rating task to train the system (10 ratings)

Choose one item from a list of recommendations

Between subjects: 2 levels (low / high diversification)

X 3 lengths (5, 10 or 20 items)

Afterwards we measured:

Perceived Diversity & Attractiveness

Choice Difficulty and Choice satisfaction

87 Participants from an online database

Paid for participation Mean age: 29Y, 41F, 46M

Page 37: Improving user experience in recommender systems

Questionnaire-items

Perceived recommendation diversity

5 items, e.g. “The list of movies was varied”

Perceived recommendation attractiveness

5 items, e.g. “The list of recommendations was attractive”

Choice satisfaction

6 items, e.g. “I think I would enjoy watching the chosen movie”

Choice difficulty

5 items, e.g.: “It was easy to select a movie”

Page 38: Improving user experience in recommender systems

Study 1b: Structural Equation Model

Page 39: Improving user experience in recommender systems

Study 1b: Structural Equation Model

-0.5

0.0

0.5

1.0

1.5

2.0

5 10 20

sta

nd

ard

ized

sco

relist length

Diversity

low diversityhigh diversity

0.0

0.5

1.0

1.5

2.0

2.5

5 10 20

sta

nd

ard

ize

d s

co

re

list length

Satisfactionlow diversity high diversity

Page 40: Improving user experience in recommender systems

Study 1b: Results

Long list is more difficult (cognitive and objective effort

(hovers)) but also more satisfying in itself

Diversity influences choice satisfaction in important ways

Diversity increases attractiveness and reduces difficulty

These can increase satisfaction (but only for short lists)

Our diversification improved the 5-item list

5 diverse items are as satisfactory as

10 or 20 items and less difficult!

Less effort needed (hovers)

Using latent feature diversification

one does not need long item lists…0.0

0.5

1.0

1.5

2.0

2.5

5 10 20

sta

nd

ard

ize

d s

co

re

list length

Satisfactionlow diversity high diversity

Page 41: Improving user experience in recommender systems

But…

Our studies show that low or high diversification from the

centroid of Top-200 works

However, these sets were optimized for diversity, not for prediction

accuracy

Most item sets thus not contain the best predicted items

So how does this compare to standard Top-N lists?

Slight modification of our algorithm: diversify starting from the best

predicted item in the top-N set (Top-1) rather than the centroid

Diversification and list length as two factors in a choice

overload experiment

list sizes: 5 and 20

Diversification: none (top 5/20), medium, high

Page 42: Improving user experience in recommender systems

Properties of the item sets

Set size diversification Avg. AFSR Avg. Rank

5

High (α = 1) 1.380 78.284

Medium (α = 0.3) 1.096 7.484

Top-N (α = 0) 0.774 3.000

20

High (α = 1) 1.793 89.380

Medium (α = 0.3) 1.486 17.849

Top-N (α = 0) 1.270 10.500

Algorithm balances between max diversity and highest rank

Every iteration: weigh (1-α) highest rank

against highest diversification (α)

α=0: top-N, α=1: max diverse.

(α=0.3 gives good medium diversity)

Page 43: Improving user experience in recommender systems

Design/procedure Study 2

159 Participants from an online database

Rating task to train the system (15 ratings)

Choose one item from a list of recommendations

Between subjects: 3 levels of diversification, 2 lengths

Afterwards we measured:

Perceptions: Perceived Diversity & Attractiveness

Experience: Choice Difficulty and Choice satisfaction

Behavior: total views / unique items considered

Page 44: Improving user experience in recommender systems

Structural Equation Model

Page 45: Improving user experience in recommender systems

Perceived Diversity & attractiveness

Perceived Diversity increases with

Diversification

Similarly for 5 and 20 items

Perc. Diversity increases attractiveness

Perceived difficulty goes down with

diversification

Perceived attractiveness goes up

with diversification

Diverse 5 item set excels…

Just as satisfying as 20 items

Less difficult to choose from

Less cognitive load…!

-0.5

0

0.5

1

1.5

none med high

sta

nd

ard

ize

d s

co

re

diversification

Perc. Diversity

5 items

20 items

-0.2

0

0.2

0.4

0.6

0.8

1

none med highsta

nd

ard

ize

d s

co

re

diversification

Choice Satisfaction

5 items

20 items

Page 46: Improving user experience in recommender systems

Choice Characteristics

Chosen option (mean and std. err)

SetDiversity

List Position Rating Rank

5 items

None (top 5) 3.60 (0.27) 4.51 (0.07) 3.60 (0.27)

Medium 4.41 (0.59) 4.41 (0.07) 14.52 (5.37)

High 4.19 (0.27) 4.30 (0.07) 77.59 (12.76)

20 items

None (top 20) 10.15 (0.92) 4.45 (0.05) 10.15 (0.92)

Medium 10.33 (1.18) 4.40 (0.08) 17.7 (2.68)

high 9.93 (1.07) 4.16 (0.07) 72.22 (11.84)

With higher diversity, no difference in position of chosen option

Resulting in less ‘optimal’ choice in terms of predicted rating

Without a reduction in choice satisfaction!

Page 47: Improving user experience in recommender systems

Conclusions

Reducing Choice difficulty and overload

Diversity reduces choice difficulty

Less uniform sets are easier to choose from

Latent feature diversification easy to implement

Diversity can improve choice satisfaction

Even when the diversified list has movies with lower

predicted ratings than standard top-N lists

No need for larger item sets

Offering personalized diversified small items sets might be the key

to help decision makers cope with too much choice!

Psychological theory can inform how to improve the

output of Recommender algorithms

Page 48: Improving user experience in recommender systems

What you should take away…

Psychological theory can inform new ways of diversifying

algorithm output or eliciting preferences

But also: working with recommenders and algorithms we could

enhance psychological theory: personalized item sets gives control

User-centric evaluation helps to assess the effectiveness

Lot of work… and we need user studies…

But linking subjective to objective measures might help future

studies that cannot do user studies

User-centric framework allows us to understand WHY

particular approaches work or not

Concept of mediation: user perception helps understanding..

Page 49: Improving user experience in recommender systems

Questions?

Contact:

Martijn Willemsen

@MCWillemsen

[email protected]

www.martijnwillemsen.nl

Thanks to my co-authors:

Mark Graus

Bart Knijnenburg