bandit based optimization of multiple objectiveson a music

10
Bandit based Optimization of Multiple Objectives on a Music Streaming Platform Rishabh Mehrotra † Spotify, London [email protected] Niannan Xue β€ βˆ— Imperial College London [email protected] Mounia Lalmas Spotify, London [email protected] ABSTRACT Recommender systems powering online multi-stakeholder plat- forms often face the challenge of jointly optimizing for multiple objectives, in an attempt to efficiently match suppliers and con- sumers. Examples of such objectives include user behavioral metrics (e.g. clicks, streams, dwell time, etc), supplier exposure objectives (e.g. diversity) and platform centric objectives (e.g. promotions). Jointly optimizing multiple metrics in online recommender sys- tems remains a challenging task. Recent work has demonstrated the prowess of contextual bandits in powering recommendation systems to serve recommendation of interest to users. This paper aims at extending contextual bandits to multi-objective setting so as to power recommendations in a multi-stakeholder platforms. Specifically, in a contextual bandit setting, we learn a recommen- dation policy that can optimize multiple objectives simultaneously in a fair way. This multi-objective online optimization problem is formalized by using the Generalized Gini index (GGI) aggrega- tion function, which combines and balances multiple objectives together. We propose an online gradient ascent learning algorithm to maximise the long-term vectorial rewards for different objectives scalarised using the GGI function. Through extensive experiments on simulated data and large scale music recommendation data from Spotify, a streaming platform, we show that the proposed algorithm learns a superior policy among the disparate objectives compared with other state-of-the-art approaches. ACM Reference Format: Rishabh Mehrotra † , Niannan Xue β€ βˆ— , and Mounia Lalmas. 2020. Bandit based Optimization of Multiple Objectives on a Music Streaming Platform. In Proceedings of the 26th ACM SIGKDD Conference on Knowledge Discovery and Data Mining (KDD ’20), August 23–27, 2020, Virtual Event, CA, USA. ACM, New York, NY, USA, 10 pages. https://doi.org/10.1145/3394486.3403374 1 INTRODUCTION Platform ecosystems have witnessed an explosive growth by facili- tating efficient interactions between multiple stakeholders, includ- ing e.g. buyers and retailers (Amazon), guests and hosts (AirBnb), † Equal contribution authors. * This work was done while the author was an intern at Spotify. Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]. KDD ’20, August 23–27, 2020, Virtual Event, CA, USA Β© 2020 Association for Computing Machinery. ACM ISBN 978-1-4503-7998-4/20/08. . . $15.00 https://doi.org/10.1145/3394486.3403374 riders and drivers (Uber), and listeners and artists (Spotify). A large number of such platforms rely on machine learning powered match- ing engines connecting consumers with suppliers by acting as a central platform, thereby finding the right fit and efficiently medi- ating economic transactions between the two sides. Recent advancements in understanding, interpreting and lever- aging user behavioral signals have enabled such recommender systems to be optimized for many different user centric objectives, including clicks [14], dwell time [43], session length time [13], streaming time, conversion [32] among others. Beyond user-centric objectives, platforms can additionally optimize their models for dif- ferent stakeholder objectives, including exposure, fairness, diversity, promotion and revenue metrics. For example, a recommender system powering a multi-stakeholder platform (e.g. Amazon, Uber) would not only aim at serving recom- mendations maximising user satisfaction, but also would optimise for fair exposure to retailers and suppliers, alongside maximising revenue-related objectives [25].Even in the case of user-centric rec- ommendation systems, predicting user interest alone is not enough to ensure user satisfaction, and notions such as engagement, nov- elty and diversity greatly enhance user experiences [8] and should be promoted as well. This motivates the need for developing multi- objective recommendation models capable of leveraging multiple user-centric as well as other stakeholder objectives. In this work we consider the task of designing multi-objective rec- ommender systems for online multi-stakeholder platforms. Specif- ically, we propose a multi-objective formulation of a contextual bandit based recommender system. Multi-armed bandits (MAB) and their variants (contextual bandits, dueling bandits, etc.) are increas- ingly popular in sequential decision making scenarios, wherein the system faces a dilemma over whether to explore to discover more about user preferences, or to exploit current knowledge and serve recommendations that pique user interests. In the contextual MAB, different from the classical MAB, the system observes a context (side information) at the beginning of each round, which gives a hint about the expected arm rewards in that round. Such contextual bandit based models are better adept at handling uncertainty of relevance of items and new information (new users, new items) than traditional recommender systems. In traditional bandit settings, only a single scalar feedback is observed after an action is performed. However, multi-objective sys- tems necessarily involve joint optimisation of a number of different criteria simultaneously. While a lot of research has gone in devel- oping contextual bandit solutions [18, 33], multi-objective variants of contextual bandits have received relatively less attention. In this paper, we focus on developing multi-objective contextual bandit models for digital multi-stakeholder platforms. We focus

Upload: others

Post on 19-Nov-2021

3 views

Category:

Documents


0 download

TRANSCRIPT

Bandit based Optimization of Multiple Objectiveson a Music Streaming Platform

Rishabh Mehrotra†Spotify, London

[email protected]

Niannan Xueβ€ βˆ—Imperial College [email protected]

Mounia LalmasSpotify, [email protected]

ABSTRACTRecommender systems powering online multi-stakeholder plat-forms often face the challenge of jointly optimizing for multipleobjectives, in an attempt to efficiently match suppliers and con-sumers. Examples of such objectives include user behavioral metrics(e.g. clicks, streams, dwell time, etc), supplier exposure objectives(e.g. diversity) and platform centric objectives (e.g. promotions).Jointly optimizing multiple metrics in online recommender sys-tems remains a challenging task. Recent work has demonstratedthe prowess of contextual bandits in powering recommendationsystems to serve recommendation of interest to users. This paperaims at extending contextual bandits to multi-objective setting soas to power recommendations in a multi-stakeholder platforms.

Specifically, in a contextual bandit setting, we learn a recommen-dation policy that can optimize multiple objectives simultaneouslyin a fair way. This multi-objective online optimization problemis formalized by using the Generalized Gini index (GGI) aggrega-tion function, which combines and balances multiple objectivestogether. We propose an online gradient ascent learning algorithmto maximise the long-term vectorial rewards for different objectivesscalarised using the GGI function. Through extensive experimentson simulated data and large scale music recommendation data fromSpotify, a streaming platform, we show that the proposed algorithmlearns a superior policy among the disparate objectives comparedwith other state-of-the-art approaches.ACM Reference Format:Rishabh Mehrotra†, Niannan Xueβ€ βˆ—, and Mounia Lalmas. 2020. Bandit basedOptimization of Multiple Objectives on a Music Streaming Platform. InProceedings of the 26th ACM SIGKDD Conference on Knowledge Discoveryand DataMining (KDD ’20), August 23–27, 2020, Virtual Event, CA, USA.ACM,New York, NY, USA, 10 pages. https://doi.org/10.1145/3394486.3403374

1 INTRODUCTIONPlatform ecosystems have witnessed an explosive growth by facili-tating efficient interactions between multiple stakeholders, includ-ing e.g. buyers and retailers (Amazon), guests and hosts (AirBnb),†Equal contribution authors.* This work was done while the author was an intern at Spotify.

Permission to make digital or hard copies of all or part of this work for personal orclassroom use is granted without fee provided that copies are not made or distributedfor profit or commercial advantage and that copies bear this notice and the full citationon the first page. Copyrights for components of this work owned by others than ACMmust be honored. Abstracting with credit is permitted. To copy otherwise, or republish,to post on servers or to redistribute to lists, requires prior specific permission and/or afee. Request permissions from [email protected] ’20, August 23–27, 2020, Virtual Event, CA, USAΒ© 2020 Association for Computing Machinery.ACM ISBN 978-1-4503-7998-4/20/08. . . $15.00https://doi.org/10.1145/3394486.3403374

riders and drivers (Uber), and listeners and artists (Spotify). A largenumber of such platforms rely on machine learning powered match-ing engines connecting consumers with suppliers by acting as acentral platform, thereby finding the right fit and efficiently medi-ating economic transactions between the two sides.

Recent advancements in understanding, interpreting and lever-aging user behavioral signals have enabled such recommendersystems to be optimized for many different user centric objectives,including clicks [14], dwell time [43], session length time [13],streaming time, conversion [32] among others. Beyond user-centricobjectives, platforms can additionally optimize their models for dif-ferent stakeholder objectives, including exposure, fairness, diversity,promotion and revenue metrics.

For example, a recommender system powering amulti-stakeholderplatform (e.g. Amazon, Uber) would not only aim at serving recom-mendations maximising user satisfaction, but also would optimisefor fair exposure to retailers and suppliers, alongside maximisingrevenue-related objectives [25].Even in the case of user-centric rec-ommendation systems, predicting user interest alone is not enoughto ensure user satisfaction, and notions such as engagement, nov-elty and diversity greatly enhance user experiences [8] and shouldbe promoted as well. This motivates the need for developing multi-objective recommendation models capable of leveraging multipleuser-centric as well as other stakeholder objectives.

In this workwe consider the task of designingmulti-objective rec-ommender systems for online multi-stakeholder platforms. Specif-ically, we propose a multi-objective formulation of a contextualbandit based recommender system. Multi-armed bandits (MAB) andtheir variants (contextual bandits, dueling bandits, etc.) are increas-ingly popular in sequential decision making scenarios, wherein thesystem faces a dilemma over whether to explore to discover moreabout user preferences, or to exploit current knowledge and serverecommendations that pique user interests. In the contextual MAB,different from the classical MAB, the system observes a context(side information) at the beginning of each round, which gives ahint about the expected arm rewards in that round. Such contextualbandit based models are better adept at handling uncertainty ofrelevance of items and new information (new users, new items)than traditional recommender systems.

In traditional bandit settings, only a single scalar feedback isobserved after an action is performed. However, multi-objective sys-tems necessarily involve joint optimisation of a number of differentcriteria simultaneously. While a lot of research has gone in devel-oping contextual bandit solutions [18, 33], multi-objective variantsof contextual bandits have received relatively less attention.

In this paper, we focus on developing multi-objective contextualbandit models for digital multi-stakeholder platforms. We focus

on the contextual bandit problem where the reward is assumed tobe a noisy linear function of context, i.e. side information, not forjust a single objective, but for a number of objectives. We proposea multi-objective contextual bandit model based on GeneralizedGini Index (GGI) [5] by introducing a model that assimilates contex-tual information and optimizes for a number of different, includingcompeting objectives. Such an extension of bandits is non-trivialbecause there is no longer a fixed mean feedback for any specificarm. The goal of the proposed GGI based optimization is to be bothefficient, i.e., minimize the cumulative cost for each objective, andfair, i.e., balance the different objectives. We propose an online gra-dient ascent method to maximise a scalarisation of the accumulatedreward over several objectives.

We consider the case of Spotify - an online music streamingplatform, and define user-centric satisfaction objectives as well assupplier-centric diversity and promotional objectives. We presentcorrelation analysis across these different user- and supplier-centricobjectives to motivate the need for multi-objective modeling. Weperform large scale experiments on simulated data as well as realworld user interaction data, and demonstrate that our proposedalgorithm performs better than a number of established baselinesand state-of-the-art multi-objective models.

Experiments with multiple user-centric objectives suggest thatoptimizing for multiple interaction metrics performs better for eachmetric than optimizing single metric. Beyond multiple user-centricobjectives, experiments around adding promotional objectives (e.g.gender promotion) demonstrate that the platform can obtain gainin such promotional-centric or supplier-centric objectives withoutsevere loss in user centric objectives. Our findings have implicationson the design of recommender systems powering online digitalmulti-stakeholder platforms and motivates future work, which webriefly discuss at the end of the paper.

2 BACKGROUND AND RELATEDWORKMulti-objectiveOptimization. Multi-objective optimisation (akaMOO) is a well studied subject in operation research and machinelearning, with applications in multi-agent systems [44], reinforce-ment learning [10] and, recently, multi-armed bandits [5]. Tradi-tional MOO approaches aimed at seeking the Pareto front directly.A more practical approach involes converting the vectorial objec-tives into a single one using some aggregation function, a processknown as scalarisation [21]. Popular aggregation functions includethe weighted sum [15] and the regularized maximin fairness pol-icy [44]. We use a concrete example of the generalised Gini func-tion in this paper, which can be seen as a generalisation to bothtotal ordering and linear priorities [4]. Furthermore, unlike pastapproaches, which dealt with MOO in traditional supervised learn-ing setup, the focus on this work is on more interactive, online,adaptive learning setting based on user feedback.

MOO in Search & Recommender Systems. MOO has been veryeffective for various business motivated applications in search andrecommender systems, including click shaping [2, 3], email volumeoptimization [11] and serving recommendations with capacity con-straints [6]. Earlier efforts also included extensions of collaborativefiltering for MOO [12] and multi-criterion user modeling [17]. In

search and information retrieval, prior work has leveraged and de-veloped MOO techniques for optimizing novelty and diversity [30]and learning to rank [35]. Recent approaches on fairness in rec-ommendation have also considered multi-objective models [40]. Arecent tutorial also covered such applications in detail [24].

A key difference between our approach and aforementionedMOO approaches is our explicit focus on bandit based settings,which perform better than traditional recommendation approachesat handling uncertainty of item relevance and new information.Our proposed model builds on top of such bandit based approachesand extends the contextual bandits to a multi-objective setting.

Bandits & Multi-objective Bandits. For multi-armed bandits,there has been extensive work to address the multi-objective na-ture. A comparison of different algorithms using multiple regretevaluation metrics can be found in [9]. Knowledge gradient [29]has been used to explore the Pareto optimal arms. Algorithms suchas Hierarchical optimistic optimisation strategy [26], Thompsonsampling [42], and combination of bi-objective optimization withcombinatorial bandits [38]. have also been proposed. While theseapproaches tackle the MOO problem in vanilla bandit setting, wefocus on the case of contextual bandits, wherein an additional con-text vector is observed at each iteration, which encodes contextualinformation about users and content. The proposed method extendsmulti-objective bandit models to a contextual bandit setting.

While a lot of research has gone in developing multi-objectivemulti-armed bandits, multi-objective variants of contextual banditshave received less attention. Only algorithms in the restrictivesettings of similarity information [37] and a dominant objective [36]exist, and rely on the assumption of access to distances betweencontext-arm pairs to the distances between expected rewards ofthese pairs. This assumption is fairly restrictive in typical industrialsettings, where rewards are derived based on user behavioral data.

3 OBJECTIVES & STAKEHOLDERSMachine learning systems powering modern recommender sys-tems are optimized for and evaluated upon an increasing numberof objectives and metrics. For example, user centric recommendersystems such as e-commerce portals optimize for different proxiesof user satisfaction, including clicks, dwell time and conversion;whereas multi-stakeholder platforms optimize metrics for its vari-ous stakeholders, including guests/hosts (Airbnb), buyers/sellers(Amazon) and listeners/artists (Spotify).

In this paper, we discuss scenarios wherein the system needs tojointly optimize for multiple metrics and propose a bandit model asa solution. In this section, we begin by motivating an industrial use-case of multi-objective modelling and present data driven analysisthat motivate the need for multi-objective modelling for recom-mender systems. We then briefly describe the contextual banditformulation of the recommender problem, which serves as the basemodel on top of which the proposed model is built.

Data Context. We consider the specific use case of a global mu-sic streaming platform where users listen to music from differentartists. The recommendation system recommends a set of tracks(i.e. playlists containing songs) to the user, each of which couldcome from different artists. Different sets have varying degree of

stre

am

click

s

song

s pla

yed

long

stre

am

dive

rsity

(g)

prom

otio

n

stre

amcli

cks

song

s pla

yed

long

stre

amdi

vers

ity (g

)pr

omot

ion

1 0.62 0.98 0.72 -0.05 -0.12

1 0.64 0.87 -0.03 -0.09

1 0.73 -0.04 -0.11

1 -0.08 -0.12

1 0.04

1 0.8

0.4

0.0

0.4

0.8

Figure 1: Heatmap of correlations between different objectives.

relevance to user’s interests, and users could be satisfied with therecommended set to varying extent.

3.1 Objective DefinitionsOften in user centric systems, system designers have access tomultiple implicit signals from the rich fine-grained user interactioninformation logged in behavior logs, which give rise to a number ofuser-centric objectives. For example, in a case of music streamingservices, a system could optimize for clicks, streams, number ofsongs played or other user engagement metrics. Often, such metricsare correlated, and optimizing for one would inherently lead themodel to improve other correlated metrics. However, this needs notbe true: an objective might be un-correlated or negatively correlatedwith user satisfaction metrics, and there exist strict trade-off inoptimizing one against the other. For example, recent work hasdemonstrated that optimizing for relevance hurts diversity [25]and vice-versa, thereby motivating the need for development ofmulti-objective models that optimize multiple metrics.

Specifically, for the case of music streaming platforms, systemdesigners can optimize for a number of user centric objectives,such as clicks, track stream, number of tracks played, long durationstreams, among others. Furthermore, when considering other stake-holders in a music streaming platform, additional objectives surface,including diversity centric and promotion centric objectives.

We conjecture that optimizing for some metrics might hurt othermetrics. For example, promoting certain artists in a recommenda-tion setting might annoy users whose taste profiles do not matchthe artist, and hence hurt user satisfaction metrics. To better un-derstand this interplay between different objectives, we consider arandom sample of user streaming data, and estimate and analyzedifferent user centric and artist centric objectives. We consider threeuser centric objectives: duration of streamed songs (stream), clickson playlists (clicks) and number of songs played. Additionally, weconsider two artist and platform centric objectives. First, diversity(g)quantifies the gender diversity present in the recommended set,which is computed as the percentage tracks in a set whose mainartist is a non-male artist. Second, promotion objective assumes ascenario wherein the platform intends to promote some niche artiststo users via its recommendations. These diversity and promotional

objectives find overlap with other corresponding metrics in otherplatforms, e.g. diversity and promotion of retailers (e-commerceplatforms), funding campaigns (crowdfunding platforms), hosts(AirBnb), among others.

Figure 1 shows the heatmap of the correlation across differentobjectives. We observe that different objectives are correlated todifferent extent, with user centric objectives strongly positivelycorrelated, while gender-diversity and promotion objectives areweakly negatively correlated with user centric objectives. This anal-ysis highlights the fact that optimizing for some objectives mighthave a detrimental effect on other objectives. In cases where therelationship between metrics of different stakeholders are simpleand correlated, optimizing one would result in gains in the othermetric. But more often than not, balancing in multi-stakeholderplatform entails a subtle trade-off between objectives. A recom-mender system built for optimizing a single metric is ill-suited in amulti-metric multi-stakeholder platform setting, as discussed next.

Before delving into a multiple metric optimization, we first de-scribe a recommendation approach based on contextual bandits,which is widely used to optimize a single user satisfaction metric.

3.2 Contextual Bandit based RecommendersWe formalize the recommendation problem as a combinatorial con-textual bandit problem, wherein the recommender system repeat-edly interacts with consumers as follows:

(1) the system observes a context (J );(2) based on the context, the system chooses an action π‘Ž ∈ 𝐴,

from the space of K possible actions (sets to recommend);(3) given a context and an action, a reward π‘₯ ∈ [0, 1] is drawn

from distribution 𝐷 (π‘₯ |,J), with rewards in different roundsbeing independent, conditioned on contexts and actions.

While the context space can be infinite, composed of informationthe system has about user’s interests, item features and other fea-tures like time, location, the action space is finite. We next describethe context and actions used for the presented research.

Context: We leveraged a large number of context signals in therecommendation problem, including: (i) features of the user, suchas age range, gender, location, affinity to genres; (ii) features of theplaylist such as its artist, its (micro and macro) genres, diversity ofsongs, popularity; (iii) affinity between the user and the playlist,taking into account past interactions, such as streams, skips, likes,and saves; and (iv) other contextual information, such as the day ofthe week and the time of day.

Actions: Each action is composed of selecting a set to recommendto the user. In our case of music streaming, we assume a set basedrecommendation strategy with the user presented with a playlist(a collection of tracks), with each track coming from specific artist.

Rewards: In a traditional user-centric system, the observed re-ward will be based on how happy the user was with the recom-mendation served, and the goal of the model is to learn an armselection strategy that maximizes user satisfaction. Such an armselection strategy is focused on a single metric, one that is generallychosen as a proxy of user satisfaction. On the other hand, in a multi-stakeholder recommender system, vectorial rewards are observed,one corresponding to each objective, and the arm selection strategy

would be decided based on the strategy that optimises for each ofthese objectives. We develop such an approach in this paper.

Multi-objective optimization gives us a mathematically princi-pled solution for the trade-off among (often competing) objectives.To this end, we next propose a multi-objective contextual bandit tosolve this specific problem.

4 MULTI-OBJECTIVE CONTEXTUALBANDITS

To jointly optimize multiple objectives for recommending content,we propose a novel multi-objective contextual bandit algorithm.We begin by defining mathematical notations (Section 4.1) anddetail the problem set-up (Section 4.2). Section 4.3 describes thearm-selection strategy given multiple objectives and Section 4.4describes the complete algorithm.

4.1 NotationsPlain letters denote scalars; bold letters denote vectors and calli-graphic letters denote matrices, unless otherwise stated. 𝒂𝑖 repre-sents the 𝑖th element of 𝒂 andA 𝑗 represents the 𝑗 th column ofA. ∧is the logical conjunction operator between two conditions. Iπ‘šΓ—π‘šdenotes theπ‘š byπ‘š identity matrix. Projection onto support set Cis given by 𝚷C. βˆ₯𝒂βˆ₯ denotes the 𝑙2-norm of vector 𝒂. For a positivedefinite matrix A ∈ R𝑑×𝑑 , the weighted 𝑙2-norm of vector 𝒙 ∈ R𝑑is defined by βˆ₯𝒙 βˆ₯A =

βˆšπ’™π‘‡A𝒙 . |π‘Ž | is the absolute value of π‘Ž.

4.2 Problem SetupWe cast the multi-objective recommendation problem in terms ofa multi-arm contextual bandit setting. Assume that we play thebandit problem for a total of 𝑇 rounds, where each round corre-sponds to a user session wherein a playlist is recommended to theuser (i.e., one bandit arm is selected). For each bandit instance atround 𝑑 , we are given features J[𝑑 ] = (J[𝑑 ],1,Β· Β· Β· ,J[𝑑 ],𝐾 ) associatedwith the 𝐾 possible arms, where J[𝑑 ],𝑖 ∈ R𝑀 and𝑀 is the featurelength. Such features encode the current user specific context, andmay include features representing user taste profiles, historic in-teraction features and other contextual signals. An arm selectionstrategy corresponds to selecting a playlist to show to the usergiven observations about the contextual features in the session.Under the linear shared model, if arm π‘˜ is chosen at round 𝑑 , weobserve reward

𝒙 [𝑑 ] = J𝑇[𝑑 ],π‘˜πœ—

βˆ— + 𝜻 [𝑑 ] , (1)

where πœ—βˆ— ∈ R𝑀×𝐷 is a fixed unknown universal parameter and𝜻 [𝑑 ] ∈ R𝐷 is an independent random noise for each objective.

4.3 Arm Selection StrategyA strategy is a way to pick an arm at each round by examiningthe features for all arms. This strategy defines which playlist (froma collection of candidate playlists) is shown to the user for eachsession. We can then calculate a strategy’s average reward after 𝑇rounds, as

𝒙 [𝑇 ] =1𝑇

π‘‡βˆ‘π‘‘=1

𝒙 [𝑑 ] . (2)

Algorithm 1 MO-LinCB1: Input: 𝐾 arms, 𝐷 objectives, 𝑇 rounds, aggregation function 𝐺 (𝒙) ,

regularisation parameter _, learning rate [, gradient ascent iterations 𝐼2: Set A = _I𝑀×𝑀3: Set B𝑑 = 0, βˆ€π‘‘ ≀ 𝐷4: for 𝑑 = 1, 2, 3, . . . ,𝑇 do5: Observe 𝐾 features, J[𝑑 ],1, J[𝑑 ],2, Β· Β· Β· , J[𝑑 ],𝐾 ∈ R𝑀6: for 𝑑 = 1, 2, 3, Β· Β· Β· , 𝐷 do7: πœ— [𝑑 ],𝑑 = Aβˆ’1B𝑑8: for π‘˜ = 1, 2, 3, Β· Β· Β· , 𝐾 do9: 𝝁 [π‘˜ ] = J𝑇[𝑑 ],π‘˜πœ— [𝑑 ]

10: 𝜢 [𝑑 ] = (1/𝐾, Β· Β· Β· , 1/𝐾)11: for 𝑖 = 1, 2, 3, Β· Β· Β· , 𝐼 do12: 𝜢 [𝑑 ] = 𝚷A (𝜢 [𝑑 ] + [βˆ‡π‘“[𝑑 ] (𝜢 [𝑑 ] ))13: Choose arm π‘˜ according to 𝜢 [𝑑 ] and observe reward 𝒙 [𝑑 ]14: A = A + J[𝑑 ],π‘˜J𝑇[𝑑 ],π‘˜15: for 𝑑 = 1, 2, 3, . . . , 𝐷 do16: B𝑑 = B𝑑 + 𝒙 [𝑑 ],𝑑 Γ— J[𝑑 ],π‘˜

We follow the scalarisation approach to multi-obejctive optimi-sation, where one usually wants to compute the Pareto front, orsearch for a particular element of the Pareto front. In practice, itmay be costly (and even infeasible depending on the size of the so-lution space) to determine all solutions of the Pareto front. One maythen prefer to directly aim for a particular solution in the Paretofront. This problem is formalized as a single objective optimizationproblem, using an aggregation function. We employ a Gini indexbased aggregation function described in detail in Section 4.3.2.

4.3.1 Optimal Policy. For an aggregation function𝐺 (𝒙), we aimto seek a strategy such that 𝐺

(𝒙 [𝑇 ]

)is as large as possible, i.e.,

arms are selected which maximize the aggregation function.Rather than considering a strategy such that only a single arm

is decided at each round, we look for a strategy that, at each round,proposes a probability distribution, A = {𝜢 ∈ R𝐾 |βˆ‘πΎ

π‘˜=1 πœΆπ‘˜ =

1 ∧ 0 ≀ πœΆπ‘˜ ,βˆ€π‘˜ ≀ 𝐾}, according to which an arm (i.e. π›Όπ‘˜ ) is tobe drawn. That is, we consider mixed strategies. For example,we can find an optimal mixed strategy arm selection policy for asingle bandit problem with known mean feedback by solving thefollowing optimisation problem,

𝜢 βˆ— ∈ arg max𝜢 ∈A

𝐺

(πΎβˆ‘π‘˜=1

πœΆπ‘˜π [π‘˜ ]

). (3)

In other words, an arm with the highest mean reward is pulledmost frequently. Nonetheless, arms with less reward values are alsopulled sometimes. This allows the model to trade-off exploitationof known arms with exploration of potentially useful arms.

In the single objective case, arms are compared in terms oftheir means, which induce a total ordering over arms. In the multi-objective setting, we use a specific form of aggregation criterion tocompare arms, which we describe next.

4.3.2 Generalised Gini Function. The aggregation function al-lows the model to scalarize inputs from different objectives. Weconsider a specific example for it, the generalised Gini function

(GGF) [39]. GGF is a non-linear but concave function. It is a spe-cial case of the ordered weighted averaging (OWA) aggregationoperators [41], which preserves impartiality with respect to theindividual criterion. For a reward vector 𝒙 = (𝒙1, . . . , 𝒙𝐷 ), GGF isdefined as

πΊπ’˜ (𝒙) =π·βˆ‘π‘‘=1

π’˜π‘‘ (π’™πœŽ )𝑑 = π’˜π‘‡ π’™πœŽ , (4)

whereπ’˜1 > π’˜2 > Β· Β· Β· > π’˜π‘‘ > 0 and 𝜎 permutes the elements of 𝒙such that (π’™πœŽ )𝑖 ≀ (π’™πœŽ )𝑖+1. GGF is strictly monotonic, which meansthat a vector that maximises Equation 4 also lies on the Pareto frontfor direct optimisation of the multiple criteria and different weights(π’˜) correspond to different points on the frontier [28]. GGF exhibitsa fairness property under the Pigou-Dalton transfer: if 𝒙𝑖 < 𝒙 𝑗 ,then πΊπ’˜ (𝒙 β€²) > πΊπ’˜ (𝒙) for 𝒙 β€²

𝑖= 𝒙𝑖 + πœ– and 𝒙 β€²

𝑗= 𝒙 𝑗 βˆ’ πœ– where

πœ– < 𝒙 𝑗 βˆ’ 𝒙𝑖 and 𝒙 β€²π‘˜= π’™π‘˜ for π‘˜ β‰  𝑖, 𝑗 . In other words, an equitable

transfer of an arbitrarily small amount from a larger componentto a smaller component is always preferable. The effect of such atransfer is to balance a reward vector.

Given the GGI formulation of the aggregation function, we nextdefine regret for the multi-objective variant of our bandit model.

4.3.3 Regret. If we know πœ—βˆ—, then after 𝑇 rounds, the optimalmixed policy𝜢 βˆ—

[𝑑 ] is provided by a solution to the following problem

max𝜢 [𝑑 ] ∈A

𝐺

(1𝑇

π‘‡βˆ‘π‘‘=1

πΎβˆ‘π‘˜=1

𝜢 [𝑑 ],π‘˜J𝑇[𝑑 ],π‘˜πœ—

βˆ—), (5)

where we have assumed that random noises 𝜻 [𝑑 ] average out atzero for large 𝑇 .

We define regret as the difference between the optimal value ofreward and reward from any strategy as:

𝑅 [𝑇 ] = 𝐺

(1𝑇

π‘‡βˆ‘π‘‘=1

πΎβˆ‘π‘˜=1

𝜢 βˆ—[𝑑 ],π‘˜J

𝑇[𝑑 ],π‘˜πœ—

βˆ—)

βˆ’πΊ(

1𝑇

π‘‡βˆ‘π‘‘=1

πΎβˆ‘π‘˜=1

𝜢 [𝑑 ],π‘˜J𝑇[𝑑 ],π‘˜πœ—

βˆ—), (6)

where 𝜢 [𝑑 ] is the action recommended by the employed strategy.Note that in our definition we use the true parameter πœ—βˆ— insteadof the true mean feedback as in [5]. Also important is to note thatperformance is measured by the function value of the averagereward instead of the average of the rewards’ function value.

The arm selection strategy presented above employs the GGFas an aggregation function to scalarize multiple metrics. Our goalnow is to find the parameters of the arm selection strategy givenby Equation 5. We next describe a gradient ascent based algorithmthat learns this policy.

4.4 Learning AlgorithmTo surface recommendations that satisfy multiple objectives, weneed to learn the arm selection strategy (𝜢 [𝑑 ] ) that recommends ap-propriate content that minimizes the regret for an aggregation func-tion 𝐺 (𝒙). We propose an online learning algorithm (MO-LinCB)to optimize the regret defined in the previous section. Our methodexploits the convexity of the GGI operator and formalizes the policy

search problem as an online convex optimization problem, whichis solved by Online Gradient Ascent algorithm.

4.4.1 Ridge Regression. The optimal policy that maximizes Equa-tion 5 assumes knowledge of πœ—βˆ—. In this section, we first estimateparameter πœ—βˆ— via 𝑙2-regularised least-squares regression with regu-larisation parameter _ > 0 at each round 𝑑 ,:

πœ— [𝑑 ] =(J1:π‘‘βˆ’1J𝑇

1:π‘‘βˆ’1 + _I𝑀×𝑀)βˆ’1

J1:π‘‘βˆ’1X1:π‘‘βˆ’1, (7)

where J1:π‘‘βˆ’1 has columns 𝒇[1] ,𝒇[2] , . . . ,𝒇[π‘‘βˆ’1] , 𝒇[𝑑 ] being the fea-ture associated with the chosen arm at round 𝑑 , andX1:π‘‘βˆ’1 has rows𝒙 [1] , 𝒙 [2] , . . . , 𝒙 [π‘‘βˆ’1] . However, if the feature length is huge, matrixinversion is prohibitively challenging and we have to use stochasticgradient descent [31] for ridge regression.1

4.4.2 Online Gradient Ascent. We are seeking an algorithm for𝜢 [𝑑 ] such that the objective below is maximised:

𝐺

(1𝑇

π‘‡βˆ‘π‘‘=1

πΎβˆ‘π‘˜=1

𝜢 [𝑑 ],π‘˜J𝑇[𝑑 ],π‘˜πœ— [𝑑 ]

). (8)

Due to the concavity of 𝐺 (𝒙), we have the following relationship

1𝑇

π‘‡βˆ‘π‘‘=1

𝐺

(πΎβˆ‘π‘˜=1

𝜢 [𝑑 ],π‘˜J𝑇[𝑑 ],π‘˜πœ— [𝑑 ]

)≀ 𝐺

(1𝑇

π‘‡βˆ‘π‘‘=1

πΎβˆ‘π‘˜=1

𝜢 [𝑑 ],π‘˜J𝑇[𝑑 ],π‘˜πœ— [𝑑 ]

). (9)

Instead of optimising Equation 8 directly, we use online gradientascent to optimise the left-hand side of Equation 9. This is becausevery rarely we have features of all the rounds simultaneously. Inthe case of personalised recommendation, users always arrive se-quentially and an instantaneous response is demanded. Therefore,at each round, we try to optimise

𝐺

(πΎβˆ‘π‘˜=1

𝜢 [𝑑 ],π‘˜J𝑇[𝑑 ],π‘˜πœ— [𝑑 ]

)(10)

which can be thought as a function of 𝜢 [𝑑 ] , 𝑓[𝑑 ] (𝜢 [𝑑 ] ). To carry outgradient ascent, we takemany steps along the gradient of 𝑓[𝑑 ] (𝜢 [𝑑 ] ).Lastly, we need to project back onto A to ensure feasibility. Theoverall algorithm is presented in Algorithm 1.

When using GGF as the aggregation function, Equation 3 can besolved by a linear program [28], so one could also try to optimiseEquation 9 by linear programming. However, in live productionthis is too expensive to deploy due to its prohibitive computationalcost. Now the gradient of 𝑓[𝑑 ] (𝜢 [𝑑 ] ) has an analytical form2

πœ•π‘“[𝑑 ]πœ•πœΆ [𝑑 ],π‘˜

= π’˜π‘‡ (J𝑇[𝑑 ],π‘˜πœ— [𝑑 ] )𝜎

and can be used to derive theoretical guarantees. We leave thisderivation of guarantees as future work.

The main component in the proposed approach is the applicationof online gradient ascent on the estimated GGF for the currentround, 𝑓[𝑑 ] (𝜢 [𝑑 ] ). While past approaches in non-contextual versionof bandit setting [5] tackle a similar problem, there are differences.1The convergence for a single pass through the data is shown in [27].2Note that 𝜎 sorts the components of

βˆ‘πΎπ‘˜=1 𝜢 [𝑑 ],π‘˜ J𝑇[𝑑 ],π‘˜πœ— [𝑑 ] in increasing order.

1.30

1.32

1.34

1.36

1.38

1.40

GGF

K = 50 and D = 5

-greedyLinUCBMO-LinCBmixed strategypure strategy

0 5000 10000Rounds

0.77

0.78

0.79

0.80

0.81

GGF

random strategyMO-OGDE

1.55

1.60

1.65

1.70

K = 100 and D = 5

-greedyLinUCBMO-LinCBmixed strategypure strategy

0 5000 10000Rounds

0.90

0.91

0.92

0.93

0.94

0.95random strategyMO-OGDE

1.40

1.45

1.50

1.55K = 200 and D = 5

-greedyLinUCBMO-LinCBmixed strategypure strategy

0 5000 10000Rounds

0.78

0.79

0.80

0.81

0.82

0.83

0.84random strategyMO-OGDE

1.82

1.84

1.86

1.88

1.90

1.92K = 500 and D = 5

-greedyLinUCBMO-LinCBmixed strategypure strategy

0 5000 10000Rounds

0.93

0.94

0.95

0.96

0.97

0.98random strategyMO-OGDE

Figure 2: [Simulation test results] GGF of average reward for different number of objectives (𝐷), and for varying number of arms (𝐾 ). Toimprove visual inspection of results, we split the results into two parts: top row shows results for 5 compared approaches, while bottom rowfor the remaining 2 baselines.

First, in the contextual bandit setting, each problem at round 𝑑 isdifferent from another due to their different mean feedback. Thus, itis not sensible to initialise 𝜢 [𝑑 ] for the current round as the outcomefrom the previous round, 𝜢 [π‘‘βˆ’1] . To counterbalance such cold start,we iterate multiple times at each round. Second, there is no forcedexploration in our algorithm as the means are estimated via ridgeregression eliminating the need to pull every arm regularly.

5 EXPERIMENTAL EVALUATIONThe proposed multi-objective contextual bandit model allows usto jointly optimize multiple objectives when serving recommen-dations. To evaluate its performance, we conduct experiments onsynthetic dataset and a large scale real world user interaction datafrom a major music streaming platform. We describe the experi-mental setup, and present a number of insights that demonstratethat the proposed multi-objective contextual bandits are well suitedas recommendation models in multi-stakeholder setting.

5.1 Baseline MethodsWe compare the proposed approach (MO-LinCB) with several estab-lished methods, including both state-of-the-art bandit recommenda-tion techniques, as well as recent multi-objective bandit algorithms.

(1) πœ–-greedy algorithm (πœ–-g (C), πœ–-g (S), πœ–-g (L)) [34]: a simplebandit approach that instead of picking the best availableoption always, randomly explores other options with a prob-ability πœ– . The baselines πœ–-g (C), πœ–-g (S), πœ–-g (S) correspondto πœ–-greedy methods for the click, stream length and totalnumber of songs played objectives respectively.

(2) LinUCB [19]: a widely used and deployed bandit approachbased on the principle of optimism in the face of uncertainty.It chooses actions by their mean payoffs and uncertaintyestimates.

(3) MO-OGDE [5]: a recently proposed multi-objective banditmodel that exploits the concavity of GGF and uses forcedexploration. The model is non-contextual.

(4) Pure strategy: A baseline model that always pulls the armwith the highest current mean reward.

(5) Mixed Strategy: A baseline model extending the pure strat-egy baseline, that, at each round, proposes a probabilitydistribution according to which an arm is to be drawn. Thisis different to the pure strategy baseline mentioned above,

wherein only a single arm is decided at each round based onthe highest mean reward.

(6) MO-epsilon (πœ–-g (MO)): The multi-objective extension of thewidely used πœ–-greedy bandit model which uses GGI function.

We additionally consider variants of our proposed approach(MO-LinCB) which is a multi-objective contextual bandit based onGGI, trained with different user satisfactionand supplier diversityobjectives:

(1) MO-LinCB: The proposed multi-objective bandit approachthat leverages Gini function for equitable distribution acrossdifferent metrics.

(2) MO-LinCB(g): Variant of the proposed approach, with genderas an additional metric to optimize.

Appendix B provides implementation details of different baselines.

5.2 DatasetsWe consider two datasets for evaluation: (i) synthetic data for con-trolled study of various algorithms and parameters, and (ii) largescale user interaction data from a major music streaming platform.

5.2.1 Simulated Dataset. Experimenting with simulated dataallows us to vary the number of objectives (𝐷) as well as the num-ber of arms (𝐾), and investigate how the different models wouldperform for recommendation problems of varying characteristics.We generate random multi-objective contextual bandit problems,with simulated features and rewards. The number of arms 𝐾 is setto {50, 100, 200, 500} and the number of objectives is taken from𝐷 ∈ {5, 10, 20}. Appendix A provides specific details of the exactsimulation setup details.

Evaluation Setup: Each of the competing algorithms, i.e. MO-LinCB, LinUCB, πœ–-greedy, MO-OGDE and random selection is re-peated 100 times. As the ground-truth (Equation 5) is too costly tocompute, we instead compare with

max𝜢 [𝑑 ] ∈A

1𝑇

π‘‡βˆ‘π‘‘=1

πΊπ’˜

(πΎβˆ‘π‘˜=1

𝜢 [𝑑 ],π‘˜J𝑇[𝑑 ],π‘˜πœ—

βˆ—)

(11)

on grounds of concavity and its pure counterpartmaxπ‘˜ πΊπ’˜

(J𝑇[𝑑 ],π‘˜πœ—

βˆ—)

for each round 𝑑 . We set the exploration probability of the πœ–-greedyalgorithm to 1%. For LinUCB and MO-OGDE, 𝛿 is set to 10%. Thestep size [ is set to 5 and the number of iterations 𝐼 is set to 5 for

Figure 3: [Music data results] Comparison of different approacheson different online metrics: clicks, total length streamed and totalsongs played. The baseline eb corresponds to the Bart [23] - a con-textual bandit recommender system, while ei, el, es correspond toπœ–-greedy methods for the click, stream length and total number ofsongs played objectives respectively.

MO-LinCB. We use this dataset in Sections 5.3 and 5.7 to compareperformance of approaches on reward obtained and time complex-ity based scalability test, respectively.

5.2.2 Music Streaming Dataset. We additionally perform ex-periments on music streaming data of over 10M users and 12Mplaylists from a major music streaming platform. The dataset iscomposed of logged feedback data from live traffic from a randomsample of users in a 7 day period, spanning over 200K user interac-tions with 12 million playlists, spread over 1 million impressions.For playlists that get impressed to users, we record three user met-rics, (i) clicks (C), (ii) total length of music streamed (L) and (iii) totalnumber of songs played (S).

The data is split into two buckets: the first bucket contains 30, 000requests and is used to tune the parameters of each competing algo-rithm (’validation data’). The second bucket contains over 165, 000requests and is used to evaluate various algorithms (’test data’).Each request is accompanied with 130 playlists. The features areconstructed from both user and playlist information. The user fea-tures include demographic information such as gender, age andgeographic information such as region, language, time; as well asaffinity scores computing the degree to which the user likes the

content based on historic interactions. For playlists, features in-clude platform, product, genre, type and tag information. Constantaugmentation and feature normalisation as in [7] are applied.

Policy Evaluation:We follow the unbiased offline evaluator in [20],and compare the proposed approach with baselines. Appendix Cdetails the policy evaluation setup in detail.

5.3 Reward based evaluationWe begin the comparison of different methods by investigating howthe approaches compare on the GGF of rewards obtained using thesimulated dataset. Reward based comparison allows us to comparemodel performance based on how high the obtained reward is forthe different approaches. Since we are trying to maximise GGF,higher GGF values imply better algorithm. We plot the mean GGFof the average reward from various algorithms along with theirstandard deviations in Figures 2 wherein we vary the number ofobjectives from 𝐷 = 5 to 𝐷 = 20 and vary the number of armsfrom 𝐾 = 50 to 𝐾 = 500 respectively. The different subplots presentresults wherein D and K are varied.

First, we observe that when the mean reward 𝒙 [𝑑 ] changes ateach round 𝑑 due to different contexts, the recently proposed multi-objective bandit approach (MO-OGDE baseline; non-contextual)performs no better than random selection; which highlights theimportance of leveraging context; thereby motivating the need formulti-objective variants of contextual bandits. Second, the mixedstrategy always performs better than the pure strategy, which high-lights the benefit of selecting an arm based on a probability distri-bution, rather than selecting only the single best arm each timein a deterministic fashion. This result highlights the benefit ofexploration-exploitation trade-off; i.e., enabling the model to trade-off exploitation of known arms with exploration of potentiallyuseful arms, is beneficial.

Third, we see that all algorithms that involve some element ofuncertainty, e.g. πœ–-greedy, LinUCB and MO-LinCB, perform betterthan the mixed strategy approach, which solves a surrogate of thetrue objective exactly. Therefore, stochastic algorithms are advanta-geous. Finally, our proposed algorithm, MO-LinCB, performs betterthan all compared approaches, including πœ–-greedy and LinUCBover the entire range of 𝐷 and 𝐾 values investigated, which demon-strates the utility of optimizing the Gini-function formulation ofdifferent objectives in a contextual bandit setting.

5.4 Optimizing for Multiple User MetricsWe use the Music Streaming Dataset to investigate performance onthe task of optimizing multiple user centric objectives. In Figure 3,we plot the reward distributions of the objectives clicks, total lengthstreamed and total songs played for the different approaches.

We observe that all strategies that leverage contextual informa-tion significantly improve upon random selection, which assertsthat the constructed features possess predictive power for all objec-tives and the online ridge regression is effective. From the resultsof single-objective algorithms, we observe that πœ–-greedy for CTRnot only performs best for CTR but also for the other two usercentric metrics (L & S). This can be explained by the fact that theconstructed features have more predictive power for clicks and thatthere is correlation between the objectives.

Figure 4: [Music data test results] Comparison of single objective andmulti-objectivemodel for the different objectives upon adding a genderbased promotion objective.

Figure 5: [Music data test results] Investigating the difference in performance across different approaches of multi-objective optimization.The results highlight that the way of doing Multi-objective matters.

Upon comparing the single-objective and multi-objective algo-rithms, we observe that both the multi-objective πœ–-greedy πœ–-g (MO)and proposed MO-LinCB algorithms perform better than all thesingle-objective algorithms (πœ–-g (C),πœ–-g (S), πœ–-g (L)) with much lessvariance. This agrees with a known result that fusing informationfrom correlated objectives is beneficial [1, 16], thereby implyingthat single-objective algorithms are not Pareto-efficient, and thatjointly optimising for correlated objectives is better than individu-ally optimising for them separately.

Most importantly, we observe that the proposed MO-LinCB ap-proach outperforms all other approaches, with over 6.9%, 11.0% and10.0% gains across the three different metrics, respectively. Thishighlights the fact that the proposed MO-LinCB is better than πœ–-g(MO) by using guided rather than random exploration.

These results show that optimizing for multiple user interactionmetrics performs better for each metric than directly optimizingthat metric alone. Considering multiple user interaction metrics hasenabled the model to have a holistic view of user experience andthe model benefits from the inherent chaining structure exhibitedby the metrics considered (i.e. click→ stream → #songs streamed).

5.5 Impact of Promotional ObjectiveBeyond users, recommender systems often have other stakeholders,e.g. artists or retailers. While the experiments so far consideredmultiple user centric objectives, many competing objectives exist,e.g. local vs global artists, popularity, gender. In this paper, we usegender to illustrate our approach on promotional objectives. Thecorrelation analysis (Figure 1) shows a correlation of -0.12 betweenpromotion and satisfaction objective, which highlights that thepromotion objective is a mildly competing objective to satisfaction.

We work with the real world music streaming data (as describedin Section 5.2.2), but for each recommended set, we additionallyconsider the gender of all the artists whose tracks belong to thisset, and compute the proportion of non-male artist, which we referto as gender metric. Figure 4 presents the results for the four usersatisfaction metrics as well as the gender metric comparing perfor-mance of the single objective πœ–-greedy approach with variants of

the proposed multi-objective linear bandit models: with and with-out the gender objective. Similar to the previous result, we observethat both variants of the proposed multi-objective models performbetter than the single objective case, across all five metrics. Further-more, we observe gains in gender metric when we add gender asan additional optimization objective. It is important to note that thegain in gender objective was achieved without hurting any of thefour satisfaction metrics. This hints at the fact that the proposedMO-LinCB is able to find better pareto-optimal solutions than otherapproaches, thereby increasing performance across all metrics.

This result highlights that optimizing for promotional metrics inaddition to user satisfaction metrics need not be a zero-sum game;i.e., it may be possible to achieve gains in promotional metricswithout hurting satisfaction metrics. One possible explanation forthis might be the fact that a large proportion of users indeed wantto listen to more gender-diverse music, and introducing gender-diversity as an additional metric satisfies those users more.

5.6 Comparing Multi-optimization MethodsThe results presented above highlight the importance of consideringmany objectives while optimizing recommendation systems. Wenow investigate the importance of the specific approach adoptedto optimize multiple metrics using the real world music streamingdata. We question whether any approach for multi-objective modelwould work. In Figure 5 we compare the proposed method withmulti-objective extension of the widely used πœ–βˆ’greedy bandit ap-proach, as well as single objective bandit model. We observe thatthe multi-objective variant of πœ–βˆ’greedy bandit approach performscomparably with the single objective bandit model, whereas theproposed multi-objective bandit model performs better than bothbaselines. This highlights the fact that the specific approach of lever-aging multiple objectives matters; and that naive multi-objectiveextensions may not perform as well.

6 CONCLUSIONSRecommender systems powering multi-stakeholder platforms oftenneed to consider different stakeholders and their objectives whenserving recommendations. Based on correlation analysis across the

different stakeholder objectives, we highlighted that trade-offs existbetween objectives, and motivated the need for multi-objectiveoptimization of recommender systems. To address this problem,we presented MO-LinCB, a multi-objective linear contextual banditmodel, which leverages Gini aggregation function to scalarize mul-tiple objectives, and proposed a scalable gradient ascent based ap-proach to learn the recommendation policy. The GGI based methodhelps the model to not only minimize the cumulative cost for eachobjective, but also balance the different objectives.

Our findings suggest that optimizing for multiple objectives isoften beneficial to all objectives involved. We observed that op-timizing for multiple satisfaction metrics result in improved per-formance for each satisfaction metric, which highlights the factthat different satisfaction metrics capture different aspects of userbehavior, and jointly considering them results in surfacing betterrecommendations. Furthermore, it is possible to obtain gains inboth complementary and competing objectives via multi-objectiveoptimization. Detailed experiments around quantifying the impactof competing objectives on satisfaction metrics highlight the ben-efits offered by the proposed model. The proposed approach wasable to obtain gains in a competing promotional objective, withouthurting user satisfaction metrics.

While our findings have implications on the design of recom-mender system powering multi-stakeholder platforms, there exista number of future research directions to pursue. First, one maybe interested in having more control over the importance of differ-ent objectives. While Gini function favors equitable distribution,other functions could be researched so that to provide fine-grainedcontrol over objectives. Second, a possible extension could includeglobal objectives that are not session-specific but aggregated overtime. Third, the ability to optimize for multiple metrics motivatesthe need to quantify other objectives of stakeholders. Finally, weonly experimented with a subset of possible objectives. It will beimportant to extend this subset with a wide range of objectives fromvarious stakeholders, in particular those very competing objectives.

REFERENCES[1] Gediminas Adomavicius and YoungOk Kwon. 2007. New Recommendation

Techniques for Multicriteria Rating Systems. IEEE Intelligent Systems (2007).[2] Deepak Agarwal, Bee-Chung, Pradheep, and Xuanhui. 2011. Click shaping to

optimize multiple objectives. In Proceedings of KDD 2011.[3] Deepak Agarwal, Bee-Chung Chen, Pradheep Elango, and Xuanhui Wang. 2012.

Personalized click shaping through lagrangian duality for online recommenda-tion. In Proceedings of SIGIR 2012.

[4] Leon Barrett and Srini Narayanan. 2008. Learning All Optimal Policies withMultiple Criteria. In ICML. 41–47.

[5] RΓ³bert Busa-Fekete, BalΓ‘zs SzΓΆrΓ©nyi, Paul Weng, and Shie Mannor. 2017. Multi-objective Bandits: Optimizing the Generalized Gini Index. In ICML.

[6] Konstantina Christakopoulou, Jaya Kawale, and Arindam Banerjee. [n.d.]. Rec-ommendation with capacity constraints. In Proceedings of CIKM 2017.

[7] Wei Chu and Seung-Taek Park. 2009. Personalized Recommendation on DynamicContent Using Predictive Bilinear Models. In WWW.

[8] Paolo Cremonesi, Yehuda Koren, and Roberto Turrin. 2010. Performance ofRecommender Algorithms on Top-n Recommendation Tasks. In RecSys.

[9] M. M. Drugan and A. Nowe. 2013. Designing multi-objective multi-armed banditsalgorithms: A study. In International Joint Conference on Neural Networks (IJCNN).

[10] ZoltΓ‘n GΓ‘bor, Zsolt KalmΓ‘r, and Csaba SzepesvΓ‘ri. 1998. Multi-criteria Reinforce-ment Learning. In ICML. 197–205.

[11] Rupesh Gupta, Guanfeng Liang, Ravi Kiran Tseng, Xiaoyu Chen, and RomerRosales. [n.d.]. Email volume optimization at LinkedIn. In KDD 2016.

[12] Tamas Jambor and Jun Wang. [n.d.]. Optimizing multiple objectives in collabora-tive filtering. In Proceedings of RecSys 2010.

[13] Dietmar Jannach and Malte Ludewig. 2017. When recurrent neural networksmeet the neighborhood for session-based recommendation. In Proceedings of theEleventh ACM Conference on Recommender Systems. ACM, 306–310.

[14] Thorsten Joachims. 2002. Optimizing search engines using clickthrough data.In Proceedings of the eighth ACM SIGKDD international conference on Knowledgediscovery and data mining. ACM, 133–142.

[15] I. Y. Kim and O. L. de Weck. 2006. Adaptive weighted sum method for multiob-jective optimization: a new method for Pareto front generation. Structural andMultidisciplinary Optimization 31, 2 (2006), 105–116.

[16] Anisio Lacerda. 2015. Contextual Bandits for Multi-objective RecommenderSystems. In Proceedings of the 2015 Brazilian Conference on Intelligent Systems(BRACIS). 68–73.

[17] Kleanthi Lakiotaki, Nikolaos F Matsatsinis, and Alexis Tsoukias. 2011. Multicri-teria user modeling in recommender systems. IEEE Intelligent Systems (2011).

[18] John Langford and Tong Zhang. 2008. The Epoch-Greedy Algorithm for Multi-armed Bandits with Side Information. In NIPS.

[19] Lihong Li, Wei Chu, John Langford, and Robert E. Schapire. 2010. A Contextual-bandit Approach to Personalized News Article Recommendation. In WWW.

[20] Lihong Li, Wei Chu, John Langford, and Xuanhui Wang. 2011. Unbiased OfflineEvaluation of Contextual-bandit-based News Article Recommendation Algo-rithms. InWSDM.

[21] C. Liu, X. Xu, and D. Hu. 2015. Multiobjective Reinforcement Learning: A Com-prehensive Overview. IEEE Transactions on Systems, Man, and Cybernetics (2015).

[22] Donald W. Marquardt and Ronald D. Snee. 1975. Ridge Regression in Practice.The American Statistician 29, 1 (1975), 3–20.

[23] James McInerney, Benjamin Lacker, Samantha Hansen, Karl Higley, HuguesBouchard, Alois Gruson, and Rishabh Mehrotra. [n.d.]. Explore, Exploit, and Ex-plain: Personalizing Explainable Recommendations with Bandits. In Proceedingsof RecSys 2018.

[24] Rishabh Mehrotra and Benjamin Carterette. 2019. Recommendations in a mar-ketplace. In Proceedings of the 13th ACM Conference on Recommender Systems.580–581.

[25] Rishabh Mehrotra, James McInerney, Hugues Bouchard, Mounia Lalmas, andFernando Diaz. 2018. Towards a Fair Marketplace: Counterfactual Evaluationof the trade-off between Relevance, Fairness & Satisfaction in RecommendationSystems. In CIKM.

[26] K. Van Moffaert, K. Van Vaerenbergh, P. Vrancx, and A. Nowe. [n.d.].[27] Eric Moulines and Francis R. Bach. 2011. Non-Asymptotic Analysis of Stochastic

Approximation Algorithms for Machine Learning. In NIPS.[28] WΕ‚odzimierz Ogryczak and Tomasz Sliwinski. 2003. On solving linear programs

with the ordered weighted averaging objective. European Journal of OperationalResearch 148, 1 (2003), 80 – 91.

[29] Saba Q. Yahyaa, Madalina M. Drugan, and Bernard Manderick. 2014. KnowledgeGradient for Multi-objective Multi-armed Bandit Algorithms. In Proceedings ofthe 6th International Conference on Agents and Artificial Intelligence - Volume 1.

[30] Marco Tulio Ribeiro, Lacerda, Veloso, and Ziviani. [n.d.]. Pareto-efficient hy-bridization for multi-objective recommender systems. In RecSys 2012.

[31] Herbert Robbins and Sutton Monro. 1951. A stochastic approximation method.The Annals of Mathematical Statistics 22, 3 (1951), 400–407.

[32] Lili Shan, Lei Lin, and Chengjie Sun. [n.d.]. Combined Regression and TripletwiseLearning for Conversion Rate Prediction in Real-Time Bidding Advertising. InSIGIR 2018.

[33] Aleksandrs Slivkins. 2014. Contextual Bandits with Similarity Information. J.Mach. Learn. Res. 15, 1 (2014), 2533–2568.

[34] Richard S. Sutton and Francis Bach. 1998. Reinforcement Learning: An Introduction.MIT Press.

[35] Krysta M Svore, Maksims N Volkovs, and Christopher JC Burges. [n.d.]. Learningto rank with multiple objective functions. In Proceedings of WWW 2011.

[36] Cem Tekin and Eralp Turğay. 2018. Multi-objective contextual multi-armedbandit with a dominant objective. IEEE Transactions on Signal Processing (2018).

[37] Eralp Turğay, Doruk Γ–ner, and Cem Tekin. [n.d.]. Multi-objective contextualbandit problem with similarity information. arXiv preprint arXiv:1803.04015, 2018([n. d.]).

[38] Umair ul Hassan and Edward Curry. 2016. Efficient task assignment for spatialcrowdsourcing: A combinatorial fractional optimization approach with semi-bandit learning. Expert Systems with Applications 58 (2016), 36–56.

[39] John A. Weymark. 1981. Generalized Gini inequality indices. Mathematical SocialSciences 1 (1981), 409–430.

[40] Lin Xiao, Minand, Zhaoquan, L Yiqun, and Ma Shaoping. [n.d.]. Fairness-awaregroup recommendation with pareto-efficiency. In RecSys 2018.

[41] R. R. Yager. [n.d.]. On ordered weighted averaging aggregation operators inmulticriteria decisionmaking. IEEE Transactions on Systems, Man, & Cybernetics([n. d.]).

[42] Saba Yahyaa, Madalina Drugan, and Bernard Manderick. 2015. Thompson Sam-pling in the Adaptive Linear Scalarized Multi Objective Multi Armed Bandit. InProceedings of the 7th International Conference on Agents and Artificial Intelligence.

[43] Xing Yi, Liangjie Hong, Erheng Zhong, Nanthan Nan Liu, and Suju Rajan. [n.d.].Beyond clicks: dwell time for personalization. In Proceedings of RecSys 2014.

[44] Chongjie Zhang and Julie A Shah. 2014. Fairness in Multi-Agent SequentialDecision-Making. In NIPS. 2636–2644.

Algorithm 2 Policy Evaluator1: Input: stream of events S, bandit algorithm Ξ, desired number

of events 𝜏2: Set H0 = βˆ… (an initially empty history)3: Set οΏ½ΜƒοΏ½ [0] = 0 (an initially zero total reward)4: Set 𝑐 = 0 (an initially zero counter for registered events)5: for 𝑑 = 1, 2, 3, . . . do6: if 𝑐 = 𝜏 then7: Stop8: else9: Get next event (J[𝑑 ] , π‘˜ , 𝒙 [𝑑 ] ) from S10: if Ξ(J[𝑑 ] ,H𝑐 ) = π‘˜ then11: H𝑐+1 = H𝑐 βˆͺ {(J[𝑑 ] , π‘˜, 𝒙 [𝑑 ] )}12: οΏ½ΜƒοΏ½ [𝑐+1] = οΏ½ΜƒοΏ½ [𝑐 ] + 𝒙 [𝑑 ]13: 𝑐 = 𝑐 + 114: else15: Skip16: Output: 𝑓 (οΏ½ΜƒοΏ½ [𝜏 ] ), e.g. πΊπ’˜ ( 1

𝜏 οΏ½ΜƒοΏ½ [𝜏 ] )

A APPENDIX A: DETAILS OF SIMULATEDDATASET

Here we describe in detail the exact setting and parameters used tocreate the simulated dataset. To generate random multi-objectivecontextual bandit problems, we draw the universal parameter, πœ—βˆ—,uniformly from [0, 1)𝑀×𝐷 . Each element of the feature with length𝑀 , J[𝑑 ],π‘˜ , for each arm π‘˜ at each round 𝑑 , is chosen independentlyfromN( 1

𝑀, 1𝑀2 ). Each element of the noise 𝜻 [𝑑 ],𝑑 for each objective

𝑑 at each round 𝑑 is generated according to N(0, 0.01𝝁2𝑑), where

𝝁𝑑 = 𝒇𝑇[𝑑 ]πœ—βˆ—π‘‘is the mean reward for objective 𝑑 of the arm chosen

at round 𝑑 . We set 𝑀 to 10 and run a total of 𝑇 = 10, 000 rounds.We set the regularisation parameter, _ to 0.1 ∈ [0, 1] [22], whichis sufficient to ensure that the estimated πœ— converges to the actualπœ—βˆ— as more rounds are played. The number of arms 𝐾 is set to{50, 100, 200, 500} and the number of objectives is taken from 𝐷 ∈{5, 10, 20}. We set the weight vector π’˜ of GGF to π’˜π‘‘ = 2βˆ’π‘‘+1, 1 ≀𝑑 ≀ 𝐷 , the same as in [5].

B APPENDIX B: IMPLEMENTATION DETAILSWe experimented with different values of exploration probabilityπœ– for the πœ–-greedy baseline, πœ– ∈ [1%,5%,10%] and picked the bestperforming estimate (πœ–=1%) for comparison. Similarly, for LinUCBand MO-OGDE, 𝛿 is set to 10% following empirical comparisonfor different tolerance values, 𝛿 ∈ [1%,5%,10%]. The step size [ isset to 5 and the number of iterations 𝐼 is set to 5 for MO-LinCB,following empirical evaluation on various step sizes and iterations.We also study single-objective πœ–-greedy algorithms using each ofthe three objectives denoted by πœ–-g (C), πœ–-g (L) and πœ–-g (S) for theclick, length and stream objectives respectively. Each algorithm isrepeated 30 times. πœ– = 0.01 is used in all πœ–-greedy algorithms inmusic data experiments as well.

C APPENDIX C: POLICY EVALUATIONGiven a bandit algorithm and a desired number of user interac-tion sessions 𝜏 , we step through the stream of recorded sessions

MO-LinCB LinUCB MO-OGDEtime/s K=50 K=200 K=500 K=50 K=200 K=500 K=50 K=200 K=500D=5 0.46 0.85 1.68 0.52 1.43 3.75 0.26 0.67 1.5D=10 0.47 0.88 1.72 0.63 1.86 4.67 0.26 0.69 1.52D=20 0.52 0.98 1.87 0.84 2.68 6.55 0.28 0.71 1.6

Table 1: Running times of different approaches.

clicksAlgorithms Random e-g (C) e-g (L) e-g (S) e-g (MO)MO-LinCB <0.1% <0.1% <0.1% <0.1% <0.1%e-g (MO) <0.1% <0.1% <0.1% <0.1%e-g (S) <0.1% <0.1% 11.8%e-g (L) <0.1% <0.1%e-g (C) <0.1%Table 2: P-values of ’clicks’ for all pairs of algorithms.

sequentially. If it happens that the algorithm selects the same armas the one that is recorded via an i.i.d. uniformly random policy,we register such an event and update our algorithm and the totalpayoff. If the algorithm selects a different arm, we ignore such anevent and move on to the next event without changing the state ofthe algorithm. The detailed procedure is presented in Algorithm 2.While this evaluation approach suffers from inefficient data usage,we defer a more detailed counterfactual evaluation for future work.We use this dataset in Sections 5.4 for evaluating the impact onmultiple user satisfaction metrics, Section 5.5 for quantifying theimpact on promotional objective and in Section 5.6 for comparingthe different multi-objective models.

D APPENDIX D: SCALABILITYWe measured the running time of different algorithms as 𝐾 and𝐷 vary using the simulation data experiments and observed thatpure and mixed strategies were orders of magnitude slower, whichrender them impractical. Running times for remaining approachesare summarise in Table 1. Our proposed MO-LinCB is roughly threetimes faster than LinUCB at large 𝐾 . This is crucial in recommen-dation as more choices (higher 𝐾 ) enables better personalisation.

E APPENDIX E: SIGNIFICANCE TESTTo test for real differences between the mean metrics of differentapproaches, we assume that scores from any algorithm are normallydistributed,3 and present the p-values under the null hypothesis(there is no difference between the mean scores between any two al-gorithms) in Table 2 for the click metric, where unbiased estimatorsfor the population variances are used. We see that the differencesbetween the results of relevant algorithms are statistically signifi-cant.

3This is a good approximation by central limit theorem given the adequate data size.