m -t learning for document ranking and q s · query suggestion for web search. it consists of two...

14
Published as a conference paper at ICLR 2018 MULTI -TASK L EARNING FOR D OCUMENT R ANKING AND QUERY S UGGESTION Wasi Uddin Ahmad & Kai-Wei Chang Department of Computer Science University of California, Los Angeles {wasiahmad,kwchang.cs}@ucla.edu Hongning Wang Department of Computer Science University of Virginia [email protected] ABSTRACT We propose a multi-task learning framework to jointly learn document ranking and query suggestion for web search. It consists of two major components, a document ranker and a query recommender. Document ranker combines current query and session information and compares the combined representation with document representation to rank the documents. Query recommender tracks users’ query re- formulation sequence considering all previous in-session queries using a sequence to sequence approach. As both tasks are driven by the users’ underlying search intent, we perform joint learning of these two components through session recur- rence, which encodes search context and intent. Extensive comparisons against state-of-the-art document ranking and query suggestion algorithms are performed on the public AOL search log, and the promising results endorse the effectiveness of the joint learning framework. 1 I NTRODUCTION Understanding users’ information need is the key to optimize a search engine for providing rele- vant search results. Search engine logs have been extensively used to mine users’ search intent, reflected in their search result click preferences and query reformulations (Baeza-Yates et al., 2004; Croft et al., 2010). Typically, user query logs are partitioned into search sessions, i.e., sequences of queries and clicks issued by the same user and within a short time interval. Search sessions provide useful contextual information about user intent and help to narrow down ambiguity while ranking documents for the current query and predicting next query that users will submit, a.k.a. context- awareness (Jiang et al., 2014). Since both a user’s click behavior and query reformulation are driven by the underlying search intent, we argue that jointly modeling both tasks can benefit each other. In this work, we propose a joint learning framework, called multi-task neural session relevance framework (M-NSRF), to predict users’ result clicks and future queries in search sessions. We model search context within a session via a recurrent latent state in a deep neural network (Collobert & Weston, 2008; Liu et al., 2015), which governs the generation of result clicks in the current query and formation of next query. By sharing the latent states across the tasks of document ranking and query suggestion, we learn the representations of queries, documents and user intent carrying over the whole session jointly, i.e., multi-task learning. This multi-task learning framework is flexible and can be incorporated with existing approaches for representing query and documents. The general workflow of M-NSRF is illustrated in Figure 1. Given a sequence of queries from the same search session, e.g., “cheap furniture” and “craig list virginia”, M-NSRF is trained to predict both the result clicks under the current query and the next query “cheap furniture for sale.” It is evident that in this search session the user kept reformulating the queries because his/her information need has not been met by the result clicks in the previously submitted queries, reflected in the added and removed query terms. And such revisions suggest what he/she might want to click next. As a result, the modeling of result clicks and query reformulations mutually reinforce each other to reveal users’ underlying search intent. In M-NSRF, we model user search intent as a session-level recurrent state, the learning of which is aided by both document ranking and query prediction tasks. We evaluate the effectiveness of the proposed framework using the publicly available AOL search log and compare with several well-known classical retrieval models, as well as various neural re- trieval models specifically designed for ad-hoc retrieval. We also compare M-NSRF with several 1

Upload: others

Post on 18-Aug-2020

4 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: M -T LEARNING FOR DOCUMENT RANKING AND Q S · query suggestion for web search. It consists of two major components, a document ranker and a query recommender. Document ranker combines

Published as a conference paper at ICLR 2018

MULTI-TASK LEARNING FOR DOCUMENT RANKINGAND QUERY SUGGESTION

Wasi Uddin Ahmad & Kai-Wei ChangDepartment of Computer ScienceUniversity of California, Los Angeles{wasiahmad,kwchang.cs}@ucla.edu

Hongning WangDepartment of Computer ScienceUniversity of [email protected]

ABSTRACT

We propose a multi-task learning framework to jointly learn document ranking andquery suggestion for web search. It consists of two major components, a documentranker and a query recommender. Document ranker combines current query andsession information and compares the combined representation with documentrepresentation to rank the documents. Query recommender tracks users’ query re-formulation sequence considering all previous in-session queries using a sequenceto sequence approach. As both tasks are driven by the users’ underlying searchintent, we perform joint learning of these two components through session recur-rence, which encodes search context and intent. Extensive comparisons againststate-of-the-art document ranking and query suggestion algorithms are performedon the public AOL search log, and the promising results endorse the effectivenessof the joint learning framework.

1 INTRODUCTION

Understanding users’ information need is the key to optimize a search engine for providing rele-vant search results. Search engine logs have been extensively used to mine users’ search intent,reflected in their search result click preferences and query reformulations (Baeza-Yates et al., 2004;Croft et al., 2010). Typically, user query logs are partitioned into search sessions, i.e., sequences ofqueries and clicks issued by the same user and within a short time interval. Search sessions provideuseful contextual information about user intent and help to narrow down ambiguity while rankingdocuments for the current query and predicting next query that users will submit, a.k.a. context-awareness (Jiang et al., 2014). Since both a user’s click behavior and query reformulation are drivenby the underlying search intent, we argue that jointly modeling both tasks can benefit each other.

In this work, we propose a joint learning framework, called multi-task neural session relevanceframework (M-NSRF), to predict users’ result clicks and future queries in search sessions. Wemodel search context within a session via a recurrent latent state in a deep neural network (Collobert& Weston, 2008; Liu et al., 2015), which governs the generation of result clicks in the current queryand formation of next query. By sharing the latent states across the tasks of document ranking andquery suggestion, we learn the representations of queries, documents and user intent carrying overthe whole session jointly, i.e., multi-task learning. This multi-task learning framework is flexibleand can be incorporated with existing approaches for representing query and documents.

The general workflow of M-NSRF is illustrated in Figure 1. Given a sequence of queries from thesame search session, e.g., “cheap furniture” and “craig list virginia”, M-NSRF is trained to predictboth the result clicks under the current query and the next query “cheap furniture for sale.” It isevident that in this search session the user kept reformulating the queries because his/her informationneed has not been met by the result clicks in the previously submitted queries, reflected in the addedand removed query terms. And such revisions suggest what he/she might want to click next. As aresult, the modeling of result clicks and query reformulations mutually reinforce each other to revealusers’ underlying search intent. In M-NSRF, we model user search intent as a session-level recurrentstate, the learning of which is aided by both document ranking and query prediction tasks.

We evaluate the effectiveness of the proposed framework using the publicly available AOL searchlog and compare with several well-known classical retrieval models, as well as various neural re-trieval models specifically designed for ad-hoc retrieval. We also compare M-NSRF with several

1

Page 2: M -T LEARNING FOR DOCUMENT RANKING AND Q S · query suggestion for web search. It consists of two major components, a document ranker and a query recommender. Document ranker combines

Published as a conference paper at ICLR 2018

Figure 1: General workflow of the proposed multi-task neural session relevance framework. Theframework is trained on search sessions to jointly predict next query and rank corresponding doc-uments. The model encodes the current query in the session, “craig list virginia”, updates corre-sponding session-level recurrent state, maximizes the probability of the next query, “cheap furniturefor sale” (i.e., the task of query suggestion), encodes the candidate documents for the followingquery and minimizes loss for clicked documents (i.e., the task of document ranking).

baseline models for the query prediction task. The empirical results show that by leveraging infor-mation in both tasks, the proposed approach outperforms existing models significantly on ad-hocdocument retrieval task and exhibits competitive performance in query suggestion.

To summarize, the key contributions of this work include:

1. We propose a novel multi-task neural session relevance model that is jointly trained on documentranking and query suggestion tasks by utilizing in-session queries and clicks in a holistic way.

2. We provide detailed experiment analysis, and release the implementation1 and the data processingtool to facilitate future research.

2 RELATED WORK AND BACKGROUND

Ad-hoc Retrieval. Traditional retrieval models such as query likelihood (Ponte & Croft, 1998) andBM25 (Robertson et al., 2009) are based on exact matching of query and document words with avariety of smoothing, weighting and normalization techniques. Recently, deep neural network basedapproaches demonstrate strong advance in ad-hoc retrieval. Existing neural ranking models fall intotwo categories: representation focused (Huang et al., 2013; Gao et al., 2014) and interaction fo-cused (Guo et al., 2016b). The earlier focus of neural ranking models was mainly on representationbased models (Hu et al., 2014; Shen et al., 2014), in which the query and documents are first em-bedded into continuous vectors, and the ranking is calculated based on their embeddings’ similarity.The interaction focused neural models (Hu et al., 2014; Pang et al., 2016; Guo et al., 2016a), on theother hand, learn query-document matching patterns from word-level interactions. Both the inter-action and representation focused models can be combined for further improvements (Mitra et al.,2017). Similarly, Jaech et al. (2017) captures both local relevance matching and global topicalitysignals when computing relevance of a document to a query. Our work falls into the representa-tion focused approach to form query and document representations and jointly models the two tasksthrough session representation learning.

Query Suggestion. In general, query suggestion algorithms define various distance metrics betweenqueries to find the most similar ones as suggestions for one another (Wen et al., 2001; Baeza-Yates

1https://github.com/wasiahmad/mnsrf ranking suggestion

2

Page 3: M -T LEARNING FOR DOCUMENT RANKING AND Q S · query suggestion for web search. It consists of two major components, a document ranker and a query recommender. Document ranker combines

Published as a conference paper at ICLR 2018

et al., 2004). The closest work to ours is the end-to-end hierarchical recurrent encoder-decoderarchitecture (HRED-qs) (Sordoni et al., 2015), which ranks candidates for context-aware query sug-gestion. Our proposed framework differs from HRED-qs as it integrates another important type ofuser search behavior, i.e., result clicks, into the joint modeling, which provides additional contex-tual information for query suggestion. Mitra & Craswell (2015) proposed a candidate generationapproach for rare prefixes using frequently observed query suffixes and suggested a neural model togenerate ranking features along with n-gram features. Similarly, a large pool of previous works in-vestigated task or context-aware approaches (Cao et al., 2008; He et al., 2009; Feild & Allan, 2013;Chen et al., 2017; Garigliotti & Balog, 2017) for query suggestion.

Multi-task Learning. The goal of multi-task learning is to improve generalization on the tar-get task by leveraging the domain-specific information contained in the training signals of relatedtasks (Caruana, 1998). Multi-task learning in combination with deep neural networks has beensuccessfully used in many application scenarios, including natural language processing (Collobert& Weston, 2008; Liu et al., 2016a;b; Peng & Dredze, 2017; Peng et al., 2017), speech recogni-tion (Deng et al., 2013; Thanda & Venkatesan, 2017) and computer vision (Girshick, 2015). How-ever, it has been less explored in the information retrieval domain. Liu et al. (2015) proposeda multi-task deep neural approach to combine query classification and document ranking, and re-ported improvement on both the tasks. Bai et al. (2009) used multi-task learning in learning to rankfor web search. We propose to jointly learn document ranking and query suggestion via multi-tasklearning to better capture latent intent embedded in users’ search behaviors.

3 MULTI-TASK NEURAL SESSION RELEVANCE FRAMEWORK

We define a session to be a sequence of queries, Q = {Q1, . . . , Qn}, submitted by a user in a shorttime interval in a chronological order to satisfy a specific search intent. Every query Qi in a sessionis associated with a set of related documents D = {D1, . . . , Dm} which need to be ranked by itsrelevance to the query, and o = {o1, . . . , om} is the set of relevance labels for each document in D.In typical search engine logs, oj is usually approximated via user clicks, e.g., oj = 1 if and only ifthe document Dj is clicked. A query Qi and a document Dj consist of a sequence of words, i.e.,Qi = {w1

i , . . . , wqi } and Dj = {w1

j , . . . , wdj }, where q = |Qi| and d = |Dj | are the query and

document lengths respectively. V is the size of vocabulary constructed over queries and relevantdocuments.

The document ranking refers to the task of ordering the retrieved results with respect to their rele-vance to the given query under the users’ search intent. Accurate modeling of relevance between adocument and a query is the key in this task. In this work, we model ranking of related documents asa pointwise classification problem where the goal is to predict the relevance label between a queryand a document. But our developed solution can be extended to pairwise or list-wise ranking models(Liu et al., 2009). On the other hand, query suggestion refers to the task of predicting next queryQi based on previous queries Q1 . . . Qi−1 by users in the same session, so as to help them explorethe search space. The key challenge is to maintain the semantic consistency between the suggestedqueries and users’ original queries, with respect to their search intent.

Traditional IR approaches consider document ranking and query suggestion tasks separately (Huanget al., 2013; Shen et al., 2014; Sordoni et al., 2015), although both the tasks are driven by users’underlying search intent. In contrast, M-NSRF is designed to jointly learn a document ranker and aquery recommender by modeling shared search context embedding in a session. Based on the latentsession state inferred from all the previous queries and clicks in the same session, the documentranker is trained to predict user clicks from the candidate documents for the current query and thequery recommender is trained in a sequence to sequence fashion (Sutskever et al., 2014) to predictthe user’s next query. The process is repeated sequentially for all the queries in the same session.The detailed architecture of the proposed multi-task neural session relevance framework (M-NSRF)is provided in Appendix A. In the following, we discuss each component of M-NSRF.

3.1 DOCUMENT RANKER

Ranking the retrieved documents for an input query requires encoding the query and documentsinto a shared representation space. In addition to the search intent carried by the current query,search context, which is reflected in the previously submitted queries in the same session, shouldalso be accounted in ranking the documents. In M-NSRF, we model the latent user search intent

3

Page 4: M -T LEARNING FOR DOCUMENT RANKING AND Q S · query suggestion for web search. It consists of two major components, a document ranker and a query recommender. Document ranker combines

Published as a conference paper at ICLR 2018

in a sequence of queries via a series of session recurrent states. As a result, the document rankercomponent in M-NSRF consists of a query encoder, a document encoder, a session encoder anda ranker. The ranker sub-component combines the latent representations of query and session andmatch them with document representations to generate the ranking score of documents, based onwhich the documents are ordered. The technical details of each constituent element are given asfollows.

Query Encoder. The query encoder encodes a query into a vector. To encode a sequence of words,various approaches have been studied. We follow (Conneau et al., 2017) to adopt a bidirectionalLSTM with max pooling (BiLSTM-max), due to its superior practical performance. Consideringquery as a sequence of words Qi = {w1

i , . . . , wqi }, the encoder composed of forward and backward

LSTM reads the sequence in two opposite directions,−→h t = LSTMt(

−→h t−1, w

ti),

←−h t = LSTMt(

←−h t+1, w

ti), ht = [

−→h t,←−h t]

where ht ∈ R2d is the query-level recurrent state, d is the dimensionality of the LSTM hidden unitinitialized to a zero vector. To form a fixed-size vector representation of variable length queries,maximum value is selected over each dimension of the hidden units,

Qi,k = maxq hk,q, k = 1, . . . , d

where Qi,k is the k-th element of the latent vector Qi.

Document Encoder. The goal of the document encoder is to encode documents into continuousvectors. We use the same BiLSTM-max technique that is utilized in encoding queries as the docu-ment encoder. The only difference is in the dimensionality of the LSTM hidden units. In general,because a document (body or title) is longer than a query, we consider dense vector of a larger sizeas the continuous representations of the documents.

Session Encoder. The session encoder generates a representation to encode the queries that thesystem has received so far. Unlike query and document encoders that can read the complete encodedqueries and documents, the session encoder does not have information from the future queries.Therefore, we use a unidirectional LSTM (Hochreiter & Schmidhuber, 1997) for session encoding.The session encoder takes the sequence of query representations Q1, ..., Qn as input and computesthe sequence of session-level recurrent states.

Si = LSTMi(Si−1, Qi),

where Si ∈ Rd is the session-level recurrent state initialized to a zero vector. As a result, each queryin the session has its session-level recurrent state, summarizing the user’s information need that hasbeen processed up to query Qi.

Ranker. We first concatenate the current query representation Qi with previous session-level re-current state Si−1 via a non-linear transformation. This combined representation reflects the searchintent reflected in the current query and the past ones in the same session. Then we compute theranking score of document Dj under the query as its probability of being relevant via a sigmoidfunction (with binary relevance labels),

P (Dj |Qi, Si−1) = σ(DT

j tanh(Wr[Qi, Si−1] + br)), j = 1, . . . ,m (1)

where Wr ∈ R(dq+ds) × dd , br ∈ Rdd , and dq , ds and dd are the dimensionality of the queryencoder, session encoder and document encoder hidden units and σ is the sigmoid function. A listof retrieved documents can therefore be ordered by this ranking score.

3.2 QUERY RECOMMENDER

Query recommender suggests related queries to users by inferring the underlying search intentthrough utilizing in-session previous queries embedded in latent vectors. Following Sutskever et al.(2014) and Bahdanau et al. (2015), the query recommender in M-NSRF predicts users’ next query ina sequence to sequence manner. Basically the query recommender module estimates the probabilityof the next queryQi = {w1

i , . . . , wqi } , given all the previous queries up to position i−1 in a session

as follows,P (Qi|Q1:i−1) =

∏q

t=1P (wt

i |w1:t−1i , Q1:i−1)

We use LSTM as a basic building block for the query recommender. Information about all theprevious queries represented through a session vector Si is passed to the query recommender. To

4

Page 5: M -T LEARNING FOR DOCUMENT RANKING AND Q S · query suggestion for web search. It consists of two major components, a document ranker and a query recommender. Document ranker combines

Published as a conference paper at ICLR 2018

this end, the recurrent state of the query recommender is initialized with a non-linear transformationof Si, h0 = tanh(WqSi + bq), where h0 ∈ Rd is the initial recurrent state. Then the queryrecommender’s recurrence is computed by ht = LSTMt(ht−1, w

ti), where ht−1 is the previous

hidden state, wt−1i is the previous query term. Finally, each recurrent state is mapped to a probability

distribution over the vocabulary of size V using a combination of linear transformation and thesoftmax function. Word with the highest probability is chosen as the next word in sequence.

P (w|w1:t−1i , Q1:i−1) = g(Wpht + b), (2)

where g is the softmax function that outputs a vector to represent the distribution of next words,where the probability assigned to the j-th element is defined as g(z)j = ezj∑K

k=1 ezk, j = 1, ...,K.

Query Suggestion. In the decoding phase, similar to (Sordoni et al., 2015), we use greedy decodingalgorithm for suggesting next query. Given a sequence of queries up to position i − 1, a suggestedquery Qi is:

Q∗ = argmaxQ′∈Q P (Q′|Q1:i−1)

where Q is the space of all possible queries. To generate Q∗ = {w1, . . . , wq}, we use a greedy ap-proach like in (Sordoni et al., 2015) where wt = argmaxw P (w|w1:t−1, Q1:i−1) To provide querysuggestions of variable lengths, we use standard word-level decoding techniques. We iterativelyconsider the best prefix w1:t up to length t and extend it by sampling the most probable word giventhe distribution in Eq. (2). The process ends when we obtain a well-formed query containing thespecial end-of-query token.

3.3 LEARNING END-TO-END

Within a session, M-NSRF ranks documents in a set of candidates and predicts next query given thecurrent query sequence. Therefore, the training objective of M-NSRF consists of two terms. Thefirst term is the binary cross entropy loss from the document ranker,

L1 ≡ −1

m

∑m

joj × log P (Dj |Qi) + (1− oj)× log(1− P (Dj |Qi))

where oj represents binary click label for Dj . The second term is the regularized negative log-likelihood loss from the query suggestion model,

L2 ≡ −∑q

tlogP (wt

i |w1:i−1i , Q1:i−1) + LR, (3)

where LR ≡ −λ∑

w∈V P (w|w1:t−1i , Q1:i−1) logP (w|w1:t−1

i , Q1:i−1) is the regularization termto avoid the distribution of words in Eq. (2) from being highly skewed, and λ is a hyper-parameterto control the regularization term. The final objective is the summation of L1 and L2 over all thequeries.2

Note that the document ranker and query recommender share the same document, query, and ses-sion encoders, and the training of M-NSRF can be done in an online manner using the followingprocedure. In the forward pass, M-NSRF computes the query and corresponding document encod-ings, updates session-level recurrent states, click probability for each candidate document and thelog-likelihood of each query in the session given the previous ones. In the backward pass, the gradi-ents are computed and the parameters are updated based on the ADAM update rule (Kingma & Ba,2014). Details of implementation can be found in Section 4.2

3.4 GENERALIZING M-NSRF FOR NEURAL IR MODELS

The proposed multi-task learning framework is general and the query encoder, document encoder,document ranker, and query recommender can be replaced by other designed architecture. In gen-eral, any neural query suggestion model working in the sequence to sequence fashion can be readilyincorporated with the proposed multi-task learning framework. Similarly, most neural IR modelsthat built on the notion of learning query and document representations in a latent space for rele-vance modeling can be Incorporated in our framework as well. However, due to distinctive nature ofdifferent neural IR models, careful study is required while adding context-awareness into the finalarchitecture.

2We can also consider optimizing the weighted sum of L1 and L2. However, our preliminary experimentsshowed that giving them equal weights already performs well.

5

Page 6: M -T LEARNING FOR DOCUMENT RANKING AND Q S · query suggestion for web search. It consists of two major components, a document ranker and a query recommender. Document ranker combines

Published as a conference paper at ICLR 2018

In this paper, we take the Match-Tensor model (Jaech et al., 2017), a recently proposed neuralrelevance model for document ranking as an example and extend it to a multi-task Match-Tensor(M-Match-Tensor) model by incorporating the query recommender component within our proposedmulti-task learning framework. Different from NSRF, which embeds queries and documents intovectors, Match-Tensor learns a contextual representation for each word in the queries and docu-ments. Therefore, documents and queries are represented as matrices. The document ranker inMatch-Tensor then computes the relevant score based on the following formulation:

P (Dj |Qi) = σ(Wr C(Qi, Dj) + br

), i = 1, . . . ,m

where C represents a sequence of convolutional operation (Lawrence et al., 1997) and max-poolingon the 2d product of the query and document vectors. The detailed architecture of the M-Match-Tensor model (M-Match-Tensor) is provided in Appendix B. To incorporate the match-tensor inour mulit-task learning framework, we can replace the document encoder, query encoder, and thedocument ranker in M-NSRF with the ones specified in the Match-Tensor model. However, as thecomputations involve in the document ranker is substantial, we do not increase the computationalcomplexity further by adding session recurrence in the document ranking. The query decodingcomponent in M-Match-Tensor is identical to M-NSRF. In the experiment, we will show that multi-task Match-Tensor achieves better performance than Match-Tensor in document ranking task.

4 EXPERIMENTS

4.1 DATA SETS AND EVALUATION METHODOLOGY

We conduct our experiments on the publicly available AOL search log (Pass et al., 2006). Thequeries in this dataset were sampled between 1 March, 2006 and 31 May, 2006. In total there are16,946,938 queries submitted by 657,426 unique users. We removed all non-alphanumeric charac-ters from the queries, applied word segmentation and lowercasing. We followed (Jansen Bernardet al., 2007) to define a session by a 30-minute window of inactive time, and filtered sessions bytheir lengths (minimum 2, maximum 10). We only kept the most frequent |V | = 100k words andmapped all other words to an<unk> token when constructing the vocabulary. We randomly selected1,032,459 sessions for training, 129,053 sessions for development and 91,108 sessions for testing,with no overlapping. In total, there are 2,987,486 queries for training, 287,138 for development, and259,117 for testing. The average length of the queries and documents (only the title field) are 3.15and 6.77 respectively. In our experiments, we set maximum allowable length of query and documentto 10 and 20 respectively.

In the document ranking task, we need to rank the most relevant (e.g., most clickable) documenton top. We used three standard ranking metrics, mean average precision (MAP), mean reciprocalrank (MRR) and normalized discounted cumulative gain (NDCG) metric computed at positions one,three, five and ten, to measure the performance. Since AOL search log only contains clicked doc-uments, we constructed the ranking candidates by the top ranked documents by BM25 (Robertsonet al., 2009). Each query in the test set consists of 50 candidate documents including the clickedones. However, to reduce the training time and memory use, each query in training and developmentset only contains 5 candidates.

In the query suggestion task, we need to suggest the most semantically related query to users. Aswe do not have user feedback on the suggested queries, we treated the users’ next submitted queryas ground-truth (Sordoni et al., 2015), and used the BLEU scores (Papineni et al., 2002) as theevaluation metric, which is a popularly used metric in machine translation and text generation tasks.In addition, following (Santos et al., 2013; Sordoni et al., 2015), we evaluated the query suggestionquality by mean reciprocal rank (MRR), i.e., to test if the algorithm can rank the users’ next queryon top of its recommendation list. In this set of experiments, given a set of candidate queries, querysuggestion models give a likelihood score of generating the candidates. We follow Sordoni et al.(2015) for dataset split and generate candidates using a co-occurrence based suggestion model. LikeSordoni et al. (2015), we give the anchor queries (second last query of a session) as an input to thequery suggestion model and evaluates the rank of the next query among the candidates.

4.2 BASELINES AND IMPLEMENTATION DETAILS

Document Ranking Baselines. We compared M-NSRF and M-Match Tensor with word-basedbaselines and neural network-based baselines. Word-based baselines include query likelihood model

6

Page 7: M -T LEARNING FOR DOCUMENT RANKING AND Q S · query suggestion for web search. It consists of two major components, a document ranker and a query recommender. Document ranker combines

Published as a conference paper at ICLR 2018

Table 1: Comparison of document ranking models over the AOL search log.

Model Type Model Name MAP MRR NDCG@1 @3 @5 @10

TraditionalIR-models

BM25 0.164 0.172 0.121 0.136 0.141 0.156QL 0.139 0.146 0.088 0.108 0.122 0.133

Embedding-based ESM 0.214 0.179 0.118 0.127 0.139 0.158

RepresentationFocused

DSSM 0.263 0.287 0.152 0.206 0.248 0.315CLSM 0.465 0.505 0.369 0.441 0.482 0.523ARC-I 0.383 0.413 0.238 0.343 0.404 0.467

InteractionFocused

DRMM 0.277 0.316 0.221 0.242 0.267 0.304ARC-II 0.423 0.455 0.294 0.386 0.442 0.501

Representation andInteraction Focused

DUET 0.272 0.301 0.152 0.212 0.263 0.341Match-Tensor 0.613 0.621 0.568 0.572 0.596 0.618

Neural Session Model(this paper) NSRF 0.553 0.568 0.481 0.526 0.555 0.574

Multi-task Model(this paper)

M-NSRF 0.581 0.603 0.523 0.568 0.583 0.614M-Match-Tensor 0.621 0.634 0.572 0.578 0.602 0.632

based on Dirichlet smoothing (QL) (Ponte & Croft, 1998) and BM25 (Robertson et al., 2009). Inaddition, following (Mitra et al., 2016), we investigated the ranking performance of a simple wordembedding-based model using GloVe word embeddings (Pennington et al., 2014). To compare withneural models, we consider baselines broadly categorized in representation-focused, interaction-focused and a combination of both. Representation-focused neural baselines include: DSSM (Huanget al., 2013), CLSM (Shen et al., 2014), ARC-I (Hu et al., 2014) and interaction-focused baselinesinclude: ARC-II (Hu et al., 2014) and DRMM (Guo et al., 2016a). A combination of representationand interaction focused models include: DUET (Mitra et al., 2017) and Match Tensor (Jaech et al.,2017). Details of these models are provided in Appendix C. We implemented all the baseline modelsin PyTorch.

Query Suggestion Baselines. To evaluate the performance on query suggestion task, we considerthree baseline methods including Seq2seq model proposed by Bahdanau et al. (2015), Seq2seq withglobal attention mechanism (Luong et al., 2015) and HRED-qs (Sordoni et al., 2015). Details ofthese models are provided in appendix D. We implemented all three baselines in PyTorch and opti-mized using negative log-likelihood loss as in Eq. (3).

Implementation Details of M-NSRF. The model was trained end-to-end and we used mini-batchSGD with Adam (Kingma & Ba, 2014) for optimization. with the two momentum parameters set to0.9 and 0.999 respectively. We use 300-dimensional word vectors trained with GloVe (Penningtonet al., 2014) on 840 billion of tokens to initialize the word embeddings. Out-of-vocabulary wordswere randomly initialized by sampling values from a zero-mean unit-variance normal distribution.All training used a mini-batch size of 32 to fit in single GPU memory. Learning rate was fixed to0.001. We used dropout (0.20) (Srivastava et al., 2014) and early stopping with a patience of 5 epochsfor regularization. We set λ = 0.1 for entropy regularization in Eq. (3). M-NSRF is implemented inPyTorch and it runs on a single GPU (TITAN X) with roughly a runtime of 90 minutes per epoch.In general, M-NSRF runs up to 20 epochs and we select the model that achieves the minimum losson the development set.

4.3 EVALUATION RESULTS

Document Ranking Quality. Table 1 shows the performance of NSRF, M-NSRF, M-Match-Tensorand other baseline models. NSRF significantly outperforms all the baselines except the Match-Tensor model. However, the model size of NSRF is much smaller than Match-Tensor. With multi-task learning, both M-NSRF and M-Match-Tensor outperform NSRF and Match-Tensor, respec-tively. To study the advantage of multi-task learning on our proposed approach, we trained M-NSRFonly on document ranking task (noted as NSRF in Table 1) and observed significant performancedrop which endorses the mutual benefit of joint learning of these two tasks via multi-task learning.

Existing neural ranking models, like DRMM and DUET architecture, achieved sub-optimal per-formance in our experiments. We believe because of the simple architecture of DRMM with fewhundreds of parameters, the model fell behind to show competitive performance on the evaluation

7

Page 8: M -T LEARNING FOR DOCUMENT RANKING AND Q S · query suggestion for web search. It consists of two major components, a document ranker and a query recommender. Document ranker combines

Published as a conference paper at ICLR 2018

Table 2: Examples of next query suggested by M-NSRF given all previous queries in a session.

Previous session queries types of weapons of mass destruction, weapons of mass destruc-tion, nuclear weapons

Next user query biological weaponsSuggested next query destructive nuclear weapons

Previous session queries resume template, resume template free, resume word perfecttemplate free, wordperfect com

Next user query wordperfect resume templates freeSuggested next query free microsoft word templates

Table 3: Comparison of different query suggestion models.

Model Name BLEU MRR2

1 2 3 4Seq2seq 24.5 9.7 4.5 1.9 0.229Seq2seq with attention 28.1 15.7 10.4 8.5 0.252HRED-qs 26.4 13.6 7.9 5.8 0.231HRED-qs w/ entropy regularizer 27.6 15.1 9.2 6.7 0.233M-NSRF 26.8 14.1 8.4 6.1 0.235M-NSRF w/ entropy regularizer 28.6 16.7 10.2 8.3 0.238

dataset. On the other hand, we believe that the use of smaller number of top character n-graphs (inour case, 5000) by the DUET architecture limits its effectiveness in modeling representation andinteraction focused features to compute matching quality between query and document. We have tonote that, some negative results have been reported for document title-based ad-hoc retrieval tasks(Guo et al., 2016a); and thus in our future work, we plan to investigate M-NSRF and all baselinemodels’ performance with document body content considered.

Query Suggestion Accuracy. Examples of the predicted queries by M-NSRF given precedingqueries from the same session is presented in Table 2 (more examples are provided in AppendixE). The quantitative comparison results in this task between our proposed framework and baselinemodels are presented in Table 3. While M-NSRF and HRED-qs consider information from allpreceding queries in the same session, the other two baselines only consider the current query topredict the next one. Table 3 shows that M-NSRF outperformed seq2seq and HRED-qs baselinesin all measured BLUE scores, and MRR; but the Seq2seq with attention baseline performed betterthan M-NSRF in terms of BLEU-3, BLEU-4 and MRR3.

We investigated the advantage of attention mechanism in this particular task and found it performswell when there is a considerable overlap between the input and output queries. Specifically, inmore than 25% of the test sessions, the input queries and their next queries are exactly the same;and the attention mechanism encourages the model to repeat words from the input query. When werestrict the experiment on test sessions that have no overlap between the input and its next query, theSeq2seq with attention baseline encountered significant performance drop, while M-NSRF providedimproved performance compared to all baselines (approx. 5%, 10% and 1% improvement in termsof MRR over Seq2seq, Seq2seq with attention and HRED-qs model). The major reason for this im-provement is that our model leverages the global session information and is therefore not restrictedby the input query and is able to generate related but totally new queries (as shown in Appendix E).

We also observed that entropy-based regularization helped M-NSRF in predicting the next query, asshown in Table 3 that M-NSRF without regularization achieved worse performance in all measuredBLUE scores. We further investigated the utility of regularization for the baselines. We found sig-nificant (∼1.2%) improvement for the HRED-qs, but the performance improvement for the seq2seqand seq2seq with attention mechanism is rather marginal (∼0.2%). However, the regularizationtechnique does not introduce any significant difference to the ranking based evaluation.

3Note that Sordoni et al. (2015) used their neural models output as a feature for a learning-to-rank model.To understand the effectiveness of these neural models in query suggestion task, we directly compared theirranking quality. Therefore, the numbers reported in Table 3 are not comparable to that in Sordoni et al. (2015).

8

Page 9: M -T LEARNING FOR DOCUMENT RANKING AND Q S · query suggestion for web search. It consists of two major components, a document ranker and a query recommender. Document ranker combines

Published as a conference paper at ICLR 2018

Table 4: Ablation study for performance analysis of M-NSRF. Statistical significances are comparedwith NSRF’s full model and presented in bold-faced. † indicates M-NSRF is trained to learn wordembeddings, no pre-trained embeddings were used.

NSRF Variant MAP NDCG@1 NDCG@3 NDCG@10Full model 0.581 0.523 0.568 0.614Fixed embeddings 0.252 (−0.329) 0.182 (−0.341) 0.216 (−0.352) 0.289 (−0.325)Learned embeddings† 0.302 (−0.279) 0.222 (−0.301) 0.261 (−0.307) 0.297 (−0.317)Mean-pool 0.576 (−0.005) 0.515 (−0.008) 0.561 (−0.007) 0.608 (−0.006)BiLSTM-last 0.563 (−0.018) 0.505 (−0.018) 0.541 (−0.027) 0.594 (−0.020)M-NRF 0.553 (−0.028) 0.494 (−0.029) 0.544 (−0.024) 0.582 (−0.032)GloVe 6B 50d 0.247 (−0.324) 0.196 (−0.329) 0.232 (−0.336) 0.296 (−0.348)GloVe 6B 100d 0.312 (−0.269) 0.241 (−0.282) 0.273 (−0.295) 0.306 (−0.308)GloVe 6B 200d 0.378 (−0.203) 0.356 (−0.167) 0.447 (−0.121) 0.498 (−0.116)Q128D256S512 0.562 (−0.019) 0.507 (−0.016) 0.544 (−0.024) 0.582 (−0.032)Q512D1024S2048 0.586 (+0.005) 0.528 (+0.005) 0.571 (+0.003) 0.617 (+0.003)

Comparing to HRED-qs, we conclude multi-task learning also helps M-NSRF achieve improvedperformance in query suggestion task, as the baseline employs a very similar model as ours for querysuggestion. The comparison with Seq2seq with attention suggests adding attention over sessioninformation is a promising direction to further improve query suggestions quality as it emphasizesthe information carried by the immediately previous query. We will pursue this in our future work.

4.4 ABLATION STUDY ON M-NSRF

We conducted experiments to better understand the effectiveness of different components in the M-NSRF. We also analyzed the impact of word embeddings and hidden units dimension in M-NSRF’sperformance. Our findings are presented in Table 4.

Impact of Different Model Components. To study the effect of different model components, wecompared the full M-NSRF with several simpler versions of the model. At first, we turned off train-ing for word embeddings and found a significant decrease in performance. In our training dataset,we have roughly |OOV | = 26k out-of-vocabulary words and thus, training the word embeddingsturned out to be very important on this data set. Also, we investigated the role of pre-trained wordembeddings (ex., GloVe embeddings) and found significant performance decrease (27.9% drop inMAP) if we train M-NSRF without any pre-trained word embeddings. Hence, we can concludethat the use of pre-trained word embeddings and training them further is important to achieve bet-ter performance in M-NSRF. Second, we investigated the advantages of using max-pooling overmean-pooling and considering the last hidden recurrent state for query and document representa-tion. We observed that max-pooling and mean-pooling provide almost the same performance whilebiLSTM-last approach lags slightly. To further analyze the features identified by the query and doc-ument encoders using the max-pooling technique, we followed the idea of visualization proposedin (Conneau et al., 2017). We provide an example in Figure 2 where document 1 is clicked by theuser (a positive example) and document 2, 3 is retrieved by BM25 but not clicked (negative exam-ples). We observed that the query and document encoders identified distinct features (e.g., the wordpriceline in the first document’s title is most important) which help differentiate between clickedand unclicked documents.

In another variant of M-NSRF, we did not use the session recurrent state when computing relevancescore for the candidate documents to examine the influence of previous queries from the same ses-sion on the ranking performance. We refer to this variant of M-NSRF as multi-task neural relevancemodel (M-NRF). From Table 4 we can find that without session information the performance dropsby 2.8% in terms of MAP. We further investigated and found the session information helps partic-ularly for longer sessions (session length > 5). In our evaluation dataset, we have roughly 5000sessions of length greater than 5 (91 thousands in total).

Impact of Dimensionality. We further study the impact of dimensionality of the word embeddings,query, document and session latent vectors. In M-NSRF, we set the dimension of query, documentand session latent vectors to 256, 512 and 1024 respectively. As shown in Table 4, decreasing the

9

Page 10: M -T LEARNING FOR DOCUMENT RANKING AND Q S · query suggestion for web search. It consists of two major components, a document ranker and a query recommender. Document ranker combines

Published as a conference paper at ICLR 2018

Figure 2: Example showing query and document term importance identified by M-NSRF whileranking candidate documents for the given query.

dimensions of latent vectors, decreases the performance; while increasing the dimensions further,does not affect the performance significantly. We also experimented with different dimensions ofpre-trained word embeddings (50d, 100d and 200d GloVe (Pennington et al., 2014) embeddings).Word embeddings of different dimensions provide different granularity of semantic similarity; withlower dimensionality, the similarity between word embeddings might be coarse and thus hard tocapture matching between two text sequences. In our experiment, we found 300 dimension forembeddings works significantly better than other dimensionality on this data set.

5 CONCLUSIONS

Existing deep neural models for ad-hoc retrieval often omit session information and are only trainedon individual query-document pairs. In this work, we propose a context-aware multi-task neuralsession relevance framework which works in a sequence to sequence fashion, and show that sharingsession-level latent recurrent states across document ranking and query suggestion tasks benefitseach other. Our experiments and analysis not only demonstrate the effectiveness of the proposedframework, but also provide useful intuitions about the advantages of multi-task learning involvingdeep neural networks for two different information retrieval tasks.

As our future work, we would like to leverage the content from document body and click sequenceto update M-NSRF (especially the session-level recurrent states) so that we can further explore thepotential of the proposed framework for ad-hoc retrieval. As attention mechanism shows promise inimproving query suggestion performance, we also also explore it in our multi-task learning setting.In addition, a broad research direction is to go beyond session boundaries to model users’ long-termsearch goals to enhance personalized search results and query suggestions.

Acknowledgement. This work was supported in part by National Science Foundation Grant IIS-1760523, IIS-1553568, IIS-1618948, and the NVIDIA Hardware Grant.

REFERENCES

Ricardo A Baeza-Yates, Carlos A Hurtado, Marcelo Mendoza, et al. Query recommendation using query logsin search engines. In EDBT workshops, volume 3268, pp. 588–596. Springer, 2004.

Dzmitry Bahdanau, Kyunghyun Cho, and Yoshua Bengio. Neural machine translation by jointly learning toalign and translate. International Conference on Learning Representations, 2015.

Jing Bai, Ke Zhou, Guirong Xue, Hongyuan Zha, Gordon Sun, Belle Tseng, Zhaohui Zheng, and Yi Chang.Multi-task learning for learning to rank in web search. In Proceedings of the 18th ACM International onConference on Information and Knowledge Management, pp. 1549–1552. ACM, 2009.

Huanhuan Cao, Daxin Jiang, Jian Pei, Qi He, Zhen Liao, Enhong Chen, and Hang Li. Context-aware querysuggestion by mining click-through and session data. In Proceedings of the 14th ACM SIGKDD internationalconference on Knowledge discovery and data mining, pp. 875–883. ACM, 2008.

Rich Caruana. Multitask learning. In Learning to learn, pp. 95–133. Springer, 1998.

Wanyu Chen, Fei Cai, Honghui Chen, and Maarten de Rijke. Personalized query suggestion diversification.Proceedings of the 37th ACM SIGIR conference on Research and development in information retrieval, 2017.

Ronan Collobert and Jason Weston. A unified architecture for natural language processing: Deep neural net-works with multitask learning. In Proceedings of the 25th international conference on Machine learning,pp. 160–167. ACM, 2008.

10

Page 11: M -T LEARNING FOR DOCUMENT RANKING AND Q S · query suggestion for web search. It consists of two major components, a document ranker and a query recommender. Document ranker combines

Published as a conference paper at ICLR 2018

Alexis Conneau, Douwe Kiela, Holger Schwenk, Loıc Barrault, and Antoine Bordes. Supervised learning ofuniversal sentence representations from natural language inference data. In Proceedings of the 2017 Confer-ence on Empirical Methods in Natural Language Processing, pp. 670–680. Association for ComputationalLinguistics, 2017.

W Bruce Croft, Donald Metzler, and Trevor Strohman. Search engines: Information retrieval in practice,volume 283. Addison-Wesley Reading, 2010.

Li Deng, Geoffrey Hinton, and Brian Kingsbury. New types of deep neural network learning for speech recog-nition and related applications: An overview. In Acoustics, Speech and Signal Processing (ICASSP), 2013IEEE International Conference on, pp. 8599–8603. IEEE, 2013.

Henry Feild and James Allan. Task-aware query recommendation. In Proceedings of the 36th ACM SIGIRconference on Research and development in information retrieval, pp. 83–92. ACM, 2013.

Jianfeng Gao, Li Deng, Michael Gamon, Xiaodong He, and Patrick Pantel. Modeling interestingness with deepneural networks, June 13 2014. US Patent App. 14/304,863.

Dario Garigliotti and Krisztian Balog. Generating query suggestions to support task-based search. In Pro-ceedings of the 40th ACM SIGIR conference on Research and development in information retrieval, pp.1153–1156. ACM, 2017.

Ross Girshick. Fast r-cnn. In Proceedings of the IEEE international conference on computer vision, pp. 1440–1448, 2015.

Jiafeng Guo, Yixing Fan, Qingyao Ai, and W Bruce Croft. A deep relevance matching model for ad-hocretrieval. In Proceedings of the 25th ACM International on Conference on Information and KnowledgeManagement, pp. 55–64. ACM, 2016a.

Jiafeng Guo, Yixing Fan, Qingyao Ai, and W Bruce Croft. Semantic matching by non-linear word transporta-tion for information retrieval. In Proceedings of the 25th ACM International on Conference on Informationand Knowledge Management, pp. 701–710. ACM, 2016b.

Qi He, Daxin Jiang, Zhen Liao, Steven CH Hoi, Kuiyu Chang, Ee-Peng Lim, and Hang Li. Web query recom-mendation via sequential query prediction. In Data Engineering, 2009. ICDE’09. IEEE 25th InternationalConference on, pp. 1443–1454. IEEE, 2009.

Sepp Hochreiter and Jurgen Schmidhuber. Long short-term memory. Neural computation, 9(8):1735–1780,1997.

Baotian Hu, Zhengdong Lu, Hang Li, and Qingcai Chen. Convolutional neural network architectures formatching natural language sentences. In Advances in neural information processing systems, pp. 2042–2050, 2014.

Po-Sen Huang, Xiaodong He, Jianfeng Gao, Li Deng, Alex Acero, and Larry Heck. Learning deep structuredsemantic models for web search using clickthrough data. Proceedings of the 22nd ACM International onConference on Information and Knowledge Management, pp. 2333–2338, 2013.

Aaron Jaech, Hetunandan Kamisetty, Eric Ringger, and Charlie Clarke. Match-tensor: a deep relevance modelfor search. arXiv preprint arXiv:1701.07795, 2017.

J Jansen Bernard, Amanda Spink, Chris Blakely, and Sherry Koshman. Defining a session on web searchengines: Research articles. Journal of the American Society for Information Science and Technology, 58(6):862–871, 2007.

Jyun-Yu Jiang, Yen-Yu Ke, Pao-Yu Chien, and Pu-Jen Cheng. Learning user reformulation behavior for queryauto-completion. In Proceedings of the 37th ACM SIGIR conference on Research and development in infor-mation retrieval, pp. 445–454. ACM, 2014.

Diederik Kingma and Jimmy Ba. Adam: A method for stochastic optimization. arXiv preprintarXiv:1412.6980, 2014.

Steve Lawrence, C Lee Giles, Ah Chung Tsoi, and Andrew D Back. Face recognition: A convolutional neural-network approach. IEEE transactions on neural networks, 8(1):98–113, 1997.

Pengfei Liu, Xipeng Qiu, and Xuanjing Huang. Deep multi-task learning with shared memory for text classifi-cation. In Proceedings of the 2016 Conference on Empirical Methods in Natural Language Processing, pp.118–127, 2016a.

Pengfei Liu, Xipeng Qiu, and Xuanjing Huang. Recurrent neural network for text classification with multi-tasklearning. Proceedings of the Twenty-Fifth International Joint Conference on Artificial Intelligence, 2016b.

11

Page 12: M -T LEARNING FOR DOCUMENT RANKING AND Q S · query suggestion for web search. It consists of two major components, a document ranker and a query recommender. Document ranker combines

Published as a conference paper at ICLR 2018

Tie-Yan Liu et al. Learning to rank for information retrieval. Foundations and Trends R© in Information Re-trieval, 3(3):225–331, 2009.

Xiaodong Liu, Jianfeng Gao, Xiaodong He, Li Deng, Kevin Duh, and Ye-Yi Wang. Representation learningusing multi-task deep neural networks for semantic classification and information retrieval. In Proceedingsof the 2015 Conference of the North American Chapter of the Association for Computational Linguistics:Human Language Technologies, pp. 912–921, 2015.

Thang Luong, Hieu Pham, and Christopher D. Manning. Effective approaches to attention-based neural ma-chine translation. In Proceedings of the 2015 Conference on Empirical Methods in Natural Language Pro-cessing, pp. 1412–1421, 2015.

Bhaskar Mitra and Nick Craswell. Query auto-completion for rare prefixes. In Proceedings of the 24th ACMInternational on Conference on Information and Knowledge Management, pp. 1755–1758. ACM, 2015.

Bhaskar Mitra, Eric Nalisnick, Nick Craswell, and Rich Caruana. A dual embedding space model for documentranking. arXiv preprint arXiv:1602.01137, 2016.

Bhaskar Mitra, Fernando Diaz, and Nick Craswell. Learning to match using local and distributed representa-tions of text for web search. pp. 1291–1299. International World Wide Web Conferences Steering Commit-tee, 2017.

Liang Pang, Yanyan Lan, Jiafeng Guo, Jun Xu, Shengxian Wan, and Xueqi Cheng. Text matching as imagerecognition. Association for the Advancement of Artificial Intelligence, pp. 2793–2799, 2016.

Kishore Papineni, Salim Roukos, Todd Ward, and Wei-Jing Zhu. Bleu: a method for automatic evaluation ofmachine translation. In Proceedings of the 40th annual meeting on association for computational linguistics,pp. 311–318. Association for Computational Linguistics, 2002.

Greg Pass, Abdur Chowdhury, and Cayley Torgeson. A picture of search. In InfoScale, volume 152, pp. 1,2006.

Nanyun Peng and Mark Dredze. Multi-task domain adaptation for sequence tagging. In Proceedings of the 2ndWorkshop on Representation Learning for NLP, pp. 91–100, 2017.

Nanyun Peng, Hoifung Poon, Chris Quirk, Kristina Toutanova, and Wen-tau Yih. Cross-sentence n-ary relationextraction with graph lstms. Transactions of the Association for Computational Linguistics, 5:101–115,2017.

Jeffrey Pennington, Richard Socher, and Christopher Manning. Glove: Global vectors for word representation.Proceedings of the 2014 conference on empirical methods in natural language processing (EMNLP), pp.1532–1543, 2014.

Jay M Ponte and W Bruce Croft. A language modeling approach to information retrieval. In Proceedings ofthe 21st ACM SIGIR conference on Research and development in information retrieval, pp. 275–281. ACM,1998.

Stephen Robertson, Hugo Zaragoza, et al. The probabilistic relevance framework: Bm25 and beyond. Founda-tions and Trends R© in Information Retrieval, 3(4):333–389, 2009.

Rodrygo LT Santos, Craig Macdonald, and Iadh Ounis. Learning to rank query suggestions for adhoc anddiversity search. Information Retrieval, 16(4):429–451, 2013.

Yelong Shen, Xiaodong He, Jianfeng Gao, Li Deng, and Gregoire Mesnil. A latent semantic model withconvolutional-pooling structure for information retrieval. In Proceedings of the 23rd ACM International onConference on Information and Knowledge Management, pp. 101–110. ACM, 2014.

Alessandro Sordoni, Yoshua Bengio, Hossein Vahabi, Christina Lioma, Jakob Grue Simonsen, and Jian-YunNie. A hierarchical recurrent encoder-decoder for generative context-aware query suggestion. Proceedingsof the 24th ACM International Conference on Information and Knowledge Management, pp. 553–562, 2015.

Nitish Srivastava, Geoffrey E Hinton, Alex Krizhevsky, Ilya Sutskever, and Ruslan Salakhutdinov. Dropout:a simple way to prevent neural networks from overfitting. Journal of machine learning research, 15(1):1929–1958, 2014.

Ilya Sutskever, Oriol Vinyals, and Quoc V Le. Sequence to sequence learning with neural networks. InAdvances in neural information processing systems, pp. 3104–3112, 2014.

Abhinav Thanda and Shankar M Venkatesan. Multi-task learning of deep neural networks for audio visualautomatic speech recognition. arXiv preprint arXiv:1701.02477, 2017.

Ji-Rong Wen, Jian-Yun Nie, and Hong-Jiang Zhang. Clustering user queries of a search engine. In Proceedingsof the 10th international conference on World Wide Web, pp. 162–168. acm, 2001.

12

Page 13: M -T LEARNING FOR DOCUMENT RANKING AND Q S · query suggestion for web search. It consists of two major components, a document ranker and a query recommender. Document ranker combines

Published as a conference paper at ICLR 2018

A MULTITASK NEURAL SESSION RELEVANCE FRAMEWORK

Figure 3: Architecture of the Multi-task Neural Session Relevance Framework (M-NSRF). M-NSRFuses bi-LSTM with max pooling to form query and document representations and use LSTM togather session-level information. These recurrent states (current query representation and session-level recurrent state, which summarizes all previous queries) are used by query decoder and docu-ment ranker for predicting next query and computing relevance scores.

B MULTITASK MATCH-TENSOR ARCHITECTURE

Figure 4: Architecture of the Multi-task Match-Tensor Model (M-Match-Tensor).

C DOCUMENT RANKING BASELINES

Deep semantic similarity model, DSSM (Huang et al., 2013) maps words to letter tri-grams us-ing a word-hashing technique and uses a feed-forward neural network to build representations forboth query and document. Similarity, convolutional latent semantic model, CLSM (Shen et al.,2014) uses word-hashing technique and uses convolutional neural networks (CNN) to build queryand document representations. To compute relevance between query and document, both DSSM and

13

Page 14: M -T LEARNING FOR DOCUMENT RANKING AND Q S · query suggestion for web search. It consists of two major components, a document ranker and a query recommender. Document ranker combines

Published as a conference paper at ICLR 2018

CLSM uses cosine similarity. ARC-I (Hu et al., 2014) uses CNN to form query and document repre-sentations and employs a multi-layer perceptron to compute relevance score. In our implementation,we used 128 convolution filters of size 1, 2 and 256 filters of size 3.

ARC-II (Hu et al., 2014) was proposed by focusing on learning hierarchical matching patternsfrom local interactions using a CNN. To keep the ARC-II model simple, we use two layers of 2dconvolution and max-pooling each and two-layer feed forward neural network to compute relevancescore. DRMM (Guo et al., 2016a) aims to perform term matching over histogram-based featuresignoring the actual position of matches. In DRMM, histogram-based features are computed usingexact term matching and pretrained word embeddings based cosine similarities. In principal, thehistogram counts the number of word pairs at different similarity levels. The counts are combinedby a feed forward network to produce final ranking scores.

The DUET (Mitra et al., 2017) model composed of a local and distributed model where the dis-tributed model projects the query and the document text into an embedding space before matching,while the local model operates over an interaction matrix comparing every query term to every doc-ument term. Similarly, Match-Tensor (Jaech et al., 2017) model incorporates both immediate andlarger contexts in a given document when comparing document to a query.

D QUERY SUGGESTION BASELINES

Seq2seq model proposed by Bahdanau et al. (2015) is a general neural network architecture that canbe applied to the task where both input and output consist of a sequence of tokens. This method havebeen shown successful in machine translation and sequential tagging. Because different input tokensmay contribute to each output token differently, attention mechanism which learns a weight betweeneach input-output token pair can further improve the Seq2seq model. In this paper, we consider aSeq2seq with global attention method proposed by Luong et al. (2015), which is suitable for shorttext such as web queries. HRED-qs suggested by Sordoni et al. (2015) is very close to our workwhich proposed to use a hierarhical recurrent encoder-decoder approach by considering sessioninformation for context-aware query suggestion.

E MORE EXAMPLES OF QUERY SUGGESTION BY M-NSRF

Previous session queries discount pet supplies, homes for rent smyrna georgiaNext user query homes for rent atlanta georgiaSuggested next query pet friendly rentals in georgiaPrevious session queries language aptitude test, foreign language aptitude testNext user query american idolSuggested next query american language associationPrevious session queries saturday night fever, saturday night fever nj bandNext user query new jersey cover bandSuggested next query saturday night livePrevious session queries pregnancy, abortion, abortion clinicsNext user query tampa abortionSuggested next query abortion clinics in florida

Previous session queries ncaa basketball, ncaa basketball trees, ncaa mens basketball bracket,sportscenter

Next user query mens ncaa basketball oddsSuggested next query espn

Previous session queries childhood autism rating scale, childhood autism rating scale free,autism screening questionnaire

Next user query pervasive developmental disorderSuggested next query how to do questionnaire

14