probabilistic latent semantic visualization: topic model ... · probabilistic latent semantic...

10

Click here to load reader

Upload: buidien

Post on 22-Mar-2019

214 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: Probabilistic Latent Semantic Visualization: Topic Model ... · Probabilistic Latent Semantic Visualization: Topic Model for Visualizing Documents ... which we call Probabilistic

Probabilistic Latent Semantic Visualization:Topic Model for Visualizing Documents

Tomoharu Iwata Takeshi Yamada Naonori UedaNTT Communication Science Laboratories

2-4 Hikaridai, Seika-cho, Soraku-gun, Kyoto, Japan{iwata,yamada,ueda}@cslab.kecl.ntt.co.jp

ABSTRACTWe propose a visualization method based on a topic modelfor discrete data such as documents. Unlike conventionalvisualization methods based on pairwise distances such asmulti-dimensional scaling, we consider a mapping from thevisualization space into the space of documents as a gener-ative process of documents. In the model, both documentsand topics are assumed to have latent coordinates in a two-or three-dimensional Euclidean space, or visualization space.The topic proportions of a document are determined by thedistances between the document and the topics in the vi-sualization space, and each word is drawn from one of thetopics according to its topic proportions. A visualization,i.e. latent coordinates of documents, can be obtained byfitting the model to a given set of documents using the EMalgorithm, resulting in documents with similar topics beingembedded close together. We demonstrate the effectivenessof the proposed model by visualizing document and moviedata sets, and quantitatively compare it with conventionalvisualization methods.

Categories and Subject DescriptorsH.2.8 [Database Management]: Database Applications—Data Mining ; I.2.6 [Artificial Intelligence]: Learning; I.5.1[Pattern Recognition]: Model—Statistical

General TermsAlgorithms

KeywordsVisualization, Topic model, Probabilistic latent semanticanalysis

1. INTRODUCTIONRecently there has been great interest in topic models for an-alyzing documents and other discrete data. A topic modelis a hierarchical probabilistic model, in which a document is

KDD’08, August 24–27, 2008, Las Vegas, Nevada, USA.

modeled as a mixture of topics, where a topic is modeled asa probability distribution over words. Probabilistic LatentSemantic Analysis (PLSA) [10] and Latent Dirichlet Alloca-tion (LDA) [4] are representative topic models, and they areused for a wide variety of applications such as informationretrieval, text clustering and collaborative filtering.

In this paper, we propose a nonlinear visualization methodbased on a topic model, which we call Probabilistic LatentSemantic Visualization (PLSV). Visualization is a usefultool for understanding complex and high dimensional data,and it enables us to browse intuitively through huge amountsof data. A number of document visualization methods havebeen proposed [8, 22], and the importance of visualizing doc-uments is increasing since documents such as web pages,blogs, e-mails, patents and scientific articles are being accu-mulated rapidly.

In PLSV, we consider a mapping from the visualizationspace into the space of documents as a generative processof documents. Both documents and topics are assumed tohave latent coordinates in a two- or three-dimensional Eu-clidean space, or visualization space. The topic proportionsof a document are determined by the Euclidean distances be-tween the document coordinate and the topic coordinates.If a document is located near a topic, the probability thatthe document has the topic becomes high. Each word in adocument is drawn from one of the topics according to itstopic proportions, as in other topic models. The parametersin PLSV including latent coordinates of documents can beestimated by fitting the model to the given set of documentsusing the EM algorithm.

A number of visualization methods have been proposed, suchas Multi-Dimensional Scaling (MDS) [20], Isomap [19] andLocally Linear Embedding (LLE) [17]. However, most ofthese methods take no account of the latent structure in thegiven data such as topics in the case of document data. Forexample, MDS embeds samples so that pairwise distances inthe visualization space accurately reflect pairwise distancesin the original space, and it does not consider topics explic-itly. On the other hand, because PLSV considers topics,documents with similar semantics are embedded close to-gether even if they do not share any words.

PLSA or LDA can extract a low-dimensional representa-tion of a document as topic proportions. However, they arenot appropriate for visualization since they cannot express

Page 2: Probabilistic Latent Semantic Visualization: Topic Model ... · Probabilistic Latent Semantic Visualization: Topic Model for Visualizing Documents ... which we call Probabilistic

more than three topics in the two-dimensional visualizationspace, and the representation is an embedding in the simplexspace but not in the Euclidean space. In contrast, PLSV canexpress any number of topics even with a two-dimensionalrepresentation, and it embeds documents in the Euclideanspace, which is a metric space that most closely correspondsto our intuitive understanding of space. The topic propor-tions estimated by PLSA or LDA can be embedded in theEuclidean space by Parametric Embedding (PE) [11], whichcan employ a set of topic proportions as input. However,the topic proportions may not be suitable when they areembedded in the visualization space since these topics areestimated in a different space from the visualization space.Moreover, errors accumulated in the topic estimation pro-cess with PLSA or LDA cannot be corrected in the embed-ding process with PE since they are modularized, and thismethod may result in poor visualization. On the other hand,since PLSV simultaneously estimates topics and visualizesdocuments in one probabilistic framework, topics are esti-mated so as to be optimal when documents are embeddedin the two- or three-dimensional visualization space.

The remainder of this paper is organized as follows. InSection 2, we formulate PLSV, and describe its parame-ter estimation procedures. In Section 3, we briefly reviewrelated work. In Section 4, we demonstrate the effective-ness of PLSV by visualizing document and movie data sets,and quantitatively compare it with conventional visualiza-tion methods. Finally, we present concluding remarks anda discussion of future work in Section 5.

2. PROBABILISTIC LATENT SEMANTICVISUALIZATION

2.1 ModelIn the following discussion, we assume that the given dataare documents. However, PLSV is applicable to other dis-crete data, such as purchase logs in collaborative filteringand gene sequences in bioinformatics.

Suppose that we have a set of N documents C = {wn}Nn=1.

Each document is represented by a sequence of Mn words de-noted by wn = (wn1, · · · , wnMn), where wnm ∈ {1, · · · , W}is the mth word in the sequence of the nth document, Mn

is the number of words in the nth document, and W is thevocabulary size.

PLSV is a probabilistic topic model for finding the embed-ding of documents with coordinates X = {xn}N

n=1, wherexn = (xn1, · · · , xnD) is a coordinate of the nth documentin the visualization space, and D is its dimensionality, usu-ally D = 2 or 3. We assume that there are Z topics in-dexed by {z}Z

z=1, and each topic has its associated coordi-nate φz = (φz1, · · · , φzD) in the visualization space. Thetopic proportion of a document is determined by its Eu-clidean distances from topics in the visualization space asfollows:

P (z|xn,Φ) =exp

`

− 12‖ xn − φz ‖2

´

PZz′=1 exp

`

− 12‖ xn − φz′ ‖2

´, (1)

where P (z|xn,Φ) is the zth topic proportion of the nth doc-

ument,PZ

z=1 P (z|xn,Φ) = 1, Φ = {φz}Zz=1 is the set of

topic coordinates, and ‖ · ‖ represents the Euclidean norm

φ θ

z wx

NM

Z

Figure 1: Graphical model representation of PLSV.

in the visualization space. If we assume this topic propor-tion, when the Euclidean distance between coordinates ofdocument xn and topic φz is small, the topic proportionP (z|xn,Φ) becomes high. Since documents located close to-gether in the visualization space are similar distances fromtopic coordinates, they have similar topic proportions.

PLSV assumes the following generative process for a set ofdocuments C :

1. For each topic z = 1, · · · , Z:

(a) Draw word probability distribution

θz ∼ Dirichlet(α).

(b) Draw topic coordinateφz ∼ Normal(0, β−1I).

2. For each document n = 1, · · · , N :

(a) Draw document coordinatexn ∼ Normal(0, γ−1I).

(b) For each word m = 1, · · · , Mn:

i. Draw topic

znm|xn,Φ ∼ Mult“

{P (z|xn,Φ)}Zz=1

.

ii. Draw wordwnm|znm,Θ ∼ Mult

{P (w|znm,Θ)}Ww=1

.

Here Θ = {θz}Zz=1 is a set of word probabilities, θz =

{θzw}Ww=1,

P

w θzw = 1, and θzw = P (w|z,Θ) is the proba-bility that the wth word occurs given the zth topic. As inLDA, each word wnm is sampled from a topic-specific multi-nomial distribution, where the multinomial parameters θz

are generated by a Dirichlet distribution that is conjugateto multinomial. The coordinates xn and φz are assumed tobe generated by zero-mean spherical Gaussian distributionsfor stabilizing the visualization. Given xn, Φ and Θ, theprobability of wn is given as follows:

P (wn|xn,Φ,Θ) =

MnY

m=1

ZX

z=1

P (z|xn,Φ)P (wnm|z,Θ). (2)

Figure 1 shows a graphical model representation of PLSV,where shaded and unshaded nodes indicate observed andlatent variables, respectively. Since each word in a documentcan be sampled from different topics, PLSV can capturemultiple topics as in PLSA and LDA.

Page 3: Probabilistic Latent Semantic Visualization: Topic Model ... · Probabilistic Latent Semantic Visualization: Topic Model for Visualizing Documents ... which we call Probabilistic

2.2 Parameter estimationWe estimate the parameters in PLSV based on maximuma posteriori (MAP) estimation. The unknown parametersare a set of document coordinates X and a set of topic co-ordinates Φ as well as a set of word probabilities Θ. Werepresent all the unknown parameters by Ψ = {X ,Φ,Θ}.The number of topics Z is assumed to be known and fixed.

The log likelihood of parameters Ψ given a set of documentsC is as follows:

L(Ψ|C) =

NX

n=1

MnX

m=1

log

ZX

z=1

P (z|xn,Φ)P (wnm|z,Θ). (3)

Following the generative process described in the previoussubsection, we use a Dirichlet prior for word probability θz:

p(θz) =Γ

`

(α + 1)W´

Γ(α + 1)W

WY

w=1

θαzw, (4)

and a Gaussian prior with a zero mean and a spherical co-variance for the coordinates of topic φz and document xn:

p(φz) =“ β

”D2

exp“

−β

2‖ φz ‖2

, (5)

p(xn) =“ γ

D2

exp“

−γ

2‖ xn ‖2

, (6)

where α, β and γ are hyper-parameters. We used α = 0.01,β = 0.1N and γ = 0.1Z in all the experiments described inSection 4.

We estimate parameters Ψ by maximizing the posterior p(Ψ|C)using the EM algorithm [5]. The conditional expectation ofthe complete-data log likelihood with priors is representedas follows:

Q(Ψ|Ψ)

=

NX

n=1

MnX

m=1

ZX

z=1

P (z|n, m; Ψ) log P (z|xn,Φ)P (wnm|z,Θ)

+

NX

n=1

log p(xn) +

ZX

z=1

log p(φz) +

ZX

z=1

log p(θz), (7)

where Ψ represents the current estimate, and P (z|n, m; Ψ)represents the class posterior probability of the nth doc-ument and the mth word given the current estimate. InE-step, we compute the class posterior probability with theBayes rule:

P (z|n, m; Ψ) =P (z|xn, Φ)P (wnm|z, Θ)

PZz′=1 P (z′|xn, Φ)P (wnm|z′, Θ)

, (8)

where P (z|xn, Φ) is calculated by (1). In M-step, we obtain

the next estimate of word probability θzw by maximizingQ(Ψ|Ψ) w.r.t. θzw subject to

PWw=1 θzw = 1:

θzw =

PNn=1

PMnm=1 I(wnm =w)P (z|n, m; Ψ) + α

PWw′=1

PNn=1

PMnm=1 I(wnm =w′)P (z|n, m; Ψ) + αW

,

(9)where I(·) represents the indicator function, i.e. I(A) = 1 ifA is true and 0 otherwise. The next estimates of documentcoordinate xn and topic coordinate φz cannot be solved in

a closed form. Therefore, we estimate them by maximiz-ing Q(Ψ|Ψ) using a gradient-based numerical optimizationmethod such as the quasi-Newton method [15]. The gradi-

ents of Q(Ψ|Ψ) w.r.t. xn and φz are respectively:

∂Q

∂xn=

MnX

m=1

ZX

z=1

P (z|xn,Φ) − P (z|n, m; Ψ)”

(xn − φz)

− γxn, (10)

∂Q

∂φz=

NX

n=1

MnX

m=1

P (z|xn,Φ) − P (z|n, m; Ψ)”

(φz − xn)

− βφz. (11)

By iterating the E-step and the M-step until convergence,we obtain a local optimum solution for Ψ. We can embed anew document with low computational cost by fixing topiccoordinates Φ and word probabilities Θ to the estimatedvalues.

3. RELATED WORK

3.1 PLSAPLSV is based on Probabilistic Latent Semantic Analysis(PLSA) [10]. An essential difference between PLSV andPLSA is that a set of topic proportions are derived from co-ordinates of documents X and topics Φ in PLSV, whereasthey are directly estimated as Λ = {λn}N

n=1 in PLSA, whereλn = {λnz}Z

z=1, and λnz = P (z|dn,Λ) is the zth topic pro-portion of the nth document. Therefore although PLSVcan express any number of topics in D-dimensional space,PLSA can express only D+1 topics in D-dimensional space.The number of parameters for topic proportions in PLSVis (N + Z)D, and it is much smaller than that in PLSAN(Z − 1) since usually D ¿ Z ¿ N . Therefore, PLSV canprevent overfitting compared with PLSA.

Under PLSA, the probability of wn given dn, Λ and Θ is asfollows:

P (wn|dn,Λ,Θ) =

MnY

m=1

ZX

z=1

P (z|dn,Λ)P (wnm|z,Θ). (12)

The unknown parameters in PLSA Υ are a set of topic pro-portions Λ and a set of word probabilities Θ. They can beestimated by maximizing the following likelihood with theEM algorithm:

L(Υ|C) =N

X

n=1

MnX

m=1

logZ

X

z=1

P (z|dn,Λ)P (wnm|z,Θ). (13)

In E-step, the class posterior probability given the currentestimate can be computed as follows:

P (z|dn, wnm; Υ) =P (z|dn, Λ)P (wnm|z, Θ)

PZz′=1 P (z′|dn, Λ)P (wnm|z′, Θ)

. (14)

In M-step, the next estimate of topic proportion λnz is givenby:

λnz =

PMnm=1 P (z|dn, wnm; Υ)

PZz′=1

PMnm=1 P (z′|dn, wnm; Υ)

, (15)

Page 4: Probabilistic Latent Semantic Visualization: Topic Model ... · Probabilistic Latent Semantic Visualization: Topic Model for Visualizing Documents ... which we call Probabilistic

and the next estimate of word probability θzw is given by:

θzw =

PNn=1

PMnm=1 I(wnm = w)P (z|dn, wnw; Υ)

PWw′=1

PNn=1

PMnm=1 I(wnm = w′)P (z|dn, wnm; Υ)

.

(16)We can use Dirichlet priors for the topic proportions andword probabilities as in PLSV.

3.2 Parametric EmbeddingParametric Embedding (PE) [11] is a nonlinear visualizationmethod, which takes a set of discrete probability distribu-tions as its input. The topic proportions estimated usingPLSA Λ with any number of topics can be embedded ina D-dimensional Euclidean space by PE. PE embeds sam-ples in a low-dimensional Euclidean space so as to preservethe input probabilities by minimizing the following sum ofKullback-Leibler divergences:

E(X ,Φ) =

NX

n=1

ZX

z=1

P (z|dn, Λ) logP (z|dn, Λ)

P (z|xn,Φ), (17)

where P (z|xn,Φ) is the probability that the zth topic is cho-sen given a coordinate xn, and it is defined in the same equa-tion as PLSV as in (1). The unknown parameters, a set ofcoordinates of documents X and topics Φ, can be obtainedwith a gradient-based numerical optimization method. Thegradients of E(X ,Φ) w.r.t. xn and φz are respectively:

∂E

∂xn=

ZX

z=1

P (z|dn, Λ) − P (z|xn,Φ)”

(xn − φz), (18)

∂E

∂φz=

NX

n=1

P (z|dn, Λ) − P (z|xn,Φ)”

(φz − xn). (19)

We can use spherical Gaussian priors with zero means forthe coordinates as in PLSV.

The parameter estimation procedures in PLSV described inSection 2.2 show that the calculation of coordinates in PE isincluded in the M-step in PLSA. These coordinates are thenmapped to the topic proportions of documents and used inthe E-step.

3.3 Other related workThere are a few visualization methods based on genera-tive models. The examples include Generative TopographicMapping (GTM) [1] and the visualization method proposedin [13]. However, these methods are not topic models, inwhich each word is assumed to be drawn from one of the top-ics according to the topic proportions. A topic model withlatent coordinates based on GTM is proposed in [12]. How-ever, it is mainly used for predicting sequences, and not forvisualization. The correlated topic model is a topic modelthat can model correlations among topics [3]. PLSV canalso model the topic correlations because topics embeddedclose together in the visualization space are likely to occurtogether. PLSV is an unsupervised visualization method,where topics are unobservable variables, and it is differentfrom supervised visualization methods, such as Fisher lineardiscriminant analysis [7].

4. EXPERIMENTS4.1 Compared methodsFor evaluation, we compared PLSV with MDS, Isomap, PLSAand PLSA+PE by the visualization in the two-dimensionalspace D = 2.

MDS is a linear dimensionality reduction method, which em-beds samples so as to minimize the discrepancy betweenpairwise distances in the visualization space and those inthe original space. We used word count vectors as inputvn = (vn1, · · · , vnW ), where vnw =

PMnm=1 I(wnm = w) is

the count of the wth word in the nth document. We nor-malized the vector so that the L-2 norm becomes one.

Isomap is a nonlinear dimensionality reduction method, inwhich a graph is first constructed by connecting h-nearestneighbors, and then samples are embedded so as to pre-serve the shortest path distances in the graph by MDS. Weused the cosine similarity between word count vectors vn

for finding neighbors, which is widely used for the similar-ity measurement between documents. We set the number ofneighbors at h = 5.

In PLSA, we visualized documents using the procedures de-scribed in Section 3.1. In order to represent topic propor-tions in a two-dimensional space, we converted the topicproportions estimated by PLSA with Z = 3 to a coordinatein the two-dimensional simplex.

PLSA+PE embeds documents using PE according to thetopic proportions that are estimated by PLSA as describedin Section 3.2.

4.2 Data setsWe used the following three data sets in the evaluations:NIPS, 20News and EachMovie.

NIPS data consist of papers from the NIPS conference from2001 to 2003 1. There were 593 documents, and the vocab-ulary size was 14,036. Each document is labeled with 13research areas, such as Neuroscience and Applications.

20News data consist of documents in the 20 Newsgroupscorpus [14]. The corpus contains about 20,000 articles cat-egorized into 20 discussion groups. We omitted stop-wordsand words that occurred fewer than 50 times, and also omit-ted documents with fewer than 50 words. The vocabularysize was 6,754. We sampled 50 documents from each of 20classes, for a total of 1000 documents.

EachMovie data consist of movie ratings, which are stan-dard benchmark data for collaborative filtering. We re-garded movies and users as documents and words, respec-tively, where a movie is represented by a sequence of ratedusers. Each movie is labeled with 10 genres, for instance Ac-tion, Comedy and Romance. We omitted users and movieswith fewer than 50 ratings and movies labeled with morethan one genre. The number of movies was 764, and thenumber of users was 7,180.

4.3 Evaluation measurement1Available at http://ai.stanford.edu/˜gal/

Page 5: Probabilistic Latent Semantic Visualization: Topic Model ... · Probabilistic Latent Semantic Visualization: Topic Model for Visualizing Documents ... which we call Probabilistic

We evaluated the visualization results quantitatively fromthe label prediction accuracy with the k-nearest neighbor(k-NN) method in the visualization space. Each samplein all three data sets is labeled with research area, discus-sion group or genre. Note that we did not use the labelinformation for visualization in any of the methods, namely,we performed visualization with fully unsupervised settings.The accuracy generally becomes high when samples withthe same labels are located close together and samples withdifferent labels are located far away from each other in thevisualization space.

The k-NN method predicts a sample label from the mostdominant label among the k nearest samples, where we usedthe Euclidean distance for finding neighbors. The accuracyis computed by acc(k) = 1

N

PNn=1 I

`

yn = yk(xn)´

× 100,where yn is the label of the nth document, and yk(xn) isthe predicted label with the k-NN method for a sample withcoordinate xn.

4.4 ResultsThe accuracies on NIPS, 20News and EachMovie data setsare shown in Figure 2 when documents are embedded in atwo-dimensional space by PLSV, MDS, Isomap, PLSA andPLSA+PE. We set the number of topics at Z = 50 for PLSVand PLSA+PE. The number of topics for PLSA is automat-ically determined as Z = 3 since D = 2. In 20News, wecreated 30 evaluation sets by random sampling, where eachset consists 1000 documents, and evaluated by the averageaccuracy over the 30 sets. In NIPS and EachMovie, the accu-racies of PLSV, PLSA+PE and PLSA are averaged over 30visualizations for one data set with different initial param-eters. Only the standard deviations for PLSV are shown.In all three data sets, the highest accuracies are achievedby PLSV. This result implies that PLSV can appropriatelyembed documents in the two-dimensional Euclidean spacewhile keeping the essential characteristics of the documents.The accuracies achieved by PLSA+PE are lower than thoseachieved by PLSV since it embeds documents through twomodularized processes, where the objective functions are dif-ferent from each other. The accuracies achieved by PLSAare low since it has only three topics, and it may be inade-quate to measure similarities among topic proportions basedon the Euclidean distance.

We also evaluated PLSV and PLSA+PE with different num-bers of topics Z = 5, 10, · · · , 50. The accuracies with theone-nearest neighbor method are shown in Figure 3. With asmall number of topics, the accuracies achieved by PLSV arenot very high, and they are comparable to those achievedby PLSA+PE. However, as the number of topics increases,PLSV outperforms PLSA+PE. This result indicates that thetopic proportions and the word probabilities overfit the highdimensional PLSA parameter space in PLSA+PE, and theymay not be appropriately represented in the two-dimensionalvisualization space.

Figures 4, 5 and 6 show visualization results obtained byPLSV (Z = 50), MDS, Isomap, PLSA (Z = 3) and PLSA+PE(Z = 50) on NIPS, 20News and EachMovie data sets, respec-tively. Here each point represents a document, and the shapeand color represents the label. In the PLSV visualizations,documents with the same label are likely to be clustered

together. On the other hand, with MDS and Isomap, docu-ments with different labels are mixed, and thus the accuracyof their visualization is low. In PLSA, many documents arelocated at the corner, and the latent topic structure of thegiven data is not fully expressed in this dimensionality. InPLSA+PE, documents are slightly more mixed than thosein PLSV as shown quantitatively by the accuracy.

Figure 7 shows PLSV visualization results for NIPS datawith different numbers of topics Z = 10, 20, 30 and 40. Al-though documents with different labels are mixed when thenumber of topics is small, the visualization with Z = 40shows similar quality to that with Z = 50.

We analyzed the PLSV visualization in detail. Figure 8shows the visualization result for 20News data obtained byPLSV with Z = 50. Here each black circle represents atopic coordinate φz, and each black × represents a meancoordinate of documents for each label, which is calculatedby µy = 1

Ny

PNn=1 I(yn = y)xn, where Ny is the number of

documents labeled with y. The number near the circle corre-sponds to the topic index in the table at the bottom, wherethe ten most probable words for each topic are shown. Docu-ments with the same label are clustered together, and closelyrelated labels are located nearby, such as rec.sport.baseballand rec.sport.hockey, rec.autos and rec.motorcycles, comp.-sys.mac.hardware and comp.sys.ibm.pc.hardware, and soc.-religion.christian and talk.religion.misc. In a topic locatednear a label mean, representative words for the label occurwith high probability. For example, probable words in top-ics near rec.sport (z = 1 and z = 2) are ’team’, ’players’and ’game’, those near com.graphics (z = 5) are ’graphics’,’points’ and ’lines’, and those near sci.crypt (z = 8) are’clipper’, ’encryption’ and ’public’.

Figure 9 shows the visualization result for certain movie ti-tles from EachMovie data obtained by PLSV with Z = 50.Movies in the same genre are likely to be located close to-gether. For example, classic movies are located in the bot-tom right, and foreign movies are located at the top. Classicmovies are tightly clustered because there may be a numberof people who see only classic movies.

The computational time of PLSV on a PC with 3.2GHzXeon CPU and 2GB memory were 117, 20, and 256 minutesfor NIPS, 20News, and EachMovie data sets, respectively.

5. CONCLUSIONSIn this paper, we proposed a visualization method based ona topic model, Probabilistic Latent Semantic Visualization(PLSV), for discrete data such as documents. We have con-firmed experimentally that PLSV can visualize documentswith the latent topic structure. The results encourage us tobelieve that our data visualization approach based on PLSVis promising and will become a useful tool for visualizingdocuments.

Since PLSV is a probabilistic topic model, we can extendPLSV easily based on other research on topic models. Topicmodels have been proposed that model not only documentsbut also other information, for example images [2], time [21],authors [16] and citations [6]. We can also visualize thisinformation with documents using the framework proposed

Page 6: Probabilistic Latent Semantic Visualization: Topic Model ... · Probabilistic Latent Semantic Visualization: Topic Model for Visualizing Documents ... which we call Probabilistic

0

10

20

30

40

50

60

70

5 10 15 20 25 30 35 40 45 50

accu

racy

#neighbors

PLSV

PLSA+PE

IsomapMDS

PLSA

0

10

20

30

40

50

60

5 10 15 20 25 30 35 40 45 50#neighbors

PLSVPLSA+PE

Isomap

MDS

PLSA

0

10

20

30

40

50

60

5 10 15 20 25 30 35 40 45 50#neighbors

PLSV

PLSA+PE

Isomap

MDS

PLSA

(a) NIPS (b) 20News (c) EachMovie

Figure 2: Accuracy with the k-nearest neighbor method in the visualization space with different numbers ofneighbors k.

0

10

20

30

40

50

60

5 10 15 20 25 30 35 40 45 50

accu

racy

#topics

PLSV

PLSA+PE

Isomap

MDS

PLSA

0

5

10

15

20

25

30

35

40

45

50

5 10 15 20 25 30 35 40 45 50#topics

PLSV

PLSA+PE

Isomap

MDS

PLSA

0

5

10

15

20

25

30

35

40

45

50

5 10 15 20 25 30 35 40 45 50#topics

PLSV

PLSA+PEIsomap

MDS

PLSA

(a) NIPS (b) 20News (c) EachMovie

Figure 3: Accuracy with the one-nearest neighbor method in the visualization space with different numbersof topics Z for PLSV and PLSA+PE.

AA: Algorithms & Architectures

AP: Applications

CN: Control & Reinforcement Learning

CS: Cognitive Science & AI

IM: Implementations

LT: Learning Theory

NS: Neuroscience

SP: Speech & Signal Processing

VS: Vison

BI: Brain Imaging

ET: Emerging Technologies

VB: Vision (Biological)

VM: Vision (Machine)

(a) PLSV (b) MDS

(c) Isomap (d) PLSA (e) PLSA+PE

Figure 4: Visualization of documents in NIPS.

Page 7: Probabilistic Latent Semantic Visualization: Topic Model ... · Probabilistic Latent Semantic Visualization: Topic Model for Visualizing Documents ... which we call Probabilistic

alt.atheismcomp.graphicscomp.os.ms−windows.misccomp.sys.ibm.pc.hardwarecomp.sys.mac.hardwarecomp.windows.xmisc.forsalerec.autosrec.motorcyclesrec.sport.baseballrec.sport.hockeysci.cryptsci.electronicssci.medsci.spacesoc.religion.christiantalk.politics.gunstalk.politics.mideasttalk.politics.misctalk.religion.misc

(a) PLSV (b) MDS

(c) Isomap (d) PLSA (e) PLSA+PE

Figure 5: Visualization of documents in 20News.

Action

Animation

Art Foreign

Classic

Comedy

Drama

Family

Horror

Romance

Thriller

(a) PLSV (b) MDS

(c) Isomap (d) PLSA (e) PLSA+PE

Figure 6: Visualization of movies in EachMovie.

Page 8: Probabilistic Latent Semantic Visualization: Topic Model ... · Probabilistic Latent Semantic Visualization: Topic Model for Visualizing Documents ... which we call Probabilistic

Z = 10 Z = 20 Z = 30 Z = 40

Figure 7: Visualization by PLSV with different numbers of topics Z on NIPS.

talk.policics.mideast

talk.policics.guns

talk.policics.misccomp.graphics

comp.os.ms−windows.misc

comp.sys.ibm.pc.hardware

comp.sys.mac.hardware

comp.windows.x

rec.autos

rec.motorcycles

rec.sport.baseball

rec.sport.hockey

sci.crypt

sci.electronics

sci.med

sci.space

alt.atheism

soc.religion.christian

talk.religion.misc

misc.forsale

1

2

3

4

5

6

7

8

9

10

alt.atheismcomp.graphicscomp.os.ms−windows.misccomp.sys.ibm.pc.hardwarecomp.sys.mac.hardwarecomp.windows.xmisc.forsalerec.autosrec.motorcyclesrec.sport.baseballrec.sport.hockeysci.cryptsci.electronicssci.medsci.spacesoc.religion.christiantalk.politics.gunstalk.politics.mideasttalk.politics.misctalk.religion.misc

1 2 3 4 5 6 7 8 9 10team game car high graphics windows scsi clipper israel god

players good bike engine points pc printer encryption israeli jesusseason games big battery lines keyboard bus public militia raleague play cars car point drive windows system arab bible

nhl baseball water stuff line mouse drivers government jewish christteams guys drive low reference mac hp keys jews heaven

average win buy bought image disk speed chip turkish peoplehall fans thing dealer access card local security armenian john

make division front kind program memory network escrow armenians scripturedave guy miles speed comp dos fonts nsa palestine spirit

Figure 8: Visualization of documents in 20News by PLSV with Z = 50. Each point represents a document,and the shape and color represent the discussion group. Each black circle indicates a topic coordinate φz,and each black × indicates a label mean µy. The table at the bottom shows the ten most probable wordsfor ten topics estimated with the visualization, i.e. the words are ordered according to P (w|z,Θ). The topicindex corresponds to the number near the circle in the visualization.

Page 9: Probabilistic Latent Semantic Visualization: Topic Model ... · Probabilistic Latent Semantic Visualization: Topic Model for Visualizing Documents ... which we call Probabilistic

Grumpier Old Men

Waiting to Exhale

Father of the Bride Part IITom and Huck

Sudden Death

Goldeneye

Dracula: Dead and Loving It

Nixon

Cutthroat Island

Casino

Ace Ventura: When Nature Calls

Money Train

Assassins

Leaving Las Vegas

The City of Lost Children

Shanghai Triad

Dangerous Minds

It Takes Two

Dead Presidents

Restoration

Mortal Kombat

Il Postino

The Indian in the Cupboard

Eye for an Eye

Two Bits

Friday

Fair Game

Les Misôþables

Bed of Roses

Big Bully

The Crossing Guard

The Juror

The White Balloon

Vampire in Brooklyn

Heidi Fleiss: Hollywood MadamMr. Wrong

Steal Big, Steal Little

The Boys of St. Vincent

Chungking ExpressFlirting With Disaster Pie in the Sky

Jade

Belle de Jour

Beyond Rangoon

Canadian Bacon

Clockers

Crimson Tide

Crumb

Die Hard: With a VengeanceHackers

Jeffrey

Judge Dredd

Kids

Living in Oblivion

Party Girl

The Prophecy

The Scarlet Letter

The Show

Smoke

Species

Under Siege 2: Dark Territory

Bushwhacked

Burnt by the Sun

Before the Rain

Before Sunrise

Billy MadisonClerks

Drop Zone

Death and the Maiden

Exit to Eden

Hideaway

Gordy

Hoop Dreams

Houseguest

The Jerky Boys

Just Cause

A Little Princess

Like Water For Chocolate

Man of the House

Mixed Nuts

My Family

Nobody’s Fool

Nemesis 2: Nebula

Pushing HandsPicture Bride

Three Colors: Red

The Specialist

Tommy Boy

Vanya on 42nd Street

Virtuosity

Cobb

It Could Happen to You

The Jungle BookThe Paper

Timecop

S.F.W.

Street Fighter8 Seconds

Another Stakeout

Bad Girls

Blink

A Bronx Tale

Carlito’s Way

Demolition Man

Farewell My Concubine

Fearless

Getting Even With Dad

Live Nude Girls

Last Action Hero

M. Butterfly

Made in America

Malice

Menace II Society

Orlando

The Ref

Renaissance Man

Robocop 3

Ruby in Paradise

The Scout

Blade Runner

Terminal Velocity

Princess Caraboo

The Wooden Man’s Bride

One Fine Day

Heavy Metal

A Family Thing

Courage Under Fire

Mission: Impossible

Barbarella

Alphaville

The Promise

Faces

Switchblade Sisters

The Visitors

MultiplicityStriptease

The Fan

Kingpin

A Time to Kill

Fled

Larger than Life

The First Wives Club

High School High

Foxfire

The Big Squeeze

The Spitfire Grill

Supercop

Gone Fishin’Killer: A Journal of Murder

Project S

The Trigger Effect

For Whom the Bell Tolls

Singing in the Rain

Funny Face

Breakfast at Tiffany’s

My Fair Lady

My Favorite Year

Gigi

Grace of My Heart

Big Night

Curdled

Escape to Witch MountainThe Parent Trap

That Darn Cat

Cool RunningsMary Poppins

Die HardA Fish Called WandaMonty Python’s Life of Brian

Victor VictoriaThe Candidate

The Old Man and the Sea

Platoon

Weekend at Bernie’s

The Doors

Glengarry Glen Ross

The Eighth Day

Private Benjamin

Ran

The Right Stuff

Local Hero

The Crucible

Michael

Fierce Creatures

Underworld

Private PartsThe Saint

Grosse Pointe BlankAustin Powers

Dirty Dancing

ActionAnimation

Art ForeignClassic

ComedyDrama

FamilyHorror

RomanceThriller

Figure 9: Visualization of movies in EachMovie by PLSV with Z = 50. Each point represents a movie, andthe shape and color represent the genre. Some examples of movie titles are also shown.

in this paper. We assumed that the number of topics wasknown. We can automatically infer the number of topics byextending PLSV to a nonparametric Bayesian model suchas the Dirichlet process mixture model [18]. In addition,we can achieve more robust estimation using the Bayesianapproach, instead of MAP estimation, as in LDA [9].

6. REFERENCES[1] C. M. Bishop, M. Svensen, and C. K. I. Williams. GTM:

The generative topographic mapping. Neural Computation,10(1):215–234, 1998.

[2] D. M. Blei and M. I. Jordan. Modeling annotated data. InSIGIR ’03: Proceedings of the 26th Annual InternationalACM SIGIR conference on Research and development ininformaion retrieval, pages 127–134, 2003.

[3] D. M. Blei and J. D. Lafferty. A correlated topic model ofscience. The Annals of Applied Statistics, 1(1):17–35, 2007.

[4] D. M. Blei, A. Y. Ng, and M. I. Jordan. Latent Dirichletallocation. Journal of Machine Learning Research,3:993–1022, 2003.

[5] A. Dempster, N. Laird, and D. Rubin. Maximum likelihoodfrom incomplete data via the EM algorithm. Journal of theRoyal Statistical Society, Series B, 39(1):1–38, 1977.

[6] L. Dietz, S. Bickel, and T. Scheffer. Unsupervisedprediction of citation influences. In ICML ’07: Proceedingsof the 24th International Conference on Machine Learning,pages 233–240, 2007.

[7] R. A. Fisher. The use of multiple measurements intaxonomic problems. Annals of Eugenics, 7:179–188, 1936.

[8] B. Fortuna, M. Grobelnik, and D. Mladenic. Visualizationof text document corpus. Informatica, 29(4):497–502, 2005.

[9] T. L. Griffiths and M. Steyvers. Finding scientific topics.Proceedings of the National Academy of Sciences, 101Suppl 1:5228–5235, 2004.

[10] T. Hofmann. Probabilistic latent semantic analysis. In UAI’99: Proceedings of 15th Conference on Uncertainty inArtificial Intelligence, pages 289–296, 1999.

[11] T. Iwata, K. Saito, N. Ueda, S. Stromsten, T. L. Griffiths,and J. B. Tenenbaum. Parametric embedding for classvisualization. Neural Computation, 19(9):2536–2556, 2007.

[12] A. Kaban. Predictive modelling of heterogeneous sequencecollections by topographic ordering of histories. MachineLearning, 68(1):63–95, 2007.

[13] A. Kaban, J. Sun, S. Raychaudhury, and L. Nolan. On classvisualisation for high dimensional data: Exploring scientificdata sets. In DS ’06: Proceedings of the 9th InternationalConference on Discovery Science, 2006.

Page 10: Probabilistic Latent Semantic Visualization: Topic Model ... · Probabilistic Latent Semantic Visualization: Topic Model for Visualizing Documents ... which we call Probabilistic

[14] K. Lang. NewsWeeder: learning to filter netnews. In ICML’95: Proceedings of the 12th International Conference onMachine Learning, pages 331–339, 1995.

[15] D. C. Liu and J. Nocedal. On the limited memory BFGSmethod for large scale optimization. Math. Programming,45(3):503–528, 1989.

[16] M. Rosen-Zvi, T. Griffiths, M. Steyvers, and P. Smyth. Theauthor-topic model for authors and documents. In UAI ’04:Proceedings of the 20th Conference on Uncertainty inArtificial Intelligence, pages 487–494, 2004.

[17] S. T. Roweis and L. K. Saul. Nonlinear dimensionalityreduction by locally linear embedding. Science,290(5500):2323–2326, 2000.

[18] Y. W. Teh, M. I. Jordan, M. J. Beal, and D. M. Blei.Hierarchical Dirichlet processes. Journal of the AmericanStatistical Association, 101(476):1566–1581, 2006.

[19] J. B. Tenenbaum, V. de Silva, and J. C. Langford. A globalgeometric framework for nonlinear dimensionalityreduction. Science, 290(5500):2319–2323, 2000.

[20] W. Torgerson. Theory and methods of scaling. Wiley, NewYork, 1958.

[21] X. Wang and A. McCallum. Topics over time: anon-Markov continuous-time model of topical trends. InKDD ’06: Proceedings of the 12th ACM SIGKDDInternational Conference on Knowledge Discovery andData mining, pages 424–433, 2006.

[22] J. A. Wise, J. J. Thomas, K. Pennock, D. Lantrip,M. Pottier, A. Schur, and V. Crow. Visualizing thenon-visual: spatial analysis and interaction withinformation from text documents. In INFOVIS ’95:Proceedings of the 1995 IEEE Symposium on InformationVisualization, pages 51–58, 1995.