graph-based methods for the automatic annotation and ...carneiro/publications/printart_icmr... ·...

8
Graph-based Methods for the Automatic Annotation and Retrieval of Art Prints Gustavo Carneiro Instituto de Sistemas e Robótica Instituto Superior Técnico Lisbon, Portugal [email protected] ABSTRACT The analysis of images taken from cultural heritage artifacts is an emerging area of research in the field of information retrieval. Cur- rent methodologies are focused on the analysis of digital images of paintings for the tasks of forgery detection and style recognition. In this paper, we introduce a graph-based method for the automatic annotation and retrieval of digital images of art prints. Such method can help art historians analyze printed art works using an annotated database of digital images of art prints. The main challenge lies in the fact that art prints generally have limited visual informa- tion. The results show that our approach produces better results in a weakly annotated database of art prints in terms of annotation and retrieval performance compared to state-of-the-art approaches based on bag of visual words. H.3.3Information storage and retrievalInformation search and re- trieval|Retrieval models I.4.8Image Processing and Computer Vi- sionScene Analysis|Object Recognition G.3Probability and Statis- ticsProbabilistic Algorithms General Terms Algorithms, Measurement, Experimentation Keywords Content-based image retrieval and annotation, Art image annota- tion and retrieval, Graph-based Learning methods 1. INTRODUCTION The analysis of digital images taken from artistic productions is a emerging field of research in the areas of information retrieval, computer vision, digital image analysis and machine learning. There are several applications being developed in this area, such as: the system designed by Google to identify paintings [1]; the artistic identification methodologies designed to classify Van Gogh’s brush This work was supported by the FCT (ISR/IST plurianual fund- ing) through the PIDDAC Program funds and Project PRINTART (PTDC/EEA-CRO/098822/2008). This project is also funded by the European Commission; Marie Curie IIF, contract number PIIF- GA-2009-236173 (IMASEG3D). Figure 1: Examples of how art print images (left) are altered in the process of becoming a tile panel (right). strokes [27]; the model to classify brushstrokes [48]; and the ap- proach to discover, recover, represent and understand cultural her- itage artifacts [21]. Therefore, the techniques developed in this area will be an essential tool for systems designed to help art historians in the task of analyzing art production. The analysis of digital images of prints is particularly important for art historians because printmaking methods have been inten- sively used over the last five centuries with the goal of replicat- ing paintings produced by artists from all over the world. This intense use of printmaking techniques is associated with the the fast and cheap production of paper and advancements in graphi- cal arts, which started in the XV th century. As a result, prints of artistic paintings have reached a vastly superior number of people compared to the original paintings. Consequently it is worth under- standing the importance of art prints in art history since they have influenced and served as a source of inspiration for generations of artists. Therefore, the proper classification, retrieval and annota- tion of art prints constitute an important activity for art historians to understand the art produced in the last five centuries. The influence of prints on visual arts can be evidenced in several artistic productions, such as the influence of Japanese art prints on impressionist artists in the XIX th century [3]. More relevant to our work is the influence of prints on artistic tile painters in Portugal [9, 12], as evidenced by the large number of panels found in Churches, public buildings and palaces. In the context of artistic tile panels,

Upload: phamcong

Post on 13-Mar-2018

220 views

Category:

Documents


4 download

TRANSCRIPT

Page 1: Graph-based Methods for the Automatic Annotation and ...carneiro/publications/printart_icmr... · Graph-based Methods for the Automatic ... In this paper, we introduce a graph-based

Graph-based Methods for the Automatic Annotation andRetrieval of Art Prints

Gustavo Carneiro∗

Instituto de Sistemas e RobóticaInstituto Superior Técnico

Lisbon, [email protected]

ABSTRACTThe analysis of images taken from cultural heritage artifacts is anemerging area of research in the field of information retrieval. Cur-rent methodologies are focused on the analysis of digital images ofpaintings for the tasks of forgery detection and style recognition.In this paper, we introduce a graph-based method for the automaticannotation and retrieval of digital images of art prints. Such methodcan help art historians analyze printed art works using an annotateddatabase of digital images of art prints. The main challengeliesin the fact that art prints generally have limited visual informa-tion. The results show that our approach produces better resultsin a weakly annotated database of art prints in terms of annotationand retrieval performance compared to state-of-the-art approachesbased on bag of visual words.H.3.3Information storage and retrievalInformation search and re-trieval|Retrieval models I.4.8Image Processing and Computer Vi-sionScene Analysis|Object Recognition G.3Probability and Statis-ticsProbabilistic Algorithms

General TermsAlgorithms, Measurement, Experimentation

KeywordsContent-based image retrieval and annotation, Art image annota-tion and retrieval, Graph-based Learning methods

1. INTRODUCTIONThe analysis of digital images taken from artistic productions isa emerging field of research in the areas of information retrieval,computer vision, digital image analysis and machine learning. Thereare several applications being developed in this area, suchas: thesystem designed by Google to identify paintings [1]; the artisticidentification methodologies designed to classify Van Gogh’s brush

∗This work was supported by the FCT (ISR/IST plurianual fund-ing) through the PIDDAC Program funds and Project PRINTART(PTDC/EEA-CRO/098822/2008). This project is also funded bythe European Commission; Marie Curie IIF, contract number PIIF-GA-2009-236173 (IMASEG3D).

Figure 1: Examples of how art print images (left) are alteredinthe process of becoming a tile panel (right).

strokes [27]; the model to classify brushstrokes [48]; and the ap-proach to discover, recover, represent and understand cultural her-itage artifacts [21]. Therefore, the techniques developedin this areawill be an essential tool for systems designed to help art historiansin the task of analyzing art production.

The analysis of digital images of prints is particularly importantfor art historians because printmaking methods have been inten-sively used over the last five centuries with the goal of replicat-ing paintings produced by artists from all over the world. Thisintense use of printmaking techniques is associated with the thefast and cheap production of paper and advancements in graphi-cal arts, which started in the XVth century. As a result, prints ofartistic paintings have reached a vastly superior number ofpeoplecompared to the original paintings. Consequently it is worth under-standing the importance of art prints in art history since they haveinfluenced and served as a source of inspiration for generations ofartists. Therefore, the proper classification, retrieval and annota-tion of art prints constitute an important activity for art historiansto understand the art produced in the last five centuries.

The influence of prints on visual arts can be evidenced in severalartistic productions, such as the influence of Japanese art prints onimpressionist artists in the XIXth century [3]. More relevant to ourwork is the influence of prints on artistic tile painters in Portugal [9,12], as evidenced by the large number of panels found in Churches,public buildings and palaces. In the context of artistic tile panels,

Page 2: Graph-based Methods for the Automatic Annotation and ...carneiro/publications/printart_icmr... · Graph-based Methods for the Automatic ... In this paper, we introduce a graph-based

Figure 2: Loss of visual information from the photographic im-age in (a), to the painting in (b), to the art print in (c).

Figure 3: Example of the manual annotation of an art printimage produced by an art historian.

the analysis of digital images of prints is extremely important forunderstanding how the Portuguese tile panel artists were influencedby such prints, which has the potential to furnish relevant informa-tion to art historians in Portugal. It is important to mention that artprints were generally used as references to produce the panels, butthe artists often produced an impression of the print (i.e.,not anexact copy of it - see Fig. 1). For this reason, the task of discov-ering the influence of one or several prints in the composition ofa tile panel requires an expert with a specialized visual knowledgeof the current databases of prints together with peculiar abilitiesin relating the tile composition with the prints. Therefore, an au-tomatic system that can retrieve a set of print images related to agiven artistic work can can be useful for art historians.

Compared to photographic digital images and paintings, artprintimages loses important visual information [11]. For example, Fig. 2shows three images of the themethe Crucifixion of Jesus Christ,which displays a large sensorial gap [40] between the photographic,painting and art print images. Notice how the low level information(texture, color, depth, etc.) is lost moving from the photographicimage to the art print image. This loss of visual informationre-duces the effectiveness of current image analysis techniques, whichusually work with photographic digital images [14]. Art imageanalysis methodologies can also be used for art prints [42, 27], butthe great majority of these techniques have been developed for theanalysis of digitized images of paintings, which contain richer vi-sual information than prints. However, note that Li and Wang[29]have developed a system that analyzes ancient Chinese paintings,which are similar to art prints.

In this work, we present a new method for the analysis of art prints.

The first goal of the work is to automatically produce global an-notations to previously unseen test images using a statistical modelwhose parameters are estimated using a database of manuallyanno-tated training images. The second goal is to retrieve un-annotatedimages given specific visual classes using the statistical model de-scribed above. The art print images in this database are constrainedin the following ways: 1) they were created between the centuriesXV and XVII, and 2) they are of religious themes. These con-straints are relevant for discovering the influences suffered by artis-tic tile panel painters, which is the long-term goal of this project.The manual annotation (Fig. 3) has been produced by art historians,who identified the image theme and the relevant subjects present inthe print (weak annotation without relation between subjects andimage regions). The method proposed in this paper follows graph-based learning algorithms, which have the assumption that the vi-sually similar images are likely to share the same annotations [50].Note that this assumption is important to uncover the aforemen-tioned influence of art prints over other artistic productions (andalso over other art prints). Specifically, we explore the followinggraph-based algorithms [6, 35, 50]: label propagation [18,28, 34,46], random walk [11], stationary solution using a stochastic ma-trix [30], and combinatorial harmonic [19]. We adapt each ofthosetechniques to a bag of visual words (BOV) approach, and we com-pare their performance with BOV approaches that use the follow-ing classifiers: support vector machines (SVM) [45], and randomforests (RF) [7]. Note that BOV approaches with these two clas-sifiers can be regarded as the state-of-the-art methods for imageretrieval and annotation. The experimental setup uses a databaseof art prints (which will soon be available for the information re-trieval and computer vision communities) containing 307 imagesand 22 labels, where one label represents a multi-class problemwith 7 classes, and the other labels represent binary problems. Theresults show that graph-based methods produce better results thanBOV models (using SVM and RF classifiers) in terms of retrievaland annotation performance.

This paper is organized as follows. Section 2 introduces relevantworks in photographic and art image retrieval, and graph-basedlearning. Section 3 describes the proposed methodology, whileSec. 4 outlines the implementation of our approach. The experi-ments are presented in Sec. 5, and the paper is concluded in Sec. 6.

2. LITERATURE REVIEWIn this section, we provide a brief review of papers in the areasof photographic and art image retrieval and annotation. We alsoreview a few relevant graph-based learning methods.

Currently in photographic image retrieval and annotation,the mostsuccessful methods [15] are based on the the bag of visual wordsrepresentation [13] combined with SVM [45] or multiple kernellearning (MKL) classifiers [41]. This methodology is effectivewhen the number of visual classes is relatively small (faster trainingand inference procedures) and the number of training imagesperclass is relatively large (better generalization). Another constraintof this methodology is that the introduction of new images and newlabels require a full (re-)training of relevant classifiers. Unfortu-nately, both constraints are too restrictive for our case where thenumber of visual classes can be large (with a limited number oftraining images per class), and the introduction of new images andlabels to the database can happen frequently. In photographic im-age analysis, there is a trend to get around the problem of thehighnumber of visual classes with the use of machine learning meth-ods that finds a sub-space of smaller dimensionality for classifica-tion [23]. However, the dynamic nature of our problem, wherenew

Page 3: Graph-based Methods for the Automatic Annotation and ...carneiro/publications/printart_icmr... · Graph-based Methods for the Automatic ... In this paper, we introduce a graph-based

classes are regularly introduced into the training database, is stillan issue in this area of research.

The area of art image retrieval has attracted the attention of re-searchers in the fields of computer vision and machine learning [27,33, 42]. The main focus of the papers is on the artistic identifica-tion problem, where the goal is to classify original and fakepaint-ings of a given artist [4, 32, 39] or to produce stylistic analysisof paintings [20, 24, 25]. Most of the methods above can be re-garded as adaptations from the content-based image retrieval sys-tems [14], where the emphasis is placed on the characterization ofbrush strokes using texture or color. The ancient Chinese paint-ing classification studied by Li and Wang [29] is more similartoours in the sense that they deal with the multi-class classificationof painting styles. Finally, the work on the automatic brushworkannotation by Yelizaveta et al. [48] is also similar to ours giventhat the authors are dealing with multi-class classification of brushstrokes. Different from the papers above, our method addresses notonly a multi-class, but also a multi-label problem [16].

Graph-based learning has been thoroughly studied by the informa-tion retrieval community to rank web pages on the World WideWeb [6, 8, 35]. Essentially, a graph is built where vertexes repre-sent web pages and the edge weights are denoted by the existenceof hyper-links. Analysis algorithms based on random walks in thisgraph have been designed to rank the vertexes (i.e., web pages) interms of their importance in this network. These graph-based tech-niques have received considerable attention by the machinelearn-ing community for the problem of semi-supervised learning [50].The random walk algorithm has also been studied in the domainsof unsupervised image segmentation [19] and multi-class classifi-cation [43]. Finally, random walk algorithms have also beenex-plored in the area of image retrieval [11, 18, 22, 26, 28, 30, 34, 46],where the main advantages of such approaches are: 1) the abilityto use visual and non-visual cues in the random walk procedure,2) the potential to extend the method to large-scale databases, and3) the relatively facility to adapt the method to dynamic problems,where new images and labels are continuously introduced into thedatabase.

3. METHODOLOGYAssume that a training set of annotated images is available,and isrepresented as follows:D = {(Ii,xi,yi)}i=1..N with xi repre-senting the feature vector of imageIi andyi denoting the anno-tation of that image. An annotated test set is also availableand isrepresented byT = {(eIi, exi, eyi)}i=1..P , but note that the anno-tation in the test set is used only for the purpose of methodologyevaluation. Each annotationy representsL multi-class and binaryproblems, soy = [y1, ..., yL] ∈ {0, 1}M , where each problemis denoted byyl ∈ {0, 1}|yl| (with |yl| denoting the dimension-ality of yl, where binary problems have|yl| = 1, and multi-classproblems have|yl| > 1 and‖yl‖1 = 1), andM represents thedimensionality of the annotation vector (i.e.,

PL

l=1 |yl| = M ). Insummary, binary problems involve an annotation that indicates thepresence or absence of a class, while multi-class annotation regardsproblems that one (and only one) of the possible classes is present.

Following the notation introduced by Estrada et al. [17], who ap-plied random walk algorithms for the problem of image de-noising,let us define a random walk sequence ofk steps asTr,k = [(x(r,1),y(r,1)), ..., (x(r,k),y(r,k))], where eachx(r,l) be-longs to the training setD, andr indexes a specific random walk.Our goal is to estimate the probability of annotationy for a test

imageex, as follows [43]:

p(y|ex) =1

ZT

RX

r=1

KX

k=1

p(Tr,k|ex)1

k p(y|x(r,k)). (1)

In (1), ZT is a normalization factor,p(y|x(r,k)) = δ(y − y(r,k))(with δ(.) being the Dirac delta function, which means that thisterm is one wheny = y(r,k)), the exponent1

kmeans that steps

taken at later stages of the random walk have higher weight,

p(Tr,k|ex) =p([(x(r,1), y(r,1)), ..., (x(r,k),y(r,k))]|ex)

=kY

j=2

p(x(r,j)|x(r,j−1), ex)p(y(r,j)|y(r,j−1))

p(x(r,1)|ex)p(y(r,1))

(2)

with the derivation made assuming a Markov process and that thetraining labels and features are independent given the testimage,p(y(r,j)|y(r,j−1)) = sy(y(r,j),y(r,j−1)) with sy(y(r,j),y(r,j−1)) =1

Zy

PM

m=1 λm × y(r,j)(m) × y(r,j−1)(m) (λm is the weight as-

sociated with the labely(m) ∈ {0, 1} andZy is a normalizationfactor),p(y(r,1)) = 1

N, andp(x(r,j)|x(r,j−1), ex) andp(x(r,1)|ex)

are defined in Sec. 4.2.

We propose the use of class mass normalization [51] to determinethe annotation of imageex. The class mass normalization takes intoconsideration the probability of a class annotation and thepropor-tion of samples annotated with that class in the training set. Specif-ically, we have

by =

"NX

i=1

yi(m) × max(p(yi(m)|ex) − p(y(m)),0)

#

m=1..M

,

(3)wherep(y(m)) = 1

N

PN

i=1 yi(m) (m indicates themth dimen-sion of label vectory). The use of class mass normalization makesthe annotation process more robust to imbalances in the trainingset with respect to the number of training images per visual class.Notice thatby in (3) represents the confidence that the image repre-sented byex is annotated with the labelsby(m) for m = {1, .., M}.Finally, we further processby for multi-class problems as follows:

∀l ∈ {1, ..., L}, with |byl| > 1

y∗l =

min(⌊byl/max(byl)⌋, 1), if max(byl) > 0.5

{0}|yl |, otherwise, (4)

and for binary problems we define:

∀l ∈ {1, ..., L}, with |byl| = 1

y∗l =

1, byl > 0.5

0, otherwise

. (5)

As a result, the final annotation for imageex is represented byy∗ =[y∗

1 , ..., y∗L].

The retrieval problem is defined as the most relevant image returnedfrom the database of test imagesT given themth visual classy(m)for m ∈ {1, ..., M}, as in:

x∗y(m) = max

ex∈Tp(ex|y(m)), (6)

wherep(ex|y(m)) = p(y(m)|ex)p(ex)/p(y(m)) with p(ex) = constantandp(y(m)) defined in (3). Furthermore,p(y(m)|ex) is definedas in (1) replacing the last term byp(y(m)|x(r,k)) = δ(y(m) −y(m)(r,k)).

Page 4: Graph-based Methods for the Automatic Annotation and ...carneiro/publications/printart_icmr... · Graph-based Methods for the Automatic ... In this paper, we introduce a graph-based

4. IMPLEMENTATIONIn this section we provide the details of the data set used, the imagerepresentation based on bags of visual words, and the implementa-tion of the following algorithms: label propagation, random walk,stationary solution, and combinatorial harmonic.

4.1 Data SetThe data set consists of 307 annotated images with one multi-classproblem (theme with seven classes) and 21 binary problems (seeFig. 5). All images have been collected from the Artstor digitalimage library [2], and annotated by art historians (Fig. 3 shows anexample of a manual annotation). For the experiments in Sec.5, werun a 10-fold cross validation in order to show the results, and foreach run, divide the data set into a training setD with 276 images(90% of the data set) and a test setT with 31 images (10% of thedata set).

4.2 Image RepresentationThe images are represented with the bag of visual words model[13],where each visual word is formed using a collection of scale in-variant feature transform (SIFT) local descriptors [31]. The visualvocabulary is built using the vocabulary tree proposed by Nistérand Stewénius [36]. The SIFT descriptor consists of a featuretransform applied to an image patch, which extracts a histogramof gradient orientations weighted by the gradient magnitudes. Inthis work, the image patch location (in the image) and scale (patchsize) are randomly determined [37], and we generate 1000 descrip-tors per image. Then, the vocabulary is built by gathering the de-scriptors from all images and running a hierarchical clustering al-gorithm with three levels, where each node in the hierarchy has10 descendants (this hierarchy is a directed tree, where each nodehas at most 10 edges) [36]. This results in a directed tree with1 + 10 + 100 + 1000 = 1111 vertexes, and the image featureis formed by using each descriptor of the image to traverse thetree and record the path (note that each descriptor generates a pathwith 4 vertexes). The histogram of visited vertexes is weighted bythe node entropy (i.e., vertexes that are visited more oftenreceivesmaller weights). As a result, an imageI is represented by thehistogramx ∈ ℜ1111 .

The probability of the transition of feature vectorx(r,j) givenx(r,j−1)

andex of (2) is then defined as:

p(x(r,j)|x(r,j−1), ex) = sx(x(r,j),x(r,j−1))sx(x(r,j), ex), (7)

and the transition probability between two feature vectorsis

p(x(r,1)|ex) = sx(x(r,1), ex), (8)

with sx(xi,xj) = 1Zx

P1111d=1 min(xi(d), xj(d)) whereZx is a

normalization factor. Consequently, we can define the followingadjacency matrix to be used in the graph-based algorithms below:

W(j, i) = sy(yi,yj) × sx(xi,xj) × sx(xj , ex), (9)

with W(i, i) = 0 for all i ∈ {1, ..., N}.

In Fig. 4 it is shown a small part of the graph which takes intoaccount the image features and annotation of the training set de-scribed by the adjacency matrixW in (9). In this part of the graph,we take a training image represented byex shown at the center (notethe enlarged image in the figure), and display the graph structurearound it. Notice that the neighboring images in the graph tend tobe similar in terms of their visual information or their annotationkeywords (for instance, notice in the graph that most of the neigh-boring images are of the themeThe Annunciation, which is the

Figure 4: Network structure in the training set shown usinga variant of the MDS algorithm [5]. The large image in thecenter represents a training image with its most similar images(in visual an annotation terms) closer in the graph.

theme of the central image). In order to show this graph, we use avariant of the multidimensional scaling algorithm for visualization(MDS) [5].

4.3 Label PropagationThe annotation and retrieval using the label propagation model con-sists of a one step random walk procedure that uses only the visualsimilarity, as follows:

p(y|ex) =1

ZLP

X

r∈N

p(x(r)|ex)p(y|x(r)), (10)

whereN ∈ D denotes the set of|N | nearest neighbors relativeto ex using the proximity measure (8), andZLP represents a nor-malization factor. The definition ofp(y|ex) in (10) is used in theannotation (Equations 4 and 5) and retrieval (6) of test images.

4.4 Random WalkThe random walk uses the adjacency matrixW in (9) in order tobuild the probability transition matrix as follows:

P = D−1

W, (11)

where the diagonal matrixD(i, i) =P

jW(i, j), which makes

the row sum ofP one. The initial distribution vector takes intoaccount the similarity between the test imageex and all images inthe database, as inu = [sx(x1, ex), ..., sx(xN , ex)]T , whereu isnormalized in order to have‖u‖1 = 1. The random walk startswith the selection of a training image (sayith training image) bysampling the distributionu. Then, the distribution denoted byπT

i A

(with πi a vector of zeros with a one at theith position) is used toselect the next training image. AfterR = 10 steps of the randomwalk, a list of visited training imagesTr,k is formed, and the testimage annotation is produced by (4) and (5), where the numberofrandom walks isK = 100. The retrieval is produced as describedin (6).

4.5 Stationary SolutionThe stationary solution estimates the result of a random walk witha large number of steps [26]. The adjacency matrixW in (9) isused to build the following normalized transition matrix:

S = D−0.5

WD−0.5, (12)

Page 5: Graph-based Methods for the Automatic Annotation and ...carneiro/publications/printart_icmr... · Graph-based Methods for the Automatic ... In this paper, we introduce a graph-based

with D defined in (11). This solution exploits the eigenvector cen-trality (i.e., the eigenvector ofS associated with the eigenvalue 1),in order to determine the rank of a node (recall that a node repre-sents a database image). This rank denotes the likelihood ofvisitingthe node after a large number of random steps using a random start-ing point, where the decision to visit graph nodes is based ontheedge weights [26].

Assuming a random initial distribution of the vertexes denoted bythe vectorv(0), and that at each iteration of the random walk, thedistribution used for the decision process is based on the weightededges and on the probability vectoru in (11) weighted by(1− α),with 0 ≤ α < 1, we compute the stationary vector as follows [49]:

v(t) = (αS)t

v(0) + (I + αS)(1 − α)u ⇒

v(∞) = (1 − α)(I − αS)−1

u,(13)

with I denoting the identity matrix.

Finally, in order to produce the annotation defined in (4) and(5),and the retrieval in (6), we define the probability of label given atest image, as follows:

p(y|ex) =1

ZSS

NX

i=1

v(∞)i δ(y − yi), (14)

wherev(∞)i is theith component ofv(∞) defined in (13), andZSS

represents a normalization factor.

4.6 Combinatorial HarmonicThe solution based on combinatorial harmonic follows the nota-tion of semi-supervised labeling [51] and image segmentation [19]problems. Consider the following extension of the adjacency ma-trix (9), which includes the similarity between the test image anddatabase images:

fW =

"W eueuT 0

#, (15)

whereW is the adjacency matrix in (9) andeu is the un-normalizedinitial distribution vectoru defined in (11). The goal is then tofind the distributionfU ∈ ℜN (‖fU‖1 = 1), which represents theprobability of first reaching each of the training images in arandomwalk procedure, where the labeling matrix representing thetrainingimages is denoted byFL = IN , whereIN is anN × N identitymatrix. In order to findfU , we minimize the following energy func-tion:

E(F) =1

2

‚‚‚‚‚[FL, fU ]eL"F

TL

fTU

#‚‚‚‚‚

2

2

, (16)

whereF = [FL, fU ], and eL = eD − fW is the Laplacian matrixcomputed from the the adjacency matrixfW (15), with diagonalmatrix eD(i, i) =

Pj

fW(i, j). This Laplacian matrix can be di-

vided into the same blocks as infW, that is

eL =

»LW B

B LU

–. (17)

Hence, in order to findfU , we solve the following optimizationproblem [19]:

minimize E(F)subject to FL = IN

(18)

which is convex given thateL is positive semi-definite. Eq. 18 issolved by setting∂E(F)

∂fU= 0, which leads to the following analyti-

cal solution:fTU = −L−1

U BT IN

The annotation defined in (4) and (5), and the retrieval in (6)arecomputed using the following probability of label given test image:

p(y|ex) =1

ZCH

NX

i=1

fU (i)δ(y − yi), (19)

wherefU (i) is theith component offU , andZCH is a normaliza-tion factor.

5. EXPERIMENTSIn this experiment we compare the model presented in (3) to modelsbased on SVM [45] and on RF [7] classifiers. For the RF model, webuildL = 22 independent classifiers (one for the multi-class themeclassification and the others for the binary problems - see Sec. 4.1),where each classifier is defined asp(yl|ex, θRF (l)), with θRF (l)representing the parameters of the random forests classifier for thelth classification problem (recall thatl = 1, ..., L). We obtainby using the notation defined in (3), but replacingp(yi(m)|ex) byp(yl(m)|ex, θRF (l)). The main parameters of the random forests,which are the number and height of trees, are determined withcrossvalidation, where the training setD is further divided into a train-ing and validation sets of90% and10% of D, respectively. Theannotation is performed with (4) and (5), and the retrieval is donewith (6). For the SVM, we trainM = 28 classifiers using theone-versus-all training method. Specifically, we train thefollowingclassifiersp(y(m)|ex, θSV M (m)), for m ∈ {1, ..., M}, and theannotation confidenceby is produced by (3), replacingp(yi(m)|ex)by p(y(m)|ex, θSV M (m)). The main parameter of the support vec-tor machine, which is the penalty factor for the slack variables, isalso determined with cross validation, where the training set D isdivided into a training and validation sets of90% and10% of D, re-spectively. Also, we perform the test image annotation with(4) and(5), and the retrieval with (6). Note that these two models (RF andSVM) roughly represent the state-of-the-art approaches for imageannotation and retrieval problems explained in Sec. 2.

In the results below, we use the following acronyms: LPi for labelpropagation withi nearest neighbors in (10), SS for stationary so-lution, CH for combinatorial harmonic, RF for random forests andSVM for support vector machines.

5.1 RetrievalWe measure the performance of the system in terms of retrievalusing the precision and recall measures [10]. For each annotationclassy(m) belonging to the set of classes in the test setT find theQ test images that produce the maximum values for (6). Out ofthoseQ images, let the setA be the images for whichy(m) = 1(note that|A| ≤ Q). Also, letB ⊂ T be the set of all test imagesthat havey(m) = 1. Then, the precision and recall are computedas follows:

precisionR =|A|

Q, andrecallR =

|A|

|B|. (20)

The performance is computed with the mean average precision[10](MAP), which is defined as the average precision over all queries,at the ranks where the recall changes. The results in Table 1 (firstcolumn) show the MAP average and standard deviation for the 10-fold cross validation experiment. Figure 6 (left) displaysthe re-trieval performance of the CH solution as a function of the number

Page 6: Graph-based Methods for the Automatic Annotation and ...carneiro/publications/printart_icmr... · Graph-based Methods for the Automatic ... In this paper, we introduce a graph-based

Table 1: Average± standard deviation of the retrieval and an-notation performance over the 10-fold cross validation experi-ment (the best performance for each measure is highlighted).

Models MAP Mean per-class Mean per-classRecall Precision

LP1 0.33 ± .03 0.43 ± .04 0.43 ± .04LP10 0.33 ± .03 0.57 ± .05 0.31 ± .03LP20 0.33 ± .02 0.58 ± .06 0.31 ± .02RW 0.31 ± .03 0.51 ± .06 0.30 ± .04CH 0.35 ± .03 0.66 ± .05 0.32 ± .03SS 0.33 ± .03 0.57 ± .03 0.30 ± .02RF 0.30 ± .03 0.23 ± .03 0.40 ± .06SVM 0.22 ± .01 0.12 ± .01 0.52 ± .03

Figure 5: Number of training images per class.

0 50 100 150 200 2500

0.2

0.4

0.6

0.8

1

# training images

MA

P

0 50 100 150 200 250 3000

0.2

0.4

0.6

0.8

1

# training images

Ann

otat

ion

perf

orm

ance

Mean PrecisionMean Recall

Figure 6: Performance of the retrieval (left) and annotation(right) as a function of the number of training images usingthe CH algorithm.

of training images. Notice that the retrieval performance is pos-itively correlated with the number of training images. Figure 7shows the the top retrieval results for four annotation classes usingthe CH algorithm.

5.2 AnnotationThe performance of the annotation procedure is evaluated bycom-paring the results of the system in (4) and (5) with the manualan-notation of the ground truth (recall that the setT also contains themanual annotation) [10]. For each annotation problem (binary ormulti-class) indexed byl ∈ 1, ..., L, assume that there arewH man-ually annotated images inT , and the system annotateswauto, of

whichwC are correct. The precision and recall are computed as:

precisionA =wC

wauto

, andrecallA =wC

wH

. (21)

Then, the values ofprecisionA andrecallA are averaged over theset of binary and multi-class problems. The results in Table1 (lasttwo columns) show the average and the standard deviation of theper-class precision and recall for the 10-fold cross validation. Fig. 6(right) displays the annotation performance (mean per class preci-sion and recall) of the CH algorithm as a function of the numberof training images. Notice the positive correlation between preci-sion and recall in terms of the number of training images. Figure 8shows the annotation produced by the CH algorithm in four testimages.

6. DISCUSSION AND CONCLUSIONSIn this work we presented a graph-based model for the annotationof art images. The retrieval experiment (Tab. 1) shows that theCH model produces better results than current state-of-the-art ap-proaches based on SVM and RF. Also notice that the LP methodsalso show competitive results, indicating that simple models canalso lead to powerful methods. The annotation results in Tab. 1show that the CH approach produces the best performance in termsof recall, but SVM is better in terms of precision (but noticethepoor result of SVM in terms of recall). This happens because SVMrarely classifies positively the test images with respect toeach la-bel, but whenever it does so, the annotation is often correct. We be-lieve that this happens due to the limited number of trainingimagesper class to estimate the parameters of the SVM model. We planto improve our method by modeling the dependencies between la-bels using, for example, structural learning methods [44, 47]. Thiswould prevent the following two issues observed in Fig. 8: 1)use oftoo many labels in the annotation, and 2) presence of pairs ofanno-tations that should never appear together (e.g., the presence ofwisemenin prints of themeAnnunciationshould not be allowed). Theincorporation of structural learning in the methodology should beassessed with more appropriate measures, such as the one describedby Nowak et al. [38]. We also intend to investigate other image fea-tures for image representation, such as wavelets and curvelets.

7. ACKNOWLEDGMENTSThe author would like to thank Duarte Lázaro and Rosário Car-valho for their help with the art history issues. I would liketoacknowledge the use of code written by T. Joachims and by A.Vedaldi. I also would like to thank David Lowe, Benito Navarrete,and the members of the PrintArt project for valuable suggestionsduring the development of this work.

8. REFERENCES[1] http://www.google.com/mobile/goggles/#artwork.[2] http://www.artstor.org.[3] F. Baumann, M. Karabelnik, and et al.Degas Portraits.

London: Merrell Holberton, 1994.[4] I. Berezhnoy, E. Postma, and H. van den Herik.

Computerized visual analysis of paintings. InInt. Conf.Association for History and Computing, pages 28–32, 2005.

[5] I. Borg and P. Groenen.Modern Multidimensional Scaling:theory and applications. Springer-Verlag, 2005.

[6] A. Borodin, G. Roberts, J. Rosenthal, and P. Tsaparas. Linkanalysis ranking: algorithms, theory, and experiments.ACMTrans. Internet Techn., 5(1):231–297, 2005.

[7] L. Breiman. Random forests.Machine Learning, 45(1):5–32,2001.

Page 7: Graph-based Methods for the Automatic Annotation and ...carneiro/publications/printart_icmr... · Graph-based Methods for the Automatic ... In this paper, we introduce a graph-based

yes yes yes yes yes

yes yes yes yes no

yes yes yes yes yes

yes yes yes no no

Figure 7: Retrieval results of the CH algorithm on the test set. Each row shows the top five matches to the following queries(fromtop to bottom): ‘rest on the flight into Egypt’, ‘Christ child’, ‘Mary’ , and ‘dove’. Below each image, it is indicated whether the image ismanually annotated with the class.

Human Theme: Annunciation Theme: Magi Theme: Baptism of Christ Theme: Rest flight Egypt

Annotation angels floating, dove, Gabriel Christ-child, Mary, st. Joseph angels, Christ, dove, angels, angels floating, Christ-child

Lilly, Mary, wing wise men st. Frances, wing donkey, Mary, miracle...palm tree

st. Joseph, wing

Comb. Harm. Theme: Annunciation Theme: Magi Theme: Baptism of Christ Theme: Rest flight Egypt

Annotation angels floating, Christ-child, dove, angels, angels floating, Christ, dove, angel, angels, angels floating angels floating, Christ-child, donkey,

Gabriel, Lilly, Mary, shepherd, Gabriel, Lilly, Melchior, st. Elizabeth Christ, dove, st. Frances, Mary, miracle...palm tree, st. Joseph

vase, wing, wise men st. Frances, vase, wing, wise men wing, wings vase, wing, wings

Zacharias

Figure 8: Comparison of CH annotations with those of a human subject on test images.

[8] S. Brin and L. Page. The anatomy of a large-scalehypertextualweb search engine. InComputer Networks andISDN Systems, pages 107–117, 1998.

[9] T. Campos. Application des regles iconographiques auxazulejos portugais du xviieme siecle. InEuropalia, pages37–40, 1991.

[10] G. Carneiro, A. Chan, P. Moreno, and N. Vasconcelos.Supervised learning of semantic classes for image annotationand retrieval.IEEE Trans. Pattern Anal. Mach. Intell.,29(3):394–410, 2007.

[11] G. Carneiro and J. Costeira. The automatic annotation and

retrieval of digital images of prints and tile panels usingnetwork link analysis algorithms. InProceedings of the SIE -Computer Vision and Image Analysis of Art II, 2011.

[12] R. Carvalho. O programa artistico da ermida do rei salvadordo mundo em castelo de vide, no contexto da arte barroca.Artis -Revista do Instituto de Historia da Arte da Faculdadede Letras de Lisboa, (2):145–180, 2003.

[13] G. Csurka, C. Dance, L. Fan, J. Willamowski, and C. Bray.Visual categorization with bags of keypoints. InIn Workshopon Statistical Learning in Computer Vision, ECCV, pages1–22, 2004.

Page 8: Graph-based Methods for the Automatic Annotation and ...carneiro/publications/printart_icmr... · Graph-based Methods for the Automatic ... In this paper, we introduce a graph-based

[14] R. Datta, D. Joshi, J. Li, and J. Wang. Image retrieval: Ideas,influences, and trends of the new age.ACM Comput. Surv.,40(2), 2008.

[15] K. V. de Sande, T. Gevers, and A. Smeulders. The universityof amsterdamâAZs concept detection system at imageclef2009. InCLEF working notes 2009, 2009.

[16] O. Dekel and O. Shamir. Multiclass-multilabel classificationwith more classes than examples. InInternationalConference on Artificial Intelligence and Statistics, 2010.

[17] F. Estrada, D. Fleet, , and A. Jepson. Stochastic imagedenoising. InBMVC, 2009.

[18] S. Feng, R. Manmatha, and V. Lavrenko. Multiple bernoullirelevance models for image and video annotation. InCVPR(2), pages 1002–1009, 2004.

[19] L. Grady. Random walks for image segmentation.IEEETrans. Pattern Anal. Mach. Intell., 28(11):1768–1783, 2006.

[20] D. Graham, J. Friedenberg, D. Rockmore, and D. Field.Mapping the similarity space of paintings: image statisticsand visual perception.Visual Cognition, 18(4):559–573,2010.

[21] S. Griffin. Recovering the past through computation: newtechniques for cultural heritage. InMultimedia InformationRetrieval, pages 13–14, 2010.

[22] X. He, W.-Y. Ma, and H.-J. Zhang. Imagerank: Spectraltechniques for structural analysis of image database. InIEEEInternational Conference on Multimedia and Expo, pages25–28, 2003.

[23] D. Hsu, S. Kakade, J. Langford, and T. Zhang. Multi-labelprediction via compressed sensing. InNeural InformationProcessing Systems, 2009.

[24] J. Hughes, D. Graham, and D. Rockmore. Stylometrics ofartwork: uses and limitations. InProceedings of SPIE:Computer Vision and Image Analysis of Art, 2010.

[25] S. Jafarpour, G. Polatkan, I. Daubechies, S. Hughes, andA. Brasoveanu. Stylistic analysis of paintings using waveletsand machine learning. InEuropean Signal ProcessingConference, 2009.

[26] Y. Jing and S. Baluja. Visualrank: Applying pagerank tolarge-scale image search.IEEE Trans. Pattern Anal. Mach.Intell., 30(11):1877–1890, 2008.

[27] C. Johnson, E. Hendriks, I. Berezhnoy, E. Brevdo,S. Hughes, I. Daubechies, J. Li, E. Postma, and J. Wang.Image processing for artistic identification: Computerizedanalysis of vincent van goghâAZs brushstrokes.IEEE SignalProcessing Magazine, pages 37–48, 2008.

[28] F. Kang, R. Jin, and R. Sukthankar. Correlated labelpropagation with application to multi-label learning. InCVPR, pages 1719–1726, 2006.

[29] J. Li and J. Wang. Studying digital imagery of ancientpaintings by mixtures of stochastic models.IEEE Trans.Image Processing, 13(3):340âAS–353, 2004.

[30] J. Liu, M. Li, Q. Liu, H. Lu, and S. Ma. Image annotation viagraph learning.Pattern Recognition, 42(2):218–228, 2009.

[31] D. Lowe. Distinctive image features from scale-invariantkeypoints.International Journal of Computer Vision, 2004.paper accepted for publication.

[32] S. Lyu, D. Rockmore, and H. Farid. A digital technique forart authentication.Proceedings of the National Academy ofSciences USA, 101(49):17006–âAS17010, 2004.

[33] H. Maitre, F. Schmitt, and C. Lahanier. 15 years of imageprocessing and the fine arts. InIEEE Int. Conf. ImageProcessing, pages 557âAS–561, 2001.

[34] A. Makadia, V. Pavlovic, and S. Kumar. A new baseline forimage annotation. InECCV (3), pages 316–329, 2008.

[35] A. Ng, A. Zheng, and M. Jordan. Link analysis, eigenvectorsand stability. InIJCAI, pages 903–910, 2001.

[36] D. Nistér and H. Stewénius. Scalable recognition with avocabulary tree. InCVPR, pages 2161–2168, 2006.

[37] E. Nowak, F. Jurie, and B. Triggs. Sampling strategies forbag-of-features image classification. InECCV, pages2161–2168, 2006.

[38] S. Nowak, H. Lukashevich, P. Dunker, and S. Rüger.Performance measures for multilabel evaluation: a casestudy in the area of image classification. InMultimediaInformation Retrieval, pages 35–44, 2010.

[39] G. Polatkan, S. Jafarpour, A. Brasoveanu, S. Hughes, andI. Daubechies. Detection of forgery in paintings usingsupervised learning. InInternational Conference on ImageProcessing, 2009.

[40] A. Smeulders, M. Worring, S. Santini, A. Gupta, and R. Jain.Content-based image retrieval at the end of the early years.IEEE Transactions on Pattern Analysis and MachineIntelligence, 22:1349–1380, 2000.

[41] S. Sonnenburg, G. Rätsch, C. Schäfer, and B. Schölkopf.Large scale multiple kernel learning.Journal of MachineLearning Research, 7:1531–1565, 2006.

[42] D. Stork. Computer image analysis of paintings anddrawings: An introduction to the literature. InProceedings ofthe Image Processing for Artist Identification Workshop,2008.

[43] M. Szummer and T. Jaakkola. Partially labeled classificationwith markov random walks. InNIPS, pages 945–952, 2001.

[44] B. Taskar, V. Chatalbashev, D. Koller, and C. Guestrin.Learning structured prediction models: a large marginapproach. InICML, pages 896–903, 2005.

[45] V. N. Vapnik.Statistical Learning Theory.Wiley, 1998.[46] J. Verbeek, M. Guillaumin, T. Mensink, and C. Schmid.

Image annotation with tagprop on the mirflickr set. InMultimedia Information Retrieval, pages 537–546, 2010.

[47] X. Xue, H. Luo, and J. Fan. Structured max-margin learningfor multi-label image annotation. InCIVR, pages 82–88,2010.

[48] M. Yelizaveta, C. Tat-Seng, , and R. Jain. Semi-supervisedannotation of brushwork in paintings domain using serialcombinations of multiple experts. InACM Multimedia, pages529âAS–538, 2006.

[49] D. Zhou, O. Bousquet, T. Lal, J. Weston, and B. Scholkopf.Learning with local and global consistency. InNIPS, pages321–328, 2004.

[50] X. Zhu. Semi-supervised learning literature survey.Technical Report 1530, Computer Sciences, University ofWisconsin-Madison, 2005.

[51] X. Zhu, Z. Ghahramani, and J. Lafferty. Semi-supervisedlearning using gaussian fields and harmonic functions. InICML, pages 912–919, 2003.