convergence rates of smooth message passing with rounding

11
Convergence Rates of Smooth Message Passing with Rounding in Entropy-Regularized MAP Inference Jonathan N. Lee Aldo Pacchiano Michael I. Jordan Stanford University UC Berkeley UC Berkeley Abstract Maximum a posteriori (MAP) inference is a fundamental computational paradigm for statistical inference. In the setting of graphi- cal models, MAP inference entails solving a combinatorial optimization problem to find the most likely configuration of the discrete- valued model. Linear programming (LP) re- laxations in the Sherali-Adams hierarchy are widely used to attempt to solve this problem, and smooth message passing algorithms have been proposed to solve regularized versions of these LPs with great success. This paper leverages recent work in entropy-regularized LPs to analyze convergence rates of a class of edge-based smooth message passing algo- rithms to -optimality in the relaxation. With an appropriately chosen regularization con- stant, we present a theoretical guarantee on the number of iterations sucient to recover the true integral MAP solution when the LP is tight and the solution is unique. 1 INTRODUCTION Undirected graphical models are a central modeling for- malism in machine learning, providing a compact and powerful way to model dependencies between variables. Here we focus on the important class of discrete-valued pairwise models. Inference in discrete-valued graphi- cal models has applications in many areas including computer vision, statistical physics, information theory, and genome research (Antonucci et al., 2014; Mezard and Montanari, 2009; Wainwright and Jordan, 2008). We focus on the problem of identifying a configuration of all variables that has highest probability, termed Equal contribution. Proceedings of the 23 rd International Conference on Artificial Intelligence and Statistics (AIS- TATS) 2020, Palermo, Italy. PMLR: Volume 108. Copyright 2020 by the author(s). maximum a posteriori (MAP) inference. This problem has an extensive literature across multiple communities, where it is described by various names, including en- ergy minimization (Kappes et al., 2013) and constraint satisfaction (Schiex et al., 1995). In the binary case, the MAP problem is sometimes described as quadratic- pseudo Boolean optimization (Hammer et al., 1984) and it is known to be NP-hard to compute exactly (Cooper, 1990; Kolmogorov and Zabin, 2004) or even to ap- proximate (Dagum and Luby, 1993). Consequently, much work has attempted to identify settings where polynomial-time methods are feasible. We call such settings “tractable” and the methods “ecient.” A general framework for obtaining tractable methodology involves “relaxation”—the MAP problem is formulated as an integer linear program (ILP) and is then relaxed to a linear program (LP). If the vertex at which the LP achieves optimality is integral, then it provides an exact solution to the original problem. In this case we say that the LP is tight. If the LP is performed over the convex hull of all integral assignments, otherwise known as the marginal polytope M, then it will always be tight. Inference over the marginal polytope is gener- ally intractable because it requires exponentially many constraints to enforce global consistency. A popular workaround is to relax the marginal poly- tope to the local polytope L 2 (Wainwright and Jordan, 2008). Instead of enforcing global consistency, the lo- cal polytope enforces consistency only over pairs of variables, thus yielding pseudo-marginals which are pairwise consistent but may not correspond to any true global distribution. The number of constraints needed to specify the local polytope is linear in the number of edges. More generally, Sherali and Adams (1990) intro- duced a series of successively tighter relaxations of the marginal polytope, or convex hull, while retaining con- trol on the number of constraints. However, even with these relaxations, it has been observed that standard LP solvers do not scale well (Yanover et al., 2006), mo- tivating the study of solvers that exploit the structure of the problem, such as message passing algorithms. Of particular interest to this paper are smooth mes-

Upload: others

Post on 02-Jan-2022

4 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: Convergence Rates of Smooth Message Passing with Rounding

Convergence Rates of Smooth Message Passing with Rounding inEntropy-Regularized MAP Inference

Jonathan N. Lee⇤ Aldo Pacchiano⇤ Michael I. JordanStanford University UC Berkeley UC Berkeley

Abstract

Maximum a posteriori (MAP) inference isa fundamental computational paradigm forstatistical inference. In the setting of graphi-cal models, MAP inference entails solving acombinatorial optimization problem to findthe most likely configuration of the discrete-valued model. Linear programming (LP) re-laxations in the Sherali-Adams hierarchy arewidely used to attempt to solve this problem,and smooth message passing algorithms havebeen proposed to solve regularized versionsof these LPs with great success. This paperleverages recent work in entropy-regularizedLPs to analyze convergence rates of a classof edge-based smooth message passing algo-rithms to ✏-optimality in the relaxation. Withan appropriately chosen regularization con-stant, we present a theoretical guarantee onthe number of iterations su�cient to recoverthe true integral MAP solution when the LPis tight and the solution is unique.

1 INTRODUCTION

Undirected graphical models are a central modeling for-malism in machine learning, providing a compact andpowerful way to model dependencies between variables.Here we focus on the important class of discrete-valuedpairwise models. Inference in discrete-valued graphi-cal models has applications in many areas includingcomputer vision, statistical physics, information theory,and genome research (Antonucci et al., 2014; Mezardand Montanari, 2009; Wainwright and Jordan, 2008).

We focus on the problem of identifying a configurationof all variables that has highest probability, termed

⇤Equal contribution. Proceedings of the 23

rdInternational

Conference on Artificial Intelligence and Statistics (AIS-

TATS) 2020, Palermo, Italy. PMLR: Volume 108. Copyright

2020 by the author(s).

maximum a posteriori (MAP) inference. This problemhas an extensive literature across multiple communities,where it is described by various names, including en-ergy minimization (Kappes et al., 2013) and constraintsatisfaction (Schiex et al., 1995). In the binary case,the MAP problem is sometimes described as quadratic-pseudo Boolean optimization (Hammer et al., 1984) andit is known to be NP-hard to compute exactly (Cooper,1990; Kolmogorov and Zabin, 2004) or even to ap-proximate (Dagum and Luby, 1993). Consequently,much work has attempted to identify settings wherepolynomial-time methods are feasible. We call suchsettings “tractable” and the methods “e�cient.” Ageneral framework for obtaining tractable methodologyinvolves “relaxation”—the MAP problem is formulatedas an integer linear program (ILP) and is then relaxedto a linear program (LP). If the vertex at which theLP achieves optimality is integral, then it provides anexact solution to the original problem. In this case wesay that the LP is tight. If the LP is performed overthe convex hull of all integral assignments, otherwiseknown as the marginal polytope M, then it will alwaysbe tight. Inference over the marginal polytope is gener-ally intractable because it requires exponentially manyconstraints to enforce global consistency.

A popular workaround is to relax the marginal poly-tope to the local polytope L2 (Wainwright and Jordan,2008). Instead of enforcing global consistency, the lo-cal polytope enforces consistency only over pairs ofvariables, thus yielding pseudo-marginals which arepairwise consistent but may not correspond to any trueglobal distribution. The number of constraints neededto specify the local polytope is linear in the number ofedges. More generally, Sherali and Adams (1990) intro-duced a series of successively tighter relaxations of themarginal polytope, or convex hull, while retaining con-trol on the number of constraints. However, even withthese relaxations, it has been observed that standardLP solvers do not scale well (Yanover et al., 2006), mo-tivating the study of solvers that exploit the structureof the problem, such as message passing algorithms.

Of particular interest to this paper are smooth mes-

Page 2: Convergence Rates of Smooth Message Passing with Rounding

Convergence Rates of Smooth Message Passing with Rounding

sage passing algorithms, i.e. algorithms derived fromregularized versions of the relaxed LP (Hazan andShashua, 2008; Meshi et al., 2012; Ravikumar et al.,2010; Savchynskyy et al., 2011, 2012). These regular-ized LPs conduce to e�cient optimization in practiceand have the special property that their fixed pointsare unique and optimal; however, this comes at the costof solving an approximation of the true MAP problemand, without rounding, they do not recover integralsolutions in general. Non-asymptotic convergence ratesto the optimal regularized function value have beenstudied (Meshi et al., 2012), but guarantees on thenumber of iterations su�cient to recover the optimalintegral assignment of the true MAP problem have notbeen considered to our knowledge.

In this work we provide a sharp analysis of the entropy-regularized MAP inference problem with Sherali-Adams relaxations. We first characterize the approxi-mation error of the regularized LP in l1 distance, basedon new results on entropy-regularized LPs (Weed, 2018).We then analyze an edge-based smooth message passingalgorithm, modified from the algorithms described inWerner (2007) and Ravikumar et al. (2010). We provea O(1/✏2) rate of convergence of iterates in l1 distance.Combining the approximation error and convergenceresults, we present a guarantee on the number of iter-ations su�cient to recover of the true integral MAPassignment using a standard vertex rounding schemewhen the LP relaxation is tight and the solution isunique.

2 RELATED WORK

The idea of entropy regularization to aid optimizationin inference problems is well studied. It is well knownthat solving a scaled and entropy-regularized linearprogram over the marginal polytope yields the scaledGibbs free energy, intimately related to the log parti-tion function, when the temperature parameter equalsone (Wainwright and Jordan, 2008). As the temper-ature parameter is driven to zero, the calculation ofthe free energy reduces to the value of the MAP prob-lem. However, this problem is intractable due to thedi�culty of both computing the exact entropy andcharacterizing the marginal polytope (Deza and Lau-rent, 2009). Therefore, there has been much work intrying to turn this observation into tractable inferencealgorithms. The standard Bethe approximation insteadminimizes an approximation of the true entropy (Bethe,1935). It was show by Yedidia et al. (2003) that fixedpoints of the loopy belief propagation correspond to itsstationary points, but still the optimization problemresulting from this approximation is non-convex andconvergence is not always guaranteed.

To alleviate convergence issues, much work has con-sidered convexifying the free energy problem leadingto classes of convergent convex belief propagation of-ten derived directly from convex regularizers (Hazanand Shashua, 2008; Heskes, 2006; Johnson and Willsky,2008; Meshi et al., 2009; Savchynskyy et al., 2012). Forinstance, Weiss et al. (2007) proposed a general con-vexified belief propagation and explored some su�cientconditions that enable heuristically recovering the MAPsolution of the LP via a convex sum-product variant.However, the approximation error was still unclear andnon-asymptotic convergence rates were not considered.A number of algorithms have also been proposed todirectly optimize the unregularized LP relaxation oftenwith only asymptotic convergence guarantees such asblock-coordinate methods (Globerson and Jaakkola,2008; Kappes et al., 2013; Kovalevsky and Koval, 1975;Tourani et al., 2018; Werner, 2007) and tree-reweightedmessage passing (Kolmogorov, 2006; Wainwright et al.,2005). The relationship between the regularized andunregularized problems can equivalently be viewed asapplying a soft-max to the dual objective typicallyconsidered in the latter to recover that of the former(Nesterov, 2005; Sontag et al., 2011). Many other con-vergent methods exist such as augmented Lagrangian(Martins et al., 2011; Meshi and Globerson, 2011), bun-dle (Kappes et al., 2012), and steepest descent (Schwinget al., 2012, 2014) approaches, but again they are di�-cult to compare without rates.

Most closely related to our work is recent work inconvergence analysis of certain smoothed message pass-ing algorithms that aim to solve the regularized LPobjective. Savchynskyy et al. (2011) proposed an accel-erated gradient method that achieves O(1/✏) conver-gence to the optimal regularized dual objective value.Convergence of the primal iterates was only shownasymptotically. Meshi et al. (2012) considered a gen-eral dual coordinate minimization algorithm based onthe entropy-regularized MAP objective. They provedupper bounds on the rate of convergence to the optimalregularized dual objective value; however, closeness tothe true MAP assignment was not formally character-ized. Furthermore, convergence in the dual objectivevalue again does not make it easy to determine whenthe true MAP assignment can be recovered. Meshiet al. (2015) later studied the benefits of adding aquadratic term to the LP objective instead and provedsimilar guarantees. Ravikumar et al. (2010) also con-sidered entropic and quadratic regularization, using aproximal minimization scheme with inner and outerloops. They additionally provided rounding guaranteesto recover true primal solutions. However, as noted bythe authors, the inexact calculation of the inner loopprevents a convergence rate analysis once combinedwith the outer loop. Additionally, rates on the inner

Page 3: Convergence Rates of Smooth Message Passing with Rounding

Jonathan N. Lee⇤, Aldo Pacchiano⇤, Michael I. Jordan

loop convergence were not addressed.

The approach of this paper can be understood asthe bridging the gap between Meshi et al. (2012)and Ravikumar et al. (2010). Our first contributionis a characterization of the approximation error ofthe entropy-regularized MAP inference problem. Wethen study an edge-based message passing algorithmthat solves the regularized LP, which is essentially asmoothed max-sum di↵usion (Werner, 2007) or theinner loop of the proximal steps of Ravikumar et al.(2010). For our main contribution, we provide non-asymptotic guarantees to the integral MAP assignmentfor this message passing algorithm when the LP is tightand the solution is unique. To our knowledge, this isthe first analysis with rates guaranteeing recovery ofthe true MAP assignment for smooth methods.

3 BACKGROUND

We denote the d-dimensional probability simplex as

⌃d

def.=

�p 2 Rd

+ :P

ipi = 1

. The set of joint dis-

tributions which give rise to p, q 2 ⌃d is defined

as Ud(p, q)def.=

�P 2 Rd⇥d

+ : P = p, P> = q . For

any two vectors or matrices p and q having the samenumber of elements, we use hp, qi to denote the dotproduct, i.e. elementwise multiplication then sumover all elements. We use kpk1 to denote the sumof absolute values of the elements of p. The Breg-man divergence between p, q 2 Rd

+ with respect to a

strictly convex function � : Rd

+ 7! R is D�(p, q)def.=

�(p)��(q)�hr�(q), p�qi. We will consider the Breg-man divergence with respect to the negative entropy

�(p) = �H(p)def.=

Pipi(log pi � 1), where p need not

be a distribution. When p is a distribution, this corre-sponds to the Kullback-Leibler (KL) divergence. TheBregman projection with respect to � of q 2 Rd

+ onto

the set X is defined as PX (q)def.= argmin

p2X D�(p, q).The Hellinger distance between p, q 2 ⌃n is defined

as h(p, q)def.= 1p

2kpp � pqk2, where k · k2 is the l2-

norm. We denote the square of the Hellinger distanceby h2(p, q). We will often deal with marginal vectors

which are ordered collections of joint and marginaldistributions in the form of matrices and vectors, re-spectively.

3.1 Pairwise Models

For a set of vertices, V = {1, . . . , n}, and edges

E , a pairwise graphical model, G def.= {V, E}, is a

Markov random field that represents the joint dis-

tribution of variables XVdef.= (Xi)i2V , taking on

values from the set of states � = {0, . . . , d � 1}.We assume that each vertex has at least one edge.

For pairwise models, the joint distribution can bewritten as a function of doubletons and singletons:

p✓(xV) / exp⇣P

i2V ✓i(xi) +P

ij2E ✓ij(xi, xj)⌘. We

wish to find maximum a posteriori (MAP) estimates ofthis model. That is, we consider the integer program:

maxxV2�n

Pi2V ✓i(xi) +

Pij2E ✓ij(xi, xj). (Int)

The maximization in (Int) can be written as a linearprogram by defining a marginal vector µ over variablevertices {µi}i2V and variable edges {µij}ij2E . Thevector µi 2 Rd

+ represents the marginal distribution

probabilities on vertex i while the matrix µij 2 Rd⇥d

+

represents the joint distribution probabilities sharedbetween vertices i and j. We follow the notation ofGloberson and Jaakkola (2008) and denote indexinginto the vector and matrix variables with parentheses,e.g. µij(xi, xj) for xi, xj 2 �. The set of marginalvectors that are valid probability distributions is knownas the marginal polytope and is defined as

M def.=

(µ : 9 P,

PXi(xi) = µi(xi), 8i, xi

PXi,Xj(xi, xj) = µij(xi, xj),

8ij, xi, xj

)(1)

We can think of M as the set of mean parameters ofthe model for which there exists a globally consistentdistribution P. We abuse notation slightly and duallyview ✓ as a potential “vector.” The edge matrix ✓ij 2Rd⇥d is indexed as ✓ij(xi, xj), indicating the elementat the xith row and xjth column. The vertex vector ✓iis indexed as ✓i(xi), indicating the xith element. TheMAP problem in (Int) can be shown to be equivalentto the following LP (Wainwright and Jordan, 2008):

max h✓,µi s.t. µ 2M

where h✓,µi =P

i2VP

xi✓i(xi)µi(xi) +P

ij2EP

xi,xj✓ij(xi, xj)µij(xi, xj).

3.2 Sherali-Adams Relaxations

The number of constraints in M is unfortunately super-polynomial (Sontag, 2010). This motivates consideringrelaxations of the marginal polytope to outer polytopesthat involve fewer constraints. For example, the local

outer polytope is obtained by enforcing consistency onlyon edges and vertices:

L2def.=

⇢µ � 0 :

µi 2 ⌃d 8i 2 Vµij 2 Ud(µi,µj) 8ij 2 E

�(2)

Relaxations of higher orders have also been studied, inparticular by Sherali and Adams (1990) who introduceda hierarchy of polytopes by enforcing consistency onjoint distributions of increasing order up to n: L2 ◆L3 ◆ . . . ◆ Ln ⌘ M. The corresponding Sherali-Adams LP relaxation of order m is then

max h✓,µi s.t. µ 2 Lm, (LP)

Page 4: Convergence Rates of Smooth Message Passing with Rounding

Convergence Rates of Smooth Message Passing with Rounding

where 1 m n. Because Lm is an outer polytopeof M, we no longer have that the solution to (LP)recovers the true MAP solution of (Int) in general.However if the solution to (LP) is integral, then xi =argmax

xµi(x) recovers the optimal solution of the true

MAP problem. In this case, we say Lm is tight.

4 ENTROPY-REGULARIZED MAP

In this section, we present our first main technicalcontribution, characterizing the approximation errorin the entropy-regularized MAP problem for Sherali-Adams relaxations. In contrast to solving the exact(LP), we aim to solve the entropy-regularized LP:

min hC,µi � 1

⌘H(µ) s.t. µ 2 Lm, (Reg)

where Cdef.= �✓ and H(µ) = hµ,� logµ + i. The

hyperparameter ⌘ adjusts the level of regularization.Denote by µ⇤

⌘the solution of (Reg) where we omit the

reference to m to alleviate notation. In addition totheir extensive history in inference problems, entropy-regularized LPs have arisen in a number of other fieldsto aid optimization when standard LP solvers are insuf-ficient. For example, recent work in optimal transporthas relied on entropy regularization to derive alternat-ing projection algorithms (Benamou et al., 2015; Cuturi,2013) which admit almost linear time convergence guar-antees in the size of the cost matrix (Altschuler et al.,2017). Some of our theoretical results draw inspirationfrom these works.

4.1 Approximation Error

When Lm is tight and the solution is unique, we showthat approximate solutions from solving (Reg) are notnecessarily detrimental because we can apply standardvertex rounding schemes to yield consistent integralsolutions. It was shown by Cominetti and San Martın(1994), and later refined by Weed (2018), that the ap-proximation error of general entropy-regularized linearprograms converges to zero at an exponential rate in⌘. Furthermore, it is possible to determine how large⌘ should be chosen in order for rounding to exactlyrecover the optimal solution to (Int). The result issummarized in the following extension of Theorem 1 ofWeed (2018)1.

Theorem 1. Let R1 = maxµ2Lmkµk1, RH =

maxµ,µ02LmH(µ) � H(µ0), Vm be the set of vertices

of Lm, and V⇤m✓ Vm the set of optimal vertices with

respect to C. Let � = minV12Vm\V⇤m,V22V⇤

mhC, V1i �

1The entropy is defined without the linear o↵set in Weed

(2018).

hC, V2i be the smallest gap in objective value be-

tween an optimal vertex and any suboptimal vertex

of Lm. Suppose Lm is tight and |V⇤m| = 1. If

⌘ � 2R1 log 64R1+2R1+2RH

� , the following rounded solu-

tion is a MAP assignment:

�round(µ⇤

⌘)�i:= argmax

x2�

(µ⇤⌘)i(x)

Proof. Define eC = C+1 1⌘, where 1 denotes an all-ones

vector with the same dimensions as C. If ⌘ � 4R1� then

eV⇤m, the set of optimal vertices of Lm with respect to eC,

satisfies eV⇤m

= V⇤m

and minV12Vm\eV⇤

m,V22eV⇤

m

hC, V1i �hC, V2i � �

2 . If V 2 eV⇤m; and V 0 2 Vm\eV⇤

m, then

h eC, V 0i � h eC, V i � � � 1⌘kV 0 � V k1 � �

2 . Let e� =�2 . If ⌘ � R1 log 64R1+R1+RH

e�, and |eV⇤

m| = 1 then

2R1 exp⇣�⌘ e�

R1+ R1+RH

R1

⌘ 1

32 . And therefore, by

Corollary 9 of Weed (2018) minµ2V⇤mkµ� µ⇤

⌘k1 1

32 .

Since Lm is assumed to be tight and eV⇤m

= V⇤m

containsa single integral vertex µ⇤, the last equation impliesround(µ⇤

⌘) = µ⇤.

Consequently, since R1 P

m

j=1

�n

j

�dj and RH P

m

j=1

�n

j

�log(dj)2, we have:

Corollary 1. If Lm is tight, |V⇤m| = 1, and ⌘ �

log(8mnmdm)+2mn

mdm

� , the rounded solution round(µ⇤⌘)

is a MAP assignment.

In general the dependence of � on ⌘ suggested by The-orem 1 is not improvable (Weed, 2018). Nevertheless,when m = 2 and d = 2, since all vertices in V2 haveentries equal to either 0, 1

2 or 1—see Padberg (1989)or Theorem 3 of Weller et al. (2016)—if the entriesof C are all integral, we have � � 1

2 , thus yielding amore concrete guarantee. The disadvantage of choos-ing exorbitantly large ⌘ is that e�cient computationof solutions often becomes more di�cult in practice(Altschuler et al., 2017; Benamou et al., 2015; Weed,2018). Thus, in practice, there exists a trade-o↵ be-tween computation time and approximation error thatis controlled by ⌘. We will provide a precise theoreticalcharacterization of the trade-o↵ in Section 6. In ourguarantees, multiplying C by a constant a (and there-fore multiplying � by a) is equivalent to multiplying ⌘by the same value.

4.2 Equivalent Bregman Projection

The objective (Reg) can be interpreted as a Bregmanprojection. This interpretation has been explored by

2For m = 2 we can get tighter bounds corresponding to

the number of edges in the graph G.

Page 5: Convergence Rates of Smooth Message Passing with Rounding

Jonathan N. Lee⇤, Aldo Pacchiano⇤, Michael I. Jordan

Ravikumar et al. (2010) as a basis for proximal up-dates and also Benamou et al. (2015) for the optimaltransport problem. The objective is equivalent to

min D� (µ, exp(�⌘C)) s.t. µ 2 Lm, (Proj)

where � := �H. The derivation, based on a mirrordescent step can be found in the appendix. The pro-jection, however, cannot be computed in closed formin general due to the complex geometry of Lm.

Ravikumar et al. (2010) proposed using the Bregman

method (Bregman, 1966), which has been applied inmany fields to solve di�cult constrained problems (Be-namou et al., 2015; Goldstein and Osher, 2009; Osheret al., 2005, 2010), to compute PLm

(exp(�⌘C)) forthe inner loop calculation of their proximal algorithm.While the outer loop proximal algorithm can be shownto converge at least linearly, the inner loop rate wasnot analyzed and the constants (possibly dependenton dimension) were not made clear. Furthermore, theBregman method is in general inexact, which makesthe approximation and the e↵ect on the outer loopunclear (Liu and Ihler, 2013).

5 SMOOTH MESSAGE PASSING

We are interested in analyzing a class of algorithmsclosely inspired by max-sum di↵usion (MSD) as pre-sented by Werner (2007) and the proximal updates ofRavikumar et al. (2010) to solve (Proj) over the L2

polytope. We describe it in detail here, with a fewminor modifications and variations to facilitate theo-retical analysis. In L2, the constraints occur only overedges between vertices3. Given an edge ij 2 E , wemust enforce the constraints prescribed by (2), whichis the intersection of the following sets:

(a) Xij!i = {µ : µij = µi}(b) Xij,i = {µ : µ>

i= 1, >µij = 1}

(c) Xij!j = {µ : µ>ij

= µj}(d) Xij,j = {µ : µ>

j= 1, >µij = 1}.

The normalization of the joint distribution µij in (b)and (d) is actually a redundant constraint, but it fa-cilitates analysis as we demonstrate in Section 6. Foreach of these a�ne constraints, we can compute theBregman projections in closed form with simple multi-plicative updates.

Proposition 1. For a given edge ij 2 E, the closed-

form solutions of the Bregman projections for each of

the above individual constraints are given below.

3Written explicitly, the constraints actually occur be-

tween any pair of vertices, but these variables play no role

in the objective or constraints.

Algorithm 1 EMP-cyclic (C, ⌘, ✏)

1: µ Normalize(exp(�⌘C))2: k 1

3: while maxij

(max{kµ(k)

ij� µ(k)

ik1,

k(µ(k)ij

)> � µ(k)jk1}

)� ✏ do

4: µ µ(k)

5: for ij 2 E do6: µ (PXij,j

� PXij!j� PXij,i

� PXij!i)(µ)

7: end for8: µ(k+1) µ9: k k + 1

10: end while11: return round(µ(k))

Figure 1: The EMP-cyclic algorithm (Ravikumar et al.,

2010) projects on all edges in order until the constraints

are satisfied up to ✏ in l1 distance. The operator � denotes

the composition of the projection operations.

(a) Left consistency: If µ0 = PXij!i(µ), then for all

xi, xj 2 �, µ0ij(xi, xj) µij(xi, xj)

qµi(xi)P

xµij(xi,x)

and µ0i(xi) µi(xi)

qPxµij(xi,x)µi(xi)

.

(b) Left normalization: If µ0 = PXij,i(µ), then

for all xi 2 �, µ0i µiP

xµi(x)

and µ0ij

µijPxi,xj

µij(xi,xj).

(c) Right consistency: If µ0 = PXij!j(µ), then for all

xi, xj 2 �, µ0ij(xi, xj) µij(xi, xj)

qµj(xj)P

xµij(x,xj)

and µ0j(xj) µj(xj)

qPxµij(x,xj)µj(xj)

.

(d) Right normalization: If µ0 = PXij,j(µ), then

for all xj 2 �, µ0j µjP

xµj(x)

and µ0ij

µijPxi,xj

µij(xi,xj).

These update rules are similar to a number of algo-rithms throughout the literature on LP relaxations.Notably, they can be viewed as a smoothed version ofMSD (Kovalevsky and Koval, 1975; Werner, 2007) inthat the updates enforce agreement between variableson the edges and vertices. Nearly identical smoothedupdates were also initially proposed by Ravikumar et al.(2010). As in MSD, it is common for message passingschemes derived from LP relaxations to operate on dualobjective instead. We presented the primal view hereas the Bregman projections lend semantic meaning tothe updates and ultimately the stopping conditions inthe algorithms. An equivalent dual view is presentedin Appendix C.1.

Based on these update rules, we formally outline thealgorithms we wish to analyze, which we call edge-based message passing (EMP) for convenience. We

Page 6: Convergence Rates of Smooth Message Passing with Rounding

Convergence Rates of Smooth Message Passing with Rounding

Algorithm 2 EMP-greedy (C, ⌘, ✏)

1: µ Normalize(exp(�⌘C))2: k 1

3: while maxij

(max{kµ(k)

ij� µ(k)

ik1,

k(µ(k)ij

)> � µ(k)jk1}

)� ✏ do

4: ij GreedyEdge(µ(k))

5: if kµ(k)ij� µ(k)

ik1 > k(µ(k)

ij)> � µ(k)

jk1 then

6: µ(k+1) (PXij,i� PXij!i

)(µ(k))7: else8: µ(k+1) (PXij,j

� PXij!j)(µ(k))

9: end if10: k k + 111: end while12: return round(µ(k))

Figure 2: The EMP-greedy algorithm selects the edge and

direction with the greatest constraint violation and projects

until all constraints are satisfied up to ✏ in l1 distance.

consider two variants: EMP-cyclic (Algorithmic 1),which cyclically applies the updates to each edge ineach iteration and EMP-greedy (Algorithmic 2), whichapplies a single projection update to only the edge withthe greatest constraint violation in each iteration. Weemphasize that these algorithms are not fundamentallynew, but our analysis in the next section is our maincontribution. EMP-cyclic is the Bregman method,almost exactly the inner loop proposed by Ravikumaret al. (2010). In both variants, µ(1) is defined as thenormalized value of exp(�⌘C). The GreedyEdgeoperation in EMP-greedy is defined as

GreedyEdge(µ) = argmaxij2E

⇢max{kµ(k)

ij � µ(k)i k1,

k(µ(k)ij )

> � µ(k)j k1}

These procedures are then repeated again until thestopping criterion is met, which is that µ(k) is ✏-closeto satisfying the constraint that the joint distributionssum to the marginals for all edges. Both algorithms alsoconclude with a rounding operation. Any fixed point ofEMP must correspond to an optimal µ⇤

⌘(see details in

appendix). Computationally, EMP-greedy requires asearch over the edges to identify the greatest constraintviolation, which can be e�ciently implemented using amax-heap (Nutini et al., 2015).

6 THEORETICAL ANALYSIS

We now present our main contribution, a theoreticalanalysis of EMP-cyclic and EMP-greedy. This resultcombines two aspects. First, we present a convergenceguarantee on the number of iterations su�cient tosolve (Proj), satisfying the L2 constraints with ✏ > 0error in l1 distance. We note that, in finite iterations,the pseudo-marginals of EMP are not primal feasible

in general due to this ✏-error. We then combine thisresult with our guarantee on the approximation error inTheorem 1 to show a bound on the number of iterationssu�cient to recover the true integral MAP assignmentby rounding, assuming the LP is tight and the solutionis unique. This holds with su�cient iterations and asu�ciently large regularization constant even though

the pseudo-marginals may not be primal feasible. Weemphasize that these theorems are a departure fromusual convergence rates in the literature (Meshi et al.,2012, 2015). Prior work has guaranteed convergencein objective value to the optimum of the regularizedobjective (Proj), making it unclear whether the optimalMAP assignment can be recovered, e.g. by rounding.We address this ambiguity in our results.

We begin with the upper bound iterations to obtain✏-close solutions, which is the result of two facts whichwe show. The first is that the updates in Proposition 1monotonically improve a Lyapunov (potential) functionby an amount proportional to the constraint violationas measured via the Hellinger distance. The secondis that the di↵erence between the initial and optimalvalues of the Lyapunov function is bounded.

Let deg(G) denote the maximum degree of graph G anddefine:

Sdef.=

X

ij2E

2

4logX

xi,xj2�

e�⌘Cij(xi,xj) +

X

xi,xj2�

⌘d2

Cij(xi, xj)

3

5

+

X

i2V

"log

X

x2�

e�⌘Ci(x) +

X

x2�

⌘dCi(x)

#.

Theorem 2. For any ✏ > 0, EMP is guaranteed to

satisfy kµij � µik1 < ✏ and kµ>ij� µjk1 < ✏ for all

ij 2 E in d 4S0(deg(Gk)+1)✏2

e iterations for EMP-cyclic

and d 4S0✏2e iterations for EMP-greedy.

Here, S0 = min(k⌘C/d+ exp(�⌘C)k1, S). In this the-orem, we give our guarantee in terms of l1 distancerather than function value convergence. As we willsee, this is significant, allowing us to relate this resultto Theorem 1 in order to derive the main result. Theproof is similar in style to Altschuler et al. (2017). Weleave the full proof for EMP-cyclic for the appendixdue to a need to handle tedious edge cases, but we stateseveral intermediate results and sketch the proof forEMP-greedy for intuition as it reveals possibly how sim-ilar message passing algorithms can be analyzed. Wefirst introduce a Lyapunov function written in terms ofdual variables (�, ⇠), indexed by the edges and verticesto which they belong in L2. We denote the iteration-indexed dual variables as (�(k), ⇠(k)). For a given edgeij 2 E , constraints enforcing row and column consis-tency correspond to �ij ,�ji 2 Rm, respectively. Nor-malizing constraints correspond to ⇠i, ⇠j , ⇠ij 2 R. TheLyapunov function, L(�, ⇠), is shown in Figure 3.

Page 7: Convergence Rates of Smooth Message Passing with Rounding

Jonathan N. Lee⇤, Aldo Pacchiano⇤, Michael I. Jordan

L(�, ⇠) = �P

ij2EP

xi,xj2�exp (�⌘Cij(xi, xj)� �ij(xi)� �ji(xj)� ⇠ij)

�P

i2VP

x2�exp

⇣�⌘Ci(x)� ⇠i +

Pj2Nr(i)

�ij(x) +P

j2Nc(i)�ji(x)

�P

ij2E ⇠ij �P

i2V ⇠i +P

ij2EP

xi,xj2�exp(�⌘Cij(xi, xj)) +

Pi

Px2�

exp(�⌘Ci(x))

(3)

Figure 3: The proposed Lyapunov function. Nr(i) denotes the set of neighboring vertices of i where row consistency is

enforced. Nc(i) is the same for column consistency. The Lyapunov function L can be derived from the dual objective

of (Proj). A full derivation is provided in the appendix.

We note that maximizing L over (�, ⇠) satisfies allconstraints and yields the solution to (Proj) by first-order optimality conditions. We now present a resultthat establishes the monotone improvement in L dueto the updates in Proposition 1.Lemma 1. For a given edge ij 2 E, let µ0

and (�0, ⇠0)denote the updated primal and dual variables after a

projection from one of (a)–(d) in Proposition 1. We

have the following improvements on L. If µ0is equal

to:

(a) PXij!i(µ), then L(�0, ⇠0)� L(�, ⇠) = 2h2

(µij ,µi)

(b) PXij,i(µ), then L(�0, ⇠0)� L(�, ⇠) � 0

(c) PXij!j(µ), then L(�0, ⇠0)� L(�, ⇠) = 2h2

(µ>ij ,µj)

(d) PXij,j(µ), then L(�0, ⇠0)� L(�, ⇠) � 0.

This result shows that L improves monotonically aftereach of the four updates in Proposition 1. Furthermore,at every update, L improves by twice the squaredHellinger distance of the constraint violation betweenthe joint and the marginals.

Lemma 2. Let �⇤, ⇠⇤ denote the maximizers of L.

The di↵erence in function value between the optimal

value of L and the first iteration value is upper bounded

L(�⇤, ⇠⇤)� L(�(1), ⇠(1)) S0.

Turning to Theorem 2, the result is obtained by observ-ing that as long as the constraints are violated by anamount ✏ > 0 (i.e., the algorithm has not terminated),then the Lyapunov function must improve by a knownpositive amount at each iteration. We provide a proofsketch for EMP-greedy.

Proof Sketch of Theorem 2 for EMP-greedy. We nowshow how to combine the results of Lemma 1 andLemma 2 to obtain Theorem 2. Let k⇤ be the first iter-ation such that the termination condition in Algorithm2 holds with respect to some ✏ > 0. Then, for any ksatisfying 1 k < k⇤, we have that GreedyEdge(µ)selects ij such that either kµij � µik1 � ✏ orkµ>

ij� µjk1 � ✏.

Without loss of generality, suppose kµij � µik1 �kµ>

ij� µjk1. Therefore, we have

✏2

4 1

4kµij � µik21 2h2(µij ,µi),

where again h2(µij ,µi) denotes the squared Hellingerdistance and the last inequality is the Hellinger in-equality. Since µij and µi are normalized for eachiteration, this inequality is valid. Thus, L improvesby 2h2(µij ,µi) when PXij!i

occurs and by a non-negative amount when PXij,i

occurs by Lemma 1.

Therefore, we can guarantee improvement of at least ✏2

4each iteration. Since the optimality gap is at most S0

by Lemma 2, this means the algorithm must terminatein d 4S0

✏2e iterations.

We now turn to our main theoretical result. We com-bine our approximation and iteration convergence guar-antees to fully characterize the convergence of EMP forL2 to the optimal MAP assignment when the relaxationis tight and the solution is unique.

Theorem 3. Let ⌘ � 2 log(16n2d2)+16|E|d2

min(�,1

128 ), and ✏�1 >

(25d deg(G)|E|)2 max (⌘kCk1, 68). If L2 is tight and

|V⇤2 | = 1, the EMP algorithm returns a MAP assign-

ment after d 4S0(deg(Gk)+1)✏2

e iterations for EMP-cyclic

and after d 4S0✏2e iterations for EMP-greedy.

When C is integral, � � 12 , yielding a bound of all

known parameters. The main technical challenge inproducing this result is to relate the termination con-dition of EMP to the l1 distance between µ(k) andµ⇤ (the MAP assignment), as this may lie outsidethe polytope L2. It does not su�ce to provide con-vergence guarantees in function value as the goal ofMAP inference is to produce integral assignments. Theproof proceeds in two steps. First we show that µ(k)

is the entropy-regularized solution to objective C over

a “slack” polytope L⌫(k)

2 . Where the slack vector ⌫(k)

corresponds to the constraint violations of µ(k). We usethis characterization to “project” µ(k) onto a nearbyL2 feasible point µ(k)(2). Second, we can use the prop-erties of the primal objective to bound µ(k)(2) and µ⇤

⌘.

The proof is in the appendix.

7 NUMERICAL EXPERIMENTS

We illustrate our theoretical results in a practical appli-cation of the EMP algorithms. Ravikumar et al. (2010)already gave empirical evidence that the basic EMP-

Page 8: Convergence Rates of Smooth Message Passing with Rounding

Convergence Rates of Smooth Message Passing with Rounding

Figure 4: A box-plot showing the e↵ect of graph size (x-axis)and regularization on the quality of rounded solutions for

both algorithm variants after 80 iterations. Thick horizontal

bars indicate the median over 20 trials each. For large ⌘(cyan and purple), the true MAP is almost always recovered.

cyclic is competitive with standard solvers. Therefore,the objective of these experiments is to understandhow graph and algorithm properties a↵ect approxima-tion (Theorem 1) and convergence (Theorem 2). Weconsider the family of multi-label Potts models (Wain-wright et al., 2005) with d = 3 labels on L2. For eachtrial, the cost vector is Ci(xi) = ↵i(xi), 8i, xi and

Cij(xi, xj) =

(�ij xi = xj

0 otherwise8ij, xi, xj

where the parameters are random ↵i(xi) ⇠Unif(�0.5, 0.5) and �ij ⇠ Unif{�0.1, 0.1}. The graphsconsidered are structured as

pn⇥pn grids (Erdogdu

et al., 2017; Globerson and Jaakkola, 2008; Ravikumaret al., 2010) and as Erdos-Renyi random graphs withedge probability p = 1.1 logn

n. To evaluate recovery

of the optimal MAP assignment, we first solved eachgraph with the ECOS LP solver (Domahidi et al., 2013)and selected graphs that were tight. Solving the LP tofind the ground-truth was the main computational bot-tleneck. Further details can be found in Appendix E.

Approximation In Figure 4, we evaluate the e↵ectof regularization and graph size on the quality of thenearly converged solution from EMP for over 80 it-erations on grids. The box-plots indicate that largechoices of ⌘ often yield the exact MAP solution (cyanand purple). Moderate choices still yield competitivesolutions but not optimal for larger graphs (orange andgreen). Low choices generally give poor solutions withhigh spread for all graph sizes (red and blue).

Convergence We then investigate the e↵ects of reg-ularization on convergence for both variants. Figure 5

Figure 5: On grids of size n = 2500, convergence rates to

the optimal MAP assignment of greedy and cyclic variants

are shown. The lines on each plot indicate choices of ⌘.

Figure 6: The algorithm variants on Erdos-Renyi random

graphs with n = 400, ⌘ = 700.0, and maximum degrees

deg(G) = 5, 10. The higher degree graphs (red and blue)

take longer to converge to the optimal MAP assignment.

illustrates the distance of the rounded solution to theoptimal MAP solution over projection steps on gridsof size n = 2500. EMP-greedy converges sharply andvarying regularization has less of an e↵ect on its conver-gence rate. Finally, in Figure 6, we look at Erdos-Renyirandom graphs to observe the e↵ect of the graph struc-ture for both variants. We considered degree-limitedrandom graphs with deg(G) = 5 and deg(G) = 10.The figure shows convergence over projection stepsfor graphs of size n = 400. For both variants, theconvergence rate deteriorates for higher degrees.

8 CONCLUSION

In this paper, we investigated the approximation e↵ectsof entropy regularization on MAP inference objectives.We combined these approximation guarantees with aconvergence analysis of an edge-based message passingalgorithm that solves the regularized objective to de-rive guarantees on the number of iterations su�cientto recover the true MAP assignment. We also showedempirically the e↵ect of regularization and graph prop-ertise on both the approximation and convergence. Infuture work, we wish to extend the analyses and prooftechniques to higher order polytopes and general block-coordinate minimization algorithms.

Page 9: Convergence Rates of Smooth Message Passing with Rounding

Jonathan N. Lee⇤, Aldo Pacchiano⇤, Michael I. Jordan

Acknowledgements We thank the anonymous re-viewers and Marco Pavone for their invaluable feed-back.

References

Jason Altschuler, Jonathan Weed, and Philippe Rigol-let. Near-linear time approximation algorithms foroptimal transport via Sinkhorn iteration. In Ad-

vances in Neural Information Processing Systems,pages 1964–1974, 2017.

Alessandro Antonucci, Cassio De Campos, and MarcoZa↵alon. Probabilistic graphical models. In Intro-

duction to Imprecise Probabilities. Wiley, 2014.

Jean-David Benamou, Guillaume Carlier, Marco Cu-turi, Luca Nenna, and Gabriel Peyre. Iterative Breg-man projections for regularized transportation prob-lems. SIAM Journal on Scientific Computing, 37(2):A1111–A1138, 2015.

Hans Bethe. Statistical theory of superlattices. Pro-

ceedings of the Royal Society of London. Series A-

Mathematical and Physical Sciences, 150(871):552–575, 1935.

Lev M Bregman. A relaxation method of finding acommon point of convex sets and its application toproblems of optimization. In Soviet Mathematics

Doklady, volume 7, pages 1578–1581, 1966.

Sebastien Bubeck. Convex optimization: Algorithmsand complexity. Foundations and Trends in Machine

Learning, 8(3-4):231–357, 2015.

Roberto Cominetti and Jaime San Martın. Asymptoticanalysis of the exponential penalty trajectory in lin-ear programming. Mathematical Programming, 67(1-3):169–187, 1994.

Gregory Cooper. The computational complexity ofprobabilistic inference using Bayesian belief networks.Artificial Intelligence, 42(2-3):393–405, 1990.

Marco Cuturi. Sinkhorn distances: Lightspeed compu-tation of optimal transport. In Advances in Neural

Information Processing Systems, pages 2292–2300,2013.

Paul Dagum and Michael Luby. Approximating proba-bilistic inference in Bayesian belief networks is NP-hard. Artificial Intelligence, 60(1):141–153, 1993.

Michel Deza and Monique Laurent. Geometry of Cuts

and Metrics, volume 15. Springer, 2009.

Alexander Domahidi, Eric Chu, and Stephen Boyd.ECOS: An SOCP solver for embedded systems. In2013 European Control Conference (ECC), pages3071–3076. IEEE, 2013.

Murat Erdogdu, Yash Deshpande, and Andrea Monta-nari. Inference in graphical models via semidefinite

programming hierarchies. In Advances in Neural In-

formation Processing Systems, pages 417–425, 2017.

Amir Globerson and Tommi Jaakkola. Fixing max-product: Convergent message passing algorithms forMAP LP-relaxations. In Advances in Neural Infor-

mation Processing Systems, pages 553–560, 2008.

Tom Goldstein and Stanley Osher. The split Bregmanmethod for L1-regularized problems. SIAM journal

on Imaging Sciences, 2(2):323–343, 2009.

Peter L Hammer, Pierre Hansen, and Bruno Simeone.Roof duality, complementation and persistency inquadratic 0–1 optimization. Mathematical Program-

ming, 28(2):121–155, 1984.

Tamir Hazan and Amnon Shashua. Convergentmessage-passing algorithms for inference over gen-eral graphs with convex free energies. In Proceedings

of the Twenty-Fourth Conference on Uncertainty in

Artificial Intelligence, pages 264–273, 2008.

Tom Heskes. Convexity arguments for e�cient mini-mization of the bethe and kikuchi free energies. Jour-nal of Artificial Intelligence Research, 26:153–190,2006.

Jason Johnson and Alan Willsky. Convex relaxation

methods for graphical models: Lagrangian and maxi-

mum entropy approaches. PhD thesis, MassachusettsInstitute of Technology, Department of ElectricalEngineering, 2008.

Joerg Kappes, Bjoern Andres, Fred Hamprecht,Christoph Schnorr, Sebastian Nowozin, Dhruv Batra,Sungwoong Kim, Bernhard Kausler, Jan Lellmann,Nikos Komodakis, et al. A comparative study ofmodern inference techniques for discrete energy min-imization problems. In Proceedings of the IEEE con-

ference on computer vision and pattern recognition,pages 1328–1335, 2013.

Jorg Hendrik Kappes, Bogdan Savchynskyy, andChristoph Schnorr. A bundle approach to e�cientmap-inference by lagrangian relaxation. In 2012

IEEE Conference on Computer Vision and Pattern

Recognition, pages 1688–1695. IEEE, 2012.

Vladimir Kolmogorov. Convergent tree-reweighted mes-sage passing for energy minimization. IEEE trans-

actions on pattern analysis and machine intelligence,28(10):1568–1583, 2006.

Vladimir Kolmogorov and Ramin Zabin. What energyfunctions can be minimized via graph cuts? IEEE

Transactions on Pattern Analysis and Machine In-

telligence, 26(2):147–159, 2004.

VA Kovalevsky and VK Koval. A di↵usion algorithmfor decreasing energy of max-sum labeling problem.Glushkov Institute of Cybernetics, Kiev, USSR, 1975.

Page 10: Convergence Rates of Smooth Message Passing with Rounding

Convergence Rates of Smooth Message Passing with Rounding

Qiang Liu and Alexander Ihler. Variational algorithmsfor marginal map. The Journal of Machine Learning

Research, 14(1):3165–3200, 2013.

Andre Martins, Mario Figueiredo, Pedro Aguiar, NoahSmith, and Eric Xing. An augmented Lagrangianapproach to constrained MAP inference. In Interna-

tional Conference on Machine Learning, volume 2,page 2, 2011.

Ofer Meshi and Amir Globerson. An alternating di-rection method for dual MAP LP relaxation. InJoint European Conference on Machine Learning and

Knowledge Discovery in Databases, pages 470–483.Springer, 2011.

Ofer Meshi, Ariel Jaimovich, Amir Globerson, andNir Friedman. Convexifying the bethe free energy.In Proceedings of the Twenty-Fifth Conference on

Uncertainty in Artificial Intelligence, pages 402–410.AUAI Press, 2009.

Ofer Meshi, Amir Globerson, and Tommi S Jaakkola.Convergence rate analysis of map coordinate mini-mization algorithms. In Advances in Neural Infor-

mation Processing Systems, pages 3014–3022, 2012.

Ofer Meshi, Mehrdad Mahdavi, and Alex Schwing.Smooth and strong: MAP inference with linear con-vergence. In Advances in Neural Information Pro-

cessing Systems, pages 298–306, 2015.

Marc Mezard and Andrea Montanari. Information,

Physics, and Computation. Oxford University Press,2009.

Arkadii Nemirovsky and David Yudin. Problem com-plexity and method e�ciency in optimization. 1983.

Yu Nesterov. Smooth minimization of non-smoothfunctions. Mathematical programming, 103(1):127–152, 2005.

Julie Nutini, Mark Schmidt, Issam Laradji, MichaelFriedlander, and Hoyt Koepke. Coordinate descentconverges faster with the gauss-southwell rule thanrandom selection. In International Conference on

Machine Learning, pages 1632–1641, 2015.

Stanley Osher, Martin Burger, Donald Goldfarb, JinjunXu, and Wotao Yin. An iterative regularizationmethod for total variation-based image restoration.Multiscale Modeling and Simulation, 4(2):460–489,2005.

Stanley Osher, Yu Mao, Bin Dong, and Wotao Yin.Fast linearized bregman iteration for compressivesensing and sparse denoising. Communications in

Mathematical Sciences, 8(1):93–111, 2010.

Manfred Padberg. The Boolean quadric polytope: somecharacteristics, facets and relatives. Mathematical

Programming, 45(1-3):139–172, 1989.

Pradeep Ravikumar, Alekh Agarwal, and Martin JWainwright. Message-passing for graph-structuredlinear programs: Proximal methods and roundingschemes. Journal of Machine Learning Research, 11(Mar):1043–1080, 2010.

Bogdan Savchynskyy, Jorg Kappes, Stefan Schmidt,and Christoph Schnorr. A study of nesterov’s schemefor lagrangian decomposition and map labeling. InCVPR 2011, pages 1817–1823. IEEE, 2011.

Bogdan Savchynskyy, Stefan Schmidt, Jorg Kappes,and Christoph Schnorr. E�cient mrf energy min-imization via adaptive diminishing smoothing. InProceedings of the Twenty-Eighth Conference on Un-

certainty in Artificial Intelligence, pages 746–755,2012.

Thomas Schiex, Helene Fargier, and Gerard Verfail-lie. Valued constraint satisfaction problems: hardand easy problems. In Proceedings of the 14th Inter-

national Joint Conference on Artificial Intelligence,pages 631–637. Morgan Kaufmann Publishers Inc.,1995.

Alex Schwing, Tamir Hazan, Marc Pollefeys, andRaquel Urtasun. Globally convergent dual MAPLP relaxation solvers using Fenchel-Young margins.In Advances in Neural Information Processing Sys-

tems, pages 2384–2392, 2012.

Alexander Schwing, Tamir Hazan, Marc Pollefeys, andRaquel Urtasun. Globally convergent parallel MAPLP relaxation solver using the Frank-Wolfe algo-rithm. In International Conference on Machine

Learning, pages 487–495, 2014.

Hanif Sherali and Warren Adams. A hierarchy of re-laxations between the continuous and convex hullrepresentations for zero-one programming problems.SIAM Journal on Discrete Mathematics, 3(3):411–430, 1990.

David Sontag, Amir Globerson, and Tommi Jaakkola.Introduction to dual composition for inference. InOptimization for Machine Learning. MIT Press, 2011.

David A. Sontag. Approximate inference in graphi-

cal models using LP relaxations. PhD thesis, Mas-sachusetts Institute of Technology, 2010.

Siddharth Tourani, Alexander Shekhovtsov, CarstenRother, and Bogdan Savchynskyy. Mplp++: Fast,parallel dual block-coordinate ascent for dense graph-ical models. In Proceedings of the European Confer-

ence on Computer Vision (ECCV), pages 251–267,2018.

Martin Wainwright and Michael Jordan. Graphicalmodels, exponential families, and variational infer-ence. Foundations and Trends in Machine Learning,1(1–2):1–305, 2008.

Page 11: Convergence Rates of Smooth Message Passing with Rounding

Jonathan N. Lee⇤, Aldo Pacchiano⇤, Michael I. Jordan

Martin J Wainwright, Tommi S Jaakkola, and Alan SWillsky. Map estimation via agreement on trees:message-passing and linear programming. IEEE

transactions on information theory, 51(11):3697–3717, 2005.

Jonathan Weed. An explicit analysis of the entropicpenalty in linear programming. In Conference On

Learning Theory, pages 1841–1855, 2018.

Yair Weiss, Chen Yanover, and Talya Meltzer. Mapestimation, linear programming and belief propaga-tion with convex free energies. In Proceedings of the

Twenty-Third Conference on Uncertainty in Artifi-

cial Intelligence, pages 416–425, 2007.

Adrian Weller, Mark Rowland, and David Sontag.Tightness of LP relaxations for almost balanced mod-els. In Artificial Intelligence and Statistics, pages47–55, 2016.

Tomas Werner. A linear programming approach tomax-sum problem: A review. IEEE transactions

on pattern analysis and machine intelligence, 29(7):1165–1179, 2007.

Chen Yanover, Talya Meltzer, and Yair Weiss. Linearprogramming relaxations and belief propagation–anempirical study. Journal of Machine Learning Re-

search, 7(Sep):1887–1907, 2006.

Jonathan Yedidia, William Freeman, and Yair Weiss.Understanding belief propagation and its generaliza-tions. Exploring Artificial Intelligence in the New

Millennium, 8:236–239, 2003.