Min-Max Problems on Factor-Graphs
←
→
Page content transcription
If your browser does not render page correctly, please read the page content below
Min-Max Problems on Factor-Graphs Siamak Ravanbakhsh∗ MRAVANBA @ UALBERTA . CA Christopher Srinivasa‡ CHRIS @ PSI . UTORONTO . CA Brendan Frey‡ FREY @ PSI . UTORONTO . CA Russell Greiner∗ RGREINER @ UALBERTA . CA ∗ Computing Science Dept., University of Alberta, Edmonton, AB T6G 2E8 Canada ‡ PSI Group, University of Toronto, ON M5S 3G4 Canada Abstract from a uniform distribution over satisfying assignments, while (b) counting and integration usually correspond to We study the min-max problem in factor graphs, estimation of the partition function and (c) maximization which seeks the assignment that minimizes the corresponds to Maximum a Posteriori (MAP) inference. maximum value over all factors. We reduce this Here we introduce and study a new class of inference over problem to both min-sum and sum-product infer- graphical models –i.e., (d) the min-max inference problem, ence, and focus on the later. In this approach where the objective is to find an assignment to minimize the min-max inference problem is reduced to the maximum value over a set of functions. a sequence of Constraint Satisfaction Problems (CSP), which allows us to solve the problem by The min-max objective appears in various fields, partic- sampling from a uniform distribution over the set ularly in building robust models under uncertain and ad- of solutions. We demonstrate how this scheme versarial settings. In the context of probabilistic graph- provides a message passing solution to several ical models, several different min-max objectives have NP-hard combinatorial problems, such as min- been previously studied (e.g., Kearns et al. 2001;Ibrahimi max clustering (a.k.a. K-clustering), asymmetric et al. 2011). In combinatorial optimization, min-max may K-center clustering problem, K-packing and the refer to the relation between maximization and minimiza- bottleneck traveling salesman problem. Further- tion in dual combinatorial objectives and their correspond- more, we theoretically relate the min-max reduc- ing linear programs (e.g., Schrijver 1983), or it may refer to tions and several NP hard decision problems such min-max settings due to uncertainty in the problem speci- as clique cover, set-cover, maximum clique and fication (e.g., Averbakh 2001;Aissi et al. 2009). Hamiltonian cycle, therefore also providing mes- Our setting is closely related to a third class of min-max sage passing solutions for these problems. Ex- combinatorial problems that are known as bottleneck prob- perimental results suggest that message passing lems. Instances of these problems include bottleneck trav- often provides near optimal min-max solutions eling salesman problem (Parker & Rardin 1984), min- for moderate size instances. max clustering (Gonzalez 1985), k-center problem (Dyer & Frieze 1985;Khuller & Sussmann 2000) and bottleneck assignment problem (Gross 1959). 1. Introduction Edmonds & Fulkerson (1970) introduce a bottleneck In recent years, message passing methods have achieved framework with a duality theorem that relates the min- a remarkable success in solving different classes of opti- max objective in one problem instance to a max-min ob- mization problems, including maximization (e.g., Frey & jective in a dual problem. An intuitive example is the du- Dueck 2007;Bayati et al. 2005), integration (e.g., Huang ality between the min-max cut separating nodes a and b – & Jebara 2009) and constraint satisfaction problems the cut with the minimum of the maximum weight – and (e.g., Mezard et al. 2002).When formulated as a graphi- min-max path between a and b, which is the path with cal model, these problems correspond to different modes the minimum of the maximum weight (Fulkerson 1966). of inference: (a) solving a CSP corresponds to sampling Hochbaum & Shmoys (1986) leverage triangle inequality in metric spaces to find constant factor approximations to Proceedings of the 31 st International Conference on Machine Learning, Beijing, China, 2014. JMLR: W&CP volume 32. Copy- several NP-hard min-max problems under a unified frame- right 2014 by the author(s). work.
Min-Max Problems on Factor-Graphs The common theme in a majority of heuristics for min-max or bottleneck problems is the relation of the min-max ob- jective with a CSP (e.g., Hochbaum & Shmoys 1986; Pan- igrahy & Vishwanathan 1998). In this paper, we establish a similar relation within the context of factor-graphs, by reducing the min-max inference problem on the original factor-graph to that of sampling (i.e., solving a CSP) on (a) (b) the reduced factor-graph. We also consider an alternative Figure 1. Factor-graphs for the bottleneck assignment problem approach where the factor-graph is transformed such that the min-sum objective produces the same optimal result as tion assume the entry Di,j is the time required by worker min-max objective on the original factor-graph. Although i to finish task j. The min-max assignment minimizes the this reduction is theoretically appealing, in its simple form maximum time required by any worker to finish his assign- it suffers from numerical problems and is not further pur- ment. This problem is also known as bottleneck bipartite sued here. matching and belongs to class P (e.g., Garfinkel 1971). Section 2 formalizes min-max problem in probabilistic Here we show two factor-graph representations of this graphical models and provides an inference procedure by problem. reduction to a sequence of CSPs on the factor graph. Sec- Categorical variable representation: Consider a factor- tion 3 reviews Perturbed Belief Propagation equations (Ra- graph with x = {x1 , . . . , xN }, where each variable xi ∈ vanbakhsh & Greiner 2014) and several forms of high- {1, . . . , N } indicates the column of the selected entry in order factors that allow efficient sum-product inference. Fi- row i of D. For example x1 = 5 indicates the fifth col- nally Section 4 uses these factors to build efficient algo- umn of the first row is selected (see Figure 1(a)). De- rithms for several important min-max problems with gen- fine the following factors: (a) local factors f{i} (xi ) = eral distance matrices. These applications include prob- Di,xi and (b) pairwise factors f{i,j} (x{i,j} ) = ∞I(xi = lems, such as K-packing, that were not previously studied xj ) − ∞I(xi 6= xj ) that enforce the constraint xi 6= xj . within the context of min-max or bottleneck problems. Here I(.) is the indicator function –i.e., I(T rue) = 1 and I(F alse) = 0. Also by convention ∞ 0 , 0. Note that 2. Factor Graphs and CSP Reductions if xi = xj , f{i,j} (x{i,j} ) = ∞, making this choice un- suitable in the min-max solution. On the other hand with Let x = {x1 , . . . , xn }, where x ∈ X , X1 × . . . × Xn de- xi 6= xj , f{i,j} (x{i,j} ) = −∞ and this factor has no im- note a set of discrete variables. Each factor fI (xI ) : XI → pact on the min-max value. YI ⊂ < is a real valued function with range YI , over a subset of variables –i.e., I ⊆ {1, . . . , n} is a set of indices. Binary variable representation: Consider a factor-graph Given the set of factors F, the min-max objective is where x = [x1−1 , . . . , x1−N , x2−1 . . . , x2−N , . . . , xN −N ] ∈ {0, 1}N ×N indicates whether each entry is selected x∗ = argx min max fI (xI ) (1) xi−j = 1 or not xi−j I∈F P = 0 (Figure 1(b)).PHere the fac- tors fI (xI ) = −∞I( i∈∂I xi = 1) + ∞I( i∈∂I xi 6= 1) This model can be conveniently represented as a bipartite ensures that only one variable in each row and column is graph, known as factor graph (Kschischang & Frey 2001), selected and local factors fi−j (xi−j ) = xi−j Di,j − ∞(1 − that includes two sets of nodes: variable nodes xi , and fac- xi−j ) have any effect only if xi−j = 1. tor nodes fI . A variable node i (note that we will often identify a variable xi with its index “i”) is connected to a The min-max assignment in both of these factor-graphs as factor node I if and only if i ∈ I. We will use ∂ to de- defined in eq. (1) gives a solution to the bottleneck assign- note the neighbors of a variable or factor node in the factor ment problem. graph– that is ∂I = {i s.t. i ∈ I} (which is the set I) and For any y ∈ Y in the range of factor values, we reduce the ∂i = {I s.t. i ∈ I}. S original min-max problem to a CSP using the following Let Y = I YI denote the union over the range of reduction. For any y ∈ Y, µy -reduction of the min-max all factors. The min-max value belongs to this set problem eq. (1), is given by maxI∈F fI (x∗I ) ∈ Y. In fact for any assignment x, 1 Y maxI∈F fI (xI ) ∈ Y. µy (x) , I(fI (xI ) ≤ y) (2) Zy I Example Bottleneck Assignment Problem: given a matrix where D ∈
Min-Max Problems on Factor-Graphs is the normalizing constant and I(.) is the indicator func- Let y (1) ≤ . . . ≤ y (N ) be an ordering of elements in Y, tion.1 This distribution defines a CSP over X , where and let r(fI (xI )) denote the rank of yI = fI (xI ) in this µy (x) > 0 iff x is a satisfying assignment. Moreover, Zy ordering. Define the min-sum reduction of {fI }I∈F as gives the number of satisfying assignments. fI0 (xI ) = 2r(fI (xI )) ∀I ∈ F We will use fIy (xI ) , I(fI (xI ) ≤ y) to refer to reduced factors. The following theorem is the basis of our approach Theorem 2.3 in solving min-max problems. X argx min fI0 (xI ) = argx min max fI (xI ) I 2 ∗ ∗ I Theorem 2.1 Let x denote the min-max solution and y be its corresponding value –i.e., y ∗ = maxI fI (x∗I ). Then where {fI0 }I is the min-sum reduction of {fI }I . µy (x) is satisfiable for all y ≥ y ∗ (in particular µy (x∗ ) > 0) and unsatisfiable for all y < y ∗ . Although this allows us to use min-sum message passing to solve min-max problems, the values in the range of factors This theorem enables us to find a min-max assignment by grow exponentially fast, resulting in numerical problems. solving a sequence of CSPs. Let y (1) ≤ . . . ≤ y (N ) be an ordering of y ∈ Y. Starting from y = y (dN/2e) , if µy 3. Solving CSP-reductions is satisfiable then y ∗ ≤ y. On the other hand, if µy is not satisfiable, y ∗ > y. Using binary search, we need to solve Previously in solving CSP-reductions, we assumed an ideal log(|Y|) CSPs to find the min-max solution. Moreover at CSP solver. However, finding an assignment x such that any time-step during the search, we have both upper and µy (x) > 0 or otherwise showing that no such assignment lower bounds on the optimal solution. That is y < y ∗ ≤ exists is in general NP-hard (Cooper 1990). However, mes- y, where µy is the latest unsatisfiable and µy is the latest sage passing methods have been successfully used to pro- satisfiable reduction. vide state-of-the-art results in solving difficult CSPs. We use Perturbed Belief Propagation (PBP Ravanbakhsh & Example Bottleneck Assignment Problem: Here we define Greiner 2014) for this purpose. By using an incomplete the µy -reduction of the binary-valued factor-graph for this solver (Kautz et al. 2009), we lose the lower-bound y on problem by reducing the constraint factors to f y (xI ) = the optimal min-max solution, as PBP may not find a so- y lution even if the instance is satisfiable. 3 However the P I( i∈∂I xi = 1) and the local factors to f{i−j} (xi−j ) = xi−j I(Di,j ≤ y). The µy -reduction can be seen as defin- following theorem states that, as we increase y from the ing a uniform distribution over the all valid assignments min-max value y ∗ , the number of satisfying assignments (i.e., each row and each column has a single entry) where to µy -reduction increases, making it potentially easier to none of the N selected entries are larger than y. solve. Proposition 3.1 2.1. Reduction to Min-Sum Kabadi & Punnen (2004) introduce a simple method to y1 < y2 ⇒ Zy1 ≤ Zy2 ∀y1 , y2 ∈ < transform instances of bottleneck TSP to TSP. Here we where Zy is defined in eq. (3). show how this results extends to min-max problems over factor-graphs. This means that the sub-optimality of our solution is re- lated to our ability to solve CSP-reductions – that is, as the Lemma 2.2 Any two sets of factors, {fI }I∈F and gap y − y ∗ increases, the µy -reduction potentially becomes {fI0 }I∈F , have identical min-max solution easier to solve. argx min max fI (xI ) = argx min max fI0 (xI ) PBP is a message passing method that interpolates between I I Belief Propagation (BP) and Gibbs Sampling. At each iteration, PBP sends a message from variables to factors if ∀I, J ∈ F, xI ∈ XI , xJ ∈ XJ (νi→I ) and vice versa (νI→i ). The factor to variable mes- sage is given by fI (xI ) < fJ (xJ ) ⇔ fI0 (xI ) < fJ0 (xJ ) X Y νI→i (xi ) ∝ fIy (xi , xI\i ) νj→I (xj ) This lemma simply states that what matters in the min-max xI\i ∈X∂I\i j∈∂I\i solution is the relative ordering in the factor-values. (4) 1 0 3 To always have a well-defined probability, we define 0 , 0. To maintain the lower-bound one should be able to correctly 2 All proofs appear in Appendix A. assert unsatisfiability.
Min-Max Problems on Factor-Graphs where the summation is over all the variables in I except on a large number of variables. The next section shows that for xi . The variable to factor message for PBP is slightly many interesting factors are sparse, which allows efficient different from BP; it is a linear combination of the BP mes- calculation of messages. sage update and an indicator function, defined based on a sample from the current estimate of marginal µ b(xi ): 3.2. High-Order Factors Y νi→I (xi ) ∝ (1 − γ) νJ→i (xi ) + γ I(bxi = xi ) The factor-graph formulation of many interesting min-max J∈∂i\I problems involves sparse high-order factors. In all such (5) factors, we are able to significantly reduce the O(|XI |) time Y complexity of calculating factor-to-variable messages. Ef- bi ∼ µ where x b(xi ) ∝ νJ→i (xi ) (6) ficient message passing over such factors is studied by sev- J∈∂i eral works in the context of sum-product and max-product where for γ = 0 we have BP updates and for γ = 1, we inference (e.g., Potetz & Lee 2008; Gupta et al. 2007;Tar- have Gibbs Sampling. PBP starts at γ = 0 and linearly low et al. 2010;Tarlow et al. 2012). The simplest form of increases γ at each iteration, ending at γ = 1 at its final sparse factor in our formulation is the so-called Potts fac- y iteration. At any iteration, PBP may encounter a contradic- tor, f{i,j} (xi , xj ) = I(xi = xj )φ(xi ). This factor assumes tion where the product of incoming messages to node i is the same domain for all the variables (Xi = Xj ∀i, j) zero for all xi ∈ Xi . This could mean that either the prob- and its tabular form is non-zero only across the diago- lem is unsatisfiable or PBP is not able to find a solution. nal. It is easy to see that this allows the marginaliza- However if it reaches the final iteration, PBP produces a tion of eq. (4) to be performed in O(|Xi |) rather than sample from µy (x), which is a solution to the correspond- O(|Xi | |Xj |). Another factor of similar form is the inverse ing CSP. The number of iterations T is the only parameter y Potts factor, f{i,j} (xi , xj ) = I(xi 6= xj ), which ensures of PBP and increasing T , increases the chance of finding a xi 6= xj . In fact any pair-wise factor that is a constant plus solution (Only downside is time complexity; nb., no chance a band-limited matrix allows O(|Xi |) inference (e.g., see of a false positive.) Section 4.4). 3.1. Computational Complexity In Section 4, we use cardinality factors, where Xi = {0, 1} and the factor is defined based on the number of non-zero y y P PBP’s time complexity per iteration is identical to that of values –i.e., fK (xK ) = fK ( i∈K xi ). The µy -reduction BP– i.e., of the factors we use in the binary representation of the X X ! bottleneck assignment problem is in this category. Gail O (|∂I| |XI |) + (|∂i| |Xi |) (7) et al. (1981) P propose a simple O(|∂I| K) method for y I i fK (xK ) = I( i∈K xi = K). We refer to this factor as K-of-N factor and use P similar algorithms for at-least-K- where the first summation accounts for all factor-to- y of-N fK (xK ) P= I( i∈K xi ≥ K) and at-most-K-of-N variable messages (eq. (4)) 4 and the second one accounts y fK (xK ) = I( i∈K xi ≤ K) factors (see Appendix B). for all variable-to-factor messages (eq. (5)). An alternative for more general forms of high order factors To perform binary search over Y we need to sort Y, which is the clique potential of Potetz & Lee (2008). For large K, requiresPO(|Y| log(|Y |)). However, since |Yi | ≤ |Xi | and more efficient methods evaluate the sum of pairs of vari- |Y| ≤ I |Yi |, the cost of sorting is already contained in ables using auxiliary variables forming a binary tree and the first term of eq. (7), and may be ignored in asymptotic use Fast Fourier Transform to reduce the complexity of K- complexity. of-N factors to O(N log(N )2 ) (see Tarlow et al. (2012) and its references). The only remaining factor is that P of binary search itself, which is O(log(|Y|)) = O(log( I (|XI |))) – i.e., at most logarithmic in the cost of PBP’s iteration (i.e., first term 4. Applications in eq. (7)). Also note that the factors added as constraints Here we introduce the factor-graph formulation for sev- only take two values of ±∞, and have no effect in the cost eral NP-hard min-max problems. Interestingly the CSP- of binary search. reduction for each case is an important NP-hard problem. As this analysis suggests, the dominant cost is that of send- Table 1 shows the relationship between the min-max and ing factor-to-variable messages, where a factor may depend the corresponding CSP and the factor α in the constant fac- 4 tor approximation available for each case. For example, The |∂I| is accounting for the number of messages that are sent from each factor. However if the messages are calculated α = 2 means the results reported by some algorithm is synchronously this factor disappears. Although more expensive, guaranteed to be within factor 2 of the optimal min-max in practice, asynchronous message updates performs better. value yb∗ when the distances satisfy the triangle inequal-
Min-Max Problems on Factor-Graphs Table 1. Min-max combinatorial problems and the corresponding CSP-reductions. The last column reports the best α-approximations when triangle inequality holds. ∗ indicates best possible approximation. min-max problem µy -reduction msg-passing cost α min-max clustering clique cover problem O(N 2 K log(N )) 2∗ (Gonzalez 1985) K-packing max-clique problem O(N 2 K log(N )) N/A (weighted) K-center problem dominating set problem O(N 3 log(N )) or O(N R2 log(N )) min(3, 1 + max i wi mini wi ) (Dyer & Frieze 1985) asymmetric K-center problem set cover problem O(N 3 log(N )) or O(N R2 log(N )) log(N )∗ (Panigrahy & Vishwanathan 1998;Chuzhoy et al. 2005) bottleneck TSP Hamiltonian cycle problem O(N 3 log(N )) 2∗ (Parker & Rardin 1984) bottleneck Asymmetric TSP directed Hamiltonian cycle O(N 3 log(N )) log(n)/ log(log(n)) (An et al. 2010) (a) (b) K-packing (binary) (c) sphere packing (d) K-center Figure 2. The factor-graphs for different applications. Factor-graph (a) is common between min-max clustering, Bottleneck TSP and K-packing (categorical). However the definition of factors is different in each case. ity. This table also includes the complexity of the mes- sage passing procedure (assuming asynchronous message updates) in finding the min-max solution. See Appendix C for details. 4.1. Min-Max Clustering Given a symmetric matrix of pairwise distances D ∈ Figure 3. Min-max clustering of 100 points with varying numbers
Min-Max Problems on Factor-Graphs imized. Here we introduce two different factor-graph for- mulations for this problem. 4.2.1. F IRST F ORMULATION : B INARY VARIABLES 2 1 2 1 0 1 1 2 2 1 2 1 2 1 0 2 1 1 1 2 1 0 0 2 2 1 1 1 0 2 1 0 Let binary variables x = {x1 , . . . , xN } ∈ {0, 1}N , in- 0 0 1 2 0 1 2 0 2 1 2 2 1 0 2 0 0 0 0 2 1 0 1 2 1 0 1 0 2 1 0 1 2 2 0 2 2 2 1 1 2 0 2 2 1 2 1 2 dicate a subset of variables of size K that are selected 0 1 2 1 0 2 0 0 0 0 2 1 0 2 2 2 0 0 0 0 2 1 1 0 0 1 0 2 0 2 0 1 2 0 2 0 2 1 0 1 1 2 0 1 1 1 2 0 as code-words P Use the factor fK (x) = (Figure 2(b)). 0 2 1 1 1 1 1 1 0 1 0 0 0 0 1 2 P 1 0 0 1 0 2 0 0 1 2 0 0 2 2 1 2 ∞I( i xi 6= K) − ∞I( i xi = K) (here K = 1 1 2 0 2 1 1 0 1 2 0 2 2 2 1 1 1 0 2 1 2 1 2 2 2 2 0 1 2 0 1 2 {1, . . . , N }) to ensure this constraint. The µy -reduction Figure 4. (left) Using message passing to choose K = 30 out of of this factor for any −∞ < y < +∞ is a K-of-N factor N = 100 random points in the Euclidean plane to maximize the as defined in Section 3.2. Furthermore, for any two vari- minimum pairwise distance (with T = 500 iterations for PBP). Touching circles show the minimum distance. (right) Example ables xi and xj , define factor fxi ,xj (xi,j ) = −Di,j xi xj − of an optimal ternary code (n = 16, y = 11, K = 12), found ∞(1 − xi xj ). This factor is effective only if both points are using the K-packing factor-graph of Section 4.2.3. Here each of selected as code-words. The use of −Di,j is to convert the K = 12 lines contains one code-word of length 16, and every initial max-min objective to min-max. pair of code-words are different in at least y = 11 digits. 4.2.2. S ECOND F ORMULATION : C ATEGORICAL VARIABLES been used to find better codes, trying either to maximize the number of keywords K or their minimum distance y Define the K-packing factor-graph as follows: Let x = (e.g., see Litsyn et al. 1999 and its references). Let {x1 , . . . , xK } be the set of variables where xi ∈ Xi = x = {x1−1 , . . . , x1−n , x2−1 , . . . , x2−n , . . . , xK−n } be the {1, . . . , N } (Figure 2(a)). For every two distinct points set of our binary variables, where xi = {xi−1 , . . . , xi−n } 1 ≤ i < j ≤ K define the factor f{i,j} (xi , xj ) = represents the ith binary vector or code-word. Additionally −Dxi ,xj I(xi 6= xj ) + ∞I(xi = xj ). Here each variable for each 1 ≤ i < j ≤ K, define an auxiliary binary vector represents a code-word and the last term of each factor en- zi,j = {zi,j,1 , . . . , zi,j,n } of length n (Figure 2(c)). sures that code-words are distinct. For each distinct pair of binary vectors xi and xj , and a Proposition 4.2 The µy -reduction of the K-packing particular bit 1 ≤ k ≤ n, the auxiliary variable zi,j,k = 1 factor-graph for the distance matrix D ∈
Min-Max Problems on Factor-Graphs (a) K-center (Euclidean) (b) Facility Location (c) BTSP (Euclidean) (d) BTSP (Random) Figure 5. (a) K-center clustering of 50 random points in a 2D plane with various numbers of clusters (x-axis). The y-axis is the ratio of the min-max value obtained by message passing (T = 500 for PBP) over the min-max value of 2-approximation of Dyer & Frieze (1985). (b) Min-max K-facility location formulated as an asymmetric K-center problem and solved using message passing. Squares indicate 20 potential facility locations and circles indicate 50 customers. The task is to select 5 facilities (red squares) to minimize the maximum distance from any customer to a facility. The radius of circles is the min-max value. (c,d) The min-max solution for Bottleneck TSP with different number of cities (x-axis) for 2D Euclidean space as well as asymmetric random distance matrices (T = 5000 for PBP). The error-bars in all figures show one standard deviation over 10 random instances. ∀j, i = 6 j define f (x{j−j,i−j} ) = ∞I(xj−j = 0 ∧ Table 2. Some optimal binary codes from Litsyn et al. 1999 re- xi−j = 1) − ∞I(xj−j = 1 ∨ xi−j = 0). covered by K-packing factor-graph in the order of increasing y. n is the length of the code, K is the number of code-words and y is the minimum distance between code-words. D. K-of-N factor: only K nodes are selected as centers. P K = {i − i | 1 ≤ i ≤ NP}, define fK (xK ) = n K y n K y n K y n K y Letting 8 4 5 11 4 7 14 4 9 16 6 9 17 4 11 19 6 11 20 8 11 20 4 13 ∞I( i−i∈K xi−i 6= K) − ∞I( i−i∈K xi−i = K). 23 6 13 24 8 13 23 4 15 26 6 15 27 8 15 28 10 15 28 5 16 26 4 17 29 6 17 29 4 19 33 6 19 34 8 19 For variants of this problem such as the capacitated K- 36 12 19 32 4 21 36 6 21 38 8 21 center, additional constraints on the maximum/minimum 39 10 21 35 4 23 39 6 23 41 8 23 39 4 25 43 6 23 46 10 25 47 12 25 points in each group may be added as the at-least/at-most 41 4 27 46 6 27 48 8 27 50 10 27 K-of-N factors. 44 4 29 49 6 29 52 8 29 53 10 29 We can significantly reduce the number of variables and the complexity (which is O((N 3 log(N ))) by bounding the partition such that the maximum distance from any node distance to the center of the cluster y. Given an upper to the center of its partition is minimized. This problem bound y, we may remove all the variables xi−j for Di,j > is known to be NP-hard, even for Euclidean distance ma- y from the factor-graph. Assuming that at most R nodes are trices (Masuyama et al. 1981). Frey & Dueck (2007) use at distance Di−j ≤ y from every node j, the complexity of max-product message passing to solve the min-sum vari- min-max inference drops to O(N R2 log(N )). ation of this problem –a.k.a. K-median problem. A bi- nary variable factor-graph for the same problem is intro- Figure 5(a) compares the performance of message-passing duced in Givoni & Frey (2009). Here we introduce a bi- and the 2-approximation of Dyer & Frieze (1985) when nary variable model for the asymmetric K-center problem. triangle inequality holds. The min-max facility location Let x = {x1−1 , . . . , x1−N , x2−1 , . . . , x2−N , . . . , xN −N } problem can also be formulated as an asymmetric K-center denote N 2 binary variables, where xi−j = 1 indicates that problem where the distance to all customers is ∞ and the point i participates in the partition that has j as its center. distance from a facility to another facility is −∞ (Fig- Now define the following factors: ure 5(b)). The following proposition establishes the relation between A. local factors: ∀i 6= j f{i−j} (xi−j ) = Di,j xi−j − the K-center factor-graph above and dominating set prob- ∞(1 − xi−j ). lem as its CSP reduction. The K-dominating set of graph B. uniqueness factors: every point is associated with ex- G = (V, E) is a subset of nodes D ⊆ V of size |D| = K actly one center (which can be itself). For every i con- such that any node in V \ D is adjacent to at least one mem- siderPI = {i − j | 1 ≤ j ≤ N }P and define fI (xI ) = ber of D – i.e., ∀i ∈ V \ D ∃j ∈ D s.t. (i, j) ∈ E. ∞I( i−j∈∂I xi−j 6= 1) − ∞I( i−j∈∂I xi−j = 1). C. consistency factors: when j is selected as a center by Proposition 4.3 For symmetric distance matrix D ∈ any node i, node j also recognizes itself as a center.
Min-Max Problems on Factor-Graphs ∞ −∞ · · · −∞ −∞ Dj,i Di,j above, is non-zero (i.e., µy (x) > 0) iff x defines a K- Dj,i ∞Di,j · · · −∞ −∞ −∞ dominating set for G(D, y). −∞ ∞ ··· Dj,i −∞ −∞ −∞ .. .. .. .. .. .. .. . . .. . . . Note that in this proposition (in contrast with Proposi- −∞ −∞ −∞ · · · ∞ Di,j −∞ tions 4.1 and 4.2) the relation between the assignments x −∞ −∞ −∞ · · · Dj,i ∞ Di,j Di,j −∞ −∞ · · · −∞ Dj,i ∞ and K-dominating sets of G(D, y) is not one-to-one as sev- eral assignments may correspond to the same dominating Figure 6. The tabular form of f{i,j} (xi , xj ) used for the bottle- set. Here we establish a similar relation between asymmet- neck TSP. ric K-center factor-graph and set-cover problem. Given universe set V and a set S = Proposition 4.5 For any distance matrix D ∈
Min-Max Problems on Factor-Graphs Averbakh, Igor. On the complexity of a class of combinatorial Kabadi, Santosh N and Punnen, Abraham P. The bottleneck tsp. optimization problems with uncertainty. Mathematical Pro- In The Traveling Salesman Problem and Its Variations, pp. gramming, 90(2):263–272, 2001. 697–735. Springer, 2004. Bayati, Mohsen, Shah, Devavrat, and Sharma, Mayank. Maxi- Kautz, Henry A, Sabharwal, Ashish, and Selman, Bart. Incom- mum weight matching via max-product belief propagation. In plete algorithms., 2009. ISIT, pp. 1763–1767. IEEE, 2005. Kearns, Michael, Littman, Michael L, and Singh, Satinder. Chuzhoy, Julia, Guha, Sudipto, Halperin, Eran, Khanna, Sanjeev, Graphical models for game theory. In Proceedings of the Sev- Kortsarz, Guy, Krauthgamer, Robert, and Naor, Joseph Seffi. enteenth conference on UAI, pp. 253–260. Morgan Kaufmann Asymmetric k-center is log* n-hard to approximate. Journal Publishers Inc., 2001. of the ACM (JACM), 52(4):538–551, 2005. Khuller, Samir and Sussmann, Yoram J. The capacitated k-center Cooper, Gregory F. The computational complexity of probabilis- problem. SIAM Journal on Discrete Mathematics, 13(3):403– tic inference using bayesian belief networks. Artificial intelli- 418, 2000. gence, 42(2):393–405, 1990. Kschischang, FR and Frey, BJ. Factor graphs and the sum-product Dyer, Martin E and Frieze, Alan M. A simple heuristic for the algorithm. Information Theory, IEEE, 2001. p-centre problem. Operations Research Letters, 3(6):285–288, 1985. Litsyn, Simon, M., Rains E., and A., Sloane N. J. The table of nonlinear binary codes. http://www.eng.tau.ac. Edmonds, Jack and Fulkerson, Delbert Ray. Bottleneck extrema. il/˜litsyn/tableand/, 1999. [Online; accessed Jan 30 Journal of Combinatorial Theory, 8(3):299–306, 1970. 2014]. Frey, Brendan J and Dueck, Delbert. Clustering by passing mes- Masuyama, Shigeru, Ibaraki, Toshihide, and Hasegawa, Toshi- sages between data points. science, 315(5814):972–976, 2007. haru. The computational complexity of the m-center problems on the plane. IEICE TRANSACTIONS (1976-1990), 64(2):57– Fulkerson, DR. Flow networks and combinatorial operations re- 64, 1981. search. The American Mathematical Monthly, 73(2):115–138, 1966. Mezard, M., Parisi, G., and Zecchina, R. Analytic and algorithmic solution of random satisfiability problems. Science, 297(5582): Gail, Mitchell H, Lubin, Jay H, and Rubinstein, Lawrence V. 812–815, August 2002. ISSN 0036-8075, 1095-9203. doi: Likelihood calculations for matched case-control studies and 10.1126/science.1073287. survival studies with tied death times. Biometrika, 68(3):703– 707, 1981. Panigrahy, Rina and Vishwanathan, Sundar. An o(log*n) approx- imation algorithm for the asymmetric p-center problem. Jour- Garfinkel, Robert S. An improved algorithm for the bottleneck nal of Algorithms, 27(2):259–268, 1998. assignment problem. Operations Research, pp. 1747–1751, 1971. Parker, R Gary and Rardin, Ronald L. Guaranteed performance heuristics for the bottleneck travelling salesman problem. Op- Givoni, Inmar E and Frey, Brendan J. A binary variable model for erations Research Letters, 2(6):269–272, 1984. affinity propagation. Neural computation, 21(6):1589–1600, 2009. Potetz, Brian and Lee, Tai Sing. Efficient belief propagation for higher-order cliques using linear constraint nodes. Computer Gonzalez, Teofilo F. Clustering to minimize the maximum inter- Vision and Image Understanding, 112(1):39–54, 2008. cluster distance. Theoretical Computer Science, 38:293–306, 1985. Ramezanpour, Abolfazl and Zecchina, Riccardo. Cavity approach to sphere packing in hamming space. Physical Review E, 85(2): Gross, O. The bottleneck assignment problem. Technical report, 021106, 2012. DTIC Document, 1959. Ravanbakhsh, Siamak and Greiner, Russell. Perturbed message Gupta, Rahul, Diwan, Ajit A, and Sarawagi, Sunita. Efficient passing for constraint satisfaction problems. arXiv preprint inference with cardinality-based clique potentials. In Proceed- arXiv:1401.6686, 2014. ings of the 24th international conference on Machine learning, pp. 329–336. ACM, 2007. Schrijver, Alexander. Min-max results in combinatorial optimiza- tion. Springer, 1983. Hochbaum, Dorit S and Shmoys, David B. A unified approach to approximation algorithms for bottleneck problems. Journal of Tarlow, Daniel, Givoni, Inmar E, and Zemel, Richard S. Hop- the ACM (JACM), 33(3):533–550, 1986. map: Efficient message passing with high order potentials. In International Conference on Artificial Intelligence and Statis- Huang, Bert and Jebara, Tony. Approximating the permanent with tics, pp. 812–819, 2010. belief propagation. arXiv preprint arXiv:0908.1769, 2009. Tarlow, Daniel, Swersky, Kevin, Zemel, Richard S, Adams, Ibrahimi, Morteza, Javanmard, Adel, Kanoria, Yashodhan, and Ryan Prescott, and Frey, Brendan J. Fast exact inference for Montanari, Andrea. Robust max-product belief propagation. recursive cardinality models. arXiv preprint arXiv:1210.4899, In Signals, Systems and Computers (ASILOMAR), 2011 Con- 2012. ference Record of the Forty Fifth Asilomar Conference on, pp. 43–49. IEEE, 2011.
You can also read