SlideShare a Scribd company logo
Biased Normalized Cuts

                         Subhransu Maji1 , Nisheeth K. Vishnoi2, and Jitendra Malik1
                                    University of California, Berkeley1
                                   Microsoft Research India, Bangalore2
          smaji@eecs.berkeley.edu, nisheeth.vishnoi@gmail.com, malik@eecs.berkeley.edu



                       Abstract                                           (a) Image    (b) Bottom-up (c) Top-down (d) Biased Ncut

   We present a modification of “Normalized Cuts” to in-
corporate priors which can be used for constrained im-
age segmentation. Compared to previous generalizations
of “Normalized Cuts” which incorporate constraints, our
technique has two advantages. First, we seek solutions                Figure 1. Segmentation using bottom-up & top-down information.
which are sufficiently “correlated” with priors which allows
us to use noisy top-down information, for example from an
object detector. Second, given the spectral solution of the           with results as shown in Figure 1(d). The formalism is that
unconstrained problem, the solution of the constrained one            of graph partitioning approached using tools and techniques
can be computed in small additional time, which allows                from spectral graph theory. As in Normalized Cuts [18], we
us to run the algorithm in an interactive mode. We com-               begin by computing eigenvectors of the Normalized Lapla-
pare our algorithm to other graph cut based algorithms and            cian of a graph where the edge weights wij correspond to
highlight the advantages.                                             the “affinity” between two pixels based on low-level simi-
                                                                      larity cues. This is illustrated in Figure 2 for the cat image,
                                                                      where the top row shows the second to fifth eigenvectors,
1. Introduction                                                       ui . Note that none of them is a particularly good indicator
   Consider Figure 1. Suppose that we want to segment the             vector from which the cat can be extracted by thresholding.
cat from its background. We can approach this                            We are interested in cuts which not only minimize the
 1. Bottom-up, we might detect contours corresponding to              normalized cut value but, at the same time, one of the sets
    significant changes in brightness, color or texture. The                                 ¯
                                                                      in the partition (S, S) is required to have a sufficient over-
    output of one such detector is shown in Figure 1(b).              lap with a “bias” or a “prior guess”. In Figure 1, this was
    Note that there will be internal contours, e.g. corre-            the output of a top-down detector. In Figure 2, bottom
    sponding to the colored spots, that are of higher con-            row, the user has specified the bias by indicating various
    trast than the external contour separating the cat from           point sets T in the different panels. We will define a seed-
    its background. This makes it difficult for any bottom-            vector sT associated to T . Let DG be the diagonal matrix
    up segmentation process to separate the cat from its              of the graph, and ui , λi be the eigenvectors and eigenval-
    background in the absence of high-level knowledge of              ues of the Normalized Laplacian. Theorem 3.1 (originally
    what cats look like.                                              due to [15]) shows that we can get cuts that meet the twin
                                                                      goals of having a small normalized cut, and being suffi-
 2. Top-down, we might look for strong activation of, say,            ciently correlated with sT by a remarkably simple proce-
    a sliding window cat detector. The output of one such             dure. Construct the biased normalized cut vector, x⋆ , where
    part-based detector is shown in Figure 1(c). But note                        n
                                                                      x⋆ = c i=2 λi1 ui uT DG sT . Intuitively the eigenvec-
                                                                                        −γ      i
    that this too is inadequate for our purpose. The outputs          tors are linearly combined such that the ones that are well
    of these detectors are typically not sharply localized;           correlated with the seed vector are up weighted, while those
    indeed that is almost a consequence of the desire to              that are inversely correlated have their sign flipped. In Fig-
    make them invariant to small deformations.                        ure 2 bottom row, these vectors are shown as images, and
   In this paper we present an approach, “Biased Normal-              the bias point sets are overlaid. This vector x⋆ can be thresh-
ized Cuts”, where we try to get the best of both worlds,              olded to find the desired “biased normalized cut”. One can

                                                               2057
run the algorithm in an interactive mode by computing the                small number of single-commodity maximum-flows to find
eigenvectors once and generating the biased normalized cut               low-conductance cuts not only inside the input subset T ,
for any bias.                                                                                                                    ¯
                                                                         but among all cuts which are well-correlated with (T, T ).
                                                                         Among spectral methods, local graph partitioning was in-
2. Previous Work                                                         troduced by Spielman and Teng [20], who were interested
                                                                         in finding a low-conductance cut in a graph in time nearly-
   Spectral graph theory originated in the 1970s with work               linear in the volume of the output cut. They used random
such as Donath and Hoffman [10] and Fiedler [13]. By the                 walk based methods to do this; and subsequently this result
1990s, there was a well developed literature in mathematics              was improved by Andersen, Chung and Lang [1] by doing
and theoretical computer science, and Chung [9] provides                 certain Personalized PageRank based random walks.
a monograph level treatment. In computer vision and ma-
chine learning, the work of Shi and Malik on normalized                  3. Biased Graph Partitioning
cuts [18, 19] was very influential, and since then a signifi-
cant body of work has emerged in these fields that we can-                Image as a Weighted Graph and Normalized Cuts. The
not possibly do justice here. The best current approaches                image is represented as a weighted undirected graph G =
to finding contours and regions bottom up are still based on              (V, E) where the nodes of the graph are the points in the
variations of this theme [16].                                           feature space, and an edge is formed between every pair of
   Given our interest in this work of biasing the solution               nodes. The weight function on the edges w : E → R≥0
to satisfy additional constraints, two papers need particular            is a function of the similarity between the end points of the
mention. The problem in the constrained setting for image                edge. The volume of a set of vertices S is the total weight
                                                                                                               def
segmentation has been studied by Yu and Shi [21] where                   of the edges incident to it: vol(S) =        i∈S,j∈V w(i, j).
they find the solution to the normalized cuts problem sub-                        def
                                                                         vol(G) = i,j∈V w(i, j) be the total weight of the edges in
ject to a set of linear constraints of the form U T x = 0. This
                                                                         the graph. Normalized cut measure, defined next, is a stan-
problem can be reduced to an eigenvalue problem which
                                                                         dard way to measure the degree of dissimilarity between
can be solved using spectral techniques as well. Ericks-                                                             ¯
                                                                         two pieces of an image. For S ⊆ V, let S denote V S.
son et al. [11] generalize the set of constraints to U T x = b.
We will see in Section 5 that enforcing constraints in this                          ¯ def
                                                                         Let cut(S, S) = i∈S,j∈S w(i, j) be the weight of edges
                                                                                                    ¯
manner is not robust when the constraints are noisy. The                                      ¯
                                                                         crossing the cut (S, S). Then, the normalized cut value cor-
computational complexity of these approaches is also sig-                                   ¯
                                                                         responding to (S, S) is defined as
nificantly higher than that of solving the basic normalized
                                                                                                   def
                                                                                                                ¯         ¯
                                                                                                         cut(S, S) cut(S, S)
cuts problem.                                                                          Ncut(S) =                  +
   In the last decade the most popular approach to inter-                                                 vol(S)        ¯ .
                                                                                                                    vol(S)
active image segmentation in computer vision and graph-                                            ¯
                                                                         Noting that vol(S) + vol(S) = vol(G), it can be seen that
ics has followed in the steps of the work of Boykov and
                                                                         this notion of a normalized cut is the same as that of the
Jolly [7] who find an image segmentation by computing a
                                                                         traditional graph conductance defined for a set S as
min-cut/max-flow on a graph which encodes both the user
constraints and pairwise pixel similarity. This line of work                                 def
                                                                                                                        ¯
                                                                                                                 cut(S, S)
has been further investigated by Blake and Rother [3] where                             φ(S) = vol(G) ·                    ¯ .
                                                                                                              vol(S) · vol(S)
they experiment with ways to model the foreground and
background regions. In the GrabCut framework [17], the                                                                      def
                                                                         The conductance of the graph G is φ(G) = minS⊆V φ(S)
process of segmentation and foreground/background mod-                   and is an extremely well studied quantity in graph theory,
eling is repeated till convergence. With the advances in the             computer science and machine learning. It is NP-hard to
min-cut/max-flow algorithms like [6], these methods have                  compute exactly and one of the earliest and most popular
become computationally attractive and can often be used in               approach to compute an approximation to it is to write down
an interactive mode. However, these methods fail when the                a relaxation which boils down to computing the second
constraints are sparse making it difficult to construct good              eigenvalue of the Laplacian matrix associated to a graph.
foreground/background models and these methods tend to                   We discuss this next.
produce isolated cuts.
   From a theoretical perspective, there has been a signif-
icant interest in the cut improvement problem: given as                  Graphs and Laplacians. For a graph G = (V, E) with
input a graph and a cut, find a subset that is a better cut.              edge weights function w, let AG ∈ RV ×V denote its ad-
This started with Gallo, Grigoriadis and Tarjan [14] and                 jacency matrix with AG (i, j) = w(i, j); DG denotes the
has attained more recent attention in the work of Ander-                 diagonal degree matrix of G, i.e., DG (i, i) = j∈V w(i, j)
                                                                                                                      def
sen and Lang [2], who gave a general algorithm that uses a               DG (i, j) = 0, for all i = j; LG = DG − AG will

                                                                  2058
Figure 2. Top Row: Input Image and the top 4 eigenvectors computed using the intervening contour cue [16]. Bottom Row: Biased
normalized cuts for various seed sets T . The marked points in each image are the set of points in T .


                                                              def
denote the (combinatorial) Laplacian of G; and LG =                        Biased Normalized Cuts. Now we move on to incorpo-
  −1/2       −1/2                                                          rating the prior information given to us about the image to
DG L G DG         will denote the normalized Laplacian of
G. We will assume G is connected, in which case the                        define the notion of biased normalized cuts. Recall that
eigenvalues of LG are 0 = λ1 < λ2 ≤ · · · ≤ λn .                           our problem is: we are given a region of interest in the
We will denote by λ2 (G) as this second eigenvalue of the                  image and we would like to segment the image so that
normalized Laplacian of G. If u1 , . . . , un are the corre-               the segment is biased towards the specified region. A re-
                                                  def  −1                  gion is modeled as a subset T ⊆ V, of the vertices of the
sponding eigenvectors of LG , then we define vi = DG 2 ui                                                                 ¯
                                                                           image. We would be interested in cuts (S, S) which not
and think of them as the associated eigenvectors of LG ;
                                                                           only minimize the normalized cut value but, at the same
LG vi = λi DG vi . The spectral decomposition of LG can
                                                                           time, have sufficient correlation with the region specified
be written as n λi ui uT . We can also define the Moore-
                i=2       i
                                       def   n
                                                                           by T. To model this, we will first associate a vector sT to
Penrose pseudo-inverse of LG as L+ =
                                 G
                                                 1      T
                                             i=2 λi ui ui .                                                                   ¯
                                                                                                                  vol(T )vol(T )      1
                                                                           the set T as follows: sT (i) =            vol(G)      · vol(T ) , if
                                                                                                                  ¯
                                                                           i ∈ T , and sT (i) =   − vol(T )vol(T ) · vol(T ) if i ∈ T ; or
                                                                                                           vol(G)
                                                                                                                          1
                                                                                                                            ¯
                                                                                                                                        ¯
The Spectral Relaxation to Computing Normalized                                             def                ¯
                                                                                                   vol(T )vol(T )     1T         1T¯
Cuts. Spectral methods approximate the solution to nor-                    equivalently, sT =         vol(G)        vol(T ) − vol(T ) . We
                                                                                                                                    ¯
malized cuts by trying to find a x ∈ RV which minimizes                     have defined it in a way such that i∈V sT (i)di = 0 and
                                                                                        2
xT LG x subject to i,j∈V di dj (xi − xj )2 = vol(G). To                       i∈V sT (i) di = 1.
see why this is a relaxation first, for a subset S of ver-                      This notion of biased normalized cuts is quite natural and
tices, let 1S be the indicator vector of S in RV . Then                    motivated from the theory of local graph partitioning where
                                                  ¯
1T LG 1S = ij∈E wi,j (1S (i) − 1S (j))2 = cut(S, S) and                    the goal is to find a low-conductance cut well correlated
  S
                                 2            ¯ Hence, if
    i,j∈V di dj (1S (i) − 1S (j)) = vol(S)vol(S).                          with a specified input set. The correlation is specified by a
                    vol(G)                                                 parameter κ ∈ (0, 1). This allows us to explore the possi-
we let x =                  ¯ · 1S , then x satisfies the above
                 vol(S)·vol(S)                                             bility of image segments which are well-correlated with the
constraints and has objective value exactly φ(S). Hence, the               prior information which may have much less normalized cut
minimum of the quadratic program above can be at-most the                  value than T itself and, hence, refine the initial guess. In
normalized cut value of G.                                                 particular, we consider the spectral relaxation in Figure 3 to
   In fact, it is easy to see that the optimal value of this               a κ-biased normalized cut around T .
optimization problem, if G is connected, is λ2 (G) and the                     Note that x = sT , is a feasible solution to this spectral
                                      −1
optimal vector to this program is DG 2 u2 , where u2 is the                relaxation. Also note that if v2 satisfies the correlation con-
eigenvector corresponding to the smallest non-zero eigen-                  straint with sT , then it will be the optimal to this program.
value (λ2 (G)) of LG . In particular, the optimal vector x⋆                What is quite interesting is that one can characterize the op-
has the property that i∈V x⋆ di = 0.
                             i                                             timal solution to this spectral program under mild condi-

                                                                    2059
Algorithm 1 Biased Normalized Cuts (G, w, sT , γ)
           minimize                             xT LG x                  Require: Graph G = (V, E), edge weight function w, seed
           s.t.              di dj (xi − xj )2 = vol(G)                      sT and a correlation parameter γ ∈ (−∞, λ2 (G))
                                                                          1: AG (i, j) ← w(i, j), DG (i, i) ←   j w(i, j)
                     i,j∈V
                                                                                                                 −1/2     −1/2
                                                                          2:   LG ← DG − AG , LG ← DG LG DG
                                (      xi sT (i)di )2 ≥ κ
                                                                          3:   Compute u1 , u2 , . . . , uK the eigenvectors of LG corre-
                                 i∈V
Figure 3. BiasedNcut(G, T, κ)- Spectral relaxation to compute                  sponding to the K smallest eigenvalues λ1 , λ2 , . . . , λK .
κ-biased normalized cuts around T
                                                                                      uT DG sT
                                                                          4:   wi ←    i
                                                                                        λi −γ    , for i = 2, . . . , K
                                                                                                                              K
tions on the graph G and the set T and, as it turns out, one              5:   Obtain the biased normalized cut, x∗ ∝         i=2   wi ui
has to do very little effort to compute the optimal solution if
one has already computed the spectrum of LG . This is cap-
tured by the following theorem which is due to [15]. We                  of a second [8]. Our method can be faster than the min-
include a proof of this in the appendix for completeness.                cut/max-flow cut based approaches in an interactive setting
                                                                         as these eigenvectors need to be computed just once. In ad-
Theorem 3.1. Let G be a connected graph and T be such                    dition the real valued solution like the one shown in Figure 2
that i∈V sT (i)v2 (i)di = 0. Further, let 1 ≥ κ ≥ 0 be a                 might provide the user better guidance than a hard segmen-
correlation parameter. Then, there is an optimal solution,               tation produced by a min-cut algorithm.
x⋆ , to the spectral relaxation to the κ-biased normalized
                                    n
cuts around T such that x⋆ = c i=2 λi1 ui uT DG sT for
                                         −γ     i                        5. Constrained Normalized Cuts
some γ ∈ (−∞, λ2 (G)) and a constant c.
                                                                             We compare our approach to the constrained normalized
4. Algorithm                                                             cut formulation of Yu and Shi [21] and Ericksson et al. [11].
                                                                         The later generalized the linear constraints to U T P x = b,
    Theorem 3.1 shows that the final solution is a weighted               where b could be an arbitrary non-zero vector. We consider
combination of the eigenvectors and the weight of each                   a toy example similar to the one used by Yu and Shi. The au-
eigenvector is proportional to the “correlation” with the                thors observe that when the set of constraints are small it is
seed, uT DG sT and inversely proportional to λi − γ. In-
        i                                                                better to use “conditioned constraints”, U T P x = 0, where
tuitively the eigenvectors that are well correlated with the             P = D−1 W , instead of the original constraints, U T x = 0.
seed vector are up weighted, while those that are inversely              The ”conditioned constraints” propagate the constraints to
correlated have their sign flipped.                                       the neighbors of the constrained points avoiding solutions
    Often for images the λi − γ grows quickly and one can                that are too close to the unconstrained one.
obtain a good approximation by considering eigenvectors                      To illustrate our approach, let us consider the points
for the K smallest eigenvalues. In our experiments with                  p1 , p2 , . . . pn grouped into three sets S1 , S2 and S3 from
natural images we set K = 26, i.e. use the top 25 eigen-                 left to right, as shown in Figure 4. We construct a graph
vectors ignoring the all ones vector. We also set γ pa-                  with edge weights between points pi and pj
rameter which controls the amount of correlation implic-
itly to γ = −τ × λavg , where τ = 1 and λavg is the av-                                   wij = exp −d(pi , pj )2 /2σ 2                     (1)
erage of the top K eigenvalues. This could also be user
defined parameter in an interactive setting. Algorithm 1                  with σ = 3, where d(x, y) is the Euclidean distance be-
describes our method for computing the biased normalized                 tween the points x and y. The unconstrained solution to
cuts. Steps 1, 2, 3 are also the steps for computing the seg-            the normalized cuts correctly groups the points as shown in
mentations using normalized cuts which involve computing                 Figure 4.
the K smallest eigenvectors of the normalized graph lapla-                   Now suppose we want to group S1 and S3 together. This
cian LG . The biased normalized cut for any seed vector                  can be done by adding a constraint that the circled points
sT is the weighted combination of eigenvectors where the                 belong to the same group and can encoded as a linear con-
weights are computed in step 4. In the interactive setting               straint of the form, xi −xj = 0, where i and j are the indices
only steps 4 and 5 need to be done when the seed vector                  of the constrained points. The constrained cut formulation
changes which is very quick.                                             is able to correctly separate S1 and S3 from S2 as shown in
    The time taken by the algorithm is dominated by the time             Figure 5(a). For our approach we construct the vector sT
taken to compute the eigenvectors. In an interactive setting             with i, j ∈ T defined earlier and use the top 16 generalized
one can use special purpose hardware accelerated methods                 eigenvectors. The biased cut solution separates S1 and S3
to compute the eigenvectors of typical images in fraction                from S2 as well.

                                                                  2060
Input Points                        Normalized Cut                                Image           Pb              Prior       Biased Ncut
9                                   0.6
8
                                    0.4
7
6                                   0.2

5                                    0
4                                  −0.2
3
                                   −0.4
2
                                   −0.6
1
                                   −0.8
       5    10      15    20              50   100   150   200   250   300

Figure 4. Input points (left) clustered using normalized cuts (right).


   Next we increase the number of such constraints by ran-
domly sampling points from S1 ∪ S3 and adding constraints
that they belong together. Given a set of n points we gen-
erate n − 1 constraints by ordering the points and adding a
constraint for each consecutive pair similar to the approach
of [21]. Instead of improving the solution, the solution to
the constrained cuts deteriorates when the number of con-
straints are large as the solution that separates S1 and S3
from S2 is no longer feasible. Our method on the other hand
gets better with more constraints as seen in Figure 5(b).                           Figure 8. Biased normalized cuts using top-down priors from an
                                                                                    object detector.
   We also test the robustness of the algorithm to outliers
by adding some points from S2 into the constraints. The
constrained cut solution deteriorates even when there is one                        Bias from object detectors. Although in our formulation
outlier as seen in Figure 5(c). Our method on the other hand                        the seed vector sT was discrete, it is not necessary. In par-
remains fairly robust even when there are two outliers Fig-                         ticular we can use the probability estimates produced by an
ure 5(d).                                                                           object detector as a bias. Figure 8 shows some examples
                                                                                    of a top down segmentation mask generated by the detector
6. Qualitative Evaluation                                                           of [4] used as a seed vector directly after normalization.

    We present qualitative results on images taken from                             7. Conclusion
PASCAL VOC 2010 dataset [12]. For all our experiments                                  We present a simple and effective method for comput-
we use the intervening contour cue [16] for computing the                           ing the biased normalized cuts of the images to incorporate
weight matrix and 25 eigenvectors to compute the biased                             top-down priors or user input. The formulation is attractive
normalized cut. We refrain from doing quantitive compari-                           as it allows one to incorporate these priors without addi-
son on segmentation benchmarks as the goal of this paper is                         tional overhead. Linear combinations of eigenvectors nat-
to introduce a new algorithm for computing biased normal-                           urally “regularizes” the combinatorial space of segmenta-
ized cuts, which is only one of the components of a good                            tions. Code for computing biased normalized cuts interac-
segmentation engine.                                                                tively on images can be downloaded at the author’s website.
                                                                                    Acknowledgements: Thanks to Pablo Arbelaez. Subhransu Maji is sup-
Effect of γ. The correlation parameter γ when set smaller                           ported by a fellowship from Google Inc. and ONR MURI N00014-06-1-
values increases the correlation κ with the seed vector. Fig-                       0734. The work was done when the authors were at Microsoft Research
ure 6 shows the effect of γ on the result of the biased nor-                        India.
malized cut. One can obtain tighter cuts around the seed by
setting γ to smaller values. In an interactive setup this could                     A. Appendix
be adjusted by the user.                                                            Proof. [of Theorem 3.1] To start off, note that
                                                                                    BiasedNcut(G, T, κ) is a non-convex program. One
Bias from user interaction. We first test the algorithm                              can relax it to SDPp (G, T, κ) of Figure 9. In the same
in the interactive setting by automatically sampling a set of                       figure the dual of this SDP appears: SDPd (G, T, κ). Here,
points inside the figure mask. Figure 7 shows the several ex-                                                               def
ample outputs. All these images are taken from the animal                           for matrices A, B ∈ Rn×n , A ◦ B =           i,j Ai,j Bi,j .
                                                                                                def           1
categories of the Pascal VOC 2010 detection dataset.                                Also Ln = DG −         vol(G) DG JDG ,   where J is the matrix


                                                                             2061
1 contraint          Biased Ncut                  Yu & Shi                 19 contraints                  Biased Ncut                  Yu & Shi
                                                    0.04
                         0.2                                                                                  2
 8                                                  0.02                       8                                                          0.1
 6                                                     0                       6
                          0                                                                                   0
                                                   −0.02                                                                                   0
 4                                                 −0.04                       4
                        −0.2                                                                              −2
 2                                                 −0.06                       2
                                                                                                                                         −0.1
                        −0.4                       −0.08
      5     10 15 20           100    200    300           100   200   300           5      10 15 20                100    200     300          100   200   300

 (a) One point each sampled from S2 and S3                                     (b) 10 points each sampled from S2 and S3
      10 contraints            Biased Ncut                  Yu & Shi                     11 contraint               Biased Ncut                  Yu & Shi
                          1                                                                                 1
 8                                                  0.05                       8                                                          0.2
                                                                                                          0.5
                                                       0
 6                        0                                                    6                            0                             0.1
                                                   −0.05
 4                                                                             4                         −0.5
                                                    −0.1                                                                                   0
 2                       −1                        −0.15                       2                          −1
                                                                                                                                         −0.1
      5     10 15 20           100    200    300           100   200   300           5     10 15 20                 100    200     300          100   200   300

 (c) 5 points each from sets S1 and S3 with 1 outlier from S2                  (d) 5 points each from sets S1 and S3 with 2 outliers from S2

Figure 5. Comparison of our biased normalized cut approach to constrained cut approach of [21, 11] for various constraints shown by
red circled points. When the number of constraints is large (b) or there are outliers (c, d), our method produces better solutions than the
solution of constrained cuts.




Figure 6. Effect of γ. Input image and biased cuts for decreasing values of γ from left to right. The local cuts are more and more correlated
with the seed set (shown as dots) as γ decreases.


with Ji,j = 1 for all i, j. Verifying Slater’s condition,
one can observe that strong duality holds for this SDP                                            minimize                                  LG ◦ X
relaxation. Then, using strong duality and the comple-                                           subject to                         Ln ◦ X = 1
mentary slackness conditions implied by it, we will argue
that the SDPp (G, T, κ) has a rank one optimal solution                                                           (DG sT )(DG sT )T ◦ X ≥ κ
under the conditions of the theorem. This implies that                                                                                X 0
the optimal solution of SDPp (G, T, κ) is the same as the
optimal solution of BiasedNcut. Combining this with the                            maximize                                               α + κβ
complementary slackness condition obtained from the dual
                                                                                   subject to LG          αLn + β(DG sT )(DG sT )T
SDPd (G, T, κ), one can derive that the optimal rank one
solution has, up to a constant, the desired form promised                                                                        α ∈ R, β ≥ 0
by the theorem. Now we expand the above steps in claims.                      Figure 9. Top: SDPp (G, T, κ) Primal SDP relaxation. Bottom:
                                                                              SDPd (G, T, κ) Dual SDP
Claim A.1. The primal SDPp (G, T, κ) is a relaxation of
the vector program BiasedNcut(G, T, κ).
                                                                              for this primal SDP. Consider X = sT sT T . Then,
Proof. Consider a vector x that is a feasible solution to                     (DG sT )(DG sT )T ◦ sT sT T = (sT T DG sT )2 = 1 > κ.
BiasedNcut(G, T, κ), and note that X = xxT is a feasible
solution to SDPp (G, T, κ).                                                   Claim A.3. The feasibility and complementary slackness
                                                                              conditions for a primal-dual pair X ⋆ , α⋆ , β ⋆ listed in Fig-
Claim A.2. Strong duality holds between SDPp (G, T, κ)
                                                                              ure 10 are sufficient for them to be an optimal solution.
and SDPd (G, T, κ).

Proof. Since SDPp (G, T, κ) is convex, if suffices to ver-                     Proof. This follows from the convexity of SDPp (G, T, κ)
ify that Slater’s constraint qualification condition is true                   and Slater’s condition [5].


                                                                       2062
Image                   Pb               Biased Ncut               Image                   Pb               Biased Ncut




Figure 7. Example biased cuts. Images shown with the “probability of boundary” map (Pb) computed using [16] and the biased normalized
cut for various seed sets (marked as red dots).



                                                                2063
References
                                   Ln ◦ X ⋆      =    1
                                                                         [1] R. Andersen, F. Chung, and K. Lang. Local graph partition-
                     (DG sT )(DG sT )T ◦ X ⋆     ≥    κ                      ing using PageRank vectors. In FOCS, pages 475–486, 2006.
        LG − α⋆ Ln − β ⋆ (DG sT )(DG sT )T            0                      2058
                                                                         [2] R. Andersen and K. Lang. An algorithm for improving graph
                                        β⋆       ≥    0
                                                                             partitions. In SODA, pages 651–660, 2008. 2058
                                                                         [3] A. Blake, C. Rother, M. Brown, P. Perez, and P. Torr. Inter-
                                α⋆ (Ln ◦ X ⋆ − 1) = 0                        active image segmentation using an adaptive gmmrf model.
               β ⋆ ((DG sT )(DG sT )T ◦ X ⋆ − κ)     = 0                     In ECCV, pages 428–441, 2004. 2058
                                                                         [4] L. Bourdev, S. Maji, T. Brox, and J. Malik. Detecting people
     X ⋆ ◦ (LG − α⋆ Ln − β ⋆ (DG sT )(DG sT )T )     = 0
                                                                             using mutually consistent poselet activations. In ECCV, sep
                                                                             2010. 2061
Figure 10. Top: Feasibility conditions. Bottom: Complementary
                                                                         [5] S. Boyd and L. Vandenberghe. Convex Optimization. Cam-
slackness conditions.
                                                                             bridge University Press, Cambridge, UK, 2004. 2062
                                                                         [6] Y. Boykov and V. Kolmogorov. An experimental comparison
Claim A.4. These feasibility and complementary slackness                     of min-cut/max-flow algorithms for energy minimization in
conditions, coupled with the assumptions of the theorem,                     vision. TPAMI, 26:1124–1137, September 2004. 2058
imply that X ⋆ must be rank 1 and β ⋆ > 0.                               [7] Y. Y. Boykov and M. P. Jolly. Interactive graph cuts for opti-
                                                                             mal boundary & region segmentation of objects in N-D im-
Now we complete the proof of the theorem. From Claim                         ages. In ICCV, 2001. 2058
A.4 it follows that, X ⋆ = x⋆ x⋆T where x⋆ satisfies the                  [8] B. Catanzaro, B.-Y. Su, N. Sundaram, Y. Lee, M. Murphy,
equation (LG −α⋆ Ln −β ⋆ (DG sT )(DG sT )T )x⋆ = 0. From                     and K. Keutzer. Efficient, high-quality image contour detec-
the second complementary slackness condition in Figure                       tion. In ICCV, pages 2381 –2388, 2009. 2060
10, and the fact that β ⋆ > 0, we obtain that i x⋆ sT (i)di =            [9] F. Chung. Spectral Graph Theory. American Mathematical
  √                          √                   i
                                                                             Society, 1997. 2058
± κ. Thus, x⋆ = ±β ⋆ κ(LG − α⋆ Ln )+ DG sT . This
                                                                        [10] W. E. Donath and A. J. Hoffman. Lower bounds for the parti-
proves the theorem.
                                                                             tioning of graphs. IBM J. Res. Dev., 17:420–425, September
A.1. Proof of Claim A.4                                                      1973. 2058
                                                                        [11] A. Eriksson, C. Olsson, and F. Kahl. Normalized cuts revis-
Proof. We start by stating two facts. The second is trivial.                 ited: A reformulation for segmentation with linear grouping
                                                                             constraints. In ICCV, oct. 2007. 2058, 2060, 2062
Fact A.5. α⋆ ≤ λ2 (G). Moreover if λ2 = α⋆ then
                                                                        [12] M. Everingham, L. Van Gool, C. K. I. Williams, J. Winn,
  i v2 (i)sT (i)di = 0.                                                      and A. Zisserman. The pascal visual object classes (voc)
                           −1                                                challenge. IJCV, 88(2):303–338, June 2010. 2061
Proof. Recall that v2 = DG 2 u2 where u2 is the unit length             [13] M. Fiedler. Algebraic connectivity of graphs. Czechoslovak
eigenvector corresponding to λ2 (G) of LG . Plugging in v2                   Mathematical Journal, 23(98):298–305, 1973. 2058
in Equation 3 from the feasibility conditions, we obtain that           [14] G. Gallo, M. D. Grigoriadis, and R. E. Tarjan. A fast para-
v2 LG v2 −α⋆ −β ⋆ ( i v2 (i)sT (i)di )2 ≥ 0. But v2 LG v2 =
 T                                                T
                                                                             metric maximum flow algorithm and applications. SIAM J.
             ⋆                            ⋆
λ2 (G) and β ≥ 0. Hence, λ2 (G) ≥ α . It follows that if                     Comput., 18(1):30–55, 1989. 2058
λ2 = α⋆ then i v2 (i)sT (i)di = 0.                                      [15] M. W. Mahoney, L. Orecchia, and N. K. Vishnoi. A
                                                                             spectral algorithm for improving graph partitions. CoRR,
Fact A.6. We may assume that the optimal X ⋆ satisfies                        abs/0912.0681, 2009. 2057, 2060
      1        1
1T DG X ⋆ DG 1 = 0, where 1 is the all ones vector.
      2        2                                                        [16] M. Maire, P. Arbelaez, C. Fowlkes, and J. Malik. Using con-
                                                                             tours to detect and localize junctions in natural images. In
Now we return to the proof of Claim A.4. If we assume
                                              ⋆                              CVPR, pages 1 –8, 2008. 2058, 2059, 2061, 2063
   i v2 (i)sT (i)di = 0, then we know that α < λ2 (G) from              [17] C. Rother, V. Kolmogorov, and A. Blake. Grabcut: Interac-
Fact A.5. Note that since G is connected and α⋆ < λ2 (G),                    tive foreground extraction using iterated graph cuts. ACM
LG − α⋆ Ln has rank exactly n − 1. From the complemen-                       Transactions on Graphics, 23:309–314, 2004. 2058
tary slackness condition 2 we can deduce that the image of              [18] J. Shi and J. Malik. Normalized cuts and image segmenta-
X ⋆ is in the kernel of LG − α⋆ Ln − β ⋆ (DG sT )(DG sT )T .                 tion. In CVPR, jun. 1997. 2057, 2058
But β ⋆ (DG sT )(DG sT )T is a rank one matrix and since                [19] J. Shi and J. Malik. Normalized cuts and image segmenta-
                                                  ⋆
   i sT (i)di = 0, it reduces the rank of LG − α Ln by one                   tion. TPAMI, 22(8):888 –905, Aug. 2000. 2058
                    ⋆         ⋆             ⋆
precisely when β > 0. If β = 0 then X must be 0 which                   [20] D. Spielman and S.-H. Teng. Nearly-linear time algorithms
is not possible if SDPp (G, T, κ) is feasible. Hence, the rank               for graph partitioning, graph sparsification, and solving lin-
of LG −α⋆ Ln −β ⋆ (DG sT )(DG sT )T must be exactly n−2                      ear systems. In STOC, pages 81–90, 2004. 2058
and since X ⋆ cannot have any component along the all ones              [21] S. X. Yu and J. Shi. Grouping with bias. In NIPS. MIT Press,
                                                                             2001. 2058, 2060, 2061, 2062
vector, X ⋆ must be rank one. This proves the claim.

                                                                 2064

More Related Content

What's hot

Steven Duplij, "Higher regularity, inverse and polyadic semigroups", Preprint...
Steven Duplij, "Higher regularity, inverse and polyadic semigroups", Preprint...Steven Duplij, "Higher regularity, inverse and polyadic semigroups", Preprint...
Steven Duplij, "Higher regularity, inverse and polyadic semigroups", Preprint...
Steven Duplij (Stepan Douplii)
 
11.final paper -0047www.iiste.org call-for_paper-58
11.final paper -0047www.iiste.org call-for_paper-5811.final paper -0047www.iiste.org call-for_paper-58
11.final paper -0047www.iiste.org call-for_paper-58
Alexander Decker
 
Cluster
ClusterCluster
Steven Duplij, "Higher regularity, inverse and polyadic semigroups", Preprint...
Steven Duplij, "Higher regularity, inverse and polyadic semigroups", Preprint...Steven Duplij, "Higher regularity, inverse and polyadic semigroups", Preprint...
Steven Duplij, "Higher regularity, inverse and polyadic semigroups", Preprint...
Steven Duplij (Stepan Douplii)
 
Higher dimensional image analysis using brunn minkowski theorem, convexity an...
Higher dimensional image analysis using brunn minkowski theorem, convexity an...Higher dimensional image analysis using brunn minkowski theorem, convexity an...
Higher dimensional image analysis using brunn minkowski theorem, convexity an...
Alexander Decker
 
Ireducible core and equal remaining obligations rule for mcst games
Ireducible core and equal remaining obligations rule for mcst gamesIreducible core and equal remaining obligations rule for mcst games
Ireducible core and equal remaining obligations rule for mcst games
vinnief
 
Steven Duplij, "Higher regularity, inverse and polyadic semigroups", Universe...
Steven Duplij, "Higher regularity, inverse and polyadic semigroups", Universe...Steven Duplij, "Higher regularity, inverse and polyadic semigroups", Universe...
Steven Duplij, "Higher regularity, inverse and polyadic semigroups", Universe...
Steven Duplij (Stepan Douplii)
 
Km2417821785
Km2417821785Km2417821785
Km2417821785
IJERA Editor
 
Common fixed point theorems in polish space for nonself mapping
Common fixed point theorems in polish space for nonself mappingCommon fixed point theorems in polish space for nonself mapping
Common fixed point theorems in polish space for nonself mapping
Alexander Decker
 
Proportional and decentralized rule mcst games
Proportional and decentralized rule mcst gamesProportional and decentralized rule mcst games
Proportional and decentralized rule mcst games
vinnief
 
3D gravity inversion by planting anomalous densities
3D gravity inversion by planting anomalous densities3D gravity inversion by planting anomalous densities
3D gravity inversion by planting anomalous densities
Leonardo Uieda
 
project final
project finalproject final
project final
Rian Rustvold
 
HW1 MIT Fall 2005
HW1 MIT Fall 2005HW1 MIT Fall 2005
HW1 MIT Fall 2005
Aditya Kapoor
 
EM Theory Term Presentation Keynotes Version
EM Theory Term Presentation Keynotes VersionEM Theory Term Presentation Keynotes Version
EM Theory Term Presentation Keynotes Version
shelling ford
 
Astaño
AstañoAstaño
Astaño
Cire Oreja
 
4515ijci01
4515ijci014515ijci01
4515ijci01
IJCI JOURNAL
 
Baum3
Baum3Baum3
Baum3
dmolina87
 
Math Conference Poster
Math Conference PosterMath Conference Poster
Math Conference Poster
mary41679
 
Polyadic integer numbers and finite (m,n) fields
Polyadic integer numbers and finite (m,n) fieldsPolyadic integer numbers and finite (m,n) fields
Polyadic integer numbers and finite (m,n) fields
Steven Duplij (Stepan Douplii)
 
Bp31457463
Bp31457463Bp31457463
Bp31457463
IJERA Editor
 

What's hot (20)

Steven Duplij, "Higher regularity, inverse and polyadic semigroups", Preprint...
Steven Duplij, "Higher regularity, inverse and polyadic semigroups", Preprint...Steven Duplij, "Higher regularity, inverse and polyadic semigroups", Preprint...
Steven Duplij, "Higher regularity, inverse and polyadic semigroups", Preprint...
 
11.final paper -0047www.iiste.org call-for_paper-58
11.final paper -0047www.iiste.org call-for_paper-5811.final paper -0047www.iiste.org call-for_paper-58
11.final paper -0047www.iiste.org call-for_paper-58
 
Cluster
ClusterCluster
Cluster
 
Steven Duplij, "Higher regularity, inverse and polyadic semigroups", Preprint...
Steven Duplij, "Higher regularity, inverse and polyadic semigroups", Preprint...Steven Duplij, "Higher regularity, inverse and polyadic semigroups", Preprint...
Steven Duplij, "Higher regularity, inverse and polyadic semigroups", Preprint...
 
Higher dimensional image analysis using brunn minkowski theorem, convexity an...
Higher dimensional image analysis using brunn minkowski theorem, convexity an...Higher dimensional image analysis using brunn minkowski theorem, convexity an...
Higher dimensional image analysis using brunn minkowski theorem, convexity an...
 
Ireducible core and equal remaining obligations rule for mcst games
Ireducible core and equal remaining obligations rule for mcst gamesIreducible core and equal remaining obligations rule for mcst games
Ireducible core and equal remaining obligations rule for mcst games
 
Steven Duplij, "Higher regularity, inverse and polyadic semigroups", Universe...
Steven Duplij, "Higher regularity, inverse and polyadic semigroups", Universe...Steven Duplij, "Higher regularity, inverse and polyadic semigroups", Universe...
Steven Duplij, "Higher regularity, inverse and polyadic semigroups", Universe...
 
Km2417821785
Km2417821785Km2417821785
Km2417821785
 
Common fixed point theorems in polish space for nonself mapping
Common fixed point theorems in polish space for nonself mappingCommon fixed point theorems in polish space for nonself mapping
Common fixed point theorems in polish space for nonself mapping
 
Proportional and decentralized rule mcst games
Proportional and decentralized rule mcst gamesProportional and decentralized rule mcst games
Proportional and decentralized rule mcst games
 
3D gravity inversion by planting anomalous densities
3D gravity inversion by planting anomalous densities3D gravity inversion by planting anomalous densities
3D gravity inversion by planting anomalous densities
 
project final
project finalproject final
project final
 
HW1 MIT Fall 2005
HW1 MIT Fall 2005HW1 MIT Fall 2005
HW1 MIT Fall 2005
 
EM Theory Term Presentation Keynotes Version
EM Theory Term Presentation Keynotes VersionEM Theory Term Presentation Keynotes Version
EM Theory Term Presentation Keynotes Version
 
Astaño
AstañoAstaño
Astaño
 
4515ijci01
4515ijci014515ijci01
4515ijci01
 
Baum3
Baum3Baum3
Baum3
 
Math Conference Poster
Math Conference PosterMath Conference Poster
Math Conference Poster
 
Polyadic integer numbers and finite (m,n) fields
Polyadic integer numbers and finite (m,n) fieldsPolyadic integer numbers and finite (m,n) fields
Polyadic integer numbers and finite (m,n) fields
 
Bp31457463
Bp31457463Bp31457463
Bp31457463
 

Similar to Biased normalized cuts

Interactive graph cuts for optimal boundary and region segmentation of objects
Interactive graph cuts for optimal boundary and region segmentation of objects Interactive graph cuts for optimal boundary and region segmentation of objects
Interactive graph cuts for optimal boundary and region segmentation of objects
yzxvvv
 
26 spanning
26 spanning26 spanning
26 spanning
Aŕŷa Ķĥaŕel
 
topological_quantum_computing
topological_quantum_computingtopological_quantum_computing
topological_quantum_computing
Zachary Forster
 
Unit II & III_uncovered topics.doc notes
Unit II & III_uncovered topics.doc notesUnit II & III_uncovered topics.doc notes
Unit II & III_uncovered topics.doc notes
smithashetty24
 
A CONVERGENCE ANALYSIS OF GRADIENT_version1
A CONVERGENCE ANALYSIS OF GRADIENT_version1A CONVERGENCE ANALYSIS OF GRADIENT_version1
A CONVERGENCE ANALYSIS OF GRADIENT_version1
thanhdowork
 
project report(1)
project report(1)project report(1)
project report(1)
Conor Bradley
 
Illustration Clamor Echelon Evaluation via Prime Piece Psychotherapy
Illustration Clamor Echelon Evaluation via Prime Piece PsychotherapyIllustration Clamor Echelon Evaluation via Prime Piece Psychotherapy
Illustration Clamor Echelon Evaluation via Prime Piece Psychotherapy
IJMER
 
Gravitational Based Hierarchical Clustering Algorithm
Gravitational Based Hierarchical Clustering AlgorithmGravitational Based Hierarchical Clustering Algorithm
Gravitational Based Hierarchical Clustering Algorithm
IDES Editor
 
Graph Image Segmentation
Graph Image SegmentationGraph Image Segmentation
Graph Image Segmentation
Manohar Kuse
 
International Journal of Engineering Research and Development (IJERD)
 International Journal of Engineering Research and Development (IJERD) International Journal of Engineering Research and Development (IJERD)
International Journal of Engineering Research and Development (IJERD)
IJERD Editor
 
Biao Hou--SAR IMAGE DESPECKLING BASED ON IMPROVED DIRECTIONLET DOMAIN GAUSSIA...
Biao Hou--SAR IMAGE DESPECKLING BASED ON IMPROVED DIRECTIONLET DOMAIN GAUSSIA...Biao Hou--SAR IMAGE DESPECKLING BASED ON IMPROVED DIRECTIONLET DOMAIN GAUSSIA...
Biao Hou--SAR IMAGE DESPECKLING BASED ON IMPROVED DIRECTIONLET DOMAIN GAUSSIA...
grssieee
 
Biao Hou--SAR IMAGE DESPECKLING BASED ON IMPROVED DIRECTIONLET DOMAIN GAUSSIA...
Biao Hou--SAR IMAGE DESPECKLING BASED ON IMPROVED DIRECTIONLET DOMAIN GAUSSIA...Biao Hou--SAR IMAGE DESPECKLING BASED ON IMPROVED DIRECTIONLET DOMAIN GAUSSIA...
Biao Hou--SAR IMAGE DESPECKLING BASED ON IMPROVED DIRECTIONLET DOMAIN GAUSSIA...
grssieee
 
Kakuro: Solving the Constraint Satisfaction Problem
Kakuro: Solving the Constraint Satisfaction ProblemKakuro: Solving the Constraint Satisfaction Problem
Kakuro: Solving the Constraint Satisfaction Problem
Varad Meru
 
[PPT]
[PPT][PPT]
[PPT]
butest
 
Spectral Clustering Report
Spectral Clustering ReportSpectral Clustering Report
Spectral Clustering Report
Miaolan Xie
 
Image Processing
Image ProcessingImage Processing
Image Processing
Tuyen Pham
 
Gupte - first year paper_approved (1)
Gupte - first year paper_approved (1)Gupte - first year paper_approved (1)
Gupte - first year paper_approved (1)
Shweta Gupte
 
論文紹介:Learning With Neighbor Consistency for Noisy Labels
論文紹介:Learning With Neighbor Consistency for Noisy Labels論文紹介:Learning With Neighbor Consistency for Noisy Labels
論文紹介:Learning With Neighbor Consistency for Noisy Labels
Toru Tamaki
 
Colloquium.pptx
Colloquium.pptxColloquium.pptx
Colloquium.pptx
Mythili680896
 
Lec11: Active Contour and Level Set for Medical Image Segmentation
Lec11: Active Contour and Level Set for Medical Image SegmentationLec11: Active Contour and Level Set for Medical Image Segmentation
Lec11: Active Contour and Level Set for Medical Image Segmentation
Ulaş Bağcı
 

Similar to Biased normalized cuts (20)

Interactive graph cuts for optimal boundary and region segmentation of objects
Interactive graph cuts for optimal boundary and region segmentation of objects Interactive graph cuts for optimal boundary and region segmentation of objects
Interactive graph cuts for optimal boundary and region segmentation of objects
 
26 spanning
26 spanning26 spanning
26 spanning
 
topological_quantum_computing
topological_quantum_computingtopological_quantum_computing
topological_quantum_computing
 
Unit II & III_uncovered topics.doc notes
Unit II & III_uncovered topics.doc notesUnit II & III_uncovered topics.doc notes
Unit II & III_uncovered topics.doc notes
 
A CONVERGENCE ANALYSIS OF GRADIENT_version1
A CONVERGENCE ANALYSIS OF GRADIENT_version1A CONVERGENCE ANALYSIS OF GRADIENT_version1
A CONVERGENCE ANALYSIS OF GRADIENT_version1
 
project report(1)
project report(1)project report(1)
project report(1)
 
Illustration Clamor Echelon Evaluation via Prime Piece Psychotherapy
Illustration Clamor Echelon Evaluation via Prime Piece PsychotherapyIllustration Clamor Echelon Evaluation via Prime Piece Psychotherapy
Illustration Clamor Echelon Evaluation via Prime Piece Psychotherapy
 
Gravitational Based Hierarchical Clustering Algorithm
Gravitational Based Hierarchical Clustering AlgorithmGravitational Based Hierarchical Clustering Algorithm
Gravitational Based Hierarchical Clustering Algorithm
 
Graph Image Segmentation
Graph Image SegmentationGraph Image Segmentation
Graph Image Segmentation
 
International Journal of Engineering Research and Development (IJERD)
 International Journal of Engineering Research and Development (IJERD) International Journal of Engineering Research and Development (IJERD)
International Journal of Engineering Research and Development (IJERD)
 
Biao Hou--SAR IMAGE DESPECKLING BASED ON IMPROVED DIRECTIONLET DOMAIN GAUSSIA...
Biao Hou--SAR IMAGE DESPECKLING BASED ON IMPROVED DIRECTIONLET DOMAIN GAUSSIA...Biao Hou--SAR IMAGE DESPECKLING BASED ON IMPROVED DIRECTIONLET DOMAIN GAUSSIA...
Biao Hou--SAR IMAGE DESPECKLING BASED ON IMPROVED DIRECTIONLET DOMAIN GAUSSIA...
 
Biao Hou--SAR IMAGE DESPECKLING BASED ON IMPROVED DIRECTIONLET DOMAIN GAUSSIA...
Biao Hou--SAR IMAGE DESPECKLING BASED ON IMPROVED DIRECTIONLET DOMAIN GAUSSIA...Biao Hou--SAR IMAGE DESPECKLING BASED ON IMPROVED DIRECTIONLET DOMAIN GAUSSIA...
Biao Hou--SAR IMAGE DESPECKLING BASED ON IMPROVED DIRECTIONLET DOMAIN GAUSSIA...
 
Kakuro: Solving the Constraint Satisfaction Problem
Kakuro: Solving the Constraint Satisfaction ProblemKakuro: Solving the Constraint Satisfaction Problem
Kakuro: Solving the Constraint Satisfaction Problem
 
[PPT]
[PPT][PPT]
[PPT]
 
Spectral Clustering Report
Spectral Clustering ReportSpectral Clustering Report
Spectral Clustering Report
 
Image Processing
Image ProcessingImage Processing
Image Processing
 
Gupte - first year paper_approved (1)
Gupte - first year paper_approved (1)Gupte - first year paper_approved (1)
Gupte - first year paper_approved (1)
 
論文紹介:Learning With Neighbor Consistency for Noisy Labels
論文紹介:Learning With Neighbor Consistency for Noisy Labels論文紹介:Learning With Neighbor Consistency for Noisy Labels
論文紹介:Learning With Neighbor Consistency for Noisy Labels
 
Colloquium.pptx
Colloquium.pptxColloquium.pptx
Colloquium.pptx
 
Lec11: Active Contour and Level Set for Medical Image Segmentation
Lec11: Active Contour and Level Set for Medical Image SegmentationLec11: Active Contour and Level Set for Medical Image Segmentation
Lec11: Active Contour and Level Set for Medical Image Segmentation
 

More from irisshicat

Object segmentation by alignment of poselet activations to image contours
Object segmentation by alignment of poselet activations to image contoursObject segmentation by alignment of poselet activations to image contours
Object segmentation by alignment of poselet activations to image contours
irisshicat
 
A probabilistic model for recursive factorized image features ppt
A probabilistic model for recursive factorized image features pptA probabilistic model for recursive factorized image features ppt
A probabilistic model for recursive factorized image features ppt
irisshicat
 
A probabilistic model for recursive factorized image features
A probabilistic model for recursive factorized image featuresA probabilistic model for recursive factorized image features
A probabilistic model for recursive factorized image features
irisshicat
 
★Mean shift a_robust_approach_to_feature_space_analysis
★Mean shift a_robust_approach_to_feature_space_analysis★Mean shift a_robust_approach_to_feature_space_analysis
★Mean shift a_robust_approach_to_feature_space_analysis
irisshicat
 
Pami meanshift
Pami meanshiftPami meanshift
Pami meanshift
irisshicat
 
Shape matching and object recognition using shape context belongie pami02
Shape matching and object recognition using shape context belongie pami02Shape matching and object recognition using shape context belongie pami02
Shape matching and object recognition using shape context belongie pami02
irisshicat
 
. Color and texture-based image segmentation using the expectation-maximizat...
. Color  and texture-based image segmentation using the expectation-maximizat.... Color  and texture-based image segmentation using the expectation-maximizat...
. Color and texture-based image segmentation using the expectation-maximizat...
irisshicat
 
Shape matching and object recognition using shape contexts
Shape matching and object recognition using shape contextsShape matching and object recognition using shape contexts
Shape matching and object recognition using shape contexts
irisshicat
 

More from irisshicat (8)

Object segmentation by alignment of poselet activations to image contours
Object segmentation by alignment of poselet activations to image contoursObject segmentation by alignment of poselet activations to image contours
Object segmentation by alignment of poselet activations to image contours
 
A probabilistic model for recursive factorized image features ppt
A probabilistic model for recursive factorized image features pptA probabilistic model for recursive factorized image features ppt
A probabilistic model for recursive factorized image features ppt
 
A probabilistic model for recursive factorized image features
A probabilistic model for recursive factorized image featuresA probabilistic model for recursive factorized image features
A probabilistic model for recursive factorized image features
 
★Mean shift a_robust_approach_to_feature_space_analysis
★Mean shift a_robust_approach_to_feature_space_analysis★Mean shift a_robust_approach_to_feature_space_analysis
★Mean shift a_robust_approach_to_feature_space_analysis
 
Pami meanshift
Pami meanshiftPami meanshift
Pami meanshift
 
Shape matching and object recognition using shape context belongie pami02
Shape matching and object recognition using shape context belongie pami02Shape matching and object recognition using shape context belongie pami02
Shape matching and object recognition using shape context belongie pami02
 
. Color and texture-based image segmentation using the expectation-maximizat...
. Color  and texture-based image segmentation using the expectation-maximizat.... Color  and texture-based image segmentation using the expectation-maximizat...
. Color and texture-based image segmentation using the expectation-maximizat...
 
Shape matching and object recognition using shape contexts
Shape matching and object recognition using shape contextsShape matching and object recognition using shape contexts
Shape matching and object recognition using shape contexts
 

Recently uploaded

Film vocab for eal 3 students: Australia the movie
Film vocab for eal 3 students: Australia the movieFilm vocab for eal 3 students: Australia the movie
Film vocab for eal 3 students: Australia the movie
Nicholas Montgomery
 
What is Digital Literacy? A guest blog from Andy McLaughlin, University of Ab...
What is Digital Literacy? A guest blog from Andy McLaughlin, University of Ab...What is Digital Literacy? A guest blog from Andy McLaughlin, University of Ab...
What is Digital Literacy? A guest blog from Andy McLaughlin, University of Ab...
GeorgeMilliken2
 
Walmart Business+ and Spark Good for Nonprofits.pdf
Walmart Business+ and Spark Good for Nonprofits.pdfWalmart Business+ and Spark Good for Nonprofits.pdf
Walmart Business+ and Spark Good for Nonprofits.pdf
TechSoup
 
MARY JANE WILSON, A “BOA MÃE” .
MARY JANE WILSON, A “BOA MÃE”           .MARY JANE WILSON, A “BOA MÃE”           .
MARY JANE WILSON, A “BOA MÃE” .
Colégio Santa Teresinha
 
Natural birth techniques - Mrs.Akanksha Trivedi Rama University
Natural birth techniques - Mrs.Akanksha Trivedi Rama UniversityNatural birth techniques - Mrs.Akanksha Trivedi Rama University
Natural birth techniques - Mrs.Akanksha Trivedi Rama University
Akanksha trivedi rama nursing college kanpur.
 
Digital Artifact 1 - 10VCD Environments Unit
Digital Artifact 1 - 10VCD Environments UnitDigital Artifact 1 - 10VCD Environments Unit
Digital Artifact 1 - 10VCD Environments Unit
chanes7
 
How to Build a Module in Odoo 17 Using the Scaffold Method
How to Build a Module in Odoo 17 Using the Scaffold MethodHow to Build a Module in Odoo 17 Using the Scaffold Method
How to Build a Module in Odoo 17 Using the Scaffold Method
Celine George
 
Liberal Approach to the Study of Indian Politics.pdf
Liberal Approach to the Study of Indian Politics.pdfLiberal Approach to the Study of Indian Politics.pdf
Liberal Approach to the Study of Indian Politics.pdf
WaniBasim
 
The Diamonds of 2023-2024 in the IGRA collection
The Diamonds of 2023-2024 in the IGRA collectionThe Diamonds of 2023-2024 in the IGRA collection
The Diamonds of 2023-2024 in the IGRA collection
Israel Genealogy Research Association
 
The basics of sentences session 6pptx.pptx
The basics of sentences session 6pptx.pptxThe basics of sentences session 6pptx.pptx
The basics of sentences session 6pptx.pptx
heathfieldcps1
 
Community pharmacy- Social and preventive pharmacy UNIT 5
Community pharmacy- Social and preventive pharmacy UNIT 5Community pharmacy- Social and preventive pharmacy UNIT 5
Community pharmacy- Social and preventive pharmacy UNIT 5
sayalidalavi006
 
Azure Interview Questions and Answers PDF By ScholarHat
Azure Interview Questions and Answers PDF By ScholarHatAzure Interview Questions and Answers PDF By ScholarHat
Azure Interview Questions and Answers PDF By ScholarHat
Scholarhat
 
Advanced Java[Extra Concepts, Not Difficult].docx
Advanced Java[Extra Concepts, Not Difficult].docxAdvanced Java[Extra Concepts, Not Difficult].docx
Advanced Java[Extra Concepts, Not Difficult].docx
adhitya5119
 
Pride Month Slides 2024 David Douglas School District
Pride Month Slides 2024 David Douglas School DistrictPride Month Slides 2024 David Douglas School District
Pride Month Slides 2024 David Douglas School District
David Douglas School District
 
ANATOMY AND BIOMECHANICS OF HIP JOINT.pdf
ANATOMY AND BIOMECHANICS OF HIP JOINT.pdfANATOMY AND BIOMECHANICS OF HIP JOINT.pdf
ANATOMY AND BIOMECHANICS OF HIP JOINT.pdf
Priyankaranawat4
 
S1-Introduction-Biopesticides in ICM.pptx
S1-Introduction-Biopesticides in ICM.pptxS1-Introduction-Biopesticides in ICM.pptx
S1-Introduction-Biopesticides in ICM.pptx
tarandeep35
 
RPMS TEMPLATE FOR SCHOOL YEAR 2023-2024 FOR TEACHER 1 TO TEACHER 3
RPMS TEMPLATE FOR SCHOOL YEAR 2023-2024 FOR TEACHER 1 TO TEACHER 3RPMS TEMPLATE FOR SCHOOL YEAR 2023-2024 FOR TEACHER 1 TO TEACHER 3
RPMS TEMPLATE FOR SCHOOL YEAR 2023-2024 FOR TEACHER 1 TO TEACHER 3
IreneSebastianRueco1
 
BÀI TẬP BỔ TRỢ TIẾNG ANH 8 CẢ NĂM - GLOBAL SUCCESS - NĂM HỌC 2023-2024 (CÓ FI...
BÀI TẬP BỔ TRỢ TIẾNG ANH 8 CẢ NĂM - GLOBAL SUCCESS - NĂM HỌC 2023-2024 (CÓ FI...BÀI TẬP BỔ TRỢ TIẾNG ANH 8 CẢ NĂM - GLOBAL SUCCESS - NĂM HỌC 2023-2024 (CÓ FI...
BÀI TẬP BỔ TRỢ TIẾNG ANH 8 CẢ NĂM - GLOBAL SUCCESS - NĂM HỌC 2023-2024 (CÓ FI...
Nguyen Thanh Tu Collection
 
BBR 2024 Summer Sessions Interview Training
BBR  2024 Summer Sessions Interview TrainingBBR  2024 Summer Sessions Interview Training
BBR 2024 Summer Sessions Interview Training
Katrina Pritchard
 
বাংলাদেশ অর্থনৈতিক সমীক্ষা (Economic Review) ২০২৪ UJS App.pdf
বাংলাদেশ অর্থনৈতিক সমীক্ষা (Economic Review) ২০২৪ UJS App.pdfবাংলাদেশ অর্থনৈতিক সমীক্ষা (Economic Review) ২০২৪ UJS App.pdf
বাংলাদেশ অর্থনৈতিক সমীক্ষা (Economic Review) ২০২৪ UJS App.pdf
eBook.com.bd (প্রয়োজনীয় বাংলা বই)
 

Recently uploaded (20)

Film vocab for eal 3 students: Australia the movie
Film vocab for eal 3 students: Australia the movieFilm vocab for eal 3 students: Australia the movie
Film vocab for eal 3 students: Australia the movie
 
What is Digital Literacy? A guest blog from Andy McLaughlin, University of Ab...
What is Digital Literacy? A guest blog from Andy McLaughlin, University of Ab...What is Digital Literacy? A guest blog from Andy McLaughlin, University of Ab...
What is Digital Literacy? A guest blog from Andy McLaughlin, University of Ab...
 
Walmart Business+ and Spark Good for Nonprofits.pdf
Walmart Business+ and Spark Good for Nonprofits.pdfWalmart Business+ and Spark Good for Nonprofits.pdf
Walmart Business+ and Spark Good for Nonprofits.pdf
 
MARY JANE WILSON, A “BOA MÃE” .
MARY JANE WILSON, A “BOA MÃE”           .MARY JANE WILSON, A “BOA MÃE”           .
MARY JANE WILSON, A “BOA MÃE” .
 
Natural birth techniques - Mrs.Akanksha Trivedi Rama University
Natural birth techniques - Mrs.Akanksha Trivedi Rama UniversityNatural birth techniques - Mrs.Akanksha Trivedi Rama University
Natural birth techniques - Mrs.Akanksha Trivedi Rama University
 
Digital Artifact 1 - 10VCD Environments Unit
Digital Artifact 1 - 10VCD Environments UnitDigital Artifact 1 - 10VCD Environments Unit
Digital Artifact 1 - 10VCD Environments Unit
 
How to Build a Module in Odoo 17 Using the Scaffold Method
How to Build a Module in Odoo 17 Using the Scaffold MethodHow to Build a Module in Odoo 17 Using the Scaffold Method
How to Build a Module in Odoo 17 Using the Scaffold Method
 
Liberal Approach to the Study of Indian Politics.pdf
Liberal Approach to the Study of Indian Politics.pdfLiberal Approach to the Study of Indian Politics.pdf
Liberal Approach to the Study of Indian Politics.pdf
 
The Diamonds of 2023-2024 in the IGRA collection
The Diamonds of 2023-2024 in the IGRA collectionThe Diamonds of 2023-2024 in the IGRA collection
The Diamonds of 2023-2024 in the IGRA collection
 
The basics of sentences session 6pptx.pptx
The basics of sentences session 6pptx.pptxThe basics of sentences session 6pptx.pptx
The basics of sentences session 6pptx.pptx
 
Community pharmacy- Social and preventive pharmacy UNIT 5
Community pharmacy- Social and preventive pharmacy UNIT 5Community pharmacy- Social and preventive pharmacy UNIT 5
Community pharmacy- Social and preventive pharmacy UNIT 5
 
Azure Interview Questions and Answers PDF By ScholarHat
Azure Interview Questions and Answers PDF By ScholarHatAzure Interview Questions and Answers PDF By ScholarHat
Azure Interview Questions and Answers PDF By ScholarHat
 
Advanced Java[Extra Concepts, Not Difficult].docx
Advanced Java[Extra Concepts, Not Difficult].docxAdvanced Java[Extra Concepts, Not Difficult].docx
Advanced Java[Extra Concepts, Not Difficult].docx
 
Pride Month Slides 2024 David Douglas School District
Pride Month Slides 2024 David Douglas School DistrictPride Month Slides 2024 David Douglas School District
Pride Month Slides 2024 David Douglas School District
 
ANATOMY AND BIOMECHANICS OF HIP JOINT.pdf
ANATOMY AND BIOMECHANICS OF HIP JOINT.pdfANATOMY AND BIOMECHANICS OF HIP JOINT.pdf
ANATOMY AND BIOMECHANICS OF HIP JOINT.pdf
 
S1-Introduction-Biopesticides in ICM.pptx
S1-Introduction-Biopesticides in ICM.pptxS1-Introduction-Biopesticides in ICM.pptx
S1-Introduction-Biopesticides in ICM.pptx
 
RPMS TEMPLATE FOR SCHOOL YEAR 2023-2024 FOR TEACHER 1 TO TEACHER 3
RPMS TEMPLATE FOR SCHOOL YEAR 2023-2024 FOR TEACHER 1 TO TEACHER 3RPMS TEMPLATE FOR SCHOOL YEAR 2023-2024 FOR TEACHER 1 TO TEACHER 3
RPMS TEMPLATE FOR SCHOOL YEAR 2023-2024 FOR TEACHER 1 TO TEACHER 3
 
BÀI TẬP BỔ TRỢ TIẾNG ANH 8 CẢ NĂM - GLOBAL SUCCESS - NĂM HỌC 2023-2024 (CÓ FI...
BÀI TẬP BỔ TRỢ TIẾNG ANH 8 CẢ NĂM - GLOBAL SUCCESS - NĂM HỌC 2023-2024 (CÓ FI...BÀI TẬP BỔ TRỢ TIẾNG ANH 8 CẢ NĂM - GLOBAL SUCCESS - NĂM HỌC 2023-2024 (CÓ FI...
BÀI TẬP BỔ TRỢ TIẾNG ANH 8 CẢ NĂM - GLOBAL SUCCESS - NĂM HỌC 2023-2024 (CÓ FI...
 
BBR 2024 Summer Sessions Interview Training
BBR  2024 Summer Sessions Interview TrainingBBR  2024 Summer Sessions Interview Training
BBR 2024 Summer Sessions Interview Training
 
বাংলাদেশ অর্থনৈতিক সমীক্ষা (Economic Review) ২০২৪ UJS App.pdf
বাংলাদেশ অর্থনৈতিক সমীক্ষা (Economic Review) ২০২৪ UJS App.pdfবাংলাদেশ অর্থনৈতিক সমীক্ষা (Economic Review) ২০২৪ UJS App.pdf
বাংলাদেশ অর্থনৈতিক সমীক্ষা (Economic Review) ২০২৪ UJS App.pdf
 

Biased normalized cuts

  • 1. Biased Normalized Cuts Subhransu Maji1 , Nisheeth K. Vishnoi2, and Jitendra Malik1 University of California, Berkeley1 Microsoft Research India, Bangalore2 smaji@eecs.berkeley.edu, nisheeth.vishnoi@gmail.com, malik@eecs.berkeley.edu Abstract (a) Image (b) Bottom-up (c) Top-down (d) Biased Ncut We present a modification of “Normalized Cuts” to in- corporate priors which can be used for constrained im- age segmentation. Compared to previous generalizations of “Normalized Cuts” which incorporate constraints, our technique has two advantages. First, we seek solutions Figure 1. Segmentation using bottom-up & top-down information. which are sufficiently “correlated” with priors which allows us to use noisy top-down information, for example from an object detector. Second, given the spectral solution of the with results as shown in Figure 1(d). The formalism is that unconstrained problem, the solution of the constrained one of graph partitioning approached using tools and techniques can be computed in small additional time, which allows from spectral graph theory. As in Normalized Cuts [18], we us to run the algorithm in an interactive mode. We com- begin by computing eigenvectors of the Normalized Lapla- pare our algorithm to other graph cut based algorithms and cian of a graph where the edge weights wij correspond to highlight the advantages. the “affinity” between two pixels based on low-level simi- larity cues. This is illustrated in Figure 2 for the cat image, where the top row shows the second to fifth eigenvectors, 1. Introduction ui . Note that none of them is a particularly good indicator Consider Figure 1. Suppose that we want to segment the vector from which the cat can be extracted by thresholding. cat from its background. We can approach this We are interested in cuts which not only minimize the 1. Bottom-up, we might detect contours corresponding to normalized cut value but, at the same time, one of the sets significant changes in brightness, color or texture. The ¯ in the partition (S, S) is required to have a sufficient over- output of one such detector is shown in Figure 1(b). lap with a “bias” or a “prior guess”. In Figure 1, this was Note that there will be internal contours, e.g. corre- the output of a top-down detector. In Figure 2, bottom sponding to the colored spots, that are of higher con- row, the user has specified the bias by indicating various trast than the external contour separating the cat from point sets T in the different panels. We will define a seed- its background. This makes it difficult for any bottom- vector sT associated to T . Let DG be the diagonal matrix up segmentation process to separate the cat from its of the graph, and ui , λi be the eigenvectors and eigenval- background in the absence of high-level knowledge of ues of the Normalized Laplacian. Theorem 3.1 (originally what cats look like. due to [15]) shows that we can get cuts that meet the twin goals of having a small normalized cut, and being suffi- 2. Top-down, we might look for strong activation of, say, ciently correlated with sT by a remarkably simple proce- a sliding window cat detector. The output of one such dure. Construct the biased normalized cut vector, x⋆ , where part-based detector is shown in Figure 1(c). But note n x⋆ = c i=2 λi1 ui uT DG sT . Intuitively the eigenvec- −γ i that this too is inadequate for our purpose. The outputs tors are linearly combined such that the ones that are well of these detectors are typically not sharply localized; correlated with the seed vector are up weighted, while those indeed that is almost a consequence of the desire to that are inversely correlated have their sign flipped. In Fig- make them invariant to small deformations. ure 2 bottom row, these vectors are shown as images, and In this paper we present an approach, “Biased Normal- the bias point sets are overlaid. This vector x⋆ can be thresh- ized Cuts”, where we try to get the best of both worlds, olded to find the desired “biased normalized cut”. One can 2057
  • 2. run the algorithm in an interactive mode by computing the small number of single-commodity maximum-flows to find eigenvectors once and generating the biased normalized cut low-conductance cuts not only inside the input subset T , for any bias. ¯ but among all cuts which are well-correlated with (T, T ). Among spectral methods, local graph partitioning was in- 2. Previous Work troduced by Spielman and Teng [20], who were interested in finding a low-conductance cut in a graph in time nearly- Spectral graph theory originated in the 1970s with work linear in the volume of the output cut. They used random such as Donath and Hoffman [10] and Fiedler [13]. By the walk based methods to do this; and subsequently this result 1990s, there was a well developed literature in mathematics was improved by Andersen, Chung and Lang [1] by doing and theoretical computer science, and Chung [9] provides certain Personalized PageRank based random walks. a monograph level treatment. In computer vision and ma- chine learning, the work of Shi and Malik on normalized 3. Biased Graph Partitioning cuts [18, 19] was very influential, and since then a signifi- cant body of work has emerged in these fields that we can- Image as a Weighted Graph and Normalized Cuts. The not possibly do justice here. The best current approaches image is represented as a weighted undirected graph G = to finding contours and regions bottom up are still based on (V, E) where the nodes of the graph are the points in the variations of this theme [16]. feature space, and an edge is formed between every pair of Given our interest in this work of biasing the solution nodes. The weight function on the edges w : E → R≥0 to satisfy additional constraints, two papers need particular is a function of the similarity between the end points of the mention. The problem in the constrained setting for image edge. The volume of a set of vertices S is the total weight def segmentation has been studied by Yu and Shi [21] where of the edges incident to it: vol(S) = i∈S,j∈V w(i, j). they find the solution to the normalized cuts problem sub- def vol(G) = i,j∈V w(i, j) be the total weight of the edges in ject to a set of linear constraints of the form U T x = 0. This the graph. Normalized cut measure, defined next, is a stan- problem can be reduced to an eigenvalue problem which dard way to measure the degree of dissimilarity between can be solved using spectral techniques as well. Ericks- ¯ two pieces of an image. For S ⊆ V, let S denote V S. son et al. [11] generalize the set of constraints to U T x = b. We will see in Section 5 that enforcing constraints in this ¯ def Let cut(S, S) = i∈S,j∈S w(i, j) be the weight of edges ¯ manner is not robust when the constraints are noisy. The ¯ crossing the cut (S, S). Then, the normalized cut value cor- computational complexity of these approaches is also sig- ¯ responding to (S, S) is defined as nificantly higher than that of solving the basic normalized def ¯ ¯ cut(S, S) cut(S, S) cuts problem. Ncut(S) = + In the last decade the most popular approach to inter- vol(S) ¯ . vol(S) active image segmentation in computer vision and graph- ¯ Noting that vol(S) + vol(S) = vol(G), it can be seen that ics has followed in the steps of the work of Boykov and this notion of a normalized cut is the same as that of the Jolly [7] who find an image segmentation by computing a traditional graph conductance defined for a set S as min-cut/max-flow on a graph which encodes both the user constraints and pairwise pixel similarity. This line of work def ¯ cut(S, S) has been further investigated by Blake and Rother [3] where φ(S) = vol(G) · ¯ . vol(S) · vol(S) they experiment with ways to model the foreground and background regions. In the GrabCut framework [17], the def The conductance of the graph G is φ(G) = minS⊆V φ(S) process of segmentation and foreground/background mod- and is an extremely well studied quantity in graph theory, eling is repeated till convergence. With the advances in the computer science and machine learning. It is NP-hard to min-cut/max-flow algorithms like [6], these methods have compute exactly and one of the earliest and most popular become computationally attractive and can often be used in approach to compute an approximation to it is to write down an interactive mode. However, these methods fail when the a relaxation which boils down to computing the second constraints are sparse making it difficult to construct good eigenvalue of the Laplacian matrix associated to a graph. foreground/background models and these methods tend to We discuss this next. produce isolated cuts. From a theoretical perspective, there has been a signif- icant interest in the cut improvement problem: given as Graphs and Laplacians. For a graph G = (V, E) with input a graph and a cut, find a subset that is a better cut. edge weights function w, let AG ∈ RV ×V denote its ad- This started with Gallo, Grigoriadis and Tarjan [14] and jacency matrix with AG (i, j) = w(i, j); DG denotes the has attained more recent attention in the work of Ander- diagonal degree matrix of G, i.e., DG (i, i) = j∈V w(i, j) def sen and Lang [2], who gave a general algorithm that uses a DG (i, j) = 0, for all i = j; LG = DG − AG will 2058
  • 3. Figure 2. Top Row: Input Image and the top 4 eigenvectors computed using the intervening contour cue [16]. Bottom Row: Biased normalized cuts for various seed sets T . The marked points in each image are the set of points in T . def denote the (combinatorial) Laplacian of G; and LG = Biased Normalized Cuts. Now we move on to incorpo- −1/2 −1/2 rating the prior information given to us about the image to DG L G DG will denote the normalized Laplacian of G. We will assume G is connected, in which case the define the notion of biased normalized cuts. Recall that eigenvalues of LG are 0 = λ1 < λ2 ≤ · · · ≤ λn . our problem is: we are given a region of interest in the We will denote by λ2 (G) as this second eigenvalue of the image and we would like to segment the image so that normalized Laplacian of G. If u1 , . . . , un are the corre- the segment is biased towards the specified region. A re- def −1 gion is modeled as a subset T ⊆ V, of the vertices of the sponding eigenvectors of LG , then we define vi = DG 2 ui ¯ image. We would be interested in cuts (S, S) which not and think of them as the associated eigenvectors of LG ; only minimize the normalized cut value but, at the same LG vi = λi DG vi . The spectral decomposition of LG can time, have sufficient correlation with the region specified be written as n λi ui uT . We can also define the Moore- i=2 i def n by T. To model this, we will first associate a vector sT to Penrose pseudo-inverse of LG as L+ = G 1 T i=2 λi ui ui . ¯ vol(T )vol(T ) 1 the set T as follows: sT (i) = vol(G) · vol(T ) , if ¯ i ∈ T , and sT (i) = − vol(T )vol(T ) · vol(T ) if i ∈ T ; or vol(G) 1 ¯ ¯ The Spectral Relaxation to Computing Normalized def ¯ vol(T )vol(T ) 1T 1T¯ Cuts. Spectral methods approximate the solution to nor- equivalently, sT = vol(G) vol(T ) − vol(T ) . We ¯ malized cuts by trying to find a x ∈ RV which minimizes have defined it in a way such that i∈V sT (i)di = 0 and 2 xT LG x subject to i,j∈V di dj (xi − xj )2 = vol(G). To i∈V sT (i) di = 1. see why this is a relaxation first, for a subset S of ver- This notion of biased normalized cuts is quite natural and tices, let 1S be the indicator vector of S in RV . Then motivated from the theory of local graph partitioning where ¯ 1T LG 1S = ij∈E wi,j (1S (i) − 1S (j))2 = cut(S, S) and the goal is to find a low-conductance cut well correlated S 2 ¯ Hence, if i,j∈V di dj (1S (i) − 1S (j)) = vol(S)vol(S). with a specified input set. The correlation is specified by a vol(G) parameter κ ∈ (0, 1). This allows us to explore the possi- we let x = ¯ · 1S , then x satisfies the above vol(S)·vol(S) bility of image segments which are well-correlated with the constraints and has objective value exactly φ(S). Hence, the prior information which may have much less normalized cut minimum of the quadratic program above can be at-most the value than T itself and, hence, refine the initial guess. In normalized cut value of G. particular, we consider the spectral relaxation in Figure 3 to In fact, it is easy to see that the optimal value of this a κ-biased normalized cut around T . optimization problem, if G is connected, is λ2 (G) and the Note that x = sT , is a feasible solution to this spectral −1 optimal vector to this program is DG 2 u2 , where u2 is the relaxation. Also note that if v2 satisfies the correlation con- eigenvector corresponding to the smallest non-zero eigen- straint with sT , then it will be the optimal to this program. value (λ2 (G)) of LG . In particular, the optimal vector x⋆ What is quite interesting is that one can characterize the op- has the property that i∈V x⋆ di = 0. i timal solution to this spectral program under mild condi- 2059
  • 4. Algorithm 1 Biased Normalized Cuts (G, w, sT , γ) minimize xT LG x Require: Graph G = (V, E), edge weight function w, seed s.t. di dj (xi − xj )2 = vol(G) sT and a correlation parameter γ ∈ (−∞, λ2 (G)) 1: AG (i, j) ← w(i, j), DG (i, i) ← j w(i, j) i,j∈V −1/2 −1/2 2: LG ← DG − AG , LG ← DG LG DG ( xi sT (i)di )2 ≥ κ 3: Compute u1 , u2 , . . . , uK the eigenvectors of LG corre- i∈V Figure 3. BiasedNcut(G, T, κ)- Spectral relaxation to compute sponding to the K smallest eigenvalues λ1 , λ2 , . . . , λK . κ-biased normalized cuts around T uT DG sT 4: wi ← i λi −γ , for i = 2, . . . , K K tions on the graph G and the set T and, as it turns out, one 5: Obtain the biased normalized cut, x∗ ∝ i=2 wi ui has to do very little effort to compute the optimal solution if one has already computed the spectrum of LG . This is cap- tured by the following theorem which is due to [15]. We of a second [8]. Our method can be faster than the min- include a proof of this in the appendix for completeness. cut/max-flow cut based approaches in an interactive setting as these eigenvectors need to be computed just once. In ad- Theorem 3.1. Let G be a connected graph and T be such dition the real valued solution like the one shown in Figure 2 that i∈V sT (i)v2 (i)di = 0. Further, let 1 ≥ κ ≥ 0 be a might provide the user better guidance than a hard segmen- correlation parameter. Then, there is an optimal solution, tation produced by a min-cut algorithm. x⋆ , to the spectral relaxation to the κ-biased normalized n cuts around T such that x⋆ = c i=2 λi1 ui uT DG sT for −γ i 5. Constrained Normalized Cuts some γ ∈ (−∞, λ2 (G)) and a constant c. We compare our approach to the constrained normalized 4. Algorithm cut formulation of Yu and Shi [21] and Ericksson et al. [11]. The later generalized the linear constraints to U T P x = b, Theorem 3.1 shows that the final solution is a weighted where b could be an arbitrary non-zero vector. We consider combination of the eigenvectors and the weight of each a toy example similar to the one used by Yu and Shi. The au- eigenvector is proportional to the “correlation” with the thors observe that when the set of constraints are small it is seed, uT DG sT and inversely proportional to λi − γ. In- i better to use “conditioned constraints”, U T P x = 0, where tuitively the eigenvectors that are well correlated with the P = D−1 W , instead of the original constraints, U T x = 0. seed vector are up weighted, while those that are inversely The ”conditioned constraints” propagate the constraints to correlated have their sign flipped. the neighbors of the constrained points avoiding solutions Often for images the λi − γ grows quickly and one can that are too close to the unconstrained one. obtain a good approximation by considering eigenvectors To illustrate our approach, let us consider the points for the K smallest eigenvalues. In our experiments with p1 , p2 , . . . pn grouped into three sets S1 , S2 and S3 from natural images we set K = 26, i.e. use the top 25 eigen- left to right, as shown in Figure 4. We construct a graph vectors ignoring the all ones vector. We also set γ pa- with edge weights between points pi and pj rameter which controls the amount of correlation implic- itly to γ = −τ × λavg , where τ = 1 and λavg is the av- wij = exp −d(pi , pj )2 /2σ 2 (1) erage of the top K eigenvalues. This could also be user defined parameter in an interactive setting. Algorithm 1 with σ = 3, where d(x, y) is the Euclidean distance be- describes our method for computing the biased normalized tween the points x and y. The unconstrained solution to cuts. Steps 1, 2, 3 are also the steps for computing the seg- the normalized cuts correctly groups the points as shown in mentations using normalized cuts which involve computing Figure 4. the K smallest eigenvectors of the normalized graph lapla- Now suppose we want to group S1 and S3 together. This cian LG . The biased normalized cut for any seed vector can be done by adding a constraint that the circled points sT is the weighted combination of eigenvectors where the belong to the same group and can encoded as a linear con- weights are computed in step 4. In the interactive setting straint of the form, xi −xj = 0, where i and j are the indices only steps 4 and 5 need to be done when the seed vector of the constrained points. The constrained cut formulation changes which is very quick. is able to correctly separate S1 and S3 from S2 as shown in The time taken by the algorithm is dominated by the time Figure 5(a). For our approach we construct the vector sT taken to compute the eigenvectors. In an interactive setting with i, j ∈ T defined earlier and use the top 16 generalized one can use special purpose hardware accelerated methods eigenvectors. The biased cut solution separates S1 and S3 to compute the eigenvectors of typical images in fraction from S2 as well. 2060
  • 5. Input Points Normalized Cut Image Pb Prior Biased Ncut 9 0.6 8 0.4 7 6 0.2 5 0 4 −0.2 3 −0.4 2 −0.6 1 −0.8 5 10 15 20 50 100 150 200 250 300 Figure 4. Input points (left) clustered using normalized cuts (right). Next we increase the number of such constraints by ran- domly sampling points from S1 ∪ S3 and adding constraints that they belong together. Given a set of n points we gen- erate n − 1 constraints by ordering the points and adding a constraint for each consecutive pair similar to the approach of [21]. Instead of improving the solution, the solution to the constrained cuts deteriorates when the number of con- straints are large as the solution that separates S1 and S3 from S2 is no longer feasible. Our method on the other hand gets better with more constraints as seen in Figure 5(b). Figure 8. Biased normalized cuts using top-down priors from an object detector. We also test the robustness of the algorithm to outliers by adding some points from S2 into the constraints. The constrained cut solution deteriorates even when there is one Bias from object detectors. Although in our formulation outlier as seen in Figure 5(c). Our method on the other hand the seed vector sT was discrete, it is not necessary. In par- remains fairly robust even when there are two outliers Fig- ticular we can use the probability estimates produced by an ure 5(d). object detector as a bias. Figure 8 shows some examples of a top down segmentation mask generated by the detector 6. Qualitative Evaluation of [4] used as a seed vector directly after normalization. We present qualitative results on images taken from 7. Conclusion PASCAL VOC 2010 dataset [12]. For all our experiments We present a simple and effective method for comput- we use the intervening contour cue [16] for computing the ing the biased normalized cuts of the images to incorporate weight matrix and 25 eigenvectors to compute the biased top-down priors or user input. The formulation is attractive normalized cut. We refrain from doing quantitive compari- as it allows one to incorporate these priors without addi- son on segmentation benchmarks as the goal of this paper is tional overhead. Linear combinations of eigenvectors nat- to introduce a new algorithm for computing biased normal- urally “regularizes” the combinatorial space of segmenta- ized cuts, which is only one of the components of a good tions. Code for computing biased normalized cuts interac- segmentation engine. tively on images can be downloaded at the author’s website. Acknowledgements: Thanks to Pablo Arbelaez. Subhransu Maji is sup- Effect of γ. The correlation parameter γ when set smaller ported by a fellowship from Google Inc. and ONR MURI N00014-06-1- values increases the correlation κ with the seed vector. Fig- 0734. The work was done when the authors were at Microsoft Research ure 6 shows the effect of γ on the result of the biased nor- India. malized cut. One can obtain tighter cuts around the seed by setting γ to smaller values. In an interactive setup this could A. Appendix be adjusted by the user. Proof. [of Theorem 3.1] To start off, note that BiasedNcut(G, T, κ) is a non-convex program. One Bias from user interaction. We first test the algorithm can relax it to SDPp (G, T, κ) of Figure 9. In the same in the interactive setting by automatically sampling a set of figure the dual of this SDP appears: SDPd (G, T, κ). Here, points inside the figure mask. Figure 7 shows the several ex- def ample outputs. All these images are taken from the animal for matrices A, B ∈ Rn×n , A ◦ B = i,j Ai,j Bi,j . def 1 categories of the Pascal VOC 2010 detection dataset. Also Ln = DG − vol(G) DG JDG , where J is the matrix 2061
  • 6. 1 contraint Biased Ncut Yu & Shi 19 contraints Biased Ncut Yu & Shi 0.04 0.2 2 8 0.02 8 0.1 6 0 6 0 0 −0.02 0 4 −0.04 4 −0.2 −2 2 −0.06 2 −0.1 −0.4 −0.08 5 10 15 20 100 200 300 100 200 300 5 10 15 20 100 200 300 100 200 300 (a) One point each sampled from S2 and S3 (b) 10 points each sampled from S2 and S3 10 contraints Biased Ncut Yu & Shi 11 contraint Biased Ncut Yu & Shi 1 1 8 0.05 8 0.2 0.5 0 6 0 6 0 0.1 −0.05 4 4 −0.5 −0.1 0 2 −1 −0.15 2 −1 −0.1 5 10 15 20 100 200 300 100 200 300 5 10 15 20 100 200 300 100 200 300 (c) 5 points each from sets S1 and S3 with 1 outlier from S2 (d) 5 points each from sets S1 and S3 with 2 outliers from S2 Figure 5. Comparison of our biased normalized cut approach to constrained cut approach of [21, 11] for various constraints shown by red circled points. When the number of constraints is large (b) or there are outliers (c, d), our method produces better solutions than the solution of constrained cuts. Figure 6. Effect of γ. Input image and biased cuts for decreasing values of γ from left to right. The local cuts are more and more correlated with the seed set (shown as dots) as γ decreases. with Ji,j = 1 for all i, j. Verifying Slater’s condition, one can observe that strong duality holds for this SDP minimize LG ◦ X relaxation. Then, using strong duality and the comple- subject to Ln ◦ X = 1 mentary slackness conditions implied by it, we will argue that the SDPp (G, T, κ) has a rank one optimal solution (DG sT )(DG sT )T ◦ X ≥ κ under the conditions of the theorem. This implies that X 0 the optimal solution of SDPp (G, T, κ) is the same as the optimal solution of BiasedNcut. Combining this with the maximize α + κβ complementary slackness condition obtained from the dual subject to LG αLn + β(DG sT )(DG sT )T SDPd (G, T, κ), one can derive that the optimal rank one solution has, up to a constant, the desired form promised α ∈ R, β ≥ 0 by the theorem. Now we expand the above steps in claims. Figure 9. Top: SDPp (G, T, κ) Primal SDP relaxation. Bottom: SDPd (G, T, κ) Dual SDP Claim A.1. The primal SDPp (G, T, κ) is a relaxation of the vector program BiasedNcut(G, T, κ). for this primal SDP. Consider X = sT sT T . Then, Proof. Consider a vector x that is a feasible solution to (DG sT )(DG sT )T ◦ sT sT T = (sT T DG sT )2 = 1 > κ. BiasedNcut(G, T, κ), and note that X = xxT is a feasible solution to SDPp (G, T, κ). Claim A.3. The feasibility and complementary slackness conditions for a primal-dual pair X ⋆ , α⋆ , β ⋆ listed in Fig- Claim A.2. Strong duality holds between SDPp (G, T, κ) ure 10 are sufficient for them to be an optimal solution. and SDPd (G, T, κ). Proof. Since SDPp (G, T, κ) is convex, if suffices to ver- Proof. This follows from the convexity of SDPp (G, T, κ) ify that Slater’s constraint qualification condition is true and Slater’s condition [5]. 2062
  • 7. Image Pb Biased Ncut Image Pb Biased Ncut Figure 7. Example biased cuts. Images shown with the “probability of boundary” map (Pb) computed using [16] and the biased normalized cut for various seed sets (marked as red dots). 2063
  • 8. References Ln ◦ X ⋆ = 1 [1] R. Andersen, F. Chung, and K. Lang. Local graph partition- (DG sT )(DG sT )T ◦ X ⋆ ≥ κ ing using PageRank vectors. In FOCS, pages 475–486, 2006. LG − α⋆ Ln − β ⋆ (DG sT )(DG sT )T 0 2058 [2] R. Andersen and K. Lang. An algorithm for improving graph β⋆ ≥ 0 partitions. In SODA, pages 651–660, 2008. 2058 [3] A. Blake, C. Rother, M. Brown, P. Perez, and P. Torr. Inter- α⋆ (Ln ◦ X ⋆ − 1) = 0 active image segmentation using an adaptive gmmrf model. β ⋆ ((DG sT )(DG sT )T ◦ X ⋆ − κ) = 0 In ECCV, pages 428–441, 2004. 2058 [4] L. Bourdev, S. Maji, T. Brox, and J. Malik. Detecting people X ⋆ ◦ (LG − α⋆ Ln − β ⋆ (DG sT )(DG sT )T ) = 0 using mutually consistent poselet activations. In ECCV, sep 2010. 2061 Figure 10. Top: Feasibility conditions. Bottom: Complementary [5] S. Boyd and L. Vandenberghe. Convex Optimization. Cam- slackness conditions. bridge University Press, Cambridge, UK, 2004. 2062 [6] Y. Boykov and V. Kolmogorov. An experimental comparison Claim A.4. These feasibility and complementary slackness of min-cut/max-flow algorithms for energy minimization in conditions, coupled with the assumptions of the theorem, vision. TPAMI, 26:1124–1137, September 2004. 2058 imply that X ⋆ must be rank 1 and β ⋆ > 0. [7] Y. Y. Boykov and M. P. Jolly. Interactive graph cuts for opti- mal boundary & region segmentation of objects in N-D im- Now we complete the proof of the theorem. From Claim ages. In ICCV, 2001. 2058 A.4 it follows that, X ⋆ = x⋆ x⋆T where x⋆ satisfies the [8] B. Catanzaro, B.-Y. Su, N. Sundaram, Y. Lee, M. Murphy, equation (LG −α⋆ Ln −β ⋆ (DG sT )(DG sT )T )x⋆ = 0. From and K. Keutzer. Efficient, high-quality image contour detec- the second complementary slackness condition in Figure tion. In ICCV, pages 2381 –2388, 2009. 2060 10, and the fact that β ⋆ > 0, we obtain that i x⋆ sT (i)di = [9] F. Chung. Spectral Graph Theory. American Mathematical √ √ i Society, 1997. 2058 ± κ. Thus, x⋆ = ±β ⋆ κ(LG − α⋆ Ln )+ DG sT . This [10] W. E. Donath and A. J. Hoffman. Lower bounds for the parti- proves the theorem. tioning of graphs. IBM J. Res. Dev., 17:420–425, September A.1. Proof of Claim A.4 1973. 2058 [11] A. Eriksson, C. Olsson, and F. Kahl. Normalized cuts revis- Proof. We start by stating two facts. The second is trivial. ited: A reformulation for segmentation with linear grouping constraints. In ICCV, oct. 2007. 2058, 2060, 2062 Fact A.5. α⋆ ≤ λ2 (G). Moreover if λ2 = α⋆ then [12] M. Everingham, L. Van Gool, C. K. I. Williams, J. Winn, i v2 (i)sT (i)di = 0. and A. Zisserman. The pascal visual object classes (voc) −1 challenge. IJCV, 88(2):303–338, June 2010. 2061 Proof. Recall that v2 = DG 2 u2 where u2 is the unit length [13] M. Fiedler. Algebraic connectivity of graphs. Czechoslovak eigenvector corresponding to λ2 (G) of LG . Plugging in v2 Mathematical Journal, 23(98):298–305, 1973. 2058 in Equation 3 from the feasibility conditions, we obtain that [14] G. Gallo, M. D. Grigoriadis, and R. E. Tarjan. A fast para- v2 LG v2 −α⋆ −β ⋆ ( i v2 (i)sT (i)di )2 ≥ 0. But v2 LG v2 = T T metric maximum flow algorithm and applications. SIAM J. ⋆ ⋆ λ2 (G) and β ≥ 0. Hence, λ2 (G) ≥ α . It follows that if Comput., 18(1):30–55, 1989. 2058 λ2 = α⋆ then i v2 (i)sT (i)di = 0. [15] M. W. Mahoney, L. Orecchia, and N. K. Vishnoi. A spectral algorithm for improving graph partitions. CoRR, Fact A.6. We may assume that the optimal X ⋆ satisfies abs/0912.0681, 2009. 2057, 2060 1 1 1T DG X ⋆ DG 1 = 0, where 1 is the all ones vector. 2 2 [16] M. Maire, P. Arbelaez, C. Fowlkes, and J. Malik. Using con- tours to detect and localize junctions in natural images. In Now we return to the proof of Claim A.4. If we assume ⋆ CVPR, pages 1 –8, 2008. 2058, 2059, 2061, 2063 i v2 (i)sT (i)di = 0, then we know that α < λ2 (G) from [17] C. Rother, V. Kolmogorov, and A. Blake. Grabcut: Interac- Fact A.5. Note that since G is connected and α⋆ < λ2 (G), tive foreground extraction using iterated graph cuts. ACM LG − α⋆ Ln has rank exactly n − 1. From the complemen- Transactions on Graphics, 23:309–314, 2004. 2058 tary slackness condition 2 we can deduce that the image of [18] J. Shi and J. Malik. Normalized cuts and image segmenta- X ⋆ is in the kernel of LG − α⋆ Ln − β ⋆ (DG sT )(DG sT )T . tion. In CVPR, jun. 1997. 2057, 2058 But β ⋆ (DG sT )(DG sT )T is a rank one matrix and since [19] J. Shi and J. Malik. Normalized cuts and image segmenta- ⋆ i sT (i)di = 0, it reduces the rank of LG − α Ln by one tion. TPAMI, 22(8):888 –905, Aug. 2000. 2058 ⋆ ⋆ ⋆ precisely when β > 0. If β = 0 then X must be 0 which [20] D. Spielman and S.-H. Teng. Nearly-linear time algorithms is not possible if SDPp (G, T, κ) is feasible. Hence, the rank for graph partitioning, graph sparsification, and solving lin- of LG −α⋆ Ln −β ⋆ (DG sT )(DG sT )T must be exactly n−2 ear systems. In STOC, pages 81–90, 2004. 2058 and since X ⋆ cannot have any component along the all ones [21] S. X. Yu and J. Shi. Grouping with bias. In NIPS. MIT Press, 2001. 2058, 2060, 2061, 2062 vector, X ⋆ must be rank one. This proves the claim. 2064