Results 1  10
of
178
Computing geodesics and minimal surfaces via graph cuts
 in International Conference on Computer Vision
, 2003
"... Geodesic active contours and graph cuts are two standard image segmentation techniques. We introduce a new segmentation method combining some of their benefits. Our main intuition is that any cut on a graph embedded in some continuous space can be interpreted as a contour (in 2D) or a surface (in 3D ..."
Abstract

Cited by 180 (22 self)
 Add to MetaCart
Geodesic active contours and graph cuts are two standard image segmentation techniques. We introduce a new segmentation method combining some of their benefits. Our main intuition is that any cut on a graph embedded in some continuous space can be interpreted as a contour (in 2D) or a surface (in 3D). We show how to build a grid graph and set its edge weights so that the cost of cuts is arbitrarily close to the length (area) of the corresponding contours (surfaces) for any anisotropic Riemannian metric. There are two interesting consequences of this technical result. First, graph cut algorithms can be used to find globally minimum geodesic contours (minimal surfaces in 3D) under arbitrary Riemannian metric for a given set of boundary conditions. Second, we show how to minimize metrication artifacts in existing graphcut based methods in vision. Theoretically speaking, our work provides an interesting link between several branches of mathematicsdifferential geometry, integral geometry, and combinatorial optimization. The main technical problem is solved using CauchyCrofton formula from integral geometry. 1.
Bounded geometries, fractals, and lowdistortion embeddings
"... The doubling constant of a metric space (X; d) is thesmallest value * such that every ball in X can be covered by * balls of half the radius. The doubling dimension of X isthen defined as dim(X) = log2 *. A metric (or sequence ofmetrics) is called doubling precisely when its doubling dimension is ..."
Abstract

Cited by 154 (31 self)
 Add to MetaCart
The doubling constant of a metric space (X; d) is thesmallest value * such that every ball in X can be covered by * balls of half the radius. The doubling dimension of X isthen defined as dim(X) = log2 *. A metric (or sequence ofmetrics) is called doubling precisely when its doubling dimension is bounded. This is a robust class of metric spaceswhich contains many families of metrics that occur in applied settings.We give tight bounds for embedding doubling metrics into (lowdimensional) normed spaces. We consider bothgeneral doubling metrics, as well as more restricted families such as those arising from trees, from graphs excludinga fixed minor, and from snowflaked metrics. Our techniques include decomposition theorems for doubling metrics, andan analysis of a fractal in the plane due to Laakso [21]. Finally, we discuss some applications and point out a centralopen question regarding dimensionality reduction in L2.
Nearestneighbor searching and metric space dimensions
 In NearestNeighbor Methods for Learning and Vision: Theory and Practice
, 2006
"... Given a set S of n sites (points), and a distance measure d, the nearest neighbor searching problem is to build a data structure so that given a query point q, the site nearest to q can be found quickly. This paper gives a data structure for this problem; the data structure is built using the distan ..."
Abstract

Cited by 87 (0 self)
 Add to MetaCart
Given a set S of n sites (points), and a distance measure d, the nearest neighbor searching problem is to build a data structure so that given a query point q, the site nearest to q can be found quickly. This paper gives a data structure for this problem; the data structure is built using the distance function as a “black box”. The structure is able to speed up nearest neighbor searching in a variety of settings, for example: points in lowdimensional or structured Euclidean space, strings under Hamming and edit distance, and bit vector data from an OCR application. The data structures are observed to need linear space, with a modest constant factor. The preprocessing time needed per site is observed to match the query time. The data structure can be viewed as an application of a “kdtree ” approach in the metric space setting, using Voronoi regions of a subset in place of axisaligned boxes. 1
A comparison of the SheraliAdams, LovászSchrijver and Lasserre relaxations for 01 programming
 Mathematics of Operations Research
, 2001
"... ..."
On Metric RamseyType Phenomena
"... The main question studied in this article may be viewed as a nonlinear analog of Dvoretzky's Theorem in Banach space theory or as part of Ramsey Theory in combinatorics. ..."
Abstract

Cited by 69 (39 self)
 Add to MetaCart
The main question studied in this article may be viewed as a nonlinear analog of Dvoretzky's Theorem in Banach space theory or as part of Ramsey Theory in combinatorics.
On Hierarchical Routing in Doubling Metrics
, 2005
"... We study the problem of routing in doubling metrics, and show how to perform hierarchical routing in such metrics with small stretch and compact routing tables (i.e., with small amount of routing information stored at each vertex). We say that a metric (X, d) has doubling dimension dim(X) at most α ..."
Abstract

Cited by 58 (8 self)
 Add to MetaCart
We study the problem of routing in doubling metrics, and show how to perform hierarchical routing in such metrics with small stretch and compact routing tables (i.e., with small amount of routing information stored at each vertex). We say that a metric (X, d) has doubling dimension dim(X) at most α if every set of diameter D can be covered by 2 α sets of diameter D/2. (A doubling metric is one whose doubling dimension dim(X) is a constant.) We show how to perform (1 + τ)stretch routing on metrics for any 0 < τ ≤ 1 with routing tables of size at most (α/τ) O(α) log 2 ∆ bits with only (α/τ) O(α) log ∆ entries, where ∆ is the diameter of the graph; hence the number of routing table entries is just τ −O(1) log ∆ for doubling metrics. These results extend and improve on those of Talwar (2004). We also give better constructions of sparse spanners for doubling metrics than those obtained from the routing tables above; for τ> 0, we give algorithms to construct (1 + τ)stretch spanners for a metric (X, d) with maximum degree at most (2 + 1/τ) O(dim(X)) , matching the results of Das et al. for Euclidean metrics.
Semidefinite Programming and Integer Programming
"... We survey how semidefinite programming can be used for finding good approximative solutions to hard combinatorial optimization problems. ..."
Abstract

Cited by 48 (7 self)
 Add to MetaCart
We survey how semidefinite programming can be used for finding good approximative solutions to hard combinatorial optimization problems.
Embeddings of negativetype metrics and an improved approximation to generalized sparsest cut
 IN SODA ‘05: PROCEEDINGS OF THE SIXTEENTH ANNUAL ACMSIAM SYMPOSIUM ON DISCRETE ALGORITHMS
, 2005
"... In this paper, we study the metrics of negative type, which are metrics (V,d) such that √ d is an Euclidean metric; these metrics are thus also known as “ℓ2squared” metrics. We show how to embed npoint negativetype metrics into Euclidean space ℓ2 with distortion D = O(log 3/4 n). This embedding re ..."
Abstract

Cited by 44 (0 self)
 Add to MetaCart
In this paper, we study the metrics of negative type, which are metrics (V,d) such that √ d is an Euclidean metric; these metrics are thus also known as “ℓ2squared” metrics. We show how to embed npoint negativetype metrics into Euclidean space ℓ2 with distortion D = O(log 3/4 n). This embedding result, in turn, implies an O(log 3/4 k)approximation algorithm for the Sparsest Cut problem with nonuniform demands. Another corollary we obtain is that npoint subsets of ℓ1 embed into ℓ2 with distortion O(log 3/4 n).
On the optimality of the random hyperplane rounding technique for MAX CUT
 Algorithms
, 2000
"... MAX CUT is the problem of partitioning the vertices of a graph into two sets, maximizing the number of edges joining these sets. This problem is NPhard. Goemans and Williamson proposed an algorithm that first uses a semidefinite programming relaxation of MAX CUT to embed the vertices of the grap ..."
Abstract

Cited by 44 (3 self)
 Add to MetaCart
MAX CUT is the problem of partitioning the vertices of a graph into two sets, maximizing the number of edges joining these sets. This problem is NPhard. Goemans and Williamson proposed an algorithm that first uses a semidefinite programming relaxation of MAX CUT to embed the vertices of the graph on the surface of an n dimensional sphere, and then uses a random hyperplane to cut the sphere in two, giving a cut of the graph. They show that the expected number of edges in the random cut is at least ff \Delta sdp, where ff ' 0:87856 and sdp is the value of the semidefinite program, which is an upper bound on opt, the number of edges in the maximum cut. This manuscript shows the following results: 1. The integrality ratio of the semidefinite program is ff. The previously known bound on the integrality ratio was roughly 0:8845. 2. In the presence of the so called "triangle constraints", the integrality ratio is no better than roughly 0:891. The previously known bound was above ...
On the Impossibility of Dimension Reduction in l_1
 In Proc. 35th Annu. ACM Sympos. Theory Comput
, 2003
"... The JohnsonLindenstrauss Lemma shows that any n points in Euclidean space (with distances measured by the L2 norm) may be mapped down to O((log n)/ep^2) dimensions such that no pairwise distance is distorted by more than a (1 ep) factor. Determining whether such dimension reduction is possible in L ..."
Abstract

Cited by 43 (1 self)
 Add to MetaCart
The JohnsonLindenstrauss Lemma shows that any n points in Euclidean space (with distances measured by the L2 norm) may be mapped down to O((log n)/ep^2) dimensions such that no pairwise distance is distorted by more than a (1 ep) factor. Determining whether such dimension reduction is possible in L1 has been an intriguing open question. Charikar and Sahai [7] recently showed lower bounds for dimension reduction in L1 that can be achieved by linear projections, and positive results for shortest path metrics of restricted graph families. However the question of general dimension reduction in L1 was still open. For example, it was not known whether it is possible to reduce the number of dimensions to O(log n) with 1 ep distortion. We show strong lower bounds for general dimension reduction in L1. We give an explicity family of n points in L1 such that any embedding with distortion d requires n^Omega(1/d^2) dimensions. This proves that there is no analog of the JohnsonLindenstrauss Lemma for L1