Kenneth L. Clarkson

Last updated
Ken Clarkson at SoCG 2011 Ken Clarkson SoCG 2011.jpg
Ken Clarkson at SoCG 2011

Kenneth Lee Clarkson is an American computer scientist known for his research in computational geometry. He is a researcher at the IBM Almaden Research Center, and co-editor-in-chief of the Journal of Computational Geometry . [1]

Contents

Biography

Clarkson received his Ph.D. from Stanford University in 1984, under the supervision of Andrew Yao. [2] Until 2007 he worked for Bell Labs. [3]

In 1998 he was co-chair of the ACM Symposium on Computational Geometry.

Research

Clarkson's primary research interests are in computational geometry.

His most highly cited paper, with Peter Shor, uses random sampling to devise optimal randomized algorithms for several problems of constructing geometric structures, following up on an earlier singly-authored paper by Clarkson on the same subject. [4] [5] It includes algorithms for finding all intersections among a set of line segments in the plane in expected time , finding the diameter of a set of points in three dimensions in expected time , and constructing the convex hull of points in -dimensional Euclidean space in expected time . The same paper also uses random sampling to prove bounds in discrete geometry, and in particular to give tight bounds on the number of k-sets.

Clarkson has also written highly cited papers on the complexity of arrangements of curves and surfaces, [6] nearest neighbor search, [7] [8] motion planning, [9] and low-dimensional linear programming and LP-type problems. [10]

Awards and honors

In 2008 Clarkson was elected a Fellow of the ACM for his "contributions to computational geometry." [11]

Related Research Articles

Polygon triangulation

In computational geometry, polygon triangulation is the decomposition of a polygonal area P into a set of triangles, i.e., finding a set of triangles with pairwise non-intersecting interiors whose union is P.

Euclidean minimum spanning tree

The Euclidean minimum spanning tree or EMST is a minimum spanning tree of a set of n points in the plane, where the weight of the edge between each pair of points is the Euclidean distance between those two points. In simpler terms, an EMST connects a set of dots using lines such that the total length of all the lines is minimized and any dot can be reached from any other by following the lines.

The Fulkerson Prize for outstanding papers in the area of discrete mathematics is sponsored jointly by the Mathematical Optimization Society (MOS) and the American Mathematical Society (AMS). Up to three awards of $1,500 each are presented at each (triennial) International Symposium of the MOS. Originally, the prizes were paid out of a memorial fund administered by the AMS that was established by friends of the late Delbert Ray Fulkerson to encourage mathematical excellence in the fields of research exemplified by his work. The prizes are now funded by an endowment administered by MPS.

Klees measure problem

In computational geometry, Klee's measure problem is the problem of determining how efficiently the measure of a union of (multidimensional) rectangular ranges can be computed. Here, a d-dimensional rectangular range is defined to be a Cartesian product of d intervals of real numbers, which is a subset of Rd.

In computational complexity theory, a computational hardness assumption is the hypothesis that a particular problem cannot be solved efficiently. It is not known how to prove (unconditional) hardness for essentially any useful problem. Instead, computer scientists rely on reductions to formally relate the hardness of a new or complicated problem to a computational hardness assumption about a problem that is better-understood.

K-set (geometry)

In discrete geometry, a k-set of a finite point set S in the Euclidean plane is a subset of k elements of S that can be strictly separated from the remaining points by a line. More generally, in Euclidean space of higher dimensions, a k-set of a finite point set is a subset of k elements that can be separated from the remaining points by a hyperplane. In particular, when k = n/2, the line or hyperplane that separates a k-set from the rest of S is a halving line or halving plane.

The dynamic convex hull problem is a class of dynamic problems in computational geometry. The problem consists in the maintenance, i.e., keeping track, of the convex hull for input data undergoing a sequence of discrete changes, i.e., when input data elements may be inserted, deleted, or modified. It should be distinguished from the kinetic convex hull, which studies similar problems for continuously moving points. Dynamic convex hull problems may be distinguished by the types of the input data and the allowed types of modification of the input data.

In mathematics, the Johnson–Lindenstrauss lemma is a result named after William B. Johnson and Joram Lindenstrauss concerning low-distortion embeddings of points from high-dimensional into low-dimensional Euclidean space. The lemma states that a set of points in a high-dimensional space can be embedded into a space of much lower dimension in such a way that distances between the points are nearly preserved. The map used for the embedding is at least Lipschitz, and can even be taken to be an orthogonal projection.

In computational geometry, the Bentley–Ottmann algorithm is a sweep line algorithm for listing all crossings in a set of line segments, i.e. it finds the intersection points of line segments. It extends the Shamos–Hoey algorithm, a similar previous algorithm for testing whether or not a set of line segments has any crossings. For an input consisting of line segments with crossings, the Bentley–Ottmann algorithm takes time . In cases where , this is an improvement on a naïve algorithm that tests every pair of segments, which takes .

In statistics and computational geometry, the notion of centerpoint is a generalization of the median to data in higher-dimensional Euclidean space. Given a set of points in d-dimensional space, a centerpoint of the set is a point such that any hyperplane that goes through that point divides the set of points in two roughly equal subsets: the smaller part should have at least a 1/(d + 1) fraction of the points. Like the median, a centerpoint need not be one of the data points. Every non-empty set of points has at least one centerpoint.

In computer science, lattice problems are a class of optimization problems related to mathematical objects called lattices. The conjectured intractability of such problems is central to the construction of secure lattice-based cryptosystems: Lattice problems are an example of NP-hard problems which have been shown to be average-case hard, providing a test case for the security of cryptographic algorithms. In addition, some lattice problems which are worst-case hard can be used as a basis for extremely secure cryptographic schemes. The use of worst-case hardness in such schemes makes them among the very few schemes that are very likely secure even against quantum computers. For applications in such cryptosystems, lattices over vector space or free modules are generally considered.

In the analysis of algorithms, several authors have studied the computation of the volume of high-dimensional convex bodies, a problem that can also be used to model many other problems in combinatorial enumeration. Often these works use a black box model of computation in which the input is given by a subroutine for testing whether a point is inside or outside of the convex body, rather than by an explicit listing of the vertices or faces of a convex polytope. It is known that, in this model, no deterministic algorithm can achieve an accurate approximation, and even for an explicit listing of faces or vertices the problem is #P-hard. However, a joint work by Martin Dyer, Alan M. Frieze and Ravindran Kannan provided a randomized polynomial time approximation scheme for the problem, providing a sharp contrast between the capabilities of randomized and deterministic algorithms.

Sauer–Shelah lemma Notion in combinatorics

In combinatorial mathematics and extremal set theory, the Sauer–Shelah lemma states that every family of sets with small VC dimension consists of a small number of sets. It is named after Norbert Sauer and Saharon Shelah, who published it independently of each other in 1972. The same result was also published slightly earlier and again independently, by Vladimir Vapnik and Alexey Chervonenkis, after whom the VC dimension is named. In his paper containing the lemma, Shelah gives credit also to Micha Perles, and for this reason the lemma has also been called the Perles–Sauer–Shelah lemma.

In the study of algorithms, an LP-type problem is an optimization problem that shares certain properties with low-dimensional linear programs and that may be solved by similar algorithms. LP-type problems include many important optimization problems that are not themselves linear programs, such as the problem of finding the smallest circle containing a given set of planar points. They may be solved by a combination of randomized algorithms in an amount of time that is linear in the number of elements defining the problem, and subexponential in the dimension of the problem.

Maxima of a point set

In computational geometry, a point p in a finite set of points S is said to be maximal or non-dominated if there is no other point q in S whose coordinates are all greater than or equal to the corresponding coordinates of p. The maxima of a point setS are all the maximal points of S. The problem of finding all maximal points, sometimes called the problem of the maxima or maxima set problem, has been studied as a variant of the convex hull and orthogonal convex hull problems. It is equivalent to finding the Pareto frontier of a collection of points, and was called the floating-currency problem by Herbert Freeman based on an application involving comparing the relative wealth of individuals with different holdings of multiple currencies.

The geometric set cover problem is the special case of the set cover problem in geometric settings. The input is a range space where is a universe of points in and is a family of subsets of called ranges, defined by the intersection of and geometric shapes such as disks and axis-parallel rectangles. The goal is to select a minimum-size subset of ranges such that every point in the universe is covered by some range in .

Greedy geometric spanner

In computational geometry, a greedy geometric spanner is an undirected graph whose vertices represent points in a Euclidean space, and whose edges are selected by a greedy algorithm to make the shortest path distances in the graph approximate the Euclidean distances between pairs of points. The construction of greedy spanners was first published by Ingo Althöfer et al. in 1993; their paper also credited Marshall Bern (unpublished) with the independent discovery of the same construction.

The sparse Fourier transform (SFT) is a kind of discrete Fourier transform (DFT) for handling big data signals. Specifically, it is used in GPS synchronization, spectrum sensing and analog-to-digital converters.:

In theoretical computer science and metric geometry, the GNRS conjecture connects the theory of graph minors, the stretch factor of embeddings, and the approximation ratio of multi-commodity flow problems. It is named after Anupam Gupta, Ilan Newman, Yuri Rabinovich, and Alistair Sinclair, who formulated it in 2004.

Relative convex hull

In discrete geometry and computational geometry, the relative convex hull or geodesic convex hull is an analogue of the convex hull for the points inside a simple polygon or a rectifiable simple closed curve.

References

  1. Editorial team, Journal of Computational Geometry. Retrieved 2009-05-30.
  2. TCS Genealogy, Association for Computing Machinery.
  3. Clarkson's page at Bell Labs Archived 2008-10-24 at the Wayback Machine , retrieved January 15, 2009.
  4. Clarkson, Kenneth L. (1987), "New applications of random sampling in computational geometry", Discrete and Computational Geometry , 2 (2): 195–222, doi: 10.1007/BF02187879 , MR   0884226 .
  5. Clarkson, Kenneth L.; Shor, Peter W. (1989), "Applications of random sampling in computational geometry. II", Discrete and Computational Geometry , 4 (5): 387–421, doi: 10.1007/BF02187740 , MR   1014736 .
  6. Clarkson, Kenneth L.; Edelsbrunner, Herbert; Guibas, Leonidas J.; Sharir, Micha; Welzl, Emo (1990), "Combinatorial complexity bounds for arrangements of curves and spheres", Discrete and Computational Geometry, 5 (2): 99–160, doi: 10.1007/BF02187783 , MR   1032370 .
  7. Clarkson, Kenneth L. (1988), "A randomized algorithm for closest-point queries", SIAM Journal on Computing, 17 (4): 830–847, doi:10.1137/0217052, MR   0953296 .
  8. Clarkson, K. L. (1999), "Nearest neighbor queries in metric spaces", Discrete and Computational Geometry , 22 (1): 63–93, doi: 10.1007/PL00009449 , MR   1692615 .
  9. Clarkson, K. (1987), "Approximation algorithms for shortest path motion planning", Proc. 19th ACM Symposium on Theory of Computing, pp. 56–65, doi:10.1145/28395.28402, S2CID   12206444 .
  10. Clarkson, Kenneth L. (1995), "Las Vegas algorithms for linear and integer programming when the dimension is small", Journal of the ACM , 42 (2): 488–499, doi:10.1145/201019.201036, MR   1409744, S2CID   6953625 .
  11. ACM Fellow citation.