Richard Hartley (scientist)

Last updated

Richard Hartley
Alma mater Australian National University
Stanford University
University of Toronto (1976, PhD)
Known for Computer Vision, Multiple-view geometry

Richard I. Hartley is an Australian computer scientist and a professor at the Australian National University, where he is head of the Computer Vision group in the Research School of Engineering.

Contents

Biography

In 1971, Hartley received a BSc degree from the Australian National University followed by MSc (1972) and PhD (1976) degrees in mathematics from the University of Toronto. He also obtained an MSc degree in computer science from Stanford University in 1983. [1]

Scientific work

His work is primarily devoted to the fields of Artificial intelligence, Image processing, and Computer vision. He is best known for his 2000 book Multiple View Geometry in computer vision, written with Andrew Zisserman, now in its second edition (2004). According to WorldCat, the book is held in 1428 libraries [2]

Publications

Hartley has published a wide variety of articles in computer science on the topics of computer vision and optimization. The following are his most highly cited works [3]

Related Research Articles

<span class="mw-page-title-main">Niklaus Wirth</span> Swiss computer scientist

Niklaus Emil Wirth is a Swiss computer scientist. He has designed several programming languages, including Pascal, and pioneered several classic topics in software engineering. In 1984, he won the Turing Award, generally recognized as the highest distinction in computer science, for developing a sequence of innovative computer languages.

<span class="mw-page-title-main">Theoretical computer science</span> Subfield of computer science and mathematics

Theoretical computer science (TCS) is a subset of general computer science and mathematics that focuses on mathematical aspects of computer science such as the theory of computation, lambda calculus, and type theory.

Random sample consensus (RANSAC) is an iterative method to estimate parameters of a mathematical model from a set of observed data that contains outliers, when outliers are to be accorded no influence on the values of the estimates. Therefore, it also can be interpreted as an outlier detection method. It is a non-deterministic algorithm in the sense that it produces a reasonable result only with a certain probability, with this probability increasing as more iterations are allowed. The algorithm was first published by Fischler and Bolles at SRI International in 1981. They used RANSAC to solve the Location Determination Problem (LDP), where the goal is to determine the points in the space that project onto an image into a set of landmarks with known locations.

In computer vision, the fundamental matrix is a 3×3 matrix which relates corresponding points in stereo images. In epipolar geometry, with homogeneous image coordinates, x and x′, of corresponding points in a stereo image pair, Fx describes a line on which the corresponding point x′ on the other image must lie. That means, for all pairs of corresponding points holds

Structure from motion (SfM) is a photogrammetric range imaging technique for estimating three-dimensional structures from two-dimensional image sequences that may be coupled with local motion signals. It is studied in the fields of computer vision and visual perception. In biological vision, SfM refers to the phenomenon by which humans can recover 3D structure from the projected 2D (retinal) motion field of a moving object or scene.

<span class="mw-page-title-main">Epipolar geometry</span> Geometry of stereo vision

Epipolar geometry is the geometry of stereo vision. When two cameras view a 3D scene from two distinct positions, there are a number of geometric relations between the 3D points and their projections onto the 2D images that lead to constraints between the image points. These relations are derived based on the assumption that the cameras can be approximated by the pinhole camera model.

In computer vision, the essential matrix is a matrix, that relates corresponding points in stereo images assuming that the cameras satisfy the pinhole camera model.

<span class="mw-page-title-main">Image rectification</span>

Image rectification is a transformation process used to project images onto a common image plane. This process has several degrees of freedom and there are many strategies for transforming images to the common plane. Image rectification is used in computer stereo vision to simplify the problem of finding matching points between images, and in geographic information systems to merge images taken from multiple perspectives into a common map coordinate system.

<span class="mw-page-title-main">Department of Engineering Science, University of Oxford</span> University department in Oxford, England

The Department of Engineering Science is the engineering department of the University of Oxford. It is part of the university's Mathematical, Physical and Life Sciences Division. The department was ranked 3rd best institute in the UK for engineering in the 2021 Research Excellence Framework.

<span class="mw-page-title-main">Bundle adjustment</span>

In photogrammetry and computer stereo vision, bundle adjustment is simultaneous refining of the 3D coordinates describing the scene geometry, the parameters of the relative motion, and the optical characteristics of the camera(s) employed to acquire the images, given a set of images depicting a number of 3D points from different viewpoints. Its name refers to the geometrical bundles of light rays originating from each 3D feature and converging on each camera's optical center, which are adjusted optimally according to an optimality criterion involving the corresponding image projections of all points.

The Hessian affine region detector is a feature detector used in the fields of computer vision and image analysis. Like other feature detectors, the Hessian affine detector is typically used as a preprocessing step to algorithms that rely on identifiable, characteristic interest points.

In computer vision, the trifocal tensor is a 3×3×3 array of numbers that incorporates all projective geometric relationships among three views. It relates the coordinates of corresponding points or lines in three views, being independent of the scene structure and depending only on the relative motion among the three views and their intrinsic calibration parameters. Hence, the trifocal tensor can be considered as the generalization of the fundamental matrix in three views. It is noted that despite the tensor being made up of 27 elements, only 18 of them are actually independent.

<span class="mw-page-title-main">Leonidas J. Guibas</span> Greek-American computer scientist

Leonidas John Guibas is the Paul Pigott Professor of Computer Science and Electrical Engineering at Stanford University. He heads the Geometric Computation group in the Computer Science Department.

Brian A. Barsky is a professor at the University of California, Berkeley, working in computer graphics and geometric modeling as well as in optometry and vision science. He is a Professor of Computer Science and Vision Science and an Affiliate Professor of Optometry. He is also a member of the Joint Graduate Group in Bioengineering, an inter-campus program, between UC Berkeley and UC San Francisco.

Andrew Zisserman is a British computer scientist and a professor at the University of Oxford, and a researcher in computer vision. As of 2014 he is affiliated with DeepMind.

Camera auto-calibration is the process of determining internal camera parameters directly from multiple uncalibrated images of unstructured scenes. In contrast to classic camera calibration, auto-calibration does not require any special calibration objects in the scene. In the visual effects industry, camera auto-calibration is often part of the "Match Moving" process where a synthetic camera trajectory and intrinsic projection model are solved to reproject synthetic content into video.

<span class="mw-page-title-main">Homography (computer vision)</span> Relation of two images with software

In the field of computer vision, any two images of the same planar surface in space are related by a homography. This has many practical applications, such as image rectification, image registration, or camera motion—rotation and translation—between two images. Once camera resectioning has been done from an estimated homography matrix, this information may be used for navigation, or to insert models of 3D objects into an image or video, so that they are rendered with the correct perspective and appear to have been part of the original scene.

<span class="mw-page-title-main">J. Michael Brady</span>

Sir John Michael Brady is an Emeritus professor of Oncological Imaging at the University of Oxford. He has been a Fellow of Keble College, Oxford since 1985 and was elected a foreign associate member of the French Academy of Sciences in 2015. He was formerly BP Professor of Information Engineering at Oxford from 1985 to 2010 and a Senior Research Scientist in the MIT Computer Science and Artificial Intelligence Laboratory (CSAIL) in Cambridge, Massachusetts, from 1980 to 1985.

<span class="mw-page-title-main">René Vidal</span>

René Vidal is a Chilean electrical engineer and computer scientist who is known for his research in machine learning, computer vision, medical image computing, robotics, and control theory. He is the Herschel L. Seder Professor of the Johns Hopkins Department of Biomedical Engineering, and the founding director of the Mathematical Institute for Data Science (MINDS).

References

  1. "Department of Computer Science - December 7, 2010 - Richard Hartley". Johns Hopkins Whiting School of Engineering. Retrieved 30 August 2014.
  2. WorldCat book entry
  3. "Richard Hartley".
  4. Multiple View Geometry in Computer Vision Second Edition. Retrieved May 20, 2009.