Hyperacuity

Last updated
Ewald Hering's model, published 1899, of how a Vernier acuity stimulus is coded by a receptor array. Receptors marked c signal a different position code along the horizontal direction from either the position a code or the position b code. Ewald Hering (1899) Fig 2.pdf
Ewald Hering's model, published 1899, of how a Vernier acuity stimulus is coded by a receptor array. Receptors marked c signal a different position code along the horizontal direction from either the position a code or the position b code.

The sharpness of our senses is defined by the finest detail we can discriminate. Visual acuity is measured by the smallest letters that can be distinguished on a chart and is governed by the anatomical spacing of the mosaic of sensory elements on the retina. Yet spatial distinctions can be made on a finer scale still: misalignment of borders can be detected with a precision up to 10 times better than visual acuity, as already shown by Ewald Hering in 1899. [1] This hyperacuity, transcending by far the size limits set by the retinal 'pixels', depends on sophisticated information processing in the brain.

Contents

How does hyperacuity differ from traditional acuity?

The best example of the distinction between acuity and hyperacuity comes from vision, for example when observing stars on a night sky. The first stage is the optical imaging of the outside world on the retina. Light impinges on the mosaic of receptor sense cells, rods and cones, which covers the retinal surface without gaps or overlap, just like the detecting pixels in the film plane of digital cameras. Each receptor accepts all the light reaching it but acts as a unit, representing a single location in visual space. This compartmentalization sets a limit to the decision whether an image came from a single or a double star (resolution). For a percept of separately articulated stars to emerge, the images of the two must be wide enough apart to leave at least one intervening pixel relatively unstimulated between them. This defines the resolution limit and the basis of visual acuity.

Acuity/Resolution versus Hyperacuity/Localization Top: Two stars imaged on the mosaic of retinal receptor cells can be resolved only if their separation leaves at least one intervening mosaic element with a detectably different intensity, otherwise the pattern is indistinguishable from a single elongated star. Bottom: Two targets can be localized relative to each other to values transcending the spacing of the mosaic units; the hyperacuity mechanism achieves this by identifying, with sub-pixel precision, the light center of each target, across all the pixels it covers Reslocfig.jpg
Acuity/Resolution versus Hyperacuity/Localization Top: Two stars imaged on the mosaic of retinal receptor cells can be resolved only if their separation leaves at least one intervening mosaic element with a detectably different intensity, otherwise the pattern is indistinguishable from a single elongated star. Bottom: Two targets can be localized relative to each other to values transcending the spacing of the mosaic units; the hyperacuity mechanism achieves this by identifying, with sub-pixel precision, the light center of each target, across all the pixels it covers

A quite different mechanism operates in hyperacuity, whose quintessential example and the one for which the word was initially coined, [2] [3] is vernier acuity: alignment of two edges or lines can be judged with a precision five or ten times better than acuity. In computer graphics the phrase "sub-pixel resolution" is sometimes used in discussions of anti-aliasing and geometrical superresolution. Though what is in fact involved is not resolution (is it one or two? – a qualitative distinction) but localization (exactly where? – a quantitative judgment) it captures the process. When an image spreads across several pixels, each with graded intensity response but only a single spatial value, the position of the image center can be located more exactly than the width of the pixel, much like the mean of a histogram can be calculated to a fraction of the bin width.

In the figure on the right, the retinal mosaic has superimposed on it, at top, the images of two stars at resolution limit when the intervening gap assures judgment that there are two stars and not a single elongated one. Shown below are the images of two separate short lines; the precision of the read-out of their location difference transcends the dimension of the mosaic elements.

Analysis of hyperacuity mechanism

Details of the neural apparatus for achieving hyperacuity still await discovery. That the hyperacuity apparatus involves signals from a range of individual receptor cells, usually in more than one location of the stimulus space, has implications concerning performance in these tasks. Low contrast, close proximity of neighboring stimuli (crowding), and temporal asynchrony of pattern components are examples of factors that cause reduced performance. [4] Of some conceptual interest are age changes [5] and susceptibility to perceptual learning [6] which can help in understanding underlying neural channeling.

Two basic algorithms have been proposed to explain mammalian visual hyperacuity: spatial, based on population firing rates, and temporal, based on temporal delays in response to miniature eye movements. While none of them gained empirical support so far, the plausibility of the former had been critically questioned by the discrete nature of neural firing [7]

The optics of the human eye are extremely simple, the main imaging component being a single element lens which can change its strength by muscular control. There is only limited facility for correction of many of the aberrations which are normally corrected in good quality instrumental optical systems. [8] Such a simple lens must inevitably have a significant amount of spherical aberration, which produces secondary lobes in the spread function. However, it has been found by experiment that light entering the pupil off-axis is less efficient in creating an image (the Stiles-Crawford effect), which has the effect of substantially reducing these unwanted side lobes. Also, the effects of diffraction limits can, with care, be used to partially compensate for the aberrations.

The retinal receptors are physically situated behind a neural layer carrying the post-retinal processing elements. Light cannot pass through this layer undistorted. In fact, measurements on the Modulation Transfer Function (MTF) suggest that the MTF degradations due to the diffusion through that neural layer are of a similar order as those due to the optics. By interplay of these different components it has been found that the overall optical quality, although poor compared to photographic optics, can remain tolerably near constant through a considerable range of pupil diameters and light levels.

When presented with colored information the optical imperfections are particularly great. The optics have residual uncorrected chromatic aberration of nearly 2 dioptres from extreme red to extreme blue/violet, mainly in the green to blue/violet region. Ophthalmologists have for many decades used this large change of focus through the spectrum in the process of providing correction spectacles. This means that such corrections can be as simple as the eye lens itself.

In addition, this large chromatic aberration has also been used to advantage within the make up of the eye itself. Instead of having the three primary colors (red, green & blue) to manipulate, nature has used this gross chromatic shift to provide a cortical visual function which is based on three sets of color opponency instead of three basic primary colors. [9] These are red / green, yellow / blue and black / white, this black / white being synonymous with brightness. Then, by using just one very high resolution opponency between red & green primaries, nature cleverly uses a mean of these two colors (i.e. yellow ), together with very low resolution blue to create a background color wash capability. In turn (by using the hyperacuity capability on the low resolution opponency) this can also serve as the source of perception of 3D depth.

The human eye has a roughly hexagonal matrix of photodetectors. [10] There is now considerable evidence that such a matrix layout provides optimum efficiency of information transfer. A number of other workers have considered using hexagonal matrices, but they then seem to subscribe to a mathematical approach and axes at 60 degrees differential orientation. In turn this must then make use of complex numbers. Overington & his team sought (and found), instead, a way to approximate to a hexagonal matrix, while at the same time retaining a conventional Cartesian layout for processing.

Although there are many and varied spatial interactions evident in the early neural networks of the human visual system, only a few are of great importance in high fidelity information sensing. The rest are predominantly associated with processes such as local adaptation. It has therefore been found that the most important interactions are of very local extent, but it is the subtleties of usage of these interactions which seem most important. For hexagonal matrices a single ring of six receptors surrounding an addressed pixel is the simplest symmetrical layout. The general finding from primate receptive field studies is that any such local group yields no output for a uniform input illumination. So this is essentially similar to one of the classical Laplacian receptive fields for square arrays - that with weightings of -1 on each side and -0.5 on each corner. The only difference is an aspect ratio of 8:7.07 (or approximately 8:7 to within 1%). Very useful further evidence of the processes going on in his area comes from the electron-microscopy studies of Kolb [11] These clearly show the neural structures which lead to difference signals being transmitted further. If one combines a point spread function having a Gaussian form and having an S.D. of 1.3 'pixels' with a single ring Laplacian - type operator, the resultant is a function with very similar properties to a DOG function as discussed by Marr. [12]

It is normally assumed, both in computer image processing and in visual science, that a local excitatory / inhibitory process is effectively a second differencing process. However, there seems to be strong psychophysical evidence for human vision that it is first differences which control human visual performance. It is necessary for the positive & negative parts of all outputs from Laplacian-like neurones to be separated for sending onwards to the cortex, since it is impossible to transmit negative signals. This means that each neurone of this type must be considered to be a set of six dipoles, such that each surround inhibition can only cancel its own portion of the central stimulation. Such a separation of positive and negative components is totally compatible with retinal physiology and is one possible function for the known pair of midget bipolar channels for each receptor. [13]

The basic evidence for orientation sensing in human vision is that it appears to be carried out (in Area 17 of the striate cortex) by banks of neurones at fairly widely spaced orientations. [14] The neurones as measured have characteristically elliptical receptive fields. [15] However, both the actual interval between the orientations and the exact form & aspect ratio of the elliptical fields is open to question, but at the same time the said receptive fields have to have been compounded with the midget receptive fields at the retina. Yet again, for probe measurements of 'single neurone' performance, the receptive field measured includes the effects of all stages of optical & neural processing that have gone before.

For orientation specific units operating on a hexagonal matrix, it makes most sense to have them with their primary & secondary axes occurring every 30 degrees of orientation. This 30 degree separation of orientations agrees with angular spacing of such units deduced to be desirable by John Canny from a mathematical approach. [16] In the absence of specific details, it seemed that a roughly best compromise between computational efficiency and simplicity on the one hand and adequate orientation al tuning on the other should be of extent 5 x 1 pixels. This again agrees with that independently suggested by Canny and also observed in primate vision studies by other researchers. The receptive field units have orientation tuning functions which bear a satisfying resemblance to the orientation tuning functions established for vision by psychophysical tests.

There is the possibility of recombining the partial difference functions arriving at the cortex in two ways. [17] It is possible to consider analysis of a second difference map - by searching for zero crossings, which was most popular until the mid-1980s. Alternatively one can sense local peaks in the first difference map, which has become increasingly popular since then. This latter then depends on finding the position of the peak of the edge image by a 3 x 1 analysis & quadratic curve fitting. In either case it has been shown that the edge position can be located to something better than 0.1 pixels because of the very broad spread of the edge due to the poor optical image, while it has also been shown that, by equally simple arithmetic, the local edge orientation can be derived to better than 1 degree. Furthermore, the interplay of first and second difference data provides very powerful means of analyzing motion, stereo, color, texture & other scene properties.

Hyperacuity in various sense modalities

The distinction between resolving power or acuity, literally sharpness, which depends on the spacing of the individual receptors through which the outside world is sampled, and the ability to identify individual locations in the sensory space is universal among modalities. There are many other examples where the organism's performance substantially surpasses the spacing of the concerned receptor cell population. The normal human has only three kinds of color receptors in the retina, yet in color vision, by subtly weighing and comparing their relative output, one can detect thousand of hues. Braille reading involves hyperacuity among touch receptors in the fingertips. [18] We can hear many more different tones than there are hair cells in the cochlea; pitch discrimination, without which a violin could not be played in tune, is a hyperacuity. [19] Hyperacuity has been identified in many animal species, for example in the detection of prey by the electric fish, [20] echolocation in the bat, [21] and in the ability of rodents to localize objects based on mechanical deformations of their whiskers. [22]

Clinical applications

In clinical vision tests, [23] hyperacuity has a special place because its processing is at the interfaces of the eye's optics, retinal functions, activation of the primary visual cortex and the perceptual apparatus. In particular, the determination of normal stereopsis is a hyperacuity task. Hyperacuity perimetry is used in clinical trials evaluating therapies for retinal degenerative changes. [24]

See also

Related Research Articles

<span class="mw-page-title-main">Retina</span> Part of the eye

The retina is the innermost, light-sensitive layer of tissue of the eye of most vertebrates and some molluscs. The optics of the eye create a focused two-dimensional image of the visual world on the retina, which then processes that image within the retina and sends nerve impulses along the optic nerve to the visual cortex to create visual perception. The retina serves a function which is in many ways analogous to that of the film or image sensor in a camera.

<span class="mw-page-title-main">Ewald Hering</span> German physiologist

Karl Ewald Konstantin Hering was a German physiologist who did much research in color vision, binocular perception, eye movements, and hyperacuity. He proposed opponent color theory in 1892.

<span class="mw-page-title-main">Eye</span> Organ that detects light and converts it into electro-chemical impulses in neurons

An eye is a sensory organ that allows an organism to perceive visual information. It detects light and converts it into electro-chemical impulses in neurons (neurones). It is part of an organism's visual system.

<span class="mw-page-title-main">Adaptive optics</span> Technique used in optical systems

Adaptive optics (AO) is a technique of precisely deforming a mirror in order to compensate for light distortion. It is used in astronomical telescopes and laser communication systems to remove the effects of atmospheric distortion, in microscopy, optical fabrication and in retinal imaging systems to reduce optical aberrations. Adaptive optics works by measuring the distortions in a wavefront and compensating for them with a device that corrects those errors such as a deformable mirror or a liquid crystal array.

<span class="mw-page-title-main">Color vision</span> Ability to perceive differences in light frequency

Color vision, a feature of visual perception, is an ability to perceive differences between light composed of different frequencies independently of light intensity.

<span class="mw-page-title-main">Visual system</span> Body parts responsible for vision

The visual system is the physiological basis of visual perception. The system detects, transduces and interprets information concerning light within the visible range to construct an image and build a mental model of the surrounding environment. The visual system is associated with the eye and functionally divided into the optical system and the neural system.

<span class="mw-page-title-main">Ommatidium</span> Component of compound eyes of arthropods

The compound eyes of arthropods like insects, crustaceans and millipedes are composed of units called ommatidia. An ommatidium contains a cluster of photoreceptor cells surrounded by support cells and pigment cells. The outer part of the ommatidium is overlaid with a transparent cornea. Each ommatidium is innervated by one axon bundle and provides the brain with one picture element. The brain forms an image from these independent picture elements. The number of ommatidia in the eye depends upon the type of arthropod and range from as low as 5 as in the Antarctic isopod Glyptonotus antarcticus, or a handful in the primitive Zygentoma, to around 30,000 in larger Anisoptera dragonflies and some Sphingidae moths.

<span class="mw-page-title-main">Visual acuity</span> Clarity of vision

Visual acuity (VA) commonly refers to the clarity of vision, but technically rates an animal's ability to recognize small details with precision. Visual acuity depends on optical and neural factors. Optical factors of the eye influence the sharpness of an image on its retina. Neural factors include the health and functioning of the retina, of the neural pathways to the brain, and of the interpretative faculty of the brain.

<span class="mw-page-title-main">Fovea centralis</span> Small pit in the retina of the eye responsible for all central vision

The fovea centralis is a small, central pit composed of closely packed cones in the eye. It is located in the center of the macula lutea of the retina.

The receptive field, or sensory space, is a delimited medium where some physiological stimuli can evoke a sensory neuronal response in specific organisms.

<span class="mw-page-title-main">Spatial frequency</span> Characteristic of any structure that is periodic across a position in space

In mathematics, physics, and engineering, spatial frequency is a characteristic of any structure that is periodic across position in space. The spatial frequency is a measure of how often sinusoidal components of the structure repeat per unit of distance.

<span class="mw-page-title-main">Defocus aberration</span> Quality of an image being out of focus

In optics, defocus is the aberration in which an image is simply out of focus. This aberration is familiar to anyone who has used a camera, videocamera, microscope, telescope, or binoculars. Optically, defocus refers to a translation of the focus along the optical axis away from the detection surface. In general, defocus reduces the sharpness and contrast of the image. What should be sharp, high-contrast edges in a scene become gradual transitions. Fine detail in the scene is blurred or even becomes invisible. Nearly all image-forming optical devices incorporate some form of focus adjustment to minimize defocus and maximize image quality.

<span class="mw-page-title-main">Contrast (vision)</span> Visible difference in brightness or color

Contrast is the difference in luminance or color that makes an object visible against a background of different luminance or color. The human visual system is more sensitive to contrast than to absolute luminance; thus, we can perceive the world similarly despite significant changes in illumination throughout the day or across different locations.

A visual prosthesis, often referred to as a bionic eye, is an experimental visual device intended to restore functional vision in those with partial or total blindness. Many devices have been developed, usually modeled on the cochlear implant or bionic ear devices, a type of neural prosthesis in use since the mid-1980s. The idea of using electrical current to provide sight dates back to the 18th century, discussed by Benjamin Franklin, Tiberius Cavallo, and Charles LeRoy.

Visual perception is the ability to interpret the surrounding environment through photopic vision, color vision, scotopic vision, and mesopic vision, using light in the visible spectrum reflected by objects in the environment. This is different from visual acuity, which refers to how clearly a person sees. A person can have problems with visual perceptual processing even if they have 20/20 vision.

Feature detection is a process by which the nervous system sorts or filters complex natural stimuli in order to extract behaviorally relevant cues that have a high probability of being associated with important objects or organisms in their environment, as opposed to irrelevant background or noise.

Preferential hyperacuity perimetry (PHP) is a psychophysical test used to identify and quantify visual abnormalities such as metamorphopsia and scotoma. It is a type of perimetry.

<span class="mw-page-title-main">Photovoltaic retinal prosthesis</span>

Photovoltaic retinal prosthesis is a technology for restoration of sight to patients blinded by degenerative retinal diseases, such as retinitis pigmentosa and age-related macular degeneration (AMD), when patients lose the 'image capturing' photoreceptors, but neurons in the 'image-processing' inner retinal layers are relatively well-preserved. This subretinal prosthesis is designed to restore sight by electrically stimulating the surviving inner retinal neurons, primarily the bipolar cells. Photovoltaic retinal implants are completely wireless and powered by near-infrared illumination (880 nm) projected from the augmented-reality glasses. Lack of trans-scleral cable greatly simplifies the implantation procedure compared to other retinal implants. Optical activation of the photovoltaic pixels allows scaling the implants to thousands of electrodes and retains natural coupling of the eye movements to visual perception. Studies in rats with retinal degeneration demonstrated that prosthetic vision with such subretinal implants preserves many features of natural vision, including flicker fusion at high frequencies (>20 Hz), adaptation to static images, antagonistic center-surround organization and non-linear summation of subunits in receptive fields, providing high spatial resolution.

<span class="mw-page-title-main">Vernier acuity</span>

Vernier acuity is a type of visual acuity – more precisely of hyperacuity – that measures the ability to discern a disalignment among two line segments or gratings. A subject's vernier acuity is the smallest visible offset between the stimuli that can be detected. Because the disalignments are often much smaller than the diameter and spacing of retinal receptors, vernier acuity requires neural processing and "pooling" to detect it. Because vernier acuity exceeds acuity by far, the phenomenon has been termed hyperacuity. Vernier acuity develops rapidly during infancy and continues to slowly develop throughout childhood. At approximately three to twelve months old, it surpasses grating acuity in foveal vision in humans. However, vernier acuity decreases more quickly than grating acuity in peripheral vision. Vernier acuity was first explained by Ewald Hering in 1899, based on earlier data by Alfred Volkmann in 1863 and results by Ernst Anton Wülfing in 1892.

<span class="mw-page-title-main">Retinal mosaic</span> Pattern of neurons on the retina

Retinal mosaic is the name given to the distribution of any particular type of neuron across any particular layer in the retina. Typically such distributions are somewhat regular; it is thought that this is so that each part of the retina is served by each type of neuron in processing visual information.

References

  1. 1 2 Strasburger, Hans; Huber, Jörg; Rose, David (2018). "Ewald Hering (1899) On the Limits of Visual Acuity: A Translation and Commentary. With a Supplement on Alfred Volkmann (1863) Physiological Investigations in the Field of Optics". i-Perception. 9 (3): 204166951876367. doi:10.1177/2041669518763675. PMC   5990881 . PMID   29899967.
  2. Westheimer, Gerald (1975). "Editorial: Visual acuity and hyperacuity". Investigative Ophthalmology. 14 (8): 570–2. PMID   1150397.
  3. Westheimer, G. (1975-08-01). "Editorial: Visual acuity and hyperacuity". Investigative Ophthalmology & Visual Science. 14 (8): 570–572. ISSN   1552-5783. PMID   1150397.
  4. Westheimer, G (2008). "Hyperacuity". In Squire, L. A. (ed.). Encyclopedia of Neuroscience. Oxford: Academic Press.[ page needed ]
  5. Wang, Yi-Zhong; Morale, Sarah E.; Cousins, Robert; Birch, Eileen E. (2009). "Course of Development of Global Hyperacuity over Lifespan". Optometry and Vision Science. 86 (6): 695–700. doi:10.1097/OPX.0b013e3181a7b0ff. PMC   2733828 . PMID   19430324.
  6. Crist, RE; Kapadia, MK; Westheimer, G; Gilbert, CD (1997). "Perceptual learning of spatial localization: Specificity for orientation, position, and context". Journal of Neurophysiology. 78 (6): 2889–94. doi:10.1152/jn.1997.78.6.2889. PMID   9405509.
  7. Rucci, M., Ahissar, E., and Burr, D. (2018) Temporal coding of visual space. Trends in cognitive sciences 22, 883-895 doi: 10.1016/j.tics.2018.07.009
  8. Computer Vision: A unified, biologically-inspired approach by Ian Overington Pages 7 to 8 and 31 to 34 Published by Elsevier North Holland 1992 ISBN   0 444 889728
  9. Computer Vision: A unified, biologically-inspired approach by Ian Overington Pages 8 to 15 and 286 - 288 Published by Elsevier North Holland 1992 ISBN   0 444 889728
  10. Computer Vision: A unified, biologically-inspired approach by Ian Overington Pages 34-36 Published by Elsevier North Holland 1992 ISBN   0 444 889728
  11. Organization of the outer plexiform layer of the primate retina: electron microscopy of Golgi-impregnated sells by H. Kolb Philosophical Transactions of the Royal Society Lond B, Soc Lond B Volume 258 page 261 1970
  12. Vision by David Marr Published by Freeman, San Francisco 1988
  13. Computer Vision: A unified, biologically-inspired approach by Ian Overington Pages 45-46 Published by Elsevier North Holland 1992 ISBN   0 444 889728
  14. Plasticity of ocular dominance columns in monkey striate cortex by D.H. Hubel, T.N Wiesel and S. LeVay, Philosophical Transactions of the Royal Society B, Volume 278, Page 377, 1977
  15. Computer Vision: A unified, biologically-inspired approach by Ian Overington Pages 46 to 49 Published by Elsevier North Holland 1992 ISBN   0 444 889728
  16. Finding edges and lines in images by J.F. Canny, MIT Tech Report No. 720, Cambridge, Mass USA 1983
  17. Computer Vision: A unified, biologically-inspired approach by Ian Overington Pages 61 to 64 and 68 to 71 Published by Elsevier North Holland 1992 ISBN   0 444 889728
  18. Loomis, JM (1979). "An investigation of tactile hyperacuity". Sensory Processes. 3 (4): 289–302. PMID   262782.
  19. Altes, Richard A. (1989). "Ubiquity of hyperacuity". The Journal of the Acoustical Society of America. 85 (2): 943–52. Bibcode:1989ASAJ...85..943A. doi:10.1121/1.397566. PMID   2926010.
  20. Heiligenberg, W (1991). "Sensory control of behavior in electric fish". Current Opinion in Neurobiology. 1 (4): 633–7. doi:10.1016/S0959-4388(05)80041-8. PMID   1822309. S2CID   38440905.
  21. Simmons, J. A.; Ferragamo, M. J.; Sanderson, M. I. (2003). "Echo delay versus spectral cues for temporal hyperacuity in the big brown bat, Eptesicus fuscus". Journal of Comparative Physiology A. 189 (9): 693–702. doi:10.1007/s00359-003-0444-9. PMID   12879356. S2CID   11442783.
  22. Knutsen, P. M.; Pietr, M; Ahissar, E (2006). "Haptic Object Localization in the Vibrissal System: Behavior and Performance". Journal of Neuroscience. 26 (33): 8451–64. doi: 10.1523/JNEUROSCI.1516-06.2006 . PMC   6674338 . PMID   16914670.
  23. Whitaker, David; Buckingham, Terry (1987). "Theory and Evidence for a Clinical Hyperacuity Test". Ophthalmic and Physiological Optics. 7 (4): 431–5. doi:10.1111/j.1475-1313.1987.tb00774.x. PMID   3454922. S2CID   27594676.
  24. Goldstein, M; Loewenstein, A; Barak, A; Pollack, A; Bukelman, A; Katz, H; Springer, A; Schachat, AP; et al. (2005). "Results of a Multicenter Clinical Trial to Evaluate the Preferential Hyperacuity Perimeter for Detection of Age-Related Macular Degeneration". Retina. 25 (3): 296–303. doi:10.1097/00006982-200504000-00008. PMID   15805906. S2CID   13711267.