Motion perception

(Redirected from Aperture problem)

Motion perception is the process of inferring the speed and direction of elements in a scene based on visual, vestibular and proprioceptive inputs. Although this process appears straightforward to most observers, it has proven to be a difficult problem from a computational perspective, and difficult to explain in terms of neural processing.

The dorsal stream (green) and ventral stream (purple) are shown. They originate from a common source in visual cortex. The dorsal stream is responsible for detection of location and motion.

Motion perception is studied by many disciplines, including psychology (i.e. visual perception), neurology, neurophysiology, engineering, and computer science.

Neuropsychology

edit

The inability to perceive motion is called akinetopsia and it may be caused by a lesion to cortical area V5 in the extrastriate cortex. Neuropsychological studies of a patient who could not see motion, seeing the world in a series of static "frames" instead, suggested that visual area V5 in humans[1] is homologous to motion processing area V5/MT in primates.[2][3][4]

First-order motion perception

edit
 
Example of Beta movement, often confused with phi phenomenon, in which a succession of still images gives the illusion of a moving ball[5]

When two or more stimuli are alternatively switched on and off, they can produce two distinct motion perceptions. The first, known as beta movement, is demonstrated in the yellow-ball figure and forms the basis for electronic news ticker displays. However, at faster alternation rates, and when the distance between the stimuli is optimal, an illusory "object"—matching the background color—appears to move between the stimuli, alternately occluding them. This phenomenon is called the phi phenomenon and is often described as an example of "pure" motion detection, uncontaminated by form cues, unlike beta movement.[5] Nevertheless, this description is somewhat paradoxical since creating such motion without figural percepts is impossible.

 
Demonstration of phi phenomenon using two black bars (SOA = 102 ms, ISI = −51 ms)

The phi phenomenon has been referred to as "first-order" motion perception. Werner E. Reichardt and Bernard Hassenstein have modelled it in terms of relatively simple "motion sensors" in the visual system, that have evolved to detect a change in luminance at one point on the retina and correlate it with a change in luminance at a neighbouring point on the retina after a short delay. Sensors that are proposed to work this way have been referred to as either Hassenstein-Reichardt detectors after the scientists Bernhard Hassenstein and Werner Reichardt, who first modelled them,[6] motion-energy sensors,[7] or Elaborated Reichardt Detectors.[8] These sensors are described as detecting motion by spatio-temporal correlation and are considered by some to be plausible models for how the visual system may detect motion. (Although, again, the notion of a "pure motion" detector suffers from the problem that there is no "pure motion" stimulus, i.e. a stimulus lacking perceived figure/ground properties). There is still considerable debate regarding the accuracy of the model and exact nature of this proposed process. It is not clear how the model distinguishes between movements of the eyes and movements of objects in the visual field, both of which produce changes in luminance on points on the retina.

Second-order motion perception

edit

Second-order motion is when the moving contour is defined by contrast, texture, flicker or some other quality that does not result in an increase in luminance or motion energy in the Fourier spectrum of the stimulus.[9][10] There is much evidence to suggest that early processing of first- and second-order motion is carried out by separate pathways.[11] Second-order mechanisms have poorer temporal resolution and are low-pass in terms of the range of spatial frequencies to which they respond. (The notion that neural responses are attuned to frequency components of stimulation suffers from the lack of a functional rationale and has been generally criticized by G. Westheimer (2001) in an article called "The Fourier Theory of Vision.") Second-order motion produces a weaker motion aftereffect unless tested with dynamically flickering stimuli.[12]

The aperture problem

edit
 
The aperture problem. The grating appears to be moving down and to the right, perpendicular to the orientation of the bars. But it could be moving in many other directions, such as only down, or only to the right. It is impossible to determine unless the ends of the bars become visible in the aperture.

The motion direction of a contour is ambiguous, because the motion component parallel to the line cannot be inferred based on the visual input. This means that a variety of contours of different orientations moving at different speeds can cause identical responses in a motion sensitive neuron in the visual system.

Motion integration

edit

Some have speculated that, having extracted the hypothesized motion signals (first- or second-order) from the retinal image, the visual system must integrate those individual local motion signals at various parts of the visual field into a 2-dimensional or global representation of moving objects and surfaces. (It is not clear how this 2D representation is then converted into the perceived 3D percept) Further processing is required to detect coherent motion or "global motion" present in a scene.[13]

The ability of a subject to detect coherent motion is commonly tested using motion coherence discrimination tasks. For these tasks, dynamic random-dot patterns (also called random dot kinematograms) are used that consist in 'signal' dots moving in one direction and 'noise' dots moving in random directions. The sensitivity to motion coherence is assessed by measuring the ratio of 'signal' to 'noise' dots required to determine the coherent motion direction. The required ratio is called the motion coherence threshold.

Motion in depth

edit

As in other aspects of vision, the observer's visual input is generally insufficient to determine the true nature of stimulus sources, in this case their velocity in the real world. In monocular vision for example, the visual input will be a 2D projection of a 3D scene. The motion cues present in the 2D projection will by default be insufficient to reconstruct the motion present in the 3D scene. Put differently, many 3D scenes will be compatible with a single 2D projection. The problem of motion estimation generalizes to binocular vision when we consider occlusion or motion perception at relatively large distances, where binocular disparity is a poor cue to depth. This fundamental difficulty is referred to as the inverse problem.[14]

Nonetheless, some humans do perceive motion in depth. There are indications that the brain uses various cues, in particular temporal changes in disparity as well as monocular velocity ratios, for producing a sensation of motion in depth.[15] Two different binocular cues of the perception motion in depth are hypothesized: Inter-ocular velocity difference (IOVD) and changing disparity (CD) over time. Motion in depth based on inter-ocular velocity differences can be tested using dedicated binocularly uncorrelated random-dot kinematograms.[16] Study results indicate that the processing of these two binocular cues – IOVD and CD – may use fundamentally different low-level stimulus features, which may be processed jointly that later stages.[17][18] Additionally, as monocular cue, also the changing size of retinal images contributes to motion in depth detection.

Detection and discrimination of motion can be improved by training with long-term results. Participants trained to detect the movements of dots on a screen in only one direction become particularly good at detecting small movements in the directions around that in which they have been trained. This improvement was still present 10 weeks later. However perceptual learning is highly specific. For example, the participants show no improvement when tested around other motion directions, or for other sorts of stimuli.[19]

Cognitive map

edit

A cognitive map is a type of mental representation which serves an individual to acquire, code, store, recall, and decode information about the relative locations and attributes of phenomena in their spatial environment. [20][21] Place cells work with other types of neurons in the hippocampus and surrounding regions of the brain to perform this kind of spatial processing,[22] but the ways in which they function within the hippocampus are still being researched.[23]

Many species of mammals can keep track of spatial location even in the absence of visual, auditory, olfactory, or tactile cues, by integrating their movements—the ability to do this is referred to in the literature as path integration. A number of theoretical models have explored mechanisms by which path integration could be performed by neural networks. In most models, such as those of Samsonovich and McNaughton (1997)[24] or Burak and Fiete (2009),[25] the principal ingredients are (1) an internal representation of position, (2) internal representations of the speed and direction of movement, and (3) a mechanism for shifting the encoded position by the right amount when the animal moves. Because cells in the Medial Entorhinal Cortex (MEC) encode information about position (grid cells[26]) and movement (head direction cells and conjunctive position-by-direction cells[27]), this area is currently viewed as the most promising candidate for the place in the brain where path integration occurs.

Neurophysiology

edit

Motion sensing using vision is crucial for detecting a potential mate, prey, or predator, and thus it is found both in vertebrates and invertebrates vision throughout a wide variety of species, although it is not universally found in all species. In vertebrates, the process takes place in retina and more specifically in retinal ganglion cells, which are neurons that receive input from bipolar cells and amacrine cells on visual information and process output to higher regions of the brain including, thalamus, hypothalamus, and mesencephalon.

The study of directionally selective units began with a discovery of such cells in the cerebral cortex of cats by David Hubel and Torsten Wiesel in 1959. Following the initial report, an attempt to understand the mechanism of directionally selective cells was pursued by Horace B. Barlow and William R. Levick in 1965.[28] Their in-depth experiments in rabbit's retina expanded the anatomical and physiological understanding of the vertebrate visual system and ignited the interest in the field. Numerous studies that followed thereafter have unveiled the mechanism of motion sensing in vision for the most part. Alexander Borst and Thomas Euler's 2011 review paper, "Seeing Things in Motion: Models, Circuits and Mechanisms".[29] discusses certain important findings from the early discoveries to the recent work on the subject, coming to the conclusion of the current status of the knowledge.

Direction selective (DS) cells

edit

Direction selective (DS) cells in the retina are defined as neurons that respond differentially to the direction of a visual stimulus. According to Barlow and Levick (1965), the term is used to describe a group of neurons that "gives a vigorous discharge of impulses when a stimulus object is moved through its receptive field in one direction."[28] This direction in which a set of neurons respond most strongly to is their "preferred direction". In contrast, they do not respond at all to the opposite direction, "null direction". The preferred direction is not dependent on the stimulus—that is, regardless of the stimulus' size, shape, or color, the neurons respond when it is moving in their preferred direction, and do not respond if it is moving in the null direction. There are three known types of DS cells in the vertebrate retina of the mouse, ON/OFF DS ganglion cells, ON DS ganglion cells, and OFF DS ganglion cells. Each has a distinctive physiology and anatomy. Analogous directionally selective cells are not thought to exist in the primate retina.[30]

ON/OFF DS ganglion cells

edit

ON/OFF DS ganglion cells act as local motion detectors. They fire at the onset and offset of a stimulus (a light source). If a stimulus is moving in the direction of the cell's preference, it will fire at the leading and the trailing edge. Their firing pattern is time-dependent and is supported by the Reichardt-Hassenstain model, which detects spatiotemporal correlation between the two adjacent points. The detailed explanation of the Reichardt-Hassenstain model will be provided later in the section. The anatomy of ON/OFF cells is such that the dendrites extend to two sublaminae of the inner plexiform layer and make synapses with bipolar and amacrine cells. They have four subtypes, each with its own preference for direction.

ON DS ganglion cells

edit

Unlike ON/OFF DS ganglion cells that respond both to the leading and the trailing edge of a stimulus, ON DS ganglion cells are responsive only to a leading edge. The dendrites of ON DS ganglion cells are monostratified and extend into the inner sublamina of the inner plexiform layer. They have three subtypes with different directional preferences.

OFF DS ganglion cells

edit

OFF DS ganglion cells act as a centripetal motion detector, and they respond only to the trailing edge of a stimulus. They are tuned to upward motion of a stimulus. The dendrites are asymmetrical and arbor in to the direction of their preference.[29]

DS cells in insects

edit

The first DS cells in invertebrates were found in flies in a brain structure called the lobula plate. The lobula plate is one of the three stacks of the neuropils in the fly's optic lobe. The "tangential cells" of the lobula plate composed of roughly about 50 neurons, and they arborize extensively in the neuropile. The tangential cells are known to be directionally selective with distinctive directional preference. One of which is Horizontally Sensitive (HS) cells, such as the H1 neuron, that depolarize most strongly in response to stimulus moving in a horizontal direction (preferred direction). On the other hand, they hyperpolarize when the direction of motion is opposite (null direction). Vertically Sensitive (VS) cells are another group of cells that are most sensitive to vertical motion. They depolarize when a stimulus is moving downward and hyperpolarize when it is moving upward. Both HS and VS cells respond with a fixed preferred direction and a null direction regardless of the color or contrast of the background or the stimulus.

The Hassenstein-Reichardt model

edit
 
Reichardt model

It is now known that motion detection in vision is based on the Hassenstein-Reichardt detector model.[31] This is a model used to detect correlation between the two adjacent points. It consists of two symmetrical subunits. Both subunits have a receptor that can be stimulated by an input (light in the case of visual system). In each subunit, when an input is received, a signal is sent to the other subunit. At the same time, the signal is delayed in time within the subunit, and after the temporal filter, is then multiplied by the signal received from the other subunit. Thus, within each subunit, the two brightness values, one received directly from its receptor with a time delay and the other received from the adjacent receptor, are multiplied. The multiplied values from the two subunits are then subtracted to produce an output. The direction of selectivity or preferred direction is determined by whether the difference is positive or negative. The direction which produces a positive outcome is the preferred direction.

In order to confirm that the Reichardt-Hassenstein model accurately describes the directional selectivity in the retina, the study was conducted using optical recordings of free cytosolic calcium levels after loading a fluorescent indicator dye into the fly tangential cells. The fly was presented uniformly moving gratings while the calcium concentration in the dendritic tips of the tangential cells was measured. The tangential cells showed modulations that matched the temporal frequency of the gratings, and the velocity of the moving gratings at which the neurons respond most strongly showed a close dependency on the pattern wavelength. This confirmed the accuracy of the model both at the cellular and the behavioral level.[32]

Although the details of the Hassenstein-Reichardt model have not been confirmed at an anatomical and physiological level, the site of subtraction in the model is now being localized to the tangential cells. When depolarizing current is injected into the tangential cell while presenting a visual stimulus, the response to the preferred direction of motion decreased, and the response to the null direction increased. The opposite was observed with hyperpolarizing current. The T4 and T5 cells, which have been selected as a strong candidate for providing input to the tangential cells, have four subtypes that each project into one of the four strata of the lobula plate that differ in the preferred orientation.[29]

DS cells in vertebrates

edit

One of the early works on DS cells in vertebrates was done on the rabbit retina by H. Barlow and W. Levick in 1965. Their experimental methods include variations to the slit-experiments and recording of the action potentials in the rabbit retina. The basic set-up of the slit experiment was they presented a moving black-white grating through a slit of various widths to a rabbit and recorded the action potentials in the retina. This early study had a large impact on the study of DS cells by laying down the foundation for later studies. The study showed that DS ganglion cells derive their property from the basis of sequence-discriminating activity of subunits, and that this activity may be the result of inhibitory mechanism in response to the motion of image in the null direction. It also showed that the DS property of retinal ganglion cells is distributed over the entire receptive field, and not limited to specific zones. Direction selectivity is contained for two adjacent points in the receptive field separated by as small as 1/4°, but selectivity decreased with larger separations. They used this to support their hypothesis that discrimination of sequences gives rise to direction selectivity because normal movement would activate adjacent points in a succession.[28]

Molecular identity and structure of DS cells in mice

edit

ON/OFF DS ganglion cells can be divided into 4 subtypes differing in their directional preference, ventral, dorsal, nasal, or temporal. The cells of different subtypes also differ in their dendritic structure and synaptic targets in the brain. The neurons that were identified to prefer ventral motion were also found to have dendritic projections in the ventral direction. Also, the neurons that prefer nasal motion had asymmetric dendritic extensions in the nasal direction. Thus, a strong association between the structural and functional asymmetry in ventral and nasal direction was observed. With a distinct property and preference for each subtype, there was an expectation that they could be selectively labeled by molecular markers. The neurons that were preferentially responsive to vertical motion were indeed shown to be selectively expressed by a specific molecular marker. However, molecular markers for other three subtypes have not been yet found.[33]

Neural mechanism: starburst amacrine cells

edit

The direction selective (DS) ganglion cells receive inputs from bipolar cells and starburst amacrine cells. The DS ganglion cells respond to their preferred direction with a large excitatory postsynaptic potential followed by a small inhibitory response. On the other hand, they respond to their null direction with a simultaneous small excitatory postsynaptic potential and a large inhibitory postsynaptic potential. Starburst amacrine cells have been viewed as a strong candidate for direction selectivity in ganglion cells because they can release both GABA and Ach. Their dendrites branch out radiantly from a soma, and there is a significant dendritic overlap. Optical measurements of Ca2+ concentration showed that they respond strongly to the centrifugal motion (the outward motion from the soma to the dendrites), while they don't respond well to the centripetal motion (the inward motion from the dendritic tips to the soma). When the starburst cells were ablated with toxins, direction selectivity was eliminated. Moreover, their release of neurotransmitters itself, specifically calcium ions, reflect direction selectivity, which may be presumably attributed to the synaptic pattern. The branching pattern is organized such that certain presynaptic input will have more influence on a given dendrite than others, creating a polarity in excitation and inhibition. Further evidence suggests that starburst cells release inhibitory neurotransmitters, GABA onto each other in a delayed and prolonged manner. This accounts for the temporal property of inhibition.[29]

In addition to spatial offset due to GABAergic synapses, the important role of chloride transporters has started to be discussed. The popular hypothesis is that starburst amacrine cells differentially express chloride transporters along the dendrites. Given this assumption, some areas along the dendrite will have a positive chloride-ion equilibrium potential relative to the resting potential while others have a negative equilibrium potential. This means that GABA at one area will be depolarizing and at another area hyperpolarizing, accounting for the spatial offset present between excitation and inhibition.[34]

Recent research (published March 2011) relying on serial block-face electron microscopy (SBEM) has led to identification of the circuitry that influences directional selectivity. This new technique provides detailed images of calcium flow and anatomy of dendrites of both starburst amacrine (SAC) and DS ganglion cells. By comparing the preferred directions of ganglion cells with their synapses on SAC's, Briggman et al. provide evidence for a mechanism primarily based on inhibitory signals from SAC's[35] based on an oversampled serial block-face scanning electron microscopy study of one sampled retina, that retinal ganglion cells may receive asymmetrical inhibitory inputs directly from starburst amacrine cells, and therefore computation of directional selectivity also occurs postsynaptically. Such postsynaptic models are unparsimonious, and so if any given starburst amacrine cells conveys motion information to retinal ganglion cells then any computing of 'local' direction selectivity postsynaptically by retinal ganglion cells is redundant and dysfunctional. An acetylcholine (ACh) transmission model of directionally selective starburst amacrine cells provides a robust topological underpinning of a motion sensing in the retina.[36]

See also

edit

References

edit
  1. ^ Watson JD, Myers R, Frackowiak RS, Hajnal JV, Woods RP, Mazziotta JC, et al. (1993). "Area V5 of the human brain: evidence from a combined study using positron emission tomography and magnetic resonance imaging". Cerebral Cortex. 3 (2): 79–94. doi:10.1093/cercor/3.2.79. PMID 8490322.
  2. ^ Zeki SM (February 1974). "Functional organization of a visual area in the posterior bank of the superior temporal sulcus of the rhesus monkey". The Journal of Physiology. 236 (3): 549–73. doi:10.1113/jphysiol.1974.sp010452. PMC 1350849. PMID 4207129.
  3. ^ Hess RH, Baker CL, Zihl J (May 1989). "The "motion-blind" patient: low-level spatial and temporal filters". The Journal of Neuroscience. 9 (5): 1628–40. doi:10.1523/JNEUROSCI.09-05-01628.1989. PMC 6569833. PMID 2723744.
  4. ^ Baker CL, Hess RF, Zihl J (February 1991). "Residual motion perception in a "motion-blind" patient, assessed with limited-lifetime random dot stimuli". The Journal of Neuroscience. 11 (2): 454–61. doi:10.1523/JNEUROSCI.11-02-00454.1991. PMC 6575225. PMID 1992012.
  5. ^ a b Steinman, Pizlo & Pizlo (2000) Phi is not Beta slideshow based on ARVO presentation.
  6. ^ Reichardt W (1961). "Autocorrelation, a principle for the evaluation of sensory information by the central nervous system". In W.A. Rosenblith (ed.). Sensory Communication. MIT Press. pp. 303–317.
  7. ^ Adelson EH, Bergen JR (February 1985). "Spatiotemporal energy models for the perception of motion". Journal of the Optical Society of America A. 2 (2): 284–99. Bibcode:1985JOSAA...2..284A. CiteSeerX 10.1.1.148.4141. doi:10.1364/JOSAA.2.000284. PMID 3973762.
  8. ^ van Santen JP, Sperling G (February 1985). "Elaborated Reichardt detectors". Journal of the Optical Society of America A. 2 (2): 300–21. Bibcode:1985JOSAA...2..300S. doi:10.1364/JOSAA.2.000300. PMID 3973763. S2CID 5699316.
  9. ^ Cavanagh P, Mather G (1989). "Motion: the long and short of it". Spatial Vision. 4 (2–3): 103–29. doi:10.1163/156856889X00077. PMID 2487159.
  10. ^ Chubb C, Sperling G (November 1988). "Drift-balanced random stimuli: a general basis for studying non-Fourier motion perception". Journal of the Optical Society of America A. 5 (11): 1986–2007. Bibcode:1988JOSAA...5.1986C. CiteSeerX 10.1.1.324.3078. doi:10.1364/JOSAA.5.001986. PMID 3210090.
  11. ^ Nishida S, Ledgeway T, Edwards M (October 1997). "Dual multiple-scale processing for motion in the human visual system". Vision Research. 37 (19): 2685–98. doi:10.1016/S0042-6989(97)00092-8. PMID 9373668. S2CID 7344938.
  12. ^ Ledgeway T, Smith AT (1994). "The duration of the motion aftereffect following adaptation to first-order and second-order motion". Perception. 23 (10): 1211–9. doi:10.1068/p231211. PMID 7899037. S2CID 22761002.
  13. ^ Burr DC, Santoro L (July 2001). "Temporal integration of optic flow, measured by contrast and coherence thresholds". Vision Research. 41 (15): 1891–9. doi:10.1016/S0042-6989(01)00072-4. PMID 11412882. S2CID 16751457.
  14. ^ Lages M, Heron S (November 2010). "On the inverse problem of binocular 3D motion perception". PLOS Computational Biology. 6 (11): e1000999. Bibcode:2010PLSCB...6E0999L. doi:10.1371/journal.pcbi.1000999. PMC 2987932. PMID 21124957.
  15. ^ Blake R, Wilson H (April 2011). "Binocular vision". Vision Research. 51 (7): 754–70. doi:10.1016/j.visres.2010.10.009. PMC 3050089. PMID 20951722.
  16. ^ Shioiri S, Saisho H, Yaguchi H (2000). "Motion in depth based on inter-ocular velocity differences". Vision Research. 40 (19): 2565–72. doi:10.1016/s0042-6989(00)00130-9. PMID 10958908. S2CID 15342293.
  17. ^ Himmelberg MM, Segala FG, Maloney RT, Harris JM, Wade AR (2020). "Decoding Neural Responses to Motion-in-Depth Using EEG". Frontiers in Neuroscience. 14: 581706. doi:10.3389/fnins.2020.581706. PMC 7758252. PMID 33362456.
  18. ^ Brooks KR (2002). "Interocular velocity difference contributes to stereomotion speed perception". Journal of Vision. 2 (3): 218–31. doi:10.1167/2.3.2. PMID 12678584.
  19. ^ Ball K, Sekuler R (November 1982). "A specific and enduring improvement in visual motion discrimination". Science. 218 (4573): 697–8. Bibcode:1982Sci...218..697B. doi:10.1126/science.7134968. PMID 7134968.
  20. ^ Kitchin RM (1994). "Cognitive Maps: What Are They and Why Study Them?" (PDF). Journal of Environmental Psychology (Submitted manuscript). 14 (1): 1–19. doi:10.1016/S0272-4944(05)80194-X.
  21. ^ O'Keefe J (1978). The Hippocampus as a Cognitive Map. Clarendon Press. ISBN 978-0198572060.
  22. ^ Muir GM, Bilkey DK (June 2001). "Instability in the place field location of hippocampal place cells after lesions centered on the perirhinal cortex". The Journal of Neuroscience. 21 (11): 4016–25. doi:10.1523/JNEUROSCI.21-11-04016.2001. PMC 6762702. PMID 11356888.
  23. ^ Redei G (2008). Encyclopedia of Genetics, Genomics, Proteomics, and Informatics. Springer. p. 1501. ISBN 978-1-4020-6753-2..
  24. ^ Samsonovich A, McNaughton BL (August 1997). "Path integration and cognitive mapping in a continuous attractor neural network model". The Journal of Neuroscience. 17 (15): 5900–20. doi:10.1523/JNEUROSCI.17-15-05900.1997. PMC 6573219. PMID 9221787.
  25. ^ Burak Y, Fiete IR (February 2009). Sporns O (ed.). "Accurate path integration in continuous attractor network models of grid cells". PLOS Computational Biology. 5 (2): e1000291. arXiv:0811.1826. Bibcode:2009PLSCB...5E0291B. doi:10.1371/journal.pcbi.1000291. PMC 2632741. PMID 19229307.
  26. ^ Hafting T, Fyhn M, Molden S, Moser MB, Moser EI (August 2005). "Microstructure of a spatial map in the entorhinal cortex". Nature. 436 (7052): 801–6. Bibcode:2005Natur.436..801H. doi:10.1038/nature03721. PMID 15965463. S2CID 4405184.
  27. ^ Sargolini F, Fyhn M, Hafting T, McNaughton BL, Witter MP, Moser MB, Moser EI (May 2006). "Conjunctive representation of position, direction, and velocity in entorhinal cortex". Science. 312 (5774): 758–62. Bibcode:2006Sci...312..758S. doi:10.1126/science.1125572. PMID 16675704. S2CID 263378884.
  28. ^ a b c Barlow HB, Levick WR (June 1965). "The mechanism of directionally selective units in rabbit's retina". The Journal of Physiology. 178 (3): 477–504. doi:10.1113/jphysiol.1965.sp007638. PMC 1357309. PMID 5827909.
  29. ^ a b c d Borst A, Euler T (September 2011). "Seeing things in motion: models, circuits, and mechanisms". Neuron. 71 (6): 974–94. doi:10.1016/j.neuron.2011.08.031. PMID 21943597. S2CID 8408814.
  30. ^ Dhande OS, Stafford BK, Franke K, El-Danaf R, Percival KA, Phan AH, et al. (January 2019). "Molecular Fingerprinting of On-Off Direction-Selective Retinal Ganglion Cells Across Species and Relevance to Primate Visual Circuits". The Journal of Neuroscience. 39 (1): 78–95. doi:10.1523/JNEUROSCI.1784-18.2018. PMC 6325260. PMID 30377226.
  31. ^ Hassenstein B, Reichardt W (1956-10-01). "Systemtheoretische Analyse der Zeit-, Reihenfolgen- und Vorzeichenauswertung bei der Bewegungsperzeption des Rüsselkäfers Chlorophanus". Zeitschrift für Naturforschung B. 11 (9–10): 513–524. doi:10.1515/znb-1956-9-1004. hdl:11858/00-001M-0000-0013-F2EA-6. ISSN 1865-7117. S2CID 98709700.
  32. ^ Haag J, Denk W, Borst A (November 2004). "Fly motion vision is based on Reichardt detectors regardless of the signal-to-noise ratio". Proceedings of the National Academy of Sciences of the United States of America. 101 (46): 16333–8. Bibcode:2004PNAS..10116333H. doi:10.1073/pnas.0407368101. PMC 526200. PMID 15534201.
  33. ^ Kay JN, De la Huerta I, Kim IJ, Zhang Y, Yamagata M, Chu MW, et al. (May 2011). "Retinal ganglion cells with distinct directional preferences differ in molecular identity, structure, and central projections". The Journal of Neuroscience. 31 (21): 7753–62. doi:10.1523/jneurosci.0907-11.2011. PMC 3108146. PMID 21613488.
  34. ^ Demb JB (July 2007). "Cellular mechanisms for direction selectivity in the retina". Neuron. 55 (2): 179–86. doi:10.1016/j.neuron.2007.07.001. PMID 17640521. S2CID 5691739.
  35. ^ Briggman KL, Helmstaedter M, Denk W (March 2011). "Wiring specificity in the direction-selectivity circuit of the retina". Nature. 471 (7337): 183–8. Bibcode:2011Natur.471..183B. doi:10.1038/nature09818. PMID 21390125. S2CID 4425160.
  36. ^ Poznanski RR (September 2010). "Cellular inhibitory behavior underlying the formation of retinal direction selectivity in the starburst network". Journal of Integrative Neuroscience. 9 (3): 299–335. doi:10.1142/s0219635210002457. PMID 21064220.

Further reading

edit
edit

Labs specialising in motion research

edit