Six Myths about Spatial Thinking
ERIC Educational Resources Information Center
Newcombe, Nora S.; Stieff, Mike
2012-01-01
Visualizations are an increasingly important part of scientific education and discovery. However, users often do not gain knowledge from them in a complete or efficient way. This article aims to direct research on visualizations in science education in productive directions by reviewing the evidence for widespread assumptions that learning styles,…
Functional Dissociation between Perception and Action Is Evident Early in Life
ERIC Educational Resources Information Center
Hadad, Bat-Sheva; Avidan, Galia; Ganel, Tzvi
2012-01-01
The functional distinction between vision for perception and vision for action is well documented in the mature visual system. Ganel and colleagues recently provided direct evidence for this dissociation, showing that while visual processing for perception follows Weber's fundamental law of psychophysics, action violates this law. We tracked the…
Buschman, Timothy J.; Miller, Earl K.
2009-01-01
Attention regulates the flood of sensory information into a manageable stream, and so understanding how attention is controlled is central to understanding cognition. Competing theories suggest visual search involves serial and/or parallel allocation of attention, but there is little direct, neural, evidence for either mechanism. Two monkeys were trained to covertly search an array for a target stimulus under visual search (endogenous) and pop-out (exogenous) conditions. Here we present neural evidence in the frontal eye fields (FEF) for serial, covert shifts of attention during search but not pop-out. Furthermore, attention shifts reflected in FEF spiking activity were correlated with 18–34 Hz oscillations in the local field potential, suggesting a ‘clocking’ signal. This provides direct neural evidence that primates can spontaneously adopt a serial search strategy and that these serial covert shifts of attention are directed by the FEF. It also suggests that neuron population oscillations may regulate the timing of cognitive processing. PMID:19679077
Causal evidence for retina dependent and independent visual motion computations in mouse cortex
Hillier, Daniel; Fiscella, Michele; Drinnenberg, Antonia; Trenholm, Stuart; Rompani, Santiago B.; Raics, Zoltan; Katona, Gergely; Juettner, Josephine; Hierlemann, Andreas; Rozsa, Balazs; Roska, Botond
2017-01-01
How neuronal computations in the sensory periphery contribute to computations in the cortex is not well understood. We examined this question in the context of visual-motion processing in the retina and primary visual cortex (V1) of mice. We disrupted retinal direction selectivity – either exclusively along the horizontal axis using FRMD7 mutants or along all directions by ablating starburst amacrine cells – and monitored neuronal activity in layer 2/3 of V1 during stimulation with visual motion. In control mice, we found an overrepresentation of cortical cells preferring posterior visual motion, the dominant motion direction an animal experiences when it moves forward. In mice with disrupted retinal direction selectivity, the overrepresentation of posterior-motion-preferring cortical cells disappeared, and their response at higher stimulus speeds was reduced. This work reveals the existence of two functionally distinct, sensory-periphery-dependent and -independent computations of visual motion in the cortex. PMID:28530661
Interlateral Asymmetry in the Time Course of the Effect of a Peripheral Prime Stimulus
ERIC Educational Resources Information Center
Castro-Barros, B. A.; Righi, L. L.; Grechi, G.; Ribeiro-do-Valle, L. E.
2008-01-01
Evidence exists that both right and left hemisphere attentional mechanisms are mobilized when attention is directed to the right visual hemifield and only right hemisphere attentional mechanisms are mobilized when attention is directed to the left visual hemifield. This arrangement might lead to a rightward bias of automatic attention. The…
NASA Astrophysics Data System (ADS)
Seki, T.; Iguchi, R.; Takanashi, K.; Uchida, K.
2018-04-01
Spatial distribution of temperature modulation due to the anomalous Ettingshausen effect (AEE) is visualized in a ferromagnetic FePt thin film with in-plane and out-of-plane magnetizations using the lock-in thermography technique. Comparing the AEE of FePt with the spin Peltier effect (SPE) of a Pt/yttrium iron garnet junction provides direct evidence of different symmetries of AEE and SPE. Our experiments and numerical calculations reveal that the distribution of heat sources induced by AEE strongly depends on the direction of magnetization, leading to the remarkable different temperature profiles in the FePt thin film between the in-plane and perpendicularly magnetized configurations.
The multisensory function of the human primary visual cortex.
Murray, Micah M; Thelen, Antonia; Thut, Gregor; Romei, Vincenzo; Martuzzi, Roberto; Matusz, Pawel J
2016-03-01
It has been nearly 10 years since Ghazanfar and Schroeder (2006) proposed that the neocortex is essentially multisensory in nature. However, it is only recently that sufficient and hard evidence that supports this proposal has accrued. We review evidence that activity within the human primary visual cortex plays an active role in multisensory processes and directly impacts behavioural outcome. This evidence emerges from a full pallet of human brain imaging and brain mapping methods with which multisensory processes are quantitatively assessed by taking advantage of particular strengths of each technique as well as advances in signal analyses. Several general conclusions about multisensory processes in primary visual cortex of humans are supported relatively solidly. First, haemodynamic methods (fMRI/PET) show that there is both convergence and integration occurring within primary visual cortex. Second, primary visual cortex is involved in multisensory processes during early post-stimulus stages (as revealed by EEG/ERP/ERFs as well as TMS). Third, multisensory effects in primary visual cortex directly impact behaviour and perception, as revealed by correlational (EEG/ERPs/ERFs) as well as more causal measures (TMS/tACS). While the provocative claim of Ghazanfar and Schroeder (2006) that the whole of neocortex is multisensory in function has yet to be demonstrated, this can now be considered established in the case of the human primary visual cortex. Copyright © 2015 Elsevier Ltd. All rights reserved.
What and where information in the caudate tail guides saccades to visual objects
Yamamoto, Shinya; Monosov, Ilya E.; Yasuda, Masaharu; Hikosaka, Okihide
2012-01-01
We understand the world by making saccadic eye movements to various objects. However, it is unclear how a saccade can be aimed at a particular object, because two kinds of visual information, what the object is and where it is, are processed separately in the dorsal and ventral visual cortical pathways. Here we provide evidence suggesting that a basal ganglia circuit through the tail of the monkey caudate nucleus (CDt) guides such object-directed saccades. First, many CDt neurons responded to visual objects depending on where and what the objects were. Second, electrical stimulation in the CDt induced saccades whose directions matched the preferred directions of neurons at the stimulation site. Third, many CDt neurons increased their activity before saccades directed to the neurons’ preferred objects and directions in a free-viewing condition. Our results suggest that CDt neurons receive both ‘what’ and ‘where’ information and guide saccades to visual objects. PMID:22875934
The role of temporal structure in human vision.
Blake, Randolph; Lee, Sang-Hun
2005-03-01
Gestalt psychologists identified several stimulus properties thought to underlie visual grouping and figure/ground segmentation, and among those properties was common fate: the tendency to group together individual objects that move together in the same direction at the same speed. Recent years have witnessed an upsurge of interest in visual grouping based on other time-dependent sources of visual information, including synchronized changes in luminance, in motion direction, and in figure/ ground relations. These various sources of temporal grouping information can be subsumed under the rubric temporal structure. In this article, the authors review evidence bearing on the effectiveness of temporal structure in visual grouping. They start with an overview of evidence bearing on temporal acuity of human vision, covering studies dealing with temporal integration and temporal differentiation. They then summarize psychophysical studies dealing with figure/ground segregation based on temporal phase differences in deterministic and stochastic events. The authors conclude with a brief discussion of neurophysiological implications of these results.
The cost of parallel consolidation into visual working memory.
Rideaux, Reuben; Edwards, Mark
2016-01-01
A growing body of evidence indicates that information can be consolidated into visual working memory in parallel. Initially, it was suggested that color information could be consolidated in parallel while orientation was strictly limited to serial consolidation (Liu & Becker, 2013). However, we recently found evidence suggesting that both orientation and motion direction items can be consolidated in parallel, with different levels of accuracy (Rideaux, Apthorp, & Edwards, 2015). Here we examine whether there is a cost associated with parallel consolidation of orientation and direction information by comparing performance, in terms of precision and guess rate, on a target recall task where items are presented either sequentially or simultaneously. The results compellingly indicate that motion direction can be consolidated in parallel, but the evidence for orientation is less conclusive. Further, we find that there is a twofold cost associated with parallel consolidation of direction: Both the probability of failing to consolidate one (or both) item/s increases and the precision at which representations are encoded is reduced. Additionally, we find evidence indicating that the increased consolidation failure may be due to interference between items presented simultaneously, and is moderated by item similarity. These findings suggest that a biased competition model may explain differences in parallel consolidation between features.
Small numbers are sensed directly, high numbers constructed from size and density.
Zimmermann, Eckart
2018-04-01
Two theories compete to explain how we estimate the numerosity of visual object sets. The first suggests that the apparent numerosity is derived from an analysis of more low-level features like size and density of the set. The second theory suggests that numbers are sensed directly. Consistent with the latter claim is the existence of neurons in parietal cortex which are specialized for processing the numerosity of elements in the visual scene. However, recent evidence suggests that only low numbers can be sensed directly whereas the perception of high numbers is supported by the analysis of low-level features. Processing of low and high numbers, being located at different levels of the neural hierarchy should involve different receptive field sizes. Here, I tested this idea with visual adaptation. I measured the spatial spread of number adaptation for low and high numerosities. A focused adaptation spread of high numerosities suggested the involvement of early neural levels where receptive fields are comparably small and the broad spread for low numerosities was consistent with processing of number neurons which have larger receptive fields. These results provide evidence for the claim that different mechanism exist generating the perception of visual numerosity. Whereas low numbers are sensed directly as a primary visual attribute, the estimation of high numbers however likely depends on the area size over which the objects are spread. Copyright © 2017 Elsevier B.V. All rights reserved.
Visual grouping under isoluminant condition: impact of mental fatigue
NASA Astrophysics Data System (ADS)
Pladere, Tatjana; Bete, Diana; Skilters, Jurgis; Krumina, Gunta
2016-09-01
Instead of selecting arbitrary elements our visual perception prefers only certain grouping of information. There is ample evidence that the visual attention and perception is substantially impaired in the presence of mental fatigue. The question is how visual grouping, which can be considered a bottom-up controlled neuronal gain mechanism, is influenced. The main purpose of our study is to determine the influence of mental fatigue on visual grouping of definite information - color and configuration of stimuli in the psychophysical experiment. Individuals provided subjective data by filling in the questionnaire about their health and general feeling. The objective evidence was obtained in the specially designed visual search task were achromatic and chromatic isoluminant stimuli were used in order to avoid so called pop-out effect due to differences in light intensity. Each individual was instructed to define the symbols with aperture in the same direction in four tasks. The color component differed in the visual search tasks according to the goals of study. The results reveal that visual grouping is completed faster when visual stimuli have the same color and aperture direction. The shortest reaction time is in the evening. What is more, the results of reaction time suggest that the analysis of two grouping processes compete for selective attention in the visual system when similarity in color conflicts with similarity in configuration of stimuli. The described effect increases significantly in the presence of mental fatigue. But it does not have strong influence on the accuracy of task accomplishment.
Federal Register 2010, 2011, 2012, 2013, 2014
2011-01-12
... eddy current and visual inspections of the upper wing strut fitting for evidence of cracks, wear and/or... permitted extending the intervals for the repetitive eddy current and visual inspections from 100 Flight... the applicability and to require repetitive eddy current and visual inspections of the upper wing...
DOE Office of Scientific and Technical Information (OSTI.GOV)
Chen Bo; State Key Laboratory of Brain and Cognitive Science, Institute of Biophysics, Chinese Academy of Science, Beijing 100101; Xia Jing
Physiological and behavioral studies have demonstrated that a number of visual functions such as visual acuity, contrast sensitivity, and motion perception can be impaired by acute alcohol exposure. The orientation- and direction-selective responses of cells in primary visual cortex are thought to participate in the perception of form and motion. To investigate how orientation selectivity and direction selectivity of neurons are influenced by acute alcohol exposure in vivo, we used the extracellular single-unit recording technique to examine the response properties of neurons in primary visual cortex (A17) of adult cats. We found that alcohol reduces spontaneous activity, visual evoked unitmore » responses, the signal-to-noise ratio, and orientation selectivity of A17 cells. In addition, small but detectable changes in both the preferred orientation/direction and the bandwidth of the orientation tuning curve of strongly orientation-biased A17 cells were observed after acute alcohol administration. Our findings may provide physiological evidence for some alcohol-related deficits in visual function observed in behavioral studies.« less
Motivationally Significant Stimuli Show Visual Prior Entry: Evidence for Attentional Capture
ERIC Educational Resources Information Center
West, Greg L.; Anderson, Adam A. K.; Pratt, Jay
2009-01-01
Previous studies that have found attentional capture effects for stimuli of motivational significance do not directly measure initial attentional deployment, leaving it unclear to what extent these items produce attentional capture. Visual prior entry, as measured by temporal order judgments (TOJs), rests on the premise that allocated attention…
Task set induces dynamic reallocation of resources in visual short-term memory.
Sheremata, Summer L; Shomstein, Sarah
2017-08-01
Successful interaction with the environment requires the ability to flexibly allocate resources to different locations in the visual field. Recent evidence suggests that visual short-term memory (VSTM) resources are distributed asymmetrically across the visual field based upon task demands. Here, we propose that context, rather than the stimulus itself, determines asymmetrical distribution of VSTM resources. To test whether context modulates the reallocation of resources to the right visual field, task set, defined by memory-load, was manipulated to influence visual short-term memory performance. Performance was measured for single-feature objects embedded within predominantly single- or two-feature memory blocks. Therefore, context was varied to determine whether task set directly predicts changes in visual field biases. In accord with the dynamic reallocation of resources hypothesis, task set, rather than aspects of the physical stimulus, drove improvements in performance in the right- visual field. Our results show, for the first time, that preparation for upcoming memory demands directly determines how resources are allocated across the visual field.
Distractor Devaluation Effect in the Attentional Blink: Direct Evidence for Distractor Inhibition
ERIC Educational Resources Information Center
Kihara, Ken; Yagi, Yoshihiko; Takeda, Yuji; Kawahara, Jun I.
2011-01-01
When two targets (T1 and T2) are embedded in rapid serial visual presentation (RSVP), T2 is often missed (attentional blink, AB) if T2 follows T1 by less than 500 ms. Some have proposed that inhibition of a distractor following T1 contributes to the AB, but no direct evidence supports this proposal. This study examined distractor inhibition by…
Verleger, Rolf; Śmigasiewicz, Kamila
2015-09-01
Everyday experience suggests that people are equally aware of events in both hemi-fields. However, when two streams of stimuli are rapidly presented left and right containing two targets, the second target is better identified in the left than in the right visual field. This might be considered evidence for a right-hemisphere advantage in generating conscious percepts. However, this putative asymmetry of conscious perception cannot be measured independently of participants' access to their conscious percepts, and there is actually evidence from split-brain patients for the reverse, left-hemisphere advantage in having access to conscious percepts. Several other topics were studied in search of the responsible mechanism, among others: Mutual inhibition of hemispheres, cooperation of hemispheres in perceiving midline stimuli, and asymmetries in processing various perceptual inputs. Directing attention by salient cues turned out to be one of the few mechanisms capable of modifying the left visual-field advantage in this paradigm. Thus, this left visual-field advantage is best explained by the notion of a right-hemisphere advantage in directing attention to salient events. Dovetailing with the pathological asymmetries of attention after right-hemisphere lesions and with asymmetries of brain activation when healthy participants shift their attention, the present results extend that body of evidence by demonstrating unusually large and reliable behavioral asymmetries for attention-directing processes in healthy participants. Copyright © 2015 Elsevier Inc. All rights reserved.
Heinen, Klaartje; Feredoes, Eva; Weiskopf, Nikolaus; Ruff, Christian C; Driver, Jon
2014-11-01
Voluntary selective attention can prioritize different features in a visual scene. The frontal eye-fields (FEF) are one potential source of such feature-specific top-down signals, but causal evidence for influences on visual cortex (as was shown for "spatial" attention) has remained elusive. Here, we show that transcranial magnetic stimulation (TMS) applied to right FEF increased the blood oxygen level-dependent (BOLD) signals in visual areas processing "target feature" but not in "distracter feature"-processing regions. TMS-induced BOLD signals increase in motion-responsive visual cortex (MT+) when motion was attended in a display with moving dots superimposed on face stimuli, but in face-responsive fusiform area (FFA) when faces were attended to. These TMS effects on BOLD signal in both regions were negatively related to performance (on the motion task), supporting the behavioral relevance of this pathway. Our findings provide new causal evidence for the human FEF in the control of nonspatial "feature"-based attention, mediated by dynamic influences on feature-specific visual cortex that vary with the currently attended property. © The Author 2013. Published by Oxford University Press.
Thaler, Lore; Goodale, Melvyn A.
2011-01-01
Neuropsychological evidence suggests that different brain areas may be involved in movements that are directed at visual targets (e.g., pointing or reaching), and movements that are based on allocentric visual information (e.g., drawing or copying). Here we used fMRI to investigate the neural correlates of these two types of movements in healthy volunteers. Subjects (n = 14) performed right hand movements in either a target-directed task (moving a cursor to a target dot) or an allocentric task (moving a cursor to reproduce the distance and direction between two distal target dots) with or without visual feedback about their hand movement. Movements were monitored with an MR compatible touch panel. A whole brain analysis revealed that movements in allocentric conditions led to an increase in activity in the fundus of the left intra-parietal sulcus (IPS), in posterior IPS, in bilateral dorsal premotor cortex (PMd), and in the lateral occipital complex (LOC). Visual feedback in both target-directed and allocentric conditions led to an increase in activity in area MT+, superior parietal–occipital cortex (SPOC), and posterior IPS (all bilateral). In addition, we found that visual feedback affected brain activity differently in target-directed as compared to allocentric conditions, particularly in the pre-supplementary motor area, PMd, IPS, and parieto-occipital cortex. Our results, in combination with previous findings, suggest that the LOC is essential for allocentric visual coding and that SPOC is involved in visual feedback control. The differences in brain activity between target-directed and allocentric visual feedback conditions may be related to behavioral differences in visual feedback control. Our results advance the understanding of the visual coordinate frame used by the LOC. In addition, because of the nature of the allocentric task, our results have relevance for the understanding of neural substrates of magnitude estimation and vector coding of movements. PMID:21941474
Illusory motion reversal is caused by rivalry, not by perceptual snapshots of the visual field.
Kline, Keith; Holcombe, Alex O; Eagleman, David M
2004-10-01
In stroboscopic conditions--such as motion pictures--rotating objects may appear to rotate in the reverse direction due to under-sampling (aliasing). A seemingly similar phenomenon occurs in constant sunlight, which has been taken as evidence that the visual system processes discrete "snapshots" of the outside world. But if snapshots are indeed taken of the visual field, then when a rotating drum appears to transiently reverse direction, its mirror image should always appeared to reverse direction simultaneously. Contrary to this hypothesis, we found that when observers watched a rotating drum and its mirror image, almost all illusory motion reversals occurred for only one image at a time. This result indicates that the motion reversal illusion cannot be explained by snapshots of the visual field. The same result is found when the two images are presented within one visual hemifield, further ruling out the possibility that discrete sampling of the visual field occurs separately in each hemisphere. The frequency distribution of illusory reversal durations approximates a gamma distribution, suggesting perceptual rivalry as a better explanation for illusory motion reversal. After adaptation of motion detectors coding for the correct direction, the activity of motion-sensitive neurons coding for motion in the reverse direction may intermittently become dominant and drive the perception of motion.
Motion Direction Biases and Decoding in Human Visual Cortex
Wang, Helena X.; Merriam, Elisha P.; Freeman, Jeremy
2014-01-01
Functional magnetic resonance imaging (fMRI) studies have relied on multivariate analysis methods to decode visual motion direction from measurements of cortical activity. Above-chance decoding has been commonly used to infer the motion-selective response properties of the underlying neural populations. Moreover, patterns of reliable response biases across voxels that underlie decoding have been interpreted to reflect maps of functional architecture. Using fMRI, we identified a direction-selective response bias in human visual cortex that: (1) predicted motion-decoding accuracy; (2) depended on the shape of the stimulus aperture rather than the absolute direction of motion, such that response amplitudes gradually decreased with distance from the stimulus aperture edge corresponding to motion origin; and 3) was present in V1, V2, V3, but not evident in MT+, explaining the higher motion-decoding accuracies reported previously in early visual cortex. These results demonstrate that fMRI-based motion decoding has little or no dependence on the underlying functional organization of motion selectivity. PMID:25209297
ERP Evidence of Hemispheric Independence in Visual Word Recognition
ERIC Educational Resources Information Center
Nemrodov, Dan; Harpaz, Yuval; Javitt, Daniel C.; Lavidor, Michal
2011-01-01
This study examined the capability of the left hemisphere (LH) and the right hemisphere (RH) to perform a visual recognition task independently as formulated by the Direct Access Model (Fernandino, Iacoboni, & Zaidel, 2007). Healthy native Hebrew speakers were asked to categorize nouns and non-words (created from nouns by transposing two middle…
Effects of visual attention on chromatic and achromatic detection sensitivities.
Uchikawa, Keiji; Sato, Masayuki; Kuwamura, Keiko
2014-05-01
Visual attention has a significant effect on various visual functions, such as response time, detection and discrimination sensitivity, and color appearance. It has been suggested that visual attention may affect visual functions in the early visual pathways. In this study we examined selective effects of visual attention on sensitivities of the chromatic and achromatic pathways to clarify whether visual attention modifies responses in the early visual system. We used a dual task paradigm in which the observer detected a peripheral test stimulus presented at 4 deg eccentricities while the observer concurrently carried out an attention task in the central visual field. In experiment 1, it was confirmed that peripheral spectral sensitivities were reduced more for short and long wavelengths than for middle wavelengths with the central attention task so that the spectral sensitivity function changed its shape by visual attention. This indicated that visual attention affected the chromatic response more strongly than the achromatic response. In experiment 2 it was obtained that the detection thresholds increased in greater degrees in the red-green and yellow-blue chromatic directions than in the white-black achromatic direction in the dual task condition. In experiment 3 we showed that the peripheral threshold elevations depended on the combination of color-directions of the central and peripheral stimuli. Since the chromatic and achromatic responses were separately processed in the early visual pathways, the present results provided additional evidence that visual attention affects responses in the early visual pathways.
Integrating visualization and interaction research to improve scientific workflows.
Keefe, Daniel F
2010-01-01
Scientific-visualization research is, nearly by necessity, interdisciplinary. In addition to their collaborators in application domains (for example, cell biology), researchers regularly build on close ties with disciplines related to visualization, such as graphics, human-computer interaction, and cognitive science. One of these ties is the connection between visualization and interaction research. This isn't a new direction for scientific visualization (see the "Early Connections" sidebar). However, momentum recently seems to be increasing toward integrating visualization research (for example, effective visual presentation of data) with interaction research (for example, innovative interactive techniques that facilitate manipulating and exploring data). We see evidence of this trend in several places, including the visualization literature and conferences.
Maffei, Vincenzo; Mazzarella, Elisabetta; Piras, Fabrizio; Spalletta, Gianfranco; Caltagirone, Carlo; Lacquaniti, Francesco; Daprati, Elena
2016-05-01
Rich behavioral evidence indicates that the brain estimates the visual direction and acceleration of gravity quite accurately, and the underlying mechanisms have begun to be unraveled. While the neuroanatomical substrates of gravity direction processing have been studied extensively in brain-damaged patients, to our knowledge no such study exists for the processing of visual gravitational motion. Here we asked 31 stroke patients to intercept a virtual ball moving along the vertical under either natural gravity or artificial reversed gravity. Twenty-seven of them also aligned a luminous bar to the vertical direction (subjective visual vertical, SVV). Using voxel-based lesion-symptom mapping as well as lesion subtraction analysis, we found that lesions mainly centered on the posterior insula are associated with greater deviations of SVV, consistent with several previous studies. Instead, lesions mainly centered on the parietal operculum decrease the ability to discriminate natural from unnatural gravitational acceleration with a timed motor response in the interception task. Both the posterior insula and the parietal operculum belong to the vestibular cortex, and presumably receive multisensory information about the gravity vector. We speculate that an internal model estimating the effects of gravity on visual objects is constructed by transforming the vestibular estimates of mechanical gravity, which are computed in the brainstem and cerebellum, into internalized estimates of virtual gravity, which are stored in the cortical vestibular network. The present lesion data suggest a specific role for the parietal operculum in detecting the mismatch between predictive signals from the internal model and the online visual signals. Copyright © 2016 Elsevier Ltd. All rights reserved.
Predictions penetrate perception: Converging insights from brain, behaviour and disorder
O’Callaghan, Claire; Kveraga, Kestutis; Shine, James M; Adams, Reginald B.; Bar, Moshe
2018-01-01
It is argued that during ongoing visual perception, the brain is generating top-down predictions to facilitate, guide and constrain the processing of incoming sensory input. Here we demonstrate that these predictions are drawn from a diverse range of cognitive processes, in order to generate the richest and most informative prediction signals. This is consistent with a central role for cognitive penetrability in visual perception. We review behavioural and mechanistic evidence that indicate a wide spectrum of domains—including object recognition, contextual associations, cognitive biases and affective state—that can directly influence visual perception. We combine these insights from the healthy brain with novel observations from neuropsychiatric disorders involving visual hallucinations, which highlight the consequences of imbalance between top-down signals and incoming sensory information. Together, these lines of evidence converge to indicate that predictive penetration, be it cognitive, social or emotional, should be considered a fundamental framework that supports visual perception. PMID:27222169
Preserved figure-ground segregation and symmetry perception in visual neglect.
Driver, J; Baylis, G C; Rafal, R D
1992-11-05
A central controversy in current research on visual attention is whether figures are segregated from their background preattentively, or whether attention is first directed to unstructured regions of the image. Here we present neurological evidence for the former view from studies of a brain-injured patient with visual neglect. His attentional impairment arises after normal segmentation of the image into figures and background has taken place. Our results indicate that information which is neglected and unavailable to higher levels of visual processing can nevertheless be processed by earlier stages in the visual system concerned with segmentation.
A Theoretical Note on Sex Linkage and Race Differences in Spatial Visualization Ability
ERIC Educational Resources Information Center
Jensen, Arthur R.
1975-01-01
Evidence on the poorer spatial visualization ability in various Negro populations compared to the White populations and on the direction and magnitude of sex differences in spatial ability relative to other abilities suggests the genetic hypothesis that spatial ability is enhanced by a sex-linked recessive gene and that, since the 20-30 percent…
ERIC Educational Resources Information Center
Thomson, Hilary J.; Thomas, Sian
2013-01-01
Visual display of reported impacts is a valuable aid to both reviewers and readers of systematic reviews. Forest plots are routinely prepared to report standardised effect sizes, but where standardised effect sizes are not available for all included studies a forest plot may misrepresent the available evidence. Tabulated data summaries to…
Vessel, Edward A; Biederman, Irving; Subramaniam, Suresh; Greene, Michelle R
2016-07-01
An L-vertex, the point at which two contours coterminate, provides highly reliable evidence that a surface terminates at that vertex, thus providing the strongest constraint on the extraction of shape from images (Guzman, 1968). Such vertices are pervasive in our visual world but the importance of a statistical regularity about them has been underappreciated: The contours defining the vertex are (almost) always of the same direction of contrast with respect to the background (i.e., both darker or both lighter). Here we show that when the two contours are of different directions of contrast, the capacity of the L-vertex to signal the termination of a surface, as reflected in object recognition, is markedly reduced. Although image statistics have been implicated in determining the connectivity in the earliest cortical visual stage (V1) and in grouping during visual search, this finding provides evidence that such statistics are involved in later stages where object representations are derived from two-dimensional images.
Chen, Chen; Schneps, Matthew H; Masyn, Katherine E; Thomson, Jennifer M
2016-11-01
Increasing evidence has shown visual attention span to be a factor, distinct from phonological skills, that explains single-word identification (pseudo-word/word reading) performance in dyslexia. Yet, little is known about how well visual attention span explains text comprehension. Observing reading comprehension in a sample of 105 high school students with dyslexia, we used a pathway analysis to examine the direct and indirect path between visual attention span and reading comprehension while controlling for other factors such as phonological awareness, letter identification, short-term memory, IQ and age. Integrating phonemic decoding efficiency skills in the analytic model, this study aimed to disentangle how visual attention span and phonological skills work together in reading comprehension for readers with dyslexia. We found visual attention span to have a significant direct effect on more difficult reading comprehension but not on an easier level. It also had a significant direct effect on pseudo-word identification but not on word identification. In addition, we found that visual attention span indirectly explains reading comprehension through pseudo-word reading and word reading skills. This study supports the hypothesis that at least part of the dyslexic profile can be explained by visual attention abilities. Copyright © 2016 John Wiley & Sons, Ltd. Copyright © 2016 John Wiley & Sons, Ltd.
Utilizing Rapid Prototyping for Architectural Modeling
ERIC Educational Resources Information Center
Kirton, E. F.; Lavoie, S. D.
2006-01-01
This paper will discuss our approach to, success with and future direction in rapid prototyping for architectural modeling. The premise that this emerging technology has broad and exciting applications in the building design and construction industry will be supported by visual and physical evidence. This evidence will be presented in the form of…
Keil, Andreas; Sabatinelli, Dean; Ding, Mingzhou; Lang, Peter J.; Ihssen, Niklas; Heim, Sabine
2013-01-01
Re-entrant modulation of visual cortex has been suggested as a critical process for enhancing perception of emotionally arousing visual stimuli. This study explores how the time information inherent in large-scale electrocortical measures can be used to examine the functional relationships among the structures involved in emotional perception. Granger causality analysis was conducted on steady-state visual evoked potentials elicited by emotionally arousing pictures flickering at a rate of 10 Hz. This procedure allows one to examine the direction of neural connections. Participants viewed pictures that varied in emotional content, depicting people in neutral contexts, erotica, or interpersonal attack scenes. Results demonstrated increased coupling between visual and cortical areas when viewing emotionally arousing content. Specifically, intraparietal to inferotemporal and precuneus to calcarine connections were stronger for emotionally arousing picture content. Thus, we provide evidence for re-entrant signal flow during emotional perception, which originates from higher tiers and enters lower tiers of visual cortex. PMID:18095279
Imitation and matching of meaningless gestures: distinct involvement from motor and visual imagery.
Lesourd, Mathieu; Navarro, Jordan; Baumard, Josselin; Jarry, Christophe; Le Gall, Didier; Osiurak, François
2017-05-01
The aim of the present study was to understand the underlying cognitive processes of imitation and matching of meaningless gestures. Neuropsychological evidence obtained in brain damaged patients, has shown that distinct cognitive processes supported imitation and matching of meaningless gestures. Left-brain damaged (LBD) patients failed to imitate while right-brain damaged (RBD) patients failed to match meaningless gestures. Moreover, other studies with brain damaged patients showed that LBD patients were impaired in motor imagery while RBD patients were impaired in visual imagery. Thus, we hypothesize that imitation of meaningless gestures might rely on motor imagery, whereas matching of meaningless gestures might be based on visual imagery. In a first experiment, using a correlational design, we demonstrated that posture imitation relies on motor imagery but not on visual imagery (Experiment 1a) and that posture matching relies on visual imagery but not on motor imagery (Experiment 1b). In a second experiment, by manipulating directly the body posture of the participants, we demonstrated that such manipulation evokes a difference only in imitation task but not in matching task. In conclusion, the present study provides direct evidence that the way we imitate or we have to compare postures depends on motor imagery or visual imagery, respectively. Our results are discussed in the light of recent findings about underlying mechanisms of meaningful and meaningless gestures.
Cognitive and psychological science insights to improve climate change data visualization
NASA Astrophysics Data System (ADS)
Harold, Jordan; Lorenzoni, Irene; Shipley, Thomas F.; Coventry, Kenny R.
2016-12-01
Visualization of climate data plays an integral role in the communication of climate change findings to both expert and non-expert audiences. The cognitive and psychological sciences can provide valuable insights into how to improve visualization of climate data based on knowledge of how the human brain processes visual and linguistic information. We review four key research areas to demonstrate their potential to make data more accessible to diverse audiences: directing visual attention, visual complexity, making inferences from visuals, and the mapping between visuals and language. We present evidence-informed guidelines to help climate scientists increase the accessibility of graphics to non-experts, and illustrate how the guidelines can work in practice in the context of Intergovernmental Panel on Climate Change graphics.
Yu, Chen; Smith, Linda B.
2013-01-01
The coordination of visual attention among social partners is central to many components of human behavior and human development. Previous research has focused on one pathway to the coordination of looking behavior by social partners, gaze following. The extant evidence shows that even very young infants follow the direction of another's gaze but they do so only in highly constrained spatial contexts because gaze direction is not a spatially precise cue as to the visual target and not easily used in spatially complex social interactions. Our findings, derived from the moment-to-moment tracking of eye gaze of one-year-olds and their parents as they actively played with toys, provide evidence for an alternative pathway, through the coordination of hands and eyes in goal-directed action. In goal-directed actions, the hands and eyes of the actor are tightly coordinated both temporally and spatially, and thus, in contexts including manual engagement with objects, hand movements and eye movements provide redundant information about where the eyes are looking. Our findings show that one-year-olds rarely look to the parent's face and eyes in these contexts but rather infants and parents coordinate looking behavior without gaze following by attending to objects held by the self or the social partner. This pathway, through eye-hand coupling, leads to coordinated joint switches in visual attention and to an overall high rate of looking at the same object at the same time, and may be the dominant pathway through which physically active toddlers align their looking behavior with a social partner. PMID:24236151
Donderi, Don C
2006-01-01
The idea of visual complexity, the history of its measurement, and its implications for behavior are reviewed, starting with structuralism and Gestalt psychology at the beginning of the 20th century and ending with visual complexity theory, perceptual learning theory, and neural circuit theory at the beginning of the 21st. Evidence is drawn from research on single forms, form and texture arrays and visual displays. Form complexity and form probability are shown to be linked through their reciprocal relationship in complexity theory, which is in turn shown to be consistent with recent developments in perceptual learning and neural circuit theory. Directions for further research are suggested.
Rideaux, Reuben; Apthorp, Deborah; Edwards, Mark
2015-02-12
Recent findings have indicated the capacity to consolidate multiple items into visual short-term memory in parallel varies as a function of the type of information. That is, while color can be consolidated in parallel, evidence suggests that orientation cannot. Here we investigated the capacity to consolidate multiple motion directions in parallel and reexamined this capacity using orientation. This was achieved by determining the shortest exposure duration necessary to consolidate a single item, then examining whether two items, presented simultaneously, could be consolidated in that time. The results show that parallel consolidation of direction and orientation information is possible, and that parallel consolidation of direction appears to be limited to two. Additionally, we demonstrate the importance of adequate separation between feature intervals used to define items when attempting to consolidate in parallel, suggesting that when multiple items are consolidated in parallel, as opposed to serially, the resolution of representations suffer. Finally, we used facilitation of spatial attention to show that the deterioration of item resolution occurs during parallel consolidation, as opposed to storage. © 2015 ARVO.
Neural Circuit to Integrate Opposing Motions in the Visual Field.
Mauss, Alex S; Pankova, Katarina; Arenz, Alexander; Nern, Aljoscha; Rubin, Gerald M; Borst, Alexander
2015-07-16
When navigating in their environment, animals use visual motion cues as feedback signals that are elicited by their own motion. Such signals are provided by wide-field neurons sampling motion directions at multiple image points as the animal maneuvers. Each one of these neurons responds selectively to a specific optic flow-field representing the spatial distribution of motion vectors on the retina. Here, we describe the discovery of a group of local, inhibitory interneurons in the fruit fly Drosophila key for filtering these cues. Using anatomy, molecular characterization, activity manipulation, and physiological recordings, we demonstrate that these interneurons convey direction-selective inhibition to wide-field neurons with opposite preferred direction and provide evidence for how their connectivity enables the computation required for integrating opposing motions. Our results indicate that, rather than sharpening directional selectivity per se, these circuit elements reduce noise by eliminating non-specific responses to complex visual information. Copyright © 2015 Elsevier Inc. All rights reserved.
The locus of impairment in English developmental letter position dyslexia
Kezilas, Yvette; Kohnen, Saskia; McKague, Meredith; Castles, Anne
2014-01-01
Many children with reading difficulties display phonological deficits and struggle to acquire non-lexical reading skills. However, not all children with reading difficulties have these problems, such as children with selective letter position dyslexia (LPD), who make excessive migration errors (such as reading slime as “smile”). Previous research has explored three possible loci for the deficit – the phonological output buffer, the orthographic input lexicon, and the orthographic-visual analysis stage of reading. While there is compelling evidence against a phonological output buffer and orthographic input lexicon deficit account of English LPD, the evidence in support of an orthographic-visual analysis deficit is currently limited. In this multiple single-case study with three English-speaking children with developmental LPD, we aimed to both replicate and extend previous findings regarding the locus of impairment in English LPD. First, we ruled out a phonological output buffer and an orthographic input lexicon deficit by administering tasks that directly assess phonological processing and lexical guessing. We then went on to directly assess whether or not children with LPD have an orthographic-visual analysis deficit by modifying two tasks that have previously been used to localize processing at this level: a same-different decision task and a non-word reading task. The results from these tasks indicate that LPD is most likely caused by a deficit specific to the coding of letter positions at the orthographic-visual analysis stage of reading. These findings provide further evidence for the heterogeneity of dyslexia and its underlying causes. PMID:24917802
Evidence for auditory-visual processing specific to biological motion.
Wuerger, Sophie M; Crocker-Buque, Alexander; Meyer, Georg F
2012-01-01
Biological motion is usually associated with highly correlated sensory signals from more than one modality: an approaching human walker will not only have a visual representation, namely an increase in the retinal size of the walker's image, but also a synchronous auditory signal since the walker's footsteps will grow louder. We investigated whether the multisensorial processing of biological motion is subject to different constraints than ecologically invalid motion. Observers were presented with a visual point-light walker and/or synchronised auditory footsteps; the walker was either approaching the observer (looming motion) or walking away (receding motion). A scrambled point-light walker served as a control. Observers were asked to detect the walker's motion as quickly and as accurately as possible. In Experiment 1 we tested whether the reaction time advantage due to redundant information in the auditory and visual modality is specific for biological motion. We found no evidence for such an effect: the reaction time reduction was accounted for by statistical facilitation for both biological and scrambled motion. In Experiment 2, we dissociated the auditory and visual information and tested whether inconsistent motion directions across the auditory and visual modality yield longer reaction times in comparison to consistent motion directions. Here we find an effect specific to biological motion: motion incongruency leads to longer reaction times only when the visual walker is intact and recognisable as a human figure. If the figure of the walker is abolished by scrambling, motion incongruency has no effect on the speed of the observers' judgments. In conjunction with Experiment 1 this suggests that conflicting auditory-visual motion information of an intact human walker leads to interference and thereby delaying the response.
Pavan, Andrea; Boyce, Matthew; Ghin, Filippo
2016-10-01
Playing action video games enhances visual motion perception. However, there is psychophysical evidence that action video games do not improve motion sensitivity for translational global moving patterns presented in fovea. This study investigates global motion perception in action video game players and compares their performance to that of non-action video game players and non-video game players. Stimuli were random dot kinematograms presented in the parafovea. Observers discriminated the motion direction of a target random dot kinematogram presented in one of the four visual quadrants. Action video game players showed lower motion coherence thresholds than the other groups. However, when the task was performed at threshold, we did not find differences between groups in terms of distributions of reaction times. These results suggest that action video games improve visual motion sensitivity in the near periphery of the visual field, rather than speed response. © The Author(s) 2016.
Absence of Visual Input Results in the Disruption of Grid Cell Firing in the Mouse.
Chen, Guifen; Manson, Daniel; Cacucci, Francesca; Wills, Thomas Joseph
2016-09-12
Grid cells are spatially modulated neurons within the medial entorhinal cortex whose firing fields are arranged at the vertices of tessellating equilateral triangles [1]. The exquisite periodicity of their firing has led to the suggestion that they represent a path integration signal, tracking the organism's position by integrating speed and direction of movement [2-10]. External sensory inputs are required to reset any errors that the path integrator would inevitably accumulate. Here we probe the nature of the external sensory inputs required to sustain grid firing, by recording grid cells as mice explore familiar environments in complete darkness. The absence of visual cues results in a significant disruption of grid cell firing patterns, even when the quality of the directional information provided by head direction cells is largely preserved. Darkness alters the expression of velocity signaling within the entorhinal cortex, with changes evident in grid cell firing rate and the local field potential theta frequency. Short-term (<1.5 s) spike timing relationships between grid cell pairs are preserved in the dark, indicating that network patterns of excitatory and inhibitory coupling between grid cells exist independently of visual input and of spatially periodic firing. However, we find no evidence of preserved hexagonal symmetry in the spatial firing of single grid cells at comparable short timescales. Taken together, these results demonstrate that visual input is required to sustain grid cell periodicity and stability in mice and suggest that grid cells in mice cannot perform accurate path integration in the absence of reliable visual cues. Copyright © 2016 The Author(s). Published by Elsevier Ltd.. All rights reserved.
Evidence for light perception in a bioluminescent organ
Tong, Deyan; Rozas, Natalia S.; Oakley, Todd H.; Mitchell, Jane; Colley, Nansi J.; McFall-Ngai, Margaret J.
2009-01-01
Here we show that bioluminescent organs of the squid Euprymna scolopes possess the molecular, biochemical, and physiological capability for light detection. Transcriptome analyses revealed expression of genes encoding key visual transduction proteins in light-organ tissues, including the same isoform of opsin that occurs in the retina. Electroretinograms demonstrated that the organ responds physiologically to light, and immunocytochemistry experiments localized multiple proteins of visual transduction cascades to tissues housing light-producing bacterial symbionts. These data provide evidence that the light-organ tissues harboring the symbionts serve as extraocular photoreceptors, with the potential to perceive directly the bioluminescence produced by their bacterial partners. PMID:19509343
Primary Visual Cortex as a Saliency Map: A Parameter-Free Prediction and Its Test by Behavioral Data
Zhaoping, Li; Zhe, Li
2015-01-01
It has been hypothesized that neural activities in the primary visual cortex (V1) represent a saliency map of the visual field to exogenously guide attention. This hypothesis has so far provided only qualitative predictions and their confirmations. We report this hypothesis’ first quantitative prediction, derived without free parameters, and its confirmation by human behavioral data. The hypothesis provides a direct link between V1 neural responses to a visual location and the saliency of that location to guide attention exogenously. In a visual input containing many bars, one of them saliently different from all the other bars which are identical to each other, saliency at the singleton’s location can be measured by the shortness of the reaction time in a visual search for singletons. The hypothesis predicts quantitatively the whole distribution of the reaction times to find a singleton unique in color, orientation, and motion direction from the reaction times to find other types of singletons. The prediction matches human reaction time data. A requirement for this successful prediction is a data-motivated assumption that V1 lacks neurons tuned simultaneously to color, orientation, and motion direction of visual inputs. Since evidence suggests that extrastriate cortices do have such neurons, we discuss the possibility that the extrastriate cortices play no role in guiding exogenous attention so that they can be devoted to other functions like visual decoding and endogenous attention. PMID:26441341
Visual imagery without visual perception: lessons from blind subjects
NASA Astrophysics Data System (ADS)
Bértolo, Helder
2014-08-01
The question regarding visual imagery and visual perception remain an open issue. Many studies have tried to understand if the two processes share the same mechanisms or if they are independent, using different neural substrates. Most research has been directed towards the need of activation of primary visual areas during imagery. Here we review some of the works providing evidence for both claims. It seems that studying visual imagery in blind subjects can be used as a way of answering some of those questions, namely if it is possible to have visual imagery without visual perception. We present results from the work of our group using visual activation in dreams and its relation with EEG's spectral components, showing that congenitally blind have visual contents in their dreams and are able to draw them; furthermore their Visual Activation Index is negatively correlated with EEG alpha power. This study supports the hypothesis that it is possible to have visual imagery without visual experience.
Majerus, Steve; Cowan, Nelson; Péters, Frédéric; Van Calster, Laurens; Phillips, Christophe; Schrouff, Jessica
2016-01-01
Recent studies suggest common neural substrates involved in verbal and visual working memory (WM), interpreted as reflecting shared attention-based, short-term retention mechanisms. We used a machine-learning approach to determine more directly the extent to which common neural patterns characterize retention in verbal WM and visual WM. Verbal WM was assessed via a standard delayed probe recognition task for letter sequences of variable length. Visual WM was assessed via a visual array WM task involving the maintenance of variable amounts of visual information in the focus of attention. We trained a classifier to distinguish neural activation patterns associated with high- and low-visual WM load and tested the ability of this classifier to predict verbal WM load (high–low) from their associated neural activation patterns, and vice versa. We observed significant between-task prediction of load effects during WM maintenance, in posterior parietal and superior frontal regions of the dorsal attention network; in contrast, between-task prediction in sensory processing cortices was restricted to the encoding stage. Furthermore, between-task prediction of load effects was strongest in those participants presenting the highest capacity for the visual WM task. This study provides novel evidence for common, attention-based neural patterns supporting verbal and visual WM. PMID:25146374
Six Myths About Spatial Thinking
NASA Astrophysics Data System (ADS)
Newcombe, Nora S.; Stieff, Mike
2012-04-01
Visualizations are an increasingly important part of scientific education and discovery. However, users often do not gain knowledge from them in a complete or efficient way. This article aims to direct research on visualizations in science education in productive directions by reviewing the evidence for widespread assumptions that learning styles, sex differences, developmental stages, and spatial language determine the impact of visualizations on science learning. First, we examine the assumption that people differ in their verbal versus visual learning style. Due to the lack of rigorous evaluation, there is no current support for this distinction. Future research should distinguish between two different kinds of visual learning style. Second, we consider the belief that there are large and intractable sex differences in spatial ability resultant from immutable biological reasons. Although there are some spatial sex differences (in some types of spatial tests although not all), there is actually only very mixed support for biological causation. Most important, there is conclusive evidence that spatial skills can be improved through training and education. Third, we explore educators' use of Piaget's ideas about spatial development to draw conclusions about 'developmental appropriateness'. However, recent research on spatial development has focused on identifying sequences that begin with early starting points of skill, and spatial education is possible in some form at all ages. Fourth, although spatial language does not determine spatial thought, it does frame attention in a way that can have impact on learning and understanding. We examine the empirical support for each assumption and its relevance to future research on visualizations in science education.
Pellicano, Antonello; Koch, Iring; Binkofski, Ferdinand
2017-09-01
An increasing number of studies have shown a close link between perception and action, which is supposed to be responsible for the automatic activation of actions compatible with objects' properties, such as the orientation of their graspable parts. It has been observed that left and right hand responses to objects (e.g., cups) are faster and more accurate if the handle orientation corresponds to the response location than when it does not. Two alternative explanations have been proposed for this handle-to-hand correspondence effect : location coding and affordance activation. The aim of the present study was to provide disambiguating evidence on the origin of this effect by employing object sets for which the visually salient portion was separated from, and opposite to the graspable 1, and vice versa. Seven experiments were conducted employing both single objects and object pairs as visual stimuli to enhance the contextual information about objects' graspability and usability. Notwithstanding these manipulations intended to favor affordance activation, results fully supported the location-coding account displaying significant Simon-like effects that involved the orientation of the visually salient portion of the object stimulus and the location of the response. Crucially, we provided evidence of Simon-like effects based on higher-level cognitive, iconic representations of action directions rather than based on lower-level spatial coding of the pure position of protruding portions of the visual stimuli. (PsycINFO Database Record (c) 2017 APA, all rights reserved).
Saccadic Corollary Discharge Underlies Stable Visual Perception
Berman, Rebecca A.; Joiner, Wilsaan M.; Wurtz, Robert H.
2016-01-01
Saccadic eye movements direct the high-resolution foveae of our retinas toward objects of interest. With each saccade, the image jumps on the retina, causing a discontinuity in visual input. Our visual perception, however, remains stable. Philosophers and scientists over centuries have proposed that visual stability depends upon an internal neuronal signal that is a copy of the neuronal signal driving the eye movement, now referred to as a corollary discharge (CD) or efference copy. In the old world monkey, such a CD circuit for saccades has been identified extending from superior colliculus through MD thalamus to frontal cortex, but there is little evidence that this circuit actually contributes to visual perception. We tested the influence of this CD circuit on visual perception by first training macaque monkeys to report their perceived eye direction, and then reversibly inactivating the CD as it passes through the thalamus. We found that the monkey's perception changed; during CD inactivation, there was a difference between where the monkey perceived its eyes to be directed and where they were actually directed. Perception and saccade were decoupled. We established that the perceived eye direction at the end of the saccade was not derived from proprioceptive input from eye muscles, and was not altered by contextual visual information. We conclude that the CD provides internal information contributing to the brain's creation of perceived visual stability. More specifically, the CD might provide the internal saccade vector used to unite separate retinal images into a stable visual scene. SIGNIFICANCE STATEMENT Visual stability is one of the most remarkable aspects of human vision. The eyes move rapidly several times per second, displacing the retinal image each time. The brain compensates for this disruption, keeping our visual perception stable. A major hypothesis explaining this stability invokes a signal within the brain, a corollary discharge, that informs visual regions of the brain when and where the eyes are about to move. Such a corollary discharge circuit for eye movements has been identified in macaque monkey. We now show that selectively inactivating this brain circuit alters the monkey's visual perception. We conclude that this corollary discharge provides a critical signal that can be used to unite jumping retinal images into a consistent visual scene. PMID:26740647
Attention and multisensory modulation argue against total encapsulation.
de Haas, Benjamin; Schwarzkopf, Dietrich Samuel; Rees, Geraint
2016-01-01
Firestone & Scholl (F&S) postulate that vision proceeds without any direct interference from cognition. We argue that this view is extreme and not in line with the available evidence. Specifically, we discuss two well-established counterexamples: Attention directly affects core aspects of visual processing, and multisensory modulations of vision originate on multiple levels, some of which are unlikely to fall "within perception."
Decoding conjunctions of direction-of-motion and binocular disparity from human visual cortex.
Seymour, Kiley J; Clifford, Colin W G
2012-05-01
Motion and binocular disparity are two features in our environment that share a common correspondence problem. Decades of psychophysical research dedicated to understanding stereopsis suggest that these features interact early in human visual processing to disambiguate depth. Single-unit recordings in the monkey also provide evidence for the joint encoding of motion and disparity across much of the dorsal visual stream. Here, we used functional MRI and multivariate pattern analysis to examine where in the human brain conjunctions of motion and disparity are encoded. Subjects sequentially viewed two stimuli that could be distinguished only by their conjunctions of motion and disparity. Specifically, each stimulus contained the same feature information (leftward and rightward motion and crossed and uncrossed disparity) but differed exclusively in the way these features were paired. Our results revealed that a linear classifier could accurately decode which stimulus a subject was viewing based on voxel activation patterns throughout the dorsal visual areas and as early as V2. This decoding success was conditional on some voxels being individually sensitive to the unique conjunctions comprising each stimulus, thus a classifier could not rely on independent information about motion and binocular disparity to distinguish these conjunctions. This study expands on evidence that disparity and motion interact at many levels of human visual processing, particularly within the dorsal stream. It also lends support to the idea that stereopsis is subserved by early mechanisms also tuned to direction of motion.
Maidenbaum, Shachar; Abboud, Sami; Amedi, Amir
2014-04-01
Sensory substitution devices (SSDs) have come a long way since first developed for visual rehabilitation. They have produced exciting experimental results, and have furthered our understanding of the human brain. Unfortunately, they are still not used for practical visual rehabilitation, and are currently considered as reserved primarily for experiments in controlled settings. Over the past decade, our understanding of the neural mechanisms behind visual restoration has changed as a result of converging evidence, much of which was gathered with SSDs. This evidence suggests that the brain is more than a pure sensory-machine but rather is a highly flexible task-machine, i.e., brain regions can maintain or regain their function in vision even with input from other senses. This complements a recent set of more promising behavioral achievements using SSDs and new promising technologies and tools. All these changes strongly suggest that the time has come to revive the focus on practical visual rehabilitation with SSDs and we chart several key steps in this direction such as training protocols and self-train tools. Copyright © 2014 The Authors. Published by Elsevier Ltd.. All rights reserved.
Directed forgetting of visual symbols: evidence for nonverbal selective rehearsal.
Hourihan, Kathleen L; Ozubko, Jason D; MacLeod, Colin M
2009-12-01
Is selective rehearsal possible for nonverbal information? Two experiments addressed this question using the item method directed forgetting paradigm, where the advantage of remember items over forget items is ascribed to selective rehearsal favoring the remember items. In both experiments, difficult-to-name abstract symbols were presented for study, followed by a recognition test. Directed forgetting effects were evident for these symbols, regardless of whether they were or were not spontaneously named. Critically, a directed forgetting effect was observed for unnamed symbols even when the symbols were studied under verbal suppression to prevent verbal rehearsal. This pattern indicates that a form of nonverbal rehearsal can be used strategically (i.e., selectively) to enhance memory, even when verbal rehearsal is not possible.
Melanopsin-based brightness discrimination in mice and humans.
Brown, Timothy M; Tsujimura, Sei-Ichi; Allen, Annette E; Wynne, Jonathan; Bedford, Robert; Vickery, Graham; Vugler, Anthony; Lucas, Robert J
2012-06-19
Photoreception in the mammalian retina is not restricted to rods and cones but extends to a small number of intrinsically photoreceptive retinal ganglion cells (ipRGCs), expressing the photopigment melanopsin. ipRGCs are known to support various accessory visual functions including circadian photoentrainment and pupillary reflexes. However, despite anatomical and physiological evidence that they contribute to the thalamocortical visual projection, no aspect of visual discrimination has been shown to rely upon ipRGCs. Based on their currently known roles, we hypothesized that ipRGCs may contribute to distinguishing brightness. This percept is related to an object's luminance-a photometric measure of light intensity relevant for cone photoreceptors. However, the perceived brightness of different sources is not always predicted by their respective luminance. Here, we used parallel behavioral and electrophysiological experiments to first show that melanopsin contributes to brightness discrimination in both retinally degenerate and fully sighted mice. We continued to use comparable paradigms in psychophysical experiments to provide evidence for a similar role in healthy human subjects. These data represent the first direct evidence that an aspect of visual discrimination in normally sighted subjects can be supported by inner retinal photoreceptors. Copyright © 2012 Elsevier Ltd. All rights reserved.
Robison, Matthew K; McGuirk, William P; Unsworth, Nash
2017-08-01
The present study examined the relative contributions of the prefrontal cortex (PFC) and posterior parietal cortex (PPC) to visual working memory. Evidence from a number of different techniques has led to the theory that the PFC controls access to working memory (i.e., filtering), determining which information is encoded and maintained for later use whereas the parietal cortex determines how much information is held at 1 given time, regardless of relevance (i.e., capacity; McNab & Klingberg, 2008; Vogel, McCollough, & Machizawa, 2005). To test this theory, we delivered transcranial DC stimulation (tDCS) to the right PFC and right PPC and measured visual working memory capacity and filtering abilities both during and immediately following stimulation. We observed no evidence that tDCS to either the PFC or PPC significantly improved visual working memory. Although the present results did not allow us to make firm theoretical conclusions about the roles of the PFC and PPC in working memory, the results add to the growing body of literature surrounding tDCS and its associated behavioral and neurophysiological effects. (PsycINFO Database Record (c) 2017 APA, all rights reserved).
Ensemble coding remains accurate under object and spatial visual working memory load.
Epstein, Michael L; Emmanouil, Tatiana A
2017-10-01
A number of studies have provided evidence that the visual system statistically summarizes large amounts of information that would exceed the limitations of attention and working memory (ensemble coding). However the necessity of working memory resources for ensemble coding has not yet been tested directly. In the current study, we used a dual task design to test the effect of object and spatial visual working memory load on size averaging accuracy. In Experiment 1, we tested participants' accuracy in comparing the mean size of two sets under various levels of object visual working memory load. Although the accuracy of average size judgments depended on the difference in mean size between the two sets, we found no effect of working memory load. In Experiment 2, we tested the same average size judgment while participants were under spatial visual working memory load, again finding no effect of load on averaging accuracy. Overall our results reveal that ensemble coding can proceed unimpeded and highly accurately under both object and spatial visual working memory load, providing further evidence that ensemble coding reflects a basic perceptual process distinct from that of individual object processing.
Visual Network Asymmetry and Default Mode Network Function in ADHD: An fMRI Study
Hale, T. Sigi; Kane, Andrea M.; Kaminsky, Olivia; Tung, Kelly L.; Wiley, Joshua F.; McGough, James J.; Loo, Sandra K.; Kaplan, Jonas T.
2014-01-01
Background: A growing body of research has identified abnormal visual information processing in attention-deficit hyperactivity disorder (ADHD). In particular, slow processing speed and increased reliance on visuo-perceptual strategies have become evident. Objective: The current study used recently developed fMRI methods to replicate and further examine abnormal rightward biased visual information processing in ADHD and to further characterize the nature of this effect; we tested its association with several large-scale distributed network systems. Method: We examined fMRI BOLD response during letter and location judgment tasks, and directly assessed visual network asymmetry and its association with large-scale networks using both a voxelwise and an averaged signal approach. Results: Initial within-group analyses revealed a pattern of left-lateralized visual cortical activity in controls but right-lateralized visual cortical activity in ADHD children. Direct analyses of visual network asymmetry confirmed atypical rightward bias in ADHD children compared to controls. This ADHD characteristic was atypically associated with reduced activation across several extra-visual networks, including the default mode network (DMN). We also found atypical associations between DMN activation and ADHD subjects’ inattentive symptoms and task performance. Conclusion: The current study demonstrated rightward VNA in ADHD during a simple letter discrimination task. This result adds an important novel consideration to the growing literature identifying abnormal visual processing in ADHD. We postulate that this characteristic reflects greater perceptual engagement of task-extraneous content, and that it may be a basic feature of less efficient top-down task-directed control over visual processing. We additionally argue that abnormal DMN function may contribute to this characteristic. PMID:25076915
Pitch perception deficits in nonverbal learning disability.
Fernández-Prieto, I; Caprile, C; Tinoco-González, D; Ristol-Orriols, B; López-Sala, A; Póo-Argüelles, P; Pons, F; Navarra, J
2016-12-01
The nonverbal learning disability (NLD) is a neurological dysfunction that affects cognitive functions predominantly related to the right hemisphere such as spatial and abstract reasoning. Previous evidence in healthy adults suggests that acoustic pitch (i.e., the relative difference in frequency between sounds) is, under certain conditions, encoded in specific areas of the right hemisphere that also encode the spatial elevation of external objects (e.g., high vs. low position). Taking this evidence into account, we explored the perception of pitch in preadolescents and adolescents with NLD and in a group of healthy participants matched by age, gender, musical knowledge and handedness. Participants performed four speeded tests: a stimulus detection test and three perceptual categorization tests based on colour, spatial position and pitch. Results revealed that both groups were equally fast at detecting visual targets and categorizing visual stimuli according to their colour. In contrast, the NLD group showed slower responses than the control group when categorizing space (direction of a visual object) and pitch (direction of a change in sound frequency). This pattern of results suggests the presence of a subtle deficit at judging pitch in NLD along with the traditionally-described difficulties in spatial processing. Copyright © 2016. Published by Elsevier Ltd.
Aytemür, Ali; Almeida, Nathalia; Lee, Kwang-Hyuk
2017-02-01
Adaptation to delayed sensory feedback following an action produces a subjective time compression between the action and the feedback (temporal recalibration effect, TRE). TRE is important for sensory delay compensation to maintain a relationship between causally related events. It is unclear whether TRE is a sensory modality-specific phenomenon. In 3 experiments employing a sensorimotor synchronization task, we investigated this question using cathodal transcranial direct-current stimulation (tDCS). We found that cathodal tDCS over the visual cortex, and to a lesser extent over the auditory cortex, produced decreased visual TRE. However, both auditory and visual cortex tDCS did not produce any measurable effects on auditory TRE. Our study revealed different nature of TRE in auditory and visual domains. Visual-motor TRE, which is more variable than auditory TRE, is a sensory modality-specific phenomenon, modulated by the auditory cortex. The robustness of auditory-motor TRE, unaffected by tDCS, suggests the dominance of the auditory system in temporal processing, by providing a frame of reference in the realignment of sensorimotor timing signals. Copyright © 2017 Elsevier Ltd. All rights reserved.
Figure-ground segregation at contours: a neural mechanism in the visual cortex of the alert monkey.
Baumann, R; van der Zwan, R; Peterhans, E
1997-06-01
An important task of vision is the segregation of figure and ground in situations of spatial occlusion. Psychophysical evidence suggests that the depth order at contours is defined early in visual processing. We have analysed this process in the visual cortex of the alert monkey. The animals were trained on a visual fixation task which reinforced foveal viewing. During periods of active visual fixation, we recorded the responses of single neurons in striate and prestriate cortex (areas V1, V2, and V3/V3A). The stimuli mimicked situations of spatial occlusion, usually a uniform light (or dark) rectangle overlaying a grating texture of opposite contrast. The direction of figure and ground at the borders of these rectangles was defined by the direction of the terminating grating lines (occlusion cues). Neuronal responses were analysed with respect to figure-ground direction and contrast polarity at such contours. Striate neurons often failed to respond to such stimuli, or were selective for contrast polarity; others were non-selective. Some neurons preferred a certain combination of figure-ground direction and contrast polarity. These neurons were rare both in striate and prestriate cortex. The majority of neurons signalled figure-ground direction independent of contrast polarity. These neurons were only found in prestriate cortex. We explain these responses in terms of a model which also explains neuronal signals of illusory contours. These results suggest that occlusion cues are used at an early level of processing to segregate figure and ground at contours.
Functional specialization and generalization for grouping of stimuli based on colour and motion
Zeki, Semir; Stutters, Jonathan
2013-01-01
This study was undertaken to learn whether the principle of functional specialization that is evident at the level of the prestriate visual cortex extends to areas that are involved in grouping visual stimuli according to attribute, and specifically according to colour and motion. Subjects viewed, in an fMRI scanner, visual stimuli composed of moving dots, which could be either coloured or achromatic; in some stimuli the moving coloured dots were randomly distributed or moved in random directions; in others, some of the moving dots were grouped together according to colour or to direction of motion, with the number of groupings varying from 1 to 3. Increased activation was observed in area V4 in response to colour grouping and in V5 in response to motion grouping while both groupings led to activity in separate though contiguous compartments within the intraparietal cortex. The activity in all the above areas was parametrically related to the number of groupings, as was the prominent activity in Crus I of the cerebellum where the activity resulting from the two types of grouping overlapped. This suggests (a) that, the specialized visual areas of the prestriate cortex have functions beyond the processing of visual signals according to attribute, namely that of grouping signals according to colour (V4) or motion (V5); (b) that the functional separation evident in visual cortical areas devoted to motion and colour, respectively, is maintained at the level of parietal cortex, at least as far as grouping according to attribute is concerned; and (c) that, by contrast, this grouping-related functional segregation is not maintained at the level of the cerebellum. PMID:23415950
Majerus, Steve; Cowan, Nelson; Péters, Frédéric; Van Calster, Laurens; Phillips, Christophe; Schrouff, Jessica
2016-01-01
Recent studies suggest common neural substrates involved in verbal and visual working memory (WM), interpreted as reflecting shared attention-based, short-term retention mechanisms. We used a machine-learning approach to determine more directly the extent to which common neural patterns characterize retention in verbal WM and visual WM. Verbal WM was assessed via a standard delayed probe recognition task for letter sequences of variable length. Visual WM was assessed via a visual array WM task involving the maintenance of variable amounts of visual information in the focus of attention. We trained a classifier to distinguish neural activation patterns associated with high- and low-visual WM load and tested the ability of this classifier to predict verbal WM load (high-low) from their associated neural activation patterns, and vice versa. We observed significant between-task prediction of load effects during WM maintenance, in posterior parietal and superior frontal regions of the dorsal attention network; in contrast, between-task prediction in sensory processing cortices was restricted to the encoding stage. Furthermore, between-task prediction of load effects was strongest in those participants presenting the highest capacity for the visual WM task. This study provides novel evidence for common, attention-based neural patterns supporting verbal and visual WM. © The Author 2014. Published by Oxford University Press. All rights reserved. For Permissions, please e-mail: journals.permissions@oup.com.
Accidental human laser retinal injuries from military laser systems
NASA Astrophysics Data System (ADS)
Stuck, Bruce E.; Zwick, Harry; Molchany, Jerome W.; Lund, David J.; Gagliano, Donald A.
1996-04-01
The time course of the ophthalmoscopic and functional consequences of eight human laser accident cases from military laser systems is described. All patients reported subjective vision loss with ophthalmoscopic evidence of retinal alteration ranging from vitreous hemorrhage to retinal burn. Five of the cases involved single or multiple exposures to Q-switched neodymium radiation at close range whereas the other three incidents occur over large ranges. Most exposures were within 5 degrees of the foveola, yet none directly in the foveola. High contrast visual activity improved with time except in the cases with progressive retinal fibrosis between lesion sites or retinal hole formation encroaching the fovea. In one patient the visual acuity recovered from 20/60 at one week to 20/25 in four months with minimal central visual field loss. Most cases showed suppression of high and low spatial frequency contrast sensitivity. Visual field measurements were enlarged relative to ophthalmoscopic lesion size observations. Deep retinal scar formation and retinal traction were evident in two of the three cases with vitreous hemorrhage. In one patient, nerve fiber layer damage to the papillo-macular bundle was clearly evident. Visual performance measured with a pursuit tracking task revealed significant performance loss relative to normal tracking observers even in cases where acuity returned to near normal levels. These functional and performance deficits may reflect secondary effects of parafoveal laser injury.
Kinesthetic information disambiguates visual motion signals.
Hu, Bo; Knill, David C
2010-05-25
Numerous studies have shown that extra-retinal signals can disambiguate motion information created by movements of the eye or head. We report a new form of cross-modal sensory integration in which the kinesthetic information generated by active hand movements essentially captures ambiguous visual motion information. Several previous studies have shown that active movement can bias observers' percepts of bi-stable stimuli; however, these effects seem to be best explained by attentional mechanisms. We show that kinesthetic information can change an otherwise stable perception of motion, providing evidence of genuine fusion between visual and kinesthetic information. The experiments take advantage of the aperture problem, in which the motion of a one-dimensional grating pattern behind an aperture, while geometrically ambiguous, appears to move stably in the grating normal direction. When actively moving the pattern, however, the observer sees the motion to be in the hand movement direction. Copyright 2010 Elsevier Ltd. All rights reserved.
Goldberg, Melissa C; Mostow, Allison J; Vecera, Shaun P; Larson, Jennifer C Gidley; Mostofsky, Stewart H; Mahone, E Mark; Denckla, Martha B
2008-09-01
We examined the ability to use static line drawings of eye gaze cues to orient visual-spatial attention in children with high functioning autism (HFA) compared to typically developing children (TD). The task was organized such that on valid trials, gaze cues were directed toward the same spatial location as the appearance of an upcoming target, while on invalid trials gaze cues were directed to an opposite location. Unlike TD children, children with HFA showed no advantage in reaction time (RT) on valid trials compared to invalid trials (i.e., no significant validity effect). The two stimulus onset asynchronies (200 ms, 700 ms) did not differentially affect these findings. The results suggest that children with HFA show impairments in utilizing static line drawings of gaze cues to orient visual-spatial attention.
Attention modulates perception of visual space
Zhou, Liu; Deng, Chenglong; Ooi, Teng Leng; He, Zijiang J.
2017-01-01
Attention readily facilitates the detection and discrimination of objects, but it is not known whether it helps to form the vast volume of visual space that contains the objects and where actions are implemented. Conventional wisdom suggests not, given the effortless ease with which we perceive three-dimensional (3D) scenes on opening our eyes. Here, we show evidence to the contrary. In Experiment 1, the observer judged the location of a briefly presented target, placed either on the textured ground or ceiling surface. Judged location was more accurate for a target on the ground, provided that the ground was visible and that the observer directed attention to the lower visual field, not the upper field. This reveals that attention facilitates space perception with reference to the ground. Experiment 2 showed that judged location of a target in mid-air, with both ground and ceiling surfaces present, was more accurate when the observer directed their attention to the lower visual field; this indicates that the attention effect extends to visual space above the ground. These findings underscore the role of attention in anchoring visual orientation in space, which is arguably a primal event that enhances one’s ability to interact with objects and surface layouts within the visual space. The fact that the effect of attention was contingent on the ground being visible suggests that our terrestrial visual system is best served by its ecological niche. PMID:29177198
Visual enhancing of tactile perception in the posterior parietal cortex.
Ro, Tony; Wallace, Ruth; Hagedorn, Judith; Farnè, Alessandro; Pienkos, Elizabeth
2004-01-01
The visual modality typically dominates over our other senses. Here we show that after inducing an extreme conflict in the left hand between vision of touch (present) and the feeling of touch (absent), sensitivity to touch increases for several minutes after the conflict. Transcranial magnetic stimulation of the posterior parietal cortex after this conflict not only eliminated the enduring visual enhancement of touch, but also impaired normal tactile perception. This latter finding demonstrates a direct role of the parietal lobe in modulating tactile perception as a result of the conflict between these senses. These results provide evidence for visual-to-tactile perceptual modulation and demonstrate effects of illusory vision of touch on touch perception through a long-lasting modulatory process in the posterior parietal cortex.
Gallivan, Jason P; Goodale, Melvyn A
2018-01-01
In 1992, Goodale and Milner proposed a division of labor in the visual pathways of the primate cerebral cortex. According to their account, the ventral pathway, which projects to occipitotemporal cortex, constructs our visual percepts, while the dorsal pathway, which projects to posterior parietal cortex, mediates the visual control of action. Although the framing of the two-visual-system hypothesis has not been without controversy, it is clear that vision for action and vision for perception have distinct computational requirements, and significant support for the proposed neuroanatomic division has continued to emerge over the last two decades from human neuropsychology, neuroimaging, behavioral psychophysics, and monkey neurophysiology. In this chapter, we review much of this evidence, with a particular focus on recent findings from human neuroimaging and monkey neurophysiology, demonstrating a specialized role for parietal cortex in visually guided behavior. But even though the available evidence suggests that dedicated circuits mediate action and perception, in order to produce adaptive goal-directed behavior there must be a close coupling and seamless integration of information processing across these two systems. We discuss such ventral-dorsal-stream interactions and argue that the two pathways play different, yet complementary, roles in the production of skilled behavior. Copyright © 2018 Elsevier B.V. All rights reserved.
Differential effects of delay upon visually and haptically guided grasping and perceptual judgments.
Pettypiece, Charles E; Culham, Jody C; Goodale, Melvyn A
2009-05-01
Experiments with visual illusions have revealed a dissociation between the systems that mediate object perception and those responsible for object-directed action. More recently, an experiment on a haptic version of the visual size-contrast illusion has provided evidence for the notion that the haptic modality shows a similar dissociation when grasping and estimating the size of objects in real-time. Here we present evidence suggesting that the similarities between the two modalities begin to break down once a delay is introduced between when people feel the target object and when they perform the grasp or estimation. In particular, when grasping after a delay in a haptic paradigm, people scale their grasps differently when the target is presented with a flanking object of a different size (although the difference does not reflect a size-contrast effect). When estimating after a delay, however, it appears that people ignore the size of the flanking objects entirely. This does not fit well with the results commonly found in visual experiments. Thus, introducing a delay reveals important differences in the way in which haptic and visual memories are stored and accessed.
Visual recovery in cortical blindness is limited by high internal noise
Cavanaugh, Matthew R.; Zhang, Ruyuan; Melnick, Michael D.; Das, Anasuya; Roberts, Mariel; Tadin, Duje; Carrasco, Marisa; Huxlin, Krystel R.
2015-01-01
Damage to the primary visual cortex typically causes cortical blindness (CB) in the hemifield contralateral to the damaged hemisphere. Recent evidence indicates that visual training can partially reverse CB at trained locations. Whereas training induces near-complete recovery of coarse direction and orientation discriminations, deficits in fine motion processing remain. Here, we systematically disentangle components of the perceptual inefficiencies present in CB fields before and after coarse direction discrimination training. In seven human CB subjects, we measured threshold versus noise functions before and after coarse direction discrimination training in the blind field and at corresponding intact field locations. Threshold versus noise functions were analyzed within the framework of the linear amplifier model and the perceptual template model. Linear amplifier model analysis identified internal noise as a key factor differentiating motion processing across the tested areas, with visual training reducing internal noise in the blind field. Differences in internal noise also explained residual perceptual deficits at retrained locations. These findings were confirmed with perceptual template model analysis, which further revealed that the major residual deficits between retrained and intact field locations could be explained by differences in internal additive noise. There were no significant differences in multiplicative noise or the ability to process external noise. Together, these results highlight the critical role of altered internal noise processing in mediating training-induced visual recovery in CB fields, and may explain residual perceptual deficits relative to intact regions of the visual field. PMID:26389544
Infants learn better from left to right: a directional bias in infants' sequence learning.
Bulf, Hermann; de Hevia, Maria Dolores; Gariboldi, Valeria; Macchi Cassia, Viola
2017-05-26
A wealth of studies show that human adults map ordered information onto a directional spatial continuum. We asked whether mapping ordinal information into a directional space constitutes an early predisposition, already functional prior to the acquisition of symbolic knowledge and language. While it is known that preverbal infants represent numerical order along a left-to-right spatial continuum, no studies have investigated yet whether infants, like adults, organize any kind of ordinal information onto a directional space. We investigated whether 7-month-olds' ability to learn high-order rule-like patterns from visual sequences of geometric shapes was affected by the spatial orientation of the sequences (left-to-right vs. right-to-left). Results showed that infants readily learn rule-like patterns when visual sequences were presented from left to right, but not when presented from right to left. This result provides evidence that spatial orientation critically determines preverbal infants' ability to perceive and learn ordered information in visual sequences, opening to the idea that a left-to-right spatially organized mental representation of ordered dimensions might be rooted in biologically-determined constraints on human brain development.
Effect of contrast on the perception of direction of a moving pattern
NASA Technical Reports Server (NTRS)
Stone, L. S.; Watson, A. B.; Mulligan, J. B.
1989-01-01
A series of experiments examining the effect of contrast on the perception of moving plaids was performed to test the hypothesis that the human visual system determines the direction of a moving plaid in a two-staged process: decomposition into component motion followed by application of the intersection-of-contraints rule. Although there is recent evidence that the first tenet of the hypothesis is correct, i.e., that plaid motion is initially decomposed into the motion of the individual grating components, the nature of the second-stage combination rule has not yet been established. It was found that when the gratings within the plaid are of different contrast the preceived direction is not predicted by the intersection-of-constraints rule. There is a strong (up to 20 deg) bias in the direction of the higher-constrast grating. A revised model, which incorporates a contrast-dependent weighting of perceived grating speed as observed for one-dimensional patterns, can quantitatively predict most of the results. The results are then discussed in the context of various models of human visual motion processing and of physiological responses of neurons in the primate visual system.
Eye Contact Is Crucial for Referential Communication in Pet Dogs.
Savalli, Carine; Resende, Briseida; Gaunet, Florence
2016-01-01
Dogs discriminate human direction of attention cues, such as body, gaze, head and eye orientation, in several circumstances. Eye contact particularly seems to provide information on human readiness to communicate; when there is such an ostensive cue, dogs tend to follow human communicative gestures more often. However, little is known about how such cues influence the production of communicative signals (e.g. gaze alternation and sustained gaze) in dogs. In the current study, in order to get an unreachable food, dogs needed to communicate with their owners in several conditions that differ according to the direction of owners' visual cues, namely gaze, head, eyes, and availability to make eye contact. Results provided evidence that pet dogs did not rely on details of owners' direction of visual attention. Instead, they relied on the whole combination of visual cues and especially on the owners' availability to make eye contact. Dogs increased visual communicative behaviors when they established eye contact with their owners, a different strategy compared to apes and baboons, that intensify vocalizations and gestures when human is not visually attending. The difference in strategy is possibly due to distinct status: domesticated vs wild. Results are discussed taking into account the ecological relevance of the task since pet dogs live in human environment and face similar situations on a daily basis during their lives.
The trait of sensory processing sensitivity and neural responses to changes in visual scenes
Xu, Xiaomeng; Aron, Arthur; Aron, Elaine; Cao, Guikang; Feng, Tingyong; Weng, Xuchu
2011-01-01
This exploratory study examined the extent to which individual differences in sensory processing sensitivity (SPS), a temperament/personality trait characterized by social, emotional and physical sensitivity, are associated with neural response in visual areas in response to subtle changes in visual scenes. Sixteen participants completed the Highly Sensitive Person questionnaire, a standard measure of SPS. Subsequently, they were tested on a change detection task while undergoing functional magnetic resonance imaging (fMRI). SPS was associated with significantly greater activation in brain areas involved in high-order visual processing (i.e. right claustrum, left occipitotemporal, bilateral temporal and medial and posterior parietal regions) as well as in the right cerebellum, when detecting minor (vs major) changes in stimuli. These findings remained strong and significant after controlling for neuroticism and introversion, traits that are often correlated with SPS. These results provide the first evidence of neural differences associated with SPS, the first direct support for the sensory aspect of this trait that has been studied primarily for its social and affective implications, and preliminary evidence for heightened sensory processing in individuals high in SPS. PMID:20203139
Self-reflection Orients Visual Attention Downward
Liu, Yi; Tong, Yu; Li, Hong
2017-01-01
Previous research has demonstrated abstract concepts associated with spatial location (e.g., God in the Heavens) could direct visual attention upward or downward, because thinking about the abstract concepts activates the corresponding vertical perceptual symbols. For self-concept, there are similar metaphors (e.g., “I am above others”). However, whether thinking about the self can induce visual attention orientation is still unknown. Therefore, the current study tested whether self-reflection can direct visual attention. Individuals often display the tendency of self-enhancement in social comparison, which reminds the individual of the higher position one possesses relative to others within the social environment. As the individual is the agent of the attention orientation, and high status tends to make an individual look down upon others to obtain a sense of pride, it was hypothesized that thinking about the self would lead to a downward attention orientation. Using reflection of personality traits and a target discrimination task, Study 1 found that, after self-reflection, visual attention was directed downward. Similar effects were also found after friend-reflection, with the level of downward attention being correlated with the likability rating scores of the friend. Thus, in Study 2, a disliked other was used as a control and the positive self-view was measured with above-average judgment task. We found downward attention orientation after self-reflection, but not after reflection upon the disliked other. Moreover, the attentional bias after self-reflection was correlated with above-average self-view. The current findings provide the first evidence that thinking about the self could direct visual-spatial attention downward, and suggest that this effect is probably derived from a positive self-view within the social context. PMID:28928694
Self-reflection Orients Visual Attention Downward.
Liu, Yi; Tong, Yu; Li, Hong
2017-01-01
Previous research has demonstrated abstract concepts associated with spatial location (e.g., God in the Heavens) could direct visual attention upward or downward, because thinking about the abstract concepts activates the corresponding vertical perceptual symbols. For self-concept, there are similar metaphors (e.g., "I am above others"). However, whether thinking about the self can induce visual attention orientation is still unknown. Therefore, the current study tested whether self-reflection can direct visual attention. Individuals often display the tendency of self-enhancement in social comparison, which reminds the individual of the higher position one possesses relative to others within the social environment. As the individual is the agent of the attention orientation, and high status tends to make an individual look down upon others to obtain a sense of pride, it was hypothesized that thinking about the self would lead to a downward attention orientation. Using reflection of personality traits and a target discrimination task, Study 1 found that, after self-reflection, visual attention was directed downward. Similar effects were also found after friend-reflection, with the level of downward attention being correlated with the likability rating scores of the friend. Thus, in Study 2, a disliked other was used as a control and the positive self-view was measured with above-average judgment task. We found downward attention orientation after self-reflection, but not after reflection upon the disliked other. Moreover, the attentional bias after self-reflection was correlated with above-average self-view. The current findings provide the first evidence that thinking about the self could direct visual-spatial attention downward, and suggest that this effect is probably derived from a positive self-view within the social context.
Evidence against the temporal subsampling account of illusory motion reversal
Kline, Keith A.; Eagleman, David M.
2010-01-01
An illusion of reversed motion may occur sporadically while viewing continuous smooth motion. This has been suggested as evidence of discrete temporal sampling by the visual system in analogy to the sampling that generates the wagon–wheel effect on film. In an alternative theory, the illusion is not the result of discrete sampling but instead of perceptual rivalry between appropriately activated and spuriously activated motion detectors. Results of the current study demonstrate that illusory reversals of two spatially overlapping and orthogonal motions often occur separately, providing evidence against the possibility that illusory motion reversal (IMR) is caused by temporal sampling within a visual region. Further, we find that IMR occurs with non-uniform and non-periodic stimuli—an observation that is not accounted for by the temporal sampling hypothesis. We propose, that a motion aftereffect is superimposed on the moving stimulus, sporadically allowing motion detectors for the reverse direction to dominate perception. PMID:18484852
Variability in visual working memory ability limits the efficiency of perceptual decision making.
Ester, Edward F; Ho, Tiffany C; Brown, Scott D; Serences, John T
2014-04-02
The ability to make rapid and accurate decisions based on limited sensory information is a critical component of visual cognition. Available evidence suggests that simple perceptual discriminations are based on the accumulation and integration of sensory evidence over time. However, the memory system(s) mediating this accumulation are unclear. One candidate system is working memory (WM), which enables the temporary maintenance of information in a readily accessible state. Here, we show that individual variability in WM capacity is strongly correlated with the speed of evidence accumulation in speeded two-alternative forced choice tasks. This relationship generalized across different decision-making tasks, and could not be easily explained by variability in general arousal or vigilance. Moreover, we show that performing a difficult discrimination task while maintaining a concurrent memory load has a deleterious effect on the latter, suggesting that WM storage and decision making are directly linked.
The Role of Direct and Visual Force Feedback in Suturing Using a 7-DOF Dual-Arm Teleoperated System.
Talasaz, Ali; Trejos, Ana Luisa; Patel, Rajni V
2017-01-01
The lack of haptic feedback in robotics-assisted surgery can result in tissue damage or accidental tool-tissue hits. This paper focuses on exploring the effect of haptic feedback via direct force reflection and visual presentation of force magnitudes on performance during suturing in robotics-assisted minimally invasive surgery (RAMIS). For this purpose, a haptics-enabled dual-arm master-slave teleoperation system capable of measuring tool-tissue interaction forces in all seven Degrees-of-Freedom (DOFs) was used. Two suturing tasks, tissue puncturing and knot-tightening, were chosen to assess user skills when suturing on phantom tissue. Sixteen subjects participated in the trials and their performance was evaluated from various points of view: force consistency, number of accidental hits with tissue, amount of tissue damage, quality of the suture knot, and the time required to accomplish the task. According to the results, visual force feedback was not very useful during the tissue puncturing task as different users needed different amounts of force depending on the penetration of the needle into the tissue. Direct force feedback, however, was more useful for this task to apply less force and to minimize the amount of damage to the tissue. Statistical results also reveal that both visual and direct force feedback were required for effective knot tightening: direct force feedback could reduce the number of accidental hits with the tissue and also the amount of tissue damage, while visual force feedback could help to securely tighten the suture knots and maintain force consistency among different trials/users. These results provide evidence of the importance of 7-DOF force reflection when performing complex tasks in a RAMIS setting.
Sex differences in adults' relative visual interest in female and male faces, toys, and play styles.
Alexander, Gerianne M; Charles, Nora
2009-06-01
An individual's reproductive potential appears to influence response to attractive faces of the opposite sex. Otherwise, relatively little is known about the characteristics of the adult observer that may influence his or her affective evaluation of male and female faces. An untested hypothesis (based on the proposed role of attractive faces in mate selection) is that most women would show greater interest in male faces whereas most men would show greater interest in female faces. Further, evidence from individuals with preferences for same-sex sexual partners suggests that response to attractive male and female faces may be influenced by gender-linked play preferences. To test these hypotheses, visual attention directed to sex-linked stimuli (faces, toys, play styles) was measured in 39 men and 44 women using eye tracking technology. Consistent with our predictions, men directed greater visual attention to all male-typical stimuli and visual attention to male and female faces was associated with visual attention to gender conforming or nonconforming stimuli in a manner consistent with previous research on sexual orientation. In contrast, women showed a visual preference for female-typical toys, but no visual preference for male faces or female-typical play styles. These findings indicate that sex differences in visual processing extend beyond stimuli associated with adult sexual behavior. We speculate that sex differences in visual processing are a component of the expression of gender phenotypes across the lifespan that may reflect sex differences in the motivational properties of gender-linked stimuli.
van Kerkoerle, Timo; Self, Matthew W.; Dagnino, Bruno; Gariel-Mathis, Marie-Alice; Poort, Jasper; van der Togt, Chris; Roelfsema, Pieter R.
2014-01-01
Cognitive functions rely on the coordinated activity of neurons in many brain regions, but the interactions between cortical areas are not yet well understood. Here we investigated whether low-frequency (α) and high-frequency (γ) oscillations characterize different directions of information flow in monkey visual cortex. We recorded from all layers of the primary visual cortex (V1) and found that γ-waves are initiated in input layer 4 and propagate to the deep and superficial layers of cortex, whereas α-waves propagate in the opposite direction. Simultaneous recordings from V1 and downstream area V4 confirmed that γ- and α-waves propagate in the feedforward and feedback direction, respectively. Microstimulation in V1 elicited γ-oscillations in V4, whereas microstimulation in V4 elicited α-oscillations in V1, thus providing causal evidence for the opposite propagation of these rhythms. Furthermore, blocking NMDA receptors, thought to be involved in feedback processing, suppressed α while boosting γ. These results provide new insights into the relation between brain rhythms and cognition. PMID:25205811
van Kerkoerle, Timo; Self, Matthew W; Dagnino, Bruno; Gariel-Mathis, Marie-Alice; Poort, Jasper; van der Togt, Chris; Roelfsema, Pieter R
2014-10-07
Cognitive functions rely on the coordinated activity of neurons in many brain regions, but the interactions between cortical areas are not yet well understood. Here we investigated whether low-frequency (α) and high-frequency (γ) oscillations characterize different directions of information flow in monkey visual cortex. We recorded from all layers of the primary visual cortex (V1) and found that γ-waves are initiated in input layer 4 and propagate to the deep and superficial layers of cortex, whereas α-waves propagate in the opposite direction. Simultaneous recordings from V1 and downstream area V4 confirmed that γ- and α-waves propagate in the feedforward and feedback direction, respectively. Microstimulation in V1 elicited γ-oscillations in V4, whereas microstimulation in V4 elicited α-oscillations in V1, thus providing causal evidence for the opposite propagation of these rhythms. Furthermore, blocking NMDA receptors, thought to be involved in feedback processing, suppressed α while boosting γ. These results provide new insights into the relation between brain rhythms and cognition.
Community screening for visual impairment in older people.
Clarke, Emily L; Evans, Jennifer R; Smeeth, Liam
2018-02-20
Visual problems in older people are common and frequently under-reported. The effects of poor vision in older people are wide reaching and include falls, confusion and reduced quality of life. Much of the visual impairment in older ages can be treated (e.g. cataract surgery, correction of refractive error). Vision screening may therefore reduce the number of older people living with sight loss. The objective of this review was to assess the effects on vision of community vision screening of older people for visual impairment. We searched the Cochrane Central Register of Controlled Trials (CENTRAL) (which contains the Cochrane Eyes and Vision Trials Register) (2017, Issue 10); Ovid MEDLINE; Ovid Embase; the ISRCTN registry; ClinicalTrials.gov and the ICTRP. The date of the search was 23 November 2017. We included randomised controlled trials (RCTs) that compared vision screening alone or as part of a multi-component screening package as compared to no vision screening or standard care, on the vision of people aged 65 years or over in a community setting. We included trials that used self-reported visual problems or visual acuity testing as the screening tool. We used standard methods expected by Cochrane. We graded the certainty of the evidence using GRADE. Visual outcome data were available for 10,608 people in 10 trials. Four trials took place in the UK, two in Australia, two in the United States and two in the Netherlands. Length of follow-up ranged from one to five years. Three of these studies were cluster-randomised trials whereby general practitioners or family physicians were randomly allocated to undertake vision screening or no vision screening. All studies were funded by government agencies. Overall we judged the studies to be at low risk of bias and only downgraded the certainty of the evidence (GRADE) for imprecision.Seven trials compared vision screening as part of a multi-component screening versus no screening. Six of these studies used self-reported vision as both screening tool and outcome measure, but did not directly measure vision. One study used a combination of self-reported vision and visual acuity measurement: participants reporting vision problems at screening were treated by the attending doctor, referred to an eye care specialist or given information about resources that were available to assist with poor vision. There was a similar risk of "not seeing well" at follow-up in people screened compared with people not screened in meta-analysis of six studies (risk ratio (RR) 1.05, 95% confidence interval (CI) 0.97 to 1.14, 4522 participants high-certainty evidence). One trial reported "improvement in vision" and this occurred slightly less frequently in the screened group (RR 0.85, 95% CI 0.52 to 1.40, 230 participants, moderate-certainty evidence).Two trials compared vision screening (visual acuity testing) alone with no vision screening. In one study, distance visual acuity was similar in the two groups at follow-up (mean difference (MD) 0.02 logMAR, 95% CI -0.02 to 0.05, 532 participants, high-certainty evidence). There was also little difference in near acuity (MD 0.02 logMAR, 95% CI -0.03 to 0.07, 532 participants, high-certainty evidence). There was no evidence of any important difference in quality of life (MD -0.06 National Eye Institute 25-item visual function questionnaire (VFQ-25) score adjusted for baseline VFQ-25 score, 95% CI -2.3 to 1.1, 532 participants, high-certainty evidence). The other study could not be included in the data analysis as the number of participants in each of the arms at follow-up could not be determined. However the authors stated that there was no significant difference in mean visual acuity in participants who had visual acuity assessed at baseline (39 letters) as compared to those who did not have their visual acuity assessed (35 letters, P = 0.25, 121 participants).One trial compared a detailed health assessment including measurement of visual acuity (intervention) with a brief health assessment including one question about vision (standard care). People given the detailed health assessment had a similar risk of visual impairment (visual acuity worse than 6/18 in either eye) at follow-up compared with people given the brief assessment (RR 1.07, 95% CI 0.84 to 1.36, 1807 participants, moderate-certainty evidence). The mean composite score of the VFQ-25 was 86.0 in the group that underwent visual acuity screening compared with 85.6 in the standard care group, a difference of 0.40 (95% CI -1.70 to 2.50, 1807 participants, high-certainty evidence). The evidence from RCTs undertaken to date does not support vision screening for older people living independently in a community setting, whether in isolation or as part of a multi-component screening package. This is true for screening programmes involving questions about visual problems, or direct measurements of visual acuity.The most likely reason for this negative review is that the populations within the trials often did not take up the offered intervention as a result of the vision screening and large proportions of those who did not have vision screening appeared to seek their own intervention. Also, trials that use questions about vision have a lower sensitivity and specificity than formal visual acuity testing. Given the importance of visual impairment among older people, further research into strategies to improve vision of older people is needed. The effectiveness of an optimised primary care-based screening intervention that overcomes possible factors contributing to the observed lack of benefit in trials to date warrants assessment; trials should consider including more dependent participants, rather than those living independently in the community.
Sounds Activate Visual Cortex and Improve Visual Discrimination
Störmer, Viola S.; Martinez, Antigona; McDonald, John J.; Hillyard, Steven A.
2014-01-01
A recent study in humans (McDonald et al., 2013) found that peripheral, task-irrelevant sounds activated contralateral visual cortex automatically as revealed by an auditory-evoked contralateral occipital positivity (ACOP) recorded from the scalp. The present study investigated the functional significance of this cross-modal activation of visual cortex, in particular whether the sound-evoked ACOP is predictive of improved perceptual processing of a subsequent visual target. A trial-by-trial analysis showed that the ACOP amplitude was markedly larger preceding correct than incorrect pattern discriminations of visual targets that were colocalized with the preceding sound. Dipole modeling of the scalp topography of the ACOP localized its neural generators to the ventrolateral extrastriate visual cortex. These results provide direct evidence that the cross-modal activation of contralateral visual cortex by a spatially nonpredictive but salient sound facilitates the discriminative processing of a subsequent visual target event at the location of the sound. Recordings of event-related potentials to the targets support the hypothesis that the ACOP is a neural consequence of the automatic orienting of visual attention to the location of the sound. PMID:25031419
Still holding after all these years: An action-perception dissociation in patient DF.
Ganel, Tzvi; Goodale, Melvyn A
2017-09-23
Patient DF, who has bilateral damage in the ventral visual stream, is perhaps the best known individual with visual form agnosia in the world, and has been the focus of scores of research papers over the past twenty-five years. The remarkable dissociation she exhibits between a profound deficit in perceptual report and a preserved ability to generate relatively normal visuomotor behaviour was early on a cornerstone in Goodale and Milner's (1992) two visual systems hypothesis. In recent years, however, there has been a greater emphasis on the damage that is evident in the posterior regions of her parietal cortex in both hemispheres. Deficits in several aspects of visuomotor control in the visual periphery have been demonstrated, leading some researchers to conclude that the double dissociation between vision-for-perception and vision-for-action in DF and patients with classic optic ataxia can no longer be assumed to be strong evidence for the division of labour between the dorsal and ventral streams of visual processing. In this short review, we argue that this is not the case. Indeed, after evaluating DF's performance and the location of her brain lesions, a clear picture of a double dissociation between DF and patients with optic ataxia is revealed. More than quarter of a century after the initial presentation of DF's unique case, she continues to provide compelling evidence for the idea that the ventral stream is critical for the perception of the shape and orientation of objects but not the visual control of skilled actions directed at those objects. Copyright © 2017 Elsevier Ltd. All rights reserved.
Situated sentence processing: the coordinated interplay account and a neurobehavioral model.
Crocker, Matthew W; Knoeferle, Pia; Mayberry, Marshall R
2010-03-01
Empirical evidence demonstrating that sentence meaning is rapidly reconciled with the visual environment has been broadly construed as supporting the seamless interaction of visual and linguistic representations during situated comprehension. Based on recent behavioral and neuroscientific findings, however, we argue for the more deeply rooted coordination of the mechanisms underlying visual and linguistic processing, and for jointly considering the behavioral and neural correlates of scene-sentence reconciliation during situated comprehension. The Coordinated Interplay Account (CIA; Knoeferle, P., & Crocker, M. W. (2007). The influence of recent scene events on spoken comprehension: Evidence from eye movements. Journal of Memory and Language, 57(4), 519-543) asserts that incremental linguistic interpretation actively directs attention in the visual environment, thereby increasing the salience of attended scene information for comprehension. We review behavioral and neuroscientific findings in support of the CIA's three processing stages: (i) incremental sentence interpretation, (ii) language-mediated visual attention, and (iii) the on-line influence of non-linguistic visual context. We then describe a recently developed connectionist model which both embodies the central CIA proposals and has been successfully applied in modeling a range of behavioral findings from the visual world paradigm (Mayberry, M. R., Crocker, M. W., & Knoeferle, P. (2009). Learning to attend: A connectionist model of situated language comprehension. Cognitive Science). Results from a new simulation suggest the model also correlates with event-related brain potentials elicited by the immediate use of visual context for linguistic disambiguation (Knoeferle, P., Habets, B., Crocker, M. W., & Münte, T. F. (2008). Visual scenes trigger immediate syntactic reanalysis: Evidence from ERPs during situated spoken comprehension. Cerebral Cortex, 18(4), 789-795). Finally, we argue that the mechanisms underlying interpretation, visual attention, and scene apprehension are not only in close temporal synchronization, but have co-adapted to optimize real-time visual grounding of situated spoken language, thus facilitating the association of linguistic, visual and motor representations that emerge during the course of our embodied linguistic experience in the world. Copyright 2009 Elsevier Inc. All rights reserved.
Reaching to virtual targets: The oblique effect reloaded in 3-D.
Kaspiris-Rousellis, Christos; Siettos, Constantinos I; Evdokimidis, Ioannis; Smyrnis, Nikolaos
2017-02-20
Perceiving and reproducing direction of visual stimuli in 2-D space produces the visual oblique effect, which manifests as increased precision in the reproduction of cardinal compared to oblique directions. A second cognitive oblique effect emerges when stimulus information is degraded (such as when reproducing stimuli from memory) and manifests as a systematic distortion where reproduced directions close to the cardinal axes deviate toward the oblique, leading to space expansion at cardinal and contraction at oblique axes. We studied the oblique effect in 3-D using a virtual reality system to present a large number of stimuli, covering the surface of an imaginary half sphere, to which subjects had to reach. We used two conditions, one with no delay (no-memory condition) and one where a three-second delay intervened between stimulus presentation and movement initiation (memory condition). A visual oblique effect was observed for the reproduction of cardinal directions compared to oblique, which did not differ with memory condition. A cognitive oblique effect also emerged, which was significantly larger in the memory compared to the no-memory condition, leading to distortion of directional space with expansion near the cardinal axes and compression near the oblique axes on the hemispherical surface. This effect provides evidence that existing models of 2-D directional space categorization could be extended in the natural 3-D space. Copyright © 2016 IBRO. Published by Elsevier Ltd. All rights reserved.
Infants' visual and auditory communication when a partner is or is not visually attending.
Liszkowski, Ulf; Albrecht, Konstanze; Carpenter, Malinda; Tomasello, Michael
2008-04-01
In the current study we investigated infants' communication in the visual and auditory modalities as a function of the recipient's visual attention. We elicited pointing at interesting events from thirty-two 12-month olds and thirty-two 18-month olds in two conditions: when the recipient either was or was not visually attending to them before and during the point. The main result was that infants initiated more pointing when the recipient's visual attention was on them than when it was not. In addition, when the recipient did not respond by sharing interest in the designated event, infants initiated more repairs (repeated pointing) than when she did, again, especially when the recipient was visually attending to them. Interestingly, accompanying vocalizations were used intentionally and increased in both experimental conditions when the recipient did not share attention and interest. However, there was little evidence that infants used their vocalizations to direct attention to their gestures when the recipient was not attending to them.
Liu, Han-Hsuan
2016-01-01
Fragile X mental retardation protein (FMRP) is thought to regulate neuronal plasticity by limiting dendritic protein synthesis, but direct demonstration of a requirement for FMRP control of local protein synthesis during behavioral plasticity is lacking. Here we tested whether FMRP knockdown in Xenopus optic tectum affects local protein synthesis in vivo and whether FMRP knockdown affects protein synthesis-dependent visual avoidance behavioral plasticity. We tagged newly synthesized proteins by incorporation of the noncanonical amino acid azidohomoalanine and visualized them with fluorescent noncanonical amino acid tagging (FUNCAT). Visual conditioning and FMRP knockdown produce similar increases in FUNCAT in tectal neuropil. Induction of visual conditioning-dependent behavioral plasticity occurs normally in FMRP knockdown animals, but plasticity degrades over 24 h. These results indicate that FMRP affects visual conditioning-induced local protein synthesis and is required to maintain the visual conditioning-induced behavioral plasticity. SIGNIFICANCE STATEMENT Fragile X syndrome (FXS) is the most common form of inherited intellectual disability. Exaggerated dendritic protein synthesis resulting from loss of fragile X mental retardation protein (FMRP) is thought to underlie cognitive deficits in FXS, but no direct evidence has demonstrated that FMRP-regulated dendritic protein synthesis affects behavioral plasticity in intact animals. Xenopus tadpoles exhibit a visual avoidance behavior that improves with visual conditioning in a protein synthesis-dependent manner. We showed that FMRP knockdown and visual conditioning dramatically increase protein synthesis in neuronal processes. Furthermore, induction of visual conditioning-dependent behavioral plasticity occurs normally after FMRP knockdown, but performance rapidly deteriorated in the absence of FMRP. These studies show that FMRP negatively regulates local protein synthesis and is required to maintain visual conditioning-induced behavioral plasticity in vivo. PMID:27383604
Liu, Han-Hsuan; Cline, Hollis T
2016-07-06
Fragile X mental retardation protein (FMRP) is thought to regulate neuronal plasticity by limiting dendritic protein synthesis, but direct demonstration of a requirement for FMRP control of local protein synthesis during behavioral plasticity is lacking. Here we tested whether FMRP knockdown in Xenopus optic tectum affects local protein synthesis in vivo and whether FMRP knockdown affects protein synthesis-dependent visual avoidance behavioral plasticity. We tagged newly synthesized proteins by incorporation of the noncanonical amino acid azidohomoalanine and visualized them with fluorescent noncanonical amino acid tagging (FUNCAT). Visual conditioning and FMRP knockdown produce similar increases in FUNCAT in tectal neuropil. Induction of visual conditioning-dependent behavioral plasticity occurs normally in FMRP knockdown animals, but plasticity degrades over 24 h. These results indicate that FMRP affects visual conditioning-induced local protein synthesis and is required to maintain the visual conditioning-induced behavioral plasticity. Fragile X syndrome (FXS) is the most common form of inherited intellectual disability. Exaggerated dendritic protein synthesis resulting from loss of fragile X mental retardation protein (FMRP) is thought to underlie cognitive deficits in FXS, but no direct evidence has demonstrated that FMRP-regulated dendritic protein synthesis affects behavioral plasticity in intact animals. Xenopus tadpoles exhibit a visual avoidance behavior that improves with visual conditioning in a protein synthesis-dependent manner. We showed that FMRP knockdown and visual conditioning dramatically increase protein synthesis in neuronal processes. Furthermore, induction of visual conditioning-dependent behavioral plasticity occurs normally after FMRP knockdown, but performance rapidly deteriorated in the absence of FMRP. These studies show that FMRP negatively regulates local protein synthesis and is required to maintain visual conditioning-induced behavioral plasticity in vivo. Copyright © 2016 the authors 0270-6474/16/367325-15$15.00/0.
Plow, Ela B; Obretenova, Souzana N; Jackson, Mary Lou; Merabet, Lotfi B
2012-07-01
We have previously reported that transcranial direct current stimulation (tDCS) delivered to the occipital cortex enhances visual functional recovery when combined with three months of computer-based rehabilitative training in patients with hemianopia. The principal objective of this study was to evaluate the temporal sequence of effects of tDCS on visual recovery as they appear over the course of training and across different indicators of visual function. Primary objective outcome measures were 1) shifts in visual field border and 2) stimulus detection accuracy within the affected hemifield. These were compared between patients randomized to either vision restoration therapy (VRT) combined with active tDCS or VRT paired with sham tDCS. Training comprised two half-hour sessions, three times a week for three months. Primary outcome measures were collected at baseline (pretest), monthly interim intervals, and at posttest (three months). As secondary outcome measures, contrast sensitivity and reading performance were collected at pretest and posttest time points only. Active tDCS combined with VRT accelerated the recovery of stimulus detection as between-group differences appeared within the first month of training. In contrast, a shift in the visual field border was only evident at posttest (after three months of training). tDCS did not affect contrast sensitivity or reading performance. These results suggest that tDCS may differentially affect the magnitude and sequence of visual recovery in a manner that is task specific to the type of visual rehabilitative training strategy employed. © 2012 International Neuromodulation Society.
Just one look: Direct gaze briefly disrupts visual working memory.
Wang, J Jessica; Apperly, Ian A
2017-04-01
Direct gaze is a salient social cue that affords rapid detection. A body of research suggests that direct gaze enhances performance on memory tasks (e.g., Hood, Macrae, Cole-Davies, & Dias, Developmental Science, 1, 67-71, 2003). Nonetheless, other studies highlight the disruptive effect direct gaze has on concurrent cognitive processes (e.g., Conty, Gimmig, Belletier, George, & Huguet, Cognition, 115(1), 133-139, 2010). This discrepancy raises questions about the effects direct gaze may have on concurrent memory tasks. We addressed this topic by employing a change detection paradigm, where participants retained information about the color of small sets of agents. Experiment 1 revealed that, despite the irrelevance of the agents' eye gaze to the memory task at hand, participants were worse at detecting changes when the agents looked directly at them compared to when the agents looked away. Experiment 2 showed that the disruptive effect was relatively short-lived. Prolonged presentation of direct gaze led to recovery from the initial disruption, rather than a sustained disruption on change detection performance. The present study provides the first evidence that direct gaze impairs visual working memory with a rapidly-developing yet short-lived effect even when there is no need to attend to agents' gaze.
Coordinates of Human Visual and Inertial Heading Perception.
Crane, Benjamin Thomas
2015-01-01
Heading estimation involves both inertial and visual cues. Inertial motion is sensed by the labyrinth, somatic sensation by the body, and optic flow by the retina. Because the eye and head are mobile these stimuli are sensed relative to different reference frames and it remains unclear if a perception occurs in a common reference frame. Recent neurophysiologic evidence has suggested the reference frames remain separate even at higher levels of processing but has not addressed the resulting perception. Seven human subjects experienced a 2s, 16 cm/s translation and/or a visual stimulus corresponding with this translation. For each condition 72 stimuli (360° in 5° increments) were delivered in random order. After each stimulus the subject identified the perceived heading using a mechanical dial. Some trial blocks included interleaved conditions in which the influence of ±28° of gaze and/or head position were examined. The observations were fit using a two degree-of-freedom population vector decoder (PVD) model which considered the relative sensitivity to lateral motion and coordinate system offset. For visual stimuli gaze shifts caused shifts in perceived head estimates in the direction opposite the gaze shift in all subjects. These perceptual shifts averaged 13 ± 2° for eye only gaze shifts and 17 ± 2° for eye-head gaze shifts. This finding indicates visual headings are biased towards retina coordinates. Similar gaze and head direction shifts prior to inertial headings had no significant influence on heading direction. Thus inertial headings are perceived in body-centered coordinates. Combined visual and inertial stimuli yielded intermediate results.
Coordinates of Human Visual and Inertial Heading Perception
Crane, Benjamin Thomas
2015-01-01
Heading estimation involves both inertial and visual cues. Inertial motion is sensed by the labyrinth, somatic sensation by the body, and optic flow by the retina. Because the eye and head are mobile these stimuli are sensed relative to different reference frames and it remains unclear if a perception occurs in a common reference frame. Recent neurophysiologic evidence has suggested the reference frames remain separate even at higher levels of processing but has not addressed the resulting perception. Seven human subjects experienced a 2s, 16 cm/s translation and/or a visual stimulus corresponding with this translation. For each condition 72 stimuli (360° in 5° increments) were delivered in random order. After each stimulus the subject identified the perceived heading using a mechanical dial. Some trial blocks included interleaved conditions in which the influence of ±28° of gaze and/or head position were examined. The observations were fit using a two degree-of-freedom population vector decoder (PVD) model which considered the relative sensitivity to lateral motion and coordinate system offset. For visual stimuli gaze shifts caused shifts in perceived head estimates in the direction opposite the gaze shift in all subjects. These perceptual shifts averaged 13 ± 2° for eye only gaze shifts and 17 ± 2° for eye-head gaze shifts. This finding indicates visual headings are biased towards retina coordinates. Similar gaze and head direction shifts prior to inertial headings had no significant influence on heading direction. Thus inertial headings are perceived in body-centered coordinates. Combined visual and inertial stimuli yielded intermediate results. PMID:26267865
Visual search, visual streams, and visual architectures.
Green, M
1991-10-01
Most psychological, physiological, and computational models of early vision suggest that retinal information is divided into a parallel set of feature modules. The dominant theories of visual search assume that these modules form a "blackboard" architecture: a set of independent representations that communicate only through a central processor. A review of research shows that blackboard-based theories, such as feature-integration theory, cannot easily explain the existing data. The experimental evidence is more consistent with a "network" architecture, which stresses that: (1) feature modules are directly connected to one another, (2) features and their locations are represented together, (3) feature detection and integration are not distinct processing stages, and (4) no executive control process, such as focal attention, is needed to integrate features. Attention is not a spotlight that synthesizes objects from raw features. Instead, it is better to conceptualize attention as an aperture which masks irrelevant visual information.
Zhang, Suge; Sun, Hongxia; Chen, Hongbo; Li, Qian; Guan, Aijiao; Wang, Lixia; Shi, Yunhua; Xu, Shujuan; Liu, Meirong; Tang, Yalin
2018-05-01
Direct detection of G-quadruplexes in human cells has become an important issue due to the vital role of G-quadruplex related to biological functions. Despite several probes have been developed for detection of the G-quadruplexes in cytoplasm or whole cells, the probe being used to monitor the nucleolar G-quadruplexes is still lacking. Formation of the nucleolar G-quadruplex structures was confirmed by using circular dichroism (CD) spectroscopy. The binding affinity and selectivity of Thioflavin T (ThT) towards various DNA/RNA motifs in solution and gel system were measured by using fluorescence spectroscopy and polyacrylamide gel electrophoresis (PAGE), respectively. G-quadruplex imaging in live cells was directly captured by using confocal laser scanning microscopy (CLSM). Formation of the rDNA and rRNA G-quadruplex structures is demonstrated in vitro. ThT is found to show much higher affinity and selectivity towards these G-quadruplex structures versus other nucleic acid motifs either in solution or in gel system. The nucleolar G-quadruplexes in living cells are visualized by using ThT as a fluorescent probe. G-quadruplex-ligand treatments in live cells lead to sharp decrease of ThT signal. The natural existence of the G-quadruplexes structure in the nucleoli of living cells is directly visualized by using ThT as an indicator. The research provides substantive evidence for formation of the rRNA G-quadruplex structures, and also offers an effective probe for direct visualization of the nucleolar G-quadruplexes in living cells. Copyright © 2018. Published by Elsevier B.V.
Touch to see: neuropsychological evidence of a sensory mirror system for touch.
Bolognini, Nadia; Olgiati, Elena; Xaiz, Annalisa; Posteraro, Lucio; Ferraro, Francesco; Maravita, Angelo
2012-09-01
The observation of touch can be grounded in the activation of brain areas underpinning direct tactile experience, namely the somatosensory cortices. What is the behavioral impact of such a mirror sensory activity on visual perception? To address this issue, we investigated the causal interplay between observed and felt touch in right brain-damaged patients, as a function of their underlying damaged visual and/or tactile modalities. Patients and healthy controls underwent a detection task, comprising visual stimuli depicting touches or without a tactile component. Touch and No-touch stimuli were presented in egocentric or allocentric perspectives. Seeing touches, regardless of the viewing perspective, differently affects visual perception depending on which sensory modality is damaged: In patients with a selective visual deficit, but without any tactile defect, the sight of touch improves the visual impairment; this effect is associated with a lesion to the supramarginal gyrus. In patients with a tactile deficit, but intact visual perception, the sight of touch disrupts visual processing, inducing a visual extinction-like phenomenon. This disruptive effect is associated with the damage of the postcentral gyrus. Hence, a damage to the somatosensory system can lead to a dysfunctional visual processing, and an intact somatosensory processing can aid visual perception.
Schindler, Andreas; Bartels, Andreas
2017-05-01
Superimposed on the visual feed-forward pathway, feedback connections convey higher level information to cortical areas lower in the hierarchy. A prominent framework for these connections is the theory of predictive coding where high-level areas send stimulus interpretations to lower level areas that compare them with sensory input. Along these lines, a growing body of neuroimaging studies shows that predictable stimuli lead to reduced blood oxygen level-dependent (BOLD) responses compared with matched nonpredictable counterparts, especially in early visual cortex (EVC) including areas V1-V3. The sources of these modulatory feedback signals are largely unknown. Here, we re-examined the robust finding of relative BOLD suppression in EVC evident during processing of coherent compared with random motion. Using functional connectivity analysis, we show an optic flow-dependent increase of functional connectivity between BOLD suppressed EVC and a network of visual motion areas including MST, V3A, V6, the cingulate sulcus visual area (CSv), and precuneus (Pc). Connectivity decreased between EVC and 2 areas known to encode heading direction: entorhinal cortex (EC) and retrosplenial cortex (RSC). Our results provide first evidence that BOLD suppression in EVC for predictable stimuli is indeed mediated by specific high-level areas, in accord with the theory of predictive coding. © The Author 2016. Published by Oxford University Press. All rights reserved. For Permissions, please e-mail: journals.permissions@oup.com.
Madsen, Sarah K.; Bohon, Cara; Feusner, Jamie D.
2013-01-01
Anorexia nervosa (AN) and body dysmorphic disorder (BDD) are psychiatric disorders that involve distortion of the experience of one’s physical appearance. In AN, individuals believe that they are overweight, perceive their body as “fat,” and are preoccupied with maintaining a low body weight. In BDD, individuals are preoccupied with misperceived defects in physical appearance, most often of the face. Distorted visual perception may contribute to these cardinal symptoms, and may be a common underlying phenotype. This review surveys the current literature on visual processing in AN and BDD, addressing lower- to higher-order stages of visual information processing and perception. We focus on peer-reviewed studies of AN and BDD that address ophthalmologic abnormalities, basic neural processing of visual input, integration of visual input with other systems, neuropsychological tests of visual processing, and representations of whole percepts (such as images of faces, bodies, and other objects). The literature suggests a pattern in both groups of over-attention to detail, reduced processing of global features, and a tendency to focus on symptom-specific details in their own images (body parts in AN, facial features in BDD), with cognitive strategy at least partially mediating the abnormalities. Visuospatial abnormalities were also evident when viewing images of others and for non-appearance related stimuli. Unfortunately no study has directly compared AN and BDD, and most studies were not designed to disentangle disease-related emotional responses from lower-order visual processing. We make recommendations for future studies to improve the understanding of visual processing abnormalities in AN and BDD. PMID:23810196
Pauers, Michael J; McKinnon, Jeffrey S; Ehlinger, Timothy J
2004-12-07
Speciation via intersexual selection on male nuptial colour pattern is thought to have been a major force in promoting the explosive speciation of African haplochromine cichlids, yet there is very little direct empirical evidence of directional preferences within populations. In this study, we used objective spectrophotometry and analyses based on visual physiology to determine whether females of the Katale population of Labeotropheus fuelleborni, a Lake Malawi haplochromine, prefer males that have higher chroma and more within-pattern colour contrast. In paired male preference tests, female Katale L. fuelleborni showed increasing preferences for males with more relatively saturated colours on their flanks. They also showed increasing preferences for males with relatively higher contrast levels among flank elements. This is the first empirical evidence, to our knowledge, for male colour as a directionally sexually selected trait within a haplochromine cichlid population.
Pauers, Michael J; McKinnon, Jeffrey S; Ehlinger, Timothy J
2004-01-01
Speciation via intersexual selection on male nuptial colour pattern is thought to have been a major force in promoting the explosive speciation of African haplochromine cichlids, yet there is very little direct empirical evidence of directional preferences within populations. In this study, we used objective spectrophotometry and analyses based on visual physiology to determine whether females of the Katale population of Labeotropheus fuelleborni, a Lake Malawi haplochromine, prefer males that have higher chroma and more within-pattern colour contrast. In paired male preference tests, female Katale L. fuelleborni showed increasing preferences for males with more relatively saturated colours on their flanks. They also showed increasing preferences for males with relatively higher contrast levels among flank elements. This is the first empirical evidence, to our knowledge, for male colour as a directionally sexually selected trait within a haplochromine cichlid population. PMID:15801599
Orienting Attention to Sound Object Representations Attenuates Change Deafness
ERIC Educational Resources Information Center
Backer, Kristina C.; Alain, Claude
2012-01-01
According to the object-based account of attention, multiple objects coexist in short-term memory (STM), and we can selectively attend to a particular object of interest. Although there is evidence that attention can be directed to visual object representations, the assumption that attention can be oriented to sound object representations has yet…
New insights into ambient and focal visual fixations using an automatic classification algorithm
Follet, Brice; Le Meur, Olivier; Baccino, Thierry
2011-01-01
Overt visual attention is the act of directing the eyes toward a given area. These eye movements are characterised by saccades and fixations. A debate currently surrounds the role of visual fixations. Do they all have the same role in the free viewing of natural scenes? Recent studies suggest that at least two types of visual fixations exist: focal and ambient. The former is believed to be used to inspect local areas accurately, whereas the latter is used to obtain the context of the scene. We investigated the use of an automated system to cluster visual fixations in two groups using four types of natural scene images. We found new evidence to support a focal–ambient dichotomy. Our data indicate that the determining factor is the saccade amplitude. The dependence on the low-level visual features and the time course of these two kinds of visual fixations were examined. Our results demonstrate that there is an interplay between both fixation populations and that focal fixations are more dependent on low-level visual features than are ambient fixations. PMID:23145248
Vernier But Not Grating Acuity Contributes to an Early Stage of Visual Word Processing.
Tan, Yufei; Tong, Xiuhong; Chen, Wei; Weng, Xuchu; He, Sheng; Zhao, Jing
2018-03-28
The process of reading words depends heavily on efficient visual skills, including analyzing and decomposing basic visual features. Surprisingly, previous reading-related studies have almost exclusively focused on gross aspects of visual skills, while only very few have investigated the role of finer skills. The present study filled this gap and examined the relations of two finer visual skills measured by grating acuity (the ability to resolve periodic luminance variations across space) and Vernier acuity (the ability to detect/discriminate relative locations of features) to Chinese character-processing as measured by character form-matching and lexical decision tasks in skilled adult readers. The results showed that Vernier acuity was significantly correlated with performance in character form-matching but not visual symbol form-matching, while no correlation was found between grating acuity and character processing. Interestingly, we found no correlation of the two visual skills with lexical decision performance. These findings provide for the first time empirical evidence that the finer visual skills, particularly as reflected in Vernier acuity, may directly contribute to an early stage of hierarchical word processing.
May, Paul J.; McHaffie, John G.; Stanford, Terrence R.; Jiang, Huai; Costello, M. Gabriela; Coizet, Veronique; Hayes, Lauren M.; Haber, Suzanne N.; Redgrave, Peter
2010-01-01
Much of the evidence linking the short-latency phasic signaling of midbrain dopaminergic neurons with reward-prediction errors used in learning and habit formation comes from recording the visual responses of monkey dopaminergic neurons. However, the information encoded by dopaminergic neuron activity is constrained by the qualities of the afferent visual signals made available to these cells. Recent evidence from rats and cats indicates the primary source of this visual input originates subcortically, via a direct tectonigral projection. The present anatomical study sought to establish whether a direct tectonigral projection is a significant feature of the primate brain. Injections of anterograde tracers into the superior colliculus of macaque monkeys labelled terminal arbors throughout the substantia nigra, with the densest terminations in the dorsal tier. Labelled boutons were found in close association (possibly indicative of synaptic contact) with ventral midbrain neurons staining positively for the dopaminergic marker tyrosine hydroxylase. Injections of retrograde tracer confined to the macaque substantia nigra retrogradely labelled small to medium sized neurons in the intermediate and deep layers of the superior colliculus. Together, these data indicate that a direct tectonigral projection is also a feature of the monkey brain, and therefore likely to have been conserved throughout mammalian evolution. Insofar as the superior colliculus is configured to detect unpredicted, biologically salient, sensory events, it may be safer to regard the phasic responses of midbrain dopaminergic neurons as ‘sensory prediction errors’ rather than ‘reward prediction errors’, in which case, dopamine-based theories of reinforcement learning will require revision. PMID:19175405
Direct evidence for a position input to the smooth pursuit system.
Blohm, Gunnar; Missal, Marcus; Lefèvre, Philippe
2005-07-01
When objects move in our environment, the orientation of the visual axis in space requires the coordination of two types of eye movements: saccades and smooth pursuit. The principal input to the saccadic system is position error, whereas it is velocity error for the smooth pursuit system. Recently, it has been shown that catch-up saccades to moving targets are triggered and programmed by using velocity error in addition to position error. Here, we show that, when a visual target is flashed during ongoing smooth pursuit, it evokes a smooth eye movement toward the flash. The velocity of this evoked smooth movement is proportional to the position error of the flash; it is neither influenced by the velocity of the ongoing smooth pursuit eye movement nor by the occurrence of a saccade, but the effect is absent if the flash is ignored by the subject. Furthermore, the response started around 85 ms after the flash presentation and decayed with an average time constant of 276 ms. Thus this is the first direct evidence of a position input to the smooth pursuit system. This study shows further evidence for a coupling between saccadic and smooth pursuit systems. It also suggests that there is an interaction between position and velocity error signals in the control of more complex movements.
The Perceptual and Social Components of Metacognition
2016-01-01
When deciding whether or not to bring an umbrella to work, your confidence will be influenced by the sky outside the window (direct evidence) as well as by, for example, whether or not people walking in the street have their own umbrella (indirect or contingent evidence). These 2 distinct aspects of decision confidence have not yet been assessed independently within the same framework. Here we study the relative contributions of stimulus-specific and social-contingent information on confidence formation. Dyads of participants made visual perceptual decisions, first individually and then together by sharing their wagers in their decisions. We independently manipulated the sensory evidence and the social consensus available to participants and found that both type of evidence contributed to wagers. Consistent with previous work, the amount people were prepared to wager covaried with the strength of sensory evidence. However, social agreements and disagreement affected wagers in opposite directions and asymmetrically. These different contributions of sensory and social evidence to wager were linearly additive. Moreover, average metacognitive sensitivity—namely the association between wagers and accuracy—between interacting dyad members positively correlated with dyadic performance and dyadic benefit above average individual performance. Our results provide a general framework that accounts for how both social context and direct sensory evidence contribute to decision confidence. PMID:27454040
Mann, David L; Abernethy, Bruce; Farrow, Damian
2010-07-01
Coupled interceptive actions are understood to be the result of neural processing-and visual information-which is distinct from that used for uncoupled perceptual responses. To examine the visual information used for action and perception, skilled cricket batters anticipated the direction of balls bowled toward them using a coupled movement (an interceptive action that preserved the natural coupling between perception and action) or an uncoupled (verbal) response, in each of four different visual blur conditions (plano, +1.00, +2.00, +3.00). Coupled responses were found to be better than uncoupled ones, with the blurring of vision found to result in different effects for the coupled and uncoupled response conditions. Low levels of visual blur did not affect coupled anticipation, a finding consistent with the comparatively poorer visual information on which online interceptive actions are proposed to rely. In contrast, some evidence was found to suggest that low levels of blur may enhance the uncoupled verbal perception of movement.
Neuro-ophthalmic manifestations of cerebrovascular accidents.
Ghannam, Alaa S Bou; Subramanian, Prem S
2017-11-01
Ocular functions can be affected in almost any type of cerebrovascular accident (CVA) creating a burden on the patient and family and limiting functionality. The present review summarizes the different ocular outcomes after stroke, divided into three categories: vision, ocular motility, and visual perception. We also discuss interventions that have been proposed to help restore vision and perception after CVA. Interventions that might help expand or compensate for visual field loss and visuospatial neglect include explorative saccade training, prisms, visual restoration therapy (VRT), and transcranial direct current stimulation (tDCS). VRT makes use of neuroplasticity, which has shown efficacy in animal models but remains controversial in human studies. CVAs can lead to decreased visual acuity, visual field loss, ocular motility abnormalities, and visuospatial perception deficits. Although ocular motility problems can be corrected with surgery, vision, and perception deficits are more difficult to overcome. Interventions to restore or compensate for visual field deficits are controversial despite theoretical underpinnings, animal model evidence, and case reports of their efficacies.
Learning Science Through Visualization
NASA Technical Reports Server (NTRS)
Chaudhury, S. Raj
2005-01-01
In the context of an introductory physical science course for non-science majors, I have been trying to understand how scientific visualizations of natural phenomena can constructively impact student learning. I have also necessarily been concerned with the instructional and assessment approaches that need to be considered when focusing on learning science through visually rich information sources. The overall project can be broken down into three distinct segments : (i) comparing students' abilities to demonstrate proportional reasoning competency on visual and verbal tasks (ii) decoding and deconstructing visualizations of an object falling under gravity (iii) the role of directed instruction to elicit alternate, valid scientific visualizations of the structure of the solar system. Evidence of student learning was collected in multiple forms for this project - quantitative analysis of student performance on written, graded assessments (tests and quizzes); qualitative analysis of videos of student 'think aloud' sessions. The results indicate that there are significant barriers for non-science majors to succeed in mastering the content of science courses, but with informed approaches to instruction and assessment, these barriers can be overcome.
Li, W; Lai, T M; Bohon, C; Loo, S K; McCurdy, D; Strober, M; Bookheimer, S; Feusner, J
2015-07-01
Anorexia nervosa (AN) and body dysmorphic disorder (BDD) are characterized by distorted body image and are frequently co-morbid with each other, although their relationship remains little studied. While there is evidence of abnormalities in visual and visuospatial processing in both disorders, no study has directly compared the two. We used two complementary modalities--event-related potentials (ERPs) and functional magnetic resonance imaging (fMRI)--to test for abnormal activity associated with early visual signaling. We acquired fMRI and ERP data in separate sessions from 15 unmedicated individuals in each of three groups (weight-restored AN, BDD, and healthy controls) while they viewed images of faces and houses of different spatial frequencies. We used joint independent component analyses to compare activity in visual systems. AN and BDD groups demonstrated similar hypoactivity in early secondary visual processing regions and the dorsal visual stream when viewing low spatial frequency faces, linked to the N170 component, as well as in early secondary visual processing regions when viewing low spatial frequency houses, linked to the P100 component. Additionally, the BDD group exhibited hyperactivity in fusiform cortex when viewing high spatial frequency houses, linked to the N170 component. Greater activity in this component was associated with lower attractiveness ratings of faces. Results provide preliminary evidence of similar abnormal spatiotemporal activation in AN and BDD for configural/holistic information for appearance- and non-appearance-related stimuli. This suggests a common phenotype of abnormal early visual system functioning, which may contribute to perceptual distortions.
Sounds activate visual cortex and improve visual discrimination.
Feng, Wenfeng; Störmer, Viola S; Martinez, Antigona; McDonald, John J; Hillyard, Steven A
2014-07-16
A recent study in humans (McDonald et al., 2013) found that peripheral, task-irrelevant sounds activated contralateral visual cortex automatically as revealed by an auditory-evoked contralateral occipital positivity (ACOP) recorded from the scalp. The present study investigated the functional significance of this cross-modal activation of visual cortex, in particular whether the sound-evoked ACOP is predictive of improved perceptual processing of a subsequent visual target. A trial-by-trial analysis showed that the ACOP amplitude was markedly larger preceding correct than incorrect pattern discriminations of visual targets that were colocalized with the preceding sound. Dipole modeling of the scalp topography of the ACOP localized its neural generators to the ventrolateral extrastriate visual cortex. These results provide direct evidence that the cross-modal activation of contralateral visual cortex by a spatially nonpredictive but salient sound facilitates the discriminative processing of a subsequent visual target event at the location of the sound. Recordings of event-related potentials to the targets support the hypothesis that the ACOP is a neural consequence of the automatic orienting of visual attention to the location of the sound. Copyright © 2014 the authors 0270-6474/14/349817-08$15.00/0.
Hoshi, Eiji
2013-01-01
Action is often executed according to information provided by a visual signal. As this type of behavior integrates two distinct neural representations, perception and action, it has been thought that identification of the neural mechanisms underlying this process will yield deeper insights into the principles underpinning goal-directed behavior. Based on a framework derived from conditional visuomotor association, prior studies have identified neural mechanisms in the dorsal premotor cortex (PMd), dorsolateral prefrontal cortex (dlPFC), ventrolateral prefrontal cortex (vlPFC), and basal ganglia (BG). However, applications resting solely on this conceptualization encounter problems related to generalization and flexibility, essential processes in executive function, because the association mode involves a direct one-to-one mapping of each visual signal onto a particular action. To overcome this problem, we extend this conceptualization and postulate a more general framework, conditional visuo-goal association. According to this new framework, the visual signal identifies an abstract behavioral goal, and an action is subsequently selected and executed to meet this goal. Neuronal activity recorded from the four key areas of the brains of monkeys performing a task involving conditional visuo-goal association revealed three major mechanisms underlying this process. First, visual-object signals are represented primarily in the vlPFC and BG. Second, all four areas are involved in initially determining the goals based on the visual signals, with the PMd and dlPFC playing major roles in maintaining the salience of the goals. Third, the cortical areas play major roles in specifying action, whereas the role of the BG in this process is restrictive. These new lines of evidence reveal that the four areas involved in conditional visuomotor association contribute to goal-directed behavior mediated by conditional visuo-goal association in an area-dependent manner. PMID:24155692
Kukona, Anuenue; Cho, Pyeong Whan; Magnuson, James S.; Tabor, Whitney
2014-01-01
Psycholinguistic research spanning a number of decades has produced diverging results with regard to the nature of constraint integration in online sentence processing. For example, evidence that language users anticipatorily fixate likely upcoming referents in advance of evidence in the speech signal supports rapid context integration. By contrast, evidence that language users activate representations that conflict with contextual constraints, or only indirectly satisfy them, supports non-integration or late integration. Here, we report on a self-organizing neural network framework that addresses one aspect of constraint integration: the integration of incoming lexical information (i.e., an incoming word) with sentence context information (i.e., from preceding words in an unfolding utterance). In two simulations, we show that the framework predicts both classic results concerned with lexical ambiguity resolution (Swinney, 1979; Tanenhaus, Leiman, & Seidenberg, 1979), which suggest late context integration, and results demonstrating anticipatory eye movements (e.g., Altmann & Kamide, 1999), which support rapid context integration. We also report two experiments using the visual world paradigm that confirm a new prediction of the framework. Listeners heard sentences like “The boy will eat the white…,” while viewing visual displays with objects like a white cake (i.e., a predictable direct object of “eat”), white car (i.e., an object not predicted by “eat,” but consistent with “white”), and distractors. Consistent with our simulation predictions, we found that while listeners fixated white cake most, they also fixated white car more than unrelated distractors in this highly constraining sentence (and visual) context. PMID:24245535
Franceschini, Sandro; Trevisan, Piergiorgio; Ronconi, Luca; Bertoni, Sara; Colmar, Susan; Double, Kit; Facoetti, Andrea; Gori, Simone
2017-07-19
Dyslexia is characterized by difficulties in learning to read and there is some evidence that action video games (AVG), without any direct phonological or orthographic stimulation, improve reading efficiency in Italian children with dyslexia. However, the cognitive mechanism underlying this improvement and the extent to which the benefits of AVG training would generalize to deep English orthography, remain two critical questions. During reading acquisition, children have to integrate written letters with speech sounds, rapidly shifting their attention from visual to auditory modality. In our study, we tested reading skills and phonological working memory, visuo-spatial attention, auditory, visual and audio-visual stimuli localization, and cross-sensory attentional shifting in two matched groups of English-speaking children with dyslexia before and after they played AVG or non-action video games. The speed of words recognition and phonological decoding increased after playing AVG, but not non-action video games. Furthermore, focused visuo-spatial attention and visual-to-auditory attentional shifting also improved only after AVG training. This unconventional reading remediation program also increased phonological short-term memory and phoneme blending skills. Our report shows that an enhancement of visuo-spatial attention and phonological working memory, and an acceleration of visual-to-auditory attentional shifting can directly translate into better reading in English-speaking children with dyslexia.
The look of royalty: visual and odour signals of reproductive status in a paper wasp
Tannure-Nascimento, Ivelize C; Nascimento, Fabio S; Zucchi, Ronaldo
2008-01-01
Reproductive conflicts within animal societies occur when all females can potentially reproduce. In social insects, these conflicts are regulated largely by behaviour and chemical signalling. There is evidence that presence of signals, which provide direct information about the quality of the reproductive females would increase the fitness of all parties. In this study, we present an association between visual and chemical signals in the paper wasp Polistes satan. Our results showed that in nest-founding phase colonies, variation of visual signals is linked to relative fertility, while chemical signals are related to dominance status. In addition, experiments revealed that higher hierarchical positions were occupied by subordinates with distinct proportions of cuticular hydrocarbons and distinct visual marks. Therefore, these wasps present cues that convey reliable information of their reproductive status. PMID:18682372
The look of royalty: visual and odour signals of reproductive status in a paper wasp.
Tannure-Nascimento, Ivelize C; Nascimento, Fabio S; Zucchi, Ronaldo
2008-11-22
Reproductive conflicts within animal societies occur when all females can potentially reproduce. In social insects, these conflicts are regulated largely by behaviour and chemical signalling. There is evidence that presence of signals, which provide direct information about the quality of the reproductive females would increase the fitness of all parties. In this study, we present an association between visual and chemical signals in the paper wasp Polistes satan. Our results showed that in nest-founding phase colonies, variation of visual signals is linked to relative fertility, while chemical signals are related to dominance status. In addition, experiments revealed that higher hierarchical positions were occupied by subordinates with distinct proportions of cuticular hydrocarbons and distinct visual marks. Therefore, these wasps present cues that convey reliable information of their reproductive status.
Binocular vision in amblyopia: structure, suppression and plasticity.
Hess, Robert F; Thompson, Benjamin; Baker, Daniel H
2014-03-01
The amblyopic visual system was once considered to be structurally monocular. However, it now evident that the capacity for binocular vision is present in many observers with amblyopia. This has led to new techniques for quantifying suppression that have provided insights into the relationship between suppression and the monocular and binocular visual deficits experienced by amblyopes. Furthermore, new treatments are emerging that directly target suppressive interactions within the visual cortex and, on the basis of initial data, appear to improve both binocular and monocular visual function, even in adults with amblyopia. The aim of this review is to provide an overview of recent studies that have investigated the structure, measurement and treatment of binocular vision in observers with strabismic, anisometropic and mixed amblyopia. © 2014 The Authors Ophthalmic & Physiological Optics © 2014 The College of Optometrists.
Spiegel, Daniel P; Byblow, Winston D; Hess, Robert F; Thompson, Benjamin
2013-10-01
Amblyopia is a neurodevelopmental disorder of vision that is associated with abnormal patterns of neural inhibition within the visual cortex. This disorder is often considered to be untreatable in adulthood because of insufficient visual cortex plasticity. There is increasing evidence that interventions that target inhibitory interactions within the visual cortex, including certain types of noninvasive brain stimulation, can improve visual function in adults with amblyopia. We tested the hypothesis that anodal transcranial direct current stimulation (a-tDCS) would improve visual function in adults with amblyopia by enhancing the neural response to inputs from the amblyopic eye. Thirteen adults with amblyopia participated and contrast sensitivity in the amblyopic and fellow fixing eye was assessed before, during and after a-tDCS or cathodal tDCS (c-tDCS). Five participants also completed a functional magnetic resonance imaging (fMRI) study designed to investigate the effect of a-tDCS on the blood oxygen level-dependent response within the visual cortex to inputs from the amblyopic versus the fellow fixing eye. A subgroup of 8/13 participants showed a transient improvement in amblyopic eye contrast sensitivity for at least 30 minutes after a-tDCS. fMRI measurements indicated that the characteristic cortical response asymmetry in amblyopes, which favors the fellow eye, was reduced by a-tDCS. These preliminary results suggest that a-tDCS deserves further investigation as a potential tool to enhance amblyopia treatment outcomes in adults.
Dudschig, Carolin; Souman, Jan; Lachmair, Martin; de la Vega, Irmgard; Kaup, Barbara
2013-01-01
Traditionally, language processing has been attributed to a separate system in the brain, which supposedly works in an abstract propositional manner. However, there is increasing evidence suggesting that language processing is strongly interrelated with sensorimotor processing. Evidence for such an interrelation is typically drawn from interactions between language and perception or action. In the current study, the effect of words that refer to entities in the world with a typical location (e.g., sun, worm) on the planning of saccadic eye movements was investigated. Participants had to perform a lexical decision task on visually presented words and non-words. They responded by moving their eyes to a target in an upper (lower) screen position for a word (non-word) or vice versa. Eye movements were faster to locations compatible with the word's referent in the real world. These results provide evidence for the importance of linguistic stimuli in directing eye movements, even if the words do not directly transfer directional information.
Cue-recruitment for extrinsic signals after training with low information stimuli.
Jain, Anshul; Fuller, Stuart; Backus, Benjamin T
2014-01-01
Cue-recruitment occurs when a previously ineffective signal comes to affect the perceptual appearance of a target object, in a manner similar to the trusted cues with which the signal was put into correlation during training. Jain, Fuller and Backus reported that extrinsic signals, those not carried by the target object itself, were not recruited even after extensive training. However, recent studies have shown that training using weakened trusted cues can facilitate recruitment of intrinsic signals. The current study was designed to examine whether extrinsic signals can be recruited by putting them in correlation with weakened trusted cues. Specifically, we tested whether an extrinsic visual signal, the rotary motion direction of an annulus of random dots, and an extrinsic auditory signal, direction of an auditory pitch glide, can be recruited as cues for the rotation direction of a Necker cube. We found learning, albeit weak, for visual but not for auditory signals. These results extend the generality of the cue-recruitment phenomenon to an extrinsic signal and provide further evidence that the visual system learns to use new signals most quickly when other, long-trusted cues are unavailable or unreliable.
Vortex locking in direct numerical simulations of quantum turbulence.
Morris, Karla; Koplik, Joel; Rouson, Damian W I
2008-07-04
Direct numerical simulations are used to examine the locking of quantized superfluid vortices and normal fluid vorticity in evolving turbulent flows. The superfluid is driven by the normal fluid, which undergoes either a decaying Taylor-Green flow or a linearly forced homogeneous isotropic turbulent flow, although the back reaction of the superfluid on the normal fluid flow is omitted. Using correlation functions and wavelet transforms, we present numerical and visual evidence for vortex locking on length scales above the intervortex spacing.
Freud, Erez; Macdonald, Scott N; Chen, Juan; Quinlan, Derek J; Goodale, Melvyn A; Culham, Jody C
2018-01-01
In the current era of touchscreen technology, humans commonly execute visually guided actions directed to two-dimensional (2D) images of objects. Although real, three-dimensional (3D), objects and images of the same objects share high degree of visual similarity, they differ fundamentally in the actions that can be performed on them. Indeed, previous behavioral studies have suggested that simulated grasping of images relies on different representations than actual grasping of real 3D objects. Yet the neural underpinnings of this phenomena have not been investigated. Here we used functional magnetic resonance imaging (fMRI) to investigate how brain activation patterns differed for grasping and reaching actions directed toward real 3D objects compared to images. Multivoxel Pattern Analysis (MVPA) revealed that the left anterior intraparietal sulcus (aIPS), a key region for visually guided grasping, discriminates between both the format in which objects were presented (real/image) and the motor task performed on them (grasping/reaching). Interestingly, during action planning, the representations of real 3D objects versus images differed more for grasping movements than reaching movements, likely because grasping real 3D objects involves fine-grained planning and anticipation of the consequences of a real interaction. Importantly, this dissociation was evident in the planning phase, before movement initiation, and was not found in any other regions, including motor and somatosensory cortices. This suggests that the dissociable representations in the left aIPS were not based on haptic, motor or proprioceptive feedback. Together, these findings provide novel evidence that actions, particularly grasping, are affected by the realness of the target objects during planning, perhaps because real targets require a more elaborate forward model based on visual cues to predict the consequences of real manipulation. Copyright © 2017 Elsevier Ltd. All rights reserved.
Plow, Ela B.; Obretenova, Souzana N.; Jackson, Mary Lou; Merabet, Lotfi B.
2012-01-01
Objectives We have previously reported that transcranial direct current stimulation (tDCS) delivered to the occipital cortex enhances visual functional recovery when combined with 3 months of computer-based rehabilitative training in patients with hemianopia. The principal objective of this study was to evaluate the temporal sequence of effects of tDCS on visual recovery as they appear over the course of training and across different indicators of visual function. Methods Primary objective outcome measures were i) shifts in visual field border and ii) stimulus detection accuracy within the affected hemifield. These were compared between patients randomized to either vision restoration therapy (VRT) combined with active tDCS or VRT paired with sham tDCS. Training comprised of 2 half hour sessions, 3 times a week for 3 months. Primary outcome measures were collected at baseline (pretest), monthly interim intervals, and at posttest (3 months). As secondary outcome measures, contrast sensitivity and reading performance were collected at pretest and posttest time-points only. Results Active tDCS combined with VRT accelerated the recovery of stimulus detection as between-group differences appeared within the first month of training. In contrast, a shift in the visual field border was only evident at posttest (after 3 months of training). TDCS did not affect contrast sensitivity or reading performance. Conclusions These results suggest that tDCS may differentially affect the magnitude and sequence of visual recovery in a manner that is task- specific to the type of visual rehabilitative training strategy employed. PMID:22376226
Separate visual representations for perception and for visually guided behavior
NASA Technical Reports Server (NTRS)
Bridgeman, Bruce
1989-01-01
Converging evidence from several sources indicates that two distinct representations of visual space mediate perception and visually guided behavior, respectively. The two maps of visual space follow different rules; spatial values in either one can be biased without affecting the other. Ordinarily the two maps give equivalent responses because both are veridically in register with the world; special techniques are required to pull them apart. One such technique is saccadic suppression: small target displacements during saccadic eye movements are not preceived, though the displacements can change eye movements or pointing to the target. A second way to separate cognitive and motor-oriented maps is with induced motion: a slowly moving frame will make a fixed target appear to drift in the opposite direction, while motor behavior toward the target is unchanged. The same result occurs with stroboscopic induced motion, where the frame jump abruptly and the target seems to jump in the opposite direction. A third method of separating cognitive and motor maps, requiring no motion of target, background or eye, is the Roelofs effect: a target surrounded by an off-center rectangular frame will appear to be off-center in the direction opposite the frame. Again the effect influences perception, but in half of the subjects it does not influence pointing to the target. This experience also reveals more characteristics of the maps and their interactions with one another, the motor map apparently has little or no memory, and must be fed from the biased cognitive map if an enforced delay occurs between stimulus presentation and motor response. In designing spatial displays, the results mean that what you see isn't necessarily what you get. Displays must be designed with either perception or visually guided behavior in mind.
Moving attention - Evidence for time-invariant shifts of visual selective attention
NASA Technical Reports Server (NTRS)
Remington, R.; Pierce, L.
1984-01-01
Two experiments measured the time to shift spatial selective attention across the visual field to targets 2 or 10 deg from central fixation. A central arrow cued the most likely target location. The direction of attention was inferred from reaction times to expected, unexpected, and neutral locations. The development of a spatial attentional set with time was examined by presenting target probes at varying times after the cue. There were no effects of distance on the time course of the attentional set. Reaction times for far locations were slower than for near, but the effects of attention were evident by 150 msec in both cases. Spatial attention does not shift with a characteristic, fixed velocity. Rather, velocity is proportional to distance, resulting in a movement time that is invariant over the distances tested.
Ouellet, Marc; Santiago, Julio; Israeli, Ziv; Gabay, Shai
2010-01-01
Spanish and English speakers tend to conceptualize time as running from left to right along a mental line. Previous research suggests that this representational strategy arises from the participants' exposure to a left-to-right writing system. However, direct evidence supporting this assertion suffers from several limitations and relies only on the visual modality. This study subjected to a direct test the reading hypothesis using an auditory task. Participants from two groups (Spanish and Hebrew) differing in the directionality of their orthographic system had to discriminate temporal reference (past or future) of verbs and adverbs (referring to either past or future) auditorily presented to either the left or right ear by pressing a left or a right key. Spanish participants were faster responding to past words with the left hand and to future words with the right hand, whereas Hebrew participants showed the opposite pattern. Our results demonstrate that the left-right mapping of time is not restricted to the visual modality and that the direction of reading accounts for the preferred directionality of the mental time line. These results are discussed in the context of a possible mechanism underlying the effects of reading direction on highly abstract conceptual representations.
Why Are Faces Denser in the Visual Experiences of Younger than Older Infants?
ERIC Educational Resources Information Center
Jayaraman, Swapnaa; Fausey, Caitlin M.; Smith, Linda B.
2017-01-01
Recent evidence from studies using head cameras suggests that the frequency of faces directly in front of infants "declines" over the first year and a half of life, a result that has implications for the development of and evolutionary constraints on face processing. Two experiments tested 2 opposing hypotheses about this observed…
Three New Z Cam Stars (Abstract)
NASA Astrophysics Data System (ADS)
Simonsen, M.
2016-12-01
(Abstract only) I will present the evidence and discovery stories of three cataclysmic variables who appear to be members of the Z Cam class of dwarf novae. One discovered by a lone visual observer and his unwavering patience and persistence, one through the directed effort of the ongoing Z CamPaign and one via survey data from the Gaia satellite.
ERIC Educational Resources Information Center
Klin, Ami; Jones, Warren
2008-01-01
Mounting clinical evidence suggests that abnormalities of social engagement in children with autism are present even during infancy. However, direct experimental documentation of these abnormalities is still limited. In this case report of a 15-month-old infant with autism, we measured visual fixation patterns to both naturalistic and ambiguous…
ERIC Educational Resources Information Center
Freeth, M.; Chapman, P.; Ropar, D.; Mitchell, P.
2010-01-01
Visual fixation patterns whilst viewing complex photographic scenes containing one person were studied in 24 high-functioning adolescents with Autism Spectrum Disorders (ASD) and 24 matched typically developing adolescents. Over two different scene presentation durations both groups spent a large, strikingly similar proportion of their viewing…
ERIC Educational Resources Information Center
Weidemann, Christoph T.; Huber, David E.; Shiffrin, Richard M.
2008-01-01
The authors conducted 4 repetition priming experiments that manipulated prime duration and prime diagnosticity in a visual forced-choice perceptual identification task. The strength and direction of prime diagnosticity produced marked effects on identification accuracy, but those effects were resistant to subsequent changes of diagnosticity.…
DeSantis, Diana
2014-06-01
Amblyopia refers to unilateral or bilateral reduction in best corrected visual acuity, not directly attributed to structural abnormality of the eye or posterior visual pathways. Early detection of amblyopia is crucial to obtaining the best response to treatment. Amblyopia responds best to treatment in the first few years of life. In the past several years a series of studies undertaken by the Pediatric Eye Disease Investigator Group (PEDIG) have been designed to evaluate traditional methods for treating amblyopia and provide evidence on which to base treatment decisions. This article summarizes and discusses the findings of the PEDIG studies to date. Copyright © 2014 Elsevier Inc. All rights reserved.
A Multi-Stage Model for Fundamental Functional Properties in Primary Visual Cortex
Hesam Shariati, Nastaran; Freeman, Alan W.
2012-01-01
Many neurons in mammalian primary visual cortex have properties such as sharp tuning for contour orientation, strong selectivity for motion direction, and insensitivity to stimulus polarity, that are not shared with their sub-cortical counterparts. Successful models have been developed for a number of these properties but in one case, direction selectivity, there is no consensus about underlying mechanisms. We here define a model that accounts for many of the empirical observations concerning direction selectivity. The model describes a single column of cat primary visual cortex and comprises a series of processing stages. Each neuron in the first cortical stage receives input from a small number of on-centre and off-centre relay cells in the lateral geniculate nucleus. Consistent with recent physiological evidence, the off-centre inputs to cortex precede the on-centre inputs by a small (∼4 ms) interval, and it is this difference that confers direction selectivity on model neurons. We show that the resulting model successfully matches the following empirical data: the proportion of cells that are direction selective; tilted spatiotemporal receptive fields; phase advance in the response to a stationary contrast-reversing grating stepped across the receptive field. The model also accounts for several other fundamental properties. Receptive fields have elongated subregions, orientation selectivity is strong, and the distribution of orientation tuning bandwidth across neurons is similar to that seen in the laboratory. Finally, neurons in the first stage have properties corresponding to simple cells, and more complex-like cells emerge in later stages. The results therefore show that a simple feed-forward model can account for a number of the fundamental properties of primary visual cortex. PMID:22496811
Visual and Non-Visual Contributions to the Perception of Object Motion during Self-Motion
Fajen, Brett R.; Matthis, Jonathan S.
2013-01-01
Many locomotor tasks involve interactions with moving objects. When observer (i.e., self-)motion is accompanied by object motion, the optic flow field includes a component due to self-motion and a component due to object motion. For moving observers to perceive the movement of other objects relative to the stationary environment, the visual system could recover the object-motion component – that is, it could factor out the influence of self-motion. In principle, this could be achieved using visual self-motion information, non-visual self-motion information, or a combination of both. In this study, we report evidence that visual information about the speed (Experiment 1) and direction (Experiment 2) of self-motion plays a role in recovering the object-motion component even when non-visual self-motion information is also available. However, the magnitude of the effect was less than one would expect if subjects relied entirely on visual self-motion information. Taken together with previous studies, we conclude that when self-motion is real and actively generated, both visual and non-visual self-motion information contribute to the perception of object motion. We also consider the possible role of this process in visually guided interception and avoidance of moving objects. PMID:23408983
Evidence for a basal temporal visual language center: cortical stimulation producing pure alexia.
Mani, J; Diehl, B; Piao, Z; Schuele, S S; Lapresto, E; Liu, P; Nair, D R; Dinner, D S; Lüders, H O
2008-11-11
Dejerine and Benson and Geschwind postulated disconnection of the dominant angular gyrus from both visual association cortices as the basis for pure alexia, emphasizing disruption of white matter tracts in the dominant temporooccipital region. Recently functional imaging studies provide evidence for direct participation of basal temporal and occipital cortices in the cognitive process of reading. The exact location and function of these areas remain a matter of debate. To confirm the participation of the basal temporal region in reading. Extraoperative electrical stimulation of the dominant hemisphere was performed in three subjects using subdural electrodes, as part of presurgical evaluation for refractory epilepsy. Pure alexia was reproduced during cortical stimulation of the dominant posterior fusiform and inferior temporal gyri in all three patients. Stimulation resulted in selective reading difficulty with intact auditory comprehension and writing. Reading difficulty involved sentences and words with intact letter by letter reading. Picture naming difficulties were also noted at some electrodes. This region is located posterior to and contiguous with the basal temporal language area (BTLA) where stimulation resulted in global language dysfunction in visual and auditory realms. The location corresponded with the visual word form area described on functional MRI. These observations support the existence of a visual language area in the dominant fusiform and occipitotemporal gyri, contiguous with basal temporal language area. A portion of visual language area was exclusively involved in lexical processing while the other part of this region processed both lexical and nonlexical symbols.
Bulf, Hermann; de Hevia, Maria Dolores; Macchi Cassia, Viola
2016-05-01
Numbers are represented as ordered magnitudes along a spatially oriented number line. While culture and formal education modulate the direction of this number-space mapping, it is a matter of debate whether its emergence is entirely driven by cultural experience. By registering 8-9-month-old infants' eye movements, this study shows that numerical cues are critical in orienting infants' visual attention towards a peripheral region of space that is congruent with the number's relative position on a left-to-right oriented representational continuum. This finding provides the first direct evidence that, in humans, the association between numbers and oriented spatial codes occurs before the acquisition of symbols or exposure to formal education, suggesting that the number line is not merely a product of human invention. © 2015 John Wiley & Sons Ltd.
Supèr, Hans; Lamme, Victor A F
2007-06-01
When and where are decisions made? In the visual system a saccade, which is a fast shift of gaze toward a target in the visual scene, is the behavioral outcome of a decision. Current neurophysiological data and reaction time models show that saccadic reaction times are determined by a build-up of activity in motor-related structures, such as the frontal eye fields. These structures depend on the sensory evidence of the stimulus. Here we use a delayed figure-ground detection task to show that late modulated activity in the visual cortex (V1) predicts saccadic reaction time. This predictive activity is part of the process of figure-ground segregation and is specific for the saccade target location. These observations indicate that sensory signals are directly involved in the decision of when and where to look.
Flow, affect and visual creativity.
Cseh, Genevieve M; Phillips, Louise H; Pearson, David G
2015-01-01
Flow (being in the zone) is purported to have positive consequences in terms of affect and performance; however, there is no empirical evidence about these links in visual creativity. Positive affect often--but inconsistently--facilitates creativity, and both may be linked to experiencing flow. This study aimed to determine relationships between these variables within visual creativity. Participants performed the creative mental synthesis task to simulate the creative process. Affect change (pre- vs. post-task) and flow were measured via questionnaires. The creativity of synthesis drawings was rated objectively and subjectively by judges. Findings empirically demonstrate that flow is related to affect improvement during visual creativity. Affect change was linked to productivity and self-rated creativity, but no other objective or subjective performance measures. Flow was unrelated to all external performance measures but was highly correlated with self-rated creativity; flow may therefore motivate perseverance towards eventual excellence rather than provide direct cognitive enhancement.
Neuropsychological implications of selective attentional functioning in psychopathic offenders.
Mayer, Andrew R; Kosson, David S; Bedrick, Edward J
2006-09-01
Several core characteristics of the psychopathic personality disorder (i.e., impulsivity, failure to attend to interpersonal cues) suggest that psychopaths suffer from disordered attention. However, there is mixed evidence from the cognitive literature as to whether they exhibit superior or deficient selective attention, which has led to the formation of several distinct theories of attentional functioning in psychopathy. The present experiment investigated participants' abilities to purposely allocate attentional resources on the basis of auditory or visual linguistic information and directly tested both theories of deficient or superior selective attention in psychopathy. Specifically, 91 male inmates at a county jail were presented with either auditory or visual linguistic cues (with and without distractors) that correctly indicated the position of an upcoming visual target in 75% of the trials. The results indicated that psychopaths did not exhibit evidence of superior selective attention in any of the conditions but were generally less efficient in shifting attention on the basis of linguistic cues, especially in regard to auditory information. Implications for understanding psychopaths' cognitive functioning and possible neuropsychological deficits are addressed. ((c) 2006 APA, all rights reserved).
Computerized EEG analysis for studying the effect of drugs on the central nervous system.
Rosadini, G; Cavazza, B; Rodriguez, G; Sannita, W G; Siccardi, A
1977-11-01
Samples of our experience in quantitative pharmaco-EEG are reviewed to discuss and define its applicability and limits. Simple processing systems, such as the computation of Hjorth's descriptors, are useful for on-line monitoring of drug-induced EEG modifications which are evident also at the visual visual analysis. Power spectral analysis is suitable to identify and quantify EEG effects not evident at the visual inspection. It demonstrated how the EEG effects of compounds in a long-acting formulation vary according to the sampling time and the explored cerebral area. EEG modifications not detected by power spectral analysis can be defined by comparing statistically (F test) the spectral values of the EEG from a single lead at the different samples (longitudinal comparison), or the spectral values from different leads at any sample (intrahemispheric comparison). The presently available procedures of quantitative pharmaco-EEG are effective when applied to the study of mutltilead EEG recordings in a statistically significant sample of population. They do not seem reliable in the monitoring of directing of neuropyschiatric therapies in single patients, due to individual variability of drug effects.
Neural Representation of Motion-In-Depth in Area MT
Sanada, Takahisa M.
2014-01-01
Neural processing of 2D visual motion has been studied extensively, but relatively little is known about how visual cortical neurons represent visual motion trajectories that include a component toward or away from the observer (motion in depth). Psychophysical studies have demonstrated that humans perceive motion in depth based on both changes in binocular disparity over time (CD cue) and interocular velocity differences (IOVD cue). However, evidence for neurons that represent motion in depth has been limited, especially in primates, and it is unknown whether such neurons make use of CD or IOVD cues. We show that approximately one-half of neurons in macaque area MT are selective for the direction of motion in depth, and that this selectivity is driven primarily by IOVD cues, with a small contribution from the CD cue. Our results establish that area MT, a central hub of the primate visual motion processing system, contains a 3D representation of visual motion. PMID:25411481
Intraorbital foreign body projectile as a consideration for unilateral pupillary defect
2012-01-01
Intraorbital foreign bodies are frequently the result of high-velocity injuries with varying clinical presentations. The resultant diagnosis, management, and outcome depend on the type of foreign body present, anatomical location, tissue disruption, and symptomatology. A patient who presented to the Emergency Department with a large intraorbital foreign body projectile that was not evident clinically, but found incidentally on computed tomography and subsequent plain films is reported. The emergency room physician needs to be aware of the differential diagnosis of a unilateral irregular pupil with or without visual acuity changes. The differential diagnosis for any trauma patient with an irregular pupil with significant visual loss must include intraorbital foreign body and associated injury to the optic nerve directly or via orbital compartment syndrome secondary to hemorrhage and/or edema. Patients with significantly decreased visual acuity may benefit from emergent surgical intervention. In patients with intact visual acuity, the patient must be monitored closely for any visual changes as this may require emergent surgical intervention. PMID:22390406
Giant honeybees ( Apis dorsata) mob wasps away from the nest by directed visual patterns
NASA Astrophysics Data System (ADS)
Kastberger, Gerald; Weihmann, Frank; Zierler, Martina; Hötzl, Thomas
2014-11-01
The open nesting behaviour of giant honeybees ( Apis dorsata) accounts for the evolution of a series of defence strategies to protect the colonies from predation. In particular, the concerted action of shimmering behaviour is known to effectively confuse and repel predators. In shimmering, bees on the nest surface flip their abdomens in a highly coordinated manner to generate Mexican wave-like patterns. The paper documents a further-going capacity of this kind of collective defence: the visual patterns of shimmering waves align regarding their directional characteristics with the projected flight manoeuvres of the wasps when preying in front of the bees' nest. The honeybees take here advantage of a threefold asymmetry intrinsic to the prey-predator interaction: (a) the visual patterns of shimmering turn faster than the wasps on their flight path, (b) they "follow" the wasps more persistently (up to 100 ms) than the wasps "follow" the shimmering patterns (up to 40 ms) and (c) the shimmering patterns align with the wasps' flight in all directions at the same strength, whereas the wasps have some preference for horizontal correspondence. The findings give evidence that shimmering honeybees utilize directional alignment to enforce their repelling power against preying wasps. This phenomenon can be identified as predator driving which is generally associated with mobbing behaviour (particularly known in selfish herds of vertebrate species), which is, until now, not reported in insects.
Beyond Control Panels: Direct Manipulation for Visual Analytics
DOE Office of Scientific and Technical Information (OSTI.GOV)
Endert, Alexander; Bradel, Lauren; North, Chris
2013-07-19
Information Visualization strives to provide visual representations through which users can think about and gain insight into information. By leveraging the visual and cognitive systems of humans, complex relationships and phenomena occurring within datasets can be uncovered by exploring information visually. Interaction metaphors for such visualizations are designed to enable users direct control over the filters, queries, and other parameters controlling how the data is visually represented. Through the evolution of information visualization, more complex mathematical and data analytic models are being used to visualize relationships and patterns in data – creating the field of Visual Analytics. However, the expectationsmore » for how users interact with these visualizations has remained largely unchanged – focused primarily on the direct manipulation of parameters of the underlying mathematical models. In this article we present an opportunity to evolve the methodology for user interaction from the direct manipulation of parameters through visual control panels, to interactions designed specifically for visual analytic systems. Instead of focusing on traditional direct manipulation of mathematical parameters, the evolution of the field can be realized through direct manipulation within the visual representation – where users can not only gain insight, but also interact. This article describes future directions and research challenges that fundamentally change the meaning of direct manipulation with regards to visual analytics, advancing the Science of Interaction.« less
ERIC Educational Resources Information Center
Lindemann, Oliver; Bekkering, Harold
2009-01-01
In 3 experiments, the authors investigated the bidirectional coupling of perception and action in the context of object manipulations and motion perception. Participants prepared to grasp an X-shaped object along one of its 2 diagonals and to rotate it in a clockwise or a counterclockwise direction. Action execution had to be delayed until the…
Interactions between dorsal and ventral streams for controlling skilled grasp
van Polanen, Vonne; Davare, Marco
2015-01-01
The two visual systems hypothesis suggests processing of visual information into two distinct routes in the brain: a dorsal stream for the control of actions and a ventral stream for the identification of objects. Recently, increasing evidence has shown that the dorsal and ventral streams are not strictly independent, but do interact with each other. In this paper, we argue that the interactions between dorsal and ventral streams are important for controlling complex object-oriented hand movements, especially skilled grasp. Anatomical studies have reported the existence of direct connections between dorsal and ventral stream areas. These physiological interconnections appear to be gradually more active as the precision demands of the grasp become higher. It is hypothesised that the dorsal stream needs to retrieve detailed information about object identity, stored in ventral stream areas, when the object properties require complex fine-tuning of the grasp. In turn, the ventral stream might receive up to date grasp-related information from dorsal stream areas to refine the object internal representation. Future research will provide direct evidence for which specific areas of the two streams interact, the timing of their interactions and in which behavioural context they occur. PMID:26169317
Top-down beta oscillatory signaling conveys behavioral context in early visual cortex.
Richter, Craig G; Coppola, Richard; Bressler, Steven L
2018-05-03
Top-down modulation of sensory processing is a critical neural mechanism subserving numerous important cognitive roles, one of which may be to inform lower-order sensory systems of the current 'task at hand' by conveying behavioral context to these systems. Accumulating evidence indicates that top-down cortical influences are carried by directed interareal synchronization of oscillatory neuronal populations, with recent results pointing to beta-frequency oscillations as particularly important for top-down processing. However, it remains to be determined if top-down beta-frequency oscillations indeed convey behavioral context. We measured spectral Granger Causality (sGC) using local field potentials recorded from microelectrodes chronically implanted in visual areas V1/V2, V4, and TEO of two rhesus macaque monkeys, and applied multivariate pattern analysis to the spatial patterns of top-down sGC. We decoded behavioral context by discriminating patterns of top-down (V4/TEO-to-V1/V2) beta-peak sGC for two different task rules governing correct responses to identical visual stimuli. The results indicate that top-down directed influences are carried to visual cortex by beta oscillations, and differentiate task demands even before visual stimulus processing. They suggest that top-down beta-frequency oscillatory processes coordinate processing of sensory information by conveying global knowledge states to early levels of the sensory cortical hierarchy independently of bottom-up stimulus-driven processing.
Visual feedback system to reduce errors while operating roof bolting machines
Steiner, Lisa J.; Burgess-Limerick, Robin; Eiter, Brianna; Porter, William; Matty, Tim
2015-01-01
Problem Operators of roof bolting machines in underground coal mines do so in confined spaces and in very close proximity to the moving equipment. Errors in the operation of these machines can have serious consequences, and the design of the equipment interface has a critical role in reducing the probability of such errors. Methods An experiment was conducted to explore coding and directional compatibility on actual roof bolting equipment and to determine the feasibility of a visual feedback system to alert operators of critical movements and to also alert other workers in close proximity to the equipment to the pending movement of the machine. The quantitative results of the study confirmed the potential for both selection errors and direction errors to be made, particularly during training. Results Subjective data confirmed a potential benefit of providing visual feedback of the intended operations and movements of the equipment. Impact This research may influence the design of these and other similar control systems to provide evidence for the use of warning systems to improve operator situational awareness. PMID:23398703
Milburn, Evelyn; Warren, Tessa; Dickey, Michael Walsh
There has been considerable debate regarding the question of whether linguistic knowledge and world knowledge are separable and used differently during processing or not (Hagoort, Hald, Bastiaansen, & Petersson, 2004; Matsuki et al., 2011; Paczynski & Kuperberg, 2012; Warren & McConnell, 2007; Warren, McConnell, & Rayner, 2008). Previous investigations into this question have provided mixed evidence as to whether violations of selectional restrictions are detected earlier than violations of world knowledge. We report a visual-world eye-tracking study comparing the timing of facilitation contributed by selectional restrictions versus world knowledge. College-aged adults (n=36) viewed photographs of natural scenes while listening to sentences. Participants anticipated upcoming direct objects similarly regardless of whether facilitation was provided by only world knowledge or a combination of selectional restrictions and world knowledge. These results suggest that selectional restrictions are not available earlier in comprehension than world knowledge.
Theta Phase Synchronization Is the Glue that Binds Human Associative Memory.
Clouter, Andrew; Shapiro, Kimron L; Hanslmayr, Simon
2017-10-23
Episodic memories are information-rich, often multisensory events that rely on binding different elements [1]. The elements that will constitute a memory episode are processed in specialized but distinct brain modules. The binding of these elements is most likely mediated by fast-acting long-term potentiation (LTP), which relies on the precise timing of neural activity [2]. Theta oscillations in the hippocampus orchestrate such timing as demonstrated by animal studies in vitro [3, 4] and in vivo [5, 6], suggesting a causal role of theta activity for the formation of complex memory episodes, but direct evidence from humans is missing. Here, we show that human episodic memory formation depends on phase synchrony between different sensory cortices at the theta frequency. By modulating the luminance of visual stimuli and the amplitude of auditory stimuli, we directly manipulated the degree of phase synchrony between visual and auditory cortices. Memory for sound-movie associations was significantly better when the stimuli were presented in phase compared to out of phase. This effect was specific to theta (4 Hz) and did not occur in slower (1.7 Hz) or faster (10.5 Hz) frequencies. These findings provide the first direct evidence that episodic memory formation in humans relies on a theta-specific synchronization mechanism. Copyright © 2017 Elsevier Ltd. All rights reserved.
Preparatory attention in visual cortex.
Battistoni, Elisa; Stein, Timo; Peelen, Marius V
2017-05-01
Top-down attention is the mechanism that allows us to selectively process goal-relevant aspects of a scene while ignoring irrelevant aspects. A large body of research has characterized the effects of attention on neural activity evoked by a visual stimulus. However, attention also includes a preparatory phase before stimulus onset in which the attended dimension is internally represented. Here, we review neurophysiological, functional magnetic resonance imaging, magnetoencephalography, electroencephalography, and transcranial magnetic stimulation (TMS) studies investigating the neural basis of preparatory attention, both when attention is directed to a location in space and when it is directed to nonspatial stimulus attributes (content-based attention) ranging from low-level features to object categories. Results show that both spatial and content-based attention lead to increased baseline activity in neural populations that selectively code for the attended attribute. TMS studies provide evidence that this preparatory activity is causally related to subsequent attentional selection and behavioral performance. Attention thus acts by preactivating selective neurons in the visual cortex before stimulus onset. This appears to be a general mechanism that can operate on multiple levels of representation. We discuss the functional relevance of this mechanism, its limitations, and its relation to working memory, imagery, and expectation. We conclude by outlining open questions and future directions. © 2017 New York Academy of Sciences.
Testing of visual field with virtual reality goggles in manual and visual grasp modes.
Wroblewski, Dariusz; Francis, Brian A; Sadun, Alfredo; Vakili, Ghazal; Chopra, Vikas
2014-01-01
Automated perimetry is used for the assessment of visual function in a variety of ophthalmic and neurologic diseases. We report development and clinical testing of a compact, head-mounted, and eye-tracking perimeter (VirtualEye) that provides a more comfortable test environment than the standard instrumentation. VirtualEye performs the equivalent of a full threshold 24-2 visual field in two modes: (1) manual, with patient response registered with a mouse click, and (2) visual grasp, where the eye tracker senses change in gaze direction as evidence of target acquisition. 59 patients successfully completed the test in manual mode and 40 in visual grasp mode, with 59 undergoing the standard Humphrey field analyzer (HFA) testing. Large visual field defects were reliably detected by VirtualEye. Point-by-point comparison between the results obtained with the different modalities indicates: (1) minimal systematic differences between measurements taken in visual grasp and manual modes, (2) the average standard deviation of the difference distributions of about 5 dB, and (3) a systematic shift (of 4-6 dB) to lower sensitivities for VirtualEye device, observed mostly in high dB range. The usability survey suggested patients' acceptance of the head-mounted device. The study appears to validate the concepts of a head-mounted perimeter and the visual grasp mode.
Goodale, M A; Murison, R C
1975-05-02
The effects of bilateral removal of the superior colliculus or visual cortex on visually guided locomotor movements in rats performing a brightness discrimination task were investigated directly with the use of cine film. Rats with collicular lesions showed patterns of locomotion comparable to or more efficient than those of normal animals when approaching one of 5 small doors located at one end of a large open area. In contrast, animals with large but incomplete lesions of visual cortex were distinctly impaired in their visual control of approach responses to the same stimuli. On the other hand, rats with collicular damage showed no orienting reflex or evidence of distraction in the same task when novel visual or auditory stimuli were presented. However, both normal and visual-decorticate rats showed various components of the orienting reflex and disturbance in task performance when the same novel stimuli were presented. These results suggest that although the superior colliculus does not appear to be essential to the visual control of locomotor orientation, this midbrain structure might participate in the mediation of shifts in visual fixation and attention. Visual cortex, while contributing to visuospatial guidance of locomotor movements, might not play a significant role in the control and integration of the orienting reflex.
Is there an age-related positivity effect in visual attention? A comparison of two methodologies.
Isaacowitz, Derek M; Wadlinger, Heather A; Goren, Deborah; Wilson, Hugh R
2006-08-01
Research suggests a positivity effect in older adults' memory for emotional material, but the evidence from the attentional domain is mixed. The present study combined 2 methodologies for studying preferences in visual attention, eye tracking, and dot-probe, as younger and older adults viewed synthetic emotional faces. Eye tracking most consistently revealed a positivity effect in older adults' attention, so that older adults showed preferential looking toward happy faces and away from sad faces. Dot-probe results were less robust, but in the same direction. Methodological and theoretical implications for the study of socioemotional aging are discussed. (c) 2006 APA, all rights reserved
Wystrach, Antoine; Dewar, Alex; Philippides, Andrew; Graham, Paul
2016-02-01
The visual systems of animals have to provide information to guide behaviour and the informational requirements of an animal's behavioural repertoire are often reflected in its sensory system. For insects, this is often evident in the optical array of the compound eye. One behaviour that insects share with many animals is the use of learnt visual information for navigation. As ants are expert visual navigators it may be that their vision is optimised for navigation. Here we take a computational approach in asking how the details of the optical array influence the informational content of scenes used in simple view matching strategies for orientation. We find that robust orientation is best achieved with low-resolution visual information and a large field of view, similar to the optical properties seen for many ant species. A lower resolution allows for a trade-off between specificity and generalisation for stored views. Additionally, our simulations show that orientation performance increases if different portions of the visual field are considered as discrete visual sensors, each giving an independent directional estimate. This suggests that ants might benefit by processing information from their two eyes independently.
Changes of Visual Pathway and Brain Connectivity in Glaucoma: A Systematic Review
Nuzzi, Raffaele; Dallorto, Laura; Rolle, Teresa
2018-01-01
Background: Glaucoma is a leading cause of irreversible blindness worldwide. The increasing interest in the involvement of the cortical visual pathway in glaucomatous patients is due to the implications in recent therapies, such as neuroprotection and neuroregeneration. Objective: In this review, we outline the current understanding of brain structural, functional, and metabolic changes detected with the modern techniques of neuroimaging in glaucomatous subjects. Methods: We screened MEDLINE, EMBASE, CINAHL, CENTRAL, LILACS, Trip Database, and NICE for original contributions published until 31 October 2017. Studies with at least six patients affected by any type of glaucoma were considered. We included studies using the following neuroimaging techniques: functional Magnetic Resonance Imaging (fMRI), resting-state fMRI (rs-fMRI), magnetic resonance spectroscopy (MRS), voxel- based Morphometry (VBM), surface-based Morphometry (SBM), diffusion tensor MRI (DTI). Results: Over a total of 1,901 studies, 56 case series with a total of 2,381 patients were included. Evidence of neurodegenerative process in glaucomatous patients was found both within and beyond the visual system. Structural alterations in visual cortex (mainly reduced cortex thickness and volume) have been demonstrated with SBM and VBM; these changes were not limited to primary visual cortex but also involved association visual areas. Other brain regions, associated with visual function, demonstrated a certain grade of increased or decreased gray matter volume. Functional and metabolic abnormalities resulted within primary visual cortex in all studies with fMRI and MRS. Studies with rs-fMRI found disrupted connectivity between the primary and higher visual cortex and between visual cortex and associative visual areas in the task-free state of glaucomatous patients. Conclusions: This review contributes to the better understanding of brain abnormalities in glaucoma. It may stimulate further speculation about brain plasticity at a later age and therapeutic strategies, such as the prevention of cortical degeneration in patients with glaucoma. Structural, functional, and metabolic neuroimaging methods provided evidence of changes throughout the visual pathway in glaucomatous patients. Other brain areas, not directly involved in the processing of visual information, also showed alterations. PMID:29896087
Space-time wiring specificity supports direction selectivity in the retina
Zlateski, Aleksandar; Lee, Kisuk; Richardson, Mark; Turaga, Srinivas C.; Purcaro, Michael; Balkam, Matthew; Robinson, Amy; Behabadi, Bardia F.; Campos, Michael; Denk, Winfried; Seung, H. Sebastian
2014-01-01
How does the mammalian retina detect motion? This classic problem in visual neuroscience has remained unsolved for 50 years. In search of clues, we reconstructed Off-type starburst amacrine cells (SACs) and bipolar cells (BCs) in serial electron microscopic images with help from EyeWire, an online community of “citizen neuroscientists.” Based on quantitative analyses of contact area and branch depth in the retina, we found evidence that one BC type prefers to wire with a SAC dendrite near the SAC soma, while another BC type prefers to wire far from the soma. The near type is known to lag the far type in time of visual response. A mathematical model shows how such “space-time wiring specificity” could endow SAC dendrites with receptive fields that are oriented in space-time and therefore respond selectively to stimuli that move in the outward direction from the soma. PMID:24805243
Space-time wiring specificity supports direction selectivity in the retina.
Kim, Jinseop S; Greene, Matthew J; Zlateski, Aleksandar; Lee, Kisuk; Richardson, Mark; Turaga, Srinivas C; Purcaro, Michael; Balkam, Matthew; Robinson, Amy; Behabadi, Bardia F; Campos, Michael; Denk, Winfried; Seung, H Sebastian
2014-05-15
How does the mammalian retina detect motion? This classic problem in visual neuroscience has remained unsolved for 50 years. In search of clues, here we reconstruct Off-type starburst amacrine cells (SACs) and bipolar cells (BCs) in serial electron microscopic images with help from EyeWire, an online community of 'citizen neuroscientists'. On the basis of quantitative analyses of contact area and branch depth in the retina, we find evidence that one BC type prefers to wire with a SAC dendrite near the SAC soma, whereas another BC type prefers to wire far from the soma. The near type is known to lag the far type in time of visual response. A mathematical model shows how such 'space-time wiring specificity' could endow SAC dendrites with receptive fields that are oriented in space-time and therefore respond selectively to stimuli that move in the outward direction from the soma.
A comparison of visual and quantitative methods to identify interstitial lung abnormalities.
Kliment, Corrine R; Araki, Tetsuro; Doyle, Tracy J; Gao, Wei; Dupuis, Josée; Latourelle, Jeanne C; Zazueta, Oscar E; Fernandez, Isis E; Nishino, Mizuki; Okajima, Yuka; Ross, James C; Estépar, Raúl San José; Diaz, Alejandro A; Lederer, David J; Schwartz, David A; Silverman, Edwin K; Rosas, Ivan O; Washko, George R; O'Connor, George T; Hatabu, Hiroto; Hunninghake, Gary M
2015-10-29
Evidence suggests that individuals with interstitial lung abnormalities (ILA) on a chest computed tomogram (CT) may have an increased risk to develop a clinically significant interstitial lung disease (ILD). Although methods used to identify individuals with ILA on chest CT have included both automated quantitative and qualitative visual inspection methods, there has been not direct comparison between these two methods. To investigate this relationship, we created lung density metrics and compared these to visual assessments of ILA. To provide a comparison between ILA detection methods based on visual assessment we generated measures of high attenuation areas (HAAs, defined by attenuation values between -600 and -250 Hounsfield Units) in >4500 participants from both the COPDGene and Framingham Heart studies (FHS). Linear and logistic regressions were used for analyses. Increased measures of HAAs (in ≥ 10 % of the lung) were significantly associated with ILA defined by visual inspection in both cohorts (P < 0.0001); however, the positive predictive values were not very high (19 % in COPDGene and 13 % in the FHS). In COPDGene, the association between HAAs and ILA defined by visual assessment were modified by the percentage of emphysema and body mass index. Although increased HAAs were associated with reductions in total lung capacity in both cohorts, there was no evidence for an association between measurement of HAAs and MUC5B promoter genotype in the FHS. Our findings demonstrate that increased measures of lung density may be helpful in determining the severity of lung volume reduction, but alone, are not strongly predictive of ILA defined by visual assessment. Moreover, HAAs were not associated with MUC5B promoter genotype.
Effects of walker gender and observer gender on biological motion walking direction discrimination.
Yang, Xiaoying; Cai, Peng; Jiang, Yi
2014-09-01
The ability to recognize the movements of other biological entities, such as whether a person is walking toward you, is essential for survival and social interaction. Previous studies have shown that the visual system is particularly sensitive to approaching biological motion. In this study, we examined whether the gender of walkers and observers influenced the walking direction discrimination of approaching point-light walkers in fine granularity. The observers were presented a walker who walked in different directions and were asked to quickly judge the walking direction (left or right). The results showed that the observers demonstrated worse direction discrimination when the walker was depicted as male than when the walker was depicted as female, probably because the observers tended to perceive the male walkers as walking straight ahead. Intriguingly, male observers performed better than female observers at judging the walking directions of female walkers but not those of male walkers, a result indicating perceptual advantage with evolutionary significance. These findings provide strong evidence that the gender of walkers and observers modulates biological motion perception and that an adaptive perceptual mechanism exists in the visual system to facilitate the survival of social organisms. © 2014 The Institute of Psychology, Chinese Academy of Sciences and Wiley Publishing Asia Pty Ltd.
Visual attentional bias for food in adolescents with binge-eating disorder.
Schmidt, Ricarda; Lüthold, Patrick; Kittel, Rebekka; Tetzlaff, Anne; Hilbert, Anja
2016-09-01
Evidence suggests that adults with binge-eating disorder (BED) are prone of having their attention interfered by food cues, and that food-related attentional biases are associated with calorie intake and eating disorder psychopathology. For adolescents with BED experimental evidence on attentional processing of food cues is lacking. Using eye-tracking and a visual search task, the present study examined visual orienting and disengagement processes of food in youth with BED. Eye-movement data and reaction times were recorded in 25 adolescents (12-20 years) with BED and 25 controls (CG) individually matched for sex, age, body mass index, and socio-economic status. During a free exploration paradigm, the BED group showed a greater gaze duration bias for food images than the CG. Groups did not differ in gaze direction biases. In a visual search task, the BED group showed a greater detection bias for food targets than the CG. Group differences were more pronounced for personally attractive than unattractive food images. Regarding clinical associations, only in the BED group the gaze duration bias for food was associated with increased hunger and lower body mass index, and the detection bias for food targets was associated with greater reward sensitivity. The study provided first evidence of an attentional bias to food in adolescents with BED. However, more research is needed for further specifying disengagement and orienting processes in adolescent BED, including overt and covert attention, and their prospective associations with binge-eating behaviors and associated psychopathology. Copyright © 2016 Elsevier Ltd. All rights reserved.
Fine lattice lines on the corneal surface after laser in situ keratomileusis (LASIK).
Carpel, E F; Carlson, K H; Shannon, S
2000-03-01
To present an example of a pattern of lines resembling fine lattice on the corneal surface subsequent to laser in situ keratomileusis (LASIK). This subtle phenomenon may be relatively common and may affect visual outcome. Case report. A 41-year-old year old man with high myopia and best-corrected visual acuity of 20/20 +2 in each eye underwent laser in situ keratomileusis (LASIK). No operative or postoperative complications occurred. No striae were evident on slit-lamp examination with direct illumination and retroillumination at the time of surgery or in the postoperative period. Postoperative uncorrected visual acuity was 20/25 with a best-corrected spectacle correction of 20/25 in both eyes. Fine lines in a lattice pattern were seen only with fluorescein dye in the precorneal tear film as areas of "negative stain" within the LASIK flap. With tear film supplementation, the lines were less evident and visual acuity improved. One year postoperatively, his uncorrected visual acuity was 20/25 in both eyes. The best-corrected spectacle visual acuity was RE: 20/20 -2, LE: 20/25. The fine lines were still present within the flap. A soft contact lens improved visual acuity to 20/20 in both eyes. Although all four puncta were occluded, he had no epiphora. Fine lines in a lattice pattern that may represent folds in the epithelium or Bowman layer may be present within the flap after LASIK and may adversely affect visual acuity. They may be visible as areas of negative stain with fluorescein dye in the precorneal tear film in the absence of any striae visible in the flap. These superficial lines have been seen more in patients with high degrees of correction and in patients with dry eye. If visual acuity is affected, it may be improved with punctal occlusion, tear supplements, or a contact lens.
Pazzaglia, Mariella; Galli, Giulia
2015-01-01
The bidirectional flow of perceptual and motor information has recently proven useful as rehabilitative tool for re-building motor memories. We analyzed how the visual-motor approach has been successfully applied in neurorehabilitation, leading to surprisingly rapid and effective improvements in action execution. We proposed that the contribution of multiple sensory channels during treatment enables individuals to predict and optimize motor behavior, having a greater effect than visual input alone. We explored how the state-of-the-art neuroscience techniques show direct evidence that employment of visual-motor approach leads to increased motor cortex excitability and synaptic and cortical map plasticity. This super-additive response to multimodal stimulation may maximize neural plasticity, potentiating the effect of conventional treatment, and will be a valuable approach when it comes to advances in innovative methodologies.
Premotor cortex is sensitive to auditory-visual congruence for biological motion.
Wuerger, Sophie M; Parkes, Laura; Lewis, Penelope A; Crocker-Buque, Alex; Rutschmann, Roland; Meyer, Georg F
2012-03-01
The auditory and visual perception systems have developed special processing strategies for ecologically valid motion stimuli, utilizing some of the statistical properties of the real world. A well-known example is the perception of biological motion, for example, the perception of a human walker. The aim of the current study was to identify the cortical network involved in the integration of auditory and visual biological motion signals. We first determined the cortical regions of auditory and visual coactivation (Experiment 1); a conjunction analysis based on unimodal brain activations identified four regions: middle temporal area, inferior parietal lobule, ventral premotor cortex, and cerebellum. The brain activations arising from bimodal motion stimuli (Experiment 2) were then analyzed within these regions of coactivation. Auditory footsteps were presented concurrently with either an intact visual point-light walker (biological motion) or a scrambled point-light walker; auditory and visual motion in depth (walking direction) could either be congruent or incongruent. Our main finding is that motion incongruency (across modalities) increases the activity in the ventral premotor cortex, but only if the visual point-light walker is intact. Our results extend our current knowledge by providing new evidence consistent with the idea that the premotor area assimilates information across the auditory and visual modalities by comparing the incoming sensory input with an internal representation.
Top-down influences on visual attention during listening are modulated by observer sex.
Shen, John; Itti, Laurent
2012-07-15
In conversation, women have a small advantage in decoding non-verbal communication compared to men. In light of these findings, we sought to determine whether sex differences also existed in visual attention during a related listening task, and if so, if the differences existed among attention to high-level aspects of the scene or to conspicuous visual features. Using eye-tracking and computational techniques, we present direct evidence that men and women orient attention differently during conversational listening. We tracked the eyes of 15 men and 19 women who watched and listened to 84 clips featuring 12 different speakers in various outdoor settings. At the fixation following each saccadic eye movement, we analyzed the type of object that was fixated. Men gazed more often at the mouth and women at the eyes of the speaker. Women more often exhibited "distracted" saccades directed away from the speaker and towards a background scene element. Examining the multi-scale center-surround variation in low-level visual features (static: color, intensity, orientation, and dynamic: motion energy), we found that men consistently selected regions which expressed more variation in dynamic features, which can be attributed to a male preference for motion and a female preference for areas that may contain nonverbal information about the speaker. In sum, significant differences were observed, which we speculate arise from different integration strategies of visual cues in selecting the final target of attention. Our findings have implications for studies of sex in nonverbal communication, as well as for more predictive models of visual attention. Published by Elsevier Ltd.
Silvanto, Juha; Cattaneo, Zaira
2010-05-01
Cortical areas involved in sensory analysis are also believed to be involved in short-term storage of that sensory information. Here we investigated whether transcranial magnetic stimulation (TMS) can reveal the content of visual short-term memory (VSTM) by bringing this information to visual awareness. Subjects were presented with two random-dot displays (moving either to the left or to the right) and they were required to maintain one of these in VSTM. In Experiment 1, TMS was applied over the motion-selective area V5/MT+ above phosphene threshold during the maintenance phase. The reported phosphene contained motion features of the memory item, when the phosphene spatially overlapped with memory item. Specifically, phosphene motion was enhanced when the memory item moved in the same direction as the subjects' V5/MT+ baseline phosphene, whereas it was reduced when the motion direction of the memory item was incongruent with that of the baseline V5/MT+ phosphene. There was no effect on phosphene reports when there was no spatial overlap between the phosphene and the memory item. In Experiment 2, VSTM maintenance did not influence the appearance of phosphenes induced from the lateral occipital region. These interactions between VSTM maintenance and phosphene appearance demonstrate that activity in V5/MT+ reflects the motion qualities of items maintained in VSTM. Furthermore, these results also demonstrate that information in VSTM can modulate the pattern of visual activation reaching awareness, providing evidence for the view that overlapping neuronal populations are involved in conscious visual perception and VSTM. 2010. Published by Elsevier Inc.
The neural basis of visual word form processing: a multivariate investigation.
Nestor, Adrian; Behrmann, Marlene; Plaut, David C
2013-07-01
Current research on the neurobiological bases of reading points to the privileged role of a ventral cortical network in visual word processing. However, the properties of this network and, in particular, its selectivity for orthographic stimuli such as words and pseudowords remain topics of significant debate. Here, we approached this issue from a novel perspective by applying pattern-based analyses to functional magnetic resonance imaging data. Specifically, we examined whether, where and how, orthographic stimuli elicit distinct patterns of activation in the human cortex. First, at the category level, multivariate mapping found extensive sensitivity throughout the ventral cortex for words relative to false-font strings. Secondly, at the identity level, the multi-voxel pattern classification provided direct evidence that different pseudowords are encoded by distinct neural patterns. Thirdly, a comparison of pseudoword and face identification revealed that both stimulus types exploit common neural resources within the ventral cortical network. These results provide novel evidence regarding the involvement of the left ventral cortex in orthographic stimulus processing and shed light on its selectivity and discriminability profile. In particular, our findings support the existence of sublexical orthographic representations within the left ventral cortex while arguing for the continuity of reading with other visual recognition skills.
Reduced Distractibility in a Remote Culture
de Fockert, Jan W.; Caparos, Serge; Linnell, Karina J.; Davidoff, Jules
2011-01-01
Background In visual processing, there are marked cultural differences in the tendency to adopt either a global or local processing style. A remote culture (the Himba) has recently been reported to have a greater local bias in visual processing than Westerners. Here we give the first evidence that a greater, and remarkable, attentional selectivity provides the basis for this local bias. Methodology/Principal Findings In Experiment 1, Eriksen-type flanker interference was measured in the Himba and in Western controls. In both groups, responses to the direction of a task-relevant target arrow were affected by the compatibility of task-irrelevant distractor arrows. However, the Himba showed a marked reduction in overall flanker interference compared to Westerners. The smaller interference effect in the Himba occurred despite their overall slower performance than Westerners, and was evident even at a low level of perceptual load of the displays. In Experiment 2, the attentional selectivity of the Himba was further demonstrated by showing that their attention was not even captured by a moving singleton distractor. Conclusions/Significance We argue that the reduced distractibility in the Himba is clearly consistent with their tendency to prioritize the analysis of local details in visual processing. PMID:22046275
Reasoning and dyslexia: a spatial strategy may impede reasoning with visually rich information.
Bacon, Alison M; Handley, Simon J; McDonald, Emma L
2007-02-01
Bacon, Handley, and Newstead (2003, 2004), have presented evidence for individual differences in reasoning strategies, with most people seeming to represent and manipulate problem information using either a verbal or a spatial strategy. There is also evidence that individuals with dyslexia are inclined to conceptualise information in a visuo-spatial, rather than a verbal, way (e.g. von Károlyi et al., 2003). If so, we might expect a higher proportion of individuals with dyslexia to be spatial reasoners, compared with individuals who do not have dyslexia. The study reported here directly compared strategies reported by these two groups of participants on a syllogistic reasoning task. Moreover, problem content was manipulated so that reasoning across concrete and abstract materials could be compared. The findings suggest that whilst most individuals without dyslexia use a verbal strategy, reasoners with dyslexia do tend to adopt a spatial approach, though their performance is impaired with visually concrete materials. However, when reasoning with more abstract content, they perform comparably with non-dyslexic controls. The paper discusses these results in the light of recent research which has suggested that visual images may impede reasoning, and considers how individuals with dyslexia may differ from other reasoners.
Prefrontal cortex modulates posterior alpha oscillations during top-down guided visual perception
Helfrich, Randolph F.; Huang, Melody; Wilson, Guy; Knight, Robert T.
2017-01-01
Conscious visual perception is proposed to arise from the selective synchronization of functionally specialized but widely distributed cortical areas. It has been suggested that different frequency bands index distinct canonical computations. Here, we probed visual perception on a fine-grained temporal scale to study the oscillatory dynamics supporting prefrontal-dependent sensory processing. We tested whether a predictive context that was embedded in a rapid visual stream modulated the perception of a subsequent near-threshold target. The rapid stream was presented either rhythmically at 10 Hz, to entrain parietooccipital alpha oscillations, or arrhythmically. We identified a 2- to 4-Hz delta signature that modulated posterior alpha activity and behavior during predictive trials. Importantly, delta-mediated top-down control diminished the behavioral effects of bottom-up alpha entrainment. Simultaneous source-reconstructed EEG and cross-frequency directionality analyses revealed that this delta activity originated from prefrontal areas and modulated posterior alpha power. Taken together, this study presents converging behavioral and electrophysiological evidence for frontal delta-mediated top-down control of posterior alpha activity, selectively facilitating visual perception. PMID:28808023
A neural measure of precision in visual working memory.
Ester, Edward F; Anderson, David E; Serences, John T; Awh, Edward
2013-05-01
Recent studies suggest that the temporary storage of visual detail in working memory is mediated by sensory recruitment or sustained patterns of stimulus-specific activation within feature-selective regions of visual cortex. According to a strong version of this hypothesis, the relative "quality" of these patterns should determine the clarity of an individual's memory. Here, we provide a direct test of this claim. We used fMRI and a forward encoding model to characterize population-level orientation-selective responses in visual cortex while human participants held an oriented grating in memory. This analysis, which enables a precise quantitative description of multivoxel, population-level activity measured during working memory storage, revealed graded response profiles whose amplitudes were greatest for the remembered orientation and fell monotonically as the angular distance from this orientation increased. Moreover, interparticipant differences in the dispersion-but not the amplitude-of these response profiles were strongly correlated with performance on a concurrent memory recall task. These findings provide important new evidence linking the precision of sustained population-level responses in visual cortex and memory acuity.
Visual cues that are effective for contextual saccade adaptation
Azadi, Reza
2014-01-01
The accuracy of saccades, as maintained by saccade adaptation, has been shown to be context dependent: able to have different amplitude movements to the same retinal displacement dependent on motor contexts such as orbital starting location. There is conflicting evidence as to whether purely visual cues also effect contextual saccade adaptation and, if so, what function this might serve. We tested what visual cues might evoke contextual adaptation. Over 5 experiments, 78 naive subjects made saccades to circularly moving targets, which stepped outward or inward during the saccade depending on target movement direction, speed, or color and shape. To test if the movement or context postsaccade were critical, we stopped the postsaccade target motion (experiment 4) or neutralized the contexts by equating postsaccade target speed to an intermediate value (experiment 5). We found contextual adaptation in all conditions except those defined by color and shape. We conclude that some, but not all, visual cues before the saccade are sufficient for contextual adaptation. We conjecture that this visual contextuality functions to allow for different motor states for different coordinated movement patterns, such as coordinated saccade and pursuit motor planning. PMID:24647429
Joint representation of translational and rotational components of optic flow in parietal cortex
Sunkara, Adhira; DeAngelis, Gregory C.; Angelaki, Dora E.
2016-01-01
Terrestrial navigation naturally involves translations within the horizontal plane and eye rotations about a vertical (yaw) axis to track and fixate targets of interest. Neurons in the macaque ventral intraparietal (VIP) area are known to represent heading (the direction of self-translation) from optic flow in a manner that is tolerant to rotational visual cues generated during pursuit eye movements. Previous studies have also reported that eye rotations modulate the response gain of heading tuning curves in VIP neurons. We tested the hypothesis that VIP neurons simultaneously represent both heading and horizontal (yaw) eye rotation velocity by measuring heading tuning curves for a range of rotational velocities of either real or simulated eye movements. Three findings support the hypothesis of a joint representation. First, we show that rotation velocity selectivity based on gain modulations of visual heading tuning is similar to that measured during pure rotations. Second, gain modulations of heading tuning are similar for self-generated eye rotations and visually simulated rotations, indicating that the representation of rotation velocity in VIP is multimodal, driven by both visual and extraretinal signals. Third, we show that roughly one-half of VIP neurons jointly represent heading and rotation velocity in a multiplicatively separable manner. These results provide the first evidence, to our knowledge, for a joint representation of translation direction and rotation velocity in parietal cortex and show that rotation velocity can be represented based on visual cues, even in the absence of efference copy signals. PMID:27095846
Perceptual grouping enhances visual plasticity.
Mastropasqua, Tommaso; Turatto, Massimo
2013-01-01
Visual perceptual learning, a manifestation of neural plasticity, refers to improvements in performance on a visual task achieved by training. Attention is known to play an important role in perceptual learning, given that the observer's discriminative ability improves only for those stimulus feature that are attended. However, the distribution of attention can be severely constrained by perceptual grouping, a process whereby the visual system organizes the initial retinal input into candidate objects. Taken together, these two pieces of evidence suggest the interesting possibility that perceptual grouping might also affect perceptual learning, either directly or via attentional mechanisms. To address this issue, we conducted two experiments. During the training phase, participants attended to the contrast of the task-relevant stimulus (oriented grating), while two similar task-irrelevant stimuli were presented in the adjacent positions. One of the two flanking stimuli was perceptually grouped with the attended stimulus as a consequence of its similar orientation (Experiment 1) or because it was part of the same perceptual object (Experiment 2). A test phase followed the training phase at each location. Compared to the task-irrelevant no-grouping stimulus, orientation discrimination improved at the attended location. Critically, a perceptual learning effect equivalent to the one observed for the attended location also emerged for the task-irrelevant grouping stimulus, indicating that perceptual grouping induced a transfer of learning to the stimulus (or feature) being perceptually grouped with the task-relevant one. Our findings indicate that no voluntary effort to direct attention to the grouping stimulus or feature is necessary to enhance visual plasticity.
Modular Representation of Luminance Polarity In the Superficial Layers Of Primary Visual Cortex
Smith, Gordon B.; Whitney, David E.; Fitzpatrick, David
2016-01-01
Summary The spatial arrangement of luminance increments (ON) and decrements (OFF) falling on the retina provides a wealth of information used by central visual pathways to construct coherent representations of visual scenes. But how the polarity of luminance change is represented in the activity of cortical circuits remains unclear. Using wide-field epifluorescence and two-photon imaging we demonstrate a robust modular representation of luminance polarity (ON or OFF) in the superficial layers of ferret primary visual cortex. Polarity-specific domains are found with both uniform changes in luminance and single light/dark edges, and include neurons selective for orientation and direction of motion. The integration of orientation and polarity preference is evident in the selectivity and discrimination capabilities of most layer 2/3 neurons. We conclude that polarity selectivity is an integral feature of layer 2/3 neurons, ensuring that the distinction between light and dark stimuli is available for further processing in downstream extrastriate areas. PMID:26590348
Visual face-movement sensitive cortex is relevant for auditory-only speech recognition.
Riedel, Philipp; Ragert, Patrick; Schelinski, Stefanie; Kiebel, Stefan J; von Kriegstein, Katharina
2015-07-01
It is commonly assumed that the recruitment of visual areas during audition is not relevant for performing auditory tasks ('auditory-only view'). According to an alternative view, however, the recruitment of visual cortices is thought to optimize auditory-only task performance ('auditory-visual view'). This alternative view is based on functional magnetic resonance imaging (fMRI) studies. These studies have shown, for example, that even if there is only auditory input available, face-movement sensitive areas within the posterior superior temporal sulcus (pSTS) are involved in understanding what is said (auditory-only speech recognition). This is particularly the case when speakers are known audio-visually, that is, after brief voice-face learning. Here we tested whether the left pSTS involvement is causally related to performance in auditory-only speech recognition when speakers are known by face. To test this hypothesis, we applied cathodal transcranial direct current stimulation (tDCS) to the pSTS during (i) visual-only speech recognition of a speaker known only visually to participants and (ii) auditory-only speech recognition of speakers they learned by voice and face. We defined the cathode as active electrode to down-regulate cortical excitability by hyperpolarization of neurons. tDCS to the pSTS interfered with visual-only speech recognition performance compared to a control group without pSTS stimulation (tDCS to BA6/44 or sham). Critically, compared to controls, pSTS stimulation additionally decreased auditory-only speech recognition performance selectively for voice-face learned speakers. These results are important in two ways. First, they provide direct evidence that the pSTS is causally involved in visual-only speech recognition; this confirms a long-standing prediction of current face-processing models. Secondly, they show that visual face-sensitive pSTS is causally involved in optimizing auditory-only speech recognition. These results are in line with the 'auditory-visual view' of auditory speech perception, which assumes that auditory speech recognition is optimized by using predictions from previously encoded speaker-specific audio-visual internal models. Copyright © 2015 Elsevier Ltd. All rights reserved.
The differential contributions of visual imagery constructs on autobiographical thinking.
Aydin, Cagla
2018-02-01
There is a growing theoretical and empirical consensus on the central role of visual imagery in autobiographical memory. However, findings from studies that explore how individual differences in visual imagery are reflected on autobiographical thinking do not present a coherent story. One reason for the mixed findings was suggested to be the treatment of visual imagery as an undifferentiated construct while evidence shows that there is more than one type of visual imagery. The present study investigates the relative contributions of different imagery constructs; namely, object and spatial imagery, on autobiographical memory processes. Additionally, it explores whether a similar relation extends to imagining the future. The results indicate that while object imagery was significantly correlated with several phenomenological characteristics, such as the level of sensory and perceptual details for past events - but not for future events - spatial imagery predicted the level of episodic specificity for both past and future events. We interpret these findings as object imagery being recruited in tasks of autobiographical memory that employ reflective processes while spatial imagery is engaged during direct retrieval of event details. Implications for the role of visual imagery in autobiographical thinking processes are discussed.
Does constraining memory maintenance reduce visual search efficiency?
Buttaccio, Daniel R; Lange, Nicholas D; Thomas, Rick P; Dougherty, Michael R
2018-03-01
We examine whether constraining memory retrieval processes affects performance in a cued recall visual search task. In the visual search task, participants are first presented with a memory prompt followed by a search array. The memory prompt provides diagnostic information regarding a critical aspect of the target (its colour). We assume that upon the presentation of the memory prompt, participants retrieve and maintain hypotheses (i.e., potential target characteristics) in working memory in order to improve their search efficiency. By constraining retrieval through the manipulation of time pressure (Experiments 1A and 1B) or a concurrent working memory task (Experiments 2A, 2B, and 2C), we directly test the involvement of working memory in visual search. We find some evidence that visual search is less efficient under conditions in which participants were likely to be maintaining fewer hypotheses in working memory (Experiments 1A, 2A, and 2C), suggesting that the retrieval of representations from long-term memory into working memory can improve visual search. However, these results should be interpreted with caution, as the data from two experiments (Experiments 1B and 2B) did not lend support for this conclusion.
Hertz, Uri; Amedi, Amir
2015-01-01
The classical view of sensory processing involves independent processing in sensory cortices and multisensory integration in associative areas. This hierarchical structure has been challenged by evidence of multisensory responses in sensory areas, and dynamic weighting of sensory inputs in associative areas, thus far reported independently. Here, we used a visual-to-auditory sensory substitution algorithm (SSA) to manipulate the information conveyed by sensory inputs while keeping the stimuli intact. During scan sessions before and after SSA learning, subjects were presented with visual images and auditory soundscapes. The findings reveal 2 dynamic processes. First, crossmodal attenuation of sensory cortices changed direction after SSA learning from visual attenuations of the auditory cortex to auditory attenuations of the visual cortex. Secondly, associative areas changed their sensory response profile from strongest response for visual to that for auditory. The interaction between these phenomena may play an important role in multisensory processing. Consistent features were also found in the sensory dominance in sensory areas and audiovisual convergence in associative area Middle Temporal Gyrus. These 2 factors allow for both stability and a fast, dynamic tuning of the system when required. PMID:24518756
Hertz, Uri; Amedi, Amir
2015-08-01
The classical view of sensory processing involves independent processing in sensory cortices and multisensory integration in associative areas. This hierarchical structure has been challenged by evidence of multisensory responses in sensory areas, and dynamic weighting of sensory inputs in associative areas, thus far reported independently. Here, we used a visual-to-auditory sensory substitution algorithm (SSA) to manipulate the information conveyed by sensory inputs while keeping the stimuli intact. During scan sessions before and after SSA learning, subjects were presented with visual images and auditory soundscapes. The findings reveal 2 dynamic processes. First, crossmodal attenuation of sensory cortices changed direction after SSA learning from visual attenuations of the auditory cortex to auditory attenuations of the visual cortex. Secondly, associative areas changed their sensory response profile from strongest response for visual to that for auditory. The interaction between these phenomena may play an important role in multisensory processing. Consistent features were also found in the sensory dominance in sensory areas and audiovisual convergence in associative area Middle Temporal Gyrus. These 2 factors allow for both stability and a fast, dynamic tuning of the system when required. © The Author 2014. Published by Oxford University Press.
Testing Neuronal Accounts of Anisotropic Motion Perception with Computational Modelling
Wong, William; Chiang Price, Nicholas Seow
2014-01-01
There is an over-representation of neurons in early visual cortical areas that respond most strongly to cardinal (horizontal and vertical) orientations and directions of visual stimuli, and cardinal- and oblique-preferring neurons are reported to have different tuning curves. Collectively, these neuronal anisotropies can explain two commonly-reported phenomena of motion perception – the oblique effect and reference repulsion – but it remains unclear whether neuronal anisotropies can simultaneously account for both perceptual effects. We show in psychophysical experiments that reference repulsion and the oblique effect do not depend on the duration of a moving stimulus, and that brief adaptation to a single direction simultaneously causes a reference repulsion in the orientation domain, and the inverse of the oblique effect in the direction domain. We attempted to link these results to underlying neuronal anisotropies by implementing a large family of neuronal decoding models with parametrically varied levels of anisotropy in neuronal direction-tuning preferences, tuning bandwidths and spiking rates. Surprisingly, no model instantiation was able to satisfactorily explain our perceptual data. We argue that the oblique effect arises from the anisotropic distribution of preferred directions evident in V1 and MT, but that reference repulsion occurs separately, perhaps reflecting a process of categorisation occurring in higher-order cortical areas. PMID:25409518
Visual landmark-directed scatter-hoarding of Siberian chipmunks Tamias sibiricus.
Zhang, Dongyuan; Li, Jia; Wang, Zhenyu; Yi, Xianfeng
2016-05-01
Spatial memory of cached food items plays an important role in cache recovery by scatter-hoarding animals. However, whether scatter-hoarding animals intentionally select cache sites with respect to visual landmarks in the environment and then rely on them to recover their cached seeds for later use has not been extensively explored. Furthermore, there is a lack of evidence on whether there are sex differences in visual landmark-based food-hoarding behaviors in small rodents even though male and female animals exhibit different spatial abilities. In the present study, we used a scatter-hoarding animal, the Siberian chipmunk, Tamias sibiricus to explore these questions in semi-natural enclosures. Our results showed that T. sibiricus preferred to establish caches in the shallow pits labeled with visual landmarks (branches of Pinus sylvestris, leaves of Athyrium brevifrons and PVC tubes). In addition, visual landmarks of P. sylvestris facilitated cache recovery by T. sibiricus. We also found significant sex differences in visual landmark-based food-hoarding strategies in Siberian chipmunks. Males, rather than females, chipmunks tended to establish their caches with respect to the visual landmarks. Our studies show that T. sibiricus rely on visual landmarks to establish and recover their caches, and that sex differences exist in visual landmark-based food hoarding in Siberian chipmunks. © 2015 International Society of Zoological Sciences, Institute of Zoology/Chinese Academy of Sciences and John Wiley & Sons Australia, Ltd.
Testing of Visual Field with Virtual Reality Goggles in Manual and Visual Grasp Modes
Wroblewski, Dariusz; Francis, Brian A.; Sadun, Alfredo; Vakili, Ghazal; Chopra, Vikas
2014-01-01
Automated perimetry is used for the assessment of visual function in a variety of ophthalmic and neurologic diseases. We report development and clinical testing of a compact, head-mounted, and eye-tracking perimeter (VirtualEye) that provides a more comfortable test environment than the standard instrumentation. VirtualEye performs the equivalent of a full threshold 24-2 visual field in two modes: (1) manual, with patient response registered with a mouse click, and (2) visual grasp, where the eye tracker senses change in gaze direction as evidence of target acquisition. 59 patients successfully completed the test in manual mode and 40 in visual grasp mode, with 59 undergoing the standard Humphrey field analyzer (HFA) testing. Large visual field defects were reliably detected by VirtualEye. Point-by-point comparison between the results obtained with the different modalities indicates: (1) minimal systematic differences between measurements taken in visual grasp and manual modes, (2) the average standard deviation of the difference distributions of about 5 dB, and (3) a systematic shift (of 4–6 dB) to lower sensitivities for VirtualEye device, observed mostly in high dB range. The usability survey suggested patients' acceptance of the head-mounted device. The study appears to validate the concepts of a head-mounted perimeter and the visual grasp mode. PMID:25050326
Looking for an accident: glider pilots' visual management and potentially dangerous final turns.
Jarvis, Steve; Harris, Don
2007-06-01
Accidents caused by spinning from low turns continue to kill glider pilots despite the introduction of specific exercises aimed at increasing pilot awareness and recognition of this issue. In-cockpit video cameras were used to analyze flying accuracy and log the areas of visual interest of 36 qualified glider pilots performing final turns in a training glider. Pilots were found to divide their attention between four areas of interest: the view directly ahead; the landing area (right); the airspeed indicator; and an area between the direct ahead view and the landing area. The mean fixation rate was 85 shifts per minute. Significant correlations were found between over-use of rudder and a lack of attention to the view ahead, as well as between the overall fixation rate and poorer coordination in the turn. The results provide some evidence that a relationship exists between pilots' visual management and making turns in a potentially dangerous manner. Pilots who monitor the view ahead for reasonable periods during the final turn while not allowing their scan to become over-busy are those who are most likely to prevent a potential spin.
Brown, G C
1999-01-01
OBJECTIVE: To determine the relationship of visual acuity loss to quality of life. DESIGN: Three hundred twenty-five patients with visual loss to a minimum of 20/40 or greater in at least 1 eye were interviewed in a standardized fashion using a modified VF-14, questionnaire. Utility values were also obtained using both the time trade-off and standard gamble methods of utility assessment. MAIN OUTCOME MEASURES: Best-corrected visual acuity was correlated with the visual function score on the modified VF-14 questionnaire, as well as with utility values obtained using both the time trade-off and standard gamble methods. RESULTS: Decreasing levels of vision in the eye with better acuity correlated directly with decreasing visual function scores on the modified VF-14 questionnaire, as did decreasing utility values using the time trade-off method of utility evaluation. The standard gamble method of utility evaluation was not as directly correlated with vision as the time trade-off method. Age, level of education, gender, race, length of time of visual loss, and the number of associated systemic comorbidities did not significantly affect the time trade-off utility values associated with visual loss in the better eye. The level of reduced vision in the better eye, rather than the specific disease process causing reduced vision, was related to mean utility values. The average person with 20/40 vision in the better seeing eye was willing to trade 2 of every 10 years of life in return for perfect vision (utility value of 0.8), while the average person with counting fingers vision in the better eye was willing to trade approximately 5 of every 10 remaining years of life (utility value of 0.52) in return for perfect vision. CONCLUSIONS: The time trade-off method of utility evaluation appears to be an effective method for assessing quality of life associated with visual loss. Time trade-off utility values decrease in direct conjunction with decreasing vision in the better-seeing eye. Unlike the modified VF-14 test and its counterparts, utility values allow the quality of life associated with visual loss to be more readily compared to the quality of life associated with other health (disease) states. This information can be employed for cost-effective analyses that objectively compare evidence-based medicine, patient-based preferences and sound econometric principles across all specialties in health care. PMID:10703139
Direction of attentional focus in biofeedback treatment for /r/ misarticulation.
McAllister Byun, Tara; Swartz, Michelle T; Halpin, Peter F; Szeredi, Daniel; Maas, Edwin
2016-07-01
Maintaining an external direction of focus during practice is reported to facilitate acquisition of non-speech motor skills, but it is not known whether these findings also apply to treatment for speech errors. This question has particular relevance for treatment incorporating visual biofeedback, where clinician cueing can direct the learner's attention either internally (i.e., to the movements of the articulators) or externally (i.e., to the visual biofeedback display). This study addressed two objectives. First, it aimed to use single-subject experimental methods to collect additional evidence regarding the efficacy of visual-acoustic biofeedback treatment for children with /r/ misarticulation. Second, it compared the efficacy of this biofeedback intervention under two cueing conditions. In the external focus (EF) condition, participants' attention was directed exclusively to the external biofeedback display. In the internal focus (IF) condition, participants viewed a biofeedback display, but they also received articulatory cues encouraging an internal direction of attentional focus. Nine school-aged children were pseudo-randomly assigned to receive either IF or EF cues during 8 weeks of visual-acoustic biofeedback intervention. Accuracy in /r/ production at the word level was probed in three to five pre-treatment baseline sessions and in three post-treatment maintenance sessions. Outcomes were assessed using visual inspection and calculation of effect sizes for individual treatment trajectories. In addition, a mixed logistic model was used to examine across-subjects effects including phase (pre/post-treatment), /r/ variant (treated/untreated), and focus cue condition (internal/external). Six out of nine participants showed sustained improvement on at least one treated /r/ variant; these six participants were evenly divided across EF and IF treatment groups. Regression results indicated that /r/ productions were significantly more likely to be rated accurate post- than pre-treatment. Internal versus external direction of focus cues was not a significant predictor of accuracy, nor did it interact significantly with other predictors. The results are consistent with previous literature reporting that visual-acoustic biofeedback can produce measurable treatment gains in children who have not responded to previous intervention. These findings are also in keeping with previous research suggesting that biofeedback may be sufficient to establish an external attentional focus, independent of verbal cues provided. The finding that explicit articulator placement cues were not necessary for progress in treatment has implications for intervention practices for speech-sound disorders in children. © 2016 Royal College of Speech and Language Therapists.
Babatunde, O O; Tan, V; Jordan, J L; Dziedzic, K; Chew-Graham, C A; Jinks, C; Protheroe, J; van der Windt, D A
2018-06-01
Barriers to dissemination and engagement with evidence pose a threat to implementing evidence-based medicine. Understanding, retention, and recall can be enhanced by visual presentation of information. The aim of this exploratory research was to develop and evaluate the accessibility and acceptability of visual summaries for presenting evidence syntheses with multiple exposures or outcomes to professional and lay audiences. "Evidence flowers" were developed as a visual method of presenting data from 4 case scenarios: 2 complex evidence syntheses with multiple outcomes, Cochrane reviews, and clinical guidelines. Petals of evidence flowers were coloured according to the GRADE evidence rating system to display key findings and recommendations from the evidence summaries. Application of evidence flowers was observed during stakeholder workshops. Evaluation and feedback were conducted via questionnaires and informal interviews. Feedback from stakeholders on the evidence flowers collected from workshops, questionnaires, and interviews was encouraging and helpful for refining the design of the flowers. Comments were made on the content and design of the flowers, as well as the usability and potential for displaying different types of evidence. Evidence flowers are a novel and visually stimulating method for presenting research evidence from evidence syntheses with multiple exposures or outcomes, Cochrane reviews, and clinical guidelines. To promote access and engagement with research evidence, evidence flowers may be used in conjunction with other evidence synthesis products, such as (lay) summaries, evidence inventories, rapid reviews, and clinical guidelines. Additional research on potential adaptations and applications of the evidence flowers may further bridge the gap between research evidence and clinical practice. Copyright © 2018 John Wiley & Sons, Ltd.
Category-Selectivity in Human Visual Cortex Follows Cortical Topology: A Grouped icEEG Study
Conner, Christopher Richard; Whaley, Meagan Lee; Baboyan, Vatche George; Tandon, Nitin
2016-01-01
Neuroimaging studies suggest that category-selective regions in higher-order visual cortex are topologically organized around specific anatomical landmarks: the mid-fusiform sulcus (MFS) in the ventral temporal cortex (VTC) and lateral occipital sulcus (LOS) in the lateral occipital cortex (LOC). To derive precise structure-function maps from direct neural signals, we collected intracranial EEG (icEEG) recordings in a large human cohort (n = 26) undergoing implantation of subdural electrodes. A surface-based approach to grouped icEEG analysis was used to overcome challenges from sparse electrode coverage within subjects and variable cortical anatomy across subjects. The topology of category-selectivity in bilateral VTC and LOC was assessed for five classes of visual stimuli—faces, animate non-face (animals/body-parts), places, tools, and words—using correlational and linear mixed effects analyses. In the LOC, selectivity for living (faces and animate non-face) and non-living (places and tools) classes was arranged in a ventral-to-dorsal axis along the LOS. In the VTC, selectivity for living and non-living stimuli was arranged in a latero-medial axis along the MFS. Written word-selectivity was reliably localized to the intersection of the left MFS and the occipito-temporal sulcus. These findings provide direct electrophysiological evidence for topological information structuring of functional representations within higher-order visual cortex. PMID:27272936
Dichoptic training enables the adult amblyopic brain to learn.
Li, Jinrong; Thompson, Benjamin; Deng, Daming; Chan, Lily Y L; Yu, Minbin; Hess, Robert F
2013-04-22
Adults with amblyopia, a common visual cortex disorder caused primarily by binocular disruption during an early critical period, do not respond to conventional therapy involving occlusion of one eye. But it is now clear that the adult human visual cortex has a significant degree of plasticity, suggesting that something must be actively preventing the adult brain from learning to see through the amblyopic eye. One possibility is an inhibitory signal from the contralateral eye that suppresses cortical inputs from the amblyopic eye. Such a gating mechanism could explain the apparent lack of plasticity within the adult amblyopic visual cortex. Here we provide direct evidence that alleviating suppression of the amblyopic eye through dichoptic stimulus presentation induces greater levels of plasticity than forced use of the amblyopic eye alone. This indicates that suppression is a key gating mechanism that prevents the amblyopic brain from learning to see. Copyright © 2013 Elsevier Ltd. All rights reserved.
Interaction between gaze and visual and proprioceptive position judgements.
Fiehler, Katja; Rösler, Frank; Henriques, Denise Y P
2010-06-01
There is considerable evidence that targets for action are represented in a dynamic gaze-centered frame of reference, such that each gaze shift requires an internal updating of the target. Here, we investigated the effect of eye movements on the spatial representation of targets used for position judgements. Participants had their hand passively placed to a location, and then judged whether this location was left or right of a remembered visual or remembered proprioceptive target, while gaze direction was varied. Estimates of position of the remembered targets relative to the unseen position of the hand were assessed with an adaptive psychophysical procedure. These positional judgements significantly varied relative to gaze for both remembered visual and remembered proprioceptive targets. Our results suggest that relative target positions may also be represented in eye-centered coordinates. This implies similar spatial reference frames for action control and space perception when positions are coded relative to the hand.
Tracking the allocation of attention using human pupillary oscillations
Naber, Marnix; Alvarez, George A.; Nakayama, Ken
2013-01-01
The muscles that control the pupil are richly innervated by the autonomic nervous system. While there are central pathways that drive pupil dilations in relation to arousal, there is no anatomical evidence that cortical centers involved with visual selective attention innervate the pupil. In this study, we show that such connections must exist. Specifically, we demonstrate a novel Pupil Frequency Tagging (PFT) method, where oscillatory changes in stimulus brightness over time are mirrored by pupil constrictions and dilations. We find that the luminance–induced pupil oscillations are enhanced when covert attention is directed to the flicker stimulus and when targets are correctly detected in an attentional tracking task. These results suggest that the amplitudes of pupil responses closely follow the allocation of focal visual attention and the encoding of stimuli. PFT provides a new opportunity to study top–down visual attention itself as well as identifying the pathways and mechanisms that support this unexpected phenomenon. PMID:24368904
Bosworth, Rain G.; Petrich, Jennifer A.; Dobkins, Karen R.
2012-01-01
In order to investigate differences in the effects of spatial attention between the left visual field (LVF) and the right visual field (RVF), we employed a full/poor attention paradigm using stimuli presented in the LVF vs. RVF. In addition, to investigate differences in the effects of spatial attention between the Dorsal and Ventral processing streams, we obtained motion thresholds (motion coherence thresholds and fine direction discrimination thresholds) and orientation thresholds, respectively. The results of this study showed negligible effects of attention on the orientation task, in either the LVF or RVF. In contrast, for both motion tasks, there was a significant effect of attention in the LVF, but not in the RVF. These data provide psychophysical evidence for greater effects of spatial attention in the LVF/right hemisphere, specifically, for motion processing in the Dorsal stream. PMID:22051893
A Limited Role for Suppression in the Central Field of Individuals with Strabismic Amblyopia
Barrett, Brendan T.; Panesar, Gurvinder K.; Scally, Andrew J.; Pacey, Ian E.
2012-01-01
Background Although their eyes are pointing in different directions, people with long-standing strabismic amblyopia typically do not experience double-vision or indeed any visual symptoms arising from their condition. It is generally believed that the phenomenon of suppression plays a major role in dealing with the consequences of amblyopia and strabismus, by preventing images from the weaker/deviating eye from reaching conscious awareness. Suppression is thus a highly sophisticated coping mechanism. Although suppression has been studied for over 100 years the literature is equivocal in relation to the extent of the retina that is suppressed, though the method used to investigate suppression is crucial to the outcome. There is growing evidence that some measurement methods lead to artefactual claims that suppression exists when it does not. Methodology/Results Here we present the results of an experiment conducted with a new method to examine the prevalence, depth and extent of suppression in ten individuals with strabismic amblyopia. Seven subjects (70%) showed no evidence whatsoever for suppression and in the three individuals who did (30%), the depth and extent of suppression was small. Conclusions Suppression may play a much smaller role in dealing with the negative consequences of strabismic amblyopia than previously thought. Whereas recent claims of this nature have been made only in those with micro-strabismus our results show extremely limited evidence for suppression across the central visual field in strabismic amblyopes more generally. Instead of suppressing the image from the weaker/deviating eye, we suggest the visual system of individuals with strabismic amblyopia may act to maximise the possibilities for binocular co-operation. This is consistent with recent evidence from strabismic and amblyopic individuals that their binocular mechanisms are intact, and that, just as in visual normals, performance with two eyes is better than with the better eye alone in these individuals. PMID:22649494
A limited role for suppression in the central field of individuals with strabismic amblyopia.
Barrett, Brendan T; Panesar, Gurvinder K; Scally, Andrew J; Pacey, Ian E
2012-01-01
Although their eyes are pointing in different directions, people with long-standing strabismic amblyopia typically do not experience double-vision or indeed any visual symptoms arising from their condition. It is generally believed that the phenomenon of suppression plays a major role in dealing with the consequences of amblyopia and strabismus, by preventing images from the weaker/deviating eye from reaching conscious awareness. Suppression is thus a highly sophisticated coping mechanism. Although suppression has been studied for over 100 years the literature is equivocal in relation to the extent of the retina that is suppressed, though the method used to investigate suppression is crucial to the outcome. There is growing evidence that some measurement methods lead to artefactual claims that suppression exists when it does not. Here we present the results of an experiment conducted with a new method to examine the prevalence, depth and extent of suppression in ten individuals with strabismic amblyopia. Seven subjects (70%) showed no evidence whatsoever for suppression and in the three individuals who did (30%), the depth and extent of suppression was small. Suppression may play a much smaller role in dealing with the negative consequences of strabismic amblyopia than previously thought. Whereas recent claims of this nature have been made only in those with micro-strabismus our results show extremely limited evidence for suppression across the central visual field in strabismic amblyopes more generally. Instead of suppressing the image from the weaker/deviating eye, we suggest the visual system of individuals with strabismic amblyopia may act to maximise the possibilities for binocular co-operation. This is consistent with recent evidence from strabismic and amblyopic individuals that their binocular mechanisms are intact, and that, just as in visual normals, performance with two eyes is better than with the better eye alone in these individuals.
A dual systems account of visual perception: Predicting candy consumption from distance estimates.
Krpan, Dario; Schnall, Simone
2017-04-01
A substantial amount of evidence shows that visual perception is influenced by forces that control human actions, ranging from motivation to physiological potential. However, studies have not yet provided convincing evidence that perception itself is directly involved in everyday behaviors such as eating. We suggest that this issue can be resolved by employing the dual systems account of human behavior. We tested the link between perceived distance to candies and their consumption for participants who were tired or depleted (impulsive system), versus those who were not (reflective system). Perception predicted eating only when participants were tired (Experiment 1) or depleted (Experiments 2 and 3). In contrast, a rational determinant of behavior-eating restraint towards candies-predicted eating for non-depleted individuals (Experiment 2). Finally, Experiment 3 established that perceived distance was correlated with participants' self-reported motivation to consume candies. Overall, these findings suggest that the dynamics between perception and behavior depend on the interplay of the two behavioral systems. Copyright © 2017 Elsevier B.V. All rights reserved.
van Oosterom, L; Montgomery, J C; Jeffs, A G; Radford, C A
2016-01-11
Soundscapes provide a new tool for the study of fish communities. Bigeyes (Pempheris adspersa) are nocturnal planktivorous reef fish, feed in loose shoals and are soniferous. These vocalisations have been suggested to be contact calls to maintain group cohesion, however direct evidence for this is absent, despite the fact that contact calls are well documented for many other vertebrates, including marine mammals. For fish, direct evidence for group cohesion signals is restricted to the use of visual and hydrodynamic cues. In support of adding vocalisation as a contributing cue, our laboratory experiments show that bigeyes significantly increased group cohesion when exposed to recordings of ambient reef sound at higher sound levels while also decreasing vocalisations. These patterns of behaviour are consistent with acoustic masking. When exposed to playback of conspecific vocalisations, the group cohesion and vocalisation rates of bigeyes both significantly increased. These results provide the first direct experimental support for the hypotheses that vocalisations are used as contact calls to maintain group cohesion in fishes, making fish the evolutionarily oldest vertebrate group in which this phenomenon has been observed, and adding a new dimension to the interpretation of nocturnal reef soundscapes.
NASA Astrophysics Data System (ADS)
van Oosterom, L.; Montgomery, J. C.; Jeffs, A. G.; Radford, C. A.
2016-01-01
Soundscapes provide a new tool for the study of fish communities. Bigeyes (Pempheris adspersa) are nocturnal planktivorous reef fish, feed in loose shoals and are soniferous. These vocalisations have been suggested to be contact calls to maintain group cohesion, however direct evidence for this is absent, despite the fact that contact calls are well documented for many other vertebrates, including marine mammals. For fish, direct evidence for group cohesion signals is restricted to the use of visual and hydrodynamic cues. In support of adding vocalisation as a contributing cue, our laboratory experiments show that bigeyes significantly increased group cohesion when exposed to recordings of ambient reef sound at higher sound levels while also decreasing vocalisations. These patterns of behaviour are consistent with acoustic masking. When exposed to playback of conspecific vocalisations, the group cohesion and vocalisation rates of bigeyes both significantly increased. These results provide the first direct experimental support for the hypotheses that vocalisations are used as contact calls to maintain group cohesion in fishes, making fish the evolutionarily oldest vertebrate group in which this phenomenon has been observed, and adding a new dimension to the interpretation of nocturnal reef soundscapes.
van Oosterom, L.; Montgomery, J. C.; Jeffs, A. G.; Radford, C. A.
2016-01-01
Soundscapes provide a new tool for the study of fish communities. Bigeyes (Pempheris adspersa) are nocturnal planktivorous reef fish, feed in loose shoals and are soniferous. These vocalisations have been suggested to be contact calls to maintain group cohesion, however direct evidence for this is absent, despite the fact that contact calls are well documented for many other vertebrates, including marine mammals. For fish, direct evidence for group cohesion signals is restricted to the use of visual and hydrodynamic cues. In support of adding vocalisation as a contributing cue, our laboratory experiments show that bigeyes significantly increased group cohesion when exposed to recordings of ambient reef sound at higher sound levels while also decreasing vocalisations. These patterns of behaviour are consistent with acoustic masking. When exposed to playback of conspecific vocalisations, the group cohesion and vocalisation rates of bigeyes both significantly increased. These results provide the first direct experimental support for the hypotheses that vocalisations are used as contact calls to maintain group cohesion in fishes, making fish the evolutionarily oldest vertebrate group in which this phenomenon has been observed, and adding a new dimension to the interpretation of nocturnal reef soundscapes. PMID:26750559
Pursey, Kirrilly M.; Stanwell, Peter; Callister, Robert J.; Brain, Katherine; Collins, Clare E.; Burrows, Tracy L.
2014-01-01
Emerging evidence from recent neuroimaging studies suggests that specific food-related behaviors contribute to the development of obesity. The aim of this review was to report the neural responses to visual food cues, as assessed by functional magnetic resonance imaging (fMRI), in humans of differing weight status. Published studies to 2014 were retrieved and included if they used visual food cues, studied humans >18 years old, reported weight status, and included fMRI outcomes. Sixty studies were identified that investigated the neural responses of healthy weight participants (n = 26), healthy weight compared to obese participants (n = 17), and weight-loss interventions (n = 12). High-calorie food images were used in the majority of studies (n = 36), however, image selection justification was only provided in 19 studies. Obese individuals had increased activation of reward-related brain areas including the insula and orbitofrontal cortex in response to visual food cues compared to healthy weight individuals, and this was particularly evident in response to energy dense cues. Additionally, obese individuals were more responsive to food images when satiated. Meta-analysis of changes in neural activation post-weight loss revealed small areas of convergence across studies in brain areas related to emotion, memory, and learning, including the cingulate gyrus, lentiform nucleus, and precuneus. Differential activation patterns to visual food cues were observed between obese, healthy weight, and weight-loss populations. Future studies require standardization of nutrition variables and fMRI outcomes to enable more direct comparisons between studies. PMID:25988110
Pursey, Kirrilly M; Stanwell, Peter; Callister, Robert J; Brain, Katherine; Collins, Clare E; Burrows, Tracy L
2014-01-01
Emerging evidence from recent neuroimaging studies suggests that specific food-related behaviors contribute to the development of obesity. The aim of this review was to report the neural responses to visual food cues, as assessed by functional magnetic resonance imaging (fMRI), in humans of differing weight status. Published studies to 2014 were retrieved and included if they used visual food cues, studied humans >18 years old, reported weight status, and included fMRI outcomes. Sixty studies were identified that investigated the neural responses of healthy weight participants (n = 26), healthy weight compared to obese participants (n = 17), and weight-loss interventions (n = 12). High-calorie food images were used in the majority of studies (n = 36), however, image selection justification was only provided in 19 studies. Obese individuals had increased activation of reward-related brain areas including the insula and orbitofrontal cortex in response to visual food cues compared to healthy weight individuals, and this was particularly evident in response to energy dense cues. Additionally, obese individuals were more responsive to food images when satiated. Meta-analysis of changes in neural activation post-weight loss revealed small areas of convergence across studies in brain areas related to emotion, memory, and learning, including the cingulate gyrus, lentiform nucleus, and precuneus. Differential activation patterns to visual food cues were observed between obese, healthy weight, and weight-loss populations. Future studies require standardization of nutrition variables and fMRI outcomes to enable more direct comparisons between studies.
Electrophysiological evidence for phenomenal consciousness.
Revonsuo, Antti; Koivisto, Mika
2010-09-01
Abstract Recent evidence from event-related brain potentials (ERPs) lends support to two central theses in Lamme's theory. The earliest ERP correlate of visual consciousness appears over posterior visual cortex around 100-200 ms after stimulus onset. Its scalp topography and time window are consistent with recurrent processing in the visual cortex. This electrophysiological correlate of visual consciousness is mostly independent of later ERPs reflecting selective attention and working memory functions. Overall, the ERP evidence supports the view that phenomenal consciousness of a visual stimulus emerges earlier than access consciousness, and that attention and awareness are served by distinct neural processes.
Internal model of gravity influences configural body processing.
Barra, Julien; Senot, Patrice; Auclair, Laurent
2017-01-01
Human bodies are processed by a configural processing mechanism. Evidence supporting this claim is the body inversion effect, in which inversion impairs recognition of bodies more than other objects. Biomechanical configuration, as well as both visual and embodied expertise, has been demonstrated to play an important role in this effect. Nevertheless, the important factor of body inversion effect may also be linked to gravity orientation since gravity is one of the most fundamental constraints of our biology, behavior, and perception on Earth. The visual presentation of an inverted body in a typical body inversion paradigm turns the observed body upside down but also inverts the implicit direction of visual gravity in the scene. The orientation of visual gravity is then in conflict with the direction of actual gravity and may influence configural processing. To test this hypothesis, we dissociated the orientations of the body and of visual gravity by manipulating body posture. In a pretest we showed that it was possible to turn an avatar upside down (inversion relative to retinal coordinates) without inverting the orientation of visual gravity when the avatar stands on his/her hands. We compared the inversion effect in typical conditions (with gravity conflict when the avatar is upside down) to the inversion effect in conditions with no conflict between visual and physical gravity. The results of our experiment revealed that the inversion effect, as measured by both error rate and reaction time, was strongly reduced when there was no gravity conflict. Our results suggest that when an observed body is upside down (inversion relative to participants' retinal coordinates) but the orientation of visual gravity is not, configural processing of bodies might still be possible. In this paper, we discuss the implications of an internal model of gravity in the configural processing of observed bodies. Copyright © 2016 Elsevier B.V. All rights reserved.
Local and Global Auditory Processing: Behavioral and ERP Evidence
Sanders, Lisa D.; Poeppel, David
2007-01-01
Differential processing of local and global visual features is well established. Global precedence effects, differences in event-related potentials (ERPs) elicited when attention is focused on local versus global levels, and hemispheric specialization for local and global features all indicate that relative scale of detail is an important distinction in visual processing. Observing analogous differential processing of local and global auditory information would suggest that scale of detail is a general organizational principle of the brain. However, to date the research on auditory local and global processing has primarily focused on music perception or on the perceptual analysis of relatively higher and lower frequencies. The study described here suggests that temporal aspects of auditory stimuli better capture the local-global distinction. By combining short (40 ms) frequency modulated tones in series to create global auditory patterns (500 ms), we independently varied whether pitch increased or decreased over short time spans (local) and longer time spans (global). Accuracy and reaction time measures revealed better performance for global judgments and asymmetric interference that were modulated by amount of pitch change. ERPs recorded while participants listened to identical sounds and indicated the direction of pitch change at the local or global levels provided evidence for differential processing similar to that found in ERP studies employing hierarchical visual stimuli. ERP measures failed to provide evidence for lateralization of local and global auditory perception, but differences in distributions suggest preferential processing in more ventral and dorsal areas respectively. PMID:17113115
Predictive Feedback and Conscious Visual Experience
Panichello, Matthew F.; Cheung, Olivia S.; Bar, Moshe
2012-01-01
The human brain continuously generates predictions about the environment based on learned regularities in the world. These predictions actively and efficiently facilitate the interpretation of incoming sensory information. We review evidence that, as a result of this facilitation, predictions directly influence conscious experience. Specifically, we propose that predictions enable rapid generation of conscious percepts and bias the contents of awareness in situations of uncertainty. The possible neural mechanisms underlying this facilitation are discussed. PMID:23346068
Nanoscale Chemical Imaging of an Individual Catalyst Particle with Soft X-ray Ptychography
Wise, Anna M.; Weker, Johanna Nelson; Kalirai, Sam; ...
2016-02-26
Understanding Fe deposition in fluid catalytic cracking (FCC) catalysis is critical for the mitigation of catalyst degradation. We employ soft X-ray ptychography to determine at the nanoscale the distribution and chemical state of Fe in an aged FCC catalyst particle. We also show that both particle swelling due to colloidal Fe deposition and Fe penetration into the matrix as a result of precracking of large organic molecules occur. Furthermore, the application of ptychography allowed us to provide direct visual evidence for these two distinct Fe-based deactivation mechanisms, which have so far been proposed only on the basis of indirect evidence.
Tanahashi, Shigehito; Ashihara, Kaoru; Ujike, Hiroyasu
2015-01-01
Recent studies have found that self-motion perception induced by simultaneous presentation of visual and auditory motion is facilitated when the directions of visual and auditory motion stimuli are identical. They did not, however, examine possible contributions of auditory motion information for determining direction of self-motion perception. To examine this, a visual stimulus projected on a hemisphere screen and an auditory stimulus presented through headphones were presented separately or simultaneously, depending on experimental conditions. The participant continuously indicated the direction and strength of self-motion during the 130-s experimental trial. When the visual stimulus with a horizontal shearing rotation and the auditory stimulus with a horizontal one-directional rotation were presented simultaneously, the duration and strength of self-motion perceived in the opposite direction of the auditory rotation stimulus were significantly longer and stronger than those perceived in the same direction of the auditory rotation stimulus. However, the auditory stimulus alone could not sufficiently induce self-motion perception, and if it did, its direction was not consistent within each experimental trial. We concluded that auditory motion information can determine perceived direction of self-motion during simultaneous presentation of visual and auditory motion information, at least when visual stimuli moved in opposing directions (around the yaw-axis). We speculate that the contribution of auditory information depends on the plausibility and information balance of visual and auditory information. PMID:26113828
Deficit in visual temporal integration in autism spectrum disorders.
Nakano, Tamami; Ota, Haruhisa; Kato, Nobumasa; Kitazawa, Shigeru
2010-04-07
Individuals with autism spectrum disorders (ASD) are superior in processing local features. Frith and Happe conceptualize this cognitive bias as 'weak central coherence', implying that a local enhancement derives from a weakness in integrating local elements into a coherent whole. The suggested deficit has been challenged, however, because individuals with ASD were not found to be inferior to normal controls in holistic perception. In these opposing studies, however, subjects were encouraged to ignore local features and attend to the whole. Therefore, no one has directly tested whether individuals with ASD are able to integrate local elements over time into a whole image. Here, we report a weakness of individuals with ASD in naming familiar objects moved behind a narrow slit, which was worsened by the absence of local salient features. The results indicate that individuals with ASD have a clear deficit in integrating local visual information over time into a global whole, providing direct evidence for the weak central coherence hypothesis.
Horschig, Jörn M; Smolders, Ruud; Bonnefond, Mathilde; Schoffelen, Jan-Mathijs; van den Munckhof, Pepijn; Schuurman, P Richard; Cools, Roshan; Denys, Damiaan; Jensen, Ole
2015-01-01
Here, we report evidence for oscillatory bi-directional interactions between the nucleus accumbens and the neocortex in humans. Six patients performed a demanding covert visual attention task while we simultaneously recorded brain activity from deep-brain electrodes implanted in the nucleus accumbens and the surface electroencephalogram (EEG). Both theta and alpha oscillations were strongly coherent with the frontal and parietal EEG during the task. Theta-band coherence increased during processing of the visual stimuli. Granger causality analysis revealed that the nucleus accumbens was communicating with the neocortex primarily in the theta-band, while the cortex was communicating the nucleus accumbens in the alpha-band. These data are consistent with a model, in which theta- and alpha-band oscillations serve dissociable roles: Prior to stimulus processing, the cortex might suppress ongoing processing in the nucleus accumbens by modulating alpha-band activity. Subsequently, upon stimulus presentation, theta oscillations might facilitate the active exchange of stimulus information from the nucleus accumbens to the cortex.
ERIC Educational Resources Information Center
Hendrickson, Homer
1988-01-01
Spelling problems arise due to problems with form discrimination and inadequate visualization. A child's sequence of visual development involves learning motor control and coordination, with vision directing and monitoring the movements; learning visual comparison of size, shape, directionality, and solidity; developing visual memory or recall;…
Extraretinal induced visual sensations during IMRT of the brain.
Wilhelm-Buchstab, Timo; Buchstab, Barbara Myrthe; Leitzen, Christina; Garbe, Stephan; Müdder, Thomas; Oberste-Beulmann, Susanne; Sprinkart, Alois Martin; Simon, Birgit; Nelles, Michael; Block, Wolfgang; Schoroth, Felix; Schild, Hans Heinz; Schüller, Heinrich
2015-01-01
We observed visual sensations (VSs) in patients undergoing intensity modulated radiotherapy (IMRT) of the brain without the beam passing through ocular structures. We analyzed this phenomenon especially with regards to reproducibility, and origin. Analyzed were ten consecutive patients (aged 41-71 years) with glioblastoma multiforme who received pulsed IMRT (total dose 60Gy) with helical tomotherapy (TT). A megavolt-CT (MVCT) was performed daily before treatment. VSs were reported and recorded using a triggered event recorder. The frequency of VSs was calculated and VSs were correlated with beam direction and couch position. Subjective patient perception was plotted on an 8x8 visual field (VF) matrix. Distance to the orbital roof (OR) from the first beam causing a VS was calculated from the Dicom radiation therapy data and MVCT data. During 175 treatment sessions (average 17.5 per patient) 5959 VSs were recorded and analyzed. VSs occurred only during the treatment session not during the MVCTs. Plotting events over time revealed patient-specific patterns. The average cranio-caudad extension of VS-inducing area was 63.4mm (range 43.24-92.1mm). The maximum distance between the first VS and the OR was 56.1mm so that direct interaction with the retina is unlikely. Data on subjective visual perception showed that VSs occurred mainly in the upper right and left quadrants of the VF. Within the visual pathways the highest probability for origin of VSs was seen in the optic chiasm and the optic tract (22%). There is clear evidence that interaction of photon irradiation with neuronal structures distant from the eye can lead to VSs.
Perceptual Grouping Enhances Visual Plasticity
Mastropasqua, Tommaso; Turatto, Massimo
2013-01-01
Visual perceptual learning, a manifestation of neural plasticity, refers to improvements in performance on a visual task achieved by training. Attention is known to play an important role in perceptual learning, given that the observer's discriminative ability improves only for those stimulus feature that are attended. However, the distribution of attention can be severely constrained by perceptual grouping, a process whereby the visual system organizes the initial retinal input into candidate objects. Taken together, these two pieces of evidence suggest the interesting possibility that perceptual grouping might also affect perceptual learning, either directly or via attentional mechanisms. To address this issue, we conducted two experiments. During the training phase, participants attended to the contrast of the task-relevant stimulus (oriented grating), while two similar task-irrelevant stimuli were presented in the adjacent positions. One of the two flanking stimuli was perceptually grouped with the attended stimulus as a consequence of its similar orientation (Experiment 1) or because it was part of the same perceptual object (Experiment 2). A test phase followed the training phase at each location. Compared to the task-irrelevant no-grouping stimulus, orientation discrimination improved at the attended location. Critically, a perceptual learning effect equivalent to the one observed for the attended location also emerged for the task-irrelevant grouping stimulus, indicating that perceptual grouping induced a transfer of learning to the stimulus (or feature) being perceptually grouped with the task-relevant one. Our findings indicate that no voluntary effort to direct attention to the grouping stimulus or feature is necessary to enhance visual plasticity. PMID:23301100
A novel examination of exposure patterns and posttraumatic stress after a university mass murder.
Liu, Sabrina R; Kia-Keating, Maryam
2018-03-05
Occurring at an alarming rate in the United States, mass violence has been linked to posttraumatic stress symptoms (PTSS) in both direct victims and community members who are indirectly exposed. Identifying what distinct exposure patterns exist and their relation to later PTSS has important clinical implications. The present study determined classes of exposure to an event of mass violence, and if PTSS differed across classes. First- and second-year college students (N = 1,189) participated in a confidential online survey following a mass murder at their university, which assessed event exposure and PTSS 3 months later. Latent class analysis (LCA) was used to empirically determine distinct classes of exposure patterns and links between class membership and PTSS. The final model yielded 4 classes: minimal exposure (55.5% of sample), auditory exposure (29.4% of sample), visual exposure (10% of sample), and interpersonal exposure (5% of sample). More severe direct exposure (i.e., the visual exposure class) was associated with significantly higher levels of PTSS than the auditory exposure or minimal exposure classes, as was the interpersonal exposure class. There were no significant differences in PTSS between the auditory exposure and minimal exposure classes or the visual exposure and interpersonal exposure classes. Results point to the differential impact of exposure categories, and provide empirical evidence for distinguishing among auditory, visual, and interpersonal exposures to events of mass violence on college campuses. Clinical implications suggest that visual and interpersonal exposure may warrant targeted efforts following mass violence. (PsycINFO Database Record (c) 2018 APA, all rights reserved).
Acton, Jennifer H; Molik, Bablin; Binns, Alison; Court, Helen; Margrain, Tom H
2016-02-24
Visual Rehabilitation Officers help people with a visual impairment maintain their independence. This intervention adopts a flexible, goal-centred approach, which may include training in mobility, use of optical and non-optical aids, and performance of activities of daily living. Although Visual Rehabilitation Officers are an integral part of the low vision service in the United Kingdom, evidence that they are effective is lacking. The purpose of this exploratory trial is to estimate the impact of a Visual Rehabilitation Officer on self-reported visual function, psychosocial and quality-of-life outcomes in individuals with low vision. In this exploratory, assessor-masked, parallel group, randomised controlled trial, participants will be allocated either to receive home visits from a Visual Rehabilitation Officer (n = 30) or to a waiting list control group (n = 30) in a 1:1 ratio. Adult volunteers with a visual impairment, who have been identified as needing rehabilitation officer input by a social worker, will take part. Those with an urgent need for a Visual Rehabilitation Officer or who have a cognitive impairment will be excluded. The primary outcome measure will be self-reported visual function (48-item Veterans Affairs Low Vision Visual Functioning Questionnaire). Secondary outcome measures will include psychological and quality-of-life metrics: the Patient Health Questionnaire (PHQ-9), the Warwick-Edinburgh Mental Well-being Scale (WEMWBS), the Adjustment to Age-related Visual Loss Scale (AVL-12), the Standardised Health-related Quality of Life Questionnaire (EQ-5D) and the UCLA Loneliness Scale. The interviewer collecting the outcomes will be masked to the group allocations. The analysis will be undertaken on a complete case and intention-to-treat basis. Analysis of covariance (ANCOVA) will be applied to follow-up questionnaire scores, with the baseline score as a covariate. This trial is expected to provide robust effect size estimates of the intervention effect. The data will be used to design a large-scale randomised controlled trial to evaluate fully the Visual Rehabilitation Officer intervention. A rigorous evaluation of Rehabilitation Officer input is vital to direct a future low vision rehabilitation strategy and to help direct government resources. The trial was registered with ( ISRCTN44807874 ) on 9 March 2015.
Gallagher, Rosemary; Damodaran, Harish; Werner, William G; Powell, Wendy; Deutsch, Judith E
2016-08-19
Evidence based virtual environments (VEs) that incorporate compensatory strategies such as cueing may change motor behavior and increase exercise intensity while also being engaging and motivating. The purpose of this study was to determine if persons with Parkinson's disease and aged matched healthy adults responded to auditory and visual cueing embedded in a bicycling VE as a method to increase exercise intensity. We tested two groups of participants, persons with Parkinson's disease (PD) (n = 15) and age-matched healthy adults (n = 13) as they cycled on a stationary bicycle while interacting with a VE. Participants cycled under two conditions: auditory cueing (provided by a metronome) and visual cueing (represented as central road markers in the VE). The auditory condition had four trials in which auditory cues or the VE were presented alone or in combination. The visual condition had five trials in which the VE and visual cue rate presentation was manipulated. Data were analyzed by condition using factorial RMANOVAs with planned t-tests corrected for multiple comparisons. There were no differences in pedaling rates between groups for both the auditory and visual cueing conditions. Persons with PD increased their pedaling rate in the auditory (F 4.78, p = 0.029) and visual cueing (F 26.48, p < 0.000) conditions. Age-matched healthy adults also increased their pedaling rate in the auditory (F = 24.72, p < 0.000) and visual cueing (F = 40.69, p < 0.000) conditions. Trial-to-trial comparisons in the visual condition in age-matched healthy adults showed a step-wise increase in pedaling rate (p = 0.003 to p < 0.000). In contrast, persons with PD increased their pedaling rate only when explicitly instructed to attend to the visual cues (p < 0.000). An evidenced based cycling VE can modify pedaling rate in persons with PD and age-matched healthy adults. Persons with PD required attention directed to the visual cues in order to obtain an increase in cycling intensity. The combination of the VE and auditory cues was neither additive nor interfering. These data serve as preliminary evidence that embedding auditory and visual cues to alter cycling speed in a VE as method to increase exercise intensity that may promote fitness.
Russell, Cristel Antonia; Swasy, John L.; Russell, Dale Wesley; Engel, Larry
2017-01-01
Risk warning or disclosure information in advertising is only effective in correcting consumers’ judgments if enough cognitive capacity is available to process that information. Hence, comprehension of verbal warnings in TV commercials may suffer if accompanied by positive visual elements. This research addresses this concern about cross-modality interference in the context of direct-to-consumer (DTC) pharmaceutical commercials in the United States by experimentally testing whether positive facial expressions reduce consumers’ understanding of the mandated health warning. A content analysis of a sample of DTC commercials reveals that positive facial expressions are more prevalent during the verbal warning act of the commercials than during the other acts. An eye-tracking experiment conducted with specially produced DTC commercials, which vary the valence of characters’ facial expressions during the health warning, provides evidence that happy faces reduce objective comprehension of the warning. PMID:29269979
Unaware Processing of Tools in the Neural System for Object-Directed Action Representation.
Tettamanti, Marco; Conca, Francesca; Falini, Andrea; Perani, Daniela
2017-11-01
The hypothesis that the brain constitutively encodes observed manipulable objects for the actions they afford is still debated. Yet, crucial evidence demonstrating that, even in the absence of perceptual awareness, the mere visual appearance of a manipulable object triggers a visuomotor coding in the action representation system including the premotor cortex, has hitherto not been provided. In this fMRI study, we instantiated reliable unaware visual perception conditions by means of continuous flash suppression, and we tested in 24 healthy human participants (13 females) whether the visuomotor object-directed action representation system that includes left-hemispheric premotor, parietal, and posterior temporal cortices is activated even under subliminal perceptual conditions. We found consistent activation in the target visuomotor cortices, both with and without perceptual awareness, specifically for pictures of manipulable versus non-manipulable objects. By means of a multivariate searchlight analysis, we also found that the brain activation patterns in this visuomotor network enabled the decoding of manipulable versus non-manipulable object picture processing, both with and without awareness. These findings demonstrate the intimate neural coupling between visual perception and motor representation that underlies manipulable object processing: manipulable object stimuli specifically engage the visuomotor object-directed action representation system, in a constitutive manner that is independent from perceptual awareness. This perceptuo-motor coupling endows the brain with an efficient mechanism for monitoring and planning reactions to external stimuli in the absence of awareness. SIGNIFICANCE STATEMENT Our brain constantly encodes the visual information that hits the retina, leading to a stimulus-specific activation of sensory and semantic representations, even for objects that we do not consciously perceive. Do these unconscious representations encompass the motor programming of actions that could be accomplished congruently with the objects' functions? In this fMRI study, we instantiated unaware visual perception conditions, by dynamically suppressing the visibility of manipulable object pictures with mondrian masks. Despite escaping conscious perception, manipulable objects activated an object-directed action representation system that includes left-hemispheric premotor, parietal, and posterior temporal cortices. This demonstrates that visuomotor encoding occurs independently of conscious object perception. Copyright © 2017 the authors 0270-6474/17/3710712-13$15.00/0.
Representational Account of Memory: Insights from Aging and Synesthesia.
Pfeifer, Gaby; Ward, Jamie; Chan, Dennis; Sigala, Natasha
2016-12-01
The representational account of memory envisages perception and memory to be on a continuum rather than in discretely divided brain systems [Bussey, T. J., & Saksida, L. M. Memory, perception, and the ventral visual-perirhinal-hippocampal stream: Thinking outside of the boxes. Hippocampus, 17, 898-908, 2007]. We tested this account using a novel between-group design with young grapheme-color synesthetes, older adults, and young controls. We investigated how the disparate sensory-perceptual abilities between these groups translated into associative memory performance for visual stimuli that do not induce synesthesia. ROI analyses of the entire ventral visual stream showed that associative retrieval (a pair-associate retrieved in the absence of a visual stimulus) yielded enhanced activity in young and older adults' visual regions relative to synesthetes, whereas associative recognition (deciding whether a visual stimulus was the correct pair-associate) was characterized by enhanced activity in synesthetes' visual regions relative to older adults. Whole-brain analyses at associative retrieval revealed an effect of age in early visual cortex, with older adults showing enhanced activity relative to synesthetes and young adults. At associative recognition, the group effect was reversed: Synesthetes showed significantly enhanced activity relative to young and older adults in early visual regions. The inverted group effects observed between retrieval and recognition indicate that reduced sensitivity in visual cortex (as in aging) comes with increased activity during top-down retrieval and decreased activity during bottom-up recognition, whereas enhanced sensitivity (as in synesthesia) shows the opposite pattern. Our results provide novel evidence for the direct contribution of perceptual mechanisms to visual associative memory based on the examples of synesthesia and aging.
Strauss, Soeren; Woodgate, Philip J.W.; Sami, Saber A.; Heinke, Dietmar
2015-01-01
We present an extension of a neurobiologically inspired robotics model, termed CoRLEGO (Choice reaching with a LEGO arm robot). CoRLEGO models experimental evidence from choice reaching tasks (CRT). In a CRT participants are asked to rapidly reach and touch an item presented on the screen. These experiments show that non-target items can divert the reaching movement away from the ideal trajectory to the target item. This is seen as evidence attentional selection of reaching targets can leak into the motor system. Using competitive target selection and topological representations of motor parameters (dynamic neural fields) CoRLEGO is able to mimic this leakage effect. Furthermore if the reaching target is determined by its colour oddity (i.e. a green square among red squares or vice versa), the reaching trajectories become straighter with repetitions of the target colour (colour streaks). This colour priming effect can also be modelled with CoRLEGO. The paper also presents an extension of CoRLEGO. This extension mimics findings that transcranial direct current stimulation (tDCS) over the motor cortex modulates the colour priming effect (Woodgate et al., 2015). The results with the new CoRLEGO suggest that feedback connections from the motor system to the brain’s attentional system (parietal cortex) guide visual attention to extract movement-relevant information (i.e. colour) from visual stimuli. This paper adds to growing evidence that there is a close interaction between the motor system and the attention system. This evidence contradicts the traditional conceptualization of the motor system as the endpoint of a serial chain of processing stages. At the end of the paper we discuss CoRLEGO’s predictions and also lessons for neurobiologically inspired robotics emerging from this work. PMID:26667353
Therapeutic potential of intravitreal pharmacotherapy in retinal vein occlusion
Shahsuvaryan, Marianne L.
2012-01-01
Retinal vein occlusion (RVO) is the most common visually disabling disease affecting the retina after diabetic retinopathy. Although the disease entity has long been known, its management is still controversial. Macular edema is the main reason for decreased visual acuity (VA) in this retinal vascular disorder. Recently the vitreous cavity has increasingly been used as a reservoir of drugs for the direct treatment of macular edema through intravitreal injection route. The most widely injected drugs so far have been triamcinolone acetonide (TA) and bevacizumab. The objective of this review is to evaluate the evidence and discuss the rationale behind the recent suggestions that intravitreal pharmacotherapy by corticosteroids and anti-vascular endothelial growth factors may be useful in the treatment of retinal vein occlusion. PMID:23275914
Sex pheromone receptor proteins. Visualization using a radiolabeled photoaffinity analog
DOE Office of Scientific and Technical Information (OSTI.GOV)
Vogt, R.G.; Prestwich, G.D.; Riddiford, L.M.
1988-03-15
A tritium-labeled photoaffinity analog of a moth pheromone was used to covalently modify pheromone-selective binding proteins in the antennal sensillum lymph and sensory dendritic membranes of the male silk moth, Antheraea polyphemus. This analog, (E,Z)-6,11-(/sup 3/H)hexadecadienyl diazoacetate, allowed visualization of a 15-kilodalton soluble protein and a 69-kilodalton membrane protein in fluorescence autoradiograms of electrophoretically separated antennal proteins. Covalent modification of these proteins was specifically reduced when incubation and UV irradiation were conducted in the presence of excess unlabeled pheromone, (E,Z)-6,11-hexadecadienyl acetate. These experiments constitute the first direct evidence for a membrane protein of a chemosensory neuron interacting in a specificmore » fashion with a biologically relevant odorant.« less
Visual cues that are effective for contextual saccade adaptation.
Azadi, Reza; Harwood, Mark R
2014-06-01
The accuracy of saccades, as maintained by saccade adaptation, has been shown to be context dependent: able to have different amplitude movements to the same retinal displacement dependent on motor contexts such as orbital starting location. There is conflicting evidence as to whether purely visual cues also effect contextual saccade adaptation and, if so, what function this might serve. We tested what visual cues might evoke contextual adaptation. Over 5 experiments, 78 naive subjects made saccades to circularly moving targets, which stepped outward or inward during the saccade depending on target movement direction, speed, or color and shape. To test if the movement or context postsaccade were critical, we stopped the postsaccade target motion (experiment 4) or neutralized the contexts by equating postsaccade target speed to an intermediate value (experiment 5). We found contextual adaptation in all conditions except those defined by color and shape. We conclude that some, but not all, visual cues before the saccade are sufficient for contextual adaptation. We conjecture that this visual contextuality functions to allow for different motor states for different coordinated movement patterns, such as coordinated saccade and pursuit motor planning. Copyright © 2014 the American Physiological Society.
The dorsal raphe modulates sensory responsiveness during arousal in zebrafish
Yokogawa, Tohei; Hannan, Markus C.; Burgess, Harold A.
2012-01-01
During waking behavior animals adapt their state of arousal in response to environmental pressures. Sensory processing is regulated in aroused states and several lines of evidence imply that this is mediated at least partly by the serotonergic system. However there is little information directly showing that serotonergic function is required for state-dependent modulation of sensory processing. Here we find that zebrafish larvae can maintain a short-term state of arousal during which neurons in the dorsal raphe modulate sensory responsiveness to behaviorally relevant visual cues. Following a brief exposure to water flow, larvae show elevated activity and heightened sensitivity to perceived motion. Calcium imaging of neuronal activity after flow revealed increased activity in serotonergic neurons of the dorsal raphe. Genetic ablation of these neurons abolished the increase in visual sensitivity during arousal without affecting baseline visual function or locomotor activity. We traced projections from the dorsal raphe to a major visual area, the optic tectum. Laser ablation of the tectum demonstrated that this structure, like the dorsal raphe, is required for improved visual sensitivity during arousal. These findings reveal that serotonergic neurons of the dorsal raphe have a state-dependent role in matching sensory responsiveness to behavioral context. PMID:23100441
Can walking motions improve visually induced rotational self-motion illusions in virtual reality?
Riecke, Bernhard E; Freiberg, Jacob B; Grechkin, Timofey Y
2015-02-04
Illusions of self-motion (vection) can provide compelling sensations of moving through virtual environments without the need for complex motion simulators or large tracked physical walking spaces. Here we explore the interaction between biomechanical cues (stepping along a rotating circular treadmill) and visual cues (viewing simulated self-rotation) for providing stationary users a compelling sensation of rotational self-motion (circular vection). When tested individually, biomechanical and visual cues were similarly effective in eliciting self-motion illusions. However, in combination they yielded significantly more intense self-motion illusions. These findings provide the first compelling evidence that walking motions can be used to significantly enhance visually induced rotational self-motion perception in virtual environments (and vice versa) without having to provide for physical self-motion or motion platforms. This is noteworthy, as linear treadmills have been found to actually impair visually induced translational self-motion perception (Ash, Palmisano, Apthorp, & Allison, 2013). Given the predominant focus on linear walking interfaces for virtual-reality locomotion, our findings suggest that investigating circular and curvilinear walking interfaces offers a promising direction for future research and development and can help to enhance self-motion illusions, presence and immersion in virtual-reality systems. © 2015 ARVO.
Octopus vulgaris uses visual information to determine the location of its arm.
Gutnick, Tamar; Byrne, Ruth A; Hochner, Binyamin; Kuba, Michael
2011-03-22
Octopuses are intelligent, soft-bodied animals with keen senses that perform reliably in a variety of visual and tactile learning tasks. However, researchers have found them disappointing in that they consistently fail in operant tasks that require them to combine central nervous system reward information with visual and peripheral knowledge of the location of their arms. Wells claimed that in order to filter and integrate an abundance of multisensory inputs that might inform the animal of the position of a single arm, octopuses would need an exceptional computing mechanism, and "There is no evidence that such a system exists in Octopus, or in any other soft bodied animal." Recent electrophysiological experiments, which found no clear somatotopic organization in the higher motor centers, support this claim. We developed a three-choice maze that required an octopus to use a single arm to reach a visually marked goal compartment. Using this operant task, we show for the first time that Octopus vulgaris is capable of guiding a single arm in a complex movement to a location. Thus, we claim that octopuses can combine peripheral arm location information with visual input to control goal-directed complex movements. Copyright © 2011 Elsevier Ltd. All rights reserved.
Fukushima, Kikuro; Fukushima, Junko; Warabi, Tateo; Barnes, Graham R.
2013-01-01
Smooth-pursuit eye movements allow primates to track moving objects. Efficient pursuit requires appropriate target selection and predictive compensation for inherent processing delays. Prediction depends on expectation of future object motion, storage of motion information and use of extra-retinal mechanisms in addition to visual feedback. We present behavioral evidence of how cognitive processes are involved in predictive pursuit in normal humans and then describe neuronal responses in monkeys and behavioral responses in patients using a new technique to test these cognitive controls. The new technique examines the neural substrate of working memory and movement preparation for predictive pursuit by using a memory-based task in macaque monkeys trained to pursue (go) or not pursue (no-go) according to a go/no-go cue, in a direction based on memory of a previously presented visual motion display. Single-unit task-related neuronal activity was examined in medial superior temporal cortex (MST), supplementary eye fields (SEF), caudal frontal eye fields (FEF), cerebellar dorsal vermis lobules VI–VII, caudal fastigial nuclei (cFN), and floccular region. Neuronal activity reflecting working memory of visual motion direction and go/no-go selection was found predominantly in SEF, cerebellar dorsal vermis and cFN, whereas movement preparation related signals were found predominantly in caudal FEF and the same cerebellar areas. Chemical inactivation produced effects consistent with differences in signals represented in each area. When applied to patients with Parkinson's disease (PD), the task revealed deficits in movement preparation but not working memory. In contrast, patients with frontal cortical or cerebellar dysfunction had high error rates, suggesting impaired working memory. We show how neuronal activity may be explained by models of retinal and extra-retinal interaction in target selection and predictive control and thus aid understanding of underlying pathophysiology. PMID:23515488
Schmidt, Joseph; MacNamara, Annmarie; Proudfit, Greg Hajcak; Zelinsky, Gregory J.
2014-01-01
The visual-search literature has assumed that the top-down target representation used to guide search resides in visual working memory (VWM). We directly tested this assumption using contralateral delay activity (CDA) to estimate the VWM load imposed by the target representation. In Experiment 1, observers previewed four photorealistic objects and were cued to remember the two objects appearing to the left or right of central fixation; Experiment 2 was identical except that observers previewed two photorealistic objects and were cued to remember one. CDA was measured during a delay following preview offset but before onset of a four-object search array. One of the targets was always present, and observers were asked to make an eye movement to it and press a button. We found lower magnitude CDA on trials when the initial search saccade was directed to the target (strong guidance) compared to when it was not (weak guidance). This difference also tended to be larger shortly before search-display onset and was largely unaffected by VWM item-capacity limits or number of previews. Moreover, the difference between mean strong- and weak-guidance CDA was proportional to the increase in search time between mean strong-and weak-guidance trials (as measured by time-to-target and reaction-time difference scores). Contrary to most search models, our data suggest that trials resulting in the maintenance of more target features results in poor search guidance to a target. We interpret these counterintuitive findings as evidence for strong search guidance using a small set of highly discriminative target features that remain after pruning from a larger set of features, with the load imposed on VWM varying with this feature-consolidation process. PMID:24599946
Schmidt, Joseph; MacNamara, Annmarie; Proudfit, Greg Hajcak; Zelinsky, Gregory J
2014-03-05
The visual-search literature has assumed that the top-down target representation used to guide search resides in visual working memory (VWM). We directly tested this assumption using contralateral delay activity (CDA) to estimate the VWM load imposed by the target representation. In Experiment 1, observers previewed four photorealistic objects and were cued to remember the two objects appearing to the left or right of central fixation; Experiment 2 was identical except that observers previewed two photorealistic objects and were cued to remember one. CDA was measured during a delay following preview offset but before onset of a four-object search array. One of the targets was always present, and observers were asked to make an eye movement to it and press a button. We found lower magnitude CDA on trials when the initial search saccade was directed to the target (strong guidance) compared to when it was not (weak guidance). This difference also tended to be larger shortly before search-display onset and was largely unaffected by VWM item-capacity limits or number of previews. Moreover, the difference between mean strong- and weak-guidance CDA was proportional to the increase in search time between mean strong-and weak-guidance trials (as measured by time-to-target and reaction-time difference scores). Contrary to most search models, our data suggest that trials resulting in the maintenance of more target features results in poor search guidance to a target. We interpret these counterintuitive findings as evidence for strong search guidance using a small set of highly discriminative target features that remain after pruning from a larger set of features, with the load imposed on VWM varying with this feature-consolidation process.
Hogendoorn, Hinze; Burkitt, Anthony N
2018-05-01
Due to the delays inherent in neuronal transmission, our awareness of sensory events necessarily lags behind the occurrence of those events in the world. If the visual system did not compensate for these delays, we would consistently mislocalize moving objects behind their actual position. Anticipatory mechanisms that might compensate for these delays have been reported in animals, and such mechanisms have also been hypothesized to underlie perceptual effects in humans such as the Flash-Lag Effect. However, to date no direct physiological evidence for anticipatory mechanisms has been found in humans. Here, we apply multivariate pattern classification to time-resolved EEG data to investigate anticipatory coding of object position in humans. By comparing the time-course of neural position representation for objects in both random and predictable apparent motion, we isolated anticipatory mechanisms that could compensate for neural delays when motion trajectories were predictable. As well as revealing an early neural position representation (lag 80-90 ms) that was unaffected by the predictability of the object's trajectory, we demonstrate a second neural position representation at 140-150 ms that was distinct from the first, and that was pre-activated ahead of the moving object when it moved on a predictable trajectory. The latency advantage for predictable motion was approximately 16 ± 2 ms. To our knowledge, this provides the first direct experimental neurophysiological evidence of anticipatory coding in human vision, revealing the time-course of predictive mechanisms without using a spatial proxy for time. The results are numerically consistent with earlier animal work, and suggest that current models of spatial predictive coding in visual cortex can be effectively extended into the temporal domain. Copyright © 2018 Elsevier Inc. All rights reserved.
Methylphenidate does not enhance visual working memory but benefits motivation in macaque monkeys.
Oemisch, Mariann; Johnston, Kevin; Paré, Martin
2016-10-01
Working memory is a limited-capacity cognitive process that retains relevant information temporarily to guide thoughts and behavior. A large body of work has suggested that catecholamines exert a major modulatory influence on cognition, but there is only equivocal evidence of a direct influence on working memory ability, which would be reflected in a dependence on working memory load. Here we tested the contribution of catecholamines to working memory by administering a wide range of acute oral doses of the dopamine and norepinephrine reuptake inhibitor methylphenidate (MPH, 0.1-9 mg/kg) to three female macaque monkeys (Macaca mulatta), whose working memory ability was measured from their performance in a visual sequential comparison task. This task allows the systematic manipulation of working memory load, and we therefore tested the specific hypothesis that MPH modulates performance in a manner that depends on both dose and memory load. We found no evidence of a dose- or memory load-dependent effect of MPH on performance. In contrast, significant effects on measures of motivation were observed. These findings suggest that an acute increase in catecholamines does not seem to affect the retention of visual information per se. As such, these results help delimit the effects of MPH on cognition. Copyright © 2016 Elsevier Ltd. All rights reserved.
Howe, Tsu-Hsin; Chen, Hao-Ling; Lee, Candy Chieh; Chen, Ying-Dar; Wang, Tien-Ni
2017-10-01
Visual perceptual motor skills have been proposed as underlying courses of handwriting difficulties. However, there is no evaluation tool currently available to assess these skills comprehensively and to serve as a sensitive measure. The purpose of this study was to validate the Computerized Perceptual Motor Skills Assessment (CPMSA), a newly developed evaluation tool for children in early elementary grades. Its test-retest reliability, concurrent validity, discriminant validity, and responsiveness were examined in 43 typically developing children and 26 children with handwriting difficulty. The CPMSA demonstrated excellent reliability across all subtests with intra-class correlation coefficients (ICCs)≥0.80. Significant moderate correlations between the domains of the CPMSA and corresponding gold standards including Beery VMI, the TVPS-3, and the eye-hand coordination subtest of the DTVP-2 demonstrated good concurrent validity. In addition, the CPMSA showed evidence of discriminant validity in samples of children with and without handwriting difficulty. This article provides evidence in support of the CPMSA. The CPMSA is a reliable, valid, and promising measure of visual perceptual motor skills for children in early elementary grades. Directions for future study and improvements to the assessment are discussed. Copyright © 2017. Published by Elsevier Ltd.
Contextual effects on smooth-pursuit eye movements.
Spering, Miriam; Gegenfurtner, Karl R
2007-02-01
Segregating a moving object from its visual context is particularly relevant for the control of smooth-pursuit eye movements. We examined the interaction between a moving object and a stationary or moving visual context to determine the role of the context motion signal in driving pursuit. Eye movements were recorded from human observers to a medium-contrast Gaussian dot that moved horizontally at constant velocity. A peripheral context consisted of two vertically oriented sinusoidal gratings, one above and one below the stimulus trajectory, that were either stationary or drifted into the same or opposite direction as that of the target at different velocities. We found that a stationary context impaired pursuit acceleration and velocity and prolonged pursuit latency. A drifting context enhanced pursuit performance, irrespective of its motion direction. This effect was modulated by context contrast and orientation. When a context was briefly perturbed to move faster or slower eye velocity changed accordingly, but only when the context was drifting along with the target. Perturbing a context into the direction orthogonal to target motion evoked a deviation of the eye opposite to the perturbation direction. We therefore provide evidence for the use of absolute and relative motion cues, or motion assimilation and motion contrast, for the control of smooth-pursuit eye movements.
Gender difference in the theta/alpha ratio during the induction of peaceful audiovisual modalities.
Yang, Chia-Yen; Lin, Ching-Po
2015-09-01
Gender differences in emotional perception have been found in numerous psychological and psychophysiological studies. The conducting modalities in diverse characteristics of different sensory systems make it interesting to determine how cooperation and competition contribute to emotional experiences. We have previously estimated the bias from the match attributes of auditory and visual modalities and revealed specific brain activity frequency patterns related to a peaceful mood. In that multimodality experiment, we focused on how inner-quiet information is processed in the human brain, and found evidence of auditory domination from the theta-band activity. However, a simple quantitative description of these three frequency bands is lacking, and no studies have assessed the effects of peacefulness on the emotional state. Therefore, the aim of this study was to use magnetoencephalography to determine if gender differences exist (and when and where) in the frequency interactions underpinning the perception of peacefulness. This study provides evidence of auditory and visual domination in perceptual bias during multimodality processing of peaceful consciousness. The results of power ratio analyses suggest that the values of the theta/alpha ratio are associated with a modality as well as hemispheric asymmetries in the anterior-to-posterior direction, which shift from right to left with the auditory to visual stimulations in a peaceful mood. This means that the theta/alpha ratio might be useful for evaluating emotion. Moreover, the difference was found to be most pronounced for auditory domination and visual sensitivity in the female group.
Evidence for discrete landmark use by pigeons during homing.
Mora, Cordula V; Ross, Jeremy D; Gorsevski, Peter V; Chowdhury, Budhaditya; Bingman, Verner P
2012-10-01
Considerable efforts have been made to investigate how homing pigeons (Columba livia f. domestica) are able to return to their loft from distant, unfamiliar sites while the mechanisms underlying navigation in familiar territory have received less attention. With the recent advent of global positioning system (GPS) data loggers small enough to be carried by pigeons, the role of visual environmental features in guiding navigation over familiar areas is beginning to be understood, yet, surprisingly, we still know very little about whether homing pigeons can rely on discrete, visual landmarks to guide navigation. To assess a possible role of discrete, visual landmarks in navigation, homing pigeons were first trained to home from a site with four wind turbines as salient landmarks as well as from a control site without any distinctive, discrete landmark features. The GPS-recorded flight paths of the pigeons on the last training release were straighter and more similar among birds from the turbine site compared with those from the control site. The pigeons were then released from both sites following a clock-shift manipulation. Vanishing bearings from the turbine site continued to be homeward oriented as 13 of 14 pigeons returned home. By contrast, at the control site the vanishing bearings were deflected in the expected clock-shift direction and only 5 of 13 pigeons returned home. Taken together, our results offer the first strong evidence that discrete, visual landmarks are one source of spatial information homing pigeons can utilize to navigate when flying over a familiar area.
Artful terms: A study on aesthetic word usage for visual art versus film and music.
Augustin, M Dorothee; Carbon, Claus-Christian; Wagemans, Johan
2012-01-01
Despite the importance of the arts in human life, psychologists still know relatively little about what characterises their experience for the recipient. The current research approaches this problem by studying people's word usage in aesthetics, with a focus on three important art forms: visual art, film, and music. The starting point was a list of 77 words known to be useful to describe aesthetic impressions of visual art (Augustin et al 2012, Acta Psychologica139 187-201). Focusing on ratings of likelihood of use, we examined to what extent word usage in aesthetic descriptions of visual art can be generalised to film and music. The results support the claim of an interplay of generality and specificity in aesthetic word usage. Terms with equal likelihood of use for all art forms included beautiful, wonderful, and terms denoting originality. Importantly, emotion-related words received higher ratings for film and music than for visual art. To our knowledge this is direct evidence that aesthetic experiences of visual art may be less affectively loaded than, for example, experiences of music. The results render important information about aesthetic word usage in the realm of the arts and may serve as a starting point to develop tailored measurement instruments for different art forms.
Artful terms: A study on aesthetic word usage for visual art versus film and music
Augustin, M Dorothee; Carbon, Claus-Christian; Wagemans, Johan
2012-01-01
Despite the importance of the arts in human life, psychologists still know relatively little about what characterises their experience for the recipient. The current research approaches this problem by studying people's word usage in aesthetics, with a focus on three important art forms: visual art, film, and music. The starting point was a list of 77 words known to be useful to describe aesthetic impressions of visual art (Augustin et al 2012, Acta Psychologica 139 187–201). Focusing on ratings of likelihood of use, we examined to what extent word usage in aesthetic descriptions of visual art can be generalised to film and music. The results support the claim of an interplay of generality and specificity in aesthetic word usage. Terms with equal likelihood of use for all art forms included beautiful, wonderful, and terms denoting originality. Importantly, emotion-related words received higher ratings for film and music than for visual art. To our knowledge this is direct evidence that aesthetic experiences of visual art may be less affectively loaded than, for example, experiences of music. The results render important information about aesthetic word usage in the realm of the arts and may serve as a starting point to develop tailored measurement instruments for different art forms. PMID:23145287
Multisensory integration across the senses in young and old adults
Mahoney, Jeannette R.; Li, Po Ching Clara; Oh-Park, Mooyeon; Verghese, Joe; Holtzer, Roee
2011-01-01
Stimuli are processed concurrently and across multiple sensory inputs. Here we directly compared the effect of multisensory integration (MSI) on reaction time across three paired sensory inputs in eighteen young (M=19.17 yrs) and eighteen old (M=76.44 yrs) individuals. Participants were determined to be non-demented and without any medical or psychiatric conditions that would affect their performance. Participants responded to randomly presented unisensory (auditory, visual, somatosensory) stimuli and three paired sensory inputs consisting of auditory-somatosensory (AS) auditory-visual (AV) and visual-somatosensory (VS) stimuli. Results revealed that reaction time (RT) to all multisensory pairings was significantly faster than those elicited to the constituent unisensory conditions across age groups; findings that could not be accounted for by simple probability summation. Both young and old participants responded the fastest to multisensory pairings containing somatosensory input. Compared to younger adults, older adults demonstrated a significantly greater RT benefit when processing concurrent VS information. In terms of co-activation, older adults demonstrated a significant increase in the magnitude of visual-somatosensory co-activation (i.e., multisensory integration), while younger adults demonstrated a significant increase in the magnitude of auditory-visual and auditory-somatosensory co-activation. This study provides first evidence in support of the facilitative effect of pairing somatosensory with visual stimuli in older adults. PMID:22024545
A Rapid Subcortical Amygdala Route for Faces Irrespective of Spatial Frequency and Emotion.
McFadyen, Jessica; Mermillod, Martial; Mattingley, Jason B; Halász, Veronika; Garrido, Marta I
2017-04-05
There is significant controversy over the existence and function of a direct subcortical visual pathway to the amygdala. It is thought that this pathway rapidly transmits low spatial frequency information to the amygdala independently of the cortex, and yet the directionality of this function has never been determined. We used magnetoencephalography to measure neural activity while human participants discriminated the gender of neutral and fearful faces filtered for low or high spatial frequencies. We applied dynamic causal modeling to demonstrate that the most likely underlying neural network consisted of a pulvinar-amygdala connection that was uninfluenced by spatial frequency or emotion, and a cortical-amygdala connection that conveyed high spatial frequencies. Crucially, data-driven neural simulations revealed a clear temporal advantage of the subcortical connection over the cortical connection in influencing amygdala activity. Thus, our findings support the existence of a rapid subcortical pathway that is nonselective in terms of the spatial frequency or emotional content of faces. We propose that that the "coarseness" of the subcortical route may be better reframed as "generalized." SIGNIFICANCE STATEMENT The human amygdala coordinates how we respond to biologically relevant stimuli, such as threat or reward. It has been postulated that the amygdala first receives visual input via a rapid subcortical route that conveys "coarse" information, namely, low spatial frequencies. For the first time, the present paper provides direction-specific evidence from computational modeling that the subcortical route plays a generalized role in visual processing by rapidly transmitting raw, unfiltered information directly to the amygdala. This calls into question a widely held assumption across human and animal research that fear responses are produced faster by low spatial frequencies. Our proposed mechanism suggests organisms quickly generate fear responses to a wide range of visual properties, heavily implicating future research on anxiety-prevention strategies. Copyright © 2017 the authors 0270-6474/17/373864-11$15.00/0.
The Premotor theory of attention: time to move on?
Smith, Daniel T; Schenk, Thomas
2012-05-01
Spatial attention and eye-movements are tightly coupled, but the precise nature of this coupling is controversial. The influential but controversial Premotor theory of attention makes four specific predictions about the relationship between motor preparation and spatial attention. Firstly, spatial attention and motor preparation use the same neural substrates. Secondly, spatial attention is functionally equivalent to planning goal directed actions such as eye-movements (i.e. planning an action is both necessary and sufficient for a shift of spatial attention). Thirdly, planning a goal directed action with any effector system is sufficient to trigger a shift of spatial attention. Fourthly, the eye-movement system has a privileged role in orienting visual spatial attention. This article reviews empirical studies that have tested these predictions. Contrary to predictions one and two there is evidence of anatomical and functional dissociations between endogenous spatial attention and motor preparation. However, there is compelling evidence that exogenous attention is reliant on activation of the oculomotor system. With respect to the third prediction, there is correlational evidence that spatial attention is directed to the endpoint of goal-directed actions but no direct evidence that this attention shift is dependent on motor preparation. The few studies to have directly tested the fourth prediction have produced conflicting results, so the extent to which the oculomotor system has a privileged role in spatial attention remains unclear. Overall, the evidence is not consistent with the view that spatial attention is functionally equivalent to motor preparation so the Premotor theory should be rejected, although a limited version of the Premotor theory in which only exogenous attention is dependent on motor preparation may still be tenable. A plausible alternative account is that activity in the motor system contributes to biased competition between different sensory representations with the winner of the competition becoming the attended item. Copyright © 2012 Elsevier Ltd. All rights reserved.
Neural Architecture for Feature Binding in Visual Working Memory.
Schneegans, Sebastian; Bays, Paul M
2017-04-05
Binding refers to the operation that groups different features together into objects. We propose a neural architecture for feature binding in visual working memory that employs populations of neurons with conjunction responses. We tested this model using cued recall tasks, in which subjects had to memorize object arrays composed of simple visual features (color, orientation, and location). After a brief delay, one feature of one item was given as a cue, and the observer had to report, on a continuous scale, one or two other features of the cued item. Binding failure in this task is associated with swap errors, in which observers report an item other than the one indicated by the cue. We observed that the probability of swapping two items strongly correlated with the items' similarity in the cue feature dimension, and found a strong correlation between swap errors occurring in spatial and nonspatial report. The neural model explains both swap errors and response variability as results of decoding noisy neural activity, and can account for the behavioral results in quantitative detail. We then used the model to compare alternative mechanisms for binding nonspatial features. We found the behavioral results fully consistent with a model in which nonspatial features are bound exclusively via their shared location, with no indication of direct binding between color and orientation. These results provide evidence for a special role of location in feature binding, and the model explains how this special role could be realized in the neural system. SIGNIFICANCE STATEMENT The problem of feature binding is of central importance in understanding the mechanisms of working memory. How do we remember not only that we saw a red and a round object, but that these features belong together to a single object rather than to different objects in our environment? Here we present evidence for a neural mechanism for feature binding in working memory, based on encoding of visual information by neurons that respond to the conjunction of features. We find clear evidence that nonspatial features are bound via space: we memorize directly where a color or an orientation appeared, but we memorize which color belonged with which orientation only indirectly by virtue of their shared location. Copyright © 2017 Schneegans and Bays.
Characteristics of visual fatigue under the effect of 3D animation.
Chang, Yu-Shuo; Hsueh, Ya-Hsin; Tung, Kwong-Chung; Jhou, Fong-Yi; Lin, David Pei-Cheng
2015-01-01
Visual fatigue is commonly encountered in modern life. Clinical visual fatigue characteristics caused by 2-D and 3-D animations may be different, but have not been characterized in detail. This study tried to distinguish the differential effects on visual fatigue caused by 2-D and 3-D animations. A total of 23 volunteers were subjected to accommodation and vergence assessments, followed by a 40-min video game program designed to aggravate their asthenopic symptoms. The volunteers were then assessed for accommodation and vergence parameters again and directed to watch a 5-min 3-D video program, and then assessed again for the parameters. The results support that the 3-D animations caused similar characteristics in vision fatigue parameters in some specific aspects as compared to that caused by 2-D animations. Furthermore, 3-D animations may lead to more exhaustion in both ciliary and extra-ocular muscles, and such differential effects were more evident in the high demand of near vision work. The current results indicated that an arbitrary set of indexes may be promoted in the design of 3-D display or equipments.
The impact of attentional, linguistic, and visual features during object naming
Clarke, Alasdair D. F.; Coco, Moreno I.; Keller, Frank
2013-01-01
Object detection and identification are fundamental to human vision, and there is mounting evidence that objects guide the allocation of visual attention. However, the role of objects in tasks involving multiple modalities is less clear. To address this question, we investigate object naming, a task in which participants have to verbally identify objects they see in photorealistic scenes. We report an eye-tracking study that investigates which features (attentional, visual, and linguistic) influence object naming. We find that the amount of visual attention directed toward an object, its position and saliency, along with linguistic factors such as word frequency, animacy, and semantic proximity, significantly influence whether the object will be named or not. We then ask how features from different modalities are combined during naming, and find significant interactions between saliency and position, saliency and linguistic features, and attention and position. We conclude that when the cognitive system performs tasks such as object naming, it uses input from one modality to constraint or enhance the processing of other modalities, rather than processing each input modality independently. PMID:24379792
Effects of Peripheral Visual Field Loss on Eye Movements During Visual Search
Wiecek, Emily; Pasquale, Louis R.; Fiser, Jozsef; Dakin, Steven; Bex, Peter J.
2012-01-01
Natural vision involves sequential eye movements that bring the fovea to locations selected by peripheral vision. How peripheral visual field loss (PVFL) affects this process is not well understood. We examine how the location and extent of PVFL affects eye movement behavior in a naturalistic visual search task. Ten patients with PVFL and 13 normally sighted subjects with full visual fields (FVF) completed 30 visual searches monocularly. Subjects located a 4° × 4° target, pseudo-randomly selected within a 26° × 11° natural image. Eye positions were recorded at 50 Hz. Search duration, fixation duration, saccade size, and number of saccades per trial were not significantly different between PVFL and FVF groups (p > 0.1). A χ2 test showed that the distributions of saccade directions for PVFL and FVL subjects were significantly different in 8 out of 10 cases (p < 0.01). Humphrey Visual Field pattern deviations for each subject were compared with the spatial distribution of eye movement directions. There were no significant correlations between saccade directional bias and visual field sensitivity across the 10 patients. Visual search performance was not significantly affected by PVFL. An analysis of eye movement directions revealed patients with PVFL show a biased directional distribution that was not directly related to the locus of vision loss, challenging feed-forward models of eye movement control. Consequently, many patients do not optimally compensate for visual field loss during visual search. PMID:23162511
Katzner, Steffen; Busse, Laura; Treue, Stefan
2009-01-01
Directing visual attention to spatial locations or to non-spatial stimulus features can strongly modulate responses of individual cortical sensory neurons. Effects of attention typically vary in magnitude, not only between visual cortical areas but also between individual neurons from the same area. Here, we investigate whether the size of attentional effects depends on the match between the tuning properties of the recorded neuron and the perceptual task at hand. We recorded extracellular responses from individual direction-selective neurons in the middle temporal area (MT) of rhesus monkeys trained to attend either to the color or the motion signal of a moving stimulus. We found that effects of spatial and feature-based attention in MT, which are typically observed in tasks allocating attention to motion, were very similar even when attention was directed to the color of the stimulus. We conclude that attentional modulation can occur in extrastriate cortex, even under conditions without a match between the tuning properties of the recorded neuron and the perceptual task at hand. Our data are consistent with theories of object-based attention describing a transfer of attention from relevant to irrelevant features, within the attended object and across the visual field. These results argue for a unified attentional system that modulates responses to a stimulus across cortical areas, even if a given area is specialized for processing task-irrelevant aspects of that stimulus.
Attentional selection of relative SF mediates global versus local processing: evidence from EEG.
Flevaris, Anastasia V; Bentin, Shlomo; Robertson, Lynn C
2011-06-13
Previous research on functional hemispheric differences in visual processing has associated global perception with low spatial frequency (LSF) processing biases of the right hemisphere (RH) and local perception with high spatial frequency (HSF) processing biases of the left hemisphere (LH). The Double Filtering by Frequency (DFF) theory expanded this hypothesis by proposing that visual attention selects and is directed to relatively LSFs by the RH and relatively HSFs by the LH, suggesting a direct causal relationship between SF selection and global versus local perception. We tested this idea in the current experiment by comparing activity in the EEG recorded at posterior right and posterior left hemisphere sites while participants' attention was directed to global or local levels of processing after selection of relatively LSFs versus HSFs in a previous stimulus. Hemispheric asymmetry in the alpha band (8-12 Hz) during preparation for global versus local processing was modulated by the selected SF. In contrast, preparatory activity associated with selection of SF was not modulated by the previously attended level (global/local). These results support the DFF theory that top-down attentional selection of SF mediates global and local processing.
Visual information for judging temporal range
NASA Technical Reports Server (NTRS)
Kaiser, Mary K.; Mowafy, Lyn
1993-01-01
Work in our laboratory suggests that pilots can extract temporal range information (i.e., the time to pass a given waypoint) directly from out-the-window motion information. This extraction does not require the use of velocity or distance, but rather operates solely on a 2-D motion cue. In this paper, we present the mathematical derivation of this information, psychophysical evidence of human observers' sensitivity, and possible advantages and limitations of basing vehicle control on this parameter.
Berger, Sue; Kaldenberg, Jennifer; Selmane, Romeissa; Carlo, Stephanie
2016-01-01
Visual and visual-perceptual impairments occur frequently with traumatic brain injury (TBI) and influence occupational performance. This systematic review examined the effectiveness of interventions within the scope of occupational therapy to improve occupational performance for adults with visual and visual-perceptual impairments as a result of TBI. Medline, PsycINFO, CINAHL, OTseeker, and the Cochrane Database of Systematic Reviews were searched, and 66 full text articles were reviewed. Sixteen articles were included in the review. Strong evidence supports the use of scanning, limited evidence supports the use of adaptive strategies, and mixed evidence supports the use of cognitive interventions to improve occupational performance for adults with TBI. Evidence related to vision therapy varies on the basis of the specific intervention implemented. Although the strength of the research varied, implications are discussed for practice, education, and research. Copyright © 2016 by the American Occupational Therapy Association, Inc.
Dasgupta, Aritra; Lee, Joon-Yong; Wilson, Ryan; Lafrance, Robert A; Cramer, Nick; Cook, Kristin; Payne, Samuel
2017-01-01
Combining interactive visualization with automated analytical methods like statistics and data mining facilitates data-driven discovery. These visual analytic methods are beginning to be instantiated within mixed-initiative systems, where humans and machines collaboratively influence evidence-gathering and decision-making. But an open research question is that, when domain experts analyze their data, can they completely trust the outputs and operations on the machine-side? Visualization potentially leads to a transparent analysis process, but do domain experts always trust what they see? To address these questions, we present results from the design and evaluation of a mixed-initiative, visual analytics system for biologists, focusing on analyzing the relationships between familiarity of an analysis medium and domain experts' trust. We propose a trust-augmented design of the visual analytics system, that explicitly takes into account domain-specific tasks, conventions, and preferences. For evaluating the system, we present the results of a controlled user study with 34 biologists where we compare the variation of the level of trust across conventional and visual analytic mediums and explore the influence of familiarity and task complexity on trust. We find that despite being unfamiliar with a visual analytic medium, scientists seem to have an average level of trust that is comparable with the same in conventional analysis medium. In fact, for complex sense-making tasks, we find that the visual analytic system is able to inspire greater trust than other mediums. We summarize the implications of our findings with directions for future research on trustworthiness of visual analytic systems.
Action video game training reduces the Simon Effect.
Hutchinson, Claire V; Barrett, Doug J K; Nitka, Aleksander; Raynes, Kerry
2016-04-01
A number of studies have shown that training on action video games improves various aspects of visual cognition including selective attention and inhibitory control. Here, we demonstrate that action video game play can also reduce the Simon Effect, and, hence, may have the potential to improve response selection during the planning and execution of goal-directed action. Non-game-players were randomly assigned to one of four groups; two trained on a first-person-shooter game (Call of Duty) on either Microsoft Xbox or Nintendo DS, one trained on a visual training game for Nintendo DS, and a control group who received no training. Response times were used to contrast performance before and after training on a behavioral assay designed to manipulate stimulus-response compatibility (the Simon Task). The results revealed significantly faster response times and a reduced cost of stimulus-response incompatibility in the groups trained on the first-person-shooter game. No benefit of training was observed in the control group or the group trained on the visual training game. These findings are consistent with previous evidence that action game play elicits plastic changes in the neural circuits that serve attentional control, and suggest training may facilitate goal-directed action by improving players' ability to resolve conflict during response selection and execution.
Spiegel, Daniel P.; Hansen, Bruce C.; Byblow, Winston D.; Thompson, Benjamin
2012-01-01
Transcranial direct current stimulation (tDCS) is a safe, non-invasive technique for transiently modulating the balance of excitation and inhibition within the human brain. It has been reported that anodal tDCS can reduce both GABA mediated inhibition and GABA concentration within the human motor cortex. As GABA mediated inhibition is thought to be a key modulator of plasticity within the adult brain, these findings have broad implications for the future use of tDCS. It is important, therefore, to establish whether tDCS can exert similar effects within non-motor brain areas. The aim of this study was to assess whether anodal tDCS could reduce inhibitory interactions within the human visual cortex. Psychophysical measures of surround suppression were used as an index of inhibition within V1. Overlay suppression, which is thought to originate within the lateral geniculate nucleus (LGN), was also measured as a control. Anodal stimulation of the occipital poles significantly reduced psychophysical surround suppression, but had no effect on overlay suppression. This effect was specific to anodal stimulation as cathodal stimulation had no effect on either measure. These psychophysical results provide the first evidence for tDCS-induced reductions of intracortical inhibition within the human visual cortex. PMID:22563485
Rallis, Austin; Fercho, Kelene A; Bosch, Taylor J; Baugh, Lee A
2018-01-31
Tool use is associated with three visual streams-dorso-dorsal, ventro-dorsal, and ventral visual streams. These streams are involved in processing online motor planning, action semantics, and tool semantics features, respectively. Little is known about the way in which the brain represents virtual tools. To directly assess this question, a virtual tool paradigm was created that provided the ability to manipulate tool components in isolation of one another. During functional magnetic resonance imaging (fMRI), adult participants performed a series of virtual tool manipulation tasks in which vision and movement kinematics of the tool were manipulated. Reaction time and hand movement direction were monitored while the tasks were performed. Functional imaging revealed that activity within all three visual streams was present, in a similar pattern to what would be expected with physical tool use. However, a previously unreported network of right-hemisphere activity was found including right inferior parietal lobule, middle and superior temporal gyri and supramarginal gyrus - regions well known to be associated with tool processing within the left hemisphere. These results provide evidence that both virtual and physical tools are processed within the same brain regions, though virtual tools recruit bilateral tool processing regions to a greater extent than physical tools. Copyright © 2017 Elsevier Ltd. All rights reserved.
No evidence for visual context-dependency of olfactory learning in Drosophila
NASA Astrophysics Data System (ADS)
Yarali, Ayse; Mayerle, Moritz; Nawroth, Christian; Gerber, Bertram
2008-08-01
How is behaviour organised across sensory modalities? Specifically, we ask concerning the fruit fly Drosophila melanogaster how visual context affects olfactory learning and recall and whether information about visual context is getting integrated into olfactory memory. We find that changing visual context between training and test does not deteriorate olfactory memory scores, suggesting that these olfactory memories can drive behaviour despite a mismatch of visual context between training and test. Rather, both the establishment and the recall of olfactory memory are generally facilitated by light. In a follow-up experiment, we find no evidence for learning about combinations of odours and visual context as predictors for reinforcement even after explicit training in a so-called biconditional discrimination task. Thus, a ‘true’ interaction between visual and olfactory modalities is not evident; instead, light seems to influence olfactory learning and recall unspecifically, for example by altering motor activity, alertness or olfactory acuity.
Biro, Dora; Guilford, Tim; Dell'Omo, Giacomo; Lipp, Hans-Peter
2002-12-01
Providing homing pigeons with a 5 min preview of the landscape at familiar sites prior to release reliably improves the birds' subsequent homing speeds. This phenomenon has been taken to suggest that the visual panorama is involved in familiar-site recognition, yet the exact nature of the improvement has never been elucidated. We employed newly developed miniature Global Positioning System (GPS) tracking technology to investigate how access to visual cues prior to release affects pigeons' flight along the length of the homing route. By applying a variety of novel analytical techniques enabled by the high-resolution GPS data (track efficiency, virtual vanishing bearings, orientation threshold), we localised the preview effect to the first 1000 m of the journey. Birds denied preview of a familiar landscape for 5 min before take-off flew an initially more tortuous path, including a high incidence of circling, possibly as part of an information-gathering strategy to determine their position. Beyond the first 1000 m, no differences were found in the performance of birds with or without preview. That the effect of the visual treatment was evident only in the early part of the journey suggests that lack of access to visual cues prior to release does not result in a non-specific effect on behaviour that is maintained throughout the flight. Instead, it seems that at least some decisions regarding the direction of home can be made prior to release and that such decisions are delayed if visual access to the landscape is denied. Overall, the variety of approaches applied here clearly highlight the potential for future applications of GPS tracking technology in navigation studies.
Transcutaneous Electrical Nerve Stimulation Effects on Neglect: A Visual-Evoked Potential Study
Pitzalis, Sabrina; Spinelli, Donatella; Vallar, Giuseppe; Di Russo, Francesco
2013-01-01
We studied the effects of transcutaneous electrical nerve stimulation (TENS) in six right-brain-damaged patients with left unilateral spatial neglect (USN), using both standard clinical tests (reading, line, and letter cancelation, and line bisection), and electrophysiological measures (steady-state visual-evoked potentials, SSVEP). TENS was applied on left neck muscles for 15′, and measures were recorded before, immediately after, and 60′ after stimulation. Behavioral results showed that the stimulation temporarily improved the deficit in all patients. In cancelation tasks, omissions and performance asymmetries between the two hand-sides were reduced, as well as the rightward deviation in line bisection. Before TENS, SSVEP average latency to stimuli displayed in the left visual half-field [LVF (160 ms)] was remarkably longer than to stimuli shown in the right visual half-field [RVF (120 ms)]. Immediately after TENS, latency to LVF stimuli was 130 ms; 1 h after stimulation the effect of TENS faded, with latency returning to baseline. TENS similarly affected also the latency SSVEP of 12 healthy participants, and their line bisection performance, with effects smaller in size. The present study, first, replicates evidence concerning the positive behavioral effects of TENS on the manifestations of left USN in right-brain-damaged patients; second, it shows putatively related electrophysiological effects on the SSVEP latency. These behavioral and novel electrophysiological results are discussed in terms of specific directional effects of left somatosensory stimulation on egocentric coordinates, which in USN patients are displaced toward the side of the cerebral lesion. Showing that visual-evoked potentials latency is modulated by proprioceptive stimulation, we provide electrophysiological evidence to the effect that TENS may improve some manifestations of USN, with implications for its rehabilitation. PMID:23966919
Sexual Orientation-Related Differences in Virtual Spatial Navigation and Spatial Search Strategies.
Rahman, Qazi; Sharp, Jonathan; McVeigh, Meadhbh; Ho, Man-Ling
2017-07-01
Spatial abilities are generally hypothesized to differ between men and women, and people with different sexual orientations. According to the cross-sex shift hypothesis, gay men are hypothesized to perform in the direction of heterosexual women and lesbian women in the direction of heterosexual men on cognitive tests. This study investigated sexual orientation differences in spatial navigation and strategy during a virtual Morris water maze task (VMWM). Forty-four heterosexual men, 43 heterosexual women, 39 gay men, and 34 lesbian/bisexual women (aged 18-54 years) navigated a desktop VMWM and completed measures of intelligence, handedness, and childhood gender nonconformity (CGN). We quantified spatial learning (hidden platform trials), probe trial performance, and cued navigation (visible platform trials). Spatial strategies during hidden and probe trials were classified into visual scanning, landmark use, thigmotaxis/circling, and enfilading. In general, heterosexual men scored better than women and gay men on some spatial learning and probe trial measures and used more visual scan strategies. However, some differences disappeared after controlling for age and estimated IQ (e.g., in visual scanning heterosexual men differed from women but not gay men). Heterosexual women did not differ from lesbian/bisexual women. For both sexes, visual scanning predicted probe trial performance. More feminine CGN scores were associated with lower performance among men and greater performance among women on specific spatial learning or probe trial measures. These results provide mixed evidence for the cross-sex shift hypothesis of sexual orientation-related differences in spatial cognition.
FGF /FGFR Signal Induces Trachea Extension in the Drosophila Visual System
Chu, Wei-Chen; Lee, Yuan-Ming; Henry Sun, Yi
2013-01-01
The Drosophila compound eye is a large sensory organ that places a high demand on oxygen supplied by the tracheal system. Although the development and function of the Drosophila visual system has been extensively studied, the development and contribution of its tracheal system has not been systematically examined. To address this issue, we studied the tracheal patterns and developmental process in the Drosophila visual system. We found that the retinal tracheae are derived from air sacs in the head, and the ingrowth of retinal trachea begin at mid-pupal stage. The tracheal development has three stages. First, the air sacs form near the optic lobe in 42-47% of pupal development (pd). Second, in 47-52% pd, air sacs extend branches along the base of the retina following a posterior-to-anterior direction and further form the tracheal network under the fenestrated membrane (TNUFM). Third, the TNUFM extend fine branches into the retina following a proximal-to-distal direction after 60% pd. Furthermore, we found that the trachea extension in both retina and TNUFM are dependent on the FGF(Bnl)/FGFR(Btl) signaling. Our results also provided strong evidence that the photoreceptors are the source of the Bnl ligand to guide the trachea ingrowth. Our work is the first systematic study of the tracheal development in the visual system, and also the first study demonstrating the interactions of two well-studied systems: the eye and trachea. PMID:23991208
Gori, Simone; Facoetti, Andrea
2015-01-14
Developmental dyslexia (DD) is the most common neurodevelopmental disorder (about 10% of children across cultures) characterized by severe difficulties in learning to read. According to the dominant view, DD is considered a phonological processing impairment that might be linked to a cross-modal, letter-to-speech sound integration deficit. However, new theories-supported by consistent data-suggest that mild deficits in low-level visual and auditory processing can lead to DD. This evidence supports the probabilistic and multifactorial approach for DD. Among others, an interesting visual deficit that is often associated with DD is excessive visual crowding. Crowding is defined as difficulty in the ability to recognize objects when surrounded by similar items. Crowding, typically observed in peripheral vision, could be modulated by attentional processes. The direct consequence of stronger crowding on reading is the inability to recognize letters when they are surrounded by other letters. This problem directly translates to reading at a slower speed and being more prone to making errors while reading. Our aim is to review the literature supporting the important role of crowding in DD. Moreover, we are interested in proposing new possible studies in order to clarify whether the observed excessive crowding could be a cause rather than an effect of DD. Finally, we also suggest possible remediation and even prevention programs that could be based on reducing the crowding in children with or at risk for DD without involving any phonological or orthographic training. © 2015 ARVO.
Carpenter, Christopher S; Nguyen, Hai V
2015-11-01
We provide new evidence on the effects of increasingly common driver cellphone bans on self-reported overall, handheld, and hands-free cellphone use while driving by studying Ontario, Canada, which instituted a 3-month education campaign in November 2009 followed by a binding driver cellphone ban in February 2010. Using residents of Alberta as a control group in a difference-in-differences framework, we find visual and regression-based evidence that Ontario's cellphone ban significantly reduced overall and handheld cellphone use. We also find that the policies significantly increased hands-free cellphone use. The reductions in overall and handheld use are driven exclusively by women, whereas the increases in hands-free use are much larger for men. Our results provide the first direct evidence that cellphone bans have the unintended effect of inducing substitution to hands-free devices. Copyright © 2014 John Wiley & Sons, Ltd.
Harris, Jill; Kamke, Marc R
2014-11-01
Selective attention fundamentally alters sensory perception, but little is known about the functioning of attention in individuals who use a cochlear implant. This study aimed to investigate visual and auditory attention in adolescent cochlear implant users. Event related potentials were used to investigate the influence of attention on visual and auditory evoked potentials in six cochlear implant users and age-matched normally-hearing children. Participants were presented with streams of alternating visual and auditory stimuli in an oddball paradigm: each modality contained frequently presented 'standard' and infrequent 'deviant' stimuli. Across different blocks attention was directed to either the visual or auditory modality. For the visual stimuli attention boosted the early N1 potential, but this effect was larger for cochlear implant users. Attention was also associated with a later P3 component for the visual deviant stimulus, but there was no difference between groups in the later attention effects. For the auditory stimuli, attention was associated with a decrease in N1 latency as well as a robust P3 for the deviant tone. Importantly, there was no difference between groups in these auditory attention effects. The results suggest that basic mechanisms of auditory attention are largely normal in children who are proficient cochlear implant users, but that visual attention may be altered. Ultimately, a better understanding of how selective attention influences sensory perception in cochlear implant users will be important for optimising habilitation strategies. Copyright © 2014 Elsevier Ireland Ltd. All rights reserved.
Goal-Directed and Habit-Like Modulations of Stimulus Processing during Reinforcement Learning.
Luque, David; Beesley, Tom; Morris, Richard W; Jack, Bradley N; Griffiths, Oren; Whitford, Thomas J; Le Pelley, Mike E
2017-03-15
Recent research has shown that perceptual processing of stimuli previously associated with high-value rewards is automatically prioritized even when rewards are no longer available. It has been hypothesized that such reward-related modulation of stimulus salience is conceptually similar to an "attentional habit." Recording event-related potentials in humans during a reinforcement learning task, we show strong evidence in favor of this hypothesis. Resistance to outcome devaluation (the defining feature of a habit) was shown by the stimulus-locked P1 component, reflecting activity in the extrastriate visual cortex. Analysis at longer latencies revealed a positive component (corresponding to the P3b, from 550-700 ms) sensitive to outcome devaluation. Therefore, distinct spatiotemporal patterns of brain activity were observed corresponding to habitual and goal-directed processes. These results demonstrate that reinforcement learning engages both attentional habits and goal-directed processes in parallel. Consequences for brain and computational models of reinforcement learning are discussed. SIGNIFICANCE STATEMENT The human attentional network adapts to detect stimuli that predict important rewards. A recent hypothesis suggests that the visual cortex automatically prioritizes reward-related stimuli, driven by cached representations of reward value; that is, stimulus-response habits. Alternatively, the neural system may track the current value of the predicted outcome. Our results demonstrate for the first time that visual cortex activity is increased for reward-related stimuli even when the rewarding event is temporarily devalued. In contrast, longer-latency brain activity was specifically sensitive to transient changes in reward value. Therefore, we show that both habit-like attention and goal-directed processes occur in the same learning episode at different latencies. This result has important consequences for computational models of reinforcement learning. Copyright © 2017 the authors 0270-6474/17/373009-09$15.00/0.
Visual attention for a desktop virtual environment with ambient scent
Toet, Alexander; van Schaik, Martin G.
2013-01-01
In the current study participants explored a desktop virtual environment (VE) representing a suburban neighborhood with signs of public disorder (neglect, vandalism, and crime), while being exposed to either room air (control group), or subliminal levels of tar (unpleasant; typically associated with burned or waste material) or freshly cut grass (pleasant; typically associated with natural or fresh material) ambient odor. They reported all signs of disorder they noticed during their walk together with their associated emotional response. Based on recent evidence that odors reflexively direct visual attention to (either semantically or affectively) congruent visual objects, we hypothesized that participants would notice more signs of disorder in the presence of ambient tar odor (since this odor may bias attention to unpleasant and negative features), and less signs of disorder in the presence of ambient grass odor (since this odor may bias visual attention toward the vegetation in the environment and away from the signs of disorder). Contrary to our expectations the results provide no indication that the presence of an ambient odor affected the participants’ visual attention for signs of disorder or their emotional response. However, the paradigm used in present study does not allow us to draw any conclusions in this respect. We conclude that a closer affective, semantic, or spatiotemporal link between the contents of a desktop VE and ambient scents may be required to effectively establish diagnostic associations that guide a user’s attention. In the absence of these direct links, ambient scent may be more diagnostic for the physical environment of the observer as a whole than for the particular items in that environment (or, in this case, items represented in the VE). PMID:24324453
Solnik, Stanislaw; Qiao, Mu; Latash, Mark L.
2017-01-01
This study tested two hypotheses on the nature of unintentional force drifts elicited by removing visual feedback during accurate force production tasks. The role of working memory (memory hypothesis) was explored in tasks with continuous force production, intermittent force production, and rest intervals over the same time interval. The assumption of unintentional drifts in referent coordinate for the fingertips was tested using manipulations of visual feedback: Young healthy subjects performed accurate steady-state force production tasks by pressing with the two index fingers on individual force sensors with visual feedback on the total force, sharing ratio, both, or none. Predictions based on the memory hypothesis have been falsified. In particular, we observed consistent force drifts to lower force values during continuous force production trials only. No force drift or drifts to higher forces were observed during intermittent force production trials and following rest intervals. The hypotheses based on the idea of drifts in referent finger coordinates have been confirmed. In particular, we observed superposition of two drift processes: A drift of total force to lower magnitudes and a drift of the sharing ratio to 50:50. When visual feedback on total force only was provided, the two finger forces showed drifts in opposite directions. We interpret the findings as evidence for the control of motor actions with changes in referent coordinates for participating effectors. Unintentional drifts in performance are viewed as natural relaxation processes in the involved systems; their typical time reflects stability in the direction of the drift. The magnitude of the drift was higher in the right (dominant) hand, which is consistent with the dynamic dominance hypothesis. PMID:28168396
Miall, R Chris; Kitchen, Nick M; Nam, Se-Ho; Lefumat, Hannah; Renault, Alix G; Ørstavik, Kristin; Cole, Jonathan D; Sarlegna, Fabrice R
2018-05-19
It is uncertain how vision and proprioception contribute to adaptation of voluntary arm movements. In normal participants, adaptation to imposed forces is possible with or without vision, suggesting that proprioception is sufficient; in participants with proprioceptive loss (PL), adaptation is possible with visual feedback, suggesting that proprioception is unnecessary. In experiment 1 adaptation to, and retention of, perturbing forces were evaluated in three chronically deafferented participants. They made rapid reaching movements to move a cursor toward a visual target, and a planar robot arm applied orthogonal velocity-dependent forces. Trial-by-trial error correction was observed in all participants. Such adaptation has been characterized with a dual-rate model: a fast process that learns quickly, but retains poorly and a slow process that learns slowly and retains well. Experiment 2 showed that the PL participants had large individual differences in learning and retention rates compared to normal controls. Experiment 3 tested participants' perception of applied forces. With visual feedback, the PL participants could report the perturbation's direction as well as controls; without visual feedback, thresholds were elevated. Experiment 4 showed, in healthy participants, that force direction could be estimated from head motion, at levels close to the no-vision threshold for the PL participants. Our results show that proprioceptive loss influences perception, motor control and adaptation but that proprioception from the moving limb is not essential for adaptation to, or detection of, force fields. The differences in learning and retention seen between the three deafferented participants suggest that they achieve these tasks in idiosyncratic ways after proprioceptive loss, possibly integrating visual and vestibular information with individual cognitive strategies.
Choi, Jong Moon; Cho, Yang Seok; Proctor, Robert W
2009-09-01
A Stroop task with separate color bar and color word stimuli was combined with an inhibition-of-return procedure to examine whether visual attention modulates color word processing. In Experiment 1, the color bar was presented at the cued location and the color word at the uncued location, or vice versa, with a 100- or 1,050-msec stimulus onset asynchrony (SOA) between cue and Stroop stimuli. In Experiment 2, on Stroop trials, the color bar was presented at a central fixated location and the color word at a cued or uncued location above or below the color bar. In both experiments, with a 100-msec SOA, the Stroop effect was numerically larger when the color word was displayed at the cued location than when it was displayed at the uncued location, but with the 1,050-msec SOA, this relation between Stroop effect magnitude and location was reversed. These results provide evidence that processing of the color word in the Stroop task is modulated by the location to which visual attention is directed.
Social Class and the Motivational Relevance of Other Human Beings: Evidence From Visual Attention.
Dietze, Pia; Knowles, Eric D
2016-11-01
We theorize that people's social class affects their appraisals of others' motivational relevance-the degree to which others are seen as potentially rewarding, threatening, or otherwise worth attending to. Supporting this account, three studies indicate that social classes differ in the amount of attention their members direct toward other human beings. In Study 1, wearable technology was used to film the visual fields of pedestrians on city streets; higher-class participants looked less at other people than did lower-class participants. In Studies 2a and 2b, participants' eye movements were tracked while they viewed street scenes; higher class was associated with reduced attention to people in the images. In Study 3, a change-detection procedure assessed the degree to which human faces spontaneously attract visual attention; faces proved less effective at drawing the attention of high-class than low-class participants, which implies that class affects spontaneous relevance appraisals. The measurement and conceptualization of social class are discussed. © The Author(s) 2016.
Eye movements during object recognition in visual agnosia.
Charles Leek, E; Patterson, Candy; Paul, Matthew A; Rafal, Robert; Cristino, Filipe
2012-07-01
This paper reports the first ever detailed study about eye movement patterns during single object recognition in visual agnosia. Eye movements were recorded in a patient with an integrative agnosic deficit during two recognition tasks: common object naming and novel object recognition memory. The patient showed normal directional biases in saccades and fixation dwell times in both tasks and was as likely as controls to fixate within object bounding contour regardless of recognition accuracy. In contrast, following initial saccades of similar amplitude to controls, the patient showed a bias for short saccades. In object naming, but not in recognition memory, the similarity of the spatial distributions of patient and control fixations was modulated by recognition accuracy. The study provides new evidence about how eye movements can be used to elucidate the functional impairments underlying object recognition deficits. We argue that the results reflect a breakdown in normal functional processes involved in the integration of shape information across object structure during the visual perception of shape. Copyright © 2012 Elsevier Ltd. All rights reserved.
Sieger, Tomáš; Serranová, Tereza; Růžička, Filip; Vostatek, Pavel; Wild, Jiří; Štastná, Daniela; Bonnet, Cecilia; Novák, Daniel; Růžička, Evžen; Urgošík, Dušan; Jech, Robert
2015-03-10
Both animal studies and studies using deep brain stimulation in humans have demonstrated the involvement of the subthalamic nucleus (STN) in motivational and emotional processes; however, participation of this nucleus in processing human emotion has not been investigated directly at the single-neuron level. We analyzed the relationship between the neuronal firing from intraoperative microrecordings from the STN during affective picture presentation in patients with Parkinson's disease (PD) and the affective ratings of emotional valence and arousal performed subsequently. We observed that 17% of neurons responded to emotional valence and arousal of visual stimuli according to individual ratings. The activity of some neurons was related to emotional valence, whereas different neurons responded to arousal. In addition, 14% of neurons responded to visual stimuli. Our results suggest the existence of neurons involved in processing or transmission of visual and emotional information in the human STN, and provide evidence of separate processing of the affective dimensions of valence and arousal at the level of single neurons as well.
Transcranial direct current stimulation enhances recovery of stereopsis in adults with amblyopia.
Spiegel, Daniel P; Li, Jinrong; Hess, Robert F; Byblow, Winston D; Deng, Daming; Yu, Minbin; Thompson, Benjamin
2013-10-01
Amblyopia is a neurodevelopmental disorder of vision caused by abnormal visual experience during early childhood that is often considered to be untreatable in adulthood. Recently, it has been shown that a novel dichoptic videogame-based treatment for amblyopia can improve visual function in adult patients, at least in part, by reducing inhibition of inputs from the amblyopic eye to the visual cortex. Non-invasive anodal transcranial direct current stimulation has been shown to reduce the activity of inhibitory cortical interneurons when applied to the primary motor or visual cortex. In this double-blind, sham-controlled cross-over study we tested the hypothesis that anodal transcranial direct current stimulation of the visual cortex would enhance the therapeutic effects of dichoptic videogame-based treatment. A homogeneous group of 16 young adults (mean age 22.1 ± 1.1 years) with amblyopia were studied to compare the effect of dichoptic treatment alone and dichoptic treatment combined with visual cortex direct current stimulation on measures of binocular (stereopsis) and monocular (visual acuity) visual function. The combined treatment led to greater improvements in stereoacuity than dichoptic treatment alone, indicating that direct current stimulation of the visual cortex boosts the efficacy of dichoptic videogame-based treatment. This intervention warrants further evaluation as a novel therapeutic approach for adults with amblyopia.
Visual Information Processing Based on Spatial Filters Constrained by Biological Data.
1978-12-01
was provided by Pantie and Sekuler ( 19681. They found that the detection (if gratings was affected most by adapting isee Section 6.1. 11 to square...evidence for certain eye scans being directed by spatial information in filtered images is given. Eye scan paths of a portrait of a young girl I Figure 08...multistable objects to more complex objects such as the man- girl figure of Fisher 119681, decision boundaries that are a natural concomitant to any pattern
Li, Siyao; Cai, Ying; Liu, Jing; Li, Dawei; Feng, Zifang; Chen, Chuansheng; Xue, Gui
2017-04-01
Mounting evidence suggests that multiple mechanisms underlie working memory capacity. Using transcranial direct current stimulation (tDCS), the current study aimed to provide causal evidence for the neural dissociation of two mechanisms underlying visual working memory (WM) capacity, namely, the scope and control of attention. A change detection task with distractors was used, where a number of colored bars (i.e., two red bars, four red bars, or two red plus two blue bars) were presented on both sides (Experiment 1) or the center (Experiment 2) of the screen for 100ms, and participants were instructed to remember the red bars and to ignore the blue bars (in both Experiments), as well as to ignore the stimuli on the un-cued side (Experiment 1 only). In both experiments, participants finished three sessions of the task after 15min of 1.5mA anodal tDCS administered on the right prefrontal cortex (PFC), the right posterior parietal cortex (PPC), and the primary visual cortex (VC), respectively. The VC stimulation served as an active control condition. We found that compared to stimulation on the VC, stimulation on the right PPC specifically increased the visual WM capacity under the no-distractor condition (i.e., 4 red bars), whereas stimulation on the right PFC specifically increased the visual WM capacity under the distractor condition (i.e., 2 red bars plus 2 blue bars). These results suggest that the PPC and PFC are involved in the scope and control of attention, respectively. We further showed that compared to central presentation of the stimuli (Experiment 2), bilateral presentation of the stimuli (on both sides of the fixation in Experiment 1) led to an additional demand for attention control. Our results emphasize the dissociated roles of the frontal and parietal lobes in visual WM capacity, and provide a deeper understanding of the neural mechanisms of WM. Copyright © 2017 Elsevier Inc. All rights reserved.
Inferring the direction of implied motion depends on visual awareness
Faivre, Nathan; Koch, Christof
2014-01-01
Visual awareness of an event, object, or scene is, by essence, an integrated experience, whereby different visual features composing an object (e.g., orientation, color, shape) appear as an unified percept and are processed as a whole. Here, we tested in human observers whether perceptual integration of static motion cues depends on awareness by measuring the capacity to infer the direction of motion implied by a static visible or invisible image under continuous flash suppression. Using measures of directional adaptation, we found that visible but not invisible implied motion adaptors biased the perception of real motion probes. In a control experiment, we found that invisible adaptors implying motion primed the perception of subsequent probes when they were identical (i.e., repetition priming), but not when they only shared the same direction (i.e., direction priming). Furthermore, using a model of visual processing, we argue that repetition priming effects are likely to arise as early as in the primary visual cortex. We conclude that although invisible images implying motion undergo some form of nonconscious processing, visual awareness is necessary to make inferences about motion direction. PMID:24706951
Inferring the direction of implied motion depends on visual awareness.
Faivre, Nathan; Koch, Christof
2014-04-04
Visual awareness of an event, object, or scene is, by essence, an integrated experience, whereby different visual features composing an object (e.g., orientation, color, shape) appear as an unified percept and are processed as a whole. Here, we tested in human observers whether perceptual integration of static motion cues depends on awareness by measuring the capacity to infer the direction of motion implied by a static visible or invisible image under continuous flash suppression. Using measures of directional adaptation, we found that visible but not invisible implied motion adaptors biased the perception of real motion probes. In a control experiment, we found that invisible adaptors implying motion primed the perception of subsequent probes when they were identical (i.e., repetition priming), but not when they only shared the same direction (i.e., direction priming). Furthermore, using a model of visual processing, we argue that repetition priming effects are likely to arise as early as in the primary visual cortex. We conclude that although invisible images implying motion undergo some form of nonconscious processing, visual awareness is necessary to make inferences about motion direction.
Matsunaka, Kumiko; Shibata, Yuki; Yamamoto, Toshikazu
2008-08-01
Study 1 investigated individual differences in spatial cognition amongst visually impaired students and sighted controls, as well as the extent to which visual status contributes to these individual differences. Fifty-eight visually impaired and 255 sighted university students evaluated their sense of direction via self-ratings. Visual impairment contributed to the factors associated with the use and understanding of maps, confirming that maps are generally unfamiliar to visually impaired people. The relationship between psychological stress associated with mobility and individual differences in sense of direction was investigated in Study 2. A stress checklist was administered to the 51 visually impaired students who participated in Study 1. Psychological stress level was related to understanding and use of maps, as well as orientation and renewal, that is, course correction after being got lost. Central visual field deficits were associated with greater mobility-related stress levels than peripheral visual field deficits.
NASA Astrophysics Data System (ADS)
Chunbo, Yuan; Ying, Wu; Yueming, Sun; Zuhong, Lu; Juzheng, Liu
1997-12-01
Molecularly resolved atomic force microscopic images of phosphatidic acid Langmuir-Blodgett bilayers show that phosphate groups in polar region of the films are packing in a distorted hexagonal organization with long-range orientational and positional order. Intermolecular hydrogen bonding interactions, which should be responsible for the ordering and stability of bilayers, are visualized directly between adjacent phosphate groups in the polar region of the bilayer. Some adjacent phosphatidic acid molecules link each other through the formation of intermolecular hydrogen bonds between phosphate groups in polar region to form local supramolecules, which provide the bilayer's potential as a functionized film in the investigation on the lateral conductions of protons in the biological bilayers.
NASA Astrophysics Data System (ADS)
Zhang, Zhi-Gang; Abe, Tomohiro; Moriyoshi, Chikako; Tanaka, Hiroshi; Kuroiwa, Yoshihiro
2018-07-01
Synchrotron-radiation X-ray diffraction studies as a function of temperature reveal the structural origin of the spontaneous polarization and related lattice strains in stoichiometric LiTaO3. Electron charge density distribution maps visualized by the maximum entropy method clearly demonstrate that ordering of the disordered Li ion in the polar direction accompanied by deformation of the oxygen octahedra lead to the ferroelectric phase transition. The ionic polarization attributed to the ionic displacements is dominant in the polar structure. The structural change occurs continuously at the phase transition temperature, which suggests a second-order phase transition.
de la Rosa, Stephan; Ekramnia, Mina; Bülthoff, Heinrich H.
2016-01-01
The ability to discriminate between different actions is essential for action recognition and social interactions. Surprisingly previous research has often probed action recognition mechanisms with tasks that did not require participants to discriminate between actions, e.g., left-right direction discrimination tasks. It is not known to what degree visual processes in direction discrimination tasks are also involved in the discrimination of actions, e.g., when telling apart a handshake from a high-five. Here, we examined whether action discrimination is influenced by movement direction and whether direction discrimination depends on the type of action. We used an action adaptation paradigm to target action and direction discrimination specific visual processes. In separate conditions participants visually adapted to forward and backward moving handshake and high-five actions. Participants subsequently categorized either the action or the movement direction of an ambiguous action. The results showed that direction discrimination adaptation effects were modulated by the type of action but action discrimination adaptation effects were unaffected by movement direction. These results suggest that action discrimination and direction categorization rely on partly different visual information. We propose that action discrimination tasks should be considered for the exploration of visual action recognition mechanisms. PMID:26941633
Valente, Dannyelle; Theurel, Anne; Gentaz, Edouard
2018-04-01
Facial expressions of emotion are nonverbal behaviors that allow us to interact efficiently in social life and respond to events affecting our welfare. This article reviews 21 studies, published between 1932 and 2015, examining the production of facial expressions of emotion by blind people. It particularly discusses the impact of visual experience on the development of this behavior from birth to adulthood. After a discussion of three methodological considerations, the review of studies reveals that blind subjects demonstrate differing capacities for producing spontaneous expressions and voluntarily posed expressions. Seventeen studies provided evidence that blind and sighted spontaneously produce the same pattern of facial expressions, even if some variations can be found, reflecting facial and body movements specific to blindness or differences in intensity and control of emotions in some specific contexts. This suggests that lack of visual experience seems to not have a major impact when this behavior is generated spontaneously in real emotional contexts. In contrast, eight studies examining voluntary expressions indicate that blind individuals have difficulty posing emotional expressions. The opportunity for prior visual observation seems to affect performance in this case. Finally, we discuss three new directions for research to provide additional and strong evidence for the debate regarding the innate or the culture-constant learning character of the production of emotional facial expressions by blind individuals: the link between perception and production of facial expressions, the impact of display rules in the absence of vision, and the role of other channels in expression of emotions in the context of blindness.
Heiser, Laura M; Berman, Rebecca A; Saunders, Richard C; Colby, Carol L
2005-11-01
With each eye movement, a new image impinges on the retina, yet we do not notice any shift in visual perception. This perceptual stability indicates that the brain must be able to update visual representations to take our eye movements into account. Neurons in the lateral intraparietal area (LIP) update visual representations when the eyes move. The circuitry that supports these updated representations remains unknown, however. In this experiment, we asked whether the forebrain commissures are necessary for updating in area LIP when stimulus representations must be updated from one visual hemifield to the other. We addressed this question by recording from LIP neurons in split-brain monkeys during two conditions: stimulus traces were updated either across or within hemifields. Our expectation was that across-hemifield updating activity in LIP would be reduced or abolished after transection of the forebrain commissures. Our principal finding is that LIP neurons can update stimulus traces from one hemifield to the other even in the absence of the forebrain commissures. This finding provides the first evidence that representations in parietal cortex can be updated without the use of direct cortico-cortical links. The second main finding is that updating activity in LIP is modified in the split-brain monkey: across-hemifield signals are reduced in magnitude and delayed in onset compared with within-hemifield signals, which indicates that the pathways for across-hemifield updating are less effective in the absence of the forebrain commissures. Together these findings reveal a dynamic circuit that contributes to updating spatial representations.
Coggan, David D; Baker, Daniel H; Andrews, Timothy J
2016-01-01
Brain-imaging studies have found distinct spatial and temporal patterns of response to different object categories across the brain. However, the extent to which these categorical patterns of response reflect higher-level semantic or lower-level visual properties of the stimulus remains unclear. To address this question, we measured patterns of EEG response to intact and scrambled images in the human brain. Our rationale for using scrambled images is that they have many of the visual properties found in intact images, but do not convey any semantic information. Images from different object categories (bottle, face, house) were briefly presented (400 ms) in an event-related design. A multivariate pattern analysis revealed categorical patterns of response to intact images emerged ∼80-100 ms after stimulus onset and were still evident when the stimulus was no longer present (∼800 ms). Next, we measured the patterns of response to scrambled images. Categorical patterns of response to scrambled images also emerged ∼80-100 ms after stimulus onset. However, in contrast to the intact images, distinct patterns of response to scrambled images were mostly evident while the stimulus was present (∼400 ms). Moreover, scrambled images were able to account only for all the variance in the intact images at early stages of processing. This direct manipulation of visual and semantic content provides new insights into the temporal dynamics of object perception and the extent to which different stages of processing are dependent on lower-level or higher-level properties of the image.
The Human Retrosplenial Cortex and Thalamus Code Head Direction in a Global Reference Frame.
Shine, Jonathan P; Valdés-Herrera, José P; Hegarty, Mary; Wolbers, Thomas
2016-06-15
Spatial navigation is a multisensory process involving integration of visual and body-based cues. In rodents, head direction (HD) cells, which are most abundant in the thalamus, integrate these cues to code facing direction. Human fMRI studies examining HD coding in virtual environments (VE) have reported effects in retrosplenial complex and (pre-)subiculum, but not the thalamus. Furthermore, HD coding appeared insensitive to global landmarks. These tasks, however, provided only visual cues for orientation, and attending to global landmarks did not benefit task performance. In the present study, participants explored a VE comprising four separate locales, surrounded by four global landmarks. To provide body-based cues, participants wore a head-mounted display so that physical rotations changed facing direction in the VE. During subsequent MRI scanning, subjects saw stationary views of the environment and judged whether their orientation was the same as in the preceding trial. Parameter estimates extracted from retrosplenial cortex and the thalamus revealed significantly reduced BOLD responses when HD was repeated. Moreover, consistent with rodent findings, the signal did not continue to adapt over repetitions of the same HD. These results were supported by a whole-brain analysis showing additional repetition suppression in the precuneus. Together, our findings suggest that: (1) consistent with the rodent literature, the human thalamus may integrate visual and body-based, orientation cues; (2) global reference frame cues can be used to integrate HD across separate individual locales; and (3) immersive training procedures providing full body-based cues may help to elucidate the neural mechanisms supporting spatial navigation. In rodents, head direction (HD) cells signal facing direction in the environment via increased firing when the animal assumes a certain orientation. Distinct brain regions, the retrosplenial cortex (RSC) and thalamus, code for visual and vestibular cues of orientation, respectively. Putative HD signals have been observed in human RSC but not the thalamus, potentially because body-based cues were not provided. Here, participants encoded HD in a novel virtual environment while wearing a head-mounted display to provide body-based cues for orientation. In subsequent fMRI scanning, we found evidence of an HD signal in RSC, thalamus, and precuneus. These findings harmonize rodent and human data, and suggest that immersive training procedures provide a viable way to examine the neural basis of navigation. Copyright © 2016 the authors 0270-6474/16/366371-11$15.00/0.
Electrophysiological evidence for biased competition in V1 for fear expressions.
West, Greg L; Anderson, Adam A K; Ferber, Susanne; Pratt, Jay
2011-11-01
When multiple stimuli are concurrently displayed in the visual field, they must compete for neural representation at the processing expense of their contemporaries. This biased competition is thought to begin as early as primary visual cortex, and can be driven by salient low-level stimulus features. Stimuli important for an organism's survival, such as facial expressions signaling environmental threat, might be similarly prioritized at this early stage of visual processing. In the present study, we used ERP recordings from striate cortex to examine whether fear expressions can bias the competition for neural representation at the earliest stage of retinotopic visuo-cortical processing when in direct competition with concurrently presented visual information of neutral valence. We found that within 50 msec after stimulus onset, information processing in primary visual cortex is biased in favor of perceptual representations of fear at the expense of competing visual information (Experiment 1). Additional experiments confirmed that the facial display's emotional content rather than low-level features is responsible for this prioritization in V1 (Experiment 2), and that this competition is reliant on a face's upright canonical orientation (Experiment 3). These results suggest that complex stimuli important for an organism's survival can indeed be prioritized at the earliest stage of cortical processing at the expense of competing information, with competition possibly beginning before encoding in V1.
Fan, Zhao; Harris, John
2010-10-12
In a recent study (Fan, Z., & Harris, J. (2008). Perceived spatial displacement of motion-defined contours in peripheral vision. Vision Research, 48(28), 2793-2804), we demonstrated that virtual contours defined by two regions of dots moving in opposite directions were displaced perceptually in the direction of motion of the dots in the more eccentric region when the contours were viewed in the right visual field. Here, we show that the magnitude and/or direction of these displacements varies in different quadrants of the visual field. When contours were presented in the lower visual field, the direction of perceived contour displacement was consistent with that when both contours were presented in the right visual field. However, this illusory motion-induced spatial displacement disappeared when both contours were presented in the upper visual field. Also, perceived contour displacement in the direction of the more eccentric dots was larger in the right than in the left visual field, perhaps because of a hemispheric asymmetry in attentional allocation. Quadrant-based analyses suggest that the pattern of results arises from opposite directions of perceived contour displacement in the upper-left and lower-right visual quadrants, which depend on the relative strengths of two effects: a greater sensitivity to centripetal motion, and an asymmetry in the allocation of spatial attention. Copyright © 2010 Elsevier Ltd. All rights reserved.
Moonlight Makes Owls More Chatty
Penteriani, Vincenzo; Delgado, María del Mar; Campioni, Letizia; Lourenço, Rui
2010-01-01
Background Lunar cycles seem to affect many of the rhythms, temporal patterns and behaviors of living things on Earth. Ambient light is known to affect visual communication in animals, with the conspicuousness of visual signals being largely determined by the light available for reflection by the sender. Although most previous studies in this context have focused on diurnal light, moonlight should not be neglected from the perspective of visual communication among nocturnal species. We recently discovered that eagle owls Bubo bubo communicate with conspecifics using a patch of white throat plumage that is repeatedly exposed during each call and is only visible during vocal displays. Methodology/Principal Findings Here we provide evidence that this species uses moonlight to increase the conspicuousness of this visual signal during call displays. We found that call displays are directly influenced by the amount of moonlight, with silent nights being more frequent during periods with no-moonlight than moonlight. Furthermore, high numbers of calling bouts were more frequent at moonlight. Finally, call posts were located on higher positions on moonlit nights. Conclusions/Significance Our results support the idea that moon phase affects the visual signaling behavior of this species, and provide a starting point for examination of this method of communication by nocturnal species. PMID:20098700
Think spatial: the representation in mental rotation is nonvisual.
Liesefeld, Heinrich R; Zimmer, Hubert D
2013-01-01
For mental rotation, introspection, theories, and interpretations of experimental results imply a certain type of mental representation, namely, visual mental images. Characteristics of the rotated representation can be examined by measuring the influence of stimulus characteristics on rotational speed. If the amount of a given type of information influences rotational speed, one can infer that it was contained in the rotated representation. In Experiment 1, rotational speed of university students (10 men, 11 women) was found to be influenced exclusively by the amount of represented orientation-dependent spatial-relational information but not by orientation-independent spatial-relational information, visual complexity, or the number of stimulus parts. As information in mental-rotation tasks is initially presented visually, this finding implies that at some point during each trial, orientation-dependent information is extracted from visual information. Searching for more direct evidence for this extraction, we recorded the EEG of another sample of university students (12 men, 12 women) during mental rotation of the same stimuli. In an early time window, the observed working memory load-dependent slow potentials were sensitive to the stimuli's visual complexity. Later, in contrast, slow potentials were sensitive to the amount of orientation-dependent information only. We conclude that only orientation-dependent information is contained in the rotated representation. (PsycINFO Database Record (c) 2013 APA, all rights reserved).
In situ AFM imaging of apolipoprotein A-I directly derived from plasma HDL.
Gan, Chaoye; Wang, Zhexuan; Chen, Yong
2017-04-01
The major apolipoproteins of plasma lipoproteins play vital roles in the structural integrity and physiological functions of lipoproteins. More than ten structural models of apolipoprotein A-I (apoA-I), the major apolipoprotein of high-density lipoprotein (HDL), have been developed successively. In these models, apoA-I was supposed to organize in a ring-shaped form. To date, however, there is no direct evidence under physiological condition. Here, atomic force microscopy (AFM) was used to in situ visualize the organization of apoA-I, which was exposed via depletion of the lipid component of plasma HDL pre-immobilized on functionalized mica sheets. For the first time, the ring-shaped coarse structure and three detailed structures (crescent-shaped, gapped "O"-shaped, and parentheses-shaped structures, respectively) of apoA-I in plasma HDL, which have the ability of binding scavenger receptors, were directly observed and quantitatively measured by AFM. The three detailed structures probably represent the different extents to which the lipid component of HDL was depleted. Data on lipid depletion of HDL may provide clues to understand lipid insertion of HDL. These data provide important information for the understanding of the structure/maturation of plasma HDL. Moreover, they suggest a powerful method for directly visualizing the major apolipoproteins of plasma lipoproteins or the protein component of lipoprotein-like lipid-protein complexes. Copyright © 2017 Elsevier B.V. All rights reserved.
Direct Visualization of Catalytically Active Sites at the FeO–Pt(111) Interface
Kudernatsch, Wilhelmine; Peng, Guowen; Zeuthen, Helene; ...
2015-05-31
Within the area of surface science, one of the “holy grails” is to directly visualize a chemical reaction at the atomic scale. Whereas this goal has been reached by high-resolution scanning tunneling microscopy (STM) in a number of cases for reactions occurring at flat surfaces, such a direct view is often inhibited for reaction occurring at steps and interfaces. Here we have studied the CO oxidation reaction at the interface between ultrathin FeO islands and a Pt(111) support by in situ STM and density functional theory (DFT) calculations. Time-lapsed STM imaging on this inverse model catalyst in O 2 andmore » CO environments revealed catalytic activity occurring at the FeO–Pt(111) interface and directly showed that the Fe-edges host the catalytically most active sites for the CO oxidation reaction. This is an important result since previous evidence for the catalytic activity of the FeO–Pt(111) interface is essentially based on averaging techniques in conjunction with DFT calculations. As a result, the presented STM results are in accord with DFT+U calculations, in which we compare possible CO oxidation pathways on oxidized Fe-edges and O-edges. We found that the CO oxidation reaction is more favorable on the oxidized Fe-edges, both thermodynamically and kinetically.« less
Evidence for negative feature guidance in visual search is explained by spatial recoding.
Beck, Valerie M; Hollingworth, Andrew
2015-10-01
Theories of attention and visual search explain how attention is guided toward objects with known target features. But can attention be directed away from objects with a feature known to be associated only with distractors? Most studies have found that the demand to maintain the to-be-avoided feature in visual working memory biases attention toward matching objects rather than away from them. In contrast, Arita, Carlisle, and Woodman (2012) claimed that attention can be configured to selectively avoid objects that match a cued distractor color, and they reported evidence that this type of negative cue generates search benefits. However, the colors of the search array items in Arita et al. (2012) were segregated by hemifield (e.g., blue items on the left, red on the right), which allowed for a strategy of translating the feature-cue information into a simple spatial template (e.g., avoid right, or attend left). In the present study, we replicated the negative cue benefit using the Arita et al. (2012), method (albeit within a subset of participants who reliably used the color cues to guide attention). Then, we eliminated the benefit by using search arrays that could not be grouped by hemifield. Our results suggest that feature-guided avoidance is implemented only indirectly, in this case by translating feature-cue information into a spatial template. (c) 2015 APA, all rights reserved).
Hertrich, Ingo; Dietrich, Susanne; Ackermann, Hermann
2011-01-01
During speech communication, visual information may interact with the auditory system at various processing stages. Most noteworthy, recent magnetoencephalography (MEG) data provided first evidence for early and preattentive phonetic/phonological encoding of the visual data stream--prior to its fusion with auditory phonological features [Hertrich, I., Mathiak, K., Lutzenberger, W., & Ackermann, H. Time course of early audiovisual interactions during speech and non-speech central-auditory processing: An MEG study. Journal of Cognitive Neuroscience, 21, 259-274, 2009]. Using functional magnetic resonance imaging, the present follow-up study aims to further elucidate the topographic distribution of visual-phonological operations and audiovisual (AV) interactions during speech perception. Ambiguous acoustic syllables--disambiguated to /pa/ or /ta/ by the visual channel (speaking face)--served as test materials, concomitant with various control conditions (nonspeech AV signals, visual-only and acoustic-only speech, and nonspeech stimuli). (i) Visual speech yielded an AV-subadditive activation of primary auditory cortex and the anterior superior temporal gyrus (STG), whereas the posterior STG responded both to speech and nonspeech motion. (ii) The inferior frontal and the fusiform gyrus of the right hemisphere showed a strong phonetic/phonological impact (differential effects of visual /pa/ vs. /ta/) upon hemodynamic activation during presentation of speaking faces. Taken together with the previous MEG data, these results point at a dual-pathway model of visual speech information processing: On the one hand, access to the auditory system via the anterior supratemporal “what" path may give rise to direct activation of "auditory objects." On the other hand, visual speech information seems to be represented in a right-hemisphere visual working memory, providing a potential basis for later interactions with auditory information such as the McGurk effect.
Bhatia, Sujata K; Yetter, Ann B
2008-08-01
Medical devices and implanted biomaterials are often assessed for biological reactivity using visual scores of cell-material interactions. In such testing, biomaterials are assigned cytotoxicity ratings based on visual evidence of morphological cellular changes, including cell lysis, rounding, spreading, and proliferation. For example, ISO 10993 cytotoxicity testing of medical devices allows the use of a visual grading scale. The present study compared visual in vitro cytotoxicity ratings to quantitative in vitro cytotoxicity measurements for biomaterials to determine the level of correlation between visual scoring and a quantitative cell viability assay. Biomaterials representing a spectrum of biological reactivity levels were evaluated, including organo-tin polyvinylchloride (PVC; a known cytotoxic material), ultra-high molecular weight polyethylene (a known non-cytotoxic material), and implantable tissue adhesives. Each material was incubated in direct contact with mouse 3T3 fibroblast cell cultures for 24 h. Visual scores were assigned to the materials using a 5-point rating scale; the scorer was blinded to the material identities. Quantitative measurements of cell viability were performed using a 3-(4,5-dimethylthiozol-2-yl)-2,5-diphenyltetrazolium bromide (MTT) colorimetric assay; again, the assay operator was blinded to material identities. The investigation revealed a high degree of correlation between visual cytotoxicity ratings and quantitative cell viability measurements; a Pearson's correlation gave a correlation coefficient of 0.90 between the visual cytotoxicity score and the percent viable cells. An equation relating the visual cytotoxicity score and the percent viable cells was derived. The results of this study are significant for the design and interpretation of in vitro cytotoxicity studies of novel biomaterials.
Noguchi, Yasuki; Tomoike, Kouta
2016-01-12
Recent studies argue that strongly-motivated positive emotions (e.g. desire) narrow a scope of attention. This argument is mainly based on an observation that, while humans normally respond faster to global than local information of a visual stimulus (global advantage), positive affects eliminated the global advantage by selectively speeding responses to local (but not global) information. In other words, narrowing of attentional scope was indirectly evidenced by the elimination of global advantage (the same speed of processing between global and local information). No study has directly shown that strongly-motivated positive affects induce faster responses to local than global information while excluding a bias for global information (global advantage) in a baseline (emotionally-neutral) condition. In the present study, we addressed this issue by eliminating the global advantage in a baseline (neutral) state. Induction of positive affects under this state resulted in faster responses to local than global information. Our results provided direct evidence that positive affects in high motivational intensity narrow a scope of attention.
NASA Technical Reports Server (NTRS)
Lackner, J. R.; Levine, M. S.
1979-01-01
Human experiments are carried out which support the observation of Goodwin (1973) and Goodwin et al. (1972) that vibration of skeletal muscles can elicit illusory limb motion. These experiments extend the class of possible myesthetic illusions by showing that vibration of the appropriate muscles can produce illusory body motion in nearly any desired direction. Such illusory changes in posture occur only when visual information about body orientation is absent; these changes in apparent posture are sometimes accompanied by a slow-phase nystagmus that compensates for the direction of apparent body motion. During illusory body motion a stationary target light that is fixated will appear to move with the body at the same apparent velocity. However, this pattern of apparent body motion and conjoint visual - defined as propriogyral illusion - is suppressed if the subject is in a fully illuminated environment providing cues about true body orientation. Persuasive evidence is thus provided for the contribution of both muscle afferent and touch-pressure information to the supraspinal mechanisms that determine apparent orientation on the basis of ongoing patterns of interoceptive and exteroceptive activity.
Visual direction finding by fishes
NASA Technical Reports Server (NTRS)
Waterman, T. H.
1972-01-01
The use of visual orientation, in the absence of landmarks, for underwater direction finding exercises by fishes is reviewed. Celestial directional clues observed directly near the water surface or indirectly at an asymptatic depth are suggested as possible orientation aids.
The case against specialized visual-spatial short-term memory.
Morey, Candice C
2018-05-24
The dominant paradigm for understanding working memory, or the combination of the perceptual, attentional, and mnemonic processes needed for thinking, subdivides short-term memory (STM) according to whether memoranda are encoded in aural-verbal or visual formats. This traditional dissociation has been supported by examples of neuropsychological patients who seem to selectively lack STM for either aural-verbal, visual, or spatial memoranda, and by experimental research using dual-task methods. Though this evidence is the foundation of assumptions of modular STM systems, the case it makes for a specialized visual STM system is surprisingly weak. I identify the key evidence supporting a distinct verbal STM system-patients with apparent selective damage to verbal STM and the resilience of verbal short-term memories to general dual-task interference-and apply these benchmarks to neuropsychological and experimental investigations of visual-spatial STM. Contrary to the evidence on verbal STM, patients with apparent visual or spatial STM deficits tend to experience a wide range of additional deficits, making it difficult to conclude that a distinct short-term store was damaged. Consistently with this, a meta-analysis of dual-task visual-spatial STM research shows that robust dual-task costs are consistently observed regardless of the domain or sensory code of the secondary task. Together, this evidence suggests that positing a specialized visual STM system is not necessary. (PsycINFO Database Record (c) 2018 APA, all rights reserved).
Kumru, Hatice; Pelayo, Raul; Vidal, Joan; Tormos, Josep Maria; Fregni, Felipe; Navarro, Xavier; Pascual-Leone, Alvaro
2010-01-01
The aim of this study was to evaluate the analgesic effect of transcranial direct current stimulation of the motor cortex and techniques of visual illusion, applied isolated or combined, in patients with neuropathic pain following spinal cord injury. In a sham controlled, double-blind, parallel group design, 39 patients were randomized into four groups receiving transcranial direct current stimulation with walking visual illusion or with control illusion and sham stimulation with visual illusion or with control illusion. For transcranial direct current stimulation, the anode was placed over the primary motor cortex. Each patient received ten treatment sessions during two consecutive weeks. Clinical assessment was performed before, after the last day of treatment, after 2 and 4 weeks follow-up and after 12 weeks. Clinical assessment included overall pain intensity perception, Neuropathic Pain Symptom Inventory and Brief Pain Inventory. The combination of transcranial direct current stimulation and visual illusion reduced the intensity of neuropathic pain significantly more than any of the single interventions. Patients receiving transcranial direct current stimulation and visual illusion experienced a significant improvement in all pain subtypes, while patients in the transcranial direct current stimulation group showed improvement in continuous and paroxysmal pain, and those in the visual illusion group improved only in continuous pain and dysaesthesias. At 12 weeks after treatment, the combined treatment group still presented significant improvement on the overall pain intensity perception, whereas no improvements were reported in the other three groups. Our results demonstrate that transcranial direct current stimulation and visual illusion can be effective in the management of neuropathic pain following spinal cord injury, with minimal side effects and with good tolerability. PMID:20685806
Neural Mechanisms Underlying Cross-Modal Phonetic Encoding.
Shahin, Antoine J; Backer, Kristina C; Rosenblum, Lawrence D; Kerlin, Jess R
2018-02-14
Audiovisual (AV) integration is essential for speech comprehension, especially in adverse listening situations. Divergent, but not mutually exclusive, theories have been proposed to explain the neural mechanisms underlying AV integration. One theory advocates that this process occurs via interactions between the auditory and visual cortices, as opposed to fusion of AV percepts in a multisensory integrator. Building upon this idea, we proposed that AV integration in spoken language reflects visually induced weighting of phonetic representations at the auditory cortex. EEG was recorded while male and female human subjects watched and listened to videos of a speaker uttering consonant vowel (CV) syllables /ba/ and /fa/, presented in Auditory-only, AV congruent or incongruent contexts. Subjects reported whether they heard /ba/ or /fa/. We hypothesized that vision alters phonetic encoding by dynamically weighting which phonetic representation in the auditory cortex is strengthened or weakened. That is, when subjects are presented with visual /fa/ and acoustic /ba/ and hear /fa/ ( illusion-fa ), the visual input strengthens the weighting of the phone /f/ representation. When subjects are presented with visual /ba/ and acoustic /fa/ and hear /ba/ ( illusion-ba ), the visual input weakens the weighting of the phone /f/ representation. Indeed, we found an enlarged N1 auditory evoked potential when subjects perceived illusion-ba , and a reduced N1 when they perceived illusion-fa , mirroring the N1 behavior for /ba/ and /fa/ in Auditory-only settings. These effects were especially pronounced in individuals with more robust illusory perception. These findings provide evidence that visual speech modifies phonetic encoding at the auditory cortex. SIGNIFICANCE STATEMENT The current study presents evidence that audiovisual integration in spoken language occurs when one modality (vision) acts on representations of a second modality (audition). Using the McGurk illusion, we show that visual context primes phonetic representations at the auditory cortex, altering the auditory percept, evidenced by changes in the N1 auditory evoked potential. This finding reinforces the theory that audiovisual integration occurs via visual networks influencing phonetic representations in the auditory cortex. We believe that this will lead to the generation of new hypotheses regarding cross-modal mapping, particularly whether it occurs via direct or indirect routes (e.g., via a multisensory mediator). Copyright © 2018 the authors 0270-6474/18/381835-15$15.00/0.
Alterations to global but not local motion processing in long-term ecstasy (MDMA) users.
White, Claire; Brown, John; Edwards, Mark
2014-07-01
Growing evidence indicates that the main psychoactive ingredient in the illegal drug "ecstasy" (methylendioxymethamphetamine) causes reduced activity in the serotonin and gamma-aminobutyric acid (GABA) systems in humans. On the basis of substantial serotonin input to the occipital lobe, recent research investigated visual processing in long-term users and found a larger magnitude of the tilt aftereffect, interpreted to reflect broadened orientation tuning bandwidths. Further research found higher orientation discrimination thresholds and reduced long-range interactions in the primary visual area of ecstasy users. The aim of the present research was to investigate whether serotonin-mediated V1 visual processing deficits in ecstasy users extend to motion processing mechanisms. Forty-five participants (21 controls, 24 drug users) completed two psychophysical studies: A direction discrimination study directly measured local motion processing in V1, while a motion coherence task tested global motion processing in area V5/MT. "Primary" ecstasy users (n = 18), those without substantial polydrug use, had significantly lower global motion thresholds than controls [p = 0.027, Cohen's d = 0.78 (large)], indicating increased sensitivity to global motion stimuli, but no difference in local motion processing (p = 0.365). These results extend on previous research investigating the long-term effects of illicit drugs on visual processing. Two possible explanations are explored: defuse attentional processes may be facilitating spatial pooling of motion signals in users. Alternatively, it may be that a GABA-mediated disruption to V5/MT processing is reducing spatial suppression and therefore improving global motion perception in ecstasy users.
Improving the discrimination of hand motor imagery via virtual reality based visual guidance.
Liang, Shuang; Choi, Kup-Sze; Qin, Jing; Pang, Wai-Man; Wang, Qiong; Heng, Pheng-Ann
2016-08-01
While research on the brain-computer interface (BCI) has been active in recent years, how to get high-quality electrical brain signals to accurately recognize human intentions for reliable communication and interaction is still a challenging task. The evidence has shown that visually guided motor imagery (MI) can modulate sensorimotor electroencephalographic (EEG) rhythms in humans, but how to design and implement efficient visual guidance during MI in order to produce better event-related desynchronization (ERD) patterns is still unclear. The aim of this paper is to investigate the effect of using object-oriented movements in a virtual environment as visual guidance on the modulation of sensorimotor EEG rhythms generated by hand MI. To improve the classification accuracy on MI, we further propose an algorithm to automatically extract subject-specific optimal frequency and time bands for the discrimination of ERD patterns produced by left and right hand MI. The experimental results show that the average classification accuracy of object-directed scenarios is much better than that of non-object-directed scenarios (76.87% vs. 69.66%). The result of the t-test measuring the difference between them is statistically significant (p = 0.0207). When compared to algorithms based on fixed frequency and time bands, contralateral dominant ERD patterns can be enhanced by using the subject-specific optimal frequency and the time bands obtained by our proposed algorithm. These findings have the potential to improve the efficacy and robustness of MI-based BCI applications. Copyright © 2016 Elsevier Ireland Ltd. All rights reserved.
Does the vestibular system contribute to head direction cell activity in the rat?
NASA Technical Reports Server (NTRS)
Brown, J. E.; Yates, B. J.; Taube, J. S.; Oman, C. M. (Principal Investigator)
2002-01-01
Head direction cells (HDC) located in several regions of the brain, including the anterior dorsal nucleus of the thalamus (ADN), postsubiculum (PoS), and lateral mammillary nuclei (LMN), provide the neural substrate for the determination of head direction. Although activity of HDC is influenced by various sensory signals and internally generated cues, lesion studies and some anatomical and physiological evidence suggest that vestibular inputs are critical for the maintenance of directional sensitivity of these cells. However, vestibular inputs must be transformed considerably in order to signal head direction, and the neuronal circuitry that accomplishes this signal processing has not been fully established. Furthermore, it is unclear why the removal of vestibular inputs abolishes the directional sensitivity of HDC, as visual and other sensory inputs and motor feedback signals strongly affect the firing of these neurons and would be expected to maintain their directional-related activity. Further physiological studies will be required to establish the role of vestibular system in producing HDC responses, and anatomical studies are needed to determine the neural circuitry that mediates vestibular influences on determination of head direction.
NASA Technical Reports Server (NTRS)
DiZio, P.; Lackner, J. R.
2000-01-01
Reaching movements made to visual targets in a rotating room are initially deviated in path and endpoint in the direction of transient Coriolis forces generated by the motion of the arm relative to the rotating environment. With additional reaches, movements become progressively straighter and more accurate. Such adaptation can occur even in the absence of visual feedback about movement progression or terminus. Here we examined whether congenitally blind and sighted subjects without visual feedback would demonstrate adaptation to Coriolis forces when they pointed to a haptically specified target location. Subjects were tested pre-, per-, and postrotation at 10 rpm counterclockwise. Reaching to straight ahead targets prerotation, both groups exhibited slightly curved paths. Per-rotation, both groups showed large initial deviations of movement path and curvature but within 12 reaches on average had returned to prerotation curvature levels and endpoints. Postrotation, both groups showed mirror image patterns of curvature and endpoint to the per-rotation pattern. The groups did not differ significantly on any of the performance measures. These results provide compelling evidence that motor adaptation to Coriolis perturbations can be achieved on the basis of proprioceptive, somatosensory, and motor information in the complete absence of visual experience.
Thomas, Emily; Murphy, Mary; Pitt, Rebecca; Rivers, Angela; Leavens, David A
2008-11-01
Povinelli, Bierschwale, and Cech (1999) reported that when tested on a visual attention task, the behavior of juvenile chimpanzees did not support a high-level understanding of visual attention. This study replicates their research using adult humans and aims to investigate the validity of their experimental design. Participants were trained to respond to pointing cues given by an experimenter, and then tested on their ability to locate hidden objects from visual cues. Povinelli et al.'s assertion that the generalization of pointing to gaze is indicative of a high-level framework was not supported by our findings: Training improved performance only on initial probe trials when the experimenter's gaze was not directed at the baited cup. Furthermore, participants performed above chance on such trials, the same result exhibited by chimpanzees and used as evidence by Povinelli et al. to support a low-level framework. These findings, together with the high performance of participants in an incongruent condition, in which the experimenter pointed to or gazed at an unbaited container, challenge the validity of their experimental design. (PsycINFO Database Record (c) 2008 APA, all rights reserved).
Loffing, Florian; Schorer, Jörg; Hagemann, Norbert; Baker, Joseph
2012-02-01
High ball speeds and close distances between competitors require athletes in interactive sports to correctly anticipate an opponent's intentions in order to render appropriate reactions. Although it is considered crucial for successful performance, such skill appears impaired when athletes are confronted with a left-handed opponent, possibly because of athletes' reduced perceptual familiarity with rarely encountered left-handed actions. To test this negative perceptual frequency effect hypothesis, we invited 18 skilled and 18 novice volleyball players to predict shot directions of left- and right-handed attacks in a video-based visual anticipation task. In accordance with our predictions, and with recent reports on laterality differences in visual perception, the outcome of left-handed actions was significantly less accurately predicted than the outcome of right-handed attacks. In addition, this left-right bias was most distinct when predictions had to be based on preimpact (i.e., before hand-ball contact) kinematic cues, and skilled players were generally more affected by the opponents' handedness than were novices. The study's findings corroborate the assumption that skilled visual perception is attuned to more frequently encountered actions.
Lateral interactions in the outer retina
Thoreson, Wallace B.; Mangel, Stuart C.
2012-01-01
Lateral interactions in the outer retina, particularly negative feedback from horizontal cells to cones and direct feed-forward input from horizontal cells to bipolar cells, play a number of important roles in early visual processing, such as generating center-surround receptive fields that enhance spatial discrimination. These circuits may also contribute to post-receptoral light adaptation and the generation of color opponency. In this review, we examine the contributions of horizontal cell feedback and feed-forward pathways to early visual processing. We begin by reviewing the properties of bipolar cell receptive fields, especially with respect to modulation of the bipolar receptive field surround by the ambient light level and to the contribution of horizontal cells to the surround. We then review evidence for and against three proposed mechanisms for negative feedback from horizontal cells to cones: 1) GABA release by horizontal cells, 2) ephaptic modulation of the cone pedicle membrane potential generated by currents flowing through hemigap junctions in horizontal cell dendrites, and 3) modulation of cone calcium currents (ICa) by changes in synaptic cleft proton levels. We also consider evidence for the presence of direct horizontal cell feed-forward input to bipolar cells and discuss a possible role for GABA at this synapse. We summarize proposed functions of horizontal cell feedback and feed-forward pathways. Finally, we examine the mechanisms and functions of two other forms of lateral interaction in the outer retina: negative feedback from horizontal cells to rods and positive feedback from horizontal cells to cones. PMID:22580106
Digital holographic interferometry applied to the investigation of ignition process.
Pérez-Huerta, J S; Saucedo-Anaya, Tonatiuh; Moreno, I; Ariza-Flores, D; Saucedo-Orozco, B
2017-06-12
We use the digital holographic interferometry (DHI) technique to display the early ignition process for a butane-air mixture flame. Because such an event occurs in a short time (few milliseconds), a fast CCD camera is used to study the event. As more detail is required for monitoring the temporal evolution of the process, less light coming from the combustion is captured by the CCD camera, resulting in a deficient and underexposed image. Therefore, the CCD's direct observation of the combustion process is limited (down to 1000 frames per second). To overcome this drawback, we propose the use of DHI along with a high power laser in order to supply enough light to increase the speed capture, thus improving the visualization of the phenomenon in the initial moments. An experimental optical setup based on DHI is used to obtain a large sequence of phase maps that allows us to observe two transitory stages in the ignition process: a first explosion which slightly emits visible light, and a second stage induced by variations in temperature when the flame is emerging. While the last stage can be directly monitored by the CCD camera, the first stage is hardly detected by direct observation, and DHI clearly evidences this process. Furthermore, our method can be easily adapted for visualizing other types of fast processes.
Experience Report: Visual Programming in the Real World
NASA Technical Reports Server (NTRS)
Baroth, E.; Hartsough, C
1994-01-01
This paper reports direct experience with two commercial, widely used visual programming environments. While neither of these systems is object oriented, the tools have transformed the development process and indicate a direction for visual object oriented tools to proceed.
Saunders, Jeffrey A.
2014-01-01
Direction of self-motion during walking is indicated by multiple cues, including optic flow, nonvisual sensory cues, and motor prediction. I measured the reliability of perceived heading from visual and nonvisual cues during walking, and whether cues are weighted in an optimal manner. I used a heading alignment task to measure perceived heading during walking. Observers walked toward a target in a virtual environment with and without global optic flow. The target was simulated to be infinitely far away, so that it did not provide direct feedback about direction of self-motion. Variability in heading direction was low even without optic flow, with average RMS error of 2.4°. Global optic flow reduced variability to 1.9°–2.1°, depending on the structure of the environment. The small amount of variance reduction was consistent with optimal use of visual information. The relative contribution of visual and nonvisual information was also measured using cue conflict conditions. Optic flow specified a conflicting heading direction (±5°), and bias in walking direction was used to infer relative weighting. Visual feedback influenced heading direction by 16%–34% depending on scene structure, with more effect with dense motion parallax. The weighting of visual feedback was close to the predictions of an optimal integration model given the observed variability measures. PMID:24648194
The contribution of LM to the neuroscience of movement vision
Zihl, Josef; Heywood, Charles A.
2015-01-01
The significance of early and sporadic reports in the 19th century of impairments of motion vision following brain damage was largely unrecognized. In the absence of satisfactory post-mortem evidence, impairments were interpreted as the consequence of a more general disturbance resulting from brain damage, the location and extent of which was unknown. Moreover, evidence that movement constituted a special visual perception and may be selectively spared was similarly dismissed. Such skepticism derived from a reluctance to acknowledge that the neural substrates of visual perception may not be confined to primary visual cortex. This view did not persist. First, it was realized that visual movement perception does not depend simply on the analysis of spatial displacements and temporal intervals, but represents a specific visual movement sensation. Second persuasive evidence for functional specialization in extrastriate cortex, and notably the discovery of cortical area V5/MT, suggested a separate region specialized for motion processing. Shortly thereafter the remarkable case of patient LM was published, providing compelling evidence for a selective and specific loss of movement vision. The case is reviewed here, along with an assessment of its contribution to visual neuroscience. PMID:25741251
Whitwell, Robert L.; Ganel, Tzvi; Byrne, Caitlin M.; Goodale, Melvyn A.
2015-01-01
Investigators study the kinematics of grasping movements (prehension) under a variety of conditions to probe visuomotor function in normal and brain-damaged individuals. “Natural” prehensile acts are directed at the goal object and are executed using real-time vision. Typically, they also entail the use of tactile, proprioceptive, and kinesthetic sources of haptic feedback about the object (“haptics-based object information”) once contact with the object has been made. Natural and simulated (pantomimed) forms of prehension are thought to recruit different cortical structures: patient DF, who has visual form agnosia following bilateral damage to her temporal-occipital cortex, loses her ability to scale her grasp aperture to the size of targets (“grip scaling”) when her prehensile movements are based on a memory of a target previewed 2 s before the cue to respond or when her grasps are directed towards a visible virtual target but she is denied haptics-based information about the target. In the first of two experiments, we show that when DF performs real-time pantomimed grasps towards a 7.5 cm displaced imagined copy of a visible object such that her fingers make contact with the surface of the table, her grip scaling is in fact quite normal. This finding suggests that real-time vision and terminal tactile feedback are sufficient to preserve DF’s grip scaling slopes. In the second experiment, we examined an “unnatural” grasping task variant in which a tangible target (along with any proxy such as the surface of the table) is denied (i.e., no terminal tactile feedback). To do this, we used a mirror-apparatus to present virtual targets with and without a spatially coincident copy for the participants to grasp. We compared the grasp kinematics from trials with and without terminal tactile feedback to a real-time-pantomimed grasping task (one without tactile feedback) in which participants visualized a copy of the visible target as instructed in our laboratory in the past. Compared to natural grasps, removing tactile feedback increased RT, slowed the velocity of the reach, reduced in-flight grip aperture, increased the slopes relating grip aperture to target width, and reduced the final grip aperture (FGA). All of these effects were also observed in the real time-pantomime grasping task. These effects seem to be independent of those that arise from using the mirror in general as we also compared grasps directed towards virtual targets to those directed at real ones viewed directly through a pane of glass. These comparisons showed that the grasps directed at virtual targets increased grip aperture, slowed the velocity of the reach, and reduced the slopes relating grip aperture to the widths of the target. Thus, using the mirror has real consequences on grasp kinematics, reflecting the importance of task-relevant sources of online visual information for the programming and updating of natural prehensile movements. Taken together, these results provide compelling support for the view that removing terminal tactile feedback, even when the grasps are target-directed, induces a switch from real-time visual control towards one that depends more on visual perception and cognitive supervision. Providing terminal tactile feedback and real-time visual information can evidently keep the dorsal visuomotor system operating normally for prehensile acts. PMID:25999834
Whitwell, Robert L; Ganel, Tzvi; Byrne, Caitlin M; Goodale, Melvyn A
2015-01-01
Investigators study the kinematics of grasping movements (prehension) under a variety of conditions to probe visuomotor function in normal and brain-damaged individuals. "Natural" prehensile acts are directed at the goal object and are executed using real-time vision. Typically, they also entail the use of tactile, proprioceptive, and kinesthetic sources of haptic feedback about the object ("haptics-based object information") once contact with the object has been made. Natural and simulated (pantomimed) forms of prehension are thought to recruit different cortical structures: patient DF, who has visual form agnosia following bilateral damage to her temporal-occipital cortex, loses her ability to scale her grasp aperture to the size of targets ("grip scaling") when her prehensile movements are based on a memory of a target previewed 2 s before the cue to respond or when her grasps are directed towards a visible virtual target but she is denied haptics-based information about the target. In the first of two experiments, we show that when DF performs real-time pantomimed grasps towards a 7.5 cm displaced imagined copy of a visible object such that her fingers make contact with the surface of the table, her grip scaling is in fact quite normal. This finding suggests that real-time vision and terminal tactile feedback are sufficient to preserve DF's grip scaling slopes. In the second experiment, we examined an "unnatural" grasping task variant in which a tangible target (along with any proxy such as the surface of the table) is denied (i.e., no terminal tactile feedback). To do this, we used a mirror-apparatus to present virtual targets with and without a spatially coincident copy for the participants to grasp. We compared the grasp kinematics from trials with and without terminal tactile feedback to a real-time-pantomimed grasping task (one without tactile feedback) in which participants visualized a copy of the visible target as instructed in our laboratory in the past. Compared to natural grasps, removing tactile feedback increased RT, slowed the velocity of the reach, reduced in-flight grip aperture, increased the slopes relating grip aperture to target width, and reduced the final grip aperture (FGA). All of these effects were also observed in the real time-pantomime grasping task. These effects seem to be independent of those that arise from using the mirror in general as we also compared grasps directed towards virtual targets to those directed at real ones viewed directly through a pane of glass. These comparisons showed that the grasps directed at virtual targets increased grip aperture, slowed the velocity of the reach, and reduced the slopes relating grip aperture to the widths of the target. Thus, using the mirror has real consequences on grasp kinematics, reflecting the importance of task-relevant sources of online visual information for the programming and updating of natural prehensile movements. Taken together, these results provide compelling support for the view that removing terminal tactile feedback, even when the grasps are target-directed, induces a switch from real-time visual control towards one that depends more on visual perception and cognitive supervision. Providing terminal tactile feedback and real-time visual information can evidently keep the dorsal visuomotor system operating normally for prehensile acts.
Localized direction selective responses in the dendrites of visual interneurons of the fly
2010-01-01
Background The various tasks of visual systems, including course control, collision avoidance and the detection of small objects, require at the neuronal level the dendritic integration and subsequent processing of many spatially distributed visual motion inputs. While much is known about the pooled output in these systems, as in the medial superior temporal cortex of monkeys or in the lobula plate of the insect visual system, the motion tuning of the elements that provide the input has yet received little attention. In order to visualize the motion tuning of these inputs we examined the dendritic activation patterns of neurons that are selective for the characteristic patterns of wide-field motion, the lobula-plate tangential cells (LPTCs) of the blowfly. These neurons are known to sample direction-selective motion information from large parts of the visual field and combine these signals into axonal and dendro-dendritic outputs. Results Fluorescence imaging of intracellular calcium concentration allowed us to take a direct look at the local dendritic activity and the resulting local preferred directions in LPTC dendrites during activation by wide-field motion in different directions. These 'calcium response fields' resembled a retinotopic dendritic map of local preferred directions in the receptive field, the layout of which is a distinguishing feature of different LPTCs. Conclusions Our study reveals how neurons acquire selectivity for distinct visual motion patterns by dendritic integration of the local inputs with different preferred directions. With their spatial layout of directional responses, the dendrites of the LPTCs we investigated thus served as matched filters for wide-field motion patterns. PMID:20384983
Conservation implications of anthropogenic impacts on visual communication and camouflage.
Delhey, Kaspar; Peters, Anne
2017-02-01
Anthropogenic environmental impacts can disrupt the sensory environment of animals and affect important processes from mate choice to predator avoidance. Currently, these effects are best understood for auditory and chemosensory modalities, and recent reviews highlight their importance for conservation. We examined how anthropogenic changes to the visual environment (ambient light, transmission, and backgrounds) affect visual communication and camouflage and considered the implications of these effects for conservation. Human changes to the visual environment can increase predation risk by affecting camouflage effectiveness, lead to maladaptive patterns of mate choice, and disrupt mutualistic interactions between pollinators and plants. Implications for conservation are particularly evident for disrupted camouflage due to its tight links with survival. The conservation importance of impaired visual communication is less documented. The effects of anthropogenic changes on visual communication and camouflage may be severe when they affect critical processes such as pollination or species recognition. However, when impaired mate choice does not lead to hybridization, the conservation consequences are less clear. We suggest that the demographic effects of human impacts on visual communication and camouflage will be particularly strong when human-induced modifications to the visual environment are evolutionarily novel (i.e., very different from natural variation); affected species and populations have low levels of intraspecific (genotypic and phenotypic) variation and behavioral, sensory, or physiological plasticity; and the processes affected are directly related to survival (camouflage), species recognition, or number of offspring produced, rather than offspring quality or attractiveness. Our findings suggest that anthropogenic effects on the visual environment may be of similar importance relative to conservation as anthropogenic effects on other sensory modalities. © 2016 Society for Conservation Biology.
Sea-floor geology of a part of Mamala Bay, Hawaii
Hampton, Monty A.; Torresan, Michael E.; Barber, John H.
1997-01-01
We surveyed the sea-floor geology within a 200-km2 area of Mamala Bay, off Honolulu, Hawaii by collecting and analyzing sidescan sonar images, 3.5-kHz profiles, video and still visual images, and box-core samples. The study area extends from 20-m water depth on the insular shelf to 600-m water depth in a southeast-trending trough. The sidescan images depict three principal types of sea-floor material: low-backscatter natural sediment, high-backscatter drowned carbonate reef, and intermediate-backscatter dredged-material deposits. Cores indicate that the natural sediment is muddy sand, composed of carbonate reef and microfauna debris with some volcanic grains. Vague areal trends in composition are evident. The dredged material comprises poorly sorted, cobble- to clay-size mixtures of reef, volcanic, and man-made debris, up to 35 cm thick. Dredged-material deposits are not evident in the 3.5-kHz profiles. In the sidescan images they appear as isolated, circular to subcircular imprints, apparently formed by individual drops, around the periphery of their occurrence, but they overlap and coalesce to a nearly continuous, intermediate-backscatter blanket toward the center of three disposal sites investigated. We did not observe significant currents during our camera surveys, but there is abundant evidence of sediment reworking: symmetrical and asymmetrical ripples in the visual images, sand waves in the 3.5-kHz profiles and side-scan images, moats around the reefs in 3.5-kHz profiles, winnowed dredged material in the visual images, and burial of dredged material by natural sediment in cores. Most current indicators imply a westerly to northwesterly transport direction, along contours or up-slope, although there are a few areas of easterly indicators. Internal waves probably drive the transport; their possible existence is implied by measured water-column density gradients.
Sea-floor geology of a part of Mamala Bay, Hawai'i
Hampton, M.A.; Torresan, M.E.; Barber, J.H.
1997-01-01
We surveyed the sea-floor geology within a 200-km2 area of Mamala Bay, off Honolulu, Hawai'i, by collecting and analyzing sidescan sonar images, 3.5kHz profiles, video and still visual images, and box-core samples. The study area extends from 20-m water depth on the insular shelf to 600-m water depth in a southeast-trending trough. The sidescan images depict three principal types of seafloor material: low-backscatter natural sediment, high-backscatter drowned carbonate reef, and intermediate-backscatter dredged-material deposits. Cores indicate that the natural sediment is muddy sand, composed of carbonate reef and microfauna debris with some volcanic grains. Vague areal trends in composition are evident. The dredged material comprises poorly sorted, cobble- to clay-size mixtures of reef, volcanic, and man-made debris, up to 35 cm thick. Dredged-material deposits are not evident in the 3.5-kHz profiles. In the sidescan images they appear as isolated, circular to subcircular imprints, apparently formed by individual drops, around the periphery of their occurrence, but they overlap and coalesce to a nearly continuous, intermediate-backscatter blanket toward the center of three disposal sites investigated. We did not observe noticeable currents during our camera surveys, but there is abundant evidence of sediment reworking: symmetrical and asymmetrical ripples in the visual images, sand waves in the 3.5-kHz profiles and side-scan images, moats around the reefs in 3.5-kHz profiles, winnowed dredged material in the visual images, and burial of dredged material by natural sediment in cores. Most current indicators imply a westerly to northwesterly transport direction, along contours or upslope, although there are a few areas of easterly indicators. Internal waves probably drive the transport; their possible existence is implied by measured water-column density gradients.
Alpha Rhythms in Audition: Cognitive and Clinical Perspectives
Weisz, Nathan; Hartmann, Thomas; Müller, Nadia; Lorenz, Isabel; Obleser, Jonas
2011-01-01
Like the visual and the sensorimotor systems, the auditory system exhibits pronounced alpha-like resting oscillatory activity. Due to the relatively small spatial extent of auditory cortical areas, this rhythmic activity is less obvious and frequently masked by non-auditory alpha-generators when recording non-invasively using magnetoencephalography (MEG) or electroencephalography (EEG). Following stimulation with sounds, marked desynchronizations can be observed between 6 and 12 Hz, which can be localized to the auditory cortex. However knowledge about the functional relevance of the auditory alpha rhythm has remained scarce so far. Results from the visual and sensorimotor system have fuelled the hypothesis of alpha activity reflecting a state of functional inhibition. The current article pursues several intentions: (1) Firstly we review and present own evidence (MEG, EEG, sEEG) for the existence of an auditory alpha-like rhythm independent of visual or motor generators, something that is occasionally met with skepticism. (2) In a second part we will discuss tinnitus and how this audiological symptom may relate to reduced background alpha. The clinical part will give an introduction into a method which aims to modulate neurophysiological activity hypothesized to underlie this distressing disorder. Using neurofeedback, one is able to directly target relevant oscillatory activity. Preliminary data point to a high potential of this approach for treating tinnitus. (3) Finally, in a cognitive neuroscientific part we will show that auditory alpha is modulated by anticipation/expectations with and without auditory stimulation. We will also introduce ideas and initial evidence that alpha oscillations are involved in the most complex capability of the auditory system, namely speech perception. The evidence presented in this article corroborates findings from other modalities, indicating that alpha-like activity functionally has an universal inhibitory role across sensory modalities. PMID:21687444
Evidence to practice: treatment of anxiety in individuals with autism spectrum disorders
Lang, Russell; Mahoney, Richard; El Zein, Farah; Delaune, Elizabeth; Amidon, Megan
2011-01-01
Clinical question What treatment improves social interactions and reduces reports of anxiety symptoms in individuals with autism spectrum disorders (ASD) and a co-occurring anxiety disorder? Results Systematic reviews and randomized clinical trials suggest that cognitive behavior therapy in tandem with direct instruction of social skills using applied behavior analysis intervention components may be effective for treating anxiety in individuals with high functioning ASD. For individuals with ASD, an anxiety disorder, and an intellectual disability, systematic desensitization may be effective. Implementation Intervention should emphasize teaching social skills. Reinforcers (ie, rewards based upon the client’s interests) should be used to encourage participation in therapy. Treatment should incorporate visual aides and family involvement. Intervention components involving abstract concepts, visualization, and discussions of emotions are less useful given difficulties in abstract reasoning and communication inherent to ASD. PMID:21326652
God: Do I have your attention?
Colzato, Lorenza S; van Beest, Ilja; van den Wildenberg, Wery P M; Scorolli, Claudia; Dorchin, Shirley; Meiran, Nachshon; Borghi, Anna M; Hommel, Bernhard
2010-10-01
Religion is commonly defined as a set of rules, developed as part of a culture. Here we provide evidence that practice in following these rules systematically changes the way people attend to visual stimuli, as indicated by the individual sizes of the global precedence effect (better performance to global than to local features). We show that this effect is significantly reduced in Calvinism, a religion emphasizing individual responsibility, and increased in Catholicism and Judaism, religions emphasizing social solidarity. We also show that this effect is long-lasting (still affecting baptized atheists) and that its size systematically varies as a function of the amount and strictness of religious practices. These findings suggest that religious practice induces particular cognitive-control styles that induce chronic, directional biases in the control of visual attention. Copyright 2010 Elsevier B.V. All rights reserved.
NASA Technical Reports Server (NTRS)
Wehner, R.
1972-01-01
Experimental data, on the visual orientation of desert ants toward astromenotactic courses and horizon landmarks involving the cooperation of different direction finding systems, are given. Attempts were made to: (1) determine if the ants choose a compromise direction between astromenotactic angles and the direction toward horizon landmarks when both angles compete with each other or whether they decide alternatively; (2) analyze adaptations of the visual system to the special demands of direction finding by astromenotactic orientation or pattern recognition; and (3) determine parameters of visual learning behavior. Results show separate orientation mechanisms are responsible for the orientation of the ant toward astromenotactic angles and horizon landmarks. If both systems compete with each other, the ants switch over from one system to the other and do not perform a compromise direction.
Strauss, Soeren; Woodgate, Philip J W; Sami, Saber A; Heinke, Dietmar
2015-12-01
We present an extension of a neurobiologically inspired robotics model, termed CoRLEGO (Choice reaching with a LEGO arm robot). CoRLEGO models experimental evidence from choice reaching tasks (CRT). In a CRT participants are asked to rapidly reach and touch an item presented on the screen. These experiments show that non-target items can divert the reaching movement away from the ideal trajectory to the target item. This is seen as evidence attentional selection of reaching targets can leak into the motor system. Using competitive target selection and topological representations of motor parameters (dynamic neural fields) CoRLEGO is able to mimic this leakage effect. Furthermore if the reaching target is determined by its colour oddity (i.e. a green square among red squares or vice versa), the reaching trajectories become straighter with repetitions of the target colour (colour streaks). This colour priming effect can also be modelled with CoRLEGO. The paper also presents an extension of CoRLEGO. This extension mimics findings that transcranial direct current stimulation (tDCS) over the motor cortex modulates the colour priming effect (Woodgate et al., 2015). The results with the new CoRLEGO suggest that feedback connections from the motor system to the brain's attentional system (parietal cortex) guide visual attention to extract movement-relevant information (i.e. colour) from visual stimuli. This paper adds to growing evidence that there is a close interaction between the motor system and the attention system. This evidence contradicts the traditional conceptualization of the motor system as the endpoint of a serial chain of processing stages. At the end of the paper we discuss CoRLEGO's predictions and also lessons for neurobiologically inspired robotics emerging from this work. Crown Copyright © 2015. Published by Elsevier Ltd. All rights reserved.
Evidence-based guidelines on the referral of visually impaired persons to low vision services.
De Boer, M R; Langelaan, M; Jansonius, N M; Van Rens, G H M B
2005-01-01
One to two percent of the population in the Western world is visually impaired or blind. For most of these people there is no curative therapy. Therefore, the Dutch Ophthalmic Society has taken the initiative to develop an evidence-based guideline for the referral of visually impaired persons to low vision services. A systematic literature search was performed in the Embase (1991-2001) and Medline (1966-2003) databases. Literature was searched for definitions of visual impairment, for physician-patient communication, and for outcome of interventions for visually impaired persons. Results of the articles that were selected were summarized and rated according to the level of evidence. Other considerations such as the current organization of rehabilitation for visually impaired persons in the Netherlands were also taken into account. The World Health Organization criteria were slightly adapted in order to include all people who experience problems with reading and other daily life activities due to visual impairment. A large number of recommendations were devised. Among these is that the complete diagnosis should be communicated to the patient and that a second appointment should be offered in which the diagnosis and potential treatment options are discussed again. Another recommendation is that in general visually impaired adults eligible for referral should be referred for the provision of low vision aids and that patients with complex problems or extensive rehabilitative demand should be referred to a rehabilitation center. This article presents a summary of the first European evidence-based guideline for the referral of visually impaired persons.
Prpic, Valter; Fumarola, Antonia; De Tommaso, Matteo; Luccio, Riccardo; Murgia, Mauro; Agostini, Tiziano
2016-08-01
The spatial-numerical association of response codes (SNARC) effect is considered an evidence of the association between numbers and space, with faster left key-press responses to small numbers and faster right key-press responses to large numbers. We examined whether visually presented note values produce a SNARC-like effect. Differently from numbers, note values are represented as a decreasing left-to-right progression, allowing us to disambiguate the contribution of order and magnitude in determining the direction of the effect. Musicians with formal education performed a note value comparison in Experiment 1 (direct task), a line orientation judgment in Experiment 2 (indirect task), and a detection task in Experiment 3 (indirect task). When note values were task relevant (direct task), participants responded faster to large note values with the left key-press, and vice versa. Conversely, when note values were task irrelevant (indirect tasks), the direction of this association was reversed. This evidence suggests the existence of separate mechanisms underlying the SNARC effect. Namely, an Order-Related Mechanism (ORM) and a Magnitude-Related Mechanism (MRM) that are revealed by different task demands. Indeed, according to a new model we proposed, ordinal and magnitude related information appears to be preferentially involved in direct and indirect tasks, respectively. (PsycINFO Database Record (c) 2016 APA, all rights reserved).
Dawson, Debra Ann; Lam, Jack; Lewis, Lindsay B.; Carbonell, Felix; Mendola, Janine D.
2016-01-01
Abstract Numerous studies have demonstrated functional magnetic resonance imaging (fMRI)-based resting-state functional connectivity (RSFC) between cortical areas. Recent evidence suggests that synchronous fluctuations in blood oxygenation level-dependent fMRI reflect functional organization at a scale finer than that of visual areas. In this study, we investigated whether RSFCs within and between lower visual areas are retinotopically organized and whether retinotopically organized RSFC merely reflects cortical distance. Subjects underwent retinotopic mapping and separately resting-state fMRI. Visual areas V1, V2, and V3, were subdivided into regions of interest (ROIs) according to quadrants and visual field eccentricity. Functional connectivity (FC) was computed based on Pearson's linear correlation (correlation), and Pearson's linear partial correlation (correlation between two time courses after the time courses from all other regions in the network are regressed out). Within a quadrant, within visual areas, all correlation and nearly all partial correlation FC measures showed statistical significance. Consistently in V1, V2, and to a lesser extent in V3, correlation decreased with increasing eccentricity separation. Consistent with previously reported monkey anatomical connectivity, correlation/partial correlation values between regions from adjacent areas (V1-V2 and V2-V3) were higher than those between nonadjacent areas (V1-V3). Within a quadrant, partial correlation showed consistent significance between regions from two different areas with the same or adjacent eccentricities. Pairs of ROIs with similar eccentricity showed higher correlation/partial correlation than pairs distant in eccentricity. Between dorsal and ventral quadrants, partial correlation between common and adjacent eccentricity regions within a visual area showed statistical significance; this extended to more distant eccentricity regions in V1. Within and between quadrants, correlation decreased approximately linearly with increasing distances separating the tested ROIs. Partial correlation showed a more complex dependence on cortical distance: it decreased exponentially with increasing distance within a quadrant, but was best fit by a quadratic function between quadrants. We conclude that RSFCs within and between lower visual areas are retinotopically organized. Correlation-based FC is nonselectively high across lower visual areas, even between regions that do not share direct anatomical connections. The mechanisms likely involve network effects caused by the dense anatomical connectivity within this network and projections from higher visual areas. FC based on partial correlation, which minimizes network effects, follows expectations based on direct anatomical connections in the monkey visual cortex better than correlation. Last, partial correlation-based retinotopically organized RSFC reflects more than cortical distance effects. PMID:26415043
Dawson, Debra Ann; Lam, Jack; Lewis, Lindsay B; Carbonell, Felix; Mendola, Janine D; Shmuel, Amir
2016-02-01
Numerous studies have demonstrated functional magnetic resonance imaging (fMRI)-based resting-state functional connectivity (RSFC) between cortical areas. Recent evidence suggests that synchronous fluctuations in blood oxygenation level-dependent fMRI reflect functional organization at a scale finer than that of visual areas. In this study, we investigated whether RSFCs within and between lower visual areas are retinotopically organized and whether retinotopically organized RSFC merely reflects cortical distance. Subjects underwent retinotopic mapping and separately resting-state fMRI. Visual areas V1, V2, and V3, were subdivided into regions of interest (ROIs) according to quadrants and visual field eccentricity. Functional connectivity (FC) was computed based on Pearson's linear correlation (correlation), and Pearson's linear partial correlation (correlation between two time courses after the time courses from all other regions in the network are regressed out). Within a quadrant, within visual areas, all correlation and nearly all partial correlation FC measures showed statistical significance. Consistently in V1, V2, and to a lesser extent in V3, correlation decreased with increasing eccentricity separation. Consistent with previously reported monkey anatomical connectivity, correlation/partial correlation values between regions from adjacent areas (V1-V2 and V2-V3) were higher than those between nonadjacent areas (V1-V3). Within a quadrant, partial correlation showed consistent significance between regions from two different areas with the same or adjacent eccentricities. Pairs of ROIs with similar eccentricity showed higher correlation/partial correlation than pairs distant in eccentricity. Between dorsal and ventral quadrants, partial correlation between common and adjacent eccentricity regions within a visual area showed statistical significance; this extended to more distant eccentricity regions in V1. Within and between quadrants, correlation decreased approximately linearly with increasing distances separating the tested ROIs. Partial correlation showed a more complex dependence on cortical distance: it decreased exponentially with increasing distance within a quadrant, but was best fit by a quadratic function between quadrants. We conclude that RSFCs within and between lower visual areas are retinotopically organized. Correlation-based FC is nonselectively high across lower visual areas, even between regions that do not share direct anatomical connections. The mechanisms likely involve network effects caused by the dense anatomical connectivity within this network and projections from higher visual areas. FC based on partial correlation, which minimizes network effects, follows expectations based on direct anatomical connections in the monkey visual cortex better than correlation. Last, partial correlation-based retinotopically organized RSFC reflects more than cortical distance effects.
Kiani, Roozbeh; Hanks, Timothy D; Shadlen, Michael N
2008-03-19
Decisions about sensory stimuli are often based on an accumulation of evidence in time. When subjects control stimulus duration, the decision terminates when the accumulated evidence reaches a criterion level. Under many natural circumstances and in many laboratory settings, the environment, rather than the subject, controls the stimulus duration. In these settings, it is generally assumed that subjects commit to a choice at the end of the stimulus stream. Indeed, failure to benefit from the full stream of information is interpreted as a sign of imperfect accumulation or memory leak. Contrary to these assumptions, we show that monkeys performing a direction discrimination task commit to a choice when the accumulated evidence reaches a threshold level (or bound), sometimes long before the end of stimulus. This bounded accumulation of evidence is reflected in the activity of neurons in the lateral intraparietal cortex. Thus, the readout of visual cortex embraces a termination rule to limit processing even when potentially useful information is available.
The Pleasantness of Visual Symmetry: Always, Never or Sometimes
Pecchinenda, Anna; Bertamini, Marco; Makin, Alexis David James; Ruta, Nicole
2014-01-01
There is evidence of a preference for visual symmetry. This is true from mate selection in the animal world to the aesthetic appreciation of works of art. It has been proposed that this preference is due to processing fluency, which engenders positive affect. But is visual symmetry pleasant? Evidence is mixed as explicit preferences show that this is the case. In contrast, implicit measures show that visual symmetry does not spontaneously engender positive affect but it depends on participants intentionally assessing visual regularities. In four experiments using variants of the affective priming paradigm, we investigated when visual symmetry engenders positive affect. Findings showed that, when no Stroop-like effects or post-lexical mechanisms enter into play, visual symmetry spontaneously elicits positive affect and results in affective congruence effects. PMID:24658112
Wang, Yan Mei; Li, Ting; Li, Lin
2017-07-19
The valence-arousal conflict theory assumes that both valence and arousal will trigger approaching or withdrawing tendencies. It also predicts that the speed of processing emotional stimuli will depend on whether valence and arousal trigger conflicting or congruent motivational tendencies. However, most previous studies have provided evidence of the interaction between valence and arousal only, and have not provided direct proof of the interactive links between valence, arousal and motivational tendencies. The present study provides direct evidence for the relationship between approach-withdrawal tendencies and the valence-arousal conflict. In an empirical test, participants were instructed to judge the valence of emotional words after visual-spatial cues that appeared to be either approaching or withdrawing from participants. A three-way interaction (valence, arousal, and approach-withdrawal tendency) was observed such that the response time was shorter if participants responded to a negative high-arousal stimulus after a withdrawing cue, or to a positive low-arousal stimulus after an approaching cue. These findings suggest that the approach-withdrawal tendency indeed plays a crucial role in valence-arousal conflict, and that the effect depends on the congruency of valence, arousal and tendency at an early stage of processing.
Sequential sampling of visual objects during sustained attention.
Jia, Jianrong; Liu, Ling; Fang, Fang; Luo, Huan
2017-06-01
In a crowded visual scene, attention must be distributed efficiently and flexibly over time and space to accommodate different contexts. It is well established that selective attention enhances the corresponding neural responses, presumably implying that attention would persistently dwell on the task-relevant item. Meanwhile, recent studies, mostly in divided attentional contexts, suggest that attention does not remain stationary but samples objects alternately over time, suggesting a rhythmic view of attention. However, it remains unknown whether the dynamic mechanism essentially mediates attentional processes at a general level. Importantly, there is also a complete lack of direct neural evidence reflecting whether and how the brain rhythmically samples multiple visual objects during stimulus processing. To address these issues, in this study, we employed electroencephalography (EEG) and a temporal response function (TRF) approach, which can dissociate responses that exclusively represent a single object from the overall neuronal activity, to examine the spatiotemporal characteristics of attention in various attentional contexts. First, attention, which is characterized by inhibitory alpha-band (approximately 10 Hz) activity in TRFs, switches between attended and unattended objects every approximately 200 ms, suggesting a sequential sampling even when attention is required to mostly stay on the attended object. Second, the attentional spatiotemporal pattern is modulated by the task context, such that alpha-mediated switching becomes increasingly prominent as the task requires a more uniform distribution of attention. Finally, the switching pattern correlates with attentional behavioral performance. Our work provides direct neural evidence supporting a generally central role of temporal organization mechanism in attention, such that multiple objects are sequentially sorted according to their priority in attentional contexts. The results suggest that selective attention, in addition to the classically posited attentional "focus," involves a dynamic mechanism for monitoring all objects outside of the focus. Our findings also suggest that attention implements a space (object)-to-time transformation by acting as a series of concatenating attentional chunks that operate on 1 object at a time.
Sequential sampling of visual objects during sustained attention
Jia, Jianrong; Liu, Ling; Fang, Fang
2017-01-01
In a crowded visual scene, attention must be distributed efficiently and flexibly over time and space to accommodate different contexts. It is well established that selective attention enhances the corresponding neural responses, presumably implying that attention would persistently dwell on the task-relevant item. Meanwhile, recent studies, mostly in divided attentional contexts, suggest that attention does not remain stationary but samples objects alternately over time, suggesting a rhythmic view of attention. However, it remains unknown whether the dynamic mechanism essentially mediates attentional processes at a general level. Importantly, there is also a complete lack of direct neural evidence reflecting whether and how the brain rhythmically samples multiple visual objects during stimulus processing. To address these issues, in this study, we employed electroencephalography (EEG) and a temporal response function (TRF) approach, which can dissociate responses that exclusively represent a single object from the overall neuronal activity, to examine the spatiotemporal characteristics of attention in various attentional contexts. First, attention, which is characterized by inhibitory alpha-band (approximately 10 Hz) activity in TRFs, switches between attended and unattended objects every approximately 200 ms, suggesting a sequential sampling even when attention is required to mostly stay on the attended object. Second, the attentional spatiotemporal pattern is modulated by the task context, such that alpha-mediated switching becomes increasingly prominent as the task requires a more uniform distribution of attention. Finally, the switching pattern correlates with attentional behavioral performance. Our work provides direct neural evidence supporting a generally central role of temporal organization mechanism in attention, such that multiple objects are sequentially sorted according to their priority in attentional contexts. The results suggest that selective attention, in addition to the classically posited attentional “focus,” involves a dynamic mechanism for monitoring all objects outside of the focus. Our findings also suggest that attention implements a space (object)-to-time transformation by acting as a series of concatenating attentional chunks that operate on 1 object at a time. PMID:28658261
47 CFR 80.293 - Check bearings by authorized ship personnel.
Code of Federal Regulations, 2010 CFR
2010-10-01
....293 Section 80.293 Telecommunication FEDERAL COMMUNICATIONS COMMISSION (CONTINUED) SAFETY AND SPECIAL... comparison of simultaneous visual and radio direction finder bearings. At least one comparison bearing must... visual bearing relative to the ship's heading and the difference between the visual and radio direction...
Evolution of community structure in the world trade web
NASA Astrophysics Data System (ADS)
Tzekina, I.; Danthi, K.; Rockmore, D. N.
2008-06-01
In this note we study the bilateral merchandise trade flows between 186 countries over the 1948 2005 period using data from the International Monetary Fund. We use the network visualization package Pajek to identify network structure and behavior across thresholds and over time. In particular, we focus on the evolution of trade “islands” in a world trade network in which countries are linked with directed edges weighted according to the fraction of total dollars sent from one country to another. We find mixed evidence for globalization.
Wavelength dependence of the bidirectional reflectance distribution function (BRDF) of beach sands.
Doctor, Katarina Z; Bachmann, Charles M; Gray, Deric J; Montes, Marcos J; Fusina, Robert A
2015-11-01
The wavelength dependence of the dominant directional reflective properties of beach sands was demonstrated using principal component analysis and the related correlation matrix. In general, we found that the hyperspectral bidirectional reflectance distribution function (BRDF) of beach sands has weak wavelength dependence. Its BRDF varies slightly in three broad wavelength regions. The variations are more evident in surfaces of greater visual roughness than in smooth surfaces. The weak wavelength dependence of the BRDF of beach sand can be captured using three broad wavelength regions instead of hundreds of individual wavelengths.
How do schizophrenia patients use visual information to decode facial emotion?
Lee, Junghee; Gosselin, Frédéric; Wynn, Jonathan K; Green, Michael F
2011-09-01
Impairment in recognizing facial emotions is a prominent feature of schizophrenia patients, but the underlying mechanism of this impairment remains unclear. This study investigated the specific aspects of visual information that are critical for schizophrenia patients to recognize emotional expression. Using the Bubbles technique, we probed the use of visual information during a facial emotion discrimination task (fear vs. happy) in 21 schizophrenia patients and 17 healthy controls. Visual information was sampled through randomly located Gaussian apertures (or "bubbles") at 5 spatial frequency scales. Online calibration of the amount of face exposed through bubbles was used to ensure 75% overall accuracy for each subject. Least-square multiple linear regression analyses between sampled information and accuracy were performed to identify critical visual information that was used to identify emotional expression. To accurately identify emotional expression, schizophrenia patients required more exposure of facial areas (i.e., more bubbles) compared with healthy controls. To identify fearful faces, schizophrenia patients relied less on bilateral eye regions at high-spatial frequency compared with healthy controls. For identification of happy faces, schizophrenia patients relied on the mouth and eye regions; healthy controls did not utilize eyes and used the mouth much less than patients did. Schizophrenia patients needed more facial information to recognize emotional expression of faces. In addition, patients differed from controls in their use of high-spatial frequency information from eye regions to identify fearful faces. This study provides direct evidence that schizophrenia patients employ an atypical strategy of using visual information to recognize emotional faces.
Transcranial electrical stimulation over visual cortex evokes phosphenes with a retinal origin.
Kar, Kohitij; Krekelberg, Bart
2012-10-01
Transcranial electrical stimulation (tES) is a promising therapeutic tool for a range of neurological diseases. Understanding how the small currents used in tES spread across the scalp and penetrate the brain will be important for the rational design of tES therapies. Alternating currents applied transcranially above visual cortex induce the perception of flashes of light (phosphenes). This makes the visual system a useful model to study tES. One hypothesis is that tES generates phosphenes by direct stimulation of the cortex underneath the transcranial electrode. Here, we provide evidence for the alternative hypothesis that phosphenes are generated in the retina by current spread from the occipital electrode. Building on the existing literature, we first confirm that phosphenes are induced at lower currents when electrodes are placed farther away from visual cortex and closer to the eye. Second, we explain the temporal frequency tuning of phosphenes based on the well-known response properties of primate retinal ganglion cells. Third, we show that there is no difference in the time it takes to evoke phosphenes in the retina or by stimulation above visual cortex. Together, these findings suggest that phosphenes induced by tES over visual cortex originate in the retina. From this, we infer that tES currents spread well beyond the area of stimulation and are unlikely to lead to focal neural activation. Novel stimulation protocols that optimize current distributions are needed to overcome these limitations of tES.
LaBonte, Michelle Lynne
2017-11-01
In 1971, Günter Blobel and David Sabatini proposed a novel and quite speculative schematic model to describe how proteins might reach the proper cellular location. According to their proposal, proteins destined to be secreted from the cell contain a "signal" to direct their release. Despite the fact that Blobel and Sabatini presented their signal hypothesis as a "beautiful idea" not grounded in experimental evidence, they received criticism from other scientists who opposed such speculation. Following the publication of the 1971 model, Blobel persisted in conducting experiments and revising the model to incorporate new data. In fact, over the period of 1975-1984, Blobel and colleagues published five subsequent schematic models of the signal hypothesis, each revised based on new laboratory evidence. I propose that the original 1971 model can be viewed as an epistemic creation. Additionally, analysis of the subsequent schematic diagrams over the period of 1975-1984 allows one to track Blobel's changing conception of an epistemic object over time. Furthermore, the entire series of schematic diagrams presented by Blobel from 1971 to 1984 allow one to visualize the initial conception and subsequent reworking of a scientific theory. In 1999, Blobel was awarded the Nobel Prize in Physiology or Medicine for his work on the signal hypothesis, which was ultimately supported by experimental evidence gathered after the speculative model was published.
37 CFR 202.3 - Registration of copyright.
Code of Federal Regulations, 2014 CFR
2014-07-01
...) Class VA: Works of the visual arts. This class includes all published and unpublished pictorial, graphic... permission and under the direction of the Visual Arts Division, the application may be submitted... published photographs after consultation and with the permission and under the direction of the Visual Arts...
Qureshi, Adam W; Apperly, Ian A; Samson, Dana
2010-11-01
Previous research suggests that perspective-taking and other "theory of mind" processes may be cognitively demanding for adult participants, and may be disrupted by concurrent performance of a secondary task. In the current study, a Level-1 visual perspective task was administered to 32 adults using a dual-task paradigm in which the secondary task tapped executive function. Results suggested that the secondary task did not affect the calculation of perspective, but did affect the selection of the relevant (Self or Other) perspective for a given trial. This is the first direct evidence of a cognitively efficient process for "theory of mind" in adults that operates independently of executive function. The contrast between this and previous findings points to a distinction between simple perspective-taking and the more complex and cognitively demanding abilities more typically examined in studies of "theory of mind". It is suggested that these findings may provide a parsimonious explanation of the success of infants on 'indirect' measures of perspective-taking that do not explicitly require selection of the relevant perspective. Copyright © 2010 Elsevier B.V. All rights reserved.
Intracranial Cortical Responses during Visual–Tactile Integration in Humans
Quinn, Brian T.; Carlson, Chad; Doyle, Werner; Cash, Sydney S.; Devinsky, Orrin; Spence, Charles; Halgren, Eric
2014-01-01
Sensory integration of touch and sight is crucial to perceiving and navigating the environment. While recent evidence from other sensory modality combinations suggests that low-level sensory areas integrate multisensory information at early processing stages, little is known about how the brain combines visual and tactile information. We investigated the dynamics of multisensory integration between vision and touch using the high spatial and temporal resolution of intracranial electrocorticography in humans. We present a novel, two-step metric for defining multisensory integration. The first step compares the sum of the unisensory responses to the bimodal response as multisensory responses. The second step eliminates the possibility that double addition of sensory responses could be misinterpreted as interactions. Using these criteria, averaged local field potentials and high-gamma-band power demonstrate a functional processing cascade whereby sensory integration occurs late, both anatomically and temporally, in the temporo–parieto–occipital junction (TPOJ) and dorsolateral prefrontal cortex. Results further suggest two neurophysiologically distinct and temporally separated integration mechanisms in TPOJ, while providing direct evidence for local suppression as a dominant mechanism for synthesizing visual and tactile input. These results tend to support earlier concepts of multisensory integration as relatively late and centered in tertiary multimodal association cortices. PMID:24381279
The psychosis spectrum in Parkinson disease
ffytche, Dominic H.; Creese, Byron; Politis, Marios; Chaudhuri, K. Ray; Weintraub, Daniel; Ballard, Clive; Aarsland, Dag
2017-01-01
In 2007, the clinical and research profile of illusions, hallucinations, delusions and related symptoms in Parkinson disease (PD) was raised with the publication of a consensus definition of PD psychosis. Symptoms that were previously deemed benign and clinically insignificant were incorporated into a continuum of severity, leading to the rapid expansion of literature focusing on clinical aspects, mechanisms and treatment. Here, we review this literature and the evolving view of PD psychosis. Key topics include the prospective risk of dementia in individuals with PD psychosis, and the causal and modifying effects of PD medication. We discuss recent developments, including recognition of an increase in the prevalence of psychosis with disease duration, addition of new visual symptoms to the psychosis continuum, and identification of frontal executive, visual perceptual and memory dysfunction at different disease stages. In addition, we highlight novel risk factors — for example, autonomic dysfunction — that have emerged from prospective studies, structural MRI evidence of frontal, parietal, occipital and hippocampal involvement, and approval of pimavanserin for the treatment of PD psychosis. The accumulating evidence raises novel questions and directions for future research to explore the clinical management and biomarker potential of PD psychosis. PMID:28106066
Fries, Pascal; Womelsdorf, Thilo; Oostenveld, Robert; Desimone, Robert
2008-04-30
Selective attention lends relevant sensory input priority access to higher-level brain areas and ultimately to behavior. Recent studies have suggested that those neurons in visual areas that are activated by an attended stimulus engage in enhanced gamma-band (30-70 Hz) synchronization compared with neurons activated by a distracter. Such precise synchronization could enhance the postsynaptic impact of cells carrying behaviorally relevant information. Previous studies have used the local field potential (LFP) power spectrum or spike-LFP coherence (SFC) to indirectly estimate spike synchronization. Here, we directly demonstrate zero-phase gamma-band coherence among spike trains of V4 neurons. This synchronization was particularly evident during visual stimulation and enhanced by selective attention, thus confirming the pattern inferred from LFP power and SFC. We therefore investigated the time course of LFP gamma-band power and found rapid dynamics consistent with interactions of top-down spatial and feature attention with bottom-up saliency. In addition to the modulation of synchronization during visual stimulation, selective attention significantly changed the prestimulus pattern of synchronization. Attention inside the receptive field of the recorded neuronal population enhanced gamma-band synchronization and strongly reduced alpha-band (9-11 Hz) synchronization in the prestimulus period. These results lend further support for a functional role of rhythmic neuronal synchronization in attentional stimulus selection.
Joint attention enhances visual working memory.
Gregory, Samantha E A; Jackson, Margaret C
2017-02-01
Joint attention-the mutual focus of 2 individuals on an item-speeds detection and discrimination of target information. However, what happens to that information beyond the initial perceptual episode? To fully comprehend and engage with our immediate environment also requires working memory (WM), which integrates information from second to second to create a coherent and fluid picture of our world. Yet, no research exists at present that examines how joint attention directly impacts WM. To investigate this, we created a unique paradigm that combines gaze cues with a traditional visual WM task. A central, direct gaze 'cue' face looked left or right, followed 500 ms later by 4, 6, or 8 colored squares presented on one side of the face for encoding. Crucially, the cue face either looked at the squares (valid cue) or looked away from them (invalid cue). A no shift (direct gaze) condition served as a baseline. After a blank 1,000 ms maintenance interval, participants stated whether a single test square color was present or not in the preceding display. WM accuracy was significantly greater for colors encoded in the valid versus invalid and direct conditions. Further experiments showed that an arrow cue and a low-level motion cue-both shown to reliably orient attention-did not reliably modulate WM, indicating that social cues are more powerful. This study provides the first direct evidence that sharing the focus of another individual establishes a point of reference from which information is advantageously encoded into WM. (PsycINFO Database Record (c) 2017 APA, all rights reserved).
Sieger, Tomáš; Serranová, Tereza; Růžička, Filip; Vostatek, Pavel; Wild, Jiří; Šťastná, Daniela; Bonnet, Cecilia; Novák, Daniel; Růžička, Evžen; Urgošík, Dušan; Jech, Robert
2015-01-01
Both animal studies and studies using deep brain stimulation in humans have demonstrated the involvement of the subthalamic nucleus (STN) in motivational and emotional processes; however, participation of this nucleus in processing human emotion has not been investigated directly at the single-neuron level. We analyzed the relationship between the neuronal firing from intraoperative microrecordings from the STN during affective picture presentation in patients with Parkinson’s disease (PD) and the affective ratings of emotional valence and arousal performed subsequently. We observed that 17% of neurons responded to emotional valence and arousal of visual stimuli according to individual ratings. The activity of some neurons was related to emotional valence, whereas different neurons responded to arousal. In addition, 14% of neurons responded to visual stimuli. Our results suggest the existence of neurons involved in processing or transmission of visual and emotional information in the human STN, and provide evidence of separate processing of the affective dimensions of valence and arousal at the level of single neurons as well. PMID:25713375
Milner, A D; Paulignan, Y; Dijkerman, H C; Michel, F; Jeannerod, M
1999-11-07
We tested a patient (A. T.) with bilateral brain damage to the parietal lobes, whose resulting 'optic ataxia' causes her to make large pointing errors when asked to locate single light emitting diodes presented in her visual field. We report here that, unlike normal individuals, A. T.'s pointing accuracy improved when she was required to wait for 5 s before responding. This counter-intuitive result is interpreted as reflecting the very brief time-scale on which visuomotor control systems in the superior parietal lobe operate. When an immediate response was required, A. T.'s damaged visuomotor system caused her to make large errors; but when a delay was required, a different, more flexible, visuospatial coding system--presumably relatively intact in her brain--came into play, resulting in much more accurate responses. The data are consistent with a dual processing theory whereby motor responses made directly to visual stimuli are guided by a dedicated system in the superior parietal and premotor cortices, while responses to remembered stimuli depend on perceptual processing and may thus crucially involve processing within the temporal neocortex.
The Pharmacological Effects of Lutein and Zeaxanthin on Visual Disorders and Cognition Diseases.
Jia, Yu-Ping; Sun, Lei; Yu, He-Shui; Liang, Li-Peng; Li, Wei; Ding, Hui; Song, Xin-Bo; Zhang, Li-Juan
2017-04-20
Lutein (L) and zeaxanthin (Z) are dietary carotenoids derived from dark green leafy vegetables, orange and yellow fruits that form the macular pigment of the human eyes. It was hypothesized that they protect against visual disorders and cognition diseases, such as age-related macular degeneration (AMD), age-related cataract (ARC), cognition diseases, ischemic/hypoxia induced retinopathy, light damage of the retina, retinitis pigmentosa, retinal detachment, uveitis and diabetic retinopathy. The mechanism by which they are involved in the prevention of eye diseases may be due their physical blue light filtration properties and local antioxidant activity. In addition to their protective roles against light-induced oxidative damage, there are increasing evidences that L and Z may also improve normal ocular function by enhancing contrast sensitivity and by reducing glare disability. Surveys about L and Z supplementation have indicated that moderate intakes of L and Z are associated with decreased AMD risk and less visual impairment. Furthermore, this review discusses the appropriate consumption quantities, the consumption safety of L, side effects and future research directions.
Right-hemispheric dominance for visual remapping in humans.
Pisella, L; Alahyane, N; Blangero, A; Thery, F; Blanc, S; Pelisson, D
2011-02-27
We review evidence showing a right-hemispheric dominance for visuo-spatial processing and representation in humans. Accordingly, visual disorganization symptoms (intuitively related to remapping impairments) are observed in both neglect and constructional apraxia. More specifically, we review findings from the intervening saccade paradigm in humans--and present additional original data--which suggest a specific role of the asymmetrical network at the temporo-parietal junction (TPJ) in the right hemisphere in visual remapping: following damage to the right dorsal posterior parietal cortex (PPC) as well as part of the corpus callosum connecting the PPC to the frontal lobes, patient OK in a double-step saccadic task exhibited an impairment when the second saccade had to be directed rightward. This singular and lateralized deficit cannot result solely from the patient's cortical lesion and, therefore, we propose that it is due to his callosal lesion that may specifically interrupt the interhemispheric transfer of information necessary to execute accurate rightward saccades towards a remapped target location. This suggests a specialized right-hemispheric network for visuo-spatial remapping that subsequently transfers target location information to downstream planning regions, which are symmetrically organized.
Tebbutt, G; Bell, V; Aislabie, J
2007-04-01
The aim of this study was to determine whether or not the assessment of surface cleanliness could make a contribution to visual inspections of food premises. Forty-five premises were studied with both rapid (ATP) and traditional microbiological swabbing being used to test surfaces that either come into direct contact with prepared foods or were likely to be touched by hands during food preparation. A significant link was found between aerobic colony counts and ATP measurements. In most cases, the visual appearance of surfaces could not be used to accurately predict either microbial or ATP results. This study suggests that ATP testing is a useful indicator of surface cleanliness and could be helpful to local authority officers as part of risk assessment inspections. This study provides further evidence that visual inspection alone may not always be adequate to assess surface cleanliness. In high-risk premises, ATP could, if appropriately targeted, help identify potential problem areas. The results are available at the time of the inspection and can be used as an on-the-spot teaching aid.
GABA predicts visual intelligence.
Cook, Emily; Hammett, Stephen T; Larsson, Jonas
2016-10-06
Early psychological researchers proposed a link between intelligence and low-level perceptual performance. It was recently suggested that this link is driven by individual variations in the ability to suppress irrelevant information, evidenced by the observation of strong correlations between perceptual surround suppression and cognitive performance. However, the neural mechanisms underlying such a link remain unclear. A candidate mechanism is neural inhibition by gamma-aminobutyric acid (GABA), but direct experimental support for GABA-mediated inhibition underlying suppression is inconsistent. Here we report evidence consistent with a global suppressive mechanism involving GABA underlying the link between sensory performance and intelligence. We measured visual cortical GABA concentration, visuo-spatial intelligence and visual surround suppression in a group of healthy adults. Levels of GABA were strongly predictive of both intelligence and surround suppression, with higher levels of intelligence associated with higher levels of GABA and stronger surround suppression. These results indicate that GABA-mediated neural inhibition may be a key factor determining cognitive performance and suggests a physiological mechanism linking surround suppression and intelligence. Copyright © 2016 The Authors. Published by Elsevier Ireland Ltd.. All rights reserved.
Object form discontinuity facilitates displacement discrimination across saccades.
Demeyer, Maarten; De Graef, Peter; Wagemans, Johan; Verfaillie, Karl
2010-06-01
Stimulus displacements coinciding with a saccadic eye movement are poorly detected by human observers. In recent years, converging evidence has shown that this phenomenon does not result from poor transsaccadic retention of presaccadic stimulus position information, but from the visual system's efforts to spatially align presaccadic and postsaccadic perception on the basis of visual landmarks. It is known that this process can be disrupted, and transsaccadic displacement detection performance can be improved, by briefly blanking the stimulus display during and immediately after the saccade. In the present study, we investigated whether this improvement could also follow from a discontinuity in the task-irrelevant form of the displaced stimulus. We observed this to be the case: Subjects more accurately identified the direction of intrasaccadic displacements when the displaced stimulus simultaneously changed form, compared to conditions without a form change. However, larger improvements were still observed under blanking conditions. In a second experiment, we show that facilitation induced by form changes and blanks can combine. We conclude that a strong assumption of visual stability underlies the suppression of transsaccadic change detection performance, the rejection of which generalizes from stimulus form to stimulus position.
Right-hemispheric dominance for visual remapping in humans
Pisella, L.; Alahyane, N.; Blangero, A.; Thery, F.; Blanc, S.; Pelisson, D.
2011-01-01
We review evidence showing a right-hemispheric dominance for visuo-spatial processing and representation in humans. Accordingly, visual disorganization symptoms (intuitively related to remapping impairments) are observed in both neglect and constructional apraxia. More specifically, we review findings from the intervening saccade paradigm in humans—and present additional original data—which suggest a specific role of the asymmetrical network at the temporo-parietal junction (TPJ) in the right hemisphere in visual remapping: following damage to the right dorsal posterior parietal cortex (PPC) as well as part of the corpus callosum connecting the PPC to the frontal lobes, patient OK in a double-step saccadic task exhibited an impairment when the second saccade had to be directed rightward. This singular and lateralized deficit cannot result solely from the patient's cortical lesion and, therefore, we propose that it is due to his callosal lesion that may specifically interrupt the interhemispheric transfer of information necessary to execute accurate rightward saccades towards a remapped target location. This suggests a specialized right-hemispheric network for visuo-spatial remapping that subsequently transfers target location information to downstream planning regions, which are symmetrically organized. PMID:21242144
Temporal processing dysfunction in schizophrenia.
Carroll, Christine A; Boggs, Jennifer; O'Donnell, Brian F; Shekhar, Anantha; Hetrick, William P
2008-07-01
Schizophrenia may be associated with a fundamental disturbance in the temporal coordination of information processing in the brain, leading to classic symptoms of schizophrenia such as thought disorder and disorganized and contextually inappropriate behavior. Despite the growing interest and centrality of time-dependent conceptualizations of the pathophysiology of schizophrenia, there remains a paucity of research directly examining overt timing performance in the disorder. Accordingly, the present study investigated timing in schizophrenia using a well-established task of time perception. Twenty-three individuals with schizophrenia and 22 non-psychiatric control participants completed a temporal bisection task, which required participants to make temporal judgments about auditory and visually presented durations ranging from 300 to 600 ms. Both schizophrenia and control groups displayed greater visual compared to auditory timing variability, with no difference between groups in the visual modality. However, individuals with schizophrenia exhibited less temporal precision than controls in the perception of auditory durations. These findings correlated with parameter estimates obtained from a quantitative model of time estimation, and provide evidence of a fundamental deficit in temporal auditory precision in schizophrenia.
ERIC Educational Resources Information Center
Zannino, Gian Daniele; Perri, Roberta; Salamone, Giovanna; Di Lorenzo, Concetta; Caltagirone, Carlo; Carlesimo, Giovanni A.
2010-01-01
There is now a large body of evidence suggesting that color and photographic detail exert an effect on recognition of visually presented familiar objects. However, an unresolved issue is whether these factors act at the visual, the semantic or lexical level of the recognition process. In the present study, we investigated this issue by having…
Ezechi, Oliver Chukwujekwu; Petterson, Karen Odberg; Gbajabiamila, Titilola A; Idigbe, Ifeoma Eugenia; Kuyoro, Olutunmike; Ujah, Innocent Achaya Otobo; Ostergren, Per Olof
2014-03-31
Increasingly evidence is emerging from south East Asia, southern and east Africa on the burden of default to follow up care after a positive cervical cancer screening/diagnosis, which impacts negatively on cervical cancer prevention and control. Unfortunately little or no information exists on the subject in the West Africa sub region. This study was designed to determine the proportion of and predictors and reasons for default from follow up care after positive cervical cancer screen. Women who screen positive at community cervical cancer screening using direct visual inspection were followed up to determine the proportion of default and associated factors. Multivariate logistic regression was used to determine independent predictors of default. One hundred and eight (16.1%) women who screened positive to direct visual inspection out of 673 were enrolled into the study. Fifty one (47.2%) out of the 108 women that screened positive defaulted from follow-up appointment. Women who were poorly educated (OR: 3.1, CI: 2.0 - 5.2), or lived more than 10 km from the clinic (OR: 2.0, CI: 1.0 - 4.1), or never screened for cervical cancer before (OR: 3.5, CI:3:1-8.4) were more likely to default from follow-up after screening positive for precancerous lesion of cervix . The main reasons for default were cost of transportation (48.6%) and time constraints (25.7%). The rate of default was high (47.2%) as a result of unaffordable transportation cost and limited time to keep the scheduled appointment. A change from the present strategy that involves multiple visits to a "see and treat" strategy in which both testing and treatment are performed at a single visit is recommended.
Visual system evolution and the nature of the ancestral snake.
Simões, B F; Sampaio, F L; Jared, C; Antoniazzi, M M; Loew, E R; Bowmaker, J K; Rodriguez, A; Hart, N S; Hunt, D M; Partridge, J C; Gower, D J
2015-07-01
The dominant hypothesis for the evolutionary origin of snakes from 'lizards' (non-snake squamates) is that stem snakes acquired many snake features while passing through a profound burrowing (fossorial) phase. To investigate this, we examined the visual pigments and their encoding opsin genes in a range of squamate reptiles, focusing on fossorial lizards and snakes. We sequenced opsin transcripts isolated from retinal cDNA and used microspectrophotometry to measure directly the spectral absorbance of the photoreceptor visual pigments in a subset of samples. In snakes, but not lizards, dedicated fossoriality (as in Scolecophidia and the alethinophidian Anilius scytale) corresponds with loss of all visual opsins other than RH1 (λmax 490-497 nm); all other snakes (including less dedicated burrowers) also have functional sws1 and lws opsin genes. In contrast, the retinas of all lizards sampled, even highly fossorial amphisbaenians with reduced eyes, express functional lws, sws1, sws2 and rh1 genes, and most also express rh2 (i.e. they express all five of the visual opsin genes present in the ancestral vertebrate). Our evidence of visual pigment complements suggests that the visual system of stem snakes was partly reduced, with two (RH2 and SWS2) of the ancestral vertebrate visual pigments being eliminated, but that this did not extend to the extreme additional loss of SWS1 and LWS that subsequently occurred (probably independently) in highly fossorial extant scolecophidians and A. scytale. We therefore consider it unlikely that the ancestral snake was as fossorial as extant scolecophidians, whether or not the latter are para- or monophyletic. © 2015 European Society For Evolutionary Biology. Journal of Evolutionary Biology © 2015 European Society For Evolutionary Biology.
Top-down modulation of ventral occipito-temporal responses during visual word recognition.
Twomey, Tae; Kawabata Duncan, Keith J; Price, Cathy J; Devlin, Joseph T
2011-04-01
Although interactivity is considered a fundamental principle of cognitive (and computational) models of reading, it has received far less attention in neural models of reading that instead focus on serial stages of feed-forward processing from visual input to orthographic processing to accessing the corresponding phonological and semantic information. In particular, the left ventral occipito-temporal (vOT) cortex is proposed to be the first stage where visual word recognition occurs prior to accessing nonvisual information such as semantics and phonology. We used functional magnetic resonance imaging (fMRI) to investigate whether there is evidence that activation in vOT is influenced top-down by the interaction of visual and nonvisual properties of the stimuli during visual word recognition tasks. Participants performed two different types of lexical decision tasks that focused on either visual or nonvisual properties of the word or word-like stimuli. The design allowed us to investigate how vOT activation during visual word recognition was influenced by a task change to the same stimuli and by a stimulus change during the same task. We found both stimulus- and task-driven modulation of vOT activation that can only be explained by top-down processing of nonvisual aspects of the task and stimuli. Our results are consistent with the hypothesis that vOT acts as an interface linking visual form with nonvisual processing in both bottom up and top down directions. Such interactive processing at the neural level is in agreement with cognitive and computational models of reading but challenges some of the assumptions made by current neuro-anatomical models of reading. Copyright © 2011 Elsevier Inc. All rights reserved.
Latham, Andrew J.; Patston, Lucy L. M.; Westermann, Christine; Kirk, Ian J.; Tippett, Lynette J.
2013-01-01
Increasing behavioural evidence suggests that expert video game players (VGPs) show enhanced visual attention and visuospatial abilities, but what underlies these enhancements remains unclear. We administered the Poffenberger paradigm with concurrent electroencephalogram (EEG) recording to assess occipital N1 latencies and interhemispheric transfer time (IHTT) in expert VGPs. Participants comprised 15 right-handed male expert VGPs and 16 non-VGP controls matched for age, handedness, IQ and years of education. Expert VGPs began playing before age 10, had a minimum 8 years experience, and maintained playtime of at least 20 hours per week over the last 6 months. Non-VGPs had little-to-no game play experience (maximum 1.5 years). Participants responded to checkerboard stimuli presented to the left and right visual fields while 128-channel EEG was recorded. Expert VGPs responded significantly more quickly than non-VGPs. Expert VGPs also had significantly earlier occipital N1s in direct visual pathways (the hemisphere contralateral to the visual field in which the stimulus was presented). IHTT was calculated by comparing the latencies of occipital N1 components between hemispheres. No significant between-group differences in electrophysiological estimates of IHTT were found. Shorter N1 latencies may enable expert VGPs to discriminate attended visual stimuli significantly earlier than non-VGPs and contribute to faster responding in visual tasks. As successful video-game play requires precise, time pressured, bimanual motor movements in response to complex visual stimuli, which in this sample began during early childhood, these differences may reflect the experience and training involved during the development of video-game expertise, but training studies are needed to test this prediction. PMID:24058667
Latham, Andrew J; Patston, Lucy L M; Westermann, Christine; Kirk, Ian J; Tippett, Lynette J
2013-01-01
Increasing behavioural evidence suggests that expert video game players (VGPs) show enhanced visual attention and visuospatial abilities, but what underlies these enhancements remains unclear. We administered the Poffenberger paradigm with concurrent electroencephalogram (EEG) recording to assess occipital N1 latencies and interhemispheric transfer time (IHTT) in expert VGPs. Participants comprised 15 right-handed male expert VGPs and 16 non-VGP controls matched for age, handedness, IQ and years of education. Expert VGPs began playing before age 10, had a minimum 8 years experience, and maintained playtime of at least 20 hours per week over the last 6 months. Non-VGPs had little-to-no game play experience (maximum 1.5 years). Participants responded to checkerboard stimuli presented to the left and right visual fields while 128-channel EEG was recorded. Expert VGPs responded significantly more quickly than non-VGPs. Expert VGPs also had significantly earlier occipital N1s in direct visual pathways (the hemisphere contralateral to the visual field in which the stimulus was presented). IHTT was calculated by comparing the latencies of occipital N1 components between hemispheres. No significant between-group differences in electrophysiological estimates of IHTT were found. Shorter N1 latencies may enable expert VGPs to discriminate attended visual stimuli significantly earlier than non-VGPs and contribute to faster responding in visual tasks. As successful video-game play requires precise, time pressured, bimanual motor movements in response to complex visual stimuli, which in this sample began during early childhood, these differences may reflect the experience and training involved during the development of video-game expertise, but training studies are needed to test this prediction.
The influence of artificial scotomas on eye movements during visual search.
Cornelissen, Frans W; Bruin, Klaas J; Kooijman, Aart C
2005-01-01
Fixation durations are normally adapted to the difficulty of the foveal analysis task. We examine to what extent artificial central and peripheral visual field defects interfere with this adaptation process. Subjects performed a visual search task while their eye movements were registered. The latter were used to drive a real-time gaze-dependent display that was used to create artificial central and peripheral visual field defects. Recorded eye movements were used to determine saccadic amplitude, number of fixations, fixation durations, return saccades, and changes in saccade direction. For central defects, although fixation duration increased with the size of the absolute central scotoma, this increase was too small to keep recognition performance optimal, evident from an associated increase in the rate of return saccades. Providing a relatively small amount of visual information in the central scotoma did substantially reduce subjects' search times but not their fixation durations. Surprisingly, reducing the size of the tunnel also prolonged fixation duration for peripheral defects. This manipulation also decreased the rate of return saccades, suggesting that the fixations were prolonged beyond the duration required by the foveal task. Although we find that adaptation of fixation duration to task difficulty clearly occurs in the presence of artificial scotomas, we also find that such field defects may render the adaptation suboptimal for the task at hand. Thus, visual field defects may not only hinder vision by limiting what the subject sees of the environment but also by limiting the visual system's ability to program efficient eye movements. We speculate this is because of how visual field defects bias the balance between saccade generation and fixation stabilization.
Gall, Carolin; Silvennoinen, Katri; Granata, Giuseppe; de Rossi, Francesca; Vecchio, Fabrizio; Brösel, Doreen; Bola, Michał; Sailer, Michael; Waleszczyk, Wioletta J; Rossini, Paolo M; Tatlisumak, Turgut; Sabel, Bernhard A
2015-07-01
Occipital stroke often leads to visual field loss, for which no effective treatment exists. Little is known about the potential of non-invasive electric current stimulation to ameliorate visual functions in patients suffering from unilateral occipital stroke. One reason is the traditional thinking that visual field loss after brain lesions is permanent. Since evidence is available documenting vision restoration by means of vision training or non-invasive electric current stimulation future studies should also consider investigating recovery processes after visual cortical strokes. Here, protocols of repetitive transorbital alternating current stimulation (rtACS) and transcranial direct current stimulation (tDCS) are presented and the European consortium for restoration of vision (REVIS) is introduced. Within the consortium different stimulation approaches will be applied to patients with unilateral occipital strokes resulting in homonymous hemianopic visual field defects. The aim of the study is to evaluate effects of current stimulation of the brain on vision parameters, vision-related quality of life, and physiological parameters that allow concluding about the mechanisms of vision restoration. These include EEG-spectra and coherence measures, and visual evoked potentials. The design of stimulation protocols involves an appropriate sham-stimulation condition and sufficient follow-up periods to test whether the effects are stable. This is the first application of non-invasive current stimulation for vision rehabilitation in stroke-related visual field deficits. Positive results of the trials could have far-reaching implications for clinical practice. The ability of non-invasive electrical current brain stimulation to modulate the activity of neuronal networks may have implications for stroke rehabilitation also in the visual domain. Copyright © 2015 Elsevier Inc. All rights reserved.
Smelling directions: Olfaction modulates ambiguous visual motion perception
Kuang, Shenbing; Zhang, Tao
2014-01-01
Senses of smells are often accompanied by simultaneous visual sensations. Previous studies have documented enhanced olfactory performance with concurrent presence of congruent color- or shape- related visual cues, and facilitated visual object perception when congruent smells are simultaneously present. These visual object-olfaction interactions suggest the existences of couplings between the olfactory pathway and the visual ventral processing stream. However, it is not known if olfaction can modulate visual motion perception, a function that is related to the visual dorsal stream. We tested this possibility by examining the influence of olfactory cues on the perceptions of ambiguous visual motion signals. We showed that, after introducing an association between motion directions and olfactory cues, olfaction could indeed bias ambiguous visual motion perceptions. Our result that olfaction modulates visual motion processing adds to the current knowledge of cross-modal interactions and implies a possible functional linkage between the olfactory system and the visual dorsal pathway. PMID:25052162
Demonstration of a neural circuit critical for imprinting behavior in chicks.
Nakamori, Tomoharu; Sato, Katsushige; Atoji, Yasuro; Kanamatsu, Tomoyuki; Tanaka, Kohichi; Ohki-Hamazaki, Hiroko
2010-03-24
Imprinting behavior in birds is elicited by visual and/or auditory cues. It has been demonstrated previously that visual cues are recognized and processed in the visual Wulst (VW), and imprinting memory is stored in the intermediate medial mesopallium (IMM) of the telencephalon. Alteration of neural responses in these two regions according to imprinting has been reported, yet direct evidence of the neural circuit linking these two regions is lacking. Thus, it remains unclear how memory is formed and expressed in this circuit. Here, we present anatomical as well as physiological evidence of the neural circuit connecting the VW and IMM and show that imprinting training during the critical period strengthens and refines this circuit. A functional connection established by imprint training resulted in an imprinting behavior. After the closure of the critical period, training could not activate this circuit nor induce the imprinting behavior. Glutamatergic neurons in the ventroposterior region of the VW, the core region of the hyperpallium densocellulare (HDCo), sent their axons to the periventricular part of the HD, just dorsal and afferent to the IMM. We found that the HDCo is important in imprinting behavior. The refinement and/or enhancement of this neural circuit are attributed to increased activity of HDCo cells, and the activity depended on NR2B-containing NMDA receptors. These findings show a neural connection in the telencephalon in Aves and demonstrate that NR2B function is indispensable for the plasticity of HDCo cells, which are key mediators of imprinting.
Brain-Stimulation Induced Blindsight: Unconscious Vision or Response Bias?
Lloyd, David A.; Abrahamyan, Arman; Harris, Justin A.
2013-01-01
A dissociation between visual awareness and visual discrimination is referred to as “blindsight”. Blindsight results from loss of function of the primary visual cortex (V1) which can occur due to cerebrovascular accidents (i.e. stroke-related lesions). There are also numerous reports of similar, though reversible, effects on vision induced by transcranial Magnetic Stimulation (TMS) to early visual cortex. These effects point to V1 as the “gate” of visual awareness and have strong implications for understanding the neurological underpinnings of consciousness. It has been argued that evidence for the dissociation between awareness of, and responses to, visual stimuli can be a measurement artifact of the use of a high response criterion under yes-no measures of visual awareness when compared with the criterion free forced-choice responses. This difference between yes-no and forced-choice measures suggests that evidence for a dissociation may actually be normal near-threshold conscious vision. Here we describe three experiments that tested visual performance in normal subjects when their visual awareness was suppressed by applying TMS to the occipital pole. The nature of subjects’ performance whilst undergoing occipital TMS was then verified by use of a psychophysical measure (d') that is independent of response criteria. This showed that there was no genuine dissociation in visual sensitivity measured by yes-no and forced-choice responses. These results highlight that evidence for visual sensitivity in the absence of awareness must be analysed using a bias-free psychophysical measure, such as d', In order to confirm whether or not visual performance is truly unconscious. PMID:24324837
Brain-stimulation induced blindsight: unconscious vision or response bias?
Lloyd, David A; Abrahamyan, Arman; Harris, Justin A
2013-01-01
A dissociation between visual awareness and visual discrimination is referred to as "blindsight". Blindsight results from loss of function of the primary visual cortex (V1) which can occur due to cerebrovascular accidents (i.e. stroke-related lesions). There are also numerous reports of similar, though reversible, effects on vision induced by transcranial Magnetic Stimulation (TMS) to early visual cortex. These effects point to V1 as the "gate" of visual awareness and have strong implications for understanding the neurological underpinnings of consciousness. It has been argued that evidence for the dissociation between awareness of, and responses to, visual stimuli can be a measurement artifact of the use of a high response criterion under yes-no measures of visual awareness when compared with the criterion free forced-choice responses. This difference between yes-no and forced-choice measures suggests that evidence for a dissociation may actually be normal near-threshold conscious vision. Here we describe three experiments that tested visual performance in normal subjects when their visual awareness was suppressed by applying TMS to the occipital pole. The nature of subjects' performance whilst undergoing occipital TMS was then verified by use of a psychophysical measure (d') that is independent of response criteria. This showed that there was no genuine dissociation in visual sensitivity measured by yes-no and forced-choice responses. These results highlight that evidence for visual sensitivity in the absence of awareness must be analysed using a bias-free psychophysical measure, such as d', In order to confirm whether or not visual performance is truly unconscious.
Invertebrate neurobiology: visual direction of arm movements in an octopus.
Niven, Jeremy E
2011-03-22
An operant task in which octopuses learn to locate food by a visual cue in a three-choice maze shows that they are capable of integrating visual and mechanosensory information to direct their arm movements to a goal. Copyright © 2011 Elsevier Ltd. All rights reserved.
The forensic validity of visual analytics
NASA Astrophysics Data System (ADS)
Erbacher, Robert F.
2008-01-01
The wider use of visualization and visual analytics in wide ranging fields has led to the need for visual analytics capabilities to be legally admissible, especially when applied to digital forensics. This brings the need to consider legal implications when performing visual analytics, an issue not traditionally examined in visualization and visual analytics techniques and research. While digital data is generally admissible under the Federal Rules of Evidence [10][21], a comprehensive validation of the digital evidence is considered prudent. A comprehensive validation requires validation of the digital data under rules for authentication, hearsay, best evidence rule, and privilege. Additional issues with digital data arise when exploring digital data related to admissibility and the validity of what information was examined, to what extent, and whether the analysis process was sufficiently covered by a search warrant. For instance, a search warrant generally covers very narrow requirements as to what law enforcement is allowed to examine and acquire during an investigation. When searching a hard drive for child pornography, how admissible is evidence of an unrelated crime, i.e. drug dealing. This is further complicated by the concept of "in plain view". When performing an analysis of a hard drive what would be considered "in plain view" when analyzing a hard drive. The purpose of this paper is to discuss the issues of digital forensics and the related issues as they apply to visual analytics and identify how visual analytics techniques fit into the digital forensics analysis process, how visual analytics techniques can improve the legal admissibility of digital data, and identify what research is needed to further improve this process. The goal of this paper is to open up consideration of legal ramifications among the visualization community; the author is not a lawyer and the discussions are not meant to be inclusive of all differences in laws between states and countries.
From genes to brain oscillations: is the visual pathway the epigenetic clue to schizophrenia?
González-Hernández, J A; Pita-Alcorta, C; Cedeño, I R
2006-01-01
Molecular data and gene expression data and recently mitochondrial genes and possible epigenetic regulation by non-coding genes is revolutionizing our views on schizophrenia. Genes and epigenetic mechanisms are triggered by cell-cell interaction and by external stimuli. A number of recent clinical and molecular observations indicate that epigenetic factors may be operational in the origin of the illness. Based on the molecular insights, gene expression profiles and epigenetic regulation of gene, we went back to the neurophysiology (brain oscillations) and found a putative role of the visual experiences (i.e. visual stimuli) as epigenetic factor. The functional evidences provided here, establish a direct link between the striate and extrastriate unimodal visual cortex and the neurobiology of the schizophrenia. This result support the hypothesis that 'visual experience' has a potential role as epigenetic factor and contribute to trigger and/or to maintain the progression of the schizophrenia. In this case, candidate genes sensible for the visual 'insult' may be located within the visual cortex including associative areas, while the integrity of the visual pathway before reaching the primary visual cortex is preserved. The same effect can be perceived if target genes are localised within the visual pathway, which actually, is more sensitive for 'insult' during the early life than the cortex per se. If this process affects gene expression at these sites a stably sensory specific 'insult', i.e. distorted visual information, is entering the visual system and expanded to fronto-temporo-parietal multimodal areas even from early maturation periods. The difference in the timing of postnatal neuroanatomical events between such areas and the primary visual cortex in humans (with the formers reaching the same development landmarks later in life than the latter) is 'optimal' to establish an abnormal 'cell- communication' mediated by the visual system that may further interfere with the local physiology. In this context the strategy to search target genes need to be rearrangement and redirected to visual-related genes. Otherwise, psychophysics studies combining functional neuroimage, and electrophysiology are strongly recommended, for the search of epigenetic clues that will allow to carrier gene association studies in schizophrenia.
Latychevskaia, Tatiana; Wicki, Flavio; Longchamp, Jean-Nicolas; Escher, Conrad; Fink, Hans-Werner
2016-09-14
Visualizing individual charges confined to molecules and observing their dynamics with high spatial resolution is a challenge for advancing various fields in science, ranging from mesoscopic physics to electron transfer events in biological molecules. We show here that the high sensitivity of low-energy electrons to local electric fields can be employed to directly visualize individual charged adsorbates and to study their behavior in a quantitative way. This makes electron holography a unique probing tool for directly visualizing charge distributions with a sensitivity of a fraction of an elementary charge. Moreover, spatial resolution in the nanometer range and fast data acquisition inherent to lens-less low-energy electron holography allows for direct visual inspection of charge transfer processes.
Chimpanzee Alarm Call Production Meets Key Criteria for Intentionality
Schel, Anne Marijke; Townsend, Simon W.; Machanda, Zarin; Zuberbühler, Klaus; Slocombe, Katie E.
2013-01-01
Determining the intentionality of primate communication is critical to understanding the evolution of human language. Although intentional signalling has been claimed for some great ape gestural signals, comparable evidence is currently lacking for their vocal signals. We presented wild chimpanzees with a python model and found that two of three alarm call types exhibited characteristics previously used to argue for intentionality in gestural communication. These alarm calls were: (i) socially directed and given to the arrival of friends, (ii) associated with visual monitoring of the audience and gaze alternations, and (iii) goal directed, as calling only stopped when recipients were safe from the predator. Our results demonstrate that certain vocalisations of our closest living relatives qualify as intentional signals, in a directly comparable way to many great ape gestures. We conclude that our results undermine a central argument of gestural theories of language evolution and instead support a multimodal origin of human language. PMID:24146908
Ben-Simon, Eti; Podlipsky, Ilana; Okon-Singer, Hadas; Gruberger, Michal; Cvetkovic, Dean; Intrator, Nathan; Hendler, Talma
2013-03-01
The unique role of the EEG alpha rhythm in different states of cortical activity is still debated. The main theories regarding alpha function posit either sensory processing or attention allocation as the main processes governing its modulation. Closing and opening eyes, a well-known manipulation of the alpha rhythm, could be regarded as attention allocation from inward to outward focus though during light is also accompanied by visual change. To disentangle the effects of attention allocation and sensory visual input on alpha modulation, 14 healthy subjects were asked to open and close their eyes during conditions of light and of complete darkness while simultaneous recordings of EEG and fMRI were acquired. Thus, during complete darkness the eyes-open condition is not related to visual input but only to attention allocation, allowing direct examination of its role in alpha modulation. A data-driven ridge regression classifier was applied to the EEG data in order to ascertain the contribution of the alpha rhythm to eyes-open/eyes-closed inference in both lighting conditions. Classifier results revealed significant alpha contribution during both light and dark conditions, suggesting that alpha rhythm modulation is closely linked to the change in the direction of attention regardless of the presence of visual sensory input. Furthermore, fMRI activation maps derived from an alpha modulation time-course during the complete darkness condition exhibited a right frontal cortical network associated with attention allocation. These findings support the importance of top-down processes such as attention allocation to alpha rhythm modulation, possibly as a prerequisite to its known bottom-up processing of sensory input. © 2012 Federation of European Neuroscience Societies and Blackwell Publishing Ltd.
Higashiyama, A
1992-03-01
Three experiments investigated anisotropic perception of visual angle outdoors. In Experiment 1, scales for vertical and horizontal visual angles ranging from 20 degrees to 80 degrees were constructed with the method of angle production (in which the subject reproduced a visual angle with a protractor) and the method of distance production (in which the subject produced a visual angle by adjusting viewing distance). In Experiment 2, scales for vertical and horizontal visual angles of 5 degrees-30 degrees were constructed with the method of angle production and were compared with scales for orientation in the frontal plane. In Experiment 3, vertical and horizontal visual angles of 3 degrees-80 degrees were judged with the method of verbal estimation. The main results of the experiments were as follows: (1) The obtained angles for visual angle are described by a quadratic equation, theta' = a + b theta + c theta 2 (where theta is the visual angle; theta', the obtained angle; a, b, and c, constants). (2) The linear coefficient b is larger than unity and is steeper for vertical direction than for horizontal direction. (3) The quadratic coefficient c is generally smaller than zero and is negatively larger for vertical direction than for horizontal direction. And (4) the obtained angle for visual angle is larger than that for orientation. From these results, it was possible to predict the horizontal-vertical illusion, over-constancy of size, and the moon illusion.
Behavioral and Neural Representations of Spatial Directions across Words, Schemas, and Images.
Weisberg, Steven M; Marchette, Steven A; Chatterjee, Anjan
2018-05-23
Modern spatial navigation requires fluency with multiple representational formats, including visual scenes, signs, and words. These formats convey different information. Visual scenes are rich and specific but contain extraneous details. Arrows, as an example of signs, are schematic representations in which the extraneous details are eliminated, but analog spatial properties are preserved. Words eliminate all spatial information and convey spatial directions in a purely abstract form. How does the human brain compute spatial directions within and across these formats? To investigate this question, we conducted two experiments on men and women: a behavioral study that was preregistered and a neuroimaging study using multivoxel pattern analysis of fMRI data to uncover similarities and differences among representational formats. Participants in the behavioral study viewed spatial directions presented as images, schemas, or words (e.g., "left"), and responded to each trial, indicating whether the spatial direction was the same or different as the one viewed previously. They responded more quickly to schemas and words than images, despite the visual complexity of stimuli being matched. Participants in the fMRI study performed the same task but responded only to occasional catch trials. Spatial directions in images were decodable in the intraparietal sulcus bilaterally but were not in schemas and words. Spatial directions were also decodable between all three formats. These results suggest that intraparietal sulcus plays a role in calculating spatial directions in visual scenes, but this neural circuitry may be bypassed when the spatial directions are presented as schemas or words. SIGNIFICANCE STATEMENT Human navigators encounter spatial directions in various formats: words ("turn left"), schematic signs (an arrow showing a left turn), and visual scenes (a road turning left). The brain must transform these spatial directions into a plan for action. Here, we investigate similarities and differences between neural representations of these formats. We found that bilateral intraparietal sulci represent spatial directions in visual scenes and across the three formats. We also found that participants respond quickest to schemas, then words, then images, suggesting that spatial directions in abstract formats are easier to interpret than concrete formats. These results support a model of spatial direction interpretation in which spatial directions are either computed for real world action or computed for efficient visual comparison. Copyright © 2018 the authors 0270-6474/18/384996-12$15.00/0.
Do Visual Illusions Probe the Visual Brain?: Illusions in Action without a Dorsal Visual Stream
ERIC Educational Resources Information Center
Coello, Yann; Danckert, James; Blangero, Annabelle; Rossetti, Yves
2007-01-01
Visual illusions have been shown to affect perceptual judgements more so than motor behaviour, which was interpreted as evidence for a functional division of labour within the visual system. The dominant perception-action theory argues that perception involves a holistic processing of visual objects or scenes, performed within the ventral,…
Location memory biases reveal the challenges of coordinating visual and kinesthetic reference frames
Simmering, Vanessa R.; Peterson, Clayton; Darling, Warren; Spencer, John P.
2008-01-01
Five experiments explored the influence of visual and kinesthetic/proprioceptive reference frames on location memory. Experiments 1 and 2 compared visual and kinesthetic reference frames in a memory task using visually-specified locations and a visually-guided response. When the environment was visible, results replicated previous findings of biases away from the midline symmetry axis of the task space, with stability for targets aligned with this axis. When the environment was not visible, results showed some evidence of bias away from a kinesthetically-specified midline (trunk anterior–posterior [a–p] axis), but there was little evidence of stability when targets were aligned with body midline. This lack of stability may reflect the challenges of coordinating visual and kinesthetic information in the absence of an environmental reference frame. Thus, Experiments 3–5 examined kinesthetic guidance of hand movement to kinesthetically-defined targets. Performance in these experiments was generally accurate with no evidence of consistent biases away from the trunk a–p axis. We discuss these results in the context of the challenges of coordinating reference frames within versus between multiple sensori-motor systems. PMID:17703284
Little, Anthony C
2013-10-01
Steroid sex hormones are responsible for some of the differences between men and women. In this article, I review evidence that steroid sex hormones impact on visual processing. Given prominent sex-differences, I focus on three topics for sex hormone effects for which there is most research available: 1. Preference and mate choice, 2. Emotion and recognition, and 3. Cerebral/perceptual asymmetries and visual-spatial abilities. For each topic, researchers have examined sex hormones and visual processing using various methods. I review indirect evidence addressing variation according to: menstrual cycle phase, pregnancy, puberty, and menopause. I further address studies of variation in testosterone and a measure of prenatal testosterone, 2D:4D, on visual processing. The most conclusive evidence, however, comes from experiments. Studies in which hormones are administrated are discussed. Overall, many studies demonstrate that sex steroids are associated with visual processing. However, findings are sometimes inconsistent, differences in methodology make strong comparisons between studies difficult, and we generally know more about activational than organizational effects. Copyright © 2013 Elsevier Inc. All rights reserved.
Ferrucci, Roberta; Giannicola, Gaia; Rosa, Manuela; Fumagalli, Manuela; Boggio, Paulo Sergio; Hallett, Mark; Zago, Stefano; Priori, Alberto
2012-01-01
Some evidence suggests that the cerebellum participates in the complex network processing emotional facial expression. To evaluate the role of the cerebellum in recognising facial expressions we delivered transcranial direct current stimulation (tDCS) over the cerebellum and prefrontal cortex. A facial emotion recognition task was administered to 21 healthy subjects before and after cerebellar tDCS; we also tested subjects with a visual attention task and a visual analogue scale (VAS) for mood. Anodal and cathodal cerebellar tDCS both significantly enhanced sensory processing in response to negative facial expressions (anodal tDCS, p=.0021; cathodal tDCS, p=.018), but left positive emotion and neutral facial expressions unchanged (p>.05). tDCS over the right prefrontal cortex left facial expressions of both negative and positive emotion unchanged. These findings suggest that the cerebellum is specifically involved in processing facial expressions of negative emotion.
Exceptional preservation of eye structure in arthropod visual predators from the Middle Jurassic
Vannier, Jean; Schoenemann, Brigitte; Gillot, Thomas; Charbonnier, Sylvain; Clarkson, Euan
2016-01-01
Vision has revolutionized the way animals explore their environment and interact with each other and rapidly became a major driving force in animal evolution. However, direct evidence of how ancient animals could perceive their environment is extremely difficult to obtain because internal eye structures are almost never fossilized. Here, we reconstruct with unprecedented resolution the three-dimensional structure of the huge compound eye of a 160-million-year-old thylacocephalan arthropod from the La Voulte exceptional fossil biota in SE France. This arthropod had about 18,000 lenses on each eye, which is a record among extinct and extant arthropods and is surpassed only by modern dragonflies. Combined information about its eyes, internal organs and gut contents obtained by X-ray microtomography lead to the conclusion that this thylacocephalan arthropod was a visual hunter probably adapted to illuminated environments, thus contradicting the hypothesis that La Voulte was a deep-water environment. PMID:26785293
Improvement in spatial imagery following sight onset late in childhood.
Gandhi, Tapan K; Ganesh, Suma; Sinha, Pawan
2014-03-01
The factors contributing to the development of spatial imagery skills are not well understood. Here, we consider whether visual experience shapes these skills. Although differences in spatial imagery between sighted and blind individuals have been reported, it is unclear whether these differences are truly due to visual deprivation or instead are due to extraneous factors, such as reduced opportunities for the blind to interact with their environment. A direct way of assessing vision's contribution to the development of spatial imagery is to determine whether spatial imagery skills change soon after the onset of sight in congenitally blind individuals. We tested 10 children who gained sight after several years of congenital blindness and found significant improvements in their spatial imagery skills following sight-restoring surgeries. These results provide evidence of vision's contribution to spatial imagery and also have implications for the nature of internal spatial representations.
Exceptional preservation of eye structure in arthropod visual predators from the Middle Jurassic.
Vannier, Jean; Schoenemann, Brigitte; Gillot, Thomas; Charbonnier, Sylvain; Clarkson, Euan
2016-01-19
Vision has revolutionized the way animals explore their environment and interact with each other and rapidly became a major driving force in animal evolution. However, direct evidence of how ancient animals could perceive their environment is extremely difficult to obtain because internal eye structures are almost never fossilized. Here, we reconstruct with unprecedented resolution the three-dimensional structure of the huge compound eye of a 160-million-year-old thylacocephalan arthropod from the La Voulte exceptional fossil biota in SE France. This arthropod had about 18,000 lenses on each eye, which is a record among extinct and extant arthropods and is surpassed only by modern dragonflies. Combined information about its eyes, internal organs and gut contents obtained by X-ray microtomography lead to the conclusion that this thylacocephalan arthropod was a visual hunter probably adapted to illuminated environments, thus contradicting the hypothesis that La Voulte was a deep-water environment.
Visual statistical learning is related to natural language ability in adults: An ERP study.
Daltrozzo, Jerome; Emerson, Samantha N; Deocampo, Joanne; Singh, Sonia; Freggens, Marjorie; Branum-Martin, Lee; Conway, Christopher M
2017-03-01
Statistical learning (SL) is believed to enable language acquisition by allowing individuals to learn regularities within linguistic input. However, neural evidence supporting a direct relationship between SL and language ability is scarce. We investigated whether there are associations between event-related potential (ERP) correlates of SL and language abilities while controlling for the general level of selective attention. Seventeen adults completed tests of visual SL, receptive vocabulary, grammatical ability, and sentence completion. Response times and ERPs showed that SL is related to receptive vocabulary and grammatical ability. ERPs indicated that the relationship between SL and grammatical ability was independent of attention while the association between SL and receptive vocabulary depended on attention. The implications of these dissociative relationships in terms of underlying mechanisms of SL and language are discussed. These results further elucidate the cognitive nature of the links between SL mechanisms and language abilities. Copyright © 2017 Elsevier Inc. All rights reserved.
Visual memory transformations in dyslexia.
Barnes, James; Hinkley, Lisa; Masters, Stuart; Boubert, Laura
2007-06-01
Representational Momentum refers to observers' distortion of recognition memory for pictures that imply motion because of an automatic mental process which extrapolates along the implied trajectory of the picture. Neuroimaging evidence suggests that activity in the magnocellular visual pathway is necessary for representational momentum to occur. It has been proposed that individuals with dyslexia have a magnocellular deficit, so it was hypothesised that these individuals would show reduced or absent representational momentum. In this study, 30 adults with dyslexia and 30 age-matched controls were compared on two tasks, one linear and one rotation, which had previously elicited the representational momentum effect. Analysis indicated significant differences in the performance of the two groups, with the dyslexia group having a reduced susceptibility to representational momentum in both linear and rotational directions. The findings highlight that deficits in temporal spatial processing may contribute to the perceptual profile of dyslexia.
Monkey Visual Short-Term Memory Directly Compared to Humans
Elmore, L. Caitlin; Wright, Anthony A.
2015-01-01
Two adult rhesus monkeys were trained to detect which item in an array of memory items had changed using the same stimuli, viewing times, and delays as used with humans. Although the monkeys were extensively trained, they were less accurate than humans with the same array sizes (2, 4, & 6 items), with both stimulus types (colored squares, clip art), and showed calculated memory capacities of about one item (or less). Nevertheless, the memory results from both monkeys and humans for both stimulus types were well characterized by the inverse power-law of display size. This characterization provides a simple and straightforward summary of a fundamental process of visual short-term memory (how VSTM declines with memory load) that emphasizes species similarities based upon similar functional relationships. By more closely matching of monkey testing parameters to those of humans, the similar functional relationships strengthen the evidence suggesting similar processes underlying monkey and human VSTM. PMID:25706544
Sexual motivation is reflected by stimulus-dependent motor cortex excitability.
Schecklmann, Martin; Engelhardt, Kristina; Konzok, Julian; Rupprecht, Rainer; Greenlee, Mark W; Mokros, Andreas; Langguth, Berthold; Poeppl, Timm B
2015-08-01
Sexual behavior involves motivational processes. Findings from both animal models and neuroimaging in humans suggest that the recruitment of neural motor networks is an integral part of the sexual response. However, no study so far has directly linked sexual motivation to physiologically measurable changes in cerebral motor systems in humans. Using transcranial magnetic stimulation in hetero- and homosexual men, we here show that sexual motivation modulates cortical excitability. More specifically, our results demonstrate that visual sexual stimuli corresponding with one's sexual orientation, compared with non-corresponding visual sexual stimuli, increase the excitability of the motor cortex. The reflection of sexual motivation in motor cortex excitability provides evidence for motor preparation processes in sexual behavior in humans. Moreover, such interrelationship links theoretical models and previous neuroimaging findings of sexual behavior. © The Author (2015). Published by Oxford University Press. For Permissions, please email: journals.permissions@oup.com.
Amblyopia and the binocular approach to its therapy.
Hess, Robert F; Thompson, Benjamin
2015-09-01
There is growing evidence that abnormal binocular interactions play a key role in amblyopia. In particular, stronger suppression of the amblyopic eye has been associated with poorer amblyopic eye visual acuity and a new therapy has been described that directly targets binocular function and has been found to improve both monocular and binocular vision in adults and children with amblyopia. Furthermore, non-invasive brain stimulation techniques that alter excitation and inhibition within the visual cortex have been shown to improve vision in the amblyopic eye. The aim of this review is to summarize this previous work and interpret the therapeutic effects of binocular therapy and non-invasive brain stimulation in the context of three potential neural mechanisms; active inhibition of signals from the amblyopic eye, attenuation of information from the amblyopic eye and metaplasticity of synaptic long term potentiation and long term depression. Copyright © 2015. Published by Elsevier Ltd.
Visual statistical learning is related to natural language ability in adults: An ERP Study
Daltrozzo, Jerome; Emerson, Samantha N.; Deocampo, Joanne; Singh, Sonia; Freggens, Marjorie; Branum-Martin, Lee; Conway, Christopher M.
2017-01-01
Statistical learning (SL) is believed to enable language acquisition by allowing individuals to learn regularities within linguistic input. However, neural evidence supporting a direct relationship between SL and language ability is scarce. We investigated whether there are associations between event-related potential (ERP) correlates of SL and language abilities while controlling for the general level of selective attention. Seventeen adults completed tests of visual SL, receptive vocabulary, grammatical ability, and sentence completion. Response times and ERPs showed that SL is related to receptive vocabulary and grammatical ability. ERPs indicated that the relationship between SL and grammatical ability was independent of attention while the association between SL and receptive vocabulary depended on attention. The implications of these dissociative relationships in terms of underlying mechanisms of SL and language are discussed. These results further elucidate the cognitive nature of the links between SL mechanisms and language abilities. PMID:28086142
(Con)text-specific effects of visual dysfunction on reading in posterior cortical atrophy.
Yong, Keir X X; Shakespeare, Timothy J; Cash, Dave; Henley, Susie M D; Warren, Jason D; Crutch, Sebastian J
2014-08-01
Reading deficits are a common early feature of the degenerative syndrome posterior cortical atrophy (PCA) but are poorly understood even at the single word level. The current study evaluated the reading accuracy and speed of 26 PCA patients, 17 typical Alzheimer's disease (tAD) patients and 14 healthy controls on a corpus of 192 single words in which the following perceptual properties were manipulated systematically: inter-letter spacing, font size, length, font type, case and confusability. PCA reading was significantly less accurate and slower than tAD patients and controls, with performance significantly adversely affected by increased letter spacing, size, length and font (cursive < non-cursive), and characterised by visual errors (69% of all error responses). By contrast, tAD and control accuracy rates were at or near ceiling, letter spacing was the only perceptual factor to influence reading speed in the same direction as controls, and, in contrast to PCA patients, control reading was faster for larger font sizes. The inverse size effect in PCA (less accurate reading of large than small font size print) was associated with lower grey matter volume in the right superior parietal lobule. Reading accuracy was associated with impairments of early visual (especially crowding), visuoperceptual and visuospatial processes. However, these deficits were not causally related to a universal impairment of reading as some patients showed preserved reading for small, unspaced words despite grave visual deficits. Rather, the impact of specific types of visual dysfunction on reading was found to be (con)text specific, being particularly evident for large, spaced, lengthy words. These findings improve the characterisation of dyslexia in PCA, shed light on the causative and associative factors, and provide clear direction for the development of reading aids and strategies to maximise and sustain reading ability in the early stages of disease. Copyright © 2014. Published by Elsevier Ltd.
(Con)text-specific effects of visual dysfunction on reading in posterior cortical atrophy
Yong, Keir X.X.; Shakespeare, Timothy J.; Cash, Dave; Henley, Susie M.D.; Warren, Jason D.; Crutch, Sebastian J.
2014-01-01
Reading deficits are a common early feature of the degenerative syndrome posterior cortical atrophy (PCA) but are poorly understood even at the single word level. The current study evaluated the reading accuracy and speed of 26 PCA patients, 17 typical Alzheimer's disease (tAD) patients and 14 healthy controls on a corpus of 192 single words in which the following perceptual properties were manipulated systematically: inter-letter spacing, font size, length, font type, case and confusability. PCA reading was significantly less accurate and slower than tAD patients and controls, with performance significantly adversely affected by increased letter spacing, size, length and font (cursive < non-cursive), and characterised by visual errors (69% of all error responses). By contrast, tAD and control accuracy rates were at or near ceiling, letter spacing was the only perceptual factor to influence reading speed in the same direction as controls, and, in contrast to PCA patients, control reading was faster for larger font sizes. The inverse size effect in PCA (less accurate reading of large than small font size print) was associated with lower grey matter volume in the right superior parietal lobule. Reading accuracy was associated with impairments of early visual (especially crowding), visuoperceptual and visuospatial processes. However, these deficits were not causally related to a universal impairment of reading as some patients showed preserved reading for small, unspaced words despite grave visual deficits. Rather, the impact of specific types of visual dysfunction on reading was found to be (con)text specific, being particularly evident for large, spaced, lengthy words. These findings improve the characterisation of dyslexia in PCA, shed light on the causative and associative factors, and provide clear direction for the development of reading aids and strategies to maximise and sustain reading ability in the early stages of disease. PMID:24841985
Visual representation of spatiotemporal structure
NASA Astrophysics Data System (ADS)
Schill, Kerstin; Zetzsche, Christoph; Brauer, Wilfried; Eisenkolb, A.; Musto, A.
1998-07-01
The processing and representation of motion information is addressed from an integrated perspective comprising low- level signal processing properties as well as higher-level cognitive aspects. For the low-level processing of motion information we argue that a fundamental requirement is the existence of a spatio-temporal memory. Its key feature, the provision of an orthogonal relation between external time and its internal representation, is achieved by a mapping of temporal structure into a locally distributed activity distribution accessible in parallel by higher-level processing stages. This leads to a reinterpretation of the classical concept of `iconic memory' and resolves inconsistencies on ultra-short-time processing and visual masking. The spatial-temporal memory is further investigated by experiments on the perception of spatio-temporal patterns. Results on the direction discrimination of motion paths provide evidence that information about direction and location are not processed and represented independent of each other. This suggests a unified representation on an early level, in the sense that motion information is internally available in form of a spatio-temporal compound. For the higher-level representation we have developed a formal framework for the qualitative description of courses of motion that may occur with moving objects.
Jansen, Petra; Kellner, Jan
2015-01-01
Mental rotation of visual images of body parts and abstract shapes can be influenced by simultaneous motor activity. Children in particular have a strong coupling between motor and cognitive processes. We investigated the influence of a rotational hand movement performed by rotating a knob on mental rotation performance in primary school-age children (N = 83; age range: 7.0–8.3 and 9.0–10.11 years). In addition, we assessed the role of motor ability in this relationship. Boys in the 7- to 8-year-old group were faster when mentally and manually rotating in the same direction than in the opposite direction. For girls and older children this effect was not found. A positive relationship was found between motor ability and accuracy on the mental rotation task: stronger motor ability related to improved mental rotation performance. In both age groups, children with more advanced motor abilities were more likely to adopt motor processes to solve mental rotation tasks if the mental rotation task was primed by a motor task. Our evidence supports the idea that an overlap between motor and visual cognitive processes in children is influenced by motor ability. PMID:26236262
Jansen, Petra; Kellner, Jan
2015-01-01
Mental rotation of visual images of body parts and abstract shapes can be influenced by simultaneous motor activity. Children in particular have a strong coupling between motor and cognitive processes. We investigated the influence of a rotational hand movement performed by rotating a knob on mental rotation performance in primary school-age children (N = 83; age range: 7.0-8.3 and 9.0-10.11 years). In addition, we assessed the role of motor ability in this relationship. Boys in the 7- to 8-year-old group were faster when mentally and manually rotating in the same direction than in the opposite direction. For girls and older children this effect was not found. A positive relationship was found between motor ability and accuracy on the mental rotation task: stronger motor ability related to improved mental rotation performance. In both age groups, children with more advanced motor abilities were more likely to adopt motor processes to solve mental rotation tasks if the mental rotation task was primed by a motor task. Our evidence supports the idea that an overlap between motor and visual cognitive processes in children is influenced by motor ability.
Suppressive mechanisms in visual motion processing: from perception to intelligence
Tadin, Duje
2015-01-01
Perception operates on an immense amount of incoming information that greatly exceeds the brain's processing capacity. Because of this fundamental limitation, the ability to suppress irrelevant information is a key determinant of perceptual efficiency. Here, I will review a series of studies investigating suppressive mechanisms in visual motion processing, namely perceptual suppression of large, background-like motions. These spatial suppression mechanisms are adaptive, operating only when sensory inputs are sufficiently robust to guarantee visibility. Converging correlational and causal evidence links these behavioral results with inhibitory center-surround mechanisms, namely those in cortical area MT. Spatial suppression is abnormally weak in several special populations, including the elderly and those with schizophrenia—a deficit that is evidenced by better-than-normal direction discriminations of large moving stimuli. Theoretical work shows that this abnormal weakening of spatial suppression should result in motion segregation deficits, but direct behavioral support of this hypothesis is lacking. Finally, I will argue that the ability to suppress information is a fundamental neural process that applies not only to perception but also to cognition in general. Supporting this argument, I will discuss recent research that shows individual differences in spatial suppression of motion signals strongly predict individual variations in IQ scores. PMID:26299386
Capture of visual direction in dynamic vergence is reduced with flashed monocular lines.
Jaschinski, Wolfgang; Jainta, Stephanie; Schürer, Michael
2006-08-01
The visual direction of a continuously presented monocular object is captured by the visual direction of a closely adjacent binocular object, which questions the reliability of nonius lines for measuring vergence. This was shown by Erkelens, C. J., and van Ee, R. (1997a,b) [Capture of the visual direction: An unexpected phenomenon in binocular vision. Vision Research, 37, 1193-1196; Capture of the visual direction of monocular objects by adjacent binocular objects. Vision Research, 37, 1735-1745] stimulating dynamic vergence by a counter phase oscillation of two square random-dot patterns (one to each eye) that contained a smaller central dot-free gap (of variable width) with a vertical monocular line oscillating in phase with the random-dot pattern of the respective eye; subjects adjusted the motion-amplitude of the line until it was perceived as (nearly) stationary. With a continuously presented monocular line, we replicated capture of visual direction provided the dot-free gap was narrow: the adjusted motion-amplitude of the line was similar as the motion-amplitude of the random-dot pattern, although large vergence errors occurred. However, when we flashed the line for 67 ms at the moments of maximal and minimal disparity of the vergence stimulus, we found that the adjusted motion-amplitude of the line was smaller; thus, the capture effect appeared to be reduced with flashed nonius lines. Accordingly, we found that the objectively measured vergence gain was significantly correlated (r=0.8) with the motion-amplitude of the flashed monocular line when the separation between the line and the fusion contour was at least 32 min arc. In conclusion, if one wishes to estimate the dynamic vergence response with psychophysical methods, effects of capture of visual direction can be reduced by using flashed nonius lines.
Redundancy reduction explains the expansion of visual direction space around the cardinal axes.
Perrone, John A; Liston, Dorion B
2015-06-01
Motion direction discrimination in humans is worse for oblique directions than for the cardinal directions (the oblique effect). For some unknown reason, the human visual system makes systematic errors in the estimation of particular motion directions; a direction displacement near a cardinal axis appears larger than it really is whereas the same displacement near an oblique axis appears to be smaller. Although the perceptual effects are robust and are clearly measurable in smooth pursuit eye movements, all attempts to identify the neural underpinnings for the oblique effect have failed. Here we show that a model of image velocity estimation based on the known properties of neurons in primary visual cortex (V1) and the middle temporal (MT) visual area of the primate brain produces the oblique effect. We also provide an explanation for the unusual asymmetric patterns of inhibition that have been found surrounding MT neurons. These patterns are consistent with a mechanism within the visual system that prevents redundant velocity signals from being passed onto the next motion-integration stage, (dorsal Medial superior temporal, MSTd). We show that model redundancy-reduction mechanisms within the MT-MSTd pathway produce the oblique effect. Copyright © 2015 Elsevier Ltd. All rights reserved.
Altieri, Nicholas; Pisoni, David B.; Townsend, James T.
2012-01-01
Summerfield (1987) proposed several accounts of audiovisual speech perception, a field of research that has burgeoned in recent years. The proposed accounts included the integration of discrete phonetic features, vectors describing the values of independent acoustical and optical parameters, the filter function of the vocal tract, and articulatory dynamics of the vocal tract. The latter two accounts assume that the representations of audiovisual speech perception are based on abstract gestures, while the former two assume that the representations consist of symbolic or featural information obtained from visual and auditory modalities. Recent converging evidence from several different disciplines reveals that the general framework of Summerfield’s feature-based theories should be expanded. An updated framework building upon the feature-based theories is presented. We propose a processing model arguing that auditory and visual brain circuits provide facilitatory information when the inputs are correctly timed, and that auditory and visual speech representations do not necessarily undergo translation into a common code during information processing. Future research on multisensory processing in speech perception should investigate the connections between auditory and visual brain regions, and utilize dynamic modeling tools to further understand the timing and information processing mechanisms involved in audiovisual speech integration. PMID:21968081
Altieri, Nicholas; Pisoni, David B; Townsend, James T
2011-01-01
Summerfield (1987) proposed several accounts of audiovisual speech perception, a field of research that has burgeoned in recent years. The proposed accounts included the integration of discrete phonetic features, vectors describing the values of independent acoustical and optical parameters, the filter function of the vocal tract, and articulatory dynamics of the vocal tract. The latter two accounts assume that the representations of audiovisual speech perception are based on abstract gestures, while the former two assume that the representations consist of symbolic or featural information obtained from visual and auditory modalities. Recent converging evidence from several different disciplines reveals that the general framework of Summerfield's feature-based theories should be expanded. An updated framework building upon the feature-based theories is presented. We propose a processing model arguing that auditory and visual brain circuits provide facilitatory information when the inputs are correctly timed, and that auditory and visual speech representations do not necessarily undergo translation into a common code during information processing. Future research on multisensory processing in speech perception should investigate the connections between auditory and visual brain regions, and utilize dynamic modeling tools to further understand the timing and information processing mechanisms involved in audiovisual speech integration.
Evidence for multisensory spatial-to-motor transformations in aiming movements of children.
King, Bradley R; Kagerer, Florian A; Contreras-Vidal, Jose L; Clark, Jane E
2009-01-01
The extant developmental literature investigating age-related differences in the execution of aiming movements has predominantly focused on visuomotor coordination, despite the fact that additional sensory modalities, such as audition and somatosensation, may contribute to motor planning, execution, and learning. The current study investigated the execution of aiming movements toward both visual and acoustic stimuli. In addition, we examined the interaction between visuomotor and auditory-motor coordination as 5- to 10-yr-old participants executed aiming movements to visual and acoustic stimuli before and after exposure to a visuomotor rotation. Children in all age groups demonstrated significant improvement in performance under the visuomotor perturbation, as indicated by decreased initial directional and root mean squared errors. Moreover, children in all age groups demonstrated significant visual aftereffects during the postexposure phase, suggesting a successful update of their spatial-to-motor transformations. Interestingly, these updated spatial-to-motor transformations also influenced auditory-motor performance, as indicated by distorted movement trajectories during the auditory postexposure phase. The distorted trajectories were present during auditory postexposure even though the auditory-motor relationship was not manipulated. Results suggest that by the age of 5 yr, children have developed a multisensory spatial-to-motor transformation for the execution of aiming movements toward both visual and acoustic targets.
Visual Working Memory Cannot Trade Quantity for Quality.
Ramaty, Ayelet; Luria, Roy
2018-01-01
Two main models have been proposed to describe how visual working memory (WM) allocates its capacity: the slot-model and the continuous resource-model. The purpose of the current study was to test a direct prediction of the resource model suggesting that WM can trade-off between the quantity and quality of the encoded information. Previous research reported equivocal results, with studies that failed to find such a trade-off and other studies that reported a trade-off. Following the design of previous studies, in Experiment 1 we replicated this trade-off, by presenting the memory array for 1200 ms. Experiment 2 failed to observe a trade-off between quantity and quality using a memory array interval of 300 ms (a standard interval for visual WM). Experiment 3 again failed to find this trade-off, when reinstating the 1200 ms memory array interval but adding an articulatory suppression manipulation. We argue that while participants can trade quantity for quality, this pattern depends on verbal encoding and transfer to long-term memory processes that were possible to perform only during the long retention interval. When these processes were eliminated, the trade-off disappeared. Thus, we didn't find any evidence that the trade-off between quantity for quality can occur within visual WM.
Audiovisual associations alter the perception of low-level visual motion
Kafaligonul, Hulusi; Oluk, Can
2015-01-01
Motion perception is a pervasive nature of vision and is affected by both immediate pattern of sensory inputs and prior experiences acquired through associations. Recently, several studies reported that an association can be established quickly between directions of visual motion and static sounds of distinct frequencies. After the association is formed, sounds are able to change the perceived direction of visual motion. To determine whether such rapidly acquired audiovisual associations and their subsequent influences on visual motion perception are dependent on the involvement of higher-order attentive tracking mechanisms, we designed psychophysical experiments using regular and reverse-phi random dot motions isolating low-level pre-attentive motion processing. Our results show that an association between the directions of low-level visual motion and static sounds can be formed and this audiovisual association alters the subsequent perception of low-level visual motion. These findings support the view that audiovisual associations are not restricted to high-level attention based motion system and early-level visual motion processing has some potential role. PMID:25873869
Azizian, Mahdi; Khoshnam, Mahta; Najmaei, Nima; Patel, Rajni V
2014-09-01
Intra-operative imaging is widely used to provide visual feedback to a clinician when he/she performs a procedure. In visual servoing, surgical instruments and parts of tissue/body are tracked by processing the acquired images. This information is then used within a control loop to manoeuvre a robotic manipulator during a procedure. A comprehensive search of electronic databases was completed for the period 2000-2013 to provide a survey of the visual servoing applications in medical robotics. The focus is on medical applications where image-based tracking is used for closed-loop control of a robotic system. Detailed classification and comparative study of various contributions in visual servoing using endoscopic or direct visual images are presented and summarized in tables and diagrams. The main challenges in using visual servoing for medical robotic applications are identified and potential future directions are suggested. 'Supervised automation of medical robotics' is found to be a major trend in this field. Copyright © 2013 John Wiley & Sons, Ltd.
Memory Performance for Everyday Motivational and Neutral Objects Is Dissociable from Attention
Schomaker, Judith; Wittmann, Bianca C.
2017-01-01
Episodic memory is typically better for items coupled with monetary reward or punishment during encoding. It is yet unclear whether memory is also enhanced for everyday objects with appetitive or aversive values learned through a lifetime of experience, and to what extent episodic memory enhancement for motivational and neutral items is attributable to attention. In a first experiment, we investigated attention to everyday motivational objects using eye-tracking during free-viewing and subsequently tested episodic memory using a remember/know procedure. Attention was directed more to aversive stimuli, as evidenced by longer viewing durations, whereas recollection was higher for both appetitive and aversive objects. In the second experiment, we manipulated the visual contrast of neutral objects through changes of contrast to further dissociate attention and memory encoding. While objects presented with high visual contrast were looked at longer, recollection was best for objects presented in unmodified, medium contrast. Generalized logistic mixed models on recollection performance showed that attention as measured by eye movements did not enhance subsequent memory, while motivational value (Experiment 1) and visual contrast (Experiment 2) had quadratic effects in opposite directions. Our findings suggest that an enhancement of incidental memory encoding for appetitive items can occur without an increase in attention and, vice versa, that enhanced attention towards salient neutral objects is not necessarily associated with memory improvement. Together, our results provide evidence for a double dissociation of attention and memory effects under certain conditions. PMID:28694774
Color, contrast sensitivity, and the cone mosaic.
Williams, D; Sekiguchi, N; Brainard, D
1993-01-01
This paper evaluates the role of various stages in the human visual system in the detection of spatial patterns. Contrast sensitivity measurements were made for interference fringe stimuli in three directions in color space with a psychophysical technique that avoided blurring by the eye's optics including chromatic aberration. These measurements were compared with the performance of an ideal observer that incorporated optical factors, such as photon catch in the cone mosaic, that influence the detection of interference fringes. The comparison of human and ideal observer performance showed that neural factors influence the shape as well as the height of the foveal contrast sensitivity function for all color directions, including those that involve luminance modulation. Furthermore, when optical factors are taken into account, the neural visual system has the same contrast sensitivity for isoluminant stimuli seen by the middle-wavelength-sensitive (M) and long-wavelength-sensitive (L) cones and isoluminant stimuli seen by the short-wavelength-sensitive (S) cones. Though the cone submosaics that feed these chromatic mechanisms have very different spatial properties, the later neural stages apparently have similar spatial properties. Finally, we review the evidence that cone sampling can produce aliasing distortion for gratings with spatial frequencies exceeding the resolution limit. Aliasing can be observed with gratings modulated in any of the three directions in color space we used. We discuss mechanisms that prevent aliasing in most ordinary viewing conditions. Images Fig. 1 Fig. 8 PMID:8234313
Attentional bias to betel quid cues: An eye tracking study.
Shen, Bin; Chiu, Meng-Chun; Li, Shuo-Heng; Huang, Guo-Joe; Liu, Ling-Jun; Ho, Ming-Chou
2016-09-01
The World Health Organization regards betel quid as a human carcinogen, and DSM-IV and ICD-10 dependence symptoms may develop with heavy use. This study, conducted in central Taiwan, investigated whether betel quid chewers can exhibit overt orienting to selectively respond to the betel quid cues. Twenty-four male chewers' and 23 male nonchewers' eye movements to betel-quid-related pictures and matched pictures were assessed during a visual probe task. The eye movement index showed that betel quid chewers were more likely to initially direct their gaze to the betel quid cues, t(23) = 3.70, p < .01, d = .75, and spent more time, F(1, 23) = 4.58, p < .05, η₂ = .17, and were more fixated, F(1, 23) = 5.18, p < .05, η₂ = .18, on them. The visual probe index (response time) failed to detect the chewers' attentional bias. The current study provided the first eye movement evidence of betel quid chewers' attentional bias. The results demonstrated that the betel quid chewers (but not the nonchewers) were more likely to initially direct their gaze to the betel quid cues, and spent more time and were more fixated on them. These findings suggested that when attention is directly measured through the eye tracking technique, this methodology may be more sensitive to detecting attentional biases in betel quid chewers. (PsycINFO Database Record (c) 2016 APA, all rights reserved).
De Sá Teixeira, Nuno Alexandre
2016-09-01
The memory for the final position of a moving object which suddenly disappears has been found to be displaced forward, in the direction of motion, and downwards, in the direction of gravity. These phenomena were coined, respectively, Representational Momentum and Representational Gravity. Although both these and similar effects have been systematically linked with the functioning of internal representations of physical variables (e.g. momentum and gravity), serious doubts have been raised for a cognitively based interpretation, favouring instead a major role of oculomotor and perceptual factors which, more often than not, were left uncontrolled and even ignored. The present work aims to determine the degree to which Representational Momentum and Representational Gravity are epiphenomenal to smooth pursuit eye movements. Observers were required to indicate the offset locations of targets moving along systematically varied directions after a variable imposed retention interval. Each participant completed the task twice, varying the eye movements' instructions: gaze was either constrained or left free to track the targets. A Fourier decomposition analysis of the localization responses was used to disentangle both phenomena. The results show unambiguously that constraining eye movements significantly eliminates the harmonic components which index Representational Momentum, but have no effect on Representational Gravity or its time course. The found outcomes offer promising prospects for the study of the visual representation of gravity and its neurological substrates.
Eye tracking to evaluate evidence recognition in crime scene investigations.
Watalingam, Renuka Devi; Richetelli, Nicole; Pelz, Jeff B; Speir, Jacqueline A
2017-11-01
Crime scene analysts are the core of criminal investigations; decisions made at the scene greatly affect the speed of analysis and the quality of conclusions, thereby directly impacting the successful resolution of a case. If an examiner fails to recognize the pertinence of an item on scene, the analyst's theory regarding the crime will be limited. Conversely, unselective evidence collection will most likely include irrelevant material, thus increasing a forensic laboratory's backlog and potentially sending the investigation into an unproductive and costly direction. Therefore, it is critical that analysts recognize and properly evaluate forensic evidence that can assess the relative support of differing hypotheses related to event reconstruction. With this in mind, the aim of this study was to determine if quantitative eye tracking data and qualitative reconstruction accuracy could be used to distinguish investigator expertise. In order to assess this, 32 participants were successfully recruited and categorized as experts or trained novices based on their practical experiences and educational backgrounds. Each volunteer then processed a mock crime scene while wearing a mobile eye tracker, wherein visual fixations, durations, search patterns, and reconstruction accuracy were evaluated. The eye tracking data (dwell time and task percentage on areas of interest or AOIs) were compared using Earth Mover's Distance (EMD) and the Needleman-Wunsch (N-W) algorithm, revealing significant group differences for both search duration (EMD), as well as search sequence (N-W). More specifically, experts exhibited greater dissimilarity in search duration, but greater similarity in search sequences than their novice counterparts. In addition to the quantitative visual assessment of examiner variability, each participant's reconstruction skill was assessed using a 22-point binary scoring system, in which significant group differences were detected as a function of total reconstruction accuracy. This result, coupled with the fact that the study failed to detect a significant difference between the groups when evaluating the total time needed to complete the investigation, indicates that experts are more efficient and effective. Finally, the results presented here provide a basis for continued research in the use of eye trackers to assess expertise in complex and distributed environments, including suggestions for future work, and cautions regarding the degree to which visual attention can infer cognitive understanding. Copyright © 2017 Elsevier B.V. All rights reserved.
The role of visual and direct force feedback in robotics-assisted mitral valve annuloplasty.
Currie, Maria E; Talasaz, Ali; Rayman, Reiza; Chu, Michael W A; Kiaii, Bob; Peters, Terry; Trejos, Ana Luisa; Patel, Rajni
2017-09-01
The objective of this work was to determine the effect of both direct force feedback and visual force feedback on the amount of force applied to mitral valve tissue during ex vivo robotics-assisted mitral valve annuloplasty. A force feedback-enabled master-slave surgical system was developed to provide both visual and direct force feedback during robotics-assisted cardiac surgery. This system measured the amount of force applied by novice and expert surgeons to cardiac tissue during ex vivo mitral valve annuloplasty repair. The addition of visual (2.16 ± 1.67), direct (1.62 ± 0.86), or both visual and direct force feedback (2.15 ± 1.08) resulted in lower mean maximum force applied to mitral valve tissue while suturing compared with no force feedback (3.34 ± 1.93 N; P < 0.05). To achieve better control of interaction forces on cardiac tissue during robotics-assisted mitral valve annuloplasty suturing, force feedback may be required. Copyright © 2016 John Wiley & Sons, Ltd.
Cortical visual dysfunction in children: a clinical study.
Dutton, G; Ballantyne, J; Boyd, G; Bradnam, M; Day, R; McCulloch, D; Mackie, R; Phillips, S; Saunders, K
1996-01-01
Damage to the cerebral cortex was responsible for impairment in vision in 90 of 130 consecutive children referred to the Vision Assessment Clinic in Glasgow. Cortical blindness was seen in 16 children. Only 2 were mobile, but both showed evidence of navigational blind-sight. Cortical visual impairment, in which it was possible to estimate visual acuity but generalised severe brain damage precluded estimation of cognitive visual function, was observed in 9 children. Complex disorders of cognitive vision were seen in 20 children. These could be divided into five categories and involved impairment of: (1) recognition, (2) orientation, (3) depth perception, (4) perception of movement and (5) simultaneous perception. These disorders were observed in a variety of combinations. The remaining children showed evidence of reduced visual acuity and/ or visual field loss, but without detectable disorders of congnitive visual function. Early recognition of disorders of cognitive vision is required if active training and remediation are to be implemented.
’What’ and ’Where’ in Visual Attention: Evidence from the Neglect Syndrome
1992-01-01
representations of the visual world, visual attention, and object representations. 24 Bauer, R. M., & Rubens, A. B. (1985). Agnosia . In K. M. Heilman, & E...visual information. Journal of Experimental Psychology: General, 1-1, 501-517. Farah, M. J. (1990). Visual Agnosia : Disorders of Object Recognition and
Stuart, Samuel; Galna, Brook; Delicato, Louise S; Lord, Sue; Rochester, Lynn
2017-07-01
Gait impairment is a core feature of Parkinson's disease (PD) which has been linked to cognitive and visual deficits, but interactions between these features are poorly understood. Monitoring saccades allows investigation of real-time cognitive and visual processes and their impact on gait when walking. This study explored: (i) saccade frequency when walking under different attentional manipulations of turning and dual-task; and (ii) direct and indirect relationships between saccades, gait impairment, vision and attention. Saccade frequency (number of fast eye movements per-second) was measured during gait in 60 PD and 40 age-matched control participants using a mobile eye-tracker. Saccade frequency was significantly reduced in PD compared to controls during all conditions. However, saccade frequency increased with a turn and decreased under dual-task for both groups. Poorer attention directly related to saccade frequency, visual function and gait impairment in PD, but not controls. Saccade frequency did not directly relate to gait in PD, but did in controls. Instead, saccade frequency and visual function deficit indirectly impacted gait impairment in PD, which was underpinned by their relationship with attention. In conclusion, our results suggest a vital role for attention with direct and indirect influences on gait impairment in PD. Attention directly impacted saccade frequency, visual function and gait impairment in PD, with connotations for falls. It also underpinned indirect impact of visual and saccadic impairment on gait. Attention therefore represents a key therapeutic target that should be considered in future research. © 2017 Federation of European Neuroscience Societies and John Wiley & Sons Ltd.
Matsui, Teppei; Ohki, Kenichi
2013-01-01
Higher order visual areas that receive input from the primary visual cortex (V1) are specialized for the processing of distinct features of visual information. However, it is still incompletely understood how this functional specialization is acquired. Here we used in vivo two photon calcium imaging in the mouse visual cortex to investigate whether this functional distinction exists at as early as the level of projections from V1 to two higher order visual areas, AL and LM. Specifically, we examined whether sharpness of orientation and direction selectivity and optimal spatial and temporal frequency of projection neurons from V1 to higher order visual areas match with that of target areas. We found that the V1 input to higher order visual areas were indeed functionally distinct: AL preferentially received inputs from V1 that were more orientation and direction selective and tuned for lower spatial frequency compared to projection of V1 to LM, consistent with functional differences between AL and LM. The present findings suggest that selective projections from V1 to higher order visual areas initiates parallel processing of sensory information in the visual cortical network. PMID:24068987
Evidence-Based Practices for Students with Visual Disabilities
ERIC Educational Resources Information Center
Ferrell, Kay Alicyn
2006-01-01
Forty years of peer-reviewed research in literacy and 50 years of research in mathematics were systematically analyzed to determine the evidence supporting instructional pedagogy for students with visual disabilities. Review teams identified 30 intervention studies that utilized an appropriate comparison group. Although some studies had large…
Kim, Ji Yeon; Kang, Eun Jin; Kwon, Oran
2010-01-01
In this study, we investigated that consumers could differentiate between levels of claims and clarify how a visual aid influences consumer understanding of the different claim levels. We interviewed 2,000 consumers in 13 shopping malls on their perception of and confidence in different levels of health claims using seven point scales. The average confidence scores given by participants were 4.17 for the probable level and 4.07 for the possible level; the score for the probable level was significantly higher than that for the possible level (P < 0.05). Scores for confidence in claims after reading labels with and without a visual aid were 5.27 and 4.43, respectively; the score for labeling with a visual aid was significantly higher than for labeling without a visual aid (P < 0.01). Our results provide compelling evidence that providing health claims with qualifying language differentiating levels of scientific evidence can help consumers understand the strength of scientific evidence behind those claims. Moreover, when a visual aid was included, consumers perceived the scientific levels more clearly and had greater confidence in their meanings than when a visual aid was not included. Although this result suggests that consumers react differently to different claim levels, it is not yet clear whether consumers understand the variations in the degree of scientific support. PMID:21103090
How visual cues for when to listen aid selective auditory attention.
Varghese, Lenny A; Ozmeral, Erol J; Best, Virginia; Shinn-Cunningham, Barbara G
2012-06-01
Visual cues are known to aid auditory processing when they provide direct information about signal content, as in lip reading. However, some studies hint that visual cues also aid auditory perception by guiding attention to the target in a mixture of similar sounds. The current study directly tests this idea for complex, nonspeech auditory signals, using a visual cue providing only timing information about the target. Listeners were asked to identify a target zebra finch bird song played at a random time within a longer, competing masker. Two different maskers were used: noise and a chorus of competing bird songs. On half of all trials, a visual cue indicated the timing of the target within the masker. For the noise masker, the visual cue did not affect performance when target and masker were from the same location, but improved performance when target and masker were in different locations. In contrast, for the chorus masker, visual cues improved performance only when target and masker were perceived as coming from the same direction. These results suggest that simple visual cues for when to listen improve target identification by enhancing sounds near the threshold of audibility when the target is energetically masked and by enhancing segregation when it is difficult to direct selective attention to the target. Visual cues help little when target and masker already differ in attributes that enable listeners to engage selective auditory attention effectively, including differences in spectrotemporal structure and in perceived location.
Gawryluk, Jodie R.; Mazerolle, Erin L.; D'Arcy, Ryan C. N.
2014-01-01
Functional magnetic resonance imaging (fMRI) is a non-invasive technique that allows for visualization of activated brain regions. Until recently, fMRI studies have focused on gray matter. There are two main reasons white matter fMRI remains controversial: (1) the blood oxygen level dependent (BOLD) fMRI signal depends on cerebral blood flow and volume, which are lower in white matter than gray matter and (2) fMRI signal has been associated with post-synaptic potentials (mainly localized in gray matter) as opposed to action potentials (the primary type of neural activity in white matter). Despite these observations, there is no direct evidence against measuring fMRI activation in white matter and reports of fMRI activation in white matter continue to increase. The questions underlying white matter fMRI activation are important. White matter fMRI activation has the potential to greatly expand the breadth of brain connectivity research, as well as improve the assessment and diagnosis of white matter and connectivity disorders. The current review provides an overview of the motivation to investigate white matter fMRI activation, as well as the published evidence of this phenomenon. We speculate on possible neurophysiologic bases of white matter fMRI signals, and discuss potential explanations for why reports of white matter fMRI activation are relatively scarce. We end with a discussion of future basic and clinical research directions in the study of white matter fMRI. PMID:25152709
Mossbridge, Julia; Zweig, Jacob; Grabowecky, Marcia; Suzuki, Satoru
2016-01-01
The perceptual system integrates synchronized auditory-visual signals in part to promote individuation of objects in cluttered environments. The processing of auditory-visual synchrony may more generally contribute to cognition by synchronizing internally generated multimodal signals. Reading is a prime example because the ability to synchronize internal phonological and/or lexical processing with visual orthographic processing may facilitate encoding of words and meanings. Consistent with this possibility, developmental and clinical research has suggested a link between reading performance and the ability to compare visual spatial/temporal patterns with auditory temporal patterns. Here, we provide converging behavioral and electrophysiological evidence suggesting that greater behavioral ability to judge auditory-visual synchrony (Experiment 1) and greater sensitivity of an electrophysiological marker of auditory-visual synchrony processing (Experiment 2) both predict superior reading comprehension performance, accounting for 16% and 25% of the variance, respectively. These results support the idea that the mechanisms that detect auditory-visual synchrony contribute to reading comprehension. PMID:28129060
Mossbridge, Julia; Zweig, Jacob; Grabowecky, Marcia; Suzuki, Satoru
2017-03-01
The perceptual system integrates synchronized auditory-visual signals in part to promote individuation of objects in cluttered environments. The processing of auditory-visual synchrony may more generally contribute to cognition by synchronizing internally generated multimodal signals. Reading is a prime example because the ability to synchronize internal phonological and/or lexical processing with visual orthographic processing may facilitate encoding of words and meanings. Consistent with this possibility, developmental and clinical research has suggested a link between reading performance and the ability to compare visual spatial/temporal patterns with auditory temporal patterns. Here, we provide converging behavioral and electrophysiological evidence suggesting that greater behavioral ability to judge auditory-visual synchrony (Experiment 1) and greater sensitivity of an electrophysiological marker of auditory-visual synchrony processing (Experiment 2) both predict superior reading comprehension performance, accounting for 16% and 25% of the variance, respectively. These results support the idea that the mechanisms that detect auditory-visual synchrony contribute to reading comprehension.
Audio–visual interactions for motion perception in depth modulate activity in visual area V3A
Ogawa, Akitoshi; Macaluso, Emiliano
2013-01-01
Multisensory signals can enhance the spatial perception of objects and events in the environment. Changes of visual size and auditory intensity provide us with the main cues about motion direction in depth. However, frequency changes in audition and binocular disparity in vision also contribute to the perception of motion in depth. Here, we presented subjects with several combinations of auditory and visual depth-cues to investigate multisensory interactions during processing of motion in depth. The task was to discriminate the direction of auditory motion in depth according to increasing or decreasing intensity. Rising or falling auditory frequency provided an additional within-audition cue that matched or did not match the intensity change (i.e. intensity-frequency (IF) “matched vs. unmatched” conditions). In two-thirds of the trials, a task-irrelevant visual stimulus moved either in the same or opposite direction of the auditory target, leading to audio–visual “congruent vs. incongruent” between-modalities depth-cues. Furthermore, these conditions were presented either with or without binocular disparity. Behavioral data showed that the best performance was observed in the audio–visual congruent condition with IF matched. Brain imaging results revealed maximal response in visual area V3A when all cues provided congruent and reliable depth information (i.e. audio–visual congruent, IF-matched condition including disparity cues). Analyses of effective connectivity revealed increased coupling from auditory cortex to V3A specifically in audio–visual congruent trials. We conclude that within- and between-modalities cues jointly contribute to the processing of motion direction in depth, and that they do so via dynamic changes of connectivity between visual and auditory cortices. PMID:23333414
Animate and Inanimate Objects in Human Visual Cortex: Evidence for Task-Independent Category Effects
ERIC Educational Resources Information Center
Wiggett, Alison J.; Pritchard, Iwan C.; Downing, Paul E.
2009-01-01
Evidence from neuropsychology suggests that the distinction between animate and inanimate kinds is fundamental to human cognition. Previous neuroimaging studies have reported that viewing animate objects activates ventrolateral visual brain regions, whereas inanimate objects activate ventromedial regions. However, these studies have typically…
The role of visual attention in multiple object tracking: evidence from ERPs.
Doran, Matthew M; Hoffman, James E
2010-01-01
We examined the role of visual attention in the multiple object tracking (MOT) task by measuring the amplitude of the N1 component of the event-related potential (ERP) to probe flashes presented on targets, distractors, or empty background areas. We found evidence that visual attention enhances targets and suppresses distractors (Experiment 1 & 3). However, we also found that when tracking load was light (two targets and two distractors), accurate tracking could be carried out without any apparent contribution from the visual attention system (Experiment 2). Our results suggest that attentional selection during MOT is flexibly determined by task demands as well as tracking load and that visual attention may not always be necessary for accurate tracking.
Executive and Perceptual Distraction in Visual Working Memory
2017-01-01
The contents of visual working memory are likely to reflect the influence of both executive control resources and information present in the environment. We investigated whether executive attention is critical in the ability to exclude unwanted stimuli by introducing concurrent potentially distracting irrelevant items to a visual working memory paradigm, and manipulating executive load using simple or more demanding secondary verbal tasks. Across 7 experiments varying in presentation format, timing, stimulus set, and distractor number, we observed clear disruptive effects of executive load and visual distraction, but relatively minimal evidence supporting an interactive relationship between these factors. These findings are in line with recent evidence using delay-based interference, and suggest that different forms of attentional selection operate relatively independently in visual working memory. PMID:28414499
Flight directions of passerine migrants in daylight and darkness: A radar and direct visual study
NASA Technical Reports Server (NTRS)
Gauthreaux, S. A., Jr.
1972-01-01
The application of radar and visual techniques to determine the migratory habits of passerine birds during daylight and darkness is discussed. The effects of wind on the direction of migration are examined. Scatter diagrams of daytime and nocturnal migration track directions correlated with wind direction are presented. It is concluded that migratory birds will fly at altitudes where wind direction and migratory direction are nearly the same. The effects of cloud cover and solar obscuration are considered negligible.
Hoffmann, Susanne; Vega-Zuniga, Tomas; Greiter, Wolfgang; Krabichler, Quirin; Bley, Alexandra; Matthes, Mariana; Zimmer, Christiane; Firzlaff, Uwe; Luksch, Harald
2016-11-01
The midbrain superior colliculus (SC) commonly features a retinotopic representation of visual space in its superficial layers, which is congruent with maps formed by multisensory neurons and motor neurons in its deep layers. Information flow between layers is suggested to enable the SC to mediate goal-directed orienting movements. While most mammals strongly rely on vision for orienting, some species such as echolocating bats have developed alternative strategies, which raises the question how sensory maps are organized in these animals. We probed the visual system of the echolocating bat Phyllostomus discolor and found that binocular high acuity vision is frontally oriented and thus aligned with the biosonar system, whereas monocular visual fields cover a large area of peripheral space. For the first time in echolocating bats, we could show that in contrast with other mammals, visual processing is restricted to the superficial layers of the SC. The topographic representation of visual space, however, followed the general mammalian pattern. In addition, we found a clear topographic representation of sound azimuth in the deeper collicular layers, which was congruent with the superficial visual space map and with a previously documented map of orienting movements. Especially for bats navigating at high speed in densely structured environments, it is vitally important to transfer and coordinate spatial information between sensors and motor systems. Here, we demonstrate first evidence for the existence of congruent maps of sensory space in the bat SC that might serve to generate a unified representation of the environment to guide motor actions. © 2016 Federation of European Neuroscience Societies and John Wiley & Sons Ltd.
Fischer-Baum, Simon; Englebretson, Robert
2016-08-01
Reading relies on the recognition of units larger than single letters and smaller than whole words. Previous research has linked sublexical structures in reading to properties of the visual system, specifically on the parallel processing of letters that the visual system enables. But whether the visual system is essential for this to happen, or whether the recognition of sublexical structures may emerge by other means, is an open question. To address this question, we investigate braille, a writing system that relies exclusively on the tactile rather than the visual modality. We provide experimental evidence demonstrating that adult readers of (English) braille are sensitive to sublexical units. Contrary to prior assumptions in the braille research literature, we find strong evidence that braille readers do indeed access sublexical structure, namely the processing of multi-cell contractions as single orthographic units and the recognition of morphemes within morphologically-complex words. Therefore, we conclude that the recognition of sublexical structure is not exclusively tied to the visual system. However, our findings also suggest that there are aspects of morphological processing on which braille and print readers differ, and that these differences may, crucially, be related to reading using the tactile rather than the visual sensory modality. Copyright © 2016 Elsevier B.V. All rights reserved.
When eyes drive hand: Influence of non-biological motion on visuo-motor coupling.
Thoret, Etienne; Aramaki, Mitsuko; Bringoux, Lionel; Ystad, Sølvi; Kronland-Martinet, Richard
2016-01-26
Many studies stressed that the human movement execution but also the perception of motion are constrained by specific kinematics. For instance, it has been shown that the visuo-manual tracking of a spotlight was optimal when the spotlight motion complies with biological rules such as the so-called 1/3 power law, establishing the co-variation between the velocity and the trajectory curvature of the movement. The visual or kinesthetic perception of a geometry induced by motion has also been shown to be constrained by such biological rules. In the present study, we investigated whether the geometry induced by the visuo-motor coupling of biological movements was also constrained by the 1/3 power law under visual open loop control, i.e. without visual feedback of arm displacement. We showed that when someone was asked to synchronize a drawing movement with a visual spotlight following a circular shape, the geometry of the reproduced shape was fooled by visual kinematics that did not respect the 1/3 power law. In particular, elliptical shapes were reproduced when the circle is trailed with a kinematics corresponding to an ellipse. Moreover, the distortions observed here were larger than in the perceptual tasks stressing the role of motor attractors in such a visuo-motor coupling. Finally, by investigating the direct influence of visual kinematics on the motor reproduction, our result conciliates previous knowledge on sensorimotor coupling of biological motions with external stimuli and gives evidence to the amodal encoding of biological motion. Copyright © 2015 Elsevier Ireland Ltd. All rights reserved.
Rauscher, Franziska G; Plant, Gordon T; James-Galton, Merle; Barbur, John L
2011-01-01
Damage to ventral occipito-temporal extrastriate visual cortex leads to the syndrome of prosopagnosia often with coexisting cerebral achromatopsia. A patient with this syndrome resulting in a left upper homonymous quadrantanopia, prosopagnosia, and incomplete achromatopsia is described. Chromatic sensitivity was assessed at a number of locations in the intact visual field using a dynamic luminance contrast masking technique that isolates the use of colour signals. In normal subjects chromatic detection thresholds form an elliptical contour when plotted in the Commission Internationale d'Eclairage, (x-y), chromaticity diagram. Because the extraction of colour signals in early visual processing involves opponent mechanisms, subjects with Daltonism (congenital red/green loss of sensitivity) show symmetric increase in thresholds towards the long wavelength ("red") and middle wavelength ("green") regions of the spectrum locus. This is also the case with acquired loss of chromatic sensitivity as a result of retinal or optic nerve disease. Our patient's results were an exception to this rule. Whilst his chromatic sensitivity in the central region of the visual field was reduced symmetrically for both "red/green" and "yellow/blue" directions in colour space, the subject's lower left quadrant showed a marked asymmetry in "red/green" thresholds with the greatest loss of sensitivity towards the "green" region of the spectrum locus. This spatially localized asymmetric loss of "green" but not "red" sensitivity has not been reported previously in human vision. Such loss is consistent with selective damage of neural substrates in the visual cortex that process colour information, but are spectrally non-opponent.
How Revisions to Mathematical Visuals Affect Cognition: Evidence from Eye Tracking
ERIC Educational Resources Information Center
Clinton, Virginia; Cooper, Jennifer L.; Michaelis, Joseph; Alibali, Martha W.; Nathan, Mitchell J.
2017-01-01
Mathematics curricula are frequently rich with visuals, but these visuals are often not designed for optimal use of students' limited cognitive resources. The authors of this study revised the visuals in a mathematics lesson based on instructional design principles. The purpose of this study is to examine the effects of these revised visuals on…
Nakashima, Ryoichi; Iwai, Ritsuko; Ueda, Sayako; Kumada, Takatsune
2015-01-01
When observers perceive several objects in a space, at the same time, they should effectively perceive their own position as a viewpoint. However, little is known about observers’ percepts of their own spatial location based on the visual scene information viewed from them. Previous studies indicate that two distinct visual spatial processes exist in the locomotion situation: the egocentric position perception and egocentric direction perception. Those studies examined such perceptions in information rich visual environments where much dynamic and static visual information was available. This study examined these two perceptions in information of impoverished environments, including only static lane edge information (i.e., limited information). We investigated the visual factors associated with static lane edge information that may affect these perceptions. Especially, we examined the effects of the two factors on egocentric direction and position perceptions. One is the “uprightness factor” that “far” visual information is seen at upper location than “near” visual information. The other is the “central vision factor” that observers usually look at “far” visual information using central vision (i.e., foveal vision) whereas ‘near’ visual information using peripheral vision. Experiment 1 examined the effect of the “uprightness factor” using normal and inverted road images. Experiment 2 examined the effect of the “central vision factor” using normal and transposed road images where the upper half of the normal image was presented under the lower half. Experiment 3 aimed to replicate the results of Experiments 1 and 2. Results showed that egocentric direction perception is interfered with image inversion or image transposition, whereas egocentric position perception is robust against these image transformations. That is, both “uprightness” and “central vision” factors are important for egocentric direction perception, but not for egocentric position perception. Therefore, the two visual spatial perceptions about observers’ own viewpoints are fundamentally dissociable. PMID:26648895
On the structural origins of ferroelectricity in HfO{sub 2} thin films
DOE Office of Scientific and Technical Information (OSTI.GOV)
Sang, Xiahan; Grimley, Everett D.; LeBeau, James M.
2015-04-20
Here, we present a structural study on the origin of ferroelectricity in Gd doped HfO{sub 2} thin films. We apply aberration corrected high-angle annular dark-field scanning transmission electron microscopy to directly determine the underlying lattice type using projected atom positions and measured lattice parameters. Furthermore, we apply nanoscale electron diffraction methods to visualize the crystal symmetry elements. Combined, the experimental results provide unambiguous evidence for the existence of a non-centrosymmetric orthorhombic phase that can support spontaneous polarization, resolving the origin of ferroelectricity in HfO{sub 2} thin films.
Why `false' colours are seen by butterflies
NASA Astrophysics Data System (ADS)
Kelber, Almut
1999-11-01
Light can be described by its intensity, spectral distribution and polarization, and normally a visual system analyses these independently to extract the maximum amount of information. Here I present behavioural evidence that this does not happen in butterflies, whose choice of oviposition substrate on the basis of its colour appears to be strongly influenced by the direction of polarization of the light reflected from the substrate. To my knowledge, this is the first record of `false' colours being perceived as a result of light polarization. This detection of false colours may help butterflies to find optimal oviposition sites.
Retinal Origin of Direction Selectivity in the Superior Colliculus
Shi, Xuefeng; Barchini, Jad; Ledesma, Hector Acaron; Koren, David; Jin, Yanjiao; Liu, Xiaorong; Wei, Wei; Cang, Jianhua
2017-01-01
Detecting visual features in the environment such as motion direction is crucial for survival. The circuit mechanisms that give rise to direction selectivity in a major visual center, the superior colliculus (SC), are entirely unknown. Here, we optogenetically isolate the retinal inputs that individual direction-selective SC neurons receive and find that they are already selective as a result of precisely converging inputs from similarly-tuned retinal ganglion cells. The direction selective retinal input is linearly amplified by the intracollicular circuits without changing its preferred direction or level of selectivity. Finally, using 2-photon calcium imaging, we show that SC direction selectivity is dramatically reduced in transgenic mice that have decreased retinal selectivity. Together, our studies demonstrate a retinal origin of direction selectivity in the SC, and reveal a central visual deficit as a consequence of altered feature selectivity in the retina. PMID:28192394
Hansen, Eva; Grimme, Britta; Reimann, Hendrik; Schöner, Gregor
2018-05-01
In a sequence of arm movements, any given segment could be influenced by its predecessors (carry-over coarticulation) and by its successor (anticipatory coarticulation). To study the interdependence of movement segments, we asked participants to move an object from an initial position to a first and then on to a second target location. The task involved ten joint angles controlling the three-dimensional spatial path of the object and hand. We applied the principle of the uncontrolled manifold (UCM) to analyze the difference between joint trajectories that either affect (non-motor equivalent) or do not affect (motor equivalent) the hand's trajectory in space. We found evidence for anticipatory coarticulation that was distributed equally in the two directions in joint space. We also found strong carry-over coarticulation, which showed clear structure in joint space: More of the difference between joint configurations observed for different preceding movements lies in directions in joint space that leaves the hand's path in space invariant than in orthogonal directions in joint space that varies the hand's path in space. We argue that the findings are consistent with anticipatory coarticulation reflecting processes of movement planning that lie at the level of the hand's trajectory in space. Carry-over coarticulation may reflect primarily processes of motor control that are governed by the principle of the UCM, according to which changes that do not affect the hand's trajectory in space are not actively delimited. Two follow-up experiments zoomed in on anticipatory coarticulation. These experiments strengthened evidence for anticipatory coarticulation. Anticipatory coarticulation was motor-equivalent when visual information supported the steering of the object to its first target, but was not motor equivalent when that information was removed. The experiments showed that visual updating of the hand's path in space when the object approaches the first target only affected the component of the joint difference vector orthogonal to the UCM, consistent with the UCM principle.
Harvey, Ben M; Dumoulin, Serge O
2016-02-15
Several studies demonstrate that visual stimulus motion affects neural receptive fields and fMRI response amplitudes. Here we unite results of these two approaches and extend them by examining the effects of visual motion on neural position preferences throughout the hierarchy of human visual field maps. We measured population receptive field (pRF) properties using high-field fMRI (7T), characterizing position preferences simultaneously over large regions of the visual cortex. We measured pRFs properties using sine wave gratings in stationary apertures, moving at various speeds in either the direction of pRF measurement or the orthogonal direction. We find direction- and speed-dependent changes in pRF preferred position and size in all visual field maps examined, including V1, V3A, and the MT+ map TO1. These effects on pRF properties increase up the hierarchy of visual field maps. However, both within and between visual field maps the extent of pRF changes was approximately proportional to pRF size. This suggests that visual motion transforms the representation of visual space similarly throughout the visual hierarchy. Visual motion can also produce an illusory displacement of perceived stimulus position. We demonstrate perceptual displacements using the same stimulus configuration. In contrast to effects on pRF properties, perceptual displacements show only weak effects of motion speed, with far larger speed-independent effects. We describe a model where low-level mechanisms could underlie the observed effects on neural position preferences. We conclude that visual motion induces similar transformations of visuo-spatial representations throughout the visual hierarchy, which may arise through low-level mechanisms. Copyright © 2015 Elsevier Inc. All rights reserved.
Priming and the guidance by visual and categorical templates in visual search.
Wilschut, Anna; Theeuwes, Jan; Olivers, Christian N L
2014-01-01
Visual search is thought to be guided by top-down templates that are held in visual working memory. Previous studies have shown that a search-guiding template can be rapidly and strongly implemented from a visual cue, whereas templates are less effective when based on categorical cues. Direct visual priming from cue to target may underlie this difference. In two experiments we first asked observers to remember two possible target colors. A postcue then indicated which of the two would be the relevant color. The task was to locate a briefly presented and masked target of the cued color among irrelevant distractor items. Experiment 1 showed that overall search accuracy improved more rapidly on the basis of a direct visual postcue that carried the target color, compared to a neutral postcue that pointed to the memorized color. However, selectivity toward the target feature, i.e., the extent to which observers searched selectively among items of the cued vs. uncued color, was found to be relatively unaffected by the presence of the visual signal. In Experiment 2 we compared search that was based on either visual or categorical information, but now controlled for direct visual priming. This resulted in no differences in overall performance nor selectivity. Altogether the results suggest that perceptual processing of visual search targets is facilitated by priming from visual cues, whereas attentional selectivity is enhanced by a working memory template that can formed from both visual and categorical input. Furthermore, if the priming is controlled for, categorical- and visual-based templates similarly enhance search guidance.
Spatial updating in area LIP is independent of saccade direction.
Heiser, Laura M; Colby, Carol L
2006-05-01
We explore the world around us by making rapid eye movements to objects of interest. Remarkably, these eye movements go unnoticed, and we perceive the world as stable. Spatial updating is one of the neural mechanisms that contributes to this perception of spatial constancy. Previous studies in macaque lateral intraparietal cortex (area LIP) have shown that individual neurons update, or "remap," the locations of salient visual stimuli at the time of an eye movement. The existence of remapping implies that neurons have access to visual information from regions far beyond the classically defined receptive field. We hypothesized that neurons have access to information located anywhere in the visual field. We tested this by recording the activity of LIP neurons while systematically varying the direction in which a stimulus location must be updated. Our primary finding is that individual neurons remap stimulus traces in multiple directions, indicating that LIP neurons have access to information throughout the visual field. At the population level, stimulus traces are updated in conjunction with all saccade directions, even when we consider direction as a function of receptive field location. These results show that spatial updating in LIP is effectively independent of saccade direction. Our findings support the hypothesis that the activity of LIP neurons contributes to the maintenance of spatial constancy throughout the visual field.
Implicit and Explicit Representations of Hand Position in Tool Use
Rand, Miya K.; Heuer, Herbert
2013-01-01
Understanding the interactions of visual and proprioceptive information in tool use is important as it is the basis for learning of the tool's kinematic transformation and thus skilled performance. This study investigated how the CNS combines seen cursor positions and felt hand positions under a visuo-motor rotation paradigm. Young and older adult participants performed aiming movements on a digitizer while looking at rotated visual feedback on a monitor. After each movement, they judged either the proprioceptively sensed hand direction or the visually sensed cursor direction. We identified asymmetric mutual biases with a strong visual dominance. Furthermore, we found a number of differences between explicit and implicit judgments of hand directions. The explicit judgments had considerably larger variability than the implicit judgments. The bias toward the cursor direction for the explicit judgments was about twice as strong as for the implicit judgments. The individual biases of explicit and implicit judgments were uncorrelated. Biases of these judgments exhibited opposite sequential effects. Moreover, age-related changes were also different between these judgments. The judgment variability was decreased and the bias toward the cursor direction was increased with increasing age only for the explicit judgments. These results indicate distinct explicit and implicit neural representations of hand direction, similar to the notion of distinct visual systems. PMID:23894307
Gaglianese, A; Costagli, M; Ueno, K; Ricciardi, E; Bernardi, G; Pietrini, P; Cheng, K
2015-01-22
The main visual pathway that conveys motion information to the middle temporal complex (hMT+) originates from the primary visual cortex (V1), which, in turn, receives spatial and temporal features of the perceived stimuli from the lateral geniculate nucleus (LGN). In addition, visual motion information reaches hMT+ directly from the thalamus, bypassing the V1, through a direct pathway. We aimed at elucidating whether this direct route between LGN and hMT+ represents a 'fast lane' reserved to high-speed motion, as proposed previously, or it is merely involved in processing motion information irrespective of speeds. We evaluated functional magnetic resonance imaging (fMRI) responses elicited by moving visual stimuli and applied connectivity analyses to investigate the effect of motion speed on the causal influence between LGN and hMT+, independent of V1, using the Conditional Granger Causality (CGC) in the presence of slow and fast visual stimuli. Our results showed that at least part of the visual motion information from LGN reaches hMT+, bypassing V1, in response to both slow and fast motion speeds of the perceived stimuli. We also investigated whether motion speeds have different effects on the connections between LGN and functional subdivisions within hMT+: direct connections between LGN and MT-proper carry mainly slow motion information, while connections between LGN and MST carry mainly fast motion information. The existence of a parallel pathway that connects the LGN directly to hMT+ in response to both slow and fast speeds may explain why MT and MST can still respond in the presence of V1 lesions. Copyright © 2014 IBRO. Published by Elsevier Ltd. All rights reserved.
Altered visual perception in long-term ecstasy (MDMA) users.
White, Claire; Brown, John; Edwards, Mark
2013-09-01
The present study investigated the long-term consequences of ecstasy use on visual processes thought to reflect serotonergic functions in the occipital lobe. Evidence indicates that the main psychoactive ingredient in ecstasy (methylendioxymethamphetamine) causes long-term changes to the serotonin system in human users. Previous research has found that amphetamine-abstinent ecstasy users have disrupted visual processing in the occipital lobe which relies on serotonin, with researchers concluding that ecstasy broadens orientation tuning bandwidths. However, other processes may have accounted for these results. The aim of the present research was to determine if amphetamine-abstinent ecstasy users have changes in occipital lobe functioning, as revealed by two studies: a masking study that directly measured the width of orientation tuning bandwidths and a contour integration task that measured the strength of long-range connections in the visual cortex of drug users compared to controls. Participants were compared on the width of orientation tuning bandwidths (26 controls, 12 ecstasy users, 10 ecstasy + amphetamine users) and the strength of long-range connections (38 controls, 15 ecstasy user, 12 ecstasy + amphetamine users) in the occipital lobe. Amphetamine-abstinent ecstasy users had significantly broader orientation tuning bandwidths than controls and significantly lower contour detection thresholds (CDTs), indicating worse performance on the task, than both controls and ecstasy + amphetamine users. These results extend on previous research, which is consistent with the proposal that ecstasy may damage the serotonin system, resulting in behavioral changes on tests of visual perception processes which are thought to reflect serotonergic functions in the occipital lobe.
Activity in human visual and parietal cortex reveals object-based attention in working memory.
Peters, Benjamin; Kaiser, Jochen; Rahm, Benjamin; Bledowski, Christoph
2015-02-25
Visual attention enables observers to select behaviorally relevant information based on spatial locations, features, or objects. Attentional selection is not limited to physically present visual information, but can also operate on internal representations maintained in working memory (WM) in service of higher-order cognition. However, only little is known about whether attention to WM contents follows the same principles as attention to sensory stimuli. To address this question, we investigated in humans whether the typically observed effects of object-based attention in perception are also evident for object-based attentional selection of internal object representations in WM. In full accordance with effects in visual perception, the key behavioral and neuronal characteristics of object-based attention were observed in WM. Specifically, we found that reaction times were shorter when shifting attention to memory positions located on the currently attended object compared with equidistant positions on a different object. Furthermore, functional magnetic resonance imaging and multivariate pattern analysis of visuotopic activity in visual (areas V1-V4) and parietal cortex revealed that directing attention to one position of an object held in WM also enhanced brain activation for other positions on the same object, suggesting that attentional selection in WM activates the entire object. This study demonstrated that all characteristic features of object-based attention are present in WM and thus follows the same principles as in perception. Copyright © 2015 the authors 0270-6474/15/353360-10$15.00/0.
Milne, Alice E; Petkov, Christopher I; Wilson, Benjamin
2017-07-05
Language flexibly supports the human ability to communicate using different sensory modalities, such as writing and reading in the visual modality and speaking and listening in the auditory domain. Although it has been argued that nonhuman primate communication abilities are inherently multisensory, direct behavioural comparisons between human and nonhuman primates are scant. Artificial grammar learning (AGL) tasks and statistical learning experiments can be used to emulate ordering relationships between words in a sentence. However, previous comparative work using such paradigms has primarily investigated sequence learning within a single sensory modality. We used an AGL paradigm to evaluate how humans and macaque monkeys learn and respond to identically structured sequences of either auditory or visual stimuli. In the auditory and visual experiments, we found that both species were sensitive to the ordering relationships between elements in the sequences. Moreover, the humans and monkeys produced largely similar response patterns to the visual and auditory sequences, indicating that the sequences are processed in comparable ways across the sensory modalities. These results provide evidence that human sequence processing abilities stem from an evolutionarily conserved capacity that appears to operate comparably across the sensory modalities in both human and nonhuman primates. The findings set the stage for future neurobiological studies to investigate the multisensory nature of these sequencing operations in nonhuman primates and how they compare to related processes in humans. Copyright © 2017 The Author(s). Published by Elsevier Ltd.. All rights reserved.
Familiarity Enhances Visual Working Memory for Faces
ERIC Educational Resources Information Center
Jackson, Margaret C.; Raymond, Jane E.
2008-01-01
Although it is intuitive that familiarity with complex visual objects should aid their preservation in visual working memory (WM), empirical evidence for this is lacking. This study used a conventional change-detection procedure to assess visual WM for unfamiliar and famous faces in healthy adults. Across experiments, faces were upright or…
Spelling: Do the Eyes Have It?
ERIC Educational Resources Information Center
Westwood, Peter
2015-01-01
This paper explores the question of whether the ability to spell depends mainly on visual perception and visual imagery, or on other equally important auditory, cognitive, and motor processes. The writer examines the evidence suggesting that accurate spelling draws on a combination of visual processing, visual memory, phonological awareness,…
ERP Evidence of Visualization at Early Stages of Visual Processing
ERIC Educational Resources Information Center
Page, Jonathan W.; Duhamel, Paul; Crognale, Michael A.
2011-01-01
Recent neuroimaging research suggests that early visual processing circuits are activated similarly during visualization and perception but have not demonstrated that the cortical activity is similar in character. We found functional equivalency in cortical activity by recording evoked potentials while color and luminance patterns were viewed and…
Potts, Geoffrey F; Wood, Susan M; Kothmann, Delia; Martin, Laura E
2008-10-21
Attention directs limited-capacity information processing resources to a subset of available perceptual representations. The mechanisms by which attention selects task-relevant representations for preferential processing are not fully known. Triesman and Gelade's [Triesman, A., Gelade, G., 1980. A feature integration theory of attention. Cognit. Psychol. 12, 97-136.] influential attention model posits that simple features are processed preattentively, in parallel, but that attention is required to serially conjoin multiple features into an object representation. Event-related potentials have provided evidence for this model showing parallel processing of perceptual features in the posterior Selection Negativity (SN) and serial, hierarchic processing of feature conjunctions in the Frontal Selection Positivity (FSP). Most prior studies have been done on conjunctions within one sensory modality while many real-world objects have multimodal features. It is not known if the same neural systems of posterior parallel processing of simple features and frontal serial processing of feature conjunctions seen within a sensory modality also operate on conjunctions between modalities. The current study used ERPs and simultaneously presented auditory and visual stimuli in three task conditions: Attend Auditory (auditory feature determines the target, visual features are irrelevant), Attend Visual (visual features relevant, auditory irrelevant), and Attend Conjunction (target defined by the co-occurrence of an auditory and a visual feature). In the Attend Conjunction condition when the auditory but not the visual feature was a target there was an SN over auditory cortex, when the visual but not auditory stimulus was a target there was an SN over visual cortex, and when both auditory and visual stimuli were targets (i.e. conjunction target) there were SNs over both auditory and visual cortex, indicating parallel processing of the simple features within each modality. In contrast, an FSP was present when either the visual only or both auditory and visual features were targets, but not when only the auditory stimulus was a target, indicating that the conjunction target determination was evaluated serially and hierarchically with visual information taking precedence. This indicates that the detection of a target defined by audio-visual conjunction is achieved via the same mechanism as within a single perceptual modality, through separate, parallel processing of the auditory and visual features and serial processing of the feature conjunction elements, rather than by evaluation of a fused multimodal percept.
Zeitoun, Jack H.; Kim, Hyungtae
2017-01-01
Binocular mechanisms for visual processing are thought to enhance spatial acuity by combining matched input from the two eyes. Studies in the primary visual cortex of carnivores and primates have confirmed that eye-specific neuronal response properties are largely matched. In recent years, the mouse has emerged as a prominent model for binocular visual processing, yet little is known about the spatial frequency tuning of binocular responses in mouse visual cortex. Using calcium imaging in awake mice of both sexes, we show that the spatial frequency preference of cortical responses to the contralateral eye is ∼35% higher than responses to the ipsilateral eye. Furthermore, we find that neurons in binocular visual cortex that respond only to the contralateral eye are tuned to higher spatial frequencies. Binocular neurons that are well matched in spatial frequency preference are also matched in orientation preference. In contrast, we observe that binocularly mismatched cells are more mismatched in orientation tuning. Furthermore, we find that contralateral responses are more direction-selective than ipsilateral responses and are strongly biased to the cardinal directions. The contralateral bias of high spatial frequency tuning was found in both awake and anesthetized recordings. The distinct properties of contralateral cortical responses may reflect the functional segregation of direction-selective, high spatial frequency-preferring neurons in earlier stages of the central visual pathway. Moreover, these results suggest that the development of binocularity and visual acuity may engage distinct circuits in the mouse visual system. SIGNIFICANCE STATEMENT Seeing through two eyes is thought to improve visual acuity by enhancing sensitivity to fine edges. Using calcium imaging of cellular responses in awake mice, we find surprising asymmetries in the spatial processing of eye-specific visual input in binocular primary visual cortex. The contralateral visual pathway is tuned to higher spatial frequencies than the ipsilateral pathway. At the highest spatial frequencies, the contralateral pathway strongly prefers to respond to visual stimuli along the cardinal (horizontal and vertical) axes. These results suggest that monocular, and not binocular, mechanisms set the limit of spatial acuity in mice. Furthermore, they suggest that the development of visual acuity and binocularity in mice involves different circuits. PMID:28924011
A Review of the Benefits of Nature Experiences: More Than Meets the Eye
Franco, Lara S.; Shanahan, Danielle F.
2017-01-01
Evidence that experiences of nature can benefit people has accumulated rapidly. Yet perhaps because of the domination of the visual sense in humans, most research has focused on the visual aspects of nature experiences. However, humans are multisensory, and it seems likely that many benefits are delivered through the non-visual senses and these are potentially avenues through which a physiological mechanism could occur. Here we review the evidence around these lesser studied sensory pathways—through sound, smell, taste, touch, and three non-sensory pathways. Natural sounds and smells underpin experiences of nature for many people, and this may well be rooted in evolutionary psychology. Tactile experiences of nature, particularly beyond animal petting, are understudied yet potentially fundamentally important. Tastes of nature, through growing and consuming natural foods, have been linked with a range of health and well-being benefits. Beyond the five senses, evidence is emerging for other non-visual pathways for nature experiences to be effective. These include ingestion or inhalation of phytoncides, negative air ions and microbes. We conclude that (i) these non-visual avenues are potentially important for delivering benefits from nature experiences; (ii) the evidence base is relatively weak and often based on correlational studies; and (iii) deeper exploration of these sensory and non-sensory avenues is needed. PMID:28763021
A Review of the Benefits of Nature Experiences: More Than Meets the Eye.
Franco, Lara S; Shanahan, Danielle F; Fuller, Richard A
2017-08-01
Evidence that experiences of nature can benefit people has accumulated rapidly. Yet perhaps because of the domination of the visual sense in humans, most research has focused on the visual aspects of nature experiences. However, humans are multisensory, and it seems likely that many benefits are delivered through the non-visual senses and these are potentially avenues through which a physiological mechanism could occur. Here we review the evidence around these lesser studied sensory pathways-through sound, smell, taste, touch, and three non-sensory pathways. Natural sounds and smells underpin experiences of nature for many people, and this may well be rooted in evolutionary psychology. Tactile experiences of nature, particularly beyond animal petting, are understudied yet potentially fundamentally important. Tastes of nature, through growing and consuming natural foods, have been linked with a range of health and well-being benefits. Beyond the five senses, evidence is emerging for other non-visual pathways for nature experiences to be effective. These include ingestion or inhalation of phytoncides, negative air ions and microbes. We conclude that (i) these non-visual avenues are potentially important for delivering benefits from nature experiences; (ii) the evidence base is relatively weak and often based on correlational studies; and (iii) deeper exploration of these sensory and non-sensory avenues is needed.
Interaction Junk: User Interaction-Based Evaluation of Visual Analytic Systems
DOE Office of Scientific and Technical Information (OSTI.GOV)
Endert, Alexander; North, Chris
2012-10-14
With the growing need for visualization to aid users in understanding large, complex datasets, the ability for users to interact and explore these datasets is critical. As visual analytic systems have advanced to leverage powerful computational models and data analytics capabilities, the modes by which users engage and interact with the information are limited. Often, users are taxed with directly manipulating parameters of these models through traditional GUIs (e.g., using sliders to directly manipulate the value of a parameter). However, the purpose of user interaction in visual analytic systems is to enable visual data exploration – where users can focusmore » on their task, as opposed to the tool or system. As a result, users can engage freely in data exploration and decision-making, for the purpose of gaining insight. In this position paper, we discuss how evaluating visual analytic systems can be approached through user interaction analysis, where the goal is to minimize the cognitive translation between the visual metaphor and the mode of interaction (i.e., reducing the “Interactionjunk”). We motivate this concept through a discussion of traditional GUIs used in visual analytics for direct manipulation of model parameters, and the importance of designing interactions the support visual data exploration.« less
On the Electrophysiological Evidence for the Capture of Visual Attention
ERIC Educational Resources Information Center
McDonald, John J.; Green, Jessica J.; Jannati, Ali; Di Lollo, Vincent
2013-01-01
The presence of a salient distractor interferes with visual search. According to the salience-driven selection hypothesis, this interference is because of an initial deployment of attention to the distractor. Three event-related potential (ERP) findings have been regarded as evidence for this hypothesis: (a) salient distractors were found to…
Priming Contour-Deleted Images: Evidence for Immediate Representations in Visual Object Recognition.
ERIC Educational Resources Information Center
Biederman, Irving; Cooper, Eric E.
1991-01-01
Speed and accuracy of identification of pictures of objects are facilitated by prior viewing. Contributions of image features, convex or concave components, and object models in a repetition priming task were explored in 2 studies involving 96 college students. Results provide evidence of intermediate representations in visual object recognition.…
NASA Astrophysics Data System (ADS)
Sargent, R.; Egge, M.; Dille, P. S.; O'Donnell, G. D.; Herwig, C.
2016-12-01
Visual evidence ignites curiosity and inspires advocacy. Zoomable imagery and video on a planetary scale provides compelling evidence of human impact on the environment. Earth Timelapse places the observable impact of 30+ years of human activity into the hands of policy makers, scientists, and advocates, with fluidity and speed that supports inquiry and exploration. Zoomability enables compelling narratives and ready apprehension of environmental changes, connecting human-scale evidence to regional and ecosystem-wide trends and changes. Leveraging the power of Google Earth Engine, join us to explore 30+ years of Landset 30m RGB imagery showing glacial retreat, agricultural deforestation, irrigation expansion, and the disappearance of lakes. These narratives are enriched with datasets showing planetary forest gain/loss, annual cycles of agricultural fires, global changes in the health of coral reefs, trends in resource extraction, and of renewable energy development. We demonstrate the intuitive and inquiry-enabling power of these planetary visualizations, and provide instruction on how scientists and advocates can create and share or contribute visualizations of their own research or topics of interest.
Fast visual prediction and slow optimization of preferred walking speed.
O'Connor, Shawn M; Donelan, J Maxwell
2012-05-01
People prefer walking speeds that minimize energetic cost. This may be accomplished by directly sensing metabolic rate and adapting gait to minimize it, but only slowly due to the compounded effects of sensing delays and iterative convergence. Visual and other sensory information is available more rapidly and could help predict which gait changes reduce energetic cost, but only approximately because it relies on prior experience and an indirect means to achieve economy. We used virtual reality to manipulate visually presented speed while 10 healthy subjects freely walked on a self-paced treadmill to test whether the nervous system beneficially combines these two mechanisms. Rather than manipulating the speed of visual flow directly, we coupled it to the walking speed selected by the subject and then manipulated the ratio between these two speeds. We then quantified the dynamics of walking speed adjustments in response to perturbations of the visual speed. For step changes in visual speed, subjects responded with rapid speed adjustments (lasting <2 s) and in a direction opposite to the perturbation and consistent with returning the visually presented speed toward their preferred walking speed, when visual speed was suddenly twice (one-half) the walking speed, subjects decreased (increased) their speed. Subjects did not maintain the new speed but instead gradually returned toward the speed preferred before the perturbation (lasting >300 s). The timing and direction of these responses strongly indicate that a rapid predictive process informed by visual feedback helps select preferred speed, perhaps to complement a slower optimization process that seeks to minimize energetic cost.
Extended Wearing Trial of Trifield Lens Device for “Tunnel Vision”
Woods, Russell L.; Giorgi, Robert G.; Berson, Eliot L.; Peli, Eli
2009-01-01
Severe visual field constriction (tunnel vision) impairs the ability to navigate and walk safely. We evaluated Trifield glasses as a mobility rehabilitation device for tunnel vision in an extended wearing trial. Twelve patients with tunnel vision (5 to 22 degrees wide) due to retinitis pigmentosa or choroideremia participated in the 5-visit wearing trial. To expand the horizontal visual field, one spectacle lens was fitted with two apex-to-apex prisms that vertically bisected the pupil on primary gaze. This provides visual field expansion at the expense of visual confusion (two objects with the same visual direction). Patients were asked to wear these spectacles as much as possible for the duration of the wearing trial (median 8, range 6 to 60, weeks). Clinical success (continued wear, indicating perceived overall benefit), visual field expansion, perceived direction and perceived visual ability were measured. Of 12 patients, 9 chose to continue wearing the Trifield glasses at the end of the wearing trial. Of those 9 patients, at long-term follow-up (35 to 78 weeks), 3 reported still wearing the Trifield glasses. Visual field expansion (median 18, range 9 to 38, degrees) was demonstrated for all patients. No patient demonstrated adaptation to the change in visual direction produced by the Trifield glasses (prisms). For difficulty with obstacles, some differences between successful and non-successful wearers were found. Trifield glasses provided reported benefits in obstacle avoidance to 7 of the 12 patients completing the wearing trial. Crowded environments were particularly difficult for most wearers. Possible reasons for long-term discontinuation and lack of adaptation to perceived direction are discussed. PMID:20444130
Extended wearing trial of Trifield lens device for 'tunnel vision'.
Woods, Russell L; Giorgi, Robert G; Berson, Eliot L; Peli, Eli
2010-05-01
Severe visual field constriction (tunnel vision) impairs the ability to navigate and walk safely. We evaluated Trifield glasses as a mobility rehabilitation device for tunnel vision in an extended wearing trial. Twelve patients with tunnel vision (5-22 degrees wide) due to retinitis pigmentosa or choroideremia participated in the 5-visit wearing trial. To expand the horizontal visual field, one spectacle lens was fitted with two apex-to-apex prisms that vertically bisected the pupil on primary gaze. This provides visual field expansion at the expense of visual confusion (two objects with the same visual direction). Patients were asked to wear these spectacles as much as possible for the duration of the wearing trial (median 8, range 6-60 weeks). Clinical success (continued wear, indicating perceived overall benefit), visual field expansion, perceived direction and perceived visual ability were measured. Of 12 patients, nine chose to continue wearing the Trifield glasses at the end of the wearing trial. Of those nine patients, at long-term follow-up (35-78 weeks), three reported still wearing the Trifield glasses. Visual field expansion (median 18, range 9-38 degrees) was demonstrated for all patients. No patient demonstrated adaptation to the change in visual direction produced by the Trifield glasses (prisms). For reported difficulty with obstacles, some differences between successful and non-successful wearers were found. Trifield glasses provided reported benefits in obstacle avoidance to 7 of the 12 patients completing the wearing trial. Crowded environments were particularly difficult for most wearers. Possible reasons for long-term discontinuation and lack of adaptation to perceived direction are discussed.
Iron homeostasis and eye disease
Loh, Allison; Hadziahmetovic, Majda; Dunaief, Joshua L.
2009-01-01
Summary Iron is necessary for life, but excess iron can be toxic to tissues. Iron is thought to damage tissues primarily by generating oxygen free radicals through the Fenton reaction. We present an overview of the evidence supporting iron's potential contribution to a broad range of eye disease using an anatomical approach. Firstly, iron can be visualized in the cornea as iron lines in the normal aging cornea as well as in diseases like keratoconus and pterygium. In the lens, we present the evidence for the role of oxidative damage in cataractogenesis. Also, we review the evidence that iron may play a role in the pathogenesis of the retinal disease age-related macular degeneration. Although currently there is no direct link between excess iron and development of optic neuropathies, ferrous iron's ability to form highly reactive oxygen species may play a role in optic nerve pathology. Lastly, we discuss recent advances in prevention and therapeutics for eye disease with antioxidants and iron chelators,. PMID:19059309
Sight Impairment registration due to stroke-A small yet significant rise?
Bunce, Catey; Zekite, Antra; Wormald, Richard; Rowe, Fiona
2017-12-01
In the United Kingdom, when an individual's sight falls to and remains at a certain threshold, they may be offered registration as sight impaired. Recent analysis of causes of registrable sight impairment in England/Wales indicated that visual impairment due to stroke had increased as a proportionate cause of sight loss. We aim to assess whether there is evidence of an increase in incidence of certification for sight impairment due to stroke in England/Wales between 2008 and 2014. The number of certifications with a main cause of sight impairment being stroke was obtained from the Certifications Office London. Directly standardized rates per 100,000 were computed with 95% confidence intervals and examined. Poisson regression was used to assess evidence of trend over time. In the year ending 31st March 2008, 992 people were newly certified with stroke with an estimated DSR of 2.1 (2.0 to 2.2) per 100,000 persons at risk. In the year ending March 31st 2014, there were 1310 certifications with a DSR of 2.5 (2.4 to 2.7). Figures were higher for men than women. Poisson regression indicated an estimated incidence rate ratio of 1.03 per year with 95% confidence intervals of 1.028 to 1.051, P < .001. These data suggest a small but statistically significant increase in the incidence of certifiable visual impairment due to stroke between 2008 and 2014. Figures are, however, considerably lower than estimated, perhaps suggesting that more should be done to address the visual needs of those who have suffered stroke.
Misperception of exocentric directions in auditory space
Arthur, Joeanna C.; Philbeck, John W.; Sargent, Jesse; Dopkins, Stephen
2008-01-01
Previous studies have demonstrated large errors (over 30°) in visually perceived exocentric directions (the direction between two objects that are both displaced from the observer’s location; e.g., Philbeck et al., in press). Here, we investigated whether a similar pattern occurs in auditory space. Blindfolded participants either attempted to aim a pointer at auditory targets (an exocentric task) or gave a verbal estimate of the egocentric target azimuth. Targets were located at 20° to 160° azimuth in the right hemispace. For comparison, we also collected pointing and verbal judgments for visual targets. We found that exocentric pointing responses exhibited sizeable undershooting errors, for both auditory and visual targets, that tended to become more strongly negative as azimuth increased (up to −19° for visual targets at 160°). Verbal estimates of the auditory and visual target azimuths, however, showed a dramatically different pattern, with relatively small overestimations of azimuths in the rear hemispace. At least some of the differences between verbal and pointing responses appear to be due to the frames of reference underlying the responses; when participants used the pointer to reproduce the egocentric target azimuth rather than the exocentric target direction relative to the pointer, the pattern of pointing errors more closely resembled that seen in verbal reports. These results show that there are similar distortions in perceiving exocentric directions in visual and auditory space. PMID:18555205
Change blindness and visual memory: visual representations get rich and act poor.
Varakin, D Alexander; Levin, Daniel T
2006-02-01
Change blindness is often taken as evidence that visual representations are impoverished, while successful recognition of specific objects is taken as evidence that they are richly detailed. In the current experiments, participants performed cover tasks that required each object in a display to be attended. Change detection trials were unexpectedly introduced and surprise recognition tests were given for nonchanging displays. For both change detection and recognition, participants had to distinguish objects from the same basic-level category, making it likely that specific visual information had to be used for successful performance. Although recognition was above chance, incidental change detection usually remained at floor. These results help reconcile demonstrations of poor change detection with demonstrations of good memory because they suggest that the capability to store visual information in memory is not reflected by the visual system's tendency to utilize these representations for purposes of detecting unexpected changes.
Douglas, Danielle; Newsome, Rachel N; Man, Louisa LY
2018-01-01
A significant body of research in cognitive neuroscience is aimed at understanding how object concepts are represented in the human brain. However, it remains unknown whether and where the visual and abstract conceptual features that define an object concept are integrated. We addressed this issue by comparing the neural pattern similarities among object-evoked fMRI responses with behavior-based models that independently captured the visual and conceptual similarities among these stimuli. Our results revealed evidence for distinctive coding of visual features in lateral occipital cortex, and conceptual features in the temporal pole and parahippocampal cortex. By contrast, we found evidence for integrative coding of visual and conceptual object features in perirhinal cortex. The neuroanatomical specificity of this effect was highlighted by results from a searchlight analysis. Taken together, our findings suggest that perirhinal cortex uniquely supports the representation of fully specified object concepts through the integration of their visual and conceptual features. PMID:29393853
Zhang, Xuetao; Huang, Jie; Yigit-Elliott, Serap; Rosenholtz, Ruth
2015-03-16
Observers can quickly search among shaded cubes for one lit from a unique direction. However, replace the cubes with similar 2-D patterns that do not appear to have a 3-D shape, and search difficulty increases. These results have challenged models of visual search and attention. We demonstrate that cube search displays differ from those with "equivalent" 2-D search items in terms of the informativeness of fairly low-level image statistics. This informativeness predicts peripheral discriminability of target-present from target-absent patches, which in turn predicts visual search performance, across a wide range of conditions. Comparing model performance on a number of classic search tasks, cube search does not appear unexpectedly easy. Easy cube search, per se, does not provide evidence for preattentive computation of 3-D scene properties. However, search asymmetries derived from rotating and/or flipping the cube search displays cannot be explained by the information in our current set of image statistics. This may merely suggest a need to modify the model's set of 2-D image statistics. Alternatively, it may be difficult cube search that provides evidence for preattentive computation of 3-D scene properties. By attributing 2-D luminance variations to a shaded 3-D shape, 3-D scene understanding may slow search for 2-D features of the target. © 2015 ARVO.
Marno, Hanna; Farroni, Teresa; Vidal Dos Santos, Yamil; Ekramnia, Milad; Nespor, Marina; Mehler, Jacques
2015-01-01
Infants’ sensitivity to selectively attend to human speech and to process it in a unique way has been widely reported in the past. However, in order to successfully acquire language, one should also understand that speech is a referential, and that words can stand for other entities in the world. While there has been some evidence showing that young infants can make inferences about the communicative intentions of a speaker, whether they would also appreciate the direct relationship between a specific word and its referent, is still unknown. In the present study we tested four-month-old infants to see whether they would expect to find a referent when they hear human speech. Our results showed that compared to other auditory stimuli or to silence, when infants were listening to speech they were more prepared to find some visual referents of the words, as signalled by their faster orienting towards the visual objects. Hence, our study is the first to report evidence that infants at a very young age already understand the referential relationship between auditory words and physical objects, thus show a precursor in appreciating the symbolic nature of language, even if they do not understand yet the meanings of words. PMID:26323990
The influence of visual motion on interceptive actions and perception.
Marinovic, Welber; Plooy, Annaliese M; Arnold, Derek H
2012-05-01
Visual information is an essential guide when interacting with moving objects, yet it can also be deceiving. For instance, motion can induce illusory position shifts, such that a moving ball can seem to have bounced past its true point of contact with the ground. Some evidence suggests illusory motion-induced position shifts bias pointing tasks to a greater extent than they do perceptual judgments. This, however, appears at odds with other findings and with our success when intercepting moving objects. Here we examined the accuracy of interceptive movements and of perceptual judgments in relation to simulated bounces. Participants were asked to intercept a moving disc at its bounce location by positioning a virtual paddle, and then to report where the disc had landed. Results showed that interceptive actions were accurate whereas perceptual judgments were inaccurate, biased in the direction of motion. Successful interceptions necessitated accurate information concerning both the location and timing of the bounce, so motor planning evidently had privileged access to an accurate forward model of bounce timing and location. This would explain why people can be accurate when intercepting a moving object, but lack insight into the accurate information that had guided their actions when asked to make a perceptual judgment. Copyright © 2012 Elsevier Ltd. All rights reserved.
Zhang, Xuetao; Huang, Jie; Yigit-Elliott, Serap; Rosenholtz, Ruth
2015-01-01
Observers can quickly search among shaded cubes for one lit from a unique direction. However, replace the cubes with similar 2-D patterns that do not appear to have a 3-D shape, and search difficulty increases. These results have challenged models of visual search and attention. We demonstrate that cube search displays differ from those with “equivalent” 2-D search items in terms of the informativeness of fairly low-level image statistics. This informativeness predicts peripheral discriminability of target-present from target-absent patches, which in turn predicts visual search performance, across a wide range of conditions. Comparing model performance on a number of classic search tasks, cube search does not appear unexpectedly easy. Easy cube search, per se, does not provide evidence for preattentive computation of 3-D scene properties. However, search asymmetries derived from rotating and/or flipping the cube search displays cannot be explained by the information in our current set of image statistics. This may merely suggest a need to modify the model's set of 2-D image statistics. Alternatively, it may be difficult cube search that provides evidence for preattentive computation of 3-D scene properties. By attributing 2-D luminance variations to a shaded 3-D shape, 3-D scene understanding may slow search for 2-D features of the target. PMID:25780063
Doerschner, K.; Boyaci, H.; Maloney, L. T.
2007-01-01
We investigated limits on the human visual system’s ability to discount directional variation in complex lights field when estimating Lambertian surface color. Directional variation in the light field was represented in the frequency domain using spherical harmonics. The bidirectional reflectance distribution function of a Lambertian surface acts as a low-pass filter on directional variation in the light field. Consequently, the visual system needs to discount only the low-pass component of the incident light corresponding to the first nine terms of a spherical harmonics expansion (Basri & Jacobs, 2001; Ramamoorthi & Hanrahan, 2001) to accurately estimate surface color. We test experimentally whether the visual system discounts directional variation in the light field up to this physical limit. Our results are consistent with the claim that the visual system can compensate for all of the complexity in the light field that affects the appearance of Lambertian surfaces. PMID:18053846
Ravens, Corvus corax, follow gaze direction of humans around obstacles.
Bugnyar, Thomas; Stöwe, Mareike; Heinrich, Bernd
2004-01-01
The ability to follow gaze (i.e. head and eye direction) has recently been shown for social mammals, particularly primates. In most studies, individuals could use gaze direction as a behavioural cue without understanding that the view of others may be different from their own. Here, we show that hand-raised ravens not only visually co-orient with the look-ups of a human experimenter but also reposition themselves to follow the experimenter's gaze around a visual barrier. Birds were capable of visual co-orientation already as fledglings but consistently tracked gaze direction behind obstacles not before six months of age. These results raise the possibility that sub-adult and adult ravens can project a line of sight for the other person into the distance. To what extent ravens may attribute mental significance to the visual behaviour of others is discussed. PMID:15306330
Mendoza-Halliday, Diego; Martinez-Trujillo, Julio C.
2017-01-01
The primate lateral prefrontal cortex (LPFC) encodes visual stimulus features while they are perceived and while they are maintained in working memory. However, it remains unclear whether perceived and memorized features are encoded by the same or different neurons and population activity patterns. Here we record LPFC neuronal activity while monkeys perceive the motion direction of a stimulus that remains visually available, or memorize the direction if the stimulus disappears. We find neurons with a wide variety of combinations of coding strength for perceived and memorized directions: some neurons encode both to similar degrees while others preferentially or exclusively encode either one. Reading out the combined activity of all neurons, a machine-learning algorithm reliably decode the motion direction and determine whether it is perceived or memorized. Our results indicate that a functionally diverse population of LPFC neurons provides a substrate for discriminating between perceptual and mnemonic representations of visual features. PMID:28569756
Zator, Krysten; Katz, Albert N
2017-07-01
Here, we examined linguistic differences in the reports of memories produced by three cueing methods. Two groups of young adults were cued visually either by words representing events or popular cultural phenomena that took place when they were 5, 10, or 16 years of age, or by words referencing a general lifetime period word cue directing them to that period in their life. A third group heard 30-second long musical clips of songs popular during the same three time periods. In each condition, participants typed a specific event memory evoked by the cue and these typed memories were subjected to analysis by the Linguistic Inquiry and Word Count (LIWC) program. Differences in the reports produced indicated that listening to music evoked memories embodied in motor-perceptual systems more so than memories evoked by our word-cueing conditions. Additionally, relative to music cues, lifetime period word cues produced memories with reliably more uses of personal pronouns, past tense terms, and negative emotions. The findings provide evidence for the embodiment of autobiographical memories, and how those differ when the cues emphasise different aspects of the encoded events.
Williams, Rebecca J; Reutens, David C; Hocking, Julia
2015-11-01
Decreased water displacement following increased neural activity has been observed using diffusion-weighted functional MRI (DfMRI) at high b-values. The physiological mechanisms underlying the diffusion signal change may be unique from the standard blood oxygenation level-dependent (BOLD) contrast and closer to the source of neural activity. Whether DfMRI reflects neural activity more directly than BOLD outside the primary cerebral regions remains unclear. Colored and achromatic Mondrian visual stimuli were statistically contrasted to functionally localize the human color center Area V4 in neurologically intact adults. Spatial and temporal properties of DfMRI and BOLD activation were examined across regions of the visual cortex. At the individual level, DfMRI activation patterns showed greater spatial specificity to V4 than BOLD. The BOLD activation patterns were more prominent in the primary visual cortex than DfMRI, where activation was localized to the ventral temporal lobe. Temporally, the diffusion signal change in V4 and V1 both preceded the corresponding hemodynamic response, however the early diffusion signal change was more evident in V1. DfMRI may be of use in imaging applications implementing cognitive subtraction paradigms, and where highly precise individual functional localization is required.
Ageing and feature binding in visual working memory: The role of presentation time.
Rhodes, Stephen; Parra, Mario A; Logie, Robert H
2016-01-01
A large body of research has clearly demonstrated that healthy ageing is accompanied by an associative memory deficit. Older adults exhibit disproportionately poor performance on memory tasks requiring the retention of associations between items (e.g., pairs of unrelated words). In contrast to this robust deficit, older adults' ability to form and temporarily hold bound representations of an object's surface features, such as colour and shape, appears to be relatively well preserved. However, the findings of one set of experiments suggest that older adults may struggle to form temporary bound representations in visual working memory when given more time to study objects. However, these findings were based on between-participant comparisons across experimental paradigms. The present study directly assesses the role of presentation time in the ability of younger and older adults to bind shape and colour in visual working memory using a within-participant design. We report new evidence that giving older adults longer to study memory objects does not differentially affect their immediate memory for feature combinations relative to individual features. This is in line with a growing body of research suggesting that there is no age-related impairment in immediate memory for colour-shape binding.
Xiao, Youping; Kavanau, Christopher; Bertin, Lauren; Kaplan, Ehud
2011-01-01
Many studies have provided evidence for the existence of universal constraints on color categorization or naming in various languages, but the biological basis of these constraints is unknown. A recent study of the pattern of color categorization across numerous languages has suggested that these patterns tend to avoid straddling a region in color space at or near the border between the English composite categories of "warm" and "cool". This fault line in color space represents a fundamental constraint on color naming. Here we report that the two-way categorization along the fault line is correlated with the sign of the L- versus M-cone contrast of a stimulus color. Moreover, we found that the sign of the L-M cone contrast also accounted for the two-way clustering of the spatially distributed neural responses in small regions of the macaque primary visual cortex, visualized with optical imaging. These small regions correspond to the hue maps, where our previous study found a spatially organized representation of stimulus hue. Altogether, these results establish a direct link between a universal constraint on color naming and the cone-specific information that is represented in the primate early visual system.
Visual navigation in starfish: first evidence for the use of vision and eyes in starfish
Garm, Anders; Nilsson, Dan-Eric
2014-01-01
Most known starfish species possess a compound eye at the tip of each arm, which, except for the lack of true optics, resembles an arthropod compound eye. Although these compound eyes have been known for about two centuries, no visually guided behaviour has ever been directly associated with their presence. There are indications that they are involved in negative phototaxis but this may also be governed by extraocular photoreceptors. Here, we show that the eyes of the coral-reef-associated starfish Linckia laevigata are slow and colour blind. The eyes are capable of true image formation although with low spatial resolution. Further, our behavioural experiments reveal that only specimens with intact eyes can navigate back to their reef habitat when displaced, demonstrating that this is a visually guided behaviour. This is, to our knowledge, the first report of a function of starfish compound eyes. We also show that the spectral sensitivity optimizes the contrast between the reef and the open ocean. Our results provide an example of an eye supporting only low-resolution vision, which is believed to be an essential stage in eye evolution, preceding the high-resolution vision required for detecting prey, predators and conspecifics. PMID:24403344
Teramoto, Wataru; Watanabe, Hiroshi; Umemura, Hiroyuki
2008-01-01
The perceived temporal order of external successive events does not always follow their physical temporal order. We examined the contribution of self-motion mechanisms in the perception of temporal order in the auditory modality. We measured perceptual biases in the judgment of the temporal order of two short sounds presented successively, while participants experienced visually induced self-motion (yaw-axis circular vection) elicited by viewing long-lasting large-field visual motion. In experiment 1, a pair of white-noise patterns was presented to participants at various stimulus-onset asynchronies through headphones, while they experienced visually induced self-motion. Perceived temporal order of auditory events was modulated by the direction of the visual motion (or self-motion). Specifically, the sound presented to the ear in the direction opposite to the visual motion (ie heading direction) was perceived prior to the sound presented to the ear in the same direction. Experiments 2A and 2B were designed to reduce the contributions of decisional and/or response processes. In experiment 2A, the directional cueing of the background (left or right) and the response dimension (high pitch or low pitch) were not spatially associated. In experiment 2B, participants were additionally asked to report which of the two sounds was perceived 'second'. Almost the same results as in experiment 1 were observed, suggesting that the change in temporal order of auditory events during large-field visual motion reflects a change in perceptual processing. Experiment 3 showed that the biases in the temporal-order judgments of auditory events were caused by concurrent actual self-motion with a rotatory chair. In experiment 4, using a small display, we showed that 'pure' long exposure to visual motion without the sensation of self-motion was not responsible for this phenomenon. These results are consistent with previous studies reporting a change in the perceived temporal order of visual or tactile events depending on the direction of self-motion. Hence, large-field induced (ie optic flow) self-motion can affect the temporal order of successive external events across various modalities.
Choice reaction time to visual motion during prolonged rotary motion in airline pilots
NASA Technical Reports Server (NTRS)
Stewart, J. D.; Clark, B.
1975-01-01
Thirteen airline pilots were studied to determine the effect of preceding rotary accelerations on the choice reaction time to the horizontal acceleration of a vertical line on a cathode-ray tube. On each trial, one of three levels of rotary and visual acceleration was presented with the rotary stimulus preceding the visual by one of seven periods. The two accelerations were always equal and were presented in the same or opposite directions. The reaction time was found to increase with increases in the time the rotary acceleration preceded the visual acceleration, and to decrease with increased levels of visual and rotary acceleration. The reaction time was found to be shorter when the accelerations were in the same direction than when they were in opposite directions. These results suggest that these findings are a special case of a general effect that the authors have termed 'gyrovisual modulation'.
Influence of Visual Prism Adaptation on Auditory Space Representation.
Pochopien, Klaudia; Fahle, Manfred
2017-01-01
Prisms shifting the visual input sideways produce a mismatch between the visual versus felt position of one's hand. Prism adaptation eliminates this mismatch, realigning hand proprioception with visual input. Whether this realignment concerns exclusively the visuo-(hand)motor system or it generalizes to acoustic inputs is controversial. We here show that there is indeed a slight influence of visual adaptation on the perceived direction of acoustic sources. However, this shift in perceived auditory direction can be fully explained by a subconscious head rotation during prism exposure and by changes in arm proprioception. Hence, prism adaptation does only indirectly generalize to auditory space perception.
NASA Astrophysics Data System (ADS)
Stork, David G.; Nagy, Gabor
2010-02-01
We explored the working methods of the Italian Baroque master Caravaggio through computer graphics reconstruction of his studio, with special focus on his use of lighting and illumination in The calling of St. Matthew. Although he surely took artistic liberties while constructing this and other works and did not strive to provide a "photographic" rendering of the tableau before him, there are nevertheless numerous visual clues to the likely studio conditions and working methods within the painting: the falloff of brightness along the rear wall, the relative brightness of the faces of figures, and the variation in sharpness of cast shadows (i.e., umbrae and penumbrae). We explored two studio lighting hypotheses: that the primary illumination was local (and hence artificial) and that it was distant solar. We find that the visual evidence can be consistent with local (artificial) illumination if Caravaggio painted his figures separately, adjusting the brightness on each to compensate for the falloff in illumination. Alternatively, the evidence is consistent with solar illumination only if the rear wall had particular reflectance properties, as described by a bi-directional reflectance distribution function, BRDF. (Ours is the first research applying computer graphics to the understanding of artists' praxis that models subtle reflectance properties of surfaces through BRDFs, a technique that may find use in studies of other artists.) A somewhat puzzling visual feature-unnoted in the scholarly literature-is the upward-slanting cast shadow in the upper-right corner of the painting. We found this shadow is naturally consistent with a local illuminant passing through a small window perpendicular to the viewer's line of sight, but could also be consistent with solar illumination if the shadow was due to a slanted, overhanging section of a roof outside the artist's studio. Our results place likely conditions upon any hypotheses concerning Caravaggio's working methods and point to new sources of evidence that could be confirmed or disconfirmed by future art historical research.
ERIC Educational Resources Information Center
Cattaneo, Zaira; Mattavelli, Giulia; Papagno, Costanza; Herbert, Andrew; Silvanto, Juha
2011-01-01
The human visual system is able to efficiently extract symmetry information from the visual environment. Prior neuroimaging evidence has revealed symmetry-preferring neuronal representations in the dorsolateral extrastriate visual cortex; the objective of the present study was to investigate the necessity of these representations in symmetry…
ERIC Educational Resources Information Center
Chen, Y.; Norton, D. J.; McBain, R.; Gold, J.; Frazier, J. A.; Coyle, J. T.
2012-01-01
An important issue for understanding visual perception in autism concerns whether individuals with this neurodevelopmental disorder possess an advantage in processing local visual information, and if so, what is the nature of this advantage. Perception of movement speed is a visual process that relies on computation of local spatiotemporal signals…
How visual working memory contents influence priming of visual attention.
Carlisle, Nancy B; Kristjánsson, Árni
2017-04-12
Recent evidence shows that when the contents of visual working memory overlap with targets and distractors in a pop-out search task, intertrial priming is inhibited (Kristjánsson, Sævarsson & Driver, Psychon Bull Rev 20(3):514-521, 2013, Experiment 2, Psychonomic Bulletin and Review). This may reflect an interesting interaction between implicit short-term memory-thought to underlie intertrial priming-and explicit visual working memory. Evidence from a non-pop-out search task suggests that it may specifically be holding distractors in visual working memory that disrupts intertrial priming (Cunningham & Egeth, Psychol Sci 27(4):476-485, 2016, Experiment 2, Psychological Science). We examined whether the inhibition of priming depends on whether feature values in visual working memory overlap with targets or distractors in the pop-out search, and we found that the inhibition of priming resulted from holding distractors in visual working memory. These results are consistent with separate mechanisms of target and distractor effects in intertrial priming, and support the notion that the impact of implicit short-term memory and explicit visual working memory can interact when each provides conflicting attentional signals.
Effects of Visual Information on Wind-Evoked Escape Behavior of the Cricket, Gryllus bimaculatus.
Kanou, Masamichi; Matsuyama, Akane; Takuwa, Hiroyuki
2014-09-01
We investigated the effects of visual information on wind-evoked escape behavior in the cricket, Gryllus bimaculatus. Most agitated crickets were found to retreat into a shelter made of cardboard installed in the test arena within a short time. As this behavior was thought to be a type of escape, we confirmed how a visual image of a shelter affected wind-evoked escape behavior. Irrespective of the brightness of the visual background (black or white) or the absence or presence of a shelter, escape jumps were oriented almost 180° opposite to the source of the air puff stimulus. Therefore, the direction of wind-evoked escape depends solely depended on the direction of the stimulus air puff. In contrast, the turning direction of the crickets during the escape was affected by the position of the visual image of the shelter. During the wind-evoked escape jump, most crickets turned in the direction in which a shelter was presented. This behavioral nature is presumably necessary for crickets to retreat into a shelter within a short time after their escape jump.
Effects of Hand Proximity and Movement Direction in Spatial and Temporal Gap Discrimination.
Wiemers, Michael; Fischer, Martin H
2016-01-01
Previous research on the interplay between static manual postures and visual attention revealed enhanced visual selection near the hands (near-hand effect). During active movements there is also superior visual performance when moving toward compared to away from the stimulus (direction effect). The "modulated visual pathways" hypothesis argues that differential involvement of magno- and parvocellular visual processing streams causes the near-hand effect. The key finding supporting this hypothesis is an increase in temporal and a reduction in spatial processing in near-hand space (Gozli et al., 2012). Since this hypothesis has, so far, only been tested with static hand postures, we provide a conceptual replication of Gozli et al.'s (2012) result with moving hands, thus also probing the generality of the direction effect. Participants performed temporal or spatial gap discriminations while their right hand was moving below the display. In contrast to Gozli et al. (2012), temporal gap discrimination was superior at intermediate and not near hand proximity. In spatial gap discrimination, a direction effect without hand proximity effect suggests that pragmatic attentional maps overshadowed temporal/spatial processing biases for far/near-hand space.
Valdois, Sylviane; Lassus-Sangosse, Delphine; Lobier, Muriel
2012-05-01
Poor parallel letter-string processing in developmental dyslexia was taken as evidence of poor visual attention (VA) span, that is, a limitation of visual attentional resources that affects multi-character processing. However, the use of letter stimuli in oral report tasks was challenged on its capacity to highlight a VA span disorder. In particular, report of poor letter/digit-string processing but preserved symbol-string processing was viewed as evidence of poor visual-to-phonology code mapping, in line with the phonological theory of developmental dyslexia. We assessed here the visual-to-phonological-code mapping disorder hypothesis. In Experiment 1, letter-string, digit-string and colour-string processing was assessed to disentangle a phonological versus visual familiarity account of the letter/digit versus symbol dissociation. Against a visual-to-phonological-code mapping disorder but in support of a familiarity account, results showed poor letter/digit-string processing but preserved colour-string processing in dyslexic children. In Experiment 2, two tasks of letter-string report were used, one of which was performed simultaneously to a high-taxing phonological task. Results show that dyslexic children are similarly impaired in letter-string report whether a concurrent phonological task is simultaneously performed or not. Taken together, these results provide strong evidence against a phonological account of poor letter-string processing in developmental dyslexia. Copyright © 2012 John Wiley & Sons, Ltd.
Kessel, Line; Andresen, Jens; Erngaard, Ditte; Flesner, Per; Tendal, Britta; Hjortdal, Jesper
2016-02-01
The need for cataract surgery is expected to rise dramatically in the future due to the increasing proportion of elderly citizens and increasing demands for optimum visual function. The aim of this study was to provide an evidence-based recommendation for the indication of cataract surgery based on which group of patients are most likely to benefit from surgery. A systematic literature search was performed in the MEDLINE, CINAHL, EMBASE and COCHRANE LIBRARY databases. Studies evaluating the outcome after cataract surgery according to preoperative visual acuity and visual complaints were included in a meta-analysis. We identified eight observational studies comparing outcome after cataract surgery in patients with poor (<20/40) and fair (>20/40) preoperative visual acuity. We could not find any studies that compared outcome after cataract surgery in patients with few or many preoperative visual complaints. A meta-analysis showed that the outcome of cataract surgery, evaluated as objective and subjective visual improvement, was independent on preoperative visual acuity. There is a lack of scientific evidence to guide the clinician in deciding which patients are most likely to benefit from surgery. To overcome this shortage of evidence, many systems have been developed internationally to prioritize patients on waiting lists for cataract surgery, but the Swedish NIKE (Nationell Indikationsmodell för Katarakt Ekstraktion) is the only system where an association to the preoperative scoring of a patient has been related to outcome of cataract surgery. We advise that clinicians are inspired by the NIKE system when they decide which patients to operate to ensure that surgery is only offered to patients who are expected to benefit from cataract surgery. © 2015 The Authors. Acta Ophthalmologica published by John Wiley & Sons Ltd on behalf of Acta Ophthalmologica Scandinavica Foundation.
Fatty acid chain length, postprandial satiety and food intake in lean men.
Poppitt, S D; Strik, C M; MacGibbon, A K H; McArdle, B H; Budgett, S C; McGill, A-T
2010-08-04
High-fat diets are associated with obesity, and the weak satiety response elicited in response to dietary lipids is likely to play a role. Preliminary evidence from studies of medium (MCT) and long chain triglycerides (LCT) supports greater appetite suppression on high-MCT diets, possibly a consequence of direct portal access, more rapid oxidation and muted lipaemia. No data is as yet available on high-SCT diets which also have direct hepatic access. In this study SCT- (dairy fats), MCT- (coconut oil) and LCT-enriched (beef tallow) test breakfasts (3.3 MJ) containing 52 g lipid (58 en% fat) were investigated in a randomized, cross-over study in 18 lean men. All participants were required to complete the 3 study days in randomised order. Participants rated appetite sensations using visual analogue scales (VAS), and energy intake (EI) was measured by covert weighing of an ad libitum lunch meal 3.5 h postprandially. Blood samples were collected by venous cannulation. There were no detectable differences between breakfasts in perceived pleasantness, visual appearance, smell, taste, aftertaste and palatability (P>0.05). There was no significant effect of fatty acid chain length on ratings of hunger, fullness, satisfaction or current thoughts of food, nor did energy (mean, sem: SCT: 4406, 366 kJ; MCT: 4422, 306 kJ; LCT: 4490, 324 kJ; P>0.05) or macronutrient intake at lunch differ between diets. The maximum difference in EI between diets was less than 2%. Postprandial lipaemia also did not differ significantly. We conclude that there was no evidence that fatty acid chain length has an effect on measures of appetite and food intake when assessed following a single high-fat test meal in lean participants. Copyright 2010 Elsevier Inc. All rights reserved.
Neural mechanisms of oculomotor abnormalities in the infantile strabismus syndrome.
Walton, Mark M G; Pallus, Adam; Fleuriet, Jérome; Mustari, Michael J; Tarczy-Hornoch, Kristina
2017-07-01
Infantile strabismus is characterized by numerous visual and oculomotor abnormalities. Recently nonhuman primate models of infantile strabismus have been established, with characteristics that closely match those observed in human patients. This has made it possible to study the neural basis for visual and oculomotor symptoms in infantile strabismus. In this review, we consider the available evidence for neural abnormalities in structures related to oculomotor pathways ranging from visual cortex to oculomotor nuclei. These studies provide compelling evidence that a disturbance of binocular vision during a sensitive period early in life, whatever the cause, results in a cascade of abnormalities through numerous brain areas involved in visual functions and eye movements. Copyright © 2017 the American Physiological Society.
ERIC Educational Resources Information Center
Parker, Amy T.; Grimmett, Eric S.; Summers, Sharon
2008-01-01
This review examines practices for building effective communication strategies for children with visual impairments, including those with additional disabilities, that have been tested by single-subject design methodology. The authors found 30 studies that met the search criteria and grouped intervention strategies to align any evidence of the…
ERIC Educational Resources Information Center
Spriggs, Amy D.; Mims, Pamela J.; van Dijk, Wilhelmina; Knight, Victoria F.
2017-01-01
We conducted a comprehensive review of the literature to establish the evidence base for using visual activity schedules (VAS) with individuals with intellectual disability. Literature published after 2005 was evaluated for quality using the criteria developed by Horner et al.; a total of 14 studies were included as acceptable. Findings suggest…
ERIC Educational Resources Information Center
Stein, Timo; Sterzer, Philipp; Peelen, Marius V.
2012-01-01
The rapid visual detection of other people in our environment is an important first step in social cognition. Here we provide evidence for selective sensitivity of the human visual system to upright depictions of conspecifics. In a series of seven experiments, we assessed the impact of stimulus inversion on the detection of person silhouettes,…
The spread of attention across features of a surface
Ernst, Zachary Raymond; Jazayeri, Mehrdad
2013-01-01
Contrasting theories of visual attention have emphasized selection by spatial location, individual features, and whole objects. We used functional magnetic resonance imaging to ask whether and how attention to one feature of an object spreads to other features of the same object. Subjects viewed two spatially superimposed surfaces of random dots that were segregated by distinct color-motion conjunctions. The color and direction of motion of each surface changed smoothly and in a cyclical fashion. Subjects were required to track one feature (e.g., color) of one of the two surfaces and detect brief moments when the attended feature diverged from its smooth trajectory. To tease apart the effect of attention to individual features on the hemodynamic response, we used a frequency-tagging scheme. In this scheme, the stimulus features (color and direction of motion) are modulated periodically at distinct frequencies so that the contribution of each feature to the hemodynamics can be inferred from the harmonic response at the corresponding frequency. We found that attention to one feature (e.g., color) of one surface increased the response modulation not only to the attended feature but also to the other feature (e.g., motion) of the same surface. This attentional modulation was evident in multiple visual areas and was present as early as V1. The spread of attention to the behaviorally irrelevant features of a surface suggests that attention may automatically select all features of a single object. Thus object-based attention may be supported by an enhancement of feature-specific sensory signals in the visual cortex. PMID:23883860
Office hysteroscopy and adenomyosis.
Molinas, Carlos Roger; Campo, Rudi
2006-08-01
Adenomyosis, the heterotopic presence of endometrial glands and stroma within the myometrium, has traditionally been diagnosed by the pathologist in hysterectomy specimens. However, the recent development of high-quality non-invasive techniques such as transvaginal sonography (TVS), magnetic resonance imaging (MRI) and hysteroscopy has renewed interest in diagnosing adenomyosis in the office prior to any treatment. Hysteroscopy offers the advantage of direct visualization of the uterine cavity, and since nowadays it is performed in the office, it can be offered as a first-line diagnostic tool for evaluation of uterine abnormalities in patients with abnormal uterine bleeding and/or infertility. The available data clearly indicate that high-quality mini-hysteroscopes, saline as a distension medium, and atraumatic insertion technique are essential for the success of office hysteroscopy. The procedure is indicated in any situation in which an intrauterine anomaly is suspected; it is immediately preceded by a physical exam and a TVS to evaluate uterine characteristics, and it is followed by a second TVS to take advantage of the intracavitary fluid for a contrast image of the uterus. Although diagnostic hysteroscopy does not provide pathognomonic signs for adenomyosis, some evidence suggests that irregular endometrium with endometrial defects, altered vascularization, and cystic haemorrhagic lesion are possibly associated with the entity. In addition to the direct visualization of the uterine cavity, the hysteroscopic approach offers the possibility of obtaining endometrial/myometrial biopsies under visual control. Since they can be performed in the office, the combination of TVS, fluid hysteroscopy and contrast sonography is therefore a powerful screening tool for detecting endometrial and myometrial abnormalities in association with adenomyosis.
Visualization of phage DNA degradation by a type I CRISPR-Cas system at the single-cell level.
Guan, Jingwen; Shi, Xu; Burgos, Roberto; Zeng, Lanying
2017-03-01
The CRISPR-Cas system is a widespread prokaryotic defense system which targets and cleaves invasive nucleic acids, such as plasmids or viruses. So far, a great number of studies have focused on the components and mechanisms of this system, however, a direct visualization of CRISPR-Cas degrading invading DNA in real-time has not yet been studied at the single-cell level. In this study, we fluorescently label phage lambda DNA in vivo , and track the labeled DNA over time to characterize DNA degradation at the single-cell level. At the bulk level, the lysogenization frequency of cells harboring CRISPR plasmids decreases significantly compared to cells with a non-CRISPR control. At the single-cell level, host cells with CRISPR activity are unperturbed by phage infection, maintaining normal growth like uninfected cells, where the efficiency of our anti-lambda CRISPR system is around 26%. During the course of time-lapse movies, the average fluorescence of invasive phage DNA in cells with CRISPR activity, decays more rapidly compared to cells without, and phage DNA is fully degraded by around 44 minutes on average. Moreover, the degradation appears to be independent of cell size or the phage DNA ejection site suggesting that Cas proteins are dispersed in sufficient quantities throughout the cell. With the CRISPR-Cas visualization system we developed, we are able to examine and characterize how a CRISPR system degrades invading phage DNA at the single-cell level. This work provides direct evidence and improves the current understanding on how CRISPR breaks down invading DNA.
Scheidt, Robert A.; Lillis, Kyle P.; Emerson, Scott J.
2010-01-01
We examined how proprioceptive contributions to perception of hand path straightness are influenced by visual, motor and attentional sources of performance variability during horizontal planar reaching. Subjects held the handle of a robot that constrained goal-directed movements of the hand to paths of controlled curvature. Subjects attempted to detect the presence of hand path curvature during both active (subject-driven) and passive (robot-driven) movements that either required active muscle force production or not. Subjects were less able to discriminate curved from straight paths when actively reaching for a target vs. when the robot moved their hand through the same curved paths. This effect was especially evident during robot-driven movements requiring concurrent activation of lengthening but not shortening muscles. Subjects were less likely to report curvature and were more variable in reporting when movements appeared straight in a novel “visual channel” condition previously shown to block adaptive updating of motor commands in response to deviations from a straight-line hand path. Similarly compromised performance was obtained when subjects simultaneously performed a distracting secondary task (key pressing with the contralateral hand). The effects compounded when these last two treatments were combined. It is concluded that environmental, intrinsic and attentional factors all impact the ability to detect deviations from a rectilinear hand path during goal-directed movement by decreasing proprioceptive contributions to limb state estimation. In contrast, response variability increased only in experimental conditions thought to impose additional attentional demands on the observer. Implications of these results for perception and other sensorimotor behaviors are discussed. PMID:20532489
Visual sensory networks and effective information transfer in animal groups.
Strandburg-Peshkin, Ariana; Twomey, Colin R; Bode, Nikolai W F; Kao, Albert B; Katz, Yael; Ioannou, Christos C; Rosenthal, Sara B; Torney, Colin J; Wu, Hai Shan; Levin, Simon A; Couzin, Iain D
2013-09-09
Social transmission of information is vital for many group-living animals, allowing coordination of motion and effective response to complex environments. Revealing the interaction networks underlying information flow within these groups is a central challenge. Previous work has modeled interactions between individuals based directly on their relative spatial positions: each individual is considered to interact with all neighbors within a fixed distance (metric range), a fixed number of nearest neighbors (topological range), a 'shell' of near neighbors (Voronoi range), or some combination (Figure 1A). However, conclusive evidence to support these assumptions is lacking. Here, we employ a novel approach that considers individual movement decisions to be based explicitly on the sensory information available to the organism. In other words, we consider that while spatial relations do inform interactions between individuals, they do so indirectly, through individuals' detection of sensory cues. We reconstruct computationally the visual field of each individual throughout experiments designed to investigate information propagation within fish schools (golden shiners, Notemigonus crysoleucas). Explicitly considering visual sensing allows us to more accurately predict the propagation of behavioral change in these groups during leadership events. Furthermore, we find that structural properties of visual interaction networks differ markedly from those of metric and topological counterparts, suggesting that previous assumptions may not appropriately reflect information flow in animal groups. Copyright © 2013 Elsevier Ltd. All rights reserved.
Summation of visual motion across eye movements reflects a nonspatial decision mechanism.
Morris, Adam P; Liu, Charles C; Cropper, Simon J; Forte, Jason D; Krekelberg, Bart; Mattingley, Jason B
2010-07-21
Human vision remains perceptually stable even though retinal inputs change rapidly with each eye movement. Although the neural basis of visual stability remains unknown, a recent psychophysical study pointed to the existence of visual feature-representations anchored in environmental rather than retinal coordinates (e.g., "spatiotopic" receptive fields; Melcher and Morrone, 2003). In that study, sensitivity to a moving stimulus presented after a saccadic eye movement was enhanced when preceded by another moving stimulus at the same spatial location before the saccade. The finding is consistent with spatiotopic sensory integration, but it could also have arisen from a probabilistic improvement in performance due to the presence of more than one motion signal for the perceptual decision. Here we show that this statistical advantage accounts completely for summation effects in this task. We first demonstrate that measurements of summation are confounded by noise related to an observer's uncertainty about motion onset times. When this uncertainty is minimized, comparable summation is observed regardless of whether two motion signals occupy the same or different locations in space, and whether they contain the same or opposite directions of motion. These results are incompatible with the tuning properties of motion-sensitive sensory neurons and provide no evidence for a spatiotopic representation of visual motion. Instead, summation in this context reflects a decision mechanism that uses abstract representations of sensory events to optimize choice behavior.
Job, Xavier E; de Fockert, Jan W; van Velzen, José
2016-08-01
Behavioural and electrophysiological evidence has demonstrated that preparation of goal-directed actions modulates sensory perception at the goal location before the action is executed. However, previous studies have focused on sensory perception in areas of peripersonal space. The present study investigated visual and tactile sensory processing at the goal location of upcoming movements towards the body, much of which is not visible, as well as visible peripersonal space. A motor task cued participants to prepare a reaching movement towards goals either in peripersonal space in front of them or personal space on the upper chest. In order to assess modulations of sensory perception during movement preparation, event-related potentials (ERPs) were recorded in response to task-irrelevant visual and tactile probe stimuli delivered randomly at one of the goal locations of the movements. In line with previous neurophysiological findings, movement preparation modulated visual processing at the goal of a movement in peripersonal space. Movement preparation also modulated somatosensory processing at the movement goal in personal space. The findings demonstrate that tactile perception in personal space is subject to similar top-down sensory modulation by motor preparation as observed for visual stimuli presented in peripersonal space. These findings show for the first time that the principles and mechanisms underlying adaptive modulation of sensory processing in the context of action extend to tactile perception in unseen personal space. Copyright © 2016 Elsevier Ltd. All rights reserved.
Social Vision: Functional Forecasting and the Integration of Compound Social Cues
Adams, Reginald B.; Kveraga, Kestutis
2017-01-01
For decades the study of social perception was largely compartmentalized by type of social cue: race, gender, emotion, eye gaze, body language, facial expression etc. This was partly due to good scientific practice (e.g., controlling for extraneous variability), and partly due to assumptions that each type of social cue was functionally distinct from others. Herein, we present a functional forecast approach to understanding compound social cue processing that emphasizes the importance of shared social affordances across various cues (see too Adams, Franklin, Nelson, & Stevenson, 2010; Adams & Nelson, 2011; Weisbuch & Adams, 2012). We review the traditional theories of emotion and face processing that argued for dissociable and noninteracting pathways (e.g., for specific emotional expressions, gaze, identity cues), as well as more recent evidence for combinatorial processing of social cues. We argue here that early, and presumably reflexive, visual integration of such cues is necessary for adaptive behavioral responding to others. In support of this claim, we review contemporary work that reveals a flexible visual system, one that readily incorporates meaningful contextual influences in even nonsocial visual processing, thereby establishing the functional and neuroanatomical bases necessary for compound social cue integration. Finally, we explicate three likely mechanisms driving such integration. Together, this work implicates a role for cognitive penetrability in visual perceptual abilities that have often been (and in some cases still are) ascribed to direct encapsulated perceptual processes. PMID:29242738
Al-Abood, Saleh A; Bennett, Simon J; Hernandez, Francisco Moreno; Ashford, Derek; Davids, Keith
2002-03-01
We assessed the effects on basketball free throw performance of two types of verbal directions with an external attentional focus. Novices (n = 16) were pre-tested on free throw performance and assigned to two groups of similar ability (n = 8 in each). Both groups received verbal instructions with an external focus on either movement dynamics (movement form) or movement effects (e.g. ball trajectory relative to basket). The participants also observed a skilled model performing the task on either a small or large screen monitor, to ascertain the effects of visual presentation mode on task performance. After observation of six videotaped trials, all participants were given a post-test. Visual search patterns were monitored during observation and cross-referenced with performance on the pre- and post-test. Group effects were noted for verbal instructions and image size on visual search strategies and free throw performance. The 'movement effects' group saw a significant improvement in outcome scores between the pre-test and post-test. These results supported evidence that this group spent more viewing time on information outside the body than the 'movement dynamics' group. Image size affected both groups equally with more fixations of shorter duration when viewing the small screen. The results support the benefits of instructions when observing a model with an external focus on movement effects, not dynamics.
Carlisle, Nancy B.; Woodman, Geoffrey F.
2014-01-01
Biased competition theory proposes that representations in working memory drive visual attention to select similar inputs. However, behavioral tests of this hypothesis have led to mixed results. These inconsistent findings could be due to the inability of behavioral measures to reliably detect the early, automatic effects on attentional deployment that the memory representations exert. Alternatively, executive mechanisms may govern how working memory representations influence attention based on higher-level goals. In the present study, we tested these hypotheses using the N2pc component of participants’ event-related potentials (ERPs) to directly measure the early deployments of covert attention. Participants searched for a target in an array that sometimes contained a memory-matching distractor. In Experiments 1–3, we manipulated the difficulty of the target discrimination and the proximity of distractors, but consistently observed that covert attention was deployed to the search targets and not the memory-matching distractors. In Experiment 4, we showed that when participants’ goal involved attending to memory-matching items that these items elicited a large and early N2pc. Our findings demonstrate that working memory representations alone are not sufficient to guide early deployments of visual attention to matching inputs and that goal-dependent executive control mediates the interactions between working memory representations and visual attention. PMID:21254796
Did Caravaggio employ optical projections? An image analysis of the parity in the artist's paintings
NASA Astrophysics Data System (ADS)
Stork, David G.
2011-03-01
We examine one class of evidence put forth in support of the recent claim that the Italian Baroque master Caravaggio secretly employed optical projectors as a direct drawing aid. Specically, we test the claims that there is an "abnormal number" of left-handed gures in his works and, more specically, that "During the Del Monte period he had too many left-handed models." We also test whether there was a reversal in the handedness of specic models in different paintings. Such evidence would be consistent with the claim that Caravaggio switched between using a convex-lens projector to using a concave-mirror projector and would support, but not prove, the claim that Caravaggio used optical projections. We estimate the parity (+ or -) of each of Caravaggio's 76 appropriate oil paintings based on the handedness of gures, the orientation of asymmetric objects, placement of scabbards, depicted text, and so on, and search for statistically significant changes in handedness in figures. We also track the direction of the illumination over time in the artist's uvre. We discuss some historical evidence as it relates to the question of his possible use of optics. We nd the proportion of left-handed figures lower than that in the general population (not higher), and no significant change in estimated handedness even of individual models. Optical proponents have argued that Bacchus (1597) portrays a left-handed gure, but we give visual and cultural evidence showing that this gure is instead right-handed, thereby rebutting this claim that the painting was executed using optical projections. Moreover, scholars recently re-discovered the image of the artist with easel and canvas reflected in the carafe of wine at the front left in the tableau in Bacchus, showing that this painting was almost surely executed using traditional (non-optical) easel methods. We conclude that there is 1) no statistically signicant abnormally high number of left-handed gures in Caravaggio's uvre, including during any limited working period, 2) no statistically significant change in handedness among all gures or even individual gures that might be consistent with a change in optical projector, and 3) the visual and cultural evidence in Bacchus shows the gure was right-handed and that the artist executed this work by traditional (non-optical) easel methods. We conclude that the general parity and handedness evidence does not support the claim that Caravaggio employed optical projections.
The effect of saccade metrics on the corollary discharge contribution to perceived eye location
Bansal, Sonia; Jayet Bray, Laurence C.; Peterson, Matthew S.
2015-01-01
Corollary discharge (CD) is hypothesized to provide the movement information (direction and amplitude) required to compensate for the saccade-induced disruptions to visual input. Here, we investigated to what extent these conveyed metrics influence perceptual stability in human subjects with a target-displacement detection task. Subjects made saccades to targets located at different amplitudes (4°, 6°, or 8°) and directions (horizontal or vertical). During the saccade, the target disappeared and then reappeared at a shifted location either in the same direction or opposite to the movement vector. Subjects reported the target displacement direction, and from these reports we determined the perceptual threshold for shift detection and estimate of target location. Our results indicate that the thresholds for all amplitudes and directions generally scaled with saccade amplitude. Additionally, subjects on average produced hypometric saccades with an estimated CD gain <1. Finally, we examined the contribution of different error signals to perceptual performance, the saccade error (movement-to-movement variability in saccade amplitude) and visual error (distance between the fovea and the shifted target location). Perceptual judgment was not influenced by the fluctuations in movement amplitude, and performance was largely the same across movement directions for different magnitudes of visual error. Importantly, subjects reported the correct direction of target displacement above chance level for very small visual errors (<0.75°), even when these errors were opposite the target-shift direction. Collectively, these results suggest that the CD-based compensatory mechanisms for visual disruptions are highly accurate and comparable for saccades with different metrics. PMID:25761955
How Ants Use Vision When Homing Backward.
Schwarz, Sebastian; Mangan, Michael; Zeil, Jochen; Webb, Barbara; Wystrach, Antoine
2017-02-06
Ants can navigate over long distances between their nest and food sites using visual cues [1, 2]. Recent studies show that this capacity is undiminished when walking backward while dragging a heavy food item [3-5]. This challenges the idea that ants use egocentric visual memories of the scene for guidance [1, 2, 6]. Can ants use their visual memories of the terrestrial cues when going backward? Our results suggest that ants do not adjust their direction of travel based on the perceived scene while going backward. Instead, they maintain a straight direction using their celestial compass. This direction can be dictated by their path integrator [5] but can also be set using terrestrial visual cues after a forward peek. If the food item is too heavy to enable body rotations, ants moving backward drop their food on occasion, rotate and walk a few steps forward, return to the food, and drag it backward in a now-corrected direction defined by terrestrial cues. Furthermore, we show that ants can maintain their direction of travel independently of their body orientation. It thus appears that egocentric retinal alignment is required for visual scene recognition, but ants can translate this acquired directional information into a holonomic frame of reference, which enables them to decouple their travel direction from their body orientation and hence navigate backward. This reveals substantial flexibility and communication between different types of navigational information: from terrestrial to celestial cues and from egocentric to holonomic directional memories. VIDEO ABSTRACT. Copyright © 2017 The Author(s). Published by Elsevier Ltd.. All rights reserved.
Scocchia, Lisa; Bolognini, Nadia; Convento, Silvia; Stucchi, Natale
2015-11-16
Human movements conform to specific kinematic laws of motion. One of such laws, the "two-thirds power law", describes the systematic co-variation between curvature and velocity of body movements. Noticeably, the same law also influences the perception of moving stimuli: the velocity of a dot moving along a curvilinear trajectory is perceived as uniform when the dot kinematics complies with the two-thirds power law. Instead, if the dot moves at constant speed, its velocity is perceived as highly non-uniform. This dynamic visual illusion points to a strong coupling between action and perception; however, how this coupling is implemented in the brain remains elusive. In this study, we tested whether the premotor cortex (PM) and the primary visual cortex (V1) play a role in the illusion by means of transcranial Direct Current Stimulation (tDCS). All participants underwent three tDCS sessions during which they received active or sham cathodal tDCS (1.5mA) over PM or V1 of the left hemisphere. During tDCS, participants were required to adjust the velocity of a dot moving along an elliptical trajectory until it looked uniform across the whole trajectory. Results show that occipital tDCS decreases the illusion variability both within and across participants, as compared to sham tDCS. This means that V1 stimulation increases individual sensitivity to the illusory motion and also increases coherence across different observers. Conversely, the illusion seems resistant to tDCS in terms of its magnitude, with cathodal stimulation of V1 or PM not affecting the amount of the illusory effect. Our results provide evidence for strong visuo-motor coupling in visual perception: the velocity of a dot moving along an elliptical trajectory is perceived as uniform only when its kinematics closely complies to the same law of motion that constrains human movement production. Occipital stimulation by cathodal tDCS can stabilize such illusory percept. Copyright © 2015 Elsevier Ireland Ltd. All rights reserved.
Visual Imagery without Visual Perception?
ERIC Educational Resources Information Center
Bertolo, Helder
2005-01-01
The question regarding visual imagery and visual perception remain an open issue. Many studies have tried to understand if the two processes share the same mechanisms or if they are independent, using different neural substrates. Most research has been directed towards the need of activation of primary visual areas during imagery. Here we review…
Binding of motion and colour is early and automatic.
Blaser, Erik; Papathomas, Thomas; Vidnyánszky, Zoltán
2005-04-01
At what stages of the human visual hierarchy different features are bound together, and whether this binding requires attention, is still highly debated. We used a colour-contingent motion after-effect (CCMAE) to study the binding of colour and motion signals. The logic of our approach was as follows: if CCMAEs can be evoked by targeted adaptation of early motion processing stages, without allowing for feedback from higher motion integration stages, then this would support our hypothesis that colour and motion are bound automatically on the basis of spatiotemporally local information. Our results show for the first time that CCMAE's can be evoked by adaptation to a locally paired opposite-motion dot display, a stimulus that, importantly, is known to trigger direction-specific responses in the primary visual cortex yet results in strong inhibition of the directional responses in area MT of macaques as well as in area MT+ in humans and, indeed, is perceived only as motionless flicker. The magnitude of the CCMAE in the locally paired condition was not significantly different from control conditions where the different directions were spatiotemporally separated (i.e. not locally paired) and therefore perceived as two moving fields. These findings provide evidence that adaptation at an early, local motion stage, and only adaptation at this stage, underlies this CCMAE, which in turn implies that spatiotemporally coincident colour and motion signals are bound automatically, most probably as early as cortical area V1, even when the association between colour and motion is perceptually inaccessible.
Template optimization and transfer in perceptual learning.
Kurki, Ilmari; Hyvärinen, Aapo; Saarinen, Jussi
2016-08-01
We studied how learning changes the processing of a low-level Gabor stimulus, using a classification-image method (psychophysical reverse correlation) and a task where observers discriminated between slight differences in the phase (relative alignment) of a target Gabor in visual noise. The method estimates the internal "template" that describes how the visual system weights the input information for decisions. One popular idea has been that learning makes the template more like an ideal Bayesian weighting; however, the evidence has been indirect. We used a new regression technique to directly estimate the template weight change and to test whether the direction of reweighting is significantly different from an optimal learning strategy. The subjects trained the task for six daily sessions, and we tested the transfer of training to a target in an orthogonal orientation. Strong learning and partial transfer were observed. We tested whether task precision (difficulty) had an effect on template change and transfer: Observers trained in either a high-precision (small, 60° phase difference) or a low-precision task (180°). Task precision did not have an effect on the amount of template change or transfer, suggesting that task precision per se does not determine whether learning generalizes. Classification images show that training made observers use more task-relevant features and unlearn some irrelevant features. The transfer templates resembled partially optimized versions of templates in training sessions. The template change direction resembles ideal learning significantly but not completely. The amount of template change was highly correlated with the amount of learning.
Brauchli, Christian; Elmer, Stefan; Rogenmoser, Lars; Burkhard, Anja; Jäncke, Lutz
2018-01-01
Auditory-visual (AV) synesthesia is a rare phenomenon in which an auditory stimulus induces a "concurrent" color sensation. Current neurophysiological models of synesthesia mainly hypothesize "hyperconnected" and "hyperactivated" brains, but differ in the directionality of signal transmission. The two-stage model proposes bottom-up signal transmission from inducer- to concurrent- to higher-order brain areas, whereas the disinhibited feedback model postulates top-down signal transmission from inducer- to higher-order- to concurrent brain areas. To test the different models of synesthesia, we estimated local current density, directed and undirected connectivity patterns in the intracranial space during 2 min of resting-state (RS) EEG in 11 AV synesthetes and 11 nonsynesthetes. AV synesthetes demonstrated increased parietal theta, alpha, and lower beta current density compared to nonsynesthetes. Furthermore, AV synesthetes were characterized by increased top-down signal transmission from the superior parietal lobe to the left color processing area V4 in the upper beta frequency band. Analyses of undirected connectivity revealed a global, synesthesia-specific hyperconnectivity in the alpha frequency band. The involvement of the superior parietal lobe even during rest is a strong indicator for its key role in AV synesthesia. By demonstrating top-down signal transmission in AV synesthetes, we provide direct support for the disinhibited feedback model of synesthesia. Finally, we suggest that synesthesia is a consequence of global hyperconnectivity. Hum Brain Mapp 39:522-531, 2018. © 2017 Wiley Periodicals, Inc. © 2017 Wiley Periodicals, Inc.