Sample records for object recognition time

  1. Real-time optical multiple object recognition and tracking system and method

    NASA Technical Reports Server (NTRS)

    Chao, Tien-Hsin (Inventor); Liu, Hua Kuang (Inventor)

    1987-01-01

    The invention relates to an apparatus and associated methods for the optical recognition and tracking of multiple objects in real time. Multiple point spatial filters are employed that pre-define the objects to be recognized at run-time. The system takes the basic technology of a Vander Lugt filter and adds a hololens. The technique replaces time, space and cost-intensive digital techniques. In place of multiple objects, the system can also recognize multiple orientations of a single object. This later capability has potential for space applications where space and weight are at a premium.

  2. Auditory-visual object recognition time suggests specific processing for animal sounds.

    PubMed

    Suied, Clara; Viaud-Delmon, Isabelle

    2009-01-01

    Recognizing an object requires binding together several cues, which may be distributed across different sensory modalities, and ignoring competing information originating from other objects. In addition, knowledge of the semantic category of an object is fundamental to determine how we should react to it. Here we investigate the role of semantic categories in the processing of auditory-visual objects. We used an auditory-visual object-recognition task (go/no-go paradigm). We compared recognition times for two categories: a biologically relevant one (animals) and a non-biologically relevant one (means of transport). Participants were asked to react as fast as possible to target objects, presented in the visual and/or the auditory modality, and to withhold their response for distractor objects. A first main finding was that, when participants were presented with unimodal or bimodal congruent stimuli (an image and a sound from the same object), similar reaction times were observed for all object categories. Thus, there was no advantage in the speed of recognition for biologically relevant compared to non-biologically relevant objects. A second finding was that, in the presence of a biologically relevant auditory distractor, the processing of a target object was slowed down, whether or not it was itself biologically relevant. It seems impossible to effectively ignore an animal sound, even when it is irrelevant to the task. These results suggest a specific and mandatory processing of animal sounds, possibly due to phylogenetic memory and consistent with the idea that hearing is particularly efficient as an alerting sense. They also highlight the importance of taking into account the auditory modality when investigating the way object concepts of biologically relevant categories are stored and retrieved.

  3. Short temporal asynchrony disrupts visual object recognition

    PubMed Central

    Singer, Jedediah M.; Kreiman, Gabriel

    2014-01-01

    Humans can recognize objects and scenes in a small fraction of a second. The cascade of signals underlying rapid recognition might be disrupted by temporally jittering different parts of complex objects. Here we investigated the time course over which shape information can be integrated to allow for recognition of complex objects. We presented fragments of object images in an asynchronous fashion and behaviorally evaluated categorization performance. We observed that visual recognition was significantly disrupted by asynchronies of approximately 30 ms, suggesting that spatiotemporal integration begins to break down with even small deviations from simultaneity. However, moderate temporal asynchrony did not completely obliterate recognition; in fact, integration of visual shape information persisted even with an asynchrony of 100 ms. We describe the data with a concise model based on the dynamic reduction of uncertainty about what image was presented. These results emphasize the importance of timing in visual processing and provide strong constraints for the development of dynamical models of visual shape recognition. PMID:24819738

  4. Decoding the time-course of object recognition in the human brain: From visual features to categorical decisions.

    PubMed

    Contini, Erika W; Wardle, Susan G; Carlson, Thomas A

    2017-10-01

    Visual object recognition is a complex, dynamic process. Multivariate pattern analysis methods, such as decoding, have begun to reveal how the brain processes complex visual information. Recently, temporal decoding methods for EEG and MEG have offered the potential to evaluate the temporal dynamics of object recognition. Here we review the contribution of M/EEG time-series decoding methods to understanding visual object recognition in the human brain. Consistent with the current understanding of the visual processing hierarchy, low-level visual features dominate decodable object representations early in the time-course, with more abstract representations related to object category emerging later. A key finding is that the time-course of object processing is highly dynamic and rapidly evolving, with limited temporal generalisation of decodable information. Several studies have examined the emergence of object category structure, and we consider to what degree category decoding can be explained by sensitivity to low-level visual features. Finally, we evaluate recent work attempting to link human behaviour to the neural time-course of object processing. Copyright © 2017 Elsevier Ltd. All rights reserved.

  5. Interactive object recognition assistance: an approach to recognition starting from target objects

    NASA Astrophysics Data System (ADS)

    Geisler, Juergen; Littfass, Michael

    1999-07-01

    Recognition of target objects in remotely sensed imagery required detailed knowledge about the target object domain as well as about mapping properties of the sensing system. The art of object recognition is to combine both worlds appropriately and to provide models of target appearance with respect to sensor characteristics. Common approaches to support interactive object recognition are either driven from the sensor point of view and address the problem of displaying images in a manner adequate to the sensing system. Or they focus on target objects and provide exhaustive encyclopedic information about this domain. Our paper discusses an approach to assist interactive object recognition based on knowledge about target objects and taking into account the significance of object features with respect to characteristics of the sensed imagery, e.g. spatial and spectral resolution. An `interactive recognition assistant' takes the image analyst through the interpretation process by indicating step-by-step the respectively most significant features of objects in an actual set of candidates. The significance of object features is expressed by pregenerated trees of significance, and by the dynamic computation of decision relevance for every feature at each step of the recognition process. In the context of this approach we discuss the question of modeling and storing the multisensorial/multispectral appearances of target objects and object classes as well as the problem of an adequate dynamic human-machine-interface that takes into account various mental models of human image interpretation.

  6. Shape and Color Features for Object Recognition Search

    NASA Technical Reports Server (NTRS)

    Duong, Tuan A.; Duong, Vu A.; Stubberud, Allen R.

    2012-01-01

    A bio-inspired shape feature of an object of interest emulates the integration of the saccadic eye movement and horizontal layer in vertebrate retina for object recognition search where a single object can be used one at a time. The optimal computational model for shape-extraction-based principal component analysis (PCA) was also developed to reduce processing time and enable the real-time adaptive system capability. A color feature of the object is employed as color segmentation to empower the shape feature recognition to solve the object recognition in the heterogeneous environment where a single technique - shape or color - may expose its difficulties. To enable the effective system, an adaptive architecture and autonomous mechanism were developed to recognize and adapt the shape and color feature of the moving object. The bio-inspired object recognition based on bio-inspired shape and color can be effective to recognize a person of interest in the heterogeneous environment where the single technique exposed its difficulties to perform effective recognition. Moreover, this work also demonstrates the mechanism and architecture of the autonomous adaptive system to enable the realistic system for the practical use in the future.

  7. Eye movements during object recognition in visual agnosia.

    PubMed

    Charles Leek, E; Patterson, Candy; Paul, Matthew A; Rafal, Robert; Cristino, Filipe

    2012-07-01

    This paper reports the first ever detailed study about eye movement patterns during single object recognition in visual agnosia. Eye movements were recorded in a patient with an integrative agnosic deficit during two recognition tasks: common object naming and novel object recognition memory. The patient showed normal directional biases in saccades and fixation dwell times in both tasks and was as likely as controls to fixate within object bounding contour regardless of recognition accuracy. In contrast, following initial saccades of similar amplitude to controls, the patient showed a bias for short saccades. In object naming, but not in recognition memory, the similarity of the spatial distributions of patient and control fixations was modulated by recognition accuracy. The study provides new evidence about how eye movements can be used to elucidate the functional impairments underlying object recognition deficits. We argue that the results reflect a breakdown in normal functional processes involved in the integration of shape information across object structure during the visual perception of shape. Copyright © 2012 Elsevier Ltd. All rights reserved.

  8. Fast neuromimetic object recognition using FPGA outperforms GPU implementations.

    PubMed

    Orchard, Garrick; Martin, Jacob G; Vogelstein, R Jacob; Etienne-Cummings, Ralph

    2013-08-01

    Recognition of objects in still images has traditionally been regarded as a difficult computational problem. Although modern automated methods for visual object recognition have achieved steadily increasing recognition accuracy, even the most advanced computational vision approaches are unable to obtain performance equal to that of humans. This has led to the creation of many biologically inspired models of visual object recognition, among them the hierarchical model and X (HMAX) model. HMAX is traditionally known to achieve high accuracy in visual object recognition tasks at the expense of significant computational complexity. Increasing complexity, in turn, increases computation time, reducing the number of images that can be processed per unit time. In this paper we describe how the computationally intensive and biologically inspired HMAX model for visual object recognition can be modified for implementation on a commercial field-programmable aate Array, specifically the Xilinx Virtex 6 ML605 evaluation board with XC6VLX240T FPGA. We show that with minor modifications to the traditional HMAX model we can perform recognition on images of size 128 × 128 pixels at a rate of 190 images per second with a less than 1% loss in recognition accuracy in both binary and multiclass visual object recognition tasks.

  9. Examining object recognition and object-in-Place memory in plateau zokors, Eospalax baileyi.

    PubMed

    Hegab, Ibrahim M; Tan, Yuchen; Wang, Chan; Yao, Baohui; Wang, Haifang; Ji, Weihong; Su, Junhu

    2018-01-01

    Recognition memory is important for the survival and fitness of subterranean rodents due to the barren underground conditions that require avoiding the burden of higher energy costs or possible conflict with conspecifics. Our study aims to examine the object and object/place recognition memories in plateau zokors (Eospalax baileyi) and test whether their underground life exerts sex-specific differences in memory functions using Novel Object Recognition (NOR) and Object-in-Place (OiP) paradigms. Animals were tested in the NOR with short (10min) and long-term (24h) inter-trial intervals (ITI) and in the OiP for a 30-min ITI between the familiarization and testing sessions. Plateau zokors showed a strong preference for novel objects manifested by a longer exploration time for the novel object after 10min ITI but failed to remember the familiar object when tested after 24h, suggesting a lack of long-term memory. In the OiP test, zokors effectively formed an association between the objects and the place where they were formerly encountered, resulting in a higher duration of exploration to the switched objects. However, both sexes showed equivalent results in exploration time during the NOR and OiP tests, which eliminates the possibility of discovering sex-specific variations in memory performance. Taken together, our study illustrates robust novelty preference and an effective short-term recognition memory without marked sex-specific differences, which might elucidate the dynamics of recognition memory formation and retrieval in plateau zokors. Copyright © 2017 Elsevier B.V. All rights reserved.

  10. Cognitive object recognition system (CORS)

    NASA Astrophysics Data System (ADS)

    Raju, Chaitanya; Varadarajan, Karthik Mahesh; Krishnamurthi, Niyant; Xu, Shuli; Biederman, Irving; Kelley, Troy

    2010-04-01

    We have developed a framework, Cognitive Object Recognition System (CORS), inspired by current neurocomputational models and psychophysical research in which multiple recognition algorithms (shape based geometric primitives, 'geons,' and non-geometric feature-based algorithms) are integrated to provide a comprehensive solution to object recognition and landmarking. Objects are defined as a combination of geons, corresponding to their simple parts, and the relations among the parts. However, those objects that are not easily decomposable into geons, such as bushes and trees, are recognized by CORS using "feature-based" algorithms. The unique interaction between these algorithms is a novel approach that combines the effectiveness of both algorithms and takes us closer to a generalized approach to object recognition. CORS allows recognition of objects through a larger range of poses using geometric primitives and performs well under heavy occlusion - about 35% of object surface is sufficient. Furthermore, geon composition of an object allows image understanding and reasoning even with novel objects. With reliable landmarking capability, the system improves vision-based robot navigation in GPS-denied environments. Feasibility of the CORS system was demonstrated with real stereo images captured from a Pioneer robot. The system can currently identify doors, door handles, staircases, trashcans and other relevant landmarks in the indoor environment.

  11. A new selective developmental deficit: Impaired object recognition with normal face recognition.

    PubMed

    Germine, Laura; Cashdollar, Nathan; Düzel, Emrah; Duchaine, Bradley

    2011-05-01

    Studies of developmental deficits in face recognition, or developmental prosopagnosia, have shown that individuals who have not suffered brain damage can show face recognition impairments coupled with normal object recognition (Duchaine and Nakayama, 2005; Duchaine et al., 2006; Nunn et al., 2001). However, no developmental cases with the opposite dissociation - normal face recognition with impaired object recognition - have been reported. The existence of a case of non-face developmental visual agnosia would indicate that the development of normal face recognition mechanisms does not rely on the development of normal object recognition mechanisms. To see whether a developmental variant of non-face visual object agnosia exists, we conducted a series of web-based object and face recognition tests to screen for individuals showing object recognition memory impairments but not face recognition impairments. Through this screening process, we identified AW, an otherwise normal 19-year-old female, who was then tested in the lab on face and object recognition tests. AW's performance was impaired in within-class visual recognition memory across six different visual categories (guns, horses, scenes, tools, doors, and cars). In contrast, she scored normally on seven tests of face recognition, tests of memory for two other object categories (houses and glasses), and tests of recall memory for visual shapes. Testing confirmed that her impairment was not related to a general deficit in lower-level perception, object perception, basic-level recognition, or memory. AW's results provide the first neuropsychological evidence that recognition memory for non-face visual object categories can be selectively impaired in individuals without brain damage or other memory impairment. These results indicate that the development of recognition memory for faces does not depend on intact object recognition memory and provide further evidence for category-specific dissociations in visual

  12. Development of a sonar-based object recognition system

    NASA Astrophysics Data System (ADS)

    Ecemis, Mustafa Ihsan

    2001-02-01

    Sonars are used extensively in mobile robotics for obstacle detection, ranging and avoidance. However, these range-finding applications do not exploit the full range of information carried in sonar echoes. In addition, mobile robots need robust object recognition systems. Therefore, a simple and robust object recognition system using ultrasonic sensors may have a wide range of applications in robotics. This dissertation develops and analyzes an object recognition system that uses ultrasonic sensors of the type commonly found on mobile robots. Three principal experiments are used to test the sonar recognition system: object recognition at various distances, object recognition during unconstrained motion, and softness discrimination. The hardware setup, consisting of an inexpensive Polaroid sonar and a data acquisition board, is described first. The software for ultrasound signal generation, echo detection, data collection, and data processing is then presented. Next, the dissertation describes two methods to extract information from the echoes, one in the frequency domain and the other in the time domain. The system uses the fuzzy ARTMAP neural network to recognize objects on the basis of the information content of their echoes. In order to demonstrate that the performance of the system does not depend on the specific classification method being used, the K- Nearest Neighbors (KNN) Algorithm is also implemented. KNN yields a test accuracy similar to fuzzy ARTMAP in all experiments. Finally, the dissertation describes a method for extracting features from the envelope function in order to reduce the dimension of the input vector used by the classifiers. Decreasing the size of the input vectors reduces the memory requirements of the system and makes it run faster. It is shown that this method does not affect the performance of the system dramatically and is more appropriate for some tasks. The results of these experiments demonstrate that sonar can be used to develop

  13. Infant visual attention and object recognition.

    PubMed

    Reynolds, Greg D

    2015-05-15

    This paper explores the role visual attention plays in the recognition of objects in infancy. Research and theory on the development of infant attention and recognition memory are reviewed in three major sections. The first section reviews some of the major findings and theory emerging from a rich tradition of behavioral research utilizing preferential looking tasks to examine visual attention and recognition memory in infancy. The second section examines research utilizing neural measures of attention and object recognition in infancy as well as research on brain-behavior relations in the early development of attention and recognition memory. The third section addresses potential areas of the brain involved in infant object recognition and visual attention. An integrated synthesis of some of the existing models of the development of visual attention is presented which may account for the observed changes in behavioral and neural measures of visual attention and object recognition that occur across infancy. Copyright © 2015 Elsevier B.V. All rights reserved.

  14. Infant Visual Attention and Object Recognition

    PubMed Central

    Reynolds, Greg D.

    2015-01-01

    This paper explores the role visual attention plays in the recognition of objects in infancy. Research and theory on the development of infant attention and recognition memory are reviewed in three major sections. The first section reviews some of the major findings and theory emerging from a rich tradition of behavioral research utilizing preferential looking tasks to examine visual attention and recognition memory in infancy. The second section examines research utilizing neural measures of attention and object recognition in infancy as well as research on brain-behavior relations in the early development of attention and recognition memory. The third section addresses potential areas of the brain involved in infant object recognition and visual attention. An integrated synthesis of some of the existing models of the development of visual attention is presented which may account for the observed changes in behavioral and neural measures of visual attention and object recognition that occur across infancy. PMID:25596333

  15. Three-dimensional object recognition based on planar images

    NASA Astrophysics Data System (ADS)

    Mital, Dinesh P.; Teoh, Eam-Khwang; Au, K. C.; Chng, E. K.

    1993-01-01

    This paper presents the development and realization of a robotic vision system for the recognition of 3-dimensional (3-D) objects. The system can recognize a single object from among a group of known regular convex polyhedron objects that is constrained to lie on a calibrated flat platform. The approach adopted comprises a series of image processing operations on a single 2-dimensional (2-D) intensity image to derive an image line drawing. Subsequently, a feature matching technique is employed to determine 2-D spatial correspondences of the image line drawing with the model in the database. Besides its identification ability, the system can also provide important position and orientation information of the recognized object. The system was implemented on an IBM-PC AT machine executing at 8 MHz without the 80287 Maths Co-processor. In our overall performance evaluation based on a 600 recognition cycles test, the system demonstrated an accuracy of above 80% with recognition time well within 10 seconds. The recognition time is, however, indirectly dependent on the number of models in the database. The reliability of the system is also affected by illumination conditions which must be clinically controlled as in any industrial robotic vision system.

  16. Exogenous temporal cues enhance recognition memory in an object-based manner.

    PubMed

    Ohyama, Junji; Watanabe, Katsumi

    2010-11-01

    Exogenous attention enhances the perception of attended items in both a space-based and an object-based manner. Exogenous attention also improves recognition memory for attended items in the space-based mode. However, it has not been examined whether object-based exogenous attention enhances recognition memory. To address this issue, we examined whether a sudden visual change in a task-irrelevant stimulus (an exogenous cue) would affect participants' recognition memory for items that were serially presented around a cued time. The results showed that recognition accuracy for an item was strongly enhanced when the visual cue occurred at the same location and time as the item (Experiments 1 and 2). The memory enhancement effect occurred when the exogenous visual cue and an item belonged to the same object (Experiments 3 and 4) and even when the cue was counterpredictive of the timing of an item to be asked about (Experiment 5). The present study suggests that an exogenous temporal cue automatically enhances the recognition accuracy for an item that is presented at close temporal proximity to the cue and that recognition memory enhancement occurs in an object-based manner.

  17. Vision-based object detection and recognition system for intelligent vehicles

    NASA Astrophysics Data System (ADS)

    Ran, Bin; Liu, Henry X.; Martono, Wilfung

    1999-01-01

    Recently, a proactive crash mitigation system is proposed to enhance the crash avoidance and survivability of the Intelligent Vehicles. Accurate object detection and recognition system is a prerequisite for a proactive crash mitigation system, as system component deployment algorithms rely on accurate hazard detection, recognition, and tracking information. In this paper, we present a vision-based approach to detect and recognize vehicles and traffic signs, obtain their information, and track multiple objects by using a sequence of color images taken from a moving vehicle. The entire system consist of two sub-systems, the vehicle detection and recognition sub-system and traffic sign detection and recognition sub-system. Both of the sub- systems consist of four models: object detection model, object recognition model, object information model, and object tracking model. In order to detect potential objects on the road, several features of the objects are investigated, which include symmetrical shape and aspect ratio of a vehicle and color and shape information of the signs. A two-layer neural network is trained to recognize different types of vehicles and a parameterized traffic sign model is established in the process of recognizing a sign. Tracking is accomplished by combining the analysis of single image frame with the analysis of consecutive image frames. The analysis of the single image frame is performed every ten full-size images. The information model will obtain the information related to the object, such as time to collision for the object vehicle and relative distance from the traffic sings. Experimental results demonstrated a robust and accurate system in real time object detection and recognition over thousands of image frames.

  18. Object Recognition using Feature- and Color-Based Methods

    NASA Technical Reports Server (NTRS)

    Duong, Tuan; Duong, Vu; Stubberud, Allen

    2008-01-01

    An improved adaptive method of processing image data in an artificial neural network has been developed to enable automated, real-time recognition of possibly moving objects under changing (including suddenly changing) conditions of illumination and perspective. The method involves a combination of two prior object-recognition methods one based on adaptive detection of shape features and one based on adaptive color segmentation to enable recognition in situations in which either prior method by itself may be inadequate. The chosen prior feature-based method is known as adaptive principal-component analysis (APCA); the chosen prior color-based method is known as adaptive color segmentation (ACOSE). These methods are made to interact with each other in a closed-loop system to obtain an optimal solution of the object-recognition problem in a dynamic environment. One of the results of the interaction is to increase, beyond what would otherwise be possible, the accuracy of the determination of a region of interest (containing an object that one seeks to recognize) within an image. Another result is to provide a minimized adaptive step that can be used to update the results obtained by the two component methods when changes of color and apparent shape occur. The net effect is to enable the neural network to update its recognition output and improve its recognition capability via an adaptive learning sequence. In principle, the improved method could readily be implemented in integrated circuitry to make a compact, low-power, real-time object-recognition system. It has been proposed to demonstrate the feasibility of such a system by integrating a 256-by-256 active-pixel sensor with APCA, ACOSE, and neural processing circuitry on a single chip. It has been estimated that such a system on a chip would have a volume no larger than a few cubic centimeters, could operate at a rate as high as 1,000 frames per second, and would consume in the order of milliwatts of power.

  19. Object, spatial and social recognition testing in a single test paradigm.

    PubMed

    Lian, Bin; Gao, Jun; Sui, Nan; Feng, Tingyong; Li, Ming

    2018-07-01

    Animals have the ability to process information about an object or a conspecific's physical features and location, and alter its behavior when such information is updated. In the laboratory, the object, spatial and social recognition are often studied in separate tasks, making them unsuitable to study the potential dissociations and interactions among various types of recognition memories. The present study introduced a single paradigm to detect the object and spatial recognition, and social recognition of a familiar and novel conspecific. Specifically, male and female Sprague-Dawley adult (>75 days old) or preadolescent (25-28 days old) rats were tested with two objects and one social partner in an open-field arena for four 10-min sessions with a 20-min inter-session interval. After the first sample session, a new object replaced one of the sampled objects in the second session, and the location of one of the old objects was changed in the third session. Finally, a new social partner was introduced in the fourth session and replaced the familiar one. Exploration time with each stimulus was recorded and measures for the three recognitions were calculated based on the discrimination ratio. Overall results show that adult and preadolescent male and female rats spent more time exploring the social partner than the objects, showing a clear preference for social stimulus over nonsocial one. They also did not differ in their abilities to discriminate a new object, a new location and a new social partner from a familiar one, and to recognize a familiar conspecific. Acute administration of MK-801 (a NMDA receptor antagonist, 0.025 and 0.10 mg/kg, i.p.) after the sample session dose-dependently reduced the total time spent on exploring the social partner and objects in the adult rats, and had a significantly larger effect in the females than in the males. MK-801 also dose-dependently increased motor activity. However, it did not alter the object, spatial and social

  20. Orientation congruency effects for familiar objects: coordinate transformations in object recognition.

    PubMed

    Graf, M; Kaping, D; Bülthoff, H H

    2005-03-01

    How do observers recognize objects after spatial transformations? Recent neurocomputational models have proposed that object recognition is based on coordinate transformations that align memory and stimulus representations. If the recognition of a misoriented object is achieved by adjusting a coordinate system (or reference frame), then recognition should be facilitated when the object is preceded by a different object in the same orientation. In the two experiments reported here, two objects were presented in brief masked displays that were in close temporal contiguity; the objects were in either congruent or incongruent picture-plane orientations. Results showed that naming accuracy was higher for congruent than for incongruent orientations. The congruency effect was independent of superordinate category membership (Experiment 1) and was found for objects with different main axes of elongation (Experiment 2). The results indicate congruency effects for common familiar objects even when they have dissimilar shapes. These findings are compatible with models in which object recognition is achieved by an adjustment of a perceptual coordinate system.

  1. Method and System for Object Recognition Search

    NASA Technical Reports Server (NTRS)

    Duong, Tuan A. (Inventor); Duong, Vu A. (Inventor); Stubberud, Allen R. (Inventor)

    2012-01-01

    A method for object recognition using shape and color features of the object to be recognized. An adaptive architecture is used to recognize and adapt the shape and color features for moving objects to enable object recognition.

  2. Neural-Network Object-Recognition Program

    NASA Technical Reports Server (NTRS)

    Spirkovska, L.; Reid, M. B.

    1993-01-01

    HONTIOR computer program implements third-order neural network exhibiting invariance under translation, change of scale, and in-plane rotation. Invariance incorporated directly into architecture of network. Only one view of each object needed to train network for two-dimensional-translation-invariant recognition of object. Also used for three-dimensional-transformation-invariant recognition by training network on only set of out-of-plane rotated views. Written in C language.

  3. Sensor agnostic object recognition using a map seeking circuit

    NASA Astrophysics Data System (ADS)

    Overman, Timothy L.; Hart, Michael

    2012-05-01

    Automatic object recognition capabilities are traditionally tuned to exploit the specific sensing modality they were designed to. Their successes (and shortcomings) are tied to object segmentation from the background, they typically require highly skilled personnel to train them, and they become cumbersome with the introduction of new objects. In this paper we describe a sensor independent algorithm based on the biologically inspired technology of map seeking circuits (MSC) which overcomes many of these obstacles. In particular, the MSC concept offers transparency in object recognition from a common interface to all sensor types, analogous to a USB device. It also provides a common core framework that is independent of the sensor and expandable to support high dimensionality decision spaces. Ease in training is assured by using commercially available 3D models from the video game community. The search time remains linear no matter how many objects are introduced, ensuring rapid object recognition. Here, we report results of an MSC algorithm applied to object recognition and pose estimation from high range resolution radar (1D), electrooptical imagery (2D), and LIDAR point clouds (3D) separately. By abstracting the sensor phenomenology from the underlying a prior knowledge base, MSC shows promise as an easily adaptable tool for incorporating additional sensor inputs.

  4. The role of perceptual load in object recognition.

    PubMed

    Lavie, Nilli; Lin, Zhicheng; Zokaei, Nahid; Thoma, Volker

    2009-10-01

    Predictions from perceptual load theory (Lavie, 1995, 2005) regarding object recognition across the same or different viewpoints were tested. Results showed that high perceptual load reduces distracter recognition levels despite always presenting distracter objects from the same view. They also showed that the levels of distracter recognition were unaffected by a change in the distracter object view under conditions of low perceptual load. These results were found both with repetition priming measures of distracter recognition and with performance on a surprise recognition memory test. The results support load theory proposals that distracter recognition critically depends on the level of perceptual load. The implications for the role of attention in object recognition theories are discussed. PsycINFO Database Record (c) 2009 APA, all rights reserved.

  5. High speed optical object recognition processor with massive holographic memory

    NASA Technical Reports Server (NTRS)

    Chao, T.; Zhou, H.; Reyes, G.

    2002-01-01

    Real-time object recognition using a compact grayscale optical correlator will be introduced. A holographic memory module for storing a large bank of optimum correlation filters, to accommodate the large data throughput rate needed for many real-world applications, has also been developed. System architecture of the optical processor and the holographic memory will be presented. Application examples of this object recognition technology will also be demonstrated.

  6. Coordinate Transformations in Object Recognition

    ERIC Educational Resources Information Center

    Graf, Markus

    2006-01-01

    A basic problem of visual perception is how human beings recognize objects after spatial transformations. Three central classes of findings have to be accounted for: (a) Recognition performance varies systematically with orientation, size, and position; (b) recognition latencies are sequentially additive, suggesting analogue transformation…

  7. Real-time object recognition in multidimensional images based on joined extended structural tensor and higher-order tensor decomposition methods

    NASA Astrophysics Data System (ADS)

    Cyganek, Boguslaw; Smolka, Bogdan

    2015-02-01

    In this paper a system for real-time recognition of objects in multidimensional video signals is proposed. Object recognition is done by pattern projection into the tensor subspaces obtained from the factorization of the signal tensors representing the input signal. However, instead of taking only the intensity signal the novelty of this paper is first to build the Extended Structural Tensor representation from the intensity signal that conveys information on signal intensities, as well as on higher-order statistics of the input signals. This way the higher-order input pattern tensors are built from the training samples. Then, the tensor subspaces are built based on the Higher-Order Singular Value Decomposition of the prototype pattern tensors. Finally, recognition relies on measurements of the distance of a test pattern projected into the tensor subspaces obtained from the training tensors. Due to high-dimensionality of the input data, tensor based methods require high memory and computational resources. However, recent achievements in the technology of the multi-core microprocessors and graphic cards allows real-time operation of the multidimensional methods as is shown and analyzed in this paper based on real examples of object detection in digital images.

  8. Integration trumps selection in object recognition.

    PubMed

    Saarela, Toni P; Landy, Michael S

    2015-03-30

    Finding and recognizing objects is a fundamental task of vision. Objects can be defined by several "cues" (color, luminance, texture, etc.), and humans can integrate sensory cues to improve detection and recognition [1-3]. Cortical mechanisms fuse information from multiple cues [4], and shape-selective neural mechanisms can display cue invariance by responding to a given shape independent of the visual cue defining it [5-8]. Selective attention, in contrast, improves recognition by isolating a subset of the visual information [9]. Humans can select single features (red or vertical) within a perceptual dimension (color or orientation), giving faster and more accurate responses to items having the attended feature [10, 11]. Attention elevates neural responses and sharpens neural tuning to the attended feature, as shown by studies in psychophysics and modeling [11, 12], imaging [13-16], and single-cell and neural population recordings [17, 18]. Besides single features, attention can select whole objects [19-21]. Objects are among the suggested "units" of attention because attention to a single feature of an object causes the selection of all of its features [19-21]. Here, we pit integration against attentional selection in object recognition. We find, first, that humans can integrate information near optimally from several perceptual dimensions (color, texture, luminance) to improve recognition. They cannot, however, isolate a single dimension even when the other dimensions provide task-irrelevant, potentially conflicting information. For object recognition, it appears that there is mandatory integration of information from multiple dimensions of visual experience. The advantage afforded by this integration, however, comes at the expense of attentional selection. Copyright © 2015 Elsevier Ltd. All rights reserved.

  9. Integration trumps selection in object recognition

    PubMed Central

    Saarela, Toni P.; Landy, Michael S.

    2015-01-01

    Summary Finding and recognizing objects is a fundamental task of vision. Objects can be defined by several “cues” (color, luminance, texture etc.), and humans can integrate sensory cues to improve detection and recognition [1–3]. Cortical mechanisms fuse information from multiple cues [4], and shape-selective neural mechanisms can display cue-invariance by responding to a given shape independent of the visual cue defining it [5–8]. Selective attention, in contrast, improves recognition by isolating a subset of the visual information [9]. Humans can select single features (red or vertical) within a perceptual dimension (color or orientation), giving faster and more accurate responses to items having the attended feature [10,11]. Attention elevates neural responses and sharpens neural tuning to the attended feature, as shown by studies in psychophysics and modeling [11,12], imaging [13–16], and single-cell and neural population recordings [17,18]. Besides single features, attention can select whole objects [19–21]. Objects are among the suggested “units” of attention because attention to a single feature of an object causes the selection of all of its features [19–21]. Here, we pit integration against attentional selection in object recognition. We find, first, that humans can integrate information near-optimally from several perceptual dimensions (color, texture, luminance) to improve recognition. They cannot, however, isolate a single dimension even when the other dimensions provide task-irrelevant, potentially conflicting information. For object recognition, it appears that there is mandatory integration of information from multiple dimensions of visual experience. The advantage afforded by this integration, however, comes at the expense of attentional selection. PMID:25802154

  10. Multi-objects recognition for distributed intelligent sensor networks

    NASA Astrophysics Data System (ADS)

    He, Haibo; Chen, Sheng; Cao, Yuan; Desai, Sachi; Hohil, Myron E.

    2008-04-01

    This paper proposes an innovative approach for multi-objects recognition for homeland security and defense based intelligent sensor networks. Unlike the conventional way of information analysis, data mining in such networks is typically characterized with high information ambiguity/uncertainty, data redundancy, high dimensionality and real-time constrains. Furthermore, since a typical military based network normally includes multiple mobile sensor platforms, ground forces, fortified tanks, combat flights, and other resources, it is critical to develop intelligent data mining approaches to fuse different information resources to understand dynamic environments, to support decision making processes, and finally to achieve the goals. This paper aims to address these issues with a focus on multi-objects recognition. Instead of classifying a single object as in the traditional image classification problems, the proposed method can automatically learn multiple objectives simultaneously. Image segmentation techniques are used to identify the interesting regions in the field, which correspond to multiple objects such as soldiers or tanks. Since different objects will come with different feature sizes, we propose a feature scaling method to represent each object in the same number of dimensions. This is achieved by linear/nonlinear scaling and sampling techniques. Finally, support vector machine (SVM) based learning algorithms are developed to learn and build the associations for different objects, and such knowledge will be adaptively accumulated for objects recognition in the testing stage. We test the effectiveness of proposed method in different simulated military environments.

  11. Real-Time pedestrian detection : layered object recognition system for pedestrian collision sensing.

    DOT National Transportation Integrated Search

    2010-01-01

    In 2005 alone, 64,000 pedestrians were injured and 4,882 were killed in the United States, with pedestrians accounting for 11 percent of all traffic fatalities and 2 percent of injuries. The focus of "Layered Object Recognition System for Pedestrian ...

  12. Object and event recognition for stroke rehabilitation

    NASA Astrophysics Data System (ADS)

    Ghali, Ahmed; Cunningham, Andrew S.; Pridmore, Tony P.

    2003-06-01

    Stroke is a major cause of disability and health care expenditure around the world. Existing stroke rehabilitation methods can be effective but are costly and need to be improved. Even modest improvements in the effectiveness of rehabilitation techniques could produce large benefits in terms of quality of life. The work reported here is part of an ongoing effort to integrate virtual reality and machine vision technologies to produce innovative stroke rehabilitation methods. We describe a combined object recognition and event detection system that provides real time feedback to stroke patients performing everyday kitchen tasks necessary for independent living, e.g. making a cup of coffee. The image plane position of each object, including the patient"s hand, is monitored using histogram-based recognition methods. The relative positions of hand and objects are then reported to a task monitor that compares the patient"s actions against a model of the target task. A prototype system has been constructed and is currently undergoing technical and clinical evaluation.

  13. Analysis and Recognition of Curve Type as The Basis of Object Recognition in Image

    NASA Astrophysics Data System (ADS)

    Nugraha, Nurma; Madenda, Sarifuddin; Indarti, Dina; Dewi Agushinta, R.; Ernastuti

    2016-06-01

    An object in an image when analyzed further will show the characteristics that distinguish one object with another object in an image. Characteristics that are used in object recognition in an image can be a color, shape, pattern, texture and spatial information that can be used to represent objects in the digital image. The method has recently been developed for image feature extraction on objects that share characteristics curve analysis (simple curve) and use the search feature of chain code object. This study will develop an algorithm analysis and the recognition of the type of curve as the basis for object recognition in images, with proposing addition of complex curve characteristics with maximum four branches that will be used for the process of object recognition in images. Definition of complex curve is the curve that has a point of intersection. By using some of the image of the edge detection, the algorithm was able to do the analysis and recognition of complex curve shape well.

  14. Object memory effects on figure assignment: conscious object recognition is not necessary or sufficient.

    PubMed

    Peterson, M A; de Gelder, B; Rapcsak, S Z; Gerhardstein, P C; Bachoud-Lévi, A

    2000-01-01

    In three experiments we investigated whether conscious object recognition is necessary or sufficient for effects of object memories on figure assignment. In experiment 1, we examined a brain-damaged participant, AD, whose conscious object recognition is severely impaired. AD's responses about figure assignment do reveal effects from memories of object structure, indicating that conscious object recognition is not necessary for these effects, and identifying the figure-ground test employed here as a new implicit test of access to memories of object structure. In experiments 2 and 3, we tested a second brain-damaged participant, WG, for whom conscious object recognition was relatively spared. Nevertheless, effects from memories of object structure on figure assignment were not evident in WG's responses about figure assignment in experiment 2, indicating that conscious object recognition is not sufficient for effects of object memories on figure assignment. WG's performance sheds light on AD's performance, and has implications for the theoretical understanding of object memory effects on figure assignment.

  15. Hippocampal histone acetylation regulates object recognition and the estradiol-induced enhancement of object recognition

    PubMed Central

    Zhao, Zaorui; Fan, Lu; Fortress, Ashley M.; Boulware, Marissa I.; Frick, Karyn M.

    2012-01-01

    Histone acetylation has recently been implicated in learning and memory processes, yet necessity of histone acetylation for such processes has not been demonstrated using pharmacological inhibitors of histone acetyltransferases (HATs). As such, the present study tested whether garcinol, a potent HAT inhibitor in vitro, could impair hippocampal memory consolidation and block the memory-enhancing effects of the modulatory hormone 17β-estradiol (E2). We first showed that bilateral infusion of garcinol (0.1, 1, or 10 μg/side) into the dorsal hippocampus (DH) immediately after training impaired object recognition memory consolidation in ovariectomized female mice. A behaviorally effective dose of garcinol (10 μg/side) also significantly decreased DH HAT activity. We next examined whether DH infusion of a behaviorally subeffective dose of garcinol (1 ng/side) could block the effects of DH E2 infusion on object recognition and epigenetic processes. Immediately after training, ovariectomized female mice received bilateral DH infusions of vehicle, E2 (5 μg/side), garcinol (1 ng/side), or E2 plus garcinol. Forty-eight hours later, garcinol blocked the memory-enhancing effects of E2. Garcinol also reversed the E2-induced increase in DH histone H3 acetylation, HAT activity, and levels of the de novo methyltransferase DNMT3B, as well as the E2-induced decrease in levels of the memory repressor protein histone deacetylase 2 (HDAC2). Collectively, these findings suggest that histone acetylation is critical for object recognition memory consolidation and the beneficial effects of E2 on object recognition. Importantly, this work demonstrates that the role of histone acetylation in memory processes can be studied using a HAT inhibitor. PMID:22396409

  16. Exploiting core knowledge for visual object recognition.

    PubMed

    Schurgin, Mark W; Flombaum, Jonathan I

    2017-03-01

    Humans recognize thousands of objects, and with relative tolerance to variable retinal inputs. The acquisition of this ability is not fully understood, and it remains an area in which artificial systems have yet to surpass people. We sought to investigate the memory process that supports object recognition. Specifically, we investigated the association of inputs that co-occur over short periods of time. We tested the hypothesis that human perception exploits expectations about object kinematics to limit the scope of association to inputs that are likely to have the same token as a source. In several experiments we exposed participants to images of objects, and we then tested recognition sensitivity. Using motion, we manipulated whether successive encounters with an image took place through kinematics that implied the same or a different token as the source of those encounters. Images were injected with noise, or shown at varying orientations, and we included 2 manipulations of motion kinematics. Across all experiments, memory performance was better for images that had been previously encountered with kinematics that implied a single token. A model-based analysis similarly showed greater memory strength when images were shown via kinematics that implied a single token. These results suggest that constraints from physics are built into the mechanisms that support memory about objects. Such constraints-often characterized as 'Core Knowledge'-are known to support perception and cognition broadly, even in young infants. But they have never been considered as a mechanism for memory with respect to recognition. (PsycINFO Database Record (c) 2017 APA, all rights reserved).

  17. General object recognition is specific: Evidence from novel and familiar objects.

    PubMed

    Richler, Jennifer J; Wilmer, Jeremy B; Gauthier, Isabel

    2017-09-01

    In tests of object recognition, individual differences typically correlate modestly but nontrivially across familiar categories (e.g. cars, faces, shoes, birds, mushrooms). In theory, these correlations could reflect either global, non-specific mechanisms, such as general intelligence (IQ), or more specific mechanisms. Here, we introduce two separate methods for effectively capturing category-general performance variation, one that uses novel objects and one that uses familiar objects. In each case, we show that category-general performance variance is unrelated to IQ, thereby implicating more specific mechanisms. The first approach examines three newly developed novel object memory tests (NOMTs). We predicted that NOMTs would exhibit more shared, category-general variance than familiar object memory tests (FOMTs) because novel objects, unlike familiar objects, lack category-specific environmental influences (e.g. exposure to car magazines or botany classes). This prediction held, and remarkably, virtually none of the substantial shared variance among NOMTs was explained by IQ. Also, while NOMTs correlated nontrivially with two FOMTs (faces, cars), these correlations were smaller than among NOMTs and no larger than between the face and car tests themselves, suggesting that the category-general variance captured by NOMTs is specific not only relative to IQ, but also, to some degree, relative to both face and car recognition. The second approach averaged performance across multiple FOMTs, which we predicted would increase category-general variance by averaging out category-specific factors. This prediction held, and as with NOMTs, virtually none of the shared variance among FOMTs was explained by IQ. Overall, these results support the existence of object recognition mechanisms that, though category-general, are specific relative to IQ and substantially separable from face and car recognition. They also add sensitive, well-normed NOMTs to the tools available to study

  18. Central administration of angiotensin IV rapidly enhances novel object recognition among mice

    PubMed Central

    Paris, Jason J.; Eans, Shainnel O.; Mizrachi, Elisa; Reilley, Kate J.; Ganno, Michelle L.; McLaughlin, Jay P.

    2013-01-01

    Angiotensin IV (Val1-Tyr2-Ile3-His4-Pro5-Phe6) has demonstrated potential cognitive-enhancing effects. The present investigation assessed and characterized: (1) dose-dependency of angiotensin IV's cognitive enhancement in a C57BL/6J mouse model of novel object recognition, (2) the time-course for these effects, (3) the identity of residues in the hexapeptide important to these effects and (4) the necessity of actions at angiotensin IV receptors for pro-cognitive activity. Assessment of C57BL/6J mice in a novel object recognition task demonstrated that prior administration of angiotensin IV (0.1, 1.0, or 10.0, but not 0.01, nmol, i.c.v.) significantly enhanced novel object recognition in a dose-dependent manner. These effects were time dependent, with improved novel object recognition observed when angiotensin IV (0.1 nmol, i.c.v.) was administered 10 or 20, but not 30, min prior to the onset of the novel object recognition testing. An alanine scan of the angiotensin IV peptide revealed that replacement of the Val1, Ile3, His4, or Phe6 residues with Ala attenuated peptide-induced improvements in novel object recognition, whereas Tyr2 or Pro5 replacement did not significantly affect performance. Administration of the angiotensin IV receptor antagonist, divalinal-Ang IV (20 nmol, i.c.v.), reduced (but did not abolish) novel object recognition; however, this antagonist completely blocked the pro-cognitive effects of angiotensin IV (0.1 nmol, i.c.v.) in this task. Rotorod testing demonstrated no locomotor effects for any angiotensin IV or divalinal-Ang IV dose tested. These data demonstrate that angiotensin IV produces a rapid enhancement of associative learning and memory performance in a mouse model that was dependent on the angiotensin IV receptor. PMID:23416700

  19. Central administration of angiotensin IV rapidly enhances novel object recognition among mice.

    PubMed

    Paris, Jason J; Eans, Shainnel O; Mizrachi, Elisa; Reilley, Kate J; Ganno, Michelle L; McLaughlin, Jay P

    2013-07-01

    Angiotensin IV (Val(1)-Tyr(2)-Ile(3)-His(4)-Pro(5)-Phe(6)) has demonstrated potential cognitive-enhancing effects. The present investigation assessed and characterized: (1) dose-dependency of angiotensin IV's cognitive enhancement in a C57BL/6J mouse model of novel object recognition, (2) the time-course for these effects, (3) the identity of residues in the hexapeptide important to these effects and (4) the necessity of actions at angiotensin IV receptors for procognitive activity. Assessment of C57BL/6J mice in a novel object recognition task demonstrated that prior administration of angiotensin IV (0.1, 1.0, or 10.0, but not 0.01 nmol, i.c.v.) significantly enhanced novel object recognition in a dose-dependent manner. These effects were time dependent, with improved novel object recognition observed when angiotensin IV (0.1 nmol, i.c.v.) was administered 10 or 20, but not 30 min prior to the onset of the novel object recognition testing. An alanine scan of the angiotensin IV peptide revealed that replacement of the Val(1), Ile(3), His(4), or Phe(6) residues with Ala attenuated peptide-induced improvements in novel object recognition, whereas Tyr(2) or Pro(5) replacement did not significantly affect performance. Administration of the angiotensin IV receptor antagonist, divalinal-Ang IV (20 nmol, i.c.v.), reduced (but did not abolish) novel object recognition; however, this antagonist completely blocked the procognitive effects of angiotensin IV (0.1 nmol, i.c.v.) in this task. Rotorod testing demonstrated no locomotor effects with any angiotensin IV or divalinal-Ang IV dose tested. These data demonstrate that angiotensin IV produces a rapid enhancement of associative learning and memory performance in a mouse model that was dependent on the angiotensin IV receptor. Copyright © 2013 Elsevier Ltd. All rights reserved.

  20. The Role of Perceptual Load in Object Recognition

    ERIC Educational Resources Information Center

    Lavie, Nilli; Lin, Zhicheng; Zokaei, Nahid; Thoma, Volker

    2009-01-01

    Predictions from perceptual load theory (Lavie, 1995, 2005) regarding object recognition across the same or different viewpoints were tested. Results showed that high perceptual load reduces distracter recognition levels despite always presenting distracter objects from the same view. They also showed that the levels of distracter recognition were…

  1. Object recognition of ladar with support vector machine

    NASA Astrophysics Data System (ADS)

    Sun, Jian-Feng; Li, Qi; Wang, Qi

    2005-01-01

    Intensity, range and Doppler images can be obtained by using laser radar. Laser radar can detect much more object information than other detecting sensor, such as passive infrared imaging and synthetic aperture radar (SAR), so it is well suited as the sensor of object recognition. Traditional method of laser radar object recognition is extracting target features, which can be influenced by noise. In this paper, a laser radar recognition method-Support Vector Machine is introduced. Support Vector Machine (SVM) is a new hotspot of recognition research after neural network. It has well performance on digital written and face recognition. Two series experiments about SVM designed for preprocessing and non-preprocessing samples are performed by real laser radar images, and the experiments results are compared.

  2. Object Recognition and Localization: The Role of Tactile Sensors

    PubMed Central

    Aggarwal, Achint; Kirchner, Frank

    2014-01-01

    Tactile sensors, because of their intrinsic insensitivity to lighting conditions and water turbidity, provide promising opportunities for augmenting the capabilities of vision sensors in applications involving object recognition and localization. This paper presents two approaches for haptic object recognition and localization for ground and underwater environments. The first approach called Batch Ransac and Iterative Closest Point augmented Particle Filter (BRICPPF) is based on an innovative combination of particle filters, Iterative-Closest-Point algorithm, and a feature-based Random Sampling and Consensus (RANSAC) algorithm for database matching. It can handle a large database of 3D-objects of complex shapes and performs a complete six-degree-of-freedom localization of static objects. The algorithms are validated by experimentation in ground and underwater environments using real hardware. To our knowledge this is the first instance of haptic object recognition and localization in underwater environments. The second approach is biologically inspired, and provides a close integration between exploration and recognition. An edge following exploration strategy is developed that receives feedback from the current state of recognition. A recognition by parts approach is developed which uses the BRICPPF for object sub-part recognition. Object exploration is either directed to explore a part until it is successfully recognized, or is directed towards new parts to endorse the current recognition belief. This approach is validated by simulation experiments. PMID:24553087

  3. Generalization between canonical and non-canonical views in object recognition

    PubMed Central

    Ghose, Tandra; Liu, Zili

    2013-01-01

    Viewpoint generalization in object recognition is the process that allows recognition of a given 3D object from many different viewpoints despite variations in its 2D projections. We used the canonical view effects as a foundation to empirically test the validity of a major theory in object recognition, the view-approximation model (Poggio & Edelman, 1990). This model predicts that generalization should be better when an object is first seen from a non-canonical view and then a canonical view than when seen in the reversed order. We also manipulated object similarity to study the degree to which this view generalization was constrained by shape details and task instructions (object vs. image recognition). Old-new recognition performance for basic and subordinate level objects was measured in separate blocks. We found that for object recognition, view generalization between canonical and non-canonical views was comparable for basic level objects. For subordinate level objects, recognition performance was more accurate from non-canonical to canonical views than the other way around. When the task was changed from object recognition to image recognition, the pattern of the results reversed. Interestingly, participants responded “old” to “new” images of “old” objects with a substantially higher rate than to “new” objects, despite instructions to the contrary, thereby indicating involuntary view generalization. Our empirical findings are incompatible with the prediction of the view-approximation theory, and argue against the hypothesis that views are stored independently. PMID:23283692

  4. Differential effects of spaced vs. massed training in long-term object-identity and object-location recognition memory.

    PubMed

    Bello-Medina, Paola C; Sánchez-Carrasco, Livia; González-Ornelas, Nadia R; Jeffery, Kathryn J; Ramírez-Amaya, Víctor

    2013-08-01

    Here we tested whether the well-known superiority of spaced training over massed training is equally evident in both object identity and object location recognition memory. We trained animals with objects placed in a variable or in a fixed location to produce a location-independent object identity memory or a location-dependent object representation. The training consisted of 5 trials that occurred either on one day (Massed) or over the course of 5 consecutive days (Spaced). The memory test was done in independent groups of animals either 24h or 7 days after the last training trial. In each test the animals were exposed to either a novel object, when trained with the objects in variable locations, or to a familiar object in a novel location, when trained with objects in fixed locations. The difference in time spent exploring the changed versus the familiar objects was used as a measure of recognition memory. For the object-identity-trained animals, spaced training produced clear evidence of recognition memory after both 24h and 7 days, but massed-training animals showed it only after 24h. In contrast, for the object-location-trained animals, recognition memory was evident after both retention intervals and with both training procedures. When objects were placed in variable locations for the two types of training and the test was done with a brand-new location, only the spaced-training animals showed recognition at 24h, but surprisingly, after 7 days, animals trained using both procedures were able to recognize the change, suggesting a post-training consolidation process. We suggest that the two training procedures trigger different neural mechanisms that may differ in the two segregated streams that process object information and that may consolidate differently. Copyright © 2013 Elsevier B.V. All rights reserved.

  5. Breaking object correspondence across saccades impairs object recognition: The role of color and luminance.

    PubMed

    Poth, Christian H; Schneider, Werner X

    2016-09-01

    Rapid saccadic eye movements bring the foveal region of the eye's retina onto objects for high-acuity vision. Saccades change the location and resolution of objects' retinal images. To perceive objects as visually stable across saccades, correspondence between the objects before and after the saccade must be established. We have previously shown that breaking object correspondence across the saccade causes a decrement in object recognition (Poth, Herwig, & Schneider, 2015). Color and luminance can establish object correspondence, but it is unknown how these surface features contribute to transsaccadic visual processing. Here, we investigated whether changing the surface features color-and-luminance and color alone across saccades impairs postsaccadic object recognition. Participants made saccades to peripheral objects, which either maintained or changed their surface features across the saccade. After the saccade, participants briefly viewed a letter within the saccade target object (terminated by a pattern mask). Postsaccadic object recognition was assessed as participants' accuracy in reporting the letter. Experiment A used the colors green and red with different luminances as surface features, Experiment B blue and yellow with approximately the same luminances. Changing the surface features across the saccade deteriorated postsaccadic object recognition in both experiments. These findings reveal a link between object recognition and object correspondence relying on the surface features colors and luminance, which is currently not addressed in theories of transsaccadic perception. We interpret the findings within a recent theory ascribing this link to visual attention (Schneider, 2013).

  6. Use of the recognition heuristic depends on the domain's recognition validity, not on the recognition validity of selected sets of objects.

    PubMed

    Pohl, Rüdiger F; Michalkiewicz, Martha; Erdfelder, Edgar; Hilbig, Benjamin E

    2017-07-01

    According to the recognition-heuristic theory, decision makers solve paired comparisons in which one object is recognized and the other not by recognition alone, inferring that recognized objects have higher criterion values than unrecognized ones. However, success-and thus usefulness-of this heuristic depends on the validity of recognition as a cue, and adaptive decision making, in turn, requires that decision makers are sensitive to it. To this end, decision makers could base their evaluation of the recognition validity either on the selected set of objects (the set's recognition validity), or on the underlying domain from which the objects were drawn (the domain's recognition validity). In two experiments, we manipulated the recognition validity both in the selected set of objects and between domains from which the sets were drawn. The results clearly show that use of the recognition heuristic depends on the domain's recognition validity, not on the set's recognition validity. In other words, participants treat all sets as roughly representative of the underlying domain and adjust their decision strategy adaptively (only) with respect to the more general environment rather than the specific items they are faced with.

  7. Resolving human object recognition in space and time

    PubMed Central

    Cichy, Radoslaw Martin; Pantazis, Dimitrios; Oliva, Aude

    2014-01-01

    A comprehensive picture of object processing in the human brain requires combining both spatial and temporal information about brain activity. Here, we acquired human magnetoencephalography (MEG) and functional magnetic resonance imaging (fMRI) responses to 92 object images. Multivariate pattern classification applied to MEG revealed the time course of object processing: whereas individual images were discriminated by visual representations early, ordinate and superordinate category levels emerged relatively later. Using representational similarity analysis, we combine human fMRI and MEG to show content-specific correspondence between early MEG responses and primary visual cortex (V1), and later MEG responses and inferior temporal (IT) cortex. We identified transient and persistent neural activities during object processing, with sources in V1 and IT., Finally, human MEG signals were correlated to single-unit responses in monkey IT. Together, our findings provide an integrated space- and time-resolved view of human object categorization during the first few hundred milliseconds of vision. PMID:24464044

  8. Orientation-Invariant Object Recognition: Evidence from Repetition Blindness

    ERIC Educational Resources Information Center

    Harris, Irina M.; Dux, Paul E.

    2005-01-01

    The question of whether object recognition is orientation-invariant or orientation-dependent was investigated using a repetition blindness (RB) paradigm. In RB, the second occurrence of a repeated stimulus is less likely to be reported, compared to the occurrence of a different stimulus, if it occurs within a short time of the first presentation.…

  9. Aniracetam restores object recognition impaired by age, scopolamine, and nucleus basalis lesions.

    PubMed

    Bartolini, L; Casamenti, F; Pepeu, G

    1996-02-01

    Object recognition was investigated in adult and aging male rats in a two-trials, unrewarded, test that assessed a form of working-episodic memory. Exploration time in the first trial, in which two copies of the same object were presented, was recorded. In the second trial, in which one of the familiar objects and a new object were presented, the time spent exploring the two objects was separately recorded and a discrimination index was calculated. Adult rats explored the new object longer than the familiar object when the intertrial time ranged from 1 to 60 min. Rats older than 20 months of age did not discriminate between familiar and new objects. Object discrimination was lost in adult rats after scopolamine (0.2 mg/kg SC) administration and with lesions of the nucleus basalis, resulting in a 40% decrease in cortical ChAT activity. Both aniracetam (25, 50, 100 mg/kg os) and oxiracetam (50 mg/kg os) restored object recognition in aging rats, in rats treated with scopolamine, and with lesions of the nucleus basalis. In the rat, object discrimination appears to depend on the integrity of the cholinergic system, and nootropic drugs can correct its disruption.

  10. Running Improves Pattern Separation during Novel Object Recognition.

    PubMed

    Bolz, Leoni; Heigele, Stefanie; Bischofberger, Josef

    2015-10-09

    Running increases adult neurogenesis and improves pattern separation in various memory tasks including context fear conditioning or touch-screen based spatial learning. However, it is unknown whether pattern separation is improved in spontaneous behavior, not emotionally biased by positive or negative reinforcement. Here we investigated the effect of voluntary running on pattern separation during novel object recognition in mice using relatively similar or substantially different objects.We show that running increases hippocampal neurogenesis but does not affect object recognition memory with 1.5 h delay after sample phase. By contrast, at 24 h delay, running significantly improves recognition memory for similar objects, whereas highly different objects can be distinguished by both, running and sedentary mice. These data show that physical exercise improves pattern separation, independent of negative or positive reinforcement. In sedentary mice there is a pronounced temporal gradient for remembering object details. In running mice, however, increased neurogenesis improves hippocampal coding and temporally preserves distinction of novel objects from familiar ones.

  11. Mechanisms of object recognition: what we have learned from pigeons

    PubMed Central

    Soto, Fabian A.; Wasserman, Edward A.

    2014-01-01

    Behavioral studies of object recognition in pigeons have been conducted for 50 years, yielding a large body of data. Recent work has been directed toward synthesizing this evidence and understanding the visual, associative, and cognitive mechanisms that are involved. The outcome is that pigeons are likely to be the non-primate species for which the computational mechanisms of object recognition are best understood. Here, we review this research and suggest that a core set of mechanisms for object recognition might be present in all vertebrates, including pigeons and people, making pigeons an excellent candidate model to study the neural mechanisms of object recognition. Behavioral and computational evidence suggests that error-driven learning participates in object category learning by pigeons and people, and recent neuroscientific research suggests that the basal ganglia, which are homologous in these species, may implement error-driven learning of stimulus-response associations. Furthermore, learning of abstract category representations can be observed in pigeons and other vertebrates. Finally, there is evidence that feedforward visual processing, a central mechanism in models of object recognition in the primate ventral stream, plays a role in object recognition by pigeons. We also highlight differences between pigeons and people in object recognition abilities, and propose candidate adaptive specializations which may explain them, such as holistic face processing and rule-based category learning in primates. From a modern comparative perspective, such specializations are to be expected regardless of the model species under study. The fact that we have a good idea of which aspects of object recognition differ in people and pigeons should be seen as an advantage over other animal models. From this perspective, we suggest that there is much to learn about human object recognition from studying the “simple” brains of pigeons. PMID:25352784

  12. Neurocomputational bases of object and face recognition.

    PubMed Central

    Biederman, I; Kalocsai, P

    1997-01-01

    A number of behavioural phenomena distinguish the recognition of faces and objects, even when members of a set of objects are highly similar. Because faces have the same parts in approximately the same relations, individuation of faces typically requires specification of the metric variation in a holistic and integral representation of the facial surface. The direct mapping of a hypercolumn-like pattern of activation onto a representation layer that preserves relative spatial filter values in a two-dimensional (2D) coordinate space, as proposed by C. von der Malsburg and his associates, may account for many of the phenomena associated with face recognition. An additional refinement, in which each column of filters (termed a 'jet') is centred on a particular facial feature (or fiducial point), allows selectivity of the input into the holistic representation to avoid incorporation of occluding or nearby surfaces. The initial hypercolumn representation also characterizes the first stage of object perception, but the image variation for objects at a given location in a 2D coordinate space may be too great to yield sufficient predictability directly from the output of spatial kernels. Consequently, objects can be represented by a structural description specifying qualitative (typically, non-accidental) characterizations of an object's parts, the attributes of the parts, and the relations among the parts, largely based on orientation and depth discontinuities (as shown by Hummel & Biederman). A series of experiments on the name priming or physical matching of complementary images (in the Fourier domain) of objects and faces documents that whereas face recognition is strongly dependent on the original spatial filter values, evidence from object recognition indicates strong invariance to these values, even when distinguishing among objects that are as similar as faces. PMID:9304687

  13. Object recognition with hierarchical discriminant saliency networks.

    PubMed

    Han, Sunhyoung; Vasconcelos, Nuno

    2014-01-01

    The benefits of integrating attention and object recognition are investigated. While attention is frequently modeled as a pre-processor for recognition, we investigate the hypothesis that attention is an intrinsic component of recognition and vice-versa. This hypothesis is tested with a recognition model, the hierarchical discriminant saliency network (HDSN), whose layers are top-down saliency detectors, tuned for a visual class according to the principles of discriminant saliency. As a model of neural computation, the HDSN has two possible implementations. In a biologically plausible implementation, all layers comply with the standard neurophysiological model of visual cortex, with sub-layers of simple and complex units that implement a combination of filtering, divisive normalization, pooling, and non-linearities. In a convolutional neural network implementation, all layers are convolutional and implement a combination of filtering, rectification, and pooling. The rectification is performed with a parametric extension of the now popular rectified linear units (ReLUs), whose parameters can be tuned for the detection of target object classes. This enables a number of functional enhancements over neural network models that lack a connection to saliency, including optimal feature denoising mechanisms for recognition, modulation of saliency responses by the discriminant power of the underlying features, and the ability to detect both feature presence and absence. In either implementation, each layer has a precise statistical interpretation, and all parameters are tuned by statistical learning. Each saliency detection layer learns more discriminant saliency templates than its predecessors and higher layers have larger pooling fields. This enables the HDSN to simultaneously achieve high selectivity to target object classes and invariance. The performance of the network in saliency and object recognition tasks is compared to those of models from the biological and

  14. Changes in Visual Object Recognition Precede the Shape Bias in Early Noun Learning

    PubMed Central

    Yee, Meagan; Jones, Susan S.; Smith, Linda B.

    2012-01-01

    Two of the most formidable skills that characterize human beings are language and our prowess in visual object recognition. They may also be developmentally intertwined. Two experiments, a large sample cross-sectional study and a smaller sample 6-month longitudinal study of 18- to 24-month-olds, tested a hypothesized developmental link between changes in visual object representation and noun learning. Previous findings in visual object recognition indicate that children’s ability to recognize common basic level categories from sparse structural shape representations of object shape emerges between the ages of 18 and 24 months, is related to noun vocabulary size, and is lacking in children with language delay. Other research shows in artificial noun learning tasks that during this same developmental period, young children systematically generalize object names by shape, that this shape bias predicts future noun learning, and is lacking in children with language delay. The two experiments examine the developmental relation between visual object recognition and the shape bias for the first time. The results show that developmental changes in visual object recognition systematically precede the emergence of the shape bias. The results suggest a developmental pathway in which early changes in visual object recognition that are themselves linked to category learning enable the discovery of higher-order regularities in category structure and thus the shape bias in novel noun learning tasks. The proposed developmental pathway has implications for understanding the role of specific experience in the development of both visual object recognition and the shape bias in early noun learning. PMID:23227015

  15. Development of visuo-haptic transfer for object recognition in typical preschool and school-aged children.

    PubMed

    Purpura, Giulia; Cioni, Giovanni; Tinelli, Francesca

    2018-07-01

    Object recognition is a long and complex adaptive process and its full maturation requires combination of many different sensory experiences as well as cognitive abilities to manipulate previous experiences in order to develop new percepts and subsequently to learn from the environment. It is well recognized that the transfer of visual and haptic information facilitates object recognition in adults, but less is known about development of this ability. In this study, we explored the developmental course of object recognition capacity in children using unimodal visual information, unimodal haptic information, and visuo-haptic information transfer in children from 4 years to 10 years and 11 months of age. Participants were tested through a clinical protocol, involving visual exploration of black-and-white photographs of common objects, haptic exploration of real objects, and visuo-haptic transfer of these two types of information. Results show an age-dependent development of object recognition abilities for visual, haptic, and visuo-haptic modalities. A significant effect of time on development of unimodal and crossmodal recognition skills was found. Moreover, our data suggest that multisensory processes for common object recognition are active at 4 years of age. They facilitate recognition of common objects, and, although not fully mature, are significant in adaptive behavior from the first years of age. The study of typical development of visuo-haptic processes in childhood is a starting point for future studies regarding object recognition in impaired populations.

  16. DORSAL HIPPOCAMPAL PROGESTERONE INFUSIONS ENHANCE OBJECT RECOGNITION IN YOUNG FEMALE MICE

    PubMed Central

    Orr, Patrick T.; Lewis, Michael C.; Frick, Karyn M.

    2009-01-01

    The effects of progesterone on memory are not nearly as well studied as the effects of estrogens. Although progesterone can reportedly enhance spatial and/or object recognition in female rodents when given immediately after training, previous studies have injected progesterone systemically, and therefore, the brain regions mediating this enhancement are not clear. As such, this study was designed to determine the role of the dorsal hippocampus in mediating the beneficial effect of progesterone on object recognition. Young ovariectomized C57BL/6 mice were trained in a hippocampal-dependent object recognition task utilizing two identical objects, and then immediately or 2 hrs afterwards, received bilateral dorsal hippocampal infusions of vehicle or 0.01, 0.1, or 1.0 μg/μl water-soluble progesterone. Forty-eight hours later, object recognition memory was tested using a previously explored object and a novel object. Relative to the vehicle group, memory for the familiar object was enhanced in all groups receiving immediate infusions of progesterone. Progesterone infusion delayed 2 hrs after training did not affect object recognition. These data suggest that the dorsal hippocampus may play a critical role in progesterone-induced enhancement of object recognition. PMID:19477194

  17. Real-time optical multiple object recognition and tracking system and method

    NASA Technical Reports Server (NTRS)

    Chao, Tien-Hsin (Inventor); Liu, Hua-Kuang (Inventor)

    1990-01-01

    System for optically recognizing and tracking a plurality of objects within a field of vision. Laser (46) produces a coherent beam (48). Beam splitter (24) splits the beam into object (26) and reference (28) beams. Beam expanders (50) and collimators (52) transform the beams (26, 28) into coherent collimated light beams (26', 28'). A two-dimensional SLM (54), disposed in the object beam (26'), modulates the object beam with optical information as a function of signals from a first camera (16) which develops X and Y signals reflecting the contents of its field of vision. A hololens (38), positioned in the object beam (26') subsequent to the modulator (54), focuses the object beam at a plurality of focal points (42). A planar transparency-forming film (32), disposed with the focal points on an exposable surface, forms a multiple position interference filter (62) upon exposure of the surface and development processing of the film (32). A reflector (53) directing the reference beam (28') onto the film (32), exposes the surface, with images focused by the hololens (38), to form interference patterns on the surface. There is apparatus (16', 64) for sensing and indicating light passage through respective ones of the positions of the filter (62), whereby recognition of objects corresponding to respective ones of the positions of the filter (62) is affected. For tracking, apparatus (64) focuses light passing through the filter (62) onto a matrix of CCD's in a second camera (16') to form a two-dimensional display of the recognized objects.

  18. Experience moderates overlap between object and face recognition, suggesting a common ability

    PubMed Central

    Gauthier, Isabel; McGugin, Rankin W.; Richler, Jennifer J.; Herzmann, Grit; Speegle, Magen; Van Gulick, Ana E.

    2014-01-01

    Some research finds that face recognition is largely independent from the recognition of other objects; a specialized and innate ability to recognize faces could therefore have little or nothing to do with our ability to recognize objects. We propose a new framework in which recognition performance for any category is the product of domain-general ability and category-specific experience. In Experiment 1, we show that the overlap between face and object recognition depends on experience with objects. In 256 subjects we measured face recognition, object recognition for eight categories, and self-reported experience with these categories. Experience predicted neither face recognition nor object recognition but moderated their relationship: Face recognition performance is increasingly similar to object recognition performance with increasing object experience. If a subject has a lot of experience with objects and is found to perform poorly, they also prove to have a low ability with faces. In a follow-up survey, we explored the dimensions of experience with objects that may have contributed to self-reported experience in Experiment 1. Different dimensions of experience appear to be more salient for different categories, with general self-reports of expertise reflecting judgments of verbal knowledge about a category more than judgments of visual performance. The complexity of experience and current limitations in its measurement support the importance of aggregating across multiple categories. Our findings imply that both face and object recognition are supported by a common, domain-general ability expressed through experience with a category and best measured when accounting for experience. PMID:24993021

  19. Experience moderates overlap between object and face recognition, suggesting a common ability.

    PubMed

    Gauthier, Isabel; McGugin, Rankin W; Richler, Jennifer J; Herzmann, Grit; Speegle, Magen; Van Gulick, Ana E

    2014-07-03

    Some research finds that face recognition is largely independent from the recognition of other objects; a specialized and innate ability to recognize faces could therefore have little or nothing to do with our ability to recognize objects. We propose a new framework in which recognition performance for any category is the product of domain-general ability and category-specific experience. In Experiment 1, we show that the overlap between face and object recognition depends on experience with objects. In 256 subjects we measured face recognition, object recognition for eight categories, and self-reported experience with these categories. Experience predicted neither face recognition nor object recognition but moderated their relationship: Face recognition performance is increasingly similar to object recognition performance with increasing object experience. If a subject has a lot of experience with objects and is found to perform poorly, they also prove to have a low ability with faces. In a follow-up survey, we explored the dimensions of experience with objects that may have contributed to self-reported experience in Experiment 1. Different dimensions of experience appear to be more salient for different categories, with general self-reports of expertise reflecting judgments of verbal knowledge about a category more than judgments of visual performance. The complexity of experience and current limitations in its measurement support the importance of aggregating across multiple categories. Our findings imply that both face and object recognition are supported by a common, domain-general ability expressed through experience with a category and best measured when accounting for experience. © 2014 ARVO.

  20. Decreased acetylcholine release delays the consolidation of object recognition memory.

    PubMed

    De Jaeger, Xavier; Cammarota, Martín; Prado, Marco A M; Izquierdo, Iván; Prado, Vania F; Pereira, Grace S

    2013-02-01

    Acetylcholine (ACh) is important for different cognitive functions such as learning, memory and attention. The release of ACh depends on its vesicular loading by the vesicular acetylcholine transporter (VAChT). It has been demonstrated that VAChT expression can modulate object recognition memory. However, the role of VAChT expression on object recognition memory persistence still remains to be understood. To address this question we used distinct mouse lines with reduced expression of VAChT, as well as pharmacological manipulations of the cholinergic system. We showed that reduction of cholinergic tone impairs object recognition memory measured at 24h. Surprisingly, object recognition memory, measured at 4 days after training, was impaired by substantial, but not moderate, reduction in VAChT expression. Our results suggest that levels of acetylcholine release strongly modulate object recognition memory consolidation and appear to be of particular importance for memory persistence 4 days after training. Copyright © 2012 Elsevier B.V. All rights reserved.

  1. New technique for real-time distortion-invariant multiobject recognition and classification

    NASA Astrophysics Data System (ADS)

    Hong, Rutong; Li, Xiaoshun; Hong, En; Wang, Zuyi; Wei, Hongan

    2001-04-01

    A real-time hybrid distortion-invariant OPR system was established to make 3D multiobject distortion-invariant automatic pattern recognition. Wavelet transform technique was used to make digital preprocessing of the input scene, to depress the noisy background and enhance the recognized object. A three-layer backpropagation artificial neural network was used in correlation signal post-processing to perform multiobject distortion-invariant recognition and classification. The C-80 and NOA real-time processing ability and the multithread programming technology were used to perform high speed parallel multitask processing and speed up the post processing rate to ROIs. The reference filter library was constructed for the distortion version of 3D object model images based on the distortion parameter tolerance measuring as rotation, azimuth and scale. The real-time optical correlation recognition testing of this OPR system demonstrates that using the preprocessing, post- processing, the nonlinear algorithm os optimum filtering, RFL construction technique and the multithread programming technology, a high possibility of recognition and recognition rate ere obtained for the real-time multiobject distortion-invariant OPR system. The recognition reliability and rate was improved greatly. These techniques are very useful to automatic target recognition.

  2. Trajectory Recognition as the Basis for Object Individuation: A Functional Model of Object File Instantiation and Object-Token Encoding

    PubMed Central

    Fields, Chris

    2011-01-01

    The perception of persisting visual objects is mediated by transient intermediate representations, object files, that are instantiated in response to some, but not all, visual trajectories. The standard object file concept does not, however, provide a mechanism sufficient to account for all experimental data on visual object persistence, object tracking, and the ability to perceive spatially disconnected stimuli as continuously existing objects. Based on relevant anatomical, functional, and developmental data, a functional model is constructed that bases visual object individuation on the recognition of temporal sequences of apparent center-of-mass positions that are specifically identified as trajectories by dedicated “trajectory recognition networks” downstream of the medial–temporal motion-detection area. This model is shown to account for a wide range of data, and to generate a variety of testable predictions. Individual differences in the recognition, abstraction, and encoding of trajectory information are expected to generate distinct object persistence judgments and object recognition abilities. Dominance of trajectory information over feature information in stored object tokens during early infancy, in particular, is expected to disrupt the ability to re-identify human and other individuals across perceptual episodes, and lead to developmental outcomes with characteristics of autism spectrum disorders. PMID:21716599

  3. A method of object recognition for single pixel imaging

    NASA Astrophysics Data System (ADS)

    Li, Boxuan; Zhang, Wenwen

    2018-01-01

    Computational ghost imaging(CGI), utilizing a single-pixel detector, has been extensively used in many fields. However, in order to achieve a high-quality reconstructed image, a large number of iterations are needed, which limits the flexibility of using CGI in practical situations, especially in the field of object recognition. In this paper, we purpose a method utilizing the feature matching to identify the number objects. In the given system, approximately 90% of accuracy of recognition rates can be achieved, which provides a new idea for the application of single pixel imaging in the field of object recognition

  4. Object Recognition Memory and the Rodent Hippocampus

    ERIC Educational Resources Information Center

    Broadbent, Nicola J.; Gaskin, Stephane; Squire, Larry R.; Clark, Robert E.

    2010-01-01

    In rodents, the novel object recognition task (NOR) has become a benchmark task for assessing recognition memory. Yet, despite its widespread use, a consensus has not developed about which brain structures are important for task performance. We assessed both the anterograde and retrograde effects of hippocampal lesions on performance in the NOR…

  5. A Taxonomy of 3D Occluded Objects Recognition Techniques

    NASA Astrophysics Data System (ADS)

    Soleimanizadeh, Shiva; Mohamad, Dzulkifli; Saba, Tanzila; Al-ghamdi, Jarallah Saleh

    2016-03-01

    The overall performances of object recognition techniques under different condition (e.g., occlusion, viewpoint, and illumination) have been improved significantly in recent years. New applications and hardware are shifted towards digital photography, and digital media. This faces an increase in Internet usage requiring object recognition for certain applications; particularly occulded objects. However occlusion is still an issue unhandled, interlacing the relations between extracted feature points through image, research is going on to develop efficient techniques and easy to use algorithms that would help users to source images; this need to overcome problems and issues regarding occlusion. The aim of this research is to review recognition occluded objects algorithms and figure out their pros and cons to solve the occlusion problem features, which are extracted from occluded object to distinguish objects from other co-existing objects by determining the new techniques, which could differentiate the occluded fragment and sections inside an image.

  6. [Recognition of visual objects under forward masking. Effects of cathegorial similarity of test and masking stimuli].

    PubMed

    Gerasimenko, N Iu; Slavutskaia, A V; Kalinin, S A; Kulikov, M A; Mikhaĭlova, E S

    2013-01-01

    In 38 healthy subjects accuracy and response time were examined during recognition of two categories of images--animals andnonliving objects--under forward masking. We revealed new data that masking effects depended of categorical similarity of target and masking stimuli. The recognition accuracy was the lowest and the response time was the most slow, when the target and masking stimuli belongs to the same category, that was combined with high dispersion of response times. The revealed effects were more clear in the task of animal recognition in comparison with the recognition of nonliving objects. We supposed that the revealed effects connected with interference between cortical representations of the target and masking stimuli and discussed our results in context of cortical interference and negative priming.

  7. Deletion of the GluA1 AMPA receptor subunit impairs recency-dependent object recognition memory

    PubMed Central

    Sanderson, David J.; Hindley, Emma; Smeaton, Emily; Denny, Nick; Taylor, Amy; Barkus, Chris; Sprengel, Rolf; Seeburg, Peter H.; Bannerman, David M.

    2011-01-01

    Deletion of the GluA1 AMPA receptor subunit impairs short-term spatial recognition memory. It has been suggested that short-term recognition depends upon memory caused by the recent presentation of a stimulus that is independent of contextual–retrieval processes. The aim of the present set of experiments was to test whether the role of GluA1 extends to nonspatial recognition memory. Wild-type and GluA1 knockout mice were tested on the standard object recognition task and a context-independent recognition task that required recency-dependent memory. In a first set of experiments it was found that GluA1 deletion failed to impair performance on either of the object recognition or recency-dependent tasks. However, GluA1 knockout mice displayed increased levels of exploration of the objects in both the sample and test phases compared to controls. In contrast, when the time that GluA1 knockout mice spent exploring the objects was yoked to control mice during the sample phase, it was found that GluA1 deletion now impaired performance on both the object recognition and the recency-dependent tasks. GluA1 deletion failed to impair performance on a context-dependent recognition task regardless of whether object exposure in knockout mice was yoked to controls or not. These results demonstrate that GluA1 is necessary for nonspatial as well as spatial recognition memory and plays an important role in recency-dependent memory processes. PMID:21378100

  8. Parallel and distributed computation for fault-tolerant object recognition

    NASA Technical Reports Server (NTRS)

    Wechsler, Harry

    1988-01-01

    The distributed associative memory (DAM) model is suggested for distributed and fault-tolerant computation as it relates to object recognition tasks. The fault-tolerance is with respect to geometrical distortions (scale and rotation), noisy inputs, occulsion/overlap, and memory faults. An experimental system was developed for fault-tolerant structure recognition which shows the feasibility of such an approach. The approach is futher extended to the problem of multisensory data integration and applied successfully to the recognition of colored polyhedral objects.

  9. Three-dimensional object recognition using similar triangles and decision trees

    NASA Technical Reports Server (NTRS)

    Spirkovska, Lilly

    1993-01-01

    A system, TRIDEC, that is capable of distinguishing between a set of objects despite changes in the objects' positions in the input field, their size, or their rotational orientation in 3D space is described. TRIDEC combines very simple yet effective features with the classification capabilities of inductive decision tree methods. The feature vector is a list of all similar triangles defined by connecting all combinations of three pixels in a coarse coded 127 x 127 pixel input field. The classification is accomplished by building a decision tree using the information provided from a limited number of translated, scaled, and rotated samples. Simulation results are presented which show that TRIDEC achieves 94 percent recognition accuracy in the 2D invariant object recognition domain and 98 percent recognition accuracy in the 3D invariant object recognition domain after training on only a small sample of transformed views of the objects.

  10. Developmental Commonalities between Object and Face Recognition in Adolescence

    PubMed Central

    Jüttner, Martin; Wakui, Elley; Petters, Dean; Davidoff, Jules

    2016-01-01

    In the visual perception literature, the recognition of faces has often been contrasted with that of non-face objects, in terms of differences with regard to the role of parts, part relations and holistic processing. However, recent evidence from developmental studies has begun to blur this sharp distinction. We review evidence for a protracted development of object recognition that is reminiscent of the well-documented slow maturation observed for faces. The prolonged development manifests itself in a retarded processing of metric part relations as opposed to that of individual parts and offers surprising parallels to developmental accounts of face recognition, even though the interpretation of the data is less clear with regard to holistic processing. We conclude that such results might indicate functional commonalities between the mechanisms underlying the recognition of faces and non-face objects, which are modulated by different task requirements in the two stimulus domains. PMID:27014176

  11. An ERP Study on Self-Relevant Object Recognition

    ERIC Educational Resources Information Center

    Miyakoshi, Makoto; Nomura, Michio; Ohira, Hideki

    2007-01-01

    We performed an event-related potential study to investigate the self-relevance effect in object recognition. Three stimulus categories were prepared: SELF (participant's own objects), FAMILIAR (disposable and public objects, defined as objects with less-self-relevant familiarity), and UNFAMILIAR (others' objects). The participants' task was to…

  12. The development of newborn object recognition in fast and slow visual worlds

    PubMed Central

    Wood, Justin N.; Wood, Samantha M. W.

    2016-01-01

    Object recognition is central to perception and cognition. Yet relatively little is known about the environmental factors that cause invariant object recognition to emerge in the newborn brain. Is this ability a hardwired property of vision? Or does the development of invariant object recognition require experience with a particular kind of visual environment? Here, we used a high-throughput controlled-rearing method to examine whether newborn chicks (Gallus gallus) require visual experience with slowly changing objects to develop invariant object recognition abilities. When newborn chicks were raised with a slowly rotating virtual object, the chicks built invariant object representations that generalized across novel viewpoints and rotation speeds. In contrast, when newborn chicks were raised with a virtual object that rotated more quickly, the chicks built viewpoint-specific object representations that failed to generalize to novel viewpoints and rotation speeds. Moreover, there was a direct relationship between the speed of the object and the amount of invariance in the chick's object representation. Thus, visual experience with slowly changing objects plays a critical role in the development of invariant object recognition. These results indicate that invariant object recognition is not a hardwired property of vision, but is learned rapidly when newborns encounter a slowly changing visual world. PMID:27097925

  13. A biologically inspired neural network model to transformation invariant object recognition

    NASA Astrophysics Data System (ADS)

    Iftekharuddin, Khan M.; Li, Yaqin; Siddiqui, Faraz

    2007-09-01

    perform a successful recognition task in general. Further, the residual critic error in DHP is generally smaller than that of HDP, and DHP achieves a 100% success rate more frequently than HDP for individual objects/subjects. On the other hand, HDP is more robust than the DHP as far as success rate across the database is concerned when applied in a stochastic and uncertain environment, and the computational time involved in DHP is more.

  14. Comparison of Object Recognition Behavior in Human and Monkey

    PubMed Central

    Rajalingham, Rishi; Schmidt, Kailyn

    2015-01-01

    Although the rhesus monkey is used widely as an animal model of human visual processing, it is not known whether invariant visual object recognition behavior is quantitatively comparable across monkeys and humans. To address this question, we systematically compared the core object recognition behavior of two monkeys with that of human subjects. To test true object recognition behavior (rather than image matching), we generated several thousand naturalistic synthetic images of 24 basic-level objects with high variation in viewing parameters and image background. Monkeys were trained to perform binary object recognition tasks on a match-to-sample paradigm. Data from 605 human subjects performing the same tasks on Mechanical Turk were aggregated to characterize “pooled human” object recognition behavior, as well as 33 separate Mechanical Turk subjects to characterize individual human subject behavior. Our results show that monkeys learn each new object in a few days, after which they not only match mean human performance but show a pattern of object confusion that is highly correlated with pooled human confusion patterns and is statistically indistinguishable from individual human subjects. Importantly, this shared human and monkey pattern of 3D object confusion is not shared with low-level visual representations (pixels, V1+; models of the retina and primary visual cortex) but is shared with a state-of-the-art computer vision feature representation. Together, these results are consistent with the hypothesis that rhesus monkeys and humans share a common neural shape representation that directly supports object perception. SIGNIFICANCE STATEMENT To date, several mammalian species have shown promise as animal models for studying the neural mechanisms underlying high-level visual processing in humans. In light of this diversity, making tight comparisons between nonhuman and human primates is particularly critical in determining the best use of nonhuman primates to

  15. From brain synapses to systems for learning and memory: Object recognition, spatial navigation, timed conditioning, and movement control.

    PubMed

    Grossberg, Stephen

    2015-09-24

    This article provides an overview of neural models of synaptic learning and memory whose expression in adaptive behavior depends critically on the circuits and systems in which the synapses are embedded. It reviews Adaptive Resonance Theory, or ART, models that use excitatory matching and match-based learning to achieve fast category learning and whose learned memories are dynamically stabilized by top-down expectations, attentional focusing, and memory search. ART clarifies mechanistic relationships between consciousness, learning, expectation, attention, resonance, and synchrony. ART models are embedded in ARTSCAN architectures that unify processes of invariant object category learning, recognition, spatial and object attention, predictive remapping, and eye movement search, and that clarify how conscious object vision and recognition may fail during perceptual crowding and parietal neglect. The generality of learned categories depends upon a vigilance process that is regulated by acetylcholine via the nucleus basalis. Vigilance can get stuck at too high or too low values, thereby causing learning problems in autism and medial temporal amnesia. Similar synaptic learning laws support qualitatively different behaviors: Invariant object category learning in the inferotemporal cortex; learning of grid cells and place cells in the entorhinal and hippocampal cortices during spatial navigation; and learning of time cells in the entorhinal-hippocampal system during adaptively timed conditioning, including trace conditioning. Spatial and temporal processes through the medial and lateral entorhinal-hippocampal system seem to be carried out with homologous circuit designs. Variations of a shared laminar neocortical circuit design have modeled 3D vision, speech perception, and cognitive working memory and learning. A complementary kind of inhibitory matching and mismatch learning controls movement. This article is part of a Special Issue entitled SI: Brain and Memory

  16. The subjective experience of object recognition: comparing metacognition for object detection and object categorization.

    PubMed

    Meuwese, Julia D I; van Loon, Anouk M; Lamme, Victor A F; Fahrenfort, Johannes J

    2014-05-01

    Perceptual decisions seem to be made automatically and almost instantly. Constructing a unitary subjective conscious experience takes more time. For example, when trying to avoid a collision with a car on a foggy road you brake or steer away in a reflex, before realizing you were in a near accident. This subjective aspect of object recognition has been given little attention. We used metacognition (assessed with confidence ratings) to measure subjective experience during object detection and object categorization for degraded and masked objects, while objective performance was matched. Metacognition was equal for degraded and masked objects, but categorization led to higher metacognition than did detection. This effect turned out to be driven by a difference in metacognition for correct rejection trials, which seemed to be caused by an asymmetry of the distractor stimulus: It does not contain object-related information in the detection task, whereas it does contain such information in the categorization task. Strikingly, this asymmetry selectively impacted metacognitive ability when objective performance was matched. This finding reveals a fundamental difference in how humans reflect versus act on information: When matching the amount of information required to perform two tasks at some objective level of accuracy (acting), metacognitive ability (reflecting) is still better in tasks that rely on positive evidence (categorization) than in tasks that rely more strongly on an absence of evidence (detection).

  17. The evolution of meaning: spatio-temporal dynamics of visual object recognition.

    PubMed

    Clarke, Alex; Taylor, Kirsten I; Tyler, Lorraine K

    2011-08-01

    Research on the spatio-temporal dynamics of visual object recognition suggests a recurrent, interactive model whereby an initial feedforward sweep through the ventral stream to prefrontal cortex is followed by recurrent interactions. However, critical questions remain regarding the factors that mediate the degree of recurrent interactions necessary for meaningful object recognition. The novel prediction we test here is that recurrent interactivity is driven by increasing semantic integration demands as defined by the complexity of semantic information required by the task and driven by the stimuli. To test this prediction, we recorded magnetoencephalography data while participants named living and nonliving objects during two naming tasks. We found that the spatio-temporal dynamics of neural activity were modulated by the level of semantic integration required. Specifically, source reconstructed time courses and phase synchronization measures showed increased recurrent interactions as a function of semantic integration demands. These findings demonstrate that the cortical dynamics of object processing are modulated by the complexity of semantic information required from the visual input.

  18. Perceptual Plasticity for Auditory Object Recognition

    PubMed Central

    Heald, Shannon L. M.; Van Hedger, Stephen C.; Nusbaum, Howard C.

    2017-01-01

    In our auditory environment, we rarely experience the exact acoustic waveform twice. This is especially true for communicative signals that have meaning for listeners. In speech and music, the acoustic signal changes as a function of the talker (or instrument), speaking (or playing) rate, and room acoustics, to name a few factors. Yet, despite this acoustic variability, we are able to recognize a sentence or melody as the same across various kinds of acoustic inputs and determine meaning based on listening goals, expectations, context, and experience. The recognition process relates acoustic signals to prior experience despite variability in signal-relevant and signal-irrelevant acoustic properties, some of which could be considered as “noise” in service of a recognition goal. However, some acoustic variability, if systematic, is lawful and can be exploited by listeners to aid in recognition. Perceivable changes in systematic variability can herald a need for listeners to reorganize perception and reorient their attention to more immediately signal-relevant cues. This view is not incorporated currently in many extant theories of auditory perception, which traditionally reduce psychological or neural representations of perceptual objects and the processes that act on them to static entities. While this reduction is likely done for the sake of empirical tractability, such a reduction may seriously distort the perceptual process to be modeled. We argue that perceptual representations, as well as the processes underlying perception, are dynamically determined by an interaction between the uncertainty of the auditory signal and constraints of context. This suggests that the process of auditory recognition is highly context-dependent in that the identity of a given auditory object may be intrinsically tied to its preceding context. To argue for the flexible neural and psychological updating of sound-to-meaning mappings across speech and music, we draw upon examples

  19. The role of color information on object recognition: a review and meta-analysis.

    PubMed

    Bramão, Inês; Reis, Alexandra; Petersson, Karl Magnus; Faísca, Luís

    2011-09-01

    In this study, we systematically review the scientific literature on the effect of color on object recognition. Thirty-five independent experiments, comprising 1535 participants, were included in a meta-analysis. We found a moderate effect of color on object recognition (d=0.28). Specific effects of moderator variables were analyzed and we found that color diagnosticity is the factor with the greatest moderator effect on the influence of color in object recognition; studies using color diagnostic objects showed a significant color effect (d=0.43), whereas a marginal color effect was found in studies that used non-color diagnostic objects (d=0.18). The present study did not permit the drawing of specific conclusions about the moderator effect of the object recognition task; while the meta-analytic review showed that color information improves object recognition mainly in studies using naming tasks (d=0.36), the literature review revealed a large body of evidence showing positive effects of color information on object recognition in studies using a large variety of visual recognition tasks. We also found that color is important for the ability to recognize artifacts and natural objects, to recognize objects presented as types (line-drawings) or as tokens (photographs), and to recognize objects that are presented without surface details, such as texture or shadow. Taken together, the results of the meta-analysis strongly support the contention that color plays a role in object recognition. This suggests that the role of color should be taken into account in models of visual object recognition. Copyright © 2011 Elsevier B.V. All rights reserved.

  20. It Takes Two–Skilled Recognition of Objects Engages Lateral Areas in Both Hemispheres

    PubMed Central

    Bilalić, Merim; Kiesel, Andrea; Pohl, Carsten; Erb, Michael; Grodd, Wolfgang

    2011-01-01

    Our object recognition abilities, a direct product of our experience with objects, are fine-tuned to perfection. Left temporal and lateral areas along the dorsal, action related stream, as well as left infero-temporal areas along the ventral, object related stream are engaged in object recognition. Here we show that expertise modulates the activity of dorsal areas in the recognition of man-made objects with clearly specified functions. Expert chess players were faster than chess novices in identifying chess objects and their functional relations. Experts' advantage was domain-specific as there were no differences between groups in a control task featuring geometrical shapes. The pattern of eye movements supported the notion that experts' extensive knowledge about domain objects and their functions enabled superior recognition even when experts were not directly fixating the objects of interest. Functional magnetic resonance imaging (fMRI) related exclusively the areas along the dorsal stream to chess specific object recognition. Besides the commonly involved left temporal and parietal lateral brain areas, we found that only in experts homologous areas on the right hemisphere were also engaged in chess specific object recognition. Based on these results, we discuss whether skilled object recognition does not only involve a more efficient version of the processes found in non-skilled recognition, but also qualitatively different cognitive processes which engage additional brain areas. PMID:21283683

  1. Measuring the Speed of Newborn Object Recognition in Controlled Visual Worlds

    ERIC Educational Resources Information Center

    Wood, Justin N.; Wood, Samantha M. W.

    2017-01-01

    How long does it take for a newborn to recognize an object? Adults can recognize objects rapidly, but measuring object recognition speed in newborns has not previously been possible. Here we introduce an automated controlled-rearing method for measuring the speed of newborn object recognition in controlled visual worlds. We raised newborn chicks…

  2. Higher-Order Neural Networks Applied to 2D and 3D Object Recognition

    NASA Technical Reports Server (NTRS)

    Spirkovska, Lilly; Reid, Max B.

    1994-01-01

    A Higher-Order Neural Network (HONN) can be designed to be invariant to geometric transformations such as scale, translation, and in-plane rotation. Invariances are built directly into the architecture of a HONN and do not need to be learned. Thus, for 2D object recognition, the network needs to be trained on just one view of each object class, not numerous scaled, translated, and rotated views. Because the 2D object recognition task is a component of the 3D object recognition task, built-in 2D invariance also decreases the size of the training set required for 3D object recognition. We present results for 2D object recognition both in simulation and within a robotic vision experiment and for 3D object recognition in simulation. We also compare our method to other approaches and show that HONNs have distinct advantages for position, scale, and rotation-invariant object recognition. The major drawback of HONNs is that the size of the input field is limited due to the memory required for the large number of interconnections in a fully connected network. We present partial connectivity strategies and a coarse-coding technique for overcoming this limitation and increasing the input field to that required by practical object recognition problems.

  3. The Timing of Visual Object Categorization

    PubMed Central

    Mack, Michael L.; Palmeri, Thomas J.

    2011-01-01

    An object can be categorized at different levels of abstraction: as natural or man-made, animal or plant, bird or dog, or as a Northern Cardinal or Pyrrhuloxia. There has been growing interest in understanding how quickly categorizations at different levels are made and how the timing of those perceptual decisions changes with experience. We specifically contrast two perspectives on the timing of object categorization at different levels of abstraction. By one account, the relative timing implies a relative timing of stages of visual processing that are tied to particular levels of object categorization: Fast categorizations are fast because they precede other categorizations within the visual processing hierarchy. By another account, the relative timing reflects when perceptual features are available over time and the quality of perceptual evidence used to drive a perceptual decision process: Fast simply means fast, it does not mean first. Understanding the short-term and long-term temporal dynamics of object categorizations is key to developing computational models of visual object recognition. We briefly review a number of models of object categorization and outline how they explain the timing of visual object categorization at different levels of abstraction. PMID:21811480

  4. Combining heterogenous features for 3D hand-held object recognition

    NASA Astrophysics Data System (ADS)

    Lv, Xiong; Wang, Shuang; Li, Xiangyang; Jiang, Shuqiang

    2014-10-01

    Object recognition has wide applications in the area of human-machine interaction and multimedia retrieval. However, due to the problem of visual polysemous and concept polymorphism, it is still a great challenge to obtain reliable recognition result for the 2D images. Recently, with the emergence and easy availability of RGB-D equipment such as Kinect, this challenge could be relieved because the depth channel could bring more information. A very special and important case of object recognition is hand-held object recognition, as hand is a straight and natural way for both human-human interaction and human-machine interaction. In this paper, we study the problem of 3D object recognition by combining heterogenous features with different modalities and extraction techniques. For hand-craft feature, although it reserves the low-level information such as shape and color, it has shown weakness in representing hiconvolutionalgh-level semantic information compared with the automatic learned feature, especially deep feature. Deep feature has shown its great advantages in large scale dataset recognition but is not always robust to rotation or scale variance compared with hand-craft feature. In this paper, we propose a method to combine hand-craft point cloud features and deep learned features in RGB and depth channle. First, hand-held object segmentation is implemented by using depth cues and human skeleton information. Second, we combine the extracted hetegerogenous 3D features in different stages using linear concatenation and multiple kernel learning (MKL). Then a training model is used to recognize 3D handheld objects. Experimental results validate the effectiveness and gerneralization ability of the proposed method.

  5. Ultra-fast Object Recognition from Few Spikes

    DTIC Science & Technology

    2005-07-06

    Computer Science and Artificial Intelligence Laboratory Ultra-fast Object Recognition from Few Spikes Chou Hung, Gabriel Kreiman , Tomaso Poggio...neural code for different kinds of object-related information. *The authors, Chou Hung and Gabriel Kreiman , contributed equally to this work...Supplementary Material is available at http://ramonycajal.mit.edu/ kreiman /resources/ultrafast

  6. Method of synthesized phase objects for pattern recognition with rotation invariance

    NASA Astrophysics Data System (ADS)

    Ostroukh, Alexander P.; Butok, Alexander M.; Shvets, Rostislav A.; Yezhov, Pavel V.; Kim, Jin-Tae; Kuzmenko, Alexander V.

    2015-11-01

    We present a development of the method of synthesized phase objects (SPO-method) [1] for the rotation-invariant pattern recognition. For the standard method of recognition and the SPO-method, the comparison of the parameters of correlation signals for a number of amplitude objects is executed at the realization of a rotation in an optical-digital correlator with the joint Fourier transformation. It is shown that not only the invariance relative to a rotation at a realization of the joint correlation for synthesized phase objects (SP-objects) but also the main advantage of the method of SP-objects over the reference one such as the unified δ-like recognition signal with the largest possible signal-to-noise ratio independent of the type of an object are attained.

  7. [Influence of object material and inter-trial interval on novel object recognition test in mice].

    PubMed

    Li, Sheng-jian; Huang, Zhu-yan; Ye, Yi-lu; Yu, Yue-ping; Zhang, Wei-ping; Wei, Er-qing; Zhang, Qi

    2014-05-01

    To investigate the efficacy of novel object recognition (NOR) test in assessment of learning and memory ability in ICR mice in different experimental conditions. One hundred and thirty male ICR mice were randomly divided into 10 groups: 4 groups for different inter-trial intervals (ITI: 10 min, 90 min, 4 h, 24 h), 4 groups for different object materials (wood-wood, plastic-plastic, plastic-wood, wood-plastic) and 2 groups for repeated test (measured once a day or every 3 days, totally three times in each group). The locomotor tracks in the open field were recorded. The amount of time spent exploring the novel and familiar objects, the discrimination ratio (DR) and the discrimination index (DI) were analyzed. Compared with familiar object, DR and DI of novel object were both increased at ITI of 10 min and 90 min (P<0.01). Exploring time, DR and DI were greatly influenced by different object materials. DR and DI remained stable by using identical object material. NOR test could be done repeatedly in the same batch of mice. NOR test can be used to assess the learning and memory ability in mice at shorter ITI and with identical material. It can be done repeatedly.

  8. Single prolonged stress impairs social and object novelty recognition in rats.

    PubMed

    Eagle, Andrew L; Fitzpatrick, Chris J; Perrine, Shane A

    2013-11-01

    Posttraumatic stress disorder (PTSD) results from exposure to a traumatic event and manifests as re-experiencing, arousal, avoidance, and negative cognition/mood symptoms. Avoidant symptoms, as well as the newly defined negative cognitions/mood, are a serious complication leading to diminished interest in once important or positive activities, such as social interaction; however, the basis of these symptoms remains poorly understood. PTSD patients also exhibit impaired object and social recognition, which may underlie the avoidance and symptoms of negative cognition, such as social estrangement or diminished interest in activities. Previous studies have demonstrated that single prolonged stress (SPS), models PTSD phenotypes, including impairments in learning and memory. Therefore, it was hypothesized that SPS would impair social and object recognition memory. Male Sprague Dawley rats were exposed to SPS then tested in the social choice test (SCT) or novel object recognition test (NOR). These tests measure recognition of novelty over familiarity, a natural preference of rodents. Results show that SPS impaired preference for both social and object novelty. In addition, SPS impairment in social recognition may be caused by impaired behavioral flexibility, or an inability to shift behavior during the SCT. These results demonstrate that traumatic stress can impair social and object recognition memory, which may underlie certain avoidant symptoms or negative cognition in PTSD and be related to impaired behavioral flexibility. Copyright © 2013 Elsevier B.V. All rights reserved.

  9. Electrophysiological evidence for effects of color knowledge in object recognition.

    PubMed

    Lu, Aitao; Xu, Guiping; Jin, Hua; Mo, Lei; Zhang, Jijia; Zhang, John X

    2010-01-29

    Knowledge about the typical colors associated with familiar everyday objects (i.e., strawberries are red) is well-known to be represented in the conceptual semantic system. Evidence that such knowledge may also play a role in early perceptual processes for object recognition is scant. In the present ERP study, participants viewed a list of object pictures and detected infrequent stimulus repetitions. Results show that shortly after stimulus onset, ERP components indexing early perceptual processes, including N1, P2, and N2, differentiated between objects in their appropriate or congruent color from these objects in an inappropriate or incongruent color. Such congruence effect also occurred in N3 associated with semantic processing of pictures but not in N4 for domain-general semantic processing. Our results demonstrate a clear effect of color knowledge in early object recognition stages and support the following proposal-color as a surface property is stored in a multiple-memory system where pre-semantic perceptual and semantic conceptual representations interact during object recognition. (c) 2009 Elsevier Ireland Ltd. All rights reserved.

  10. Nicotine Administration Attenuates Methamphetamine-Induced Novel Object Recognition Deficits

    PubMed Central

    Vieira-Brock, Paula L.; McFadden, Lisa M.; Nielsen, Shannon M.; Smith, Misty D.; Hanson, Glen R.

    2015-01-01

    Background: Previous studies have demonstrated that methamphetamine abuse leads to memory deficits and these are associated with relapse. Furthermore, extensive evidence indicates that nicotine prevents and/or improves memory deficits in different models of cognitive dysfunction and these nicotinic effects might be mediated by hippocampal or cortical nicotinic acetylcholine receptors. The present study investigated whether nicotine attenuates methamphetamine-induced novel object recognition deficits in rats and explored potential underlying mechanisms. Methods: Adolescent or adult male Sprague-Dawley rats received either nicotine water (10–75 μg/mL) or tap water for several weeks. Methamphetamine (4×7.5mg/kg/injection) or saline was administered either before or after chronic nicotine exposure. Novel object recognition was evaluated 6 days after methamphetamine or saline. Serotonin transporter function and density and α4β2 nicotinic acetylcholine receptor density were assessed on the following day. Results: Chronic nicotine intake via drinking water beginning during either adolescence or adulthood attenuated the novel object recognition deficits caused by a high-dose methamphetamine administration. Similarly, nicotine attenuated methamphetamine-induced deficits in novel object recognition when administered after methamphetamine treatment. However, nicotine did not attenuate the serotonergic deficits caused by methamphetamine in adults. Conversely, nicotine attenuated methamphetamine-induced deficits in α4β2 nicotinic acetylcholine receptor density in the hippocampal CA1 region. Furthermore, nicotine increased α4β2 nicotinic acetylcholine receptor density in the hippocampal CA3, dentate gyrus and perirhinal cortex in both saline- and methamphetamine-treated rats. Conclusions: Overall, these findings suggest that nicotine-induced increases in α4β2 nicotinic acetylcholine receptors in the hippocampus and perirhinal cortex might be one mechanism by which

  11. Track Everything: Limiting Prior Knowledge in Online Multi-Object Recognition.

    PubMed

    Wong, Sebastien C; Stamatescu, Victor; Gatt, Adam; Kearney, David; Lee, Ivan; McDonnell, Mark D

    2017-10-01

    This paper addresses the problem of online tracking and classification of multiple objects in an image sequence. Our proposed solution is to first track all objects in the scene without relying on object-specific prior knowledge, which in other systems can take the form of hand-crafted features or user-based track initialization. We then classify the tracked objects with a fast-learning image classifier, that is based on a shallow convolutional neural network architecture and demonstrate that object recognition improves when this is combined with object state information from the tracking algorithm. We argue that by transferring the use of prior knowledge from the detection and tracking stages to the classification stage, we can design a robust, general purpose object recognition system with the ability to detect and track a variety of object types. We describe our biologically inspired implementation, which adaptively learns the shape and motion of tracked objects, and apply it to the Neovision2 Tower benchmark data set, which contains multiple object types. An experimental evaluation demonstrates that our approach is competitive with the state-of-the-art video object recognition systems that do make use of object-specific prior knowledge in detection and tracking, while providing additional practical advantages by virtue of its generality.

  12. Affective and contextual values modulate spatial frequency use in object recognition

    PubMed Central

    Caplette, Laurent; West, Gregory; Gomot, Marie; Gosselin, Frédéric; Wicker, Bruno

    2014-01-01

    Visual object recognition is of fundamental importance in our everyday interaction with the environment. Recent models of visual perception emphasize the role of top-down predictions facilitating object recognition via initial guesses that limit the number of object representations that need to be considered. Several results suggest that this rapid and efficient object processing relies on the early extraction and processing of low spatial frequencies (LSF). The present study aimed to investigate the SF content of visual object representations and its modulation by contextual and affective values of the perceived object during a picture-name verification task. Stimuli consisted of pictures of objects equalized in SF content and categorized as having low or high affective and contextual values. To access the SF content of stored visual representations of objects, SFs of each image were then randomly sampled on a trial-by-trial basis. Results reveal that intermediate SFs between 14 and 24 cycles per object (2.3–4 cycles per degree) are correlated with fast and accurate identification for all categories of objects. Moreover, there was a significant interaction between affective and contextual values over the SFs correlating with fast recognition. These results suggest that affective and contextual values of a visual object modulate the SF content of its internal representation, thus highlighting the flexibility of the visual recognition system. PMID:24904514

  13. Lateral Entorhinal Cortex is Critical for Novel Object-Context Recognition

    PubMed Central

    Wilson, David IG; Langston, Rosamund F; Schlesiger, Magdalene I; Wagner, Monica; Watanabe, Sakurako; Ainge, James A

    2013-01-01

    Episodic memory incorporates information about specific events or occasions including spatial locations and the contextual features of the environment in which the event took place. It has been modeled in rats using spontaneous exploration of novel configurations of objects, their locations, and the contexts in which they are presented. While we have a detailed understanding of how spatial location is processed in the brain relatively little is known about where the nonspatial contextual components of episodic memory are processed. Initial experiments measured c-fos expression during an object-context recognition (OCR) task to examine which networks within the brain process contextual features of an event. Increased c-fos expression was found in the lateral entorhinal cortex (LEC; a major hippocampal afferent) during OCR relative to control conditions. In a subsequent experiment it was demonstrated that rats with lesions of LEC were unable to recognize object-context associations yet showed normal object recognition and normal context recognition. These data suggest that contextual features of the environment are integrated with object identity in LEC and demonstrate that recognition of such object-context associations requires the LEC. This is consistent with the suggestion that contextual features of an event are processed in LEC and that this information is combined with spatial information from medial entorhinal cortex to form episodic memory in the hippocampus. © 2013 Wiley Periodicals, Inc. PMID:23389958

  14. The uncrowded window of object recognition

    PubMed Central

    Pelli, Denis G; Tillman, Katharine A

    2009-01-01

    It is now emerging that vision is usually limited by object spacing rather than size. The visual system recognizes an object by detecting and then combining its features. ‘Crowding’ occurs when objects are too close together and features from several objects are combined into a jumbled percept. Here, we review the explosion of studies on crowding—in grating discrimination, letter and face recognition, visual search, selective attention, and reading—and find a universal principle, the Bouma law. The critical spacing required to prevent crowding is equal for all objects, although the effect is weaker between dissimilar objects. Furthermore, critical spacing at the cortex is independent of object position, and critical spacing at the visual field is proportional to object distance from fixation. The region where object spacing exceeds critical spacing is the ‘uncrowded window’. Observers cannot recognize objects outside of this window and its size limits the speed of reading and search. PMID:18828191

  15. Recognition Of Complex Three Dimensional Objects Using Three Dimensional Moment Invariants

    NASA Astrophysics Data System (ADS)

    Sadjadi, Firooz A.

    1985-01-01

    A technique for the recognition of complex three dimensional objects is presented. The complex 3-D objects are represented in terms of their 3-D moment invariants, algebraic expressions that remain invariant independent of the 3-D objects' orientations and locations in the field of view. The technique of 3-D moment invariants has been used successfully for simple 3-D object recognition in the past. In this work we have extended this method for the representation of more complex objects. Two complex objects are represented digitally; their 3-D moment invariants have been calculated, and then the invariancy of these 3-D invariant moment expressions is verified by changing the orientation and the location of the objects in the field of view. The results of this study have significant impact on 3-D robotic vision, 3-D target recognition, scene analysis and artificial intelligence.

  16. Metric invariance in object recognition: a review and further evidence.

    PubMed

    Cooper, E E; Biederman, I; Hummel, J E

    1992-06-01

    Phenomenologically, human shape recognition appears to be invariant with changes of orientation in depth (up to parts occlusion), position in the visual field, and size. Recent versions of template theories (e.g., Ullman, 1989; Lowe, 1987) assume that these invariances are achieved through the application of transformations such as rotation, translation, and scaling of the image so that it can be matched metrically to a stored template. Presumably, such transformations would require time for their execution. We describe recent priming experiments in which the effects of a prior brief presentation of an image on its subsequent recognition are assessed. The results of these experiments indicate that the invariance is complete: The magnitude of visual priming (as distinct from name or basic level concept priming) is not affected by a change in position, size, orientation in depth, or the particular lines and vertices present in the image, as long as representations of the same components can be activated. An implemented seven layer neural network model (Hummel & Biederman, 1992) that captures these fundamental properties of human object recognition is described. Given a line drawing of an object, the model activates a viewpoint-invariant structural description of the object, specifying its parts and their interrelations. Visual priming is interpreted as a change in the connection weights for the activation of: a) cells, termed geon feature assemblies (GFAs), that conjoin the output of units that represent invariant, independent properties of a single geon and its relations (such as its type, aspect ratio, relations to other geons), or b) a change in the connection weights by which several GFAs activate a cell representing an object.

  17. Differing time dependencies of object recognition memory impairments produced by nicotinic and muscarinic cholinergic antagonism in perirhinal cortex

    PubMed Central

    Tinsley, Chris J.; Fontaine-Palmer, Nadine S.; Vincent, Maria; Endean, Emma P.E.; Aggleton, John P.; Brown, Malcolm W.; Warburton, E. Clea

    2011-01-01

    The roles of muscarinic and nicotinic cholinergic receptors in perirhinal cortex in object recognition memory were compared. Rats' discrimination of a novel object preference test (NOP) test was measured after either systemic or local infusion into the perirhinal cortex of the nicotinic receptor antagonist methyllycaconitine (MLA), which targets alpha-7 (α7) amongst other nicotinic receptors or the muscarinic receptor antagonists scopolamine, AFDX-384, and pirenzepine. Methyllycaconitine administered systemically or intraperirhinally before acquisition impaired recognition memory tested after a 24-h, but not a 20-min delay. In contrast, all three muscarinic antagonists produced a similar, unusual pattern of impairment with amnesia after a 20-min delay, but remembrance after a 24-h delay. Thus, the amnesic effects of nicotinic and muscarinic antagonism were doubly dissociated across the 20-min and 24-h delays. The same pattern of shorter-term but not longer-term memory impairment was found for scopolamine whether the object preference test was carried out in a square arena or a Y-maze and whether rats of the Dark Agouti or Lister-hooded strains were used. Coinfusion of MLA and either scopolamine or AFDX-384 produced an impairment profile matching that for MLA. Hence, the antagonists did not act additively when coadministered. These findings establish an important role in recognition memory for both nicotinic and muscarinic cholinergic receptors in perirhinal cortex, and provide a challenge to simple ideas about the role of cholinergic processes in recognition memory: The effects of muscarinic and nicotinic antagonism are neither independent nor additive. PMID:21693636

  18. Differing time dependencies of object recognition memory impairments produced by nicotinic and muscarinic cholinergic antagonism in perirhinal cortex.

    PubMed

    Tinsley, Chris J; Fontaine-Palmer, Nadine S; Vincent, Maria; Endean, Emma P E; Aggleton, John P; Brown, Malcolm W; Warburton, E Clea

    2011-01-01

    The roles of muscarinic and nicotinic cholinergic receptors in perirhinal cortex in object recognition memory were compared. Rats' discrimination of a novel object preference test (NOP) test was measured after either systemic or local infusion into the perirhinal cortex of the nicotinic receptor antagonist methyllycaconitine (MLA), which targets alpha-7 (α7) amongst other nicotinic receptors or the muscarinic receptor antagonists scopolamine, AFDX-384, and pirenzepine. Methyllycaconitine administered systemically or intraperirhinally before acquisition impaired recognition memory tested after a 24-h, but not a 20-min delay. In contrast, all three muscarinic antagonists produced a similar, unusual pattern of impairment with amnesia after a 20-min delay, but remembrance after a 24-h delay. Thus, the amnesic effects of nicotinic and muscarinic antagonism were doubly dissociated across the 20-min and 24-h delays. The same pattern of shorter-term but not longer-term memory impairment was found for scopolamine whether the object preference test was carried out in a square arena or a Y-maze and whether rats of the Dark Agouti or Lister-hooded strains were used. Coinfusion of MLA and either scopolamine or AFDX-384 produced an impairment profile matching that for MLA. Hence, the antagonists did not act additively when coadministered. These findings establish an important role in recognition memory for both nicotinic and muscarinic cholinergic receptors in perirhinal cortex, and provide a challenge to simple ideas about the role of cholinergic processes in recognition memory: The effects of muscarinic and nicotinic antagonism are neither independent nor additive.

  19. Regulation of object recognition and object placement by ovarian sex steroid hormones

    PubMed Central

    Tuscher, Jennifer J.; Fortress, Ashley M.; Kim, Jaekyoon; Frick, Karyn M.

    2014-01-01

    The ovarian hormones 17β-estradiol (E2) and progesterone (P4) are potent modulators of hippocampal memory formation. Both hormones have been demonstrated to enhance hippocampal memory by regulating the cellular and molecular mechanisms thought to underlie memory formation. Behavioral neuroendocrinologists have increasingly used the object recognition and object placement (object location) tasks to investigate the role of E2 and P4 in regulating hippocampal memory formation in rodents. These one-trial learning tasks are ideal for studying acute effects of hormone treatments on different phases of memory because they can be administered during acquisition (pre-training), consolidation (post-training), or retrieval (pre-testing). This review synthesizes the rodent literature testing the effects of E2 and P4 on object recognition (OR) and object placement (OP), and the molecular mechanisms in the hippocampus supporting memory formation in these tasks. Some general trends emerge from the data. Among gonadally intact females, object memory tends to be best when E2 and P4 levels are elevated during the estrous cycle, pregnancy, and in middle age. In ovariectomized females, E2 given before or immediately after testing generally enhances OR and OP in young and middle-aged rats and mice, although effects are mixed in aged rodents. Effects of E2 treatment on OR 7and OP memory consolidation can be mediated by both classical estrogen receptors (ERα and ERβ), and depend on glutamate receptors (NMDA, mGluR1) and activation of numerous cell signaling cascades (e.g., ERK, PI3K/Akt, mTOR) and epigenetic processes (e.g., histone H3 acetylation, DNA methylation). Acute P4 treatment given immediately after training also enhances OR and OP in young and middle-aged ovariectomized females by activating similar cell signaling pathways as E2 (e.g., ERK, mTOR). The few studies that have administered both hormones in combination suggest that treatment can enhance OR and OP, but that

  20. Distinct roles of basal forebrain cholinergic neurons in spatial and object recognition memory.

    PubMed

    Okada, Kana; Nishizawa, Kayo; Kobayashi, Tomoko; Sakata, Shogo; Kobayashi, Kazuto

    2015-08-06

    Recognition memory requires processing of various types of information such as objects and locations. Impairment in recognition memory is a prominent feature of amnesia and a symptom of Alzheimer's disease (AD). Basal forebrain cholinergic neurons contain two major groups, one localized in the medial septum (MS)/vertical diagonal band of Broca (vDB), and the other in the nucleus basalis magnocellularis (NBM). The roles of these cell groups in recognition memory have been debated, and it remains unclear how they contribute to it. We use a genetic cell targeting technique to selectively eliminate cholinergic cell groups and then test spatial and object recognition memory through different behavioural tasks. Eliminating MS/vDB neurons impairs spatial but not object recognition memory in the reference and working memory tasks, whereas NBM elimination undermines only object recognition memory in the working memory task. These impairments are restored by treatment with acetylcholinesterase inhibitors, anti-dementia drugs for AD. Our results highlight that MS/vDB and NBM cholinergic neurons are not only implicated in recognition memory but also have essential roles in different types of recognition memory.

  1. On the three-quarter view advantage of familiar object recognition.

    PubMed

    Nonose, Kohei; Niimi, Ryosuke; Yokosawa, Kazuhiko

    2016-11-01

    A three-quarter view, i.e., an oblique view, of familiar objects often leads to a higher subjective goodness rating when compared with other orientations. What is the source of the high goodness for oblique views? First, we confirmed that object recognition performance was also best for oblique views around 30° view, even when the foreshortening disadvantage of front- and side-views was minimized (Experiments 1 and 2). In Experiment 3, we measured subjective ratings of view goodness and two possible determinants of view goodness: familiarity of view, and subjective impression of three-dimensionality. Three-dimensionality was measured as the subjective saliency of visual depth information. The oblique views were rated best, most familiar, and as approximating greatest three-dimensionality on average; however, the cluster analyses showed that the "best" orientation systematically varied among objects. We found three clusters of objects: front-preferred objects, oblique-preferred objects, and side-preferred objects. Interestingly, recognition performance and the three-dimensionality rating were higher for oblique views irrespective of the clusters. It appears that recognition efficiency is not the major source of the three-quarter view advantage. There are multiple determinants and variability among objects. This study suggests that the classical idea that a canonical view has a unique advantage in object perception requires further discussion.

  2. Superior voice recognition in a patient with acquired prosopagnosia and object agnosia.

    PubMed

    Hoover, Adria E N; Démonet, Jean-François; Steeves, Jennifer K E

    2010-11-01

    Anecdotally, it has been reported that individuals with acquired prosopagnosia compensate for their inability to recognize faces by using other person identity cues such as hair, gait or the voice. Are they therefore superior at the use of non-face cues, specifically voices, to person identity? Here, we empirically measure person and object identity recognition in a patient with acquired prosopagnosia and object agnosia. We quantify person identity (face and voice) and object identity (car and horn) recognition for visual, auditory, and bimodal (visual and auditory) stimuli. The patient is unable to recognize faces or cars, consistent with his prosopagnosia and object agnosia, respectively. He is perfectly able to recognize people's voices and car horns and bimodal stimuli. These data show a reverse shift in the typical weighting of visual over auditory information for audiovisual stimuli in a compromised visual recognition system. Moreover, the patient shows selectively superior voice recognition compared to the controls revealing that two different stimulus domains, persons and objects, may not be equally affected by sensory adaptation effects. This also implies that person and object identity recognition are processed in separate pathways. These data demonstrate that an individual with acquired prosopagnosia and object agnosia can compensate for the visual impairment and become quite skilled at using spared aspects of sensory processing. In the case of acquired prosopagnosia it is advantageous to develop a superior use of voices for person identity recognition in everyday life. Copyright © 2010 Elsevier Ltd. All rights reserved.

  3. Environmental enrichment improves novel object recognition and enhances agonistic behavior in male mice.

    PubMed

    Mesa-Gresa, Patricia; Pérez-Martinez, Asunción; Redolat, Rosa

    2013-01-01

    Environmental enrichment (EE) is an experimental paradigm in which rodents are housed in complex environments containing objects that provide stimulation, the effects of which are expected to improve the welfare of these subjects. EE has been shown to considerably improve learning and memory in rodents. However, knowledge about the effects of EE on social interaction is generally limited and rather controversial. Thus, our aim was to evaluate both novel object recognition and agonistic behavior in NMRI mice receiving EE, hypothesizing enhanced cognition and slightly enhanced agonistic interaction upon EE rearing. During a 4-week period half the mice (n = 16) were exposed to EE and the other half (n = 16) remained in a standard environment (SE). On PND 56-57, animals performed the object recognition test, in which recognition memory was measured using a discrimination index. The social interaction test consisted of an encounter between an experimental animal and a standard opponent. Results indicated that EE mice explored the new object for longer periods than SE animals (P < .05). During social encounters, EE mice devoted more time to sociability and agonistic behavior (P < .05) than their non-EE counterparts. In conclusion, EE has been shown to improve object recognition and increase agonistic behavior in adolescent/early adulthood mice. In the future we intend to extend this study on a longitudinal basis in order to assess in more depth the effect of EE and the consistency of the above-mentioned observations in NMRI mice. Copyright © 2013 Wiley Periodicals, Inc.

  4. 3-D Object Recognition from Point Cloud Data

    NASA Astrophysics Data System (ADS)

    Smith, W.; Walker, A. S.; Zhang, B.

    2011-09-01

    The market for real-time 3-D mapping includes not only traditional geospatial applications but also navigation of unmanned autonomous vehicles (UAVs). Massively parallel processes such as graphics processing unit (GPU) computing make real-time 3-D object recognition and mapping achievable. Geospatial technologies such as digital photogrammetry and GIS offer advanced capabilities to produce 2-D and 3-D static maps using UAV data. The goal is to develop real-time UAV navigation through increased automation. It is challenging for a computer to identify a 3-D object such as a car, a tree or a house, yet automatic 3-D object recognition is essential to increasing the productivity of geospatial data such as 3-D city site models. In the past three decades, researchers have used radiometric properties to identify objects in digital imagery with limited success, because these properties vary considerably from image to image. Consequently, our team has developed software that recognizes certain types of 3-D objects within 3-D point clouds. Although our software is developed for modeling, simulation and visualization, it has the potential to be valuable in robotics and UAV applications. The locations and shapes of 3-D objects such as buildings and trees are easily recognizable by a human from a brief glance at a representation of a point cloud such as terrain-shaded relief. The algorithms to extract these objects have been developed and require only the point cloud and minimal human inputs such as a set of limits on building size and a request to turn on a squaring option. The algorithms use both digital surface model (DSM) and digital elevation model (DEM), so software has also been developed to derive the latter from the former. The process continues through the following steps: identify and group 3-D object points into regions; separate buildings and houses from trees; trace region boundaries; regularize and simplify boundary polygons; construct complex roofs. Several case

  5. Visual Object Detection, Categorization, and Identification Tasks Are Associated with Different Time Courses and Sensitivities

    ERIC Educational Resources Information Center

    de la Rosa, Stephan; Choudhery, Rabia N.; Chatziastros, Astros

    2011-01-01

    Recent evidence suggests that the recognition of an object's presence and its explicit recognition are temporally closely related. Here we re-examined the time course (using a fine and a coarse temporal resolution) and the sensitivity of three possible component processes of visual object recognition. In particular, participants saw briefly…

  6. Object recognition and pose estimation of planar objects from range data

    NASA Technical Reports Server (NTRS)

    Pendleton, Thomas W.; Chien, Chiun Hong; Littlefield, Mark L.; Magee, Michael

    1994-01-01

    The Extravehicular Activity Helper/Retriever (EVAHR) is a robotic device currently under development at the NASA Johnson Space Center that is designed to fetch objects or to assist in retrieving an astronaut who may have become inadvertently de-tethered. The EVAHR will be required to exhibit a high degree of intelligent autonomous operation and will base much of its reasoning upon information obtained from one or more three-dimensional sensors that it will carry and control. At the highest level of visual cognition and reasoning, the EVAHR will be required to detect objects, recognize them, and estimate their spatial orientation and location. The recognition phase and estimation of spatial pose will depend on the ability of the vision system to reliably extract geometric features of the objects such as whether the surface topologies observed are planar or curved and the spatial relationships between the component surfaces. In order to achieve these tasks, three-dimensional sensing of the operational environment and objects in the environment will therefore be essential. One of the sensors being considered to provide image data for object recognition and pose estimation is a phase-shift laser scanner. The characteristics of the data provided by this scanner have been studied and algorithms have been developed for segmenting range images into planar surfaces, extracting basic features such as surface area, and recognizing the object based on the characteristics of extracted features. Also, an approach has been developed for estimating the spatial orientation and location of the recognized object based on orientations of extracted planes and their intersection points. This paper presents some of the algorithms that have been developed for the purpose of recognizing and estimating the pose of objects as viewed by the laser scanner, and characterizes the desirability and utility of these algorithms within the context of the scanner itself, considering data quality and

  7. The Functional Architecture of Visual Object Recognition

    DTIC Science & Technology

    1991-07-01

    different forms of agnosia can provide clues to the representations underlying normal object recognition (Farah, 1990). For example, the pair-wise...patterns of deficit and sparing occur. In a review of 99 published cases of agnosia , the observed patterns of co- occurrence implicated two underlying

  8. DIAC object recognition system

    NASA Astrophysics Data System (ADS)

    Buurman, Johannes

    1992-03-01

    This paper describes the object recognition system used in an intelligent robot cell. It is used to recognize and estimate pose and orientation of parts as they enter the cell. The parts are mostly metal and consist of polyhedral and cylindrical shapes. The system uses feature-based stereo vision to acquire a wireframe of the observed part. Features are defined as straight lines and ellipses, which lead to a wireframe of straight lines and circular arcs (the latter using a new algorithm). This wireframe is compared to a number of wire frame models obtained from the CAD database. Experimental results show that image processing hardware and parallelization may add considerably to the speed of the system.

  9. Parts and Relations in Young Children's Shape-Based Object Recognition

    ERIC Educational Resources Information Center

    Augustine, Elaine; Smith, Linda B.; Jones, Susan S.

    2011-01-01

    The ability to recognize common objects from sparse information about geometric shape emerges during the same period in which children learn object names and object categories. Hummel and Biederman's (1992) theory of object recognition proposes that the geometric shapes of objects have two components--geometric volumes representing major object…

  10. Combining color and shape information for illumination-viewpoint invariant object recognition.

    PubMed

    Diplaros, Aristeidis; Gevers, Theo; Patras, Ioannis

    2006-01-01

    In this paper, we propose a new scheme that merges color- and shape-invariant information for object recognition. To obtain robustness against photometric changes, color-invariant derivatives are computed first. Color invariance is an important aspect of any object recognition scheme, as color changes considerably with the variation in illumination, object pose, and camera viewpoint. These color invariant derivatives are then used to obtain similarity invariant shape descriptors. Shape invariance is equally important as, under a change in camera viewpoint and object pose, the shape of a rigid object undergoes a perspective projection on the image plane. Then, the color and shape invariants are combined in a multidimensional color-shape context which is subsequently used as an index. As the indexing scheme makes use of a color-shape invariant context, it provides a high-discriminative information cue robust against varying imaging conditions. The matching function of the color-shape context allows for fast recognition, even in the presence of object occlusion and cluttering. From the experimental results, it is shown that the method recognizes rigid objects with high accuracy in 3-D complex scenes and is robust against changing illumination, camera viewpoint, object pose, and noise.

  11. Object recognition with severe spatial deficits in Williams syndrome: sparing and breakdown.

    PubMed

    Landau, Barbara; Hoffman, James E; Kurz, Nicole

    2006-07-01

    Williams syndrome (WS) is a rare genetic disorder that results in severe visual-spatial cognitive deficits coupled with relative sparing in language, face recognition, and certain aspects of motion processing. Here, we look for evidence for sparing or impairment in another cognitive system-object recognition. Children with WS, normal mental-age (MA) and chronological age-matched (CA) children, and normal adults viewed pictures of a large range of objects briefly presented under various conditions of degradation, including canonical and unusual orientations, and clear or blurred contours. Objects were shown as either full-color views (Experiment 1) or line drawings (Experiment 2). Across both experiments, WS and MA children performed similarly in all conditions while CA children performed better than both WS group and MA groups with unusual views. This advantage, however, was eliminated when images were also blurred. The error types and relative difficulty of different objects were similar across all participant groups. The results indicate selective sparing of basic mechanisms of object recognition in WS, together with developmental delay or arrest in recognition of objects from unusual viewpoints. These findings are consistent with the growing literature on brain abnormalities in WS which points to selective impairment in the parietal areas of the brain. As a whole, the results lend further support to the growing literature on the functional separability of object recognition mechanisms from other spatial functions, and raise intriguing questions about the link between genetic deficits and cognition.

  12. Spatiotemporal information during unsupervised learning enhances viewpoint invariant object recognition

    PubMed Central

    Tian, Moqian; Grill-Spector, Kalanit

    2015-01-01

    Recognizing objects is difficult because it requires both linking views of an object that can be different and distinguishing objects with similar appearance. Interestingly, people can learn to recognize objects across views in an unsupervised way, without feedback, just from the natural viewing statistics. However, there is intense debate regarding what information during unsupervised learning is used to link among object views. Specifically, researchers argue whether temporal proximity, motion, or spatiotemporal continuity among object views during unsupervised learning is beneficial. Here, we untangled the role of each of these factors in unsupervised learning of novel three-dimensional (3-D) objects. We found that after unsupervised training with 24 object views spanning a 180° view space, participants showed significant improvement in their ability to recognize 3-D objects across rotation. Surprisingly, there was no advantage to unsupervised learning with spatiotemporal continuity or motion information than training with temporal proximity. However, we discovered that when participants were trained with just a third of the views spanning the same view space, unsupervised learning via spatiotemporal continuity yielded significantly better recognition performance on novel views than learning via temporal proximity. These results suggest that while it is possible to obtain view-invariant recognition just from observing many views of an object presented in temporal proximity, spatiotemporal information enhances performance by producing representations with broader view tuning than learning via temporal association. Our findings have important implications for theories of object recognition and for the development of computational algorithms that learn from examples. PMID:26024454

  13. Post-Training Reversible Inactivation of the Hippocampus Enhances Novel Object Recognition Memory

    ERIC Educational Resources Information Center

    Oliveira, Ana M. M.; Hawk, Joshua D.; Abel, Ted; Havekes, Robbert

    2010-01-01

    Research on the role of the hippocampus in object recognition memory has produced conflicting results. Previous studies have used permanent hippocampal lesions to assess the requirement for the hippocampus in the object recognition task. However, permanent hippocampal lesions may impact performance through effects on processes besides memory…

  14. Invariant visual object recognition and shape processing in rats

    PubMed Central

    Zoccolan, Davide

    2015-01-01

    Invariant visual object recognition is the ability to recognize visual objects despite the vastly different images that each object can project onto the retina during natural vision, depending on its position and size within the visual field, its orientation relative to the viewer, etc. Achieving invariant recognition represents such a formidable computational challenge that is often assumed to be a unique hallmark of primate vision. Historically, this has limited the invasive investigation of its neuronal underpinnings to monkey studies, in spite of the narrow range of experimental approaches that these animal models allow. Meanwhile, rodents have been largely neglected as models of object vision, because of the widespread belief that they are incapable of advanced visual processing. However, the powerful array of experimental tools that have been developed to dissect neuronal circuits in rodents has made these species very attractive to vision scientists too, promoting a new tide of studies that have started to systematically explore visual functions in rats and mice. Rats, in particular, have been the subjects of several behavioral studies, aimed at assessing how advanced object recognition and shape processing is in this species. Here, I review these recent investigations, as well as earlier studies of rat pattern vision, to provide an historical overview and a critical summary of the status of the knowledge about rat object vision. The picture emerging from this survey is very encouraging with regard to the possibility of using rats as complementary models to monkeys in the study of higher-level vision. PMID:25561421

  15. Object recognition contributions to figure-ground organization: operations on outlines and subjective contours.

    PubMed

    Peterson, M A; Gibson, B S

    1994-11-01

    In previous research, replicated here, we found that some object recognition processes influence figure-ground organization. We have proposed that these object recognition processes operate on edges (or contours) detected early in visual processing, rather than on regions. Consistent with this proposal, influences from object recognition on figure-ground organization were previously observed in both pictures and stereograms depicting regions of different luminance, but not in random-dot stereograms, where edges arise late in processing (Peterson & Gibson, 1993). In the present experiments, we examined whether or not two other types of contours--outlines and subjective contours--enable object recognition influences on figure-ground organization. For both types of contours we observed a pattern of effects similar to that originally obtained with luminance edges. The results of these experiments are valuable for distinguishing between alternative views of the mechanisms mediating object recognition influences on figure-ground organization. In addition, in both Experiments 1 and 2, fixated regions were seen as figure longer than nonfixated regions, suggesting that fixation location must be included among the variables relevant to figure-ground organization.

  16. Progestogens’ effects and mechanisms for object recognition memory across the lifespan

    PubMed Central

    Walf, Alicia A.; Koonce, Carolyn J.; Frye, Cheryl A.

    2016-01-01

    This review explores the effects of female reproductive hormones, estrogens and progestogens, with a focus on progesterone and allopregnanolone, on object memory. Progesterone and its metabolites, in particular allopregnanolone, exert various effects on both cognitive and non-mnemonic functions in females. The well-known object recognition task is a valuable experimental paradigm that can be used to determine the effects and mechanisms of progestogens for mnemonic effects across the lifespan, which will be discussed herein. In this task there is little test-decay when different objects are used as targets and baseline valance for objects is controlled. This allows repeated testing, within-subjects designs, and longitudinal assessments, which aid understanding of changes in hormonal milieu. Objects are not aversive or food-based, which are hormone-sensitive factors. This review focuses on published data from our laboratory, and others, using the object recognition task in rodents to assess the role and mechanisms of progestogens throughout the lifespan. Improvements in object recognition performance of rodents are often associated with higher hormone levels in the hippocampus and prefrontal cortex during natural cycles, with hormone replacement following ovariectomy in young animals, or with aging. The capacity for reversal of age- and reproductive senescence-related decline in cognitive performance, and changes in neural plasticity that may be dissociated from peripheral effects with such decline, are discussed. The focus here will be on the effects of brain-derived factors, such as the neurosteroid, allopregnanolone, and other hormones, for enhancing object recognition across the lifespan. PMID:26235328

  17. A new method of edge detection for object recognition

    USGS Publications Warehouse

    Maddox, Brian G.; Rhew, Benjamin

    2004-01-01

    Traditional edge detection systems function by returning every edge in an input image. This can result in a large amount of clutter and make certain vectorization algorithms less accurate. Accuracy problems can then have a large impact on automated object recognition systems that depend on edge information. A new method of directed edge detection can be used to limit the number of edges returned based on a particular feature. This results in a cleaner image that is easier for vectorization. Vectorized edges from this process could then feed an object recognition system where the edge data would also contain information as to what type of feature it bordered.

  18. Figure-ground organization and object recognition processes: an interactive account.

    PubMed

    Vecera, S P; O'Reilly, R C

    1998-04-01

    Traditional bottom-up models of visual processing assume that figure-ground organization precedes object recognition. This assumption seems logically necessary: How can object recognition occur before a region is labeled as figure? However, some behavioral studies find that familiar regions are more likely to be labeled figure than less familiar regions, a problematic finding for bottom-up models. An interactive account is proposed in which figure-ground processes receive top-down input from object representations in a hierarchical system. A graded, interactive computational model is presented that accounts for behavioral results in which familiarity effects are found. The interactive model offers an alternative conception of visual processing to bottom-up models.

  19. Modeling guidance and recognition in categorical search: bridging human and computer object detection.

    PubMed

    Zelinsky, Gregory J; Peng, Yifan; Berg, Alexander C; Samaras, Dimitris

    2013-10-08

    Search is commonly described as a repeating cycle of guidance to target-like objects, followed by the recognition of these objects as targets or distractors. Are these indeed separate processes using different visual features? We addressed this question by comparing observer behavior to that of support vector machine (SVM) models trained on guidance and recognition tasks. Observers searched for a categorically defined teddy bear target in four-object arrays. Target-absent trials consisted of random category distractors rated in their visual similarity to teddy bears. Guidance, quantified as first-fixated objects during search, was strongest for targets, followed by target-similar, medium-similarity, and target-dissimilar distractors. False positive errors to first-fixated distractors also decreased with increasing dissimilarity to the target category. To model guidance, nine teddy bear detectors, using features ranging in biological plausibility, were trained on unblurred bears then tested on blurred versions of the same objects appearing in each search display. Guidance estimates were based on target probabilities obtained from these detectors. To model recognition, nine bear/nonbear classifiers, trained and tested on unblurred objects, were used to classify the object that would be fixated first (based on the detector estimates) as a teddy bear or a distractor. Patterns of categorical guidance and recognition accuracy were modeled almost perfectly by an HMAX model in combination with a color histogram feature. We conclude that guidance and recognition in the context of search are not separate processes mediated by different features, and that what the literature knows as guidance is really recognition performed on blurred objects viewed in the visual periphery.

  20. Ball-scale based hierarchical multi-object recognition in 3D medical images

    NASA Astrophysics Data System (ADS)

    Bağci, Ulas; Udupa, Jayaram K.; Chen, Xinjian

    2010-03-01

    This paper investigates, using prior shape models and the concept of ball scale (b-scale), ways of automatically recognizing objects in 3D images without performing elaborate searches or optimization. That is, the goal is to place the model in a single shot close to the right pose (position, orientation, and scale) in a given image so that the model boundaries fall in the close vicinity of object boundaries in the image. This is achieved via the following set of key ideas: (a) A semi-automatic way of constructing a multi-object shape model assembly. (b) A novel strategy of encoding, via b-scale, the pose relationship between objects in the training images and their intensity patterns captured in b-scale images. (c) A hierarchical mechanism of positioning the model, in a one-shot way, in a given image from a knowledge of the learnt pose relationship and the b-scale image of the given image to be segmented. The evaluation results on a set of 20 routine clinical abdominal female and male CT data sets indicate the following: (1) Incorporating a large number of objects improves the recognition accuracy dramatically. (2) The recognition algorithm can be thought as a hierarchical framework such that quick replacement of the model assembly is defined as coarse recognition and delineation itself is known as finest recognition. (3) Scale yields useful information about the relationship between the model assembly and any given image such that the recognition results in a placement of the model close to the actual pose without doing any elaborate searches or optimization. (4) Effective object recognition can make delineation most accurate.

  1. Timing, timing, timing: Fast decoding of object information from intracranial field potentials in human visual cortex

    PubMed Central

    Liu, Hesheng; Agam, Yigal; Madsen, Joseph R.; Kreiman, Gabriel

    2010-01-01

    Summary The difficulty of visual recognition stems from the need to achieve high selectivity while maintaining robustness to object transformations within hundreds of milliseconds. Theories of visual recognition differ in whether the neuronal circuits invoke recurrent feedback connections or not. The timing of neurophysiological responses in visual cortex plays a key role in distinguishing between bottom-up and top-down theories. Here we quantified at millisecond resolution the amount of visual information conveyed by intracranial field potentials from 912 electrodes in 11 human subjects. We could decode object category information from human visual cortex in single trials as early as 100 ms post-stimulus. Decoding performance was robust to depth rotation and scale changes. The results suggest that physiological activity in the temporal lobe can account for key properties of visual recognition. The fast decoding in single trials is compatible with feed-forward theories and provides strong constraints for computational models of human vision. PMID:19409272

  2. Knowledge-based object recognition for different morphological classes of plants

    NASA Astrophysics Data System (ADS)

    Brendel, Thorsten; Schwanke, Joerg; Jensch, Peter F.; Megnet, Roland

    1995-01-01

    Micropropagation of plants is done by cutting juvenile plants and placing them into special container-boxes with nutrient-solution where the pieces can grow up and be cut again several times. To produce high amounts of biomass it is necessary to do plant micropropagation by a robotic syshoot. In this paper we describe parts of the vision syshoot that recognizes plants and their particular cutting points. Therefore, it is necessary to extract elements of the plants and relations between these elements (for example root, shoot, leaf). Different species vary in their morphological appearance, variation is also immanent in plants of the same species. Therefore, we introduce several morphological classes of plants from that we expect same recognition methods. As a result of our work we present rules which help users to create specific algorithms for object recognition of plant species.

  3. An Effective 3D Shape Descriptor for Object Recognition with RGB-D Sensors

    PubMed Central

    Liu, Zhong; Zhao, Changchen; Wu, Xingming; Chen, Weihai

    2017-01-01

    RGB-D sensors have been widely used in various areas of computer vision and graphics. A good descriptor will effectively improve the performance of operation. This article further analyzes the recognition performance of shape features extracted from multi-modality source data using RGB-D sensors. A hybrid shape descriptor is proposed as a representation of objects for recognition. We first extracted five 2D shape features from contour-based images and five 3D shape features over point cloud data to capture the global and local shape characteristics of an object. The recognition performance was tested for category recognition and instance recognition. Experimental results show that the proposed shape descriptor outperforms several common global-to-global shape descriptors and is comparable to some partial-to-global shape descriptors that achieved the best accuracies in category and instance recognition. Contribution of partial features and computational complexity were also analyzed. The results indicate that the proposed shape features are strong cues for object recognition and can be combined with other features to boost accuracy. PMID:28245553

  4. The implementation of aerial object recognition algorithm based on contour descriptor in FPGA-based on-board vision system

    NASA Astrophysics Data System (ADS)

    Babayan, Pavel; Smirnov, Sergey; Strotov, Valery

    2017-10-01

    This paper describes the aerial object recognition algorithm for on-board and stationary vision system. Suggested algorithm is intended to recognize the objects of a specific kind using the set of the reference objects defined by 3D models. The proposed algorithm based on the outer contour descriptor building. The algorithm consists of two stages: learning and recognition. Learning stage is devoted to the exploring of reference objects. Using 3D models we can build the database containing training images by rendering the 3D model from viewpoints evenly distributed on a sphere. Sphere points distribution is made by the geosphere principle. Gathered training image set is used for calculating descriptors, which will be used in the recognition stage of the algorithm. The recognition stage is focusing on estimating the similarity of the captured object and the reference objects by matching an observed image descriptor and the reference object descriptors. The experimental research was performed using a set of the models of the aircraft of the different types (airplanes, helicopters, UAVs). The proposed orientation estimation algorithm showed good accuracy in all case studies. The real-time performance of the algorithm in FPGA-based vision system was demonstrated.

  5. Size-Sensitive Perceptual Representations Underlie Visual and Haptic Object Recognition

    PubMed Central

    Craddock, Matt; Lawson, Rebecca

    2009-01-01

    A variety of similarities between visual and haptic object recognition suggests that the two modalities may share common representations. However, it is unclear whether such common representations preserve low-level perceptual features or whether transfer between vision and haptics is mediated by high-level, abstract representations. Two experiments used a sequential shape-matching task to examine the effects of size changes on unimodal and crossmodal visual and haptic object recognition. Participants felt or saw 3D plastic models of familiar objects. The two objects presented on a trial were either the same size or different sizes and were the same shape or different but similar shapes. Participants were told to ignore size changes and to match on shape alone. In Experiment 1, size changes on same-shape trials impaired performance similarly for both visual-to-visual and haptic-to-haptic shape matching. In Experiment 2, size changes impaired performance on both visual-to-haptic and haptic-to-visual shape matching and there was no interaction between the cost of size changes and direction of transfer. Together the unimodal and crossmodal matching results suggest that the same, size-specific perceptual representations underlie both visual and haptic object recognition, and indicate that crossmodal memory for objects must be at least partly based on common perceptual representations. PMID:19956685

  6. Orientation estimation of anatomical structures in medical images for object recognition

    NASA Astrophysics Data System (ADS)

    Bağci, Ulaş; Udupa, Jayaram K.; Chen, Xinjian

    2011-03-01

    Recognition of anatomical structures is an important step in model based medical image segmentation. It provides pose estimation of objects and information about "where" roughly the objects are in the image and distinguishing them from other object-like entities. In,1 we presented a general method of model-based multi-object recognition to assist in segmentation (delineation) tasks. It exploits the pose relationship that can be encoded, via the concept of ball scale (b-scale), between the binary training objects and their associated grey images. The goal was to place the model, in a single shot, close to the right pose (position, orientation, and scale) in a given image so that the model boundaries fall in the close vicinity of object boundaries in the image. Unlike position and scale parameters, we observe that orientation parameters require more attention when estimating the pose of the model as even small differences in orientation parameters can lead to inappropriate recognition. Motivated from the non-Euclidean nature of the pose information, we propose in this paper the use of non-Euclidean metrics to estimate orientation of the anatomical structures for more accurate recognition and segmentation. We statistically analyze and evaluate the following metrics for orientation estimation: Euclidean, Log-Euclidean, Root-Euclidean, Procrustes Size-and-Shape, and mean Hermitian metrics. The results show that mean Hermitian and Cholesky decomposition metrics provide more accurate orientation estimates than other Euclidean and non-Euclidean metrics.

  7. Cultural differences in visual object recognition in 3-year-old children

    PubMed Central

    Kuwabara, Megumi; Smith, Linda B.

    2016-01-01

    Recent research indicates that culture penetrates fundamental processes of perception and cognition (e.g. Nisbett & Miyamoto, 2005). Here, we provide evidence that these influences begin early and influence how preschool children recognize common objects. The three tasks (n=128) examined the degree to which nonface object recognition by 3 year olds was based on individual diagnostic features versus more configural and holistic processing. Task 1 used a 6-alternative forced choice task in which children were asked to find a named category in arrays of masked objects in which only 3 diagnostic features were visible for each object. U.S. children outperformed age-matched Japanese children. Task 2 presented pictures of objects to children piece by piece. U.S. children recognized the objects given fewer pieces than Japanese children and likelihood of recognition increased for U.S., but not Japanese children when the piece added was rated by both U.S. and Japanese adults as highly defining. Task 3 used a standard measure of configural progressing, asking the degree to which recognition of matching pictures was disrupted by the rotation of one picture. Japanese children’s recognition was more disrupted by inversion than was that of U.S. children, indicating more configural processing by Japanese than U.S. children. The pattern suggests early cross-cultural differences in visual processing; findings that raise important questions about how visual experiences differ across cultures and about universal patterns of cognitive development. PMID:26985576

  8. Cultural differences in visual object recognition in 3-year-old children.

    PubMed

    Kuwabara, Megumi; Smith, Linda B

    2016-07-01

    Recent research indicates that culture penetrates fundamental processes of perception and cognition. Here, we provide evidence that these influences begin early and influence how preschool children recognize common objects. The three tasks (N=128) examined the degree to which nonface object recognition by 3-year-olds was based on individual diagnostic features versus more configural and holistic processing. Task 1 used a 6-alternative forced choice task in which children were asked to find a named category in arrays of masked objects where only three diagnostic features were visible for each object. U.S. children outperformed age-matched Japanese children. Task 2 presented pictures of objects to children piece by piece. U.S. children recognized the objects given fewer pieces than Japanese children, and the likelihood of recognition increased for U.S. children, but not Japanese children, when the piece added was rated by both U.S. and Japanese adults as highly defining. Task 3 used a standard measure of configural progressing, asking the degree to which recognition of matching pictures was disrupted by the rotation of one picture. Japanese children's recognition was more disrupted by inversion than was that of U.S. children, indicating more configural processing by Japanese than U.S. children. The pattern suggests early cross-cultural differences in visual processing; findings that raise important questions about how visual experiences differ across cultures and about universal patterns of cognitive development. Copyright © 2016 Elsevier Inc. All rights reserved.

  9. Leveraging Cognitive Context for Object Recognition

    DTIC Science & Technology

    2014-06-01

    learned from large image databases. We build upon this concept by exploring cognitive context, demonstrating how rich dynamic context provided by...context that people rely upon as they perceive the world. Context in ACT-R/E takes the form of associations between related concepts that are learned ...and accuracy of object recognition. Context is most often viewed as a static concept, learned from large image databases. We build upon this concept by

  10. Ultra-FDst Object Recognition from Few Spikes

    DTIC Science & Technology

    2005-07-01

    Ultra-fast Object Recognition from Few Spikes Chou Hung, Gabriel Kreiman , Tomaso Poggio & James J. DiCarlo AI Memo 2005-022 July 2005 CBCL Memo 253...authors, Chou Hung and Gabriel Kreiman , contributed equally to this work. Supplementary Material is available at http://ramonycajal.mit.edu... kreiman /resources/ultrafast/. _____________________________________________________________________________ This report describes research done at

  11. Reversibility of object recognition but not spatial memory impairment following binge-like alcohol exposure in rats

    PubMed Central

    Cippitelli, Andrea; Zook, Michelle; Bell, Lauren; Damadzic, Ruslan; Eskay, Robert L.; Schwandt, Melanie; Heilig, Markus

    2010-01-01

    Excessive alcohol use leads to neurodegeneration in several brain structures including the hippocampal dentate gyrus and the entorhinal cortex. Cognitive deficits that result are among the most insidious and debilitating consequences of alcoholism. The object exploration task (OET) provides a sensitive measurement of spatial memory impairment induced by hippocampal and cortical damage. In this study, we examine whether the observed neurotoxicity produced by a 4-day binge ethanol treatment results in long-term memory impairment by observing the time course of reactions to spatial change (object configuration) and non-spatial change (object recognition). Wistar rats were assessed for their abilities to detect spatial configuration in the OET at 1 week and 10 weeks following the ethanol treatment, in which ethanol groups received 9–15 g/kg/day and achieved blood alcohol levels over 300 mg/dl. At 1 week, results indicated that the binge alcohol treatment produced impairment in both spatial memory and non-spatial object recognition performance. Unlike the controls, ethanol treated rats did not increase the duration or number of contacts with the displaced object in the spatial memory task, nor did they increase the duration of contacts with the novel object in the object recognition task. After 10 weeks, spatial memory remained impaired in the ethanol treated rats but object recognition ability was recovered. Our data suggest that episodes of binge-like alcohol exposure result in long-term and possibly permanent impairments in memory for the configuration of objects during exploration, whereas the ability to detect non-spatial changes is only temporarily affected. PMID:20849966

  12. Augmented reality three-dimensional object visualization and recognition with axially distributed sensing.

    PubMed

    Markman, Adam; Shen, Xin; Hua, Hong; Javidi, Bahram

    2016-01-15

    An augmented reality (AR) smartglass display combines real-world scenes with digital information enabling the rapid growth of AR-based applications. We present an augmented reality-based approach for three-dimensional (3D) optical visualization and object recognition using axially distributed sensing (ADS). For object recognition, the 3D scene is reconstructed, and feature extraction is performed by calculating the histogram of oriented gradients (HOG) of a sliding window. A support vector machine (SVM) is then used for classification. Once an object has been identified, the 3D reconstructed scene with the detected object is optically displayed in the smartglasses allowing the user to see the object, remove partial occlusions of the object, and provide critical information about the object such as 3D coordinates, which are not possible with conventional AR devices. To the best of our knowledge, this is the first report on combining axially distributed sensing with 3D object visualization and recognition for applications to augmented reality. The proposed approach can have benefits for many applications, including medical, military, transportation, and manufacturing.

  13. Implicit and Explicit Contributions to Object Recognition: Evidence from Rapid Perceptual Learning

    PubMed Central

    Hassler, Uwe; Friese, Uwe; Gruber, Thomas

    2012-01-01

    The present study investigated implicit and explicit recognition processes of rapidly perceptually learned objects by means of steady-state visual evoked potentials (SSVEP). Participants were initially exposed to object pictures within an incidental learning task (living/non-living categorization). Subsequently, degraded versions of some of these learned pictures were presented together with degraded versions of unlearned pictures and participants had to judge, whether they recognized an object or not. During this test phase, stimuli were presented at 15 Hz eliciting an SSVEP at the same frequency. Source localizations of SSVEP effects revealed for implicit and explicit processes overlapping activations in orbito-frontal and temporal regions. Correlates of explicit object recognition were additionally found in the superior parietal lobe. These findings are discussed to reflect facilitation of object-specific processing areas within the temporal lobe by an orbito-frontal top-down signal as proposed by bi-directional accounts of object recognition. PMID:23056558

  14. Modeling guidance and recognition in categorical search: Bridging human and computer object detection

    PubMed Central

    Zelinsky, Gregory J.; Peng, Yifan; Berg, Alexander C.; Samaras, Dimitris

    2013-01-01

    Search is commonly described as a repeating cycle of guidance to target-like objects, followed by the recognition of these objects as targets or distractors. Are these indeed separate processes using different visual features? We addressed this question by comparing observer behavior to that of support vector machine (SVM) models trained on guidance and recognition tasks. Observers searched for a categorically defined teddy bear target in four-object arrays. Target-absent trials consisted of random category distractors rated in their visual similarity to teddy bears. Guidance, quantified as first-fixated objects during search, was strongest for targets, followed by target-similar, medium-similarity, and target-dissimilar distractors. False positive errors to first-fixated distractors also decreased with increasing dissimilarity to the target category. To model guidance, nine teddy bear detectors, using features ranging in biological plausibility, were trained on unblurred bears then tested on blurred versions of the same objects appearing in each search display. Guidance estimates were based on target probabilities obtained from these detectors. To model recognition, nine bear/nonbear classifiers, trained and tested on unblurred objects, were used to classify the object that would be fixated first (based on the detector estimates) as a teddy bear or a distractor. Patterns of categorical guidance and recognition accuracy were modeled almost perfectly by an HMAX model in combination with a color histogram feature. We conclude that guidance and recognition in the context of search are not separate processes mediated by different features, and that what the literature knows as guidance is really recognition performed on blurred objects viewed in the visual periphery. PMID:24105460

  15. Estrous cycle, pregnancy, and parity enhance performance of rats in object recognition or object placement tasks

    PubMed Central

    Paris, Jason J; Frye, Cheryl A

    2008-01-01

    Ovarian hormone elevations are associated with enhanced learning/memory. During behavioral estrus or pregnancy, progestins, such as progesterone (P4) and its metabolite 5α-pregnan-3α-ol-20-one (3α,5α-THP), are elevated due, in part, to corpora luteal and placental secretion. During ‘pseudopregnancy’, the induction of corpora luteal functioning results in a hormonal milieu analogous to pregnancy, which ceases after about 12 days, due to the lack of placental formation. Multiparity is also associated with enhanced learning/memory, perhaps due to prior steroid exposure during pregnancy. Given evidence that progestins and/or parity may influence cognition, we investigated how natural alterations in the progestin milieu influence cognitive performance. In Experiment 1, virgin rats (nulliparous) or rats with two prior pregnancies (multiparous) were assessed on the object placement and recognition tasks, when in high-estrogen/P4 (behavioral estrus) or low-estrogen/P4 (diestrus) phases of the estrous cycle. In Experiment 2, primiparous or multiparous rats were tested in the object placement and recognition tasks when not pregnant, pseudopregnant, or pregnant (between gestational days (GDs) 6 and 12). In Experiment 3, pregnant primiparous or multiparous rats were assessed daily in the object placement or recognition tasks. Females in natural states associated with higher endogenous progestins (behavioral estrus, pregnancy, multiparity) outperformed rats in low progestin states (diestrus, non-pregnancy, nulliparity) on the object placement and recognition tasks. In earlier pregnancy, multiparous, compared with primiparous, rats had a lower corticosterone, but higher estrogen levels, concomitant with better object placement performance. From GD 13 until post partum, primiparous rats had higher 3α,5α-THP levels and improved object placement performance compared with multiparous rats. PMID:18390689

  16. Neural Substrates of View-Invariant Object Recognition Developed without Experiencing Rotations of the Objects

    PubMed Central

    Okamura, Jun-ya; Yamaguchi, Reona; Honda, Kazunari; Tanaka, Keiji

    2014-01-01

    One fails to recognize an unfamiliar object across changes in viewing angle when it must be discriminated from similar distractor objects. View-invariant recognition gradually develops as the viewer repeatedly sees the objects in rotation. It is assumed that different views of each object are associated with one another while their successive appearance is experienced in rotation. However, natural experience of objects also contains ample opportunities to discriminate among objects at each of the multiple viewing angles. Our previous behavioral experiments showed that after experiencing a new set of object stimuli during a task that required only discrimination at each of four viewing angles at 30° intervals, monkeys could recognize the objects across changes in viewing angle up to 60°. By recording activities of neurons from the inferotemporal cortex after various types of preparatory experience, we here found a possible neural substrate for the monkeys' performance. For object sets that the monkeys had experienced during the task that required only discrimination at each of four viewing angles, many inferotemporal neurons showed object selectivity covering multiple views. The degree of view generalization found for these object sets was similar to that found for stimulus sets with which the monkeys had been trained to conduct view-invariant recognition. These results suggest that the experience of discriminating new objects in each of several viewing angles develops the partially view-generalized object selectivity distributed over many neurons in the inferotemporal cortex, which in turn bases the monkeys' emergent capability to discriminate the objects across changes in viewing angle. PMID:25378169

  17. View-invariant object recognition ability develops after discrimination, not mere exposure, at several viewing angles.

    PubMed

    Yamashita, Wakayo; Wang, Gang; Tanaka, Keiji

    2010-01-01

    One usually fails to recognize an unfamiliar object across changes in viewing angle when it has to be discriminated from similar distractor objects. Previous work has demonstrated that after long-term experience in discriminating among a set of objects seen from the same viewing angle, immediate recognition of the objects across 30-60 degrees changes in viewing angle becomes possible. The capability for view-invariant object recognition should develop during the within-viewing-angle discrimination, which includes two kinds of experience: seeing individual views and discriminating among the objects. The aim of the present study was to determine the relative contribution of each factor to the development of view-invariant object recognition capability. Monkeys were first extensively trained in a task that required view-invariant object recognition (Object task) with several sets of objects. The animals were then exposed to a new set of objects over 26 days in one of two preparatory tasks: one in which each object view was seen individually, and a second that required discrimination among the objects at each of four viewing angles. After the preparatory period, we measured the monkeys' ability to recognize the objects across changes in viewing angle, by introducing the object set to the Object task. Results indicated significant view-invariant recognition after the second but not first preparatory task. These results suggest that discrimination of objects from distractors at each of several viewing angles is required for the development of view-invariant recognition of the objects when the distractors are similar to the objects.

  18. Neuropeptide Trefoil factor 3 improves learning and retention of novel object recognition memory in mice.

    PubMed

    Shi, Hai-Shui; Yin, Xi; Song, Li; Guo, Qing-Jun; Luo, Xiang-Heng

    2012-02-01

    Accumulating evidence has implicated neuropeptides in modulating recognition, learning and memory. However, to date, no study has investigated the effects of neuropeptide Trefoil factor 3 (TFF3) on the process of learning and memory. In the present study, we evaluated the acute effects of TFF3 administration (0.1 and 0.5mg/kg, i.p.) on the acquisition and retention of object recognition memory in mice. We found that TFF3 administration significantly enhanced both short-term and long-term memory during the retention test, conducted 90 min and 24h after training respectively. Remarkably, acute TFF3 administration transformed a learning event that would not normally result in long-term memory into an event retained for a long-term period and produced no effect on locomotor activity in mice. In conclusion, the present results provide an important role of TFF3 in improving object recognition memory and reserving it for a longer time, which suggests a potential therapeutic application for diseases with recognition and memory impairment. Copyright © 2011 Elsevier B.V. All rights reserved.

  19. Ignorance- versus evidence-based decision making: a decision time analysis of the recognition heuristic.

    PubMed

    Hilbig, Benjamin E; Pohl, Rüdiger F

    2009-09-01

    According to part of the adaptive toolbox notion of decision making known as the recognition heuristic (RH), the decision process in comparative judgments-and its duration-is determined by whether recognition discriminates between objects. By contrast, some recently proposed alternative models predict that choices largely depend on the amount of evidence speaking for each of the objects and that decision times thus depend on the evidential difference between objects, or the degree of conflict between options. This article presents 3 experiments that tested predictions derived from the RH against those from alternative models. All experiments used naturally recognized objects without teaching participants any information and thus provided optimal conditions for application of the RH. However, results supported the alternative, evidence-based models and often conflicted with the RH. Recognition was not the key determinant of decision times, whereas differences between objects with respect to (both positive and negative) evidence predicted effects well. In sum, alternative models that allow for the integration of different pieces of information may well provide a better account of comparative judgments. (c) 2009 APA, all rights reserved.

  20. Object similarity affects the perceptual strategy underlying invariant visual object recognition in rats

    PubMed Central

    Rosselli, Federica B.; Alemi, Alireza; Ansuini, Alessio; Zoccolan, Davide

    2015-01-01

    In recent years, a number of studies have explored the possible use of rats as models of high-level visual functions. One central question at the root of such an investigation is to understand whether rat object vision relies on the processing of visual shape features or, rather, on lower-order image properties (e.g., overall brightness). In a recent study, we have shown that rats are capable of extracting multiple features of an object that are diagnostic of its identity, at least when those features are, structure-wise, distinct enough to be parsed by the rat visual system. In the present study, we have assessed the impact of object structure on rat perceptual strategy. We trained rats to discriminate between two structurally similar objects, and compared their recognition strategies with those reported in our previous study. We found that, under conditions of lower stimulus discriminability, rat visual discrimination strategy becomes more view-dependent and subject-dependent. Rats were still able to recognize the target objects, in a way that was largely tolerant (i.e., invariant) to object transformation; however, the larger structural and pixel-wise similarity affected the way objects were processed. Compared to the findings of our previous study, the patterns of diagnostic features were: (i) smaller and more scattered; (ii) only partially preserved across object views; and (iii) only partially reproducible across rats. On the other hand, rats were still found to adopt a multi-featural processing strategy and to make use of part of the optimal discriminatory information afforded by the two objects. Our findings suggest that, as in humans, rat invariant recognition can flexibly rely on either view-invariant representations of distinctive object features or view-specific object representations, acquired through learning. PMID:25814936

  1. Global precedence effects account for individual differences in both face and object recognition performance.

    PubMed

    Gerlach, Christian; Starrfelt, Randi

    2018-03-20

    There has been an increase in studies adopting an individual difference approach to examine visual cognition and in particular in studies trying to relate face recognition performance with measures of holistic processing (the face composite effect and the part-whole effect). In the present study we examine whether global precedence effects, measured by means of non-face stimuli in Navon's paradigm, can also account for individual differences in face recognition and, if so, whether the effect is of similar magnitude for faces and objects. We find evidence that global precedence effects facilitate both face and object recognition, and to a similar extent. Our results suggest that both face and object recognition are characterized by a coarse-to-fine temporal dynamic, where global shape information is derived prior to local shape information, and that the efficiency of face and object recognition is related to the magnitude of the global precedence effect.

  2. SEMI-SUPERVISED OBJECT RECOGNITION USING STRUCTURE KERNEL

    PubMed Central

    Wang, Botao; Xiong, Hongkai; Jiang, Xiaoqian; Ling, Fan

    2013-01-01

    Object recognition is a fundamental problem in computer vision. Part-based models offer a sparse, flexible representation of objects, but suffer from difficulties in training and often use standard kernels. In this paper, we propose a positive definite kernel called “structure kernel”, which measures the similarity of two part-based represented objects. The structure kernel has three terms: 1) the global term that measures the global visual similarity of two objects; 2) the part term that measures the visual similarity of corresponding parts; 3) the spatial term that measures the spatial similarity of geometric configuration of parts. The contribution of this paper is to generalize the discriminant capability of local kernels to complex part-based object models. Experimental results show that the proposed kernel exhibit higher accuracy than state-of-art approaches using standard kernels. PMID:23666108

  3. Plastic modifications induced by object recognition memory processing

    PubMed Central

    Clarke, Julia Rosauro; Cammarota, Martín; Gruart, Agnès; Izquierdo, Iván; Delgado-García, José María

    2010-01-01

    Long-term potentiation (LTP) phenomenon is widely accepted as a cellular model of memory consolidation. Object recognition (OR) is a particularly useful way of studying declarative memory in rodents because it makes use of their innate preference for novel over familiar objects. In this study, mice had electrodes implanted in the hippocampal Schaffer collaterals–pyramidal CA1 pathway and were trained for OR. Field EPSPs evoked at the CA3-CA1 synapse were recorded at the moment of training and at different times thereafter. LTP-like synaptic enhancement was found 6 h posttraining. A testing session was conducted 24 h after training, in the presence of one familiar and one novel object. Hippocampal synaptic facilitation was observed during exploration of familiar and novel objects. A short depotentiation period was observed early after the test and was followed by a later phase of synaptic efficacy enhancement. Here, we show that OR memory consolidation is accompanied by transient potentiation in the hippocampal CA3-CA1 synapses, while reconsolidation of this memory requires a short-lasting phase of depotentiation that could account for its well described vulnerability. The late synaptic enhancement phase, on the other hand, would be a consequence of memory restabilization. PMID:20133798

  4. Automatic anatomy recognition using neural network learning of object relationships via virtual landmarks

    NASA Astrophysics Data System (ADS)

    Yan, Fengxia; Udupa, Jayaram K.; Tong, Yubing; Xu, Guoping; Odhner, Dewey; Torigian, Drew A.

    2018-03-01

    The recently developed body-wide Automatic Anatomy Recognition (AAR) methodology depends on fuzzy modeling of individual objects, hierarchically arranging objects, constructing an anatomy ensemble of these models, and a dichotomous object recognition-delineation process. The parent-to-offspring spatial relationship in the object hierarchy is crucial in the AAR method. We have found this relationship to be quite complex, and as such any improvement in capturing this relationship information in the anatomy model will improve the process of recognition itself. Currently, the method encodes this relationship based on the layout of the geometric centers of the objects. Motivated by the concept of virtual landmarks (VLs), this paper presents a new one-shot AAR recognition method that utilizes the VLs to learn object relationships by training a neural network to predict the pose and the VLs of an offspring object given the VLs of the parent object in the hierarchy. We set up two neural networks for each parent-offspring object pair in a body region, one for predicting the VLs and another for predicting the pose parameters. The VL-based learning/prediction method is evaluated on two object hierarchies involving 14 objects. We utilize 54 computed tomography (CT) image data sets of head and neck cancer patients and the associated object contours drawn by dosimetrists for routine radiation therapy treatment planning. The VL neural network method is found to yield more accurate object localization than the currently used simple AAR method.

  5. Object recognition for autonomous robot utilizing distributed knowledge database

    NASA Astrophysics Data System (ADS)

    Takatori, Jiro; Suzuki, Kenji; Hartono, Pitoyo; Hashimoto, Shuji

    2003-10-01

    In this paper we present a novel method of object recognition utilizing a remote knowledge database for an autonomous robot. The developed robot has three robot arms with different sensors; two CCD cameras and haptic sensors. It can see, touch and move the target object from different directions. Referring to remote knowledge database of geometry and material, the robot observes and handles the objects to understand them including their physical characteristics.

  6. Neural substrates of view-invariant object recognition developed without experiencing rotations of the objects.

    PubMed

    Okamura, Jun-Ya; Yamaguchi, Reona; Honda, Kazunari; Wang, Gang; Tanaka, Keiji

    2014-11-05

    One fails to recognize an unfamiliar object across changes in viewing angle when it must be discriminated from similar distractor objects. View-invariant recognition gradually develops as the viewer repeatedly sees the objects in rotation. It is assumed that different views of each object are associated with one another while their successive appearance is experienced in rotation. However, natural experience of objects also contains ample opportunities to discriminate among objects at each of the multiple viewing angles. Our previous behavioral experiments showed that after experiencing a new set of object stimuli during a task that required only discrimination at each of four viewing angles at 30° intervals, monkeys could recognize the objects across changes in viewing angle up to 60°. By recording activities of neurons from the inferotemporal cortex after various types of preparatory experience, we here found a possible neural substrate for the monkeys' performance. For object sets that the monkeys had experienced during the task that required only discrimination at each of four viewing angles, many inferotemporal neurons showed object selectivity covering multiple views. The degree of view generalization found for these object sets was similar to that found for stimulus sets with which the monkeys had been trained to conduct view-invariant recognition. These results suggest that the experience of discriminating new objects in each of several viewing angles develops the partially view-generalized object selectivity distributed over many neurons in the inferotemporal cortex, which in turn bases the monkeys' emergent capability to discriminate the objects across changes in viewing angle. Copyright © 2014 the authors 0270-6474/14/3415047-13$15.00/0.

  7. Viewpoint-Specific Representations in Three-Dimensional Object Recognition

    DTIC Science & Technology

    1990-08-01

    for useful suggestions and illuminating discuc- sions, and Ellen Hildreth for her comments on a draft of this repcrt. References [1] 1. Biederman ...1982. [24] I. Rock and J. DiVita. A case of viewer-centered object perception. Cognitive Psychology, 19:280-293, 1987 . [25] I. Rock, D. Wheeler, and...Raleigh, NC, 1987 . [30] S. Ullman. Aligning pictorial descriptions: an approach to object recognition. Cognition, 32:193-254, 1989. [31] S. UUman and R

  8. Reversibility of object recognition but not spatial memory impairment following binge-like alcohol exposure in rats.

    PubMed

    Cippitelli, Andrea; Zook, Michelle; Bell, Lauren; Damadzic, Ruslan; Eskay, Robert L; Schwandt, Melanie; Heilig, Markus

    2010-11-01

    Excessive alcohol use leads to neurodegeneration in several brain structures including the hippocampal dentate gyrus and the entorhinal cortex. Cognitive deficits that result are among the most insidious and debilitating consequences of alcoholism. The object exploration task (OET) provides a sensitive measurement of spatial memory impairment induced by hippocampal and cortical damage. In this study, we examine whether the observed neurotoxicity produced by a 4-day binge ethanol treatment results in long-term memory impairment by observing the time course of reactions to spatial change (object configuration) and non-spatial change (object recognition). Wistar rats were assessed for their abilities to detect spatial configuration in the OET at 1 week and 10 weeks following the ethanol treatment, in which ethanol groups received 9-15 g/kg/day and achieved blood alcohol levels over 300 mg/dl. At 1 week, results indicated that the binge alcohol treatment produced impairment in both spatial memory and non-spatial object recognition performance. Unlike the controls, ethanol treated rats did not increase the duration or number of contacts with the displaced object in the spatial memory task, nor did they increase the duration of contacts with the novel object in the object recognition task. After 10 weeks, spatial memory remained impaired in the ethanol treated rats but object recognition ability was recovered. Our data suggest that episodes of binge-like alcohol exposure result in long-term and possibly permanent impairments in memory for the configuration of objects during exploration, whereas the ability to detect non-spatial changes is only temporarily affected. Copyright © 2010 Elsevier Inc. All rights reserved.

  9. Single-pixel non-imaging object recognition by means of Fourier spectrum acquisition

    NASA Astrophysics Data System (ADS)

    Chen, Huichao; Shi, Jianhong; Liu, Xialin; Niu, Zhouzhou; Zeng, Guihua

    2018-04-01

    Single-pixel imaging has emerged over recent years as a novel imaging technique, which has significant application prospects. In this paper, we propose and experimentally demonstrate a scheme that can achieve single-pixel non-imaging object recognition by acquiring the Fourier spectrum. In an experiment, a four-step phase-shifting sinusoid illumination light is used to irradiate the object image, the value of the light intensity is measured with a single-pixel detection unit, and the Fourier coefficients of the object image are obtained by a differential measurement. The Fourier coefficients are first cast into binary numbers to obtain the hash value. We propose a new method of perceptual hashing algorithm, which is combined with a discrete Fourier transform to calculate the hash value. The hash distance is obtained by calculating the difference of the hash value between the object image and the contrast images. By setting an appropriate threshold, the object image can be quickly and accurately recognized. The proposed scheme realizes single-pixel non-imaging perceptual hashing object recognition by using fewer measurements. Our result might open a new path for realizing object recognition with non-imaging.

  10. Speckle-learning-based object recognition through scattering media.

    PubMed

    Ando, Takamasa; Horisaki, Ryoichi; Tanida, Jun

    2015-12-28

    We experimentally demonstrated object recognition through scattering media based on direct machine learning of a number of speckle intensity images. In the experiments, speckle intensity images of amplitude or phase objects on a spatial light modulator between scattering plates were captured by a camera. We used the support vector machine for binary classification of the captured speckle intensity images of face and non-face data. The experimental results showed that speckles are sufficient for machine learning.

  11. Online Feature Transformation Learning for Cross-Domain Object Category Recognition.

    PubMed

    Zhang, Xuesong; Zhuang, Yan; Wang, Wei; Pedrycz, Witold

    2017-06-09

    In this paper, we introduce a new research problem termed online feature transformation learning in the context of multiclass object category recognition. The learning of a feature transformation is viewed as learning a global similarity metric function in an online manner. We first consider the problem of online learning a feature transformation matrix expressed in the original feature space and propose an online passive aggressive feature transformation algorithm. Then these original features are mapped to kernel space and an online single kernel feature transformation (OSKFT) algorithm is developed to learn a nonlinear feature transformation. Based on the OSKFT and the existing Hedge algorithm, a novel online multiple kernel feature transformation algorithm is also proposed, which can further improve the performance of online feature transformation learning in large-scale application. The classifier is trained with k nearest neighbor algorithm together with the learned similarity metric function. Finally, we experimentally examined the effect of setting different parameter values in the proposed algorithms and evaluate the model performance on several multiclass object recognition data sets. The experimental results demonstrate the validity and good performance of our methods on cross-domain and multiclass object recognition application.

  12. Mechanisms and neural basis of object and pattern recognition: a study with chess experts.

    PubMed

    Bilalić, Merim; Langner, Robert; Erb, Michael; Grodd, Wolfgang

    2010-11-01

    Comparing experts with novices offers unique insights into the functioning of cognition, based on the maximization of individual differences. Here we used this expertise approach to disentangle the mechanisms and neural basis behind two processes that contribute to everyday expertise: object and pattern recognition. We compared chess experts and novices performing chess-related and -unrelated (visual) search tasks. As expected, the superiority of experts was limited to the chess-specific task, as there were no differences in a control task that used the same chess stimuli but did not require chess-specific recognition. The analysis of eye movements showed that experts immediately and exclusively focused on the relevant aspects in the chess task, whereas novices also examined irrelevant aspects. With random chess positions, when pattern knowledge could not be used to guide perception, experts nevertheless maintained an advantage. Experts' superior domain-specific parafoveal vision, a consequence of their knowledge about individual domain-specific symbols, enabled improved object recognition. Functional magnetic resonance imaging corroborated this differentiation between object and pattern recognition and showed that chess-specific object recognition was accompanied by bilateral activation of the occipitotemporal junction, whereas chess-specific pattern recognition was related to bilateral activations in the middle part of the collateral sulci. Using the expertise approach together with carefully chosen controls and multiple dependent measures, we identified object and pattern recognition as two essential cognitive processes in expert visual cognition, which may also help to explain the mechanisms of everyday perception.

  13. Critical object recognition in millimeter-wave images with robustness to rotation and scale.

    PubMed

    Mohammadzade, Hoda; Ghojogh, Benyamin; Faezi, Sina; Shabany, Mahdi

    2017-06-01

    Locating critical objects is crucial in various security applications and industries. For example, in security applications, such as in airports, these objects might be hidden or covered under shields or secret sheaths. Millimeter-wave images can be utilized to discover and recognize the critical objects out of the hidden cases without any health risk due to their non-ionizing features. However, millimeter-wave images usually have waves in and around the detected objects, making object recognition difficult. Thus, regular image processing and classification methods cannot be used for these images and additional pre-processings and classification methods should be introduced. This paper proposes a novel pre-processing method for canceling rotation and scale using principal component analysis. In addition, a two-layer classification method is introduced and utilized for recognition. Moreover, a large dataset of millimeter-wave images is collected and created for experiments. Experimental results show that a typical classification method such as support vector machines can recognize 45.5% of a type of critical objects at 34.2% false alarm rate (FAR), which is a drastically poor recognition. The same method within the proposed recognition framework achieves 92.9% recognition rate at 0.43% FAR, which indicates a highly significant improvement. The significant contribution of this work is to introduce a new method for analyzing millimeter-wave images based on machine vision and learning approaches, which is not yet widely noted in the field of millimeter-wave image analysis.

  14. Object recognition in images via a factor graph model

    NASA Astrophysics Data System (ADS)

    He, Yong; Wang, Long; Wu, Zhaolin; Zhang, Haisu

    2018-04-01

    Object recognition in images suffered from huge search space and uncertain object profile. Recently, the Bag-of- Words methods are utilized to solve these problems, especially the 2-dimension CRF(Conditional Random Field) model. In this paper we suggest the method based on a general and flexible fact graph model, which can catch the long-range correlation in Bag-of-Words by constructing a network learning framework contrasted from lattice in CRF. Furthermore, we explore a parameter learning algorithm based on the gradient descent and Loopy Sum-Product algorithms for the factor graph model. Experimental results on Graz 02 dataset show that, the recognition performance of our method in precision and recall is better than a state-of-art method and the original CRF model, demonstrating the effectiveness of the proposed method.

  15. The Vanderbilt Expertise Test Reveals Domain-General and Domain-Specific Sex Effects in Object Recognition

    PubMed Central

    McGugin, Rankin W.; Richler, Jennifer J.; Herzmann, Grit; Speegle, Magen; Gauthier, Isabel

    2012-01-01

    Individual differences in face recognition are often contrasted with differences in object recognition using a single object category. Likewise, individual differences in perceptual expertise for a given object domain have typically been measured relative to only a single category baseline. In Experiment 1, we present a new test of object recognition, the Vanderbilt Expertise Test (VET), which is comparable in methods to the Cambridge Face Memory Task (CFMT) but uses eight different object categories. Principal component analysis reveals that the underlying structure of the VET can be largely explained by two independent factors, which demonstrate good reliability and capture interesting sex differences inherent in the VET structure. In Experiment 2, we show how the VET can be used to separate domain-specific from domain-general contributions to a standard measure of perceptual expertise. While domain-specific contributions are found for car matching for both men and women and for plane matching in men, women in this sample appear to use more domain-general strategies to match planes. In Experiment 3, we use the VET to demonstrate that holistic processing of faces predicts face recognition independently of general object recognition ability, which has a sex-specific contribution to face recognition. Overall, the results suggest that the VET is a reliable and valid measure of object recognition abilities and can measure both domain-general skills and domain-specific expertise, which were both found to depend on the sex of observers. PMID:22877929

  16. Intraperirhinal cortex administration of the synthetic cannabinoid, HU210, disrupts object recognition memory in rats.

    PubMed

    Sticht, Martin A; Jacklin, Derek L; Mechoulam, Raphael; Parker, Linda A; Winters, Boyer D

    2015-03-25

    Cannabinoids disrupt learning and memory in human and nonhuman participants. Object recognition memory, which is particularly susceptible to the impairing effects of cannabinoids, relies critically on the perirhinal cortex (PRh); however, to date, the effects of cannabinoids within PRh have not been assessed. In the present study, we evaluated the effects of localized administration of the synthetic cannabinoid, HU210 (0.01, 1.0 μg/hemisphere), into PRh on spontaneous object recognition in Long-Evans rats. Animals received intra-PRh infusions of HU210 before the sample phase, and object recognition memory was assessed at various delays in a subsequent retention test. We found that presample intra-PRh HU210 dose dependently (1.0 μg but not 0.01 μg) interfered with spontaneous object recognition performance, exerting an apparently more pronounced effect when memory demands were increased. These novel findings show that cannabinoid agonists in PRh disrupt object recognition memory. Copyright © 2015 Wolters Kluwer Health, Inc. All rights reserved.

  17. Three-dimensional model-based object recognition and segmentation in cluttered scenes.

    PubMed

    Mian, Ajmal S; Bennamoun, Mohammed; Owens, Robyn

    2006-10-01

    Viewpoint independent recognition of free-form objects and their segmentation in the presence of clutter and occlusions is a challenging task. We present a novel 3D model-based algorithm which performs this task automatically and efficiently. A 3D model of an object is automatically constructed offline from its multiple unordered range images (views). These views are converted into multidimensional table representations (which we refer to as tensors). Correspondences are automatically established between these views by simultaneously matching the tensors of a view with those of the remaining views using a hash table-based voting scheme. This results in a graph of relative transformations used to register the views before they are integrated into a seamless 3D model. These models and their tensor representations constitute the model library. During online recognition, a tensor from the scene is simultaneously matched with those in the library by casting votes. Similarity measures are calculated for the model tensors which receive the most votes. The model with the highest similarity is transformed to the scene and, if it aligns accurately with an object in the scene, that object is declared as recognized and is segmented. This process is repeated until the scene is completely segmented. Experiments were performed on real and synthetic data comprised of 55 models and 610 scenes and an overall recognition rate of 95 percent was achieved. Comparison with the spin images revealed that our algorithm is superior in terms of recognition rate and efficiency.

  18. Neurophysiological indices of perceptual object priming in the absence of explicit recognition memory.

    PubMed

    Harris, Jill D; Cutmore, Tim R H; O'Gorman, John; Finnigan, Simon; Shum, David

    2009-02-01

    The aim of this study was to identify ERP correlates of perceptual object priming that are insensitive to factors affecting explicit, episodic memory. EEG was recorded from 21 participants while they performed a visual object recognition test on a combination of unstudied items and old items that were previously encountered during either a 'deep' or 'shallow' levels-of-processing (LOP) study task. The results demonstrated a midline P150 old/new effect which was sensitive only to objects' old/new status and not to the accuracy of recognition responses to old items, or to the LOP manipulation. Similar outcomes were observed for the subsequent P200 and N400 effects, the former of which had a parietal scalp maximum and the latter, a broadly distributed topography. In addition an LPC old/new effect typical of those reported in past ERP recognition studies was observed. These outcomes support the proposal that the P150 effect is reflective of perceptual object priming and moreover, provide novel evidence that this and the P200 effect are independent of explicit recognition memory process(es).

  19. Dietary effects on object recognition: The impact of high-fat high-sugar diets on recollection and familiarity-based memory.

    PubMed

    Tran, Dominic M D; Westbrook, R Frederick

    2018-05-31

    Exposure to a high-fat high-sugar (HFHS) diet rapidly impairs novel-place- but not novel-object-recognition memory in rats (Tran & Westbrook, 2015, 2017). Three experiments sought to investigate the generality of diet-induced cognitive deficits by examining whether there are conditions under which object-recognition memory is impaired. Experiments 1 and 3 tested the strength of short- and long-term object-memory trace, respectively, by varying the interval of time between object familiarization and subsequent novel object test. Experiment 2 tested the effect of increasing working memory load on object-recognition memory by interleaving additional object exposures between familiarization and test in an n-back style task. Experiments 1-3 failed to detect any differences in object recognition between HFHS and control rats. Experiment 4 controlled for object novelty by separately familiarizing both objects presented at test, which included one remote-familiar and one recent-familiar object. Under these conditions, when test objects differed in their relative recency, HFHS rats showed a weaker memory trace for the remote object compared to chow rats. This result suggests that the diet leaves intact recollection judgments, but impairs familiarity judgments. We speculate that the HFHS diet adversely affects "where" memories as well as the quality of "what" memories, and discuss these effects in relation to recollection and familiarity memory models, hippocampal-dependent functions, and episodic food memories. (PsycINFO Database Record (c) 2018 APA, all rights reserved).

  20. The relationship between change detection and recognition of centrally attended objects in motion pictures.

    PubMed

    Angelone, Bonnie L; Levin, Daniel T; Simons, Daniel J

    2003-01-01

    Observers typically detect changes to central objects more readily than changes to marginal objects, but they sometimes miss changes to central, attended objects as well. However, even if observers do not report such changes, they may be able to recognize the changed object. In three experiments we explored change detection and recognition memory for several types of changes to central objects in motion pictures. Observers who failed to detect a change still performed at above chance levels on a recognition task in almost all conditions. In addition, observers who detected the change were no more accurate in their recognition than those who did not detect the change. Despite large differences in the detectability of changes across conditions, those observers who missed the change did not vary in their ability to recognize the changing object.

  1. The Neural Regions Sustaining Episodic Encoding and Recognition of Objects

    ERIC Educational Resources Information Center

    Hofer, Alex; Siedentopf, Christian M.; Ischebeck, Anja; Rettenbacher, Maria A.; Widschwendter, Christian G.; Verius, Michael; Golaszewski, Stefan M.; Koppelstaetter, Florian; Felber, Stephan; Wolfgang Fleischhacker, W.

    2007-01-01

    In this functional MRI experiment, encoding of objects was associated with activation in left ventrolateral prefrontal/insular and right dorsolateral prefrontal and fusiform regions as well as in the left putamen. By contrast, correct recognition of previously learned objects (R judgments) produced activation in left superior frontal, bilateral…

  2. Development of novel tasks for studying view-invariant object recognition in rodents: Sensitivity to scopolamine.

    PubMed

    Mitchnick, Krista A; Wideman, Cassidy E; Huff, Andrew E; Palmer, Daniel; McNaughton, Bruce L; Winters, Boyer D

    2018-05-15

    The capacity to recognize objects from different view-points or angles, referred to as view-invariance, is an essential process that humans engage in daily. Currently, the ability to investigate the neurobiological underpinnings of this phenomenon is limited, as few ethologically valid view-invariant object recognition tasks exist for rodents. Here, we report two complementary, novel view-invariant object recognition tasks in which rodents physically interact with three-dimensional objects. Prior to experimentation, rats and mice were given extensive experience with a set of 'pre-exposure' objects. In a variant of the spontaneous object recognition task, novelty preference for pre-exposed or new objects was assessed at various angles of rotation (45°, 90° or 180°); unlike control rodents, for whom the objects were novel, rats and mice tested with pre-exposed objects did not discriminate between rotated and un-rotated objects in the choice phase, indicating substantial view-invariant object recognition. Secondly, using automated operant touchscreen chambers, rats were tested on pre-exposed or novel objects in a pairwise discrimination task, where the rewarded stimulus (S+) was rotated (180°) once rats had reached acquisition criterion; rats tested with pre-exposed objects re-acquired the pairwise discrimination following S+ rotation more effectively than those tested with new objects. Systemic scopolamine impaired performance on both tasks, suggesting involvement of acetylcholine at muscarinic receptors in view-invariant object processing. These tasks present novel means of studying the behavioral and neural bases of view-invariant object recognition in rodents. Copyright © 2018 Elsevier B.V. All rights reserved.

  3. Standard object recognition memory and "what" and "where" components: Improvement by post-training epinephrine in highly habituated rats.

    PubMed

    Jurado-Berbel, Patricia; Costa-Miserachs, David; Torras-Garcia, Meritxell; Coll-Andreu, Margalida; Portell-Cortés, Isabel

    2010-02-11

    The present work examined whether post-training systemic epinephrine (EPI) is able to modulate short-term (3h) and long-term (24 h and 48 h) memory of standard object recognition, as well as long-term (24 h) memory of separate "what" (object identity) and "where" (object location) components of object recognition. Although object recognition training is associated to low arousal levels, all the animals received habituation to the training box in order to further reduce emotional arousal. Post-training EPI improved long-term (24 h and 48 h), but not short-term (3 h), memory in the standard object recognition task, as well as 24 h memory for both object identity and object location. These data indicate that post-training epinephrine: (1) facilitates long-term memory for standard object recognition; (2) exerts separate facilitatory effects on "what" (object identity) and "where" (object location) components of object recognition; and (3) is capable of improving memory for a low arousing task even in highly habituated rats.

  4. Toward a unified model of face and object recognition in the human visual system

    PubMed Central

    Wallis, Guy

    2013-01-01

    Our understanding of the mechanisms and neural substrates underlying visual recognition has made considerable progress over the past 30 years. During this period, accumulating evidence has led many scientists to conclude that objects and faces are recognised in fundamentally distinct ways, and in fundamentally distinct cortical areas. In the psychological literature, in particular, this dissociation has led to a palpable disconnect between theories of how we process and represent the two classes of object. This paper follows a trend in part of the recognition literature to try to reconcile what we know about these two forms of recognition by considering the effects of learning. Taking a widely accepted, self-organizing model of object recognition, this paper explains how such a system is affected by repeated exposure to specific stimulus classes. In so doing, it explains how many aspects of recognition generally regarded as unusual to faces (holistic processing, configural processing, sensitivity to inversion, the other-race effect, the prototype effect, etc.) are emergent properties of category-specific learning within such a system. Overall, the paper describes how a single model of recognition learning can and does produce the seemingly very different types of representation associated with faces and objects. PMID:23966963

  5. It's all connected: Pathways in visual object recognition and early noun learning.

    PubMed

    Smith, Linda B

    2013-11-01

    A developmental pathway may be defined as the route, or chain of events, through which a new structure or function forms. For many human behaviors, including object name learning and visual object recognition, these pathways are often complex and multicausal and include unexpected dependencies. This article presents three principles of development that suggest the value of a developmental psychology that explicitly seeks to trace these pathways and uses empirical evidence on developmental dependencies among motor development, action on objects, visual object recognition, and object name learning in 12- to 24-month-old infants to make the case. The article concludes with a consideration of the theoretical implications of this approach. (PsycINFO Database Record (c) 2013 APA, all rights reserved).

  6. Image processing strategies based on saliency segmentation for object recognition under simulated prosthetic vision.

    PubMed

    Li, Heng; Su, Xiaofan; Wang, Jing; Kan, Han; Han, Tingting; Zeng, Yajie; Chai, Xinyu

    2018-01-01

    Current retinal prostheses can only generate low-resolution visual percepts constituted of limited phosphenes which are elicited by an electrode array and with uncontrollable color and restricted grayscale. Under this visual perception, prosthetic recipients can just complete some simple visual tasks, but more complex tasks like face identification/object recognition are extremely difficult. Therefore, it is necessary to investigate and apply image processing strategies for optimizing the visual perception of the recipients. This study focuses on recognition of the object of interest employing simulated prosthetic vision. We used a saliency segmentation method based on a biologically plausible graph-based visual saliency model and a grabCut-based self-adaptive-iterative optimization framework to automatically extract foreground objects. Based on this, two image processing strategies, Addition of Separate Pixelization and Background Pixel Shrink, were further utilized to enhance the extracted foreground objects. i) The results showed by verification of psychophysical experiments that under simulated prosthetic vision, both strategies had marked advantages over Direct Pixelization in terms of recognition accuracy and efficiency. ii) We also found that recognition performance under two strategies was tied to the segmentation results and was affected positively by the paired-interrelated objects in the scene. The use of the saliency segmentation method and image processing strategies can automatically extract and enhance foreground objects, and significantly improve object recognition performance towards recipients implanted a high-density implant. Copyright © 2017 Elsevier B.V. All rights reserved.

  7. A role for calcium-calmodulin-dependent protein kinase II in the consolidation of visual object recognition memory.

    PubMed

    Tinsley, C J; Narduzzo, K E; Ho, J W; Barker, G R; Brown, M W; Warburton, E C

    2009-09-01

    The aim was to investigate the role of calcium-calmodulin-dependent protein kinase (CAMK)II in object recognition memory. The performance of rats in a preferential object recognition test was examined after local infusion of the CAMKII inhibitors KN-62 or autocamtide-2-related inhibitory peptide (AIP) into the perirhinal cortex. KN-62 or AIP infused after acquisition impaired memory tested at 24 h, indicating an involvement of CAMKII in the consolidation of recognition memory. Memory was impaired when KN-62 was infused at 20 min after acquisition or when AIP was infused at 20, 40, 60 or 100 min after acquisition. The time-course of CAMKII activation in rats was further examined by immunohistochemical staining for phospho-CAMKII(Thre286)alpha at 10, 40, 70 and 100 min following the viewing of novel and familiar images. At 70 min, processing novel images resulted in more phospho-CAMKII(Thre286)alpha-stained neurons in the perirhinal cortex than did the processing of familiar images, consistent with the viewing of novel images increasing the activity of CAMKII at this time. This difference was eliminated by prior infusion of AIP. These findings establish that CAMKII is active within the perirhinal region between approximately 20 and 100 min following learning and then returns to baseline. Thus, increased CAMKII activity is essential for the consolidation of long-term object recognition memory but continuation of that increased activity throughout the 24 h memory delay is not necessary for maintenance of the memory.

  8. Ventromedial prefrontal cortex is obligatory for consolidation and reconsolidation of object recognition memory.

    PubMed

    Akirav, Irit; Maroun, Mouna

    2006-12-01

    Once consolidated, a long-term memory item could regain susceptibility to consolidation blockers, that is, reconsolidate, upon its reactivation. Both consolidation and reconsolidation require protein synthesis, but it is not yet known how similar these processes are in terms of molecular, cellular, and neural circuit mechanisms. Whereas most previous studies focused on aversive conditioning in the amygdala and the hippocampus, here we examine the role of the ventromedial prefrontal cortex (vmPFC) in consolidation and reconsolidation of object recognition memory. Object recognition memory is the ability to discriminate the familiarity of previously encountered objects. We found that microinfusion of the protein synthesis inhibitor anisomycin or the N-methyl-D-aspartate (NMDA) receptor antagonist D,L-2-amino-5-phosphonovaleric acid (APV) into the vmPFC, immediately after training, resulted in impairment of long-term (24 h) but not short-term (3 h) recognition memory. Similarly, microinfusion of anisomycin or APV into the vmPFC immediately after reactivation of the long-term memory impaired recognition memory 24 h, but not 3 h, post-reactivation. These results indicate that both protein synthesis and NMDA receptors are required for consolidation and reconsolidation of recognition memory in the vmPFC.

  9. A role for the CAMKK pathway in visual object recognition memory.

    PubMed

    Tinsley, Chris J; Narduzzo, Katherine E; Brown, Malcolm W; Warburton, E Clea

    2012-03-01

    The role of the CAMKK pathway in object recognition memory was investigated. Rats' performance in a preferential object recognition test was examined after local infusion into the perirhinal cortex of the CAMKK inhibitor STO-609. STO-609 infused either before or immediately after acquisition impaired memory tested after a 24 h but not a 20-min delay. Memory was not impaired when STO-609 was infused 20 min after acquisition. The expression of a downstream reaction product of CAMKK was measured by immunohistochemical staining for phospho-CAMKI(Thr177) at 10, 40, 70, and 100 min following the viewing of novel and familiar images of objects. Processing familiar images resulted in more pCAMKI stained neurons in the perirhinal cortex than processing novel images at the 10- and 40-min delays. Prior infusion of STO-609 caused a reduction in pCAMKI stained neurons in response to viewing either novel or familiar images, consistent with its role as an inhibitor of CAMKK. The results establish that the CAMKK pathway within the perirhinal cortex is important for the consolidation of object recognition memory. The activation of pCAMKI after acquisition is earlier than previously reported for pCAMKII. Copyright © 2011 Wiley Periodicals, Inc.

  10. Lexical leverage: Category knowledge boosts real-time novel word recognition in two-year- olds

    PubMed Central

    Borovsky, Arielle; Ellis, Erica M.; Evans, Julia L.; Elman, Jeffrey L.

    2016-01-01

    Recent research suggests that infants tend to add words to their vocabulary that are semantically related to other known words, though it is not clear why this pattern emerges. In this paper, we explore whether infants to leverage their existing vocabulary and semantic knowledge when interpreting novel label-object mappings in real-time. We initially identified categorical domains for which individual 24-month-old infants have relatively higher and lower levels of knowledge, irrespective of overall vocabulary size. Next, we taught infants novel words in these higher and lower knowledge domains and then asked if their subsequent real-time recognition of these items varied as a function of their category knowledge. While our participants successfully acquired the novel label -object mappings in our task, there were important differences in the way infants recognized these words in real time. Namely, infants showed more robust recognition of high (vs. low) domain knowledge words. These findings suggest that dense semantic structure facilitates early word learning and real-time novel word recognition. PMID:26452444

  11. Neuropeptide S interacts with the basolateral amygdala noradrenergic system in facilitating object recognition memory consolidation.

    PubMed

    Han, Ren-Wen; Xu, Hong-Jiao; Zhang, Rui-San; Wang, Pei; Chang, Min; Peng, Ya-Li; Deng, Ke-Yu; Wang, Rui

    2014-01-01

    The noradrenergic activity in the basolateral amygdala (BLA) was reported to be involved in the regulation of object recognition memory. As the BLA expresses high density of receptors for Neuropeptide S (NPS), we investigated whether the BLA is involved in mediating NPS's effects on object recognition memory consolidation and whether such effects require noradrenergic activity. Intracerebroventricular infusion of NPS (1nmol) post training facilitated 24-h memory in a mouse novel object recognition task. The memory-enhancing effect of NPS could be blocked by the β-adrenoceptor antagonist propranolol. Furthermore, post-training intra-BLA infusions of NPS (0.5nmol/side) improved 24-h memory for objects, which was impaired by co-administration of propranolol (0.5μg/side). Taken together, these results indicate that NPS interacts with the BLA noradrenergic system in improving object recognition memory during consolidation. Copyright © 2013 Elsevier Inc. All rights reserved.

  12. Dopamine D1 receptor activation leads to object recognition memory in a coral reef fish.

    PubMed

    Hamilton, Trevor J; Tresguerres, Martin; Kline, David I

    2017-07-01

    Object recognition memory is the ability to identify previously seen objects and is an adaptive mechanism that increases survival for many species throughout the animal kingdom. Previously believed to be possessed by only the highest order mammals, it is now becoming clear that fish are also capable of this type of memory formation. Similar to the mammalian hippocampus, the dorsolateral pallium regulates distinct memory processes and is modulated by neurotransmitters such as dopamine. Caribbean bicolour damselfish ( Stegastes partitus ) live in complex environments dominated by coral reef structures and thus likely possess many types of complex memory abilities including object recognition. This study used a novel object recognition test in which fish were first presented two identical objects, then after a retention interval of 10 min with no objects, the fish were presented with a novel object and one of the objects they had previously encountered in the first trial. We demonstrate that the dopamine D 1 -receptor agonist (SKF 38393) induces the formation of object recognition memories in these fish. Thus, our results suggest that dopamine-receptor mediated enhancement of spatial memory formation in fish represents an evolutionarily conserved mechanism in vertebrates. © 2017 The Author(s).

  13. Picture object recognition in an American black bear (Ursus americanus).

    PubMed

    Johnson-Ulrich, Zoe; Vonk, Jennifer; Humbyrd, Mary; Crowley, Marilyn; Wojtkowski, Ela; Yates, Florence; Allard, Stephanie

    2016-11-01

    Many animals have been tested for conceptual discriminations using two-dimensional images as stimuli, and many of these species appear to transfer knowledge from 2D images to analogous real life objects. We tested an American black bear for picture-object recognition using a two alternative forced choice task. She was presented with four unique sets of objects and corresponding pictures. The bear showed generalization from both objects to pictures and pictures to objects; however, her transfer was superior when transferring from real objects to pictures, suggesting that bears can recognize visual features from real objects within photographic images during discriminations.

  14. Multiscale moment-based technique for object matching and recognition

    NASA Astrophysics Data System (ADS)

    Thio, HweeLi; Chen, Liya; Teoh, Eam-Khwang

    2000-03-01

    A new method is proposed to extract features from an object for matching and recognition. The features proposed are a combination of local and global characteristics -- local characteristics from the 1-D signature function that is defined to each pixel on the object boundary, global characteristics from the moments that are generated from the signature function. The boundary of the object is first extracted, then the signature function is generated by computing the angle between two lines from every point on the boundary as a function of position along the boundary. This signature function is position, scale and rotation invariant (PSRI). The shape of the signature function is then described quantitatively by using moments. The moments of the signature function are the global characters of a local feature set. Using moments as the eventual features instead of the signature function reduces the time and complexity of an object matching application. Multiscale moments are implemented to produce several sets of moments that will generate more accurate matching. Basically multiscale technique is a coarse to fine procedure and makes the proposed method more robust to noise. This method is proposed to match and recognize objects under simple transformation, such as translation, scale changes, rotation and skewing. A simple logo indexing system is implemented to illustrate the performance of the proposed method.

  15. An Approach to Object Recognition: Aligning Pictorial Descriptions.

    DTIC Science & Technology

    1986-12-01

    PERFORMING 0RGANIZATION NAMIE ANDORS IS551. PROGRAM ELEMENT. PROJECT. TASK Artificial Inteligence Laboratory AREKA A WORK UNIT NUMBERS ( 545 Technology... ARTIFICIAL INTELLIGENCE LABORATORY A.I. Memo No. 931 December, 1986 AN APPROACH TO OBJECT RECOGNITION: ALIGNING PICTORIAL DESCRIPTIONS Shimon Ullman...within the Artificial Intelligence Laboratory at the Massachusetts Institute of Technology. Support for the A.I. Laboratory’s artificial intelligence

  16. Automatic image database generation from CAD for 3D object recognition

    NASA Astrophysics Data System (ADS)

    Sardana, Harish K.; Daemi, Mohammad F.; Ibrahim, Mohammad K.

    1993-06-01

    The development and evaluation of Multiple-View 3-D object recognition systems is based on a large set of model images. Due to the various advantages of using CAD, it is becoming more and more practical to use existing CAD data in computer vision systems. Current PC- level CAD systems are capable of providing physical image modelling and rendering involving positional variations in cameras, light sources etc. We have formulated a modular scheme for automatic generation of various aspects (views) of the objects in a model based 3-D object recognition system. These views are generated at desired orientations on the unit Gaussian sphere. With a suitable network file sharing system (NFS), the images can directly be stored on a database located on a file server. This paper presents the image modelling solutions using CAD in relation to multiple-view approach. Our modular scheme for data conversion and automatic image database storage for such a system is discussed. We have used this approach in 3-D polyhedron recognition. An overview of the results, advantages and limitations of using CAD data and conclusions using such as scheme are also presented.

  17. A cortical framework for invariant object categorization and recognition.

    PubMed

    Rodrigues, João; Hans du Buf, J M

    2009-08-01

    In this paper we present a new model for invariant object categorization and recognition. It is based on explicit multi-scale features: lines, edges and keypoints are extracted from responses of simple, complex and end-stopped cells in cortical area V1, and keypoints are used to construct saliency maps for Focus-of-Attention. The model is a functional but dichotomous one, because keypoints are employed to model the "where" data stream, with dynamic routing of features from V1 to higher areas to obtain translation, rotation and size invariance, whereas lines and edges are employed in the "what" stream for object categorization and recognition. Furthermore, both the "where" and "what" pathways are dynamic in that information at coarse scales is employed first, after which information at progressively finer scales is added in order to refine the processes, i.e., both the dynamic feature routing and the categorization level. The construction of group and object templates, which are thought to be available in the prefrontal cortex with "what" and "where" components in PF46d and PF46v, is also illustrated. The model was tested in the framework of an integrated and biologically plausible architecture.

  18. Multi-channel feature dictionaries for RGB-D object recognition

    NASA Astrophysics Data System (ADS)

    Lan, Xiaodong; Li, Qiming; Chong, Mina; Song, Jian; Li, Jun

    2018-04-01

    Hierarchical matching pursuit (HMP) is a popular feature learning method for RGB-D object recognition. However, the feature representation with only one dictionary for RGB channels in HMP does not capture sufficient visual information. In this paper, we propose multi-channel feature dictionaries based feature learning method for RGB-D object recognition. The process of feature extraction in the proposed method consists of two layers. The K-SVD algorithm is used to learn dictionaries in sparse coding of these two layers. In the first-layer, we obtain features by performing max pooling on sparse codes of pixels in a cell. And the obtained features of cells in a patch are concatenated to generate patch jointly features. Then, patch jointly features in the first-layer are used to learn the dictionary and sparse codes in the second-layer. Finally, spatial pyramid pooling can be applied to the patch jointly features of any layer to generate the final object features in our method. Experimental results show that our method with first or second-layer features can obtain a comparable or better performance than some published state-of-the-art methods.

  19. Time and timing in the acoustic recognition system of crickets

    PubMed Central

    Hennig, R. Matthias; Heller, Klaus-Gerhard; Clemens, Jan

    2014-01-01

    The songs of many insects exhibit precise timing as the result of repetitive and stereotyped subunits on several time scales. As these signals encode the identity of a species, time and timing are important for the recognition system that analyzes these signals. Crickets are a prominent example as their songs are built from sound pulses that are broadcast in a long trill or as a chirped song. This pattern appears to be analyzed on two timescales, short and long. Recent evidence suggests that song recognition in crickets relies on two computations with respect to time; a short linear-nonlinear (LN) model that operates as a filter for pulse rate and a longer integration time window for monitoring song energy over time. Therefore, there is a twofold role for timing. A filter for pulse rate shows differentiating properties for which the specific timing of excitation and inhibition is important. For an integrator, however, the duration of the time window is more important than the precise timing of events. Here, we first review evidence for the role of LN-models and integration time windows for song recognition in crickets. We then parameterize the filter part by Gabor functions and explore the effects of duration, frequency, phase, and offset as these will correspond to differently timed patterns of excitation and inhibition. These filter properties were compared with known preference functions of crickets and katydids. In a comparative approach, the power for song discrimination by LN-models was tested with the songs of over 100 cricket species. It is demonstrated how the acoustic signals of crickets occupy a simple 2-dimensional space for song recognition that arises from timing, described by a Gabor function, and time, the integration window. Finally, we discuss the evolution of recognition systems in insects based on simple sensory computations. PMID:25161622

  20. Grouping in object recognition: the role of a Gestalt law in letter identification.

    PubMed

    Pelli, Denis G; Majaj, Najib J; Raizman, Noah; Christian, Christopher J; Kim, Edward; Palomares, Melanie C

    2009-02-01

    The Gestalt psychologists reported a set of laws describing how vision groups elements to recognize objects. The Gestalt laws "prescribe for us what we are to recognize 'as one thing'" (Kohler, 1920). Were they right? Does object recognition involve grouping? Tests of the laws of grouping have been favourable, but mostly assessed only detection, not identification, of the compound object. The grouping of elements seen in the detection experiments with lattices and "snakes in the grass" is compelling, but falls far short of the vivid everyday experience of recognizing a familiar, meaningful, named thing, which mediates the ordinary identification of an object. Thus, after nearly a century, there is hardly any evidence that grouping plays a role in ordinary object recognition. To assess grouping in object recognition, we made letters out of grating patches and measured threshold contrast for identifying these letters in visual noise as a function of perturbation of grating orientation, phase, and offset. We define a new measure, "wiggle", to characterize the degree to which these various perturbations violate the Gestalt law of good continuation. We find that efficiency for letter identification is inversely proportional to wiggle and is wholly determined by wiggle, independent of how the wiggle was produced. Thus the effects of three different kinds of shape perturbation on letter identifiability are predicted by a single measure of goodness of continuation. This shows that letter identification obeys the Gestalt law of good continuation and may be the first confirmation of the original Gestalt claim that object recognition involves grouping.

  1. Grouping in object recognition: The role of a Gestalt law in letter identification

    PubMed Central

    Pelli, Denis G.; Majaj, Najib J.; Raizman, Noah; Christian, Christopher J.; Kim, Edward; Palomares, Melanie C.

    2009-01-01

    The Gestalt psychologists reported a set of laws describing how vision groups elements to recognize objects. The Gestalt laws “prescribe for us what we are to recognize ‘as one thing’” (Köhler, 1920). Were they right? Does object recognition involve grouping? Tests of the laws of grouping have been favourable, but mostly assessed only detection, not identification, of the compound object. The grouping of elements seen in the detection experiments with lattices and “snakes in the grass” is compelling, but falls far short of the vivid everyday experience of recognizing a familiar, meaningful, named thing, which mediates the ordinary identification of an object. Thus, after nearly a century, there is hardly any evidence that grouping plays a role in ordinary object recognition. To assess grouping in object recognition, we made letters out of grating patches and measured threshold contrast for identifying these letters in visual noise as a function of perturbation of grating orientation, phase, and offset. We define a new measure, “wiggle”, to characterize the degree to which these various perturbations violate the Gestalt law of good continuation. We find that efficiency for letter identification is inversely proportional to wiggle and is wholly determined by wiggle, independent of how the wiggle was produced. Thus the effects of three different kinds of shape perturbation on letter identifiability are predicted by a single measure of goodness of continuation. This shows that letter identification obeys the Gestalt law of good continuation and may be the first confirmation of the original Gestalt claim that object recognition involves grouping. PMID:19424881

  2. Using Markov Chains and Multi-Objective Optimization for Energy-Efficient Context Recognition.

    PubMed

    Janko, Vito; Luštrek, Mitja

    2017-12-29

    The recognition of the user's context with wearable sensing systems is a common problem in ubiquitous computing. However, the typically small battery of such systems often makes continuous recognition impractical. The strain on the battery can be reduced if the sensor setting is adapted to each context. We propose a method that efficiently finds near-optimal sensor settings for each context. It uses Markov chains to simulate the behavior of the system in different configurations and the multi-objective genetic algorithm to find a set of good non-dominated configurations. The method was evaluated on three real-life datasets and found good trade-offs between the system's energy expenditure and the system's accuracy. One of the solutions, for example, consumed five-times less energy than the default one, while sacrificing only two percentage points of accuracy.

  3. Object recognition of real targets using modelled SAR images

    NASA Astrophysics Data System (ADS)

    Zherdev, D. A.

    2017-12-01

    In this work the problem of recognition is studied using SAR images. The algorithm of recognition is based on the computation of conjugation indices with vectors of class. The support subspaces for each class are constructed by exception of the most and the less correlated vectors in a class. In the study we examine the ability of a significant feature vector size reduce that leads to recognition time decrease. The images of targets form the feature vectors that are transformed using pre-trained convolutional neural network (CNN).

  4. Three Dimensional Object Recognition Using an Unsupervised Neural Network: Understanding the Distinguishing Features

    DTIC Science & Technology

    1992-12-23

    predominance of structural models of recognition, of which a recent example is the Recognition By Components (RBC) theory ( Biederman , 1987 ). Structural...related to recent statistical theory (Huber, 1985; Friedman, 1987 ) and is derived from a biologically motivated computational theory (Bienenstock et...dimensional object recognition (Intrator and Gold, 1991). The method is related to recent statistical theory (Huber, 1985; Friedman, 1987 ) and is derived

  5. Intrinsic and contextual features in object recognition.

    PubMed

    Schlangen, Derrick; Barenholtz, Elan

    2015-01-28

    The context in which an object is found can facilitate its recognition. Yet, it is not known how effective this contextual information is relative to the object's intrinsic visual features, such as color and shape. To address this, we performed four experiments using rendered scenes with novel objects. In each experiment, participants first performed a visual search task, searching for a uniquely shaped target object whose color and location within the scene was experimentally manipulated. We then tested participants' tendency to use their knowledge of the location and color information in an identification task when the objects' images were degraded due to blurring, thus eliminating the shape information. In Experiment 1, we found that, in the absence of any diagnostic intrinsic features, participants identified objects based purely on their locations within the scene. In Experiment 2, we found that participants combined an intrinsic feature, color, with contextual location in order to uniquely specify an object. In Experiment 3, we found that when an object's color and location information were in conflict, participants identified the object using both sources of information equally. Finally, in Experiment 4, we found that participants used whichever source of information-either color or location-was more statistically reliable in order to identify the target object. Overall, these experiments show that the context in which objects are found can play as important a role as intrinsic features in identifying the objects. © 2015 ARVO.

  6. Activity Recognition for Personal Time Management

    NASA Astrophysics Data System (ADS)

    Prekopcsák, Zoltán; Soha, Sugárka; Henk, Tamás; Gáspár-Papanek, Csaba

    We describe an accelerometer based activity recognition system for mobile phones with a special focus on personal time management. We compare several data mining algorithms for the automatic recognition task in the case of single user and multiuser scenario, and improve accuracy with heuristics and advanced data mining methods. The results show that daily activities can be recognized with high accuracy and the integration with the RescueTime software can give good insights for personal time management.

  7. Computing with Connections in Visual Recognition of Origami Objects.

    ERIC Educational Resources Information Center

    Sabbah, Daniel

    1985-01-01

    Summarizes an initial foray in tackling artificial intelligence problems using a connectionist approach. The task chosen is visual recognition of Origami objects, and the questions answered are how to construct a connectionist network to represent and recognize projected Origami line drawings and the advantages such an approach would have. (30…

  8. Humans and Deep Networks Largely Agree on Which Kinds of Variation Make Object Recognition Harder.

    PubMed

    Kheradpisheh, Saeed R; Ghodrati, Masoud; Ganjtabesh, Mohammad; Masquelier, Timothée

    2016-01-01

    View-invariant object recognition is a challenging problem that has attracted much attention among the psychology, neuroscience, and computer vision communities. Humans are notoriously good at it, even if some variations are presumably more difficult to handle than others (e.g., 3D rotations). Humans are thought to solve the problem through hierarchical processing along the ventral stream, which progressively extracts more and more invariant visual features. This feed-forward architecture has inspired a new generation of bio-inspired computer vision systems called deep convolutional neural networks (DCNN), which are currently the best models for object recognition in natural images. Here, for the first time, we systematically compared human feed-forward vision and DCNNs at view-invariant object recognition task using the same set of images and controlling the kinds of transformation (position, scale, rotation in plane, and rotation in depth) as well as their magnitude, which we call "variation level." We used four object categories: car, ship, motorcycle, and animal. In total, 89 human subjects participated in 10 experiments in which they had to discriminate between two or four categories after rapid presentation with backward masking. We also tested two recent DCNNs (proposed respectively by Hinton's group and Zisserman's group) on the same tasks. We found that humans and DCNNs largely agreed on the relative difficulties of each kind of variation: rotation in depth is by far the hardest transformation to handle, followed by scale, then rotation in plane, and finally position (much easier). This suggests that DCNNs would be reasonable models of human feed-forward vision. In addition, our results show that the variation levels in rotation in depth and scale strongly modulate both humans' and DCNNs' recognition performances. We thus argue that these variations should be controlled in the image datasets used in vision research.

  9. Adrenergic enhancement of consolidation of object recognition memory.

    PubMed

    Dornelles, Arethuza; de Lima, Maria Noemia Martins; Grazziotin, Manoela; Presti-Torres, Juliana; Garcia, Vanessa Athaide; Scalco, Felipe Siciliani; Roesler, Rafael; Schröder, Nadja

    2007-07-01

    Extensive evidence indicates that epinephrine (EPI) modulates memory consolidation for emotionally arousing tasks in animals and human subjects. However, previous studies have not examined the effects of EPI on consolidation of recognition memory. Here we report that systemic administration of EPI enhances consolidation of memory for a novel object recognition (NOR) task under different training conditions. Control male rats given a systemic injection of saline (0.9% NaCl) immediately after NOR training showed significant memory retention when tested at 1.5 or 24, but not 96h after training. In contrast, rats given a post-training injection of EPI showed significant retention of NOR at all delays. In a second experiment using a different training condition, rats treated with EPI, but not SAL-treated animals, showed significant NOR retention at both 1.5 and 24-h delays. We next showed that the EPI-induced enhancement of retention tested at 96h after training was prevented by pretraining systemic administration of the beta-adrenoceptor antagonist propranolol. The findings suggest that, as previously observed in experiments using aversively motivated tasks, epinephrine modulates consolidation of recognition memory and that the effects require activation of beta-adrenoceptors.

  10. Neural network application for thermal image recognition of low-resolution objects

    NASA Astrophysics Data System (ADS)

    Fang, Yi-Chin; Wu, Bo-Wen

    2007-02-01

    In the ever-changing situation on a battle field, accurate recognition of a distant object is critical to a commander's decision-making and the general public's safety. Efficiently distinguishing between an enemy's armoured vehicles and ordinary civilian houses under all weather conditions has become an important research topic. This study presents a system for recognizing an armoured vehicle by distinguishing marks and contours. The characteristics of 12 different shapes and 12 characters are used to explore thermal image recognition under the circumstance of long distance and low resolution. Although the recognition capability of human eyes is superior to that of artificial intelligence under normal conditions, it tends to deteriorate substantially under long-distance and low-resolution scenarios. This study presents an effective method for choosing features and processing images. The artificial neural network technique is applied to further improve the probability of accurate recognition well beyond the limit of the recognition capability of human eyes.

  11. Comparison of deep neural networks to spatio-temporal cortical dynamics of human visual object recognition reveals hierarchical correspondence

    PubMed Central

    Cichy, Radoslaw Martin; Khosla, Aditya; Pantazis, Dimitrios; Torralba, Antonio; Oliva, Aude

    2016-01-01

    The complex multi-stage architecture of cortical visual pathways provides the neural basis for efficient visual object recognition in humans. However, the stage-wise computations therein remain poorly understood. Here, we compared temporal (magnetoencephalography) and spatial (functional MRI) visual brain representations with representations in an artificial deep neural network (DNN) tuned to the statistics of real-world visual recognition. We showed that the DNN captured the stages of human visual processing in both time and space from early visual areas towards the dorsal and ventral streams. Further investigation of crucial DNN parameters revealed that while model architecture was important, training on real-world categorization was necessary to enforce spatio-temporal hierarchical relationships with the brain. Together our results provide an algorithmically informed view on the spatio-temporal dynamics of visual object recognition in the human visual brain. PMID:27282108

  12. Comparison of deep neural networks to spatio-temporal cortical dynamics of human visual object recognition reveals hierarchical correspondence.

    PubMed

    Cichy, Radoslaw Martin; Khosla, Aditya; Pantazis, Dimitrios; Torralba, Antonio; Oliva, Aude

    2016-06-10

    The complex multi-stage architecture of cortical visual pathways provides the neural basis for efficient visual object recognition in humans. However, the stage-wise computations therein remain poorly understood. Here, we compared temporal (magnetoencephalography) and spatial (functional MRI) visual brain representations with representations in an artificial deep neural network (DNN) tuned to the statistics of real-world visual recognition. We showed that the DNN captured the stages of human visual processing in both time and space from early visual areas towards the dorsal and ventral streams. Further investigation of crucial DNN parameters revealed that while model architecture was important, training on real-world categorization was necessary to enforce spatio-temporal hierarchical relationships with the brain. Together our results provide an algorithmically informed view on the spatio-temporal dynamics of visual object recognition in the human visual brain.

  13. The Consolidation of Object and Context Recognition Memory Involve Different Regions of the Temporal Lobe

    ERIC Educational Resources Information Center

    Balderas, Israela; Rodriguez-Ortiz, Carlos J.; Salgado-Tonda, Paloma; Chavez-Hurtado, Julio; McGaugh, James L.; Bermudez-Rattoni, Federico

    2008-01-01

    These experiments investigated the involvement of several temporal lobe regions in consolidation of recognition memory. Anisomycin, a protein synthesis inhibitor, was infused into the hippocampus, perirhinal cortex, insular cortex, or basolateral amygdala of rats immediately after the sample phase of object or object-in-context recognition memory…

  14. Comparing object recognition from binary and bipolar edge images for visual prostheses.

    PubMed

    Jung, Jae-Hyun; Pu, Tian; Peli, Eli

    2016-11-01

    Visual prostheses require an effective representation method due to the limited display condition which has only 2 or 3 levels of grayscale in low resolution. Edges derived from abrupt luminance changes in images carry essential information for object recognition. Typical binary (black and white) edge images have been used to represent features to convey essential information. However, in scenes with a complex cluttered background, the recognition rate of the binary edge images by human observers is limited and additional information is required. The polarity of edges and cusps (black or white features on a gray background) carries important additional information; the polarity may provide shape from shading information missing in the binary edge image. This depth information may be restored by using bipolar edges. We compared object recognition rates from 16 binary edge images and bipolar edge images by 26 subjects to determine the possible impact of bipolar filtering in visual prostheses with 3 or more levels of grayscale. Recognition rates were higher with bipolar edge images and the improvement was significant in scenes with complex backgrounds. The results also suggest that erroneous shape from shading interpretation of bipolar edges resulting from pigment rather than boundaries of shape may confound the recognition.

  15. Comparing object recognition from binary and bipolar edge images for visual prostheses

    PubMed Central

    Jung, Jae-Hyun; Pu, Tian; Peli, Eli

    2017-01-01

    Visual prostheses require an effective representation method due to the limited display condition which has only 2 or 3 levels of grayscale in low resolution. Edges derived from abrupt luminance changes in images carry essential information for object recognition. Typical binary (black and white) edge images have been used to represent features to convey essential information. However, in scenes with a complex cluttered background, the recognition rate of the binary edge images by human observers is limited and additional information is required. The polarity of edges and cusps (black or white features on a gray background) carries important additional information; the polarity may provide shape from shading information missing in the binary edge image. This depth information may be restored by using bipolar edges. We compared object recognition rates from 16 binary edge images and bipolar edge images by 26 subjects to determine the possible impact of bipolar filtering in visual prostheses with 3 or more levels of grayscale. Recognition rates were higher with bipolar edge images and the improvement was significant in scenes with complex backgrounds. The results also suggest that erroneous shape from shading interpretation of bipolar edges resulting from pigment rather than boundaries of shape may confound the recognition. PMID:28458481

  16. Glucocorticoid effects on object recognition memory require training-associated emotional arousal.

    PubMed

    Okuda, Shoki; Roozendaal, Benno; McGaugh, James L

    2004-01-20

    Considerable evidence implicates glucocorticoid hormones in the regulation of memory consolidation and memory retrieval. The present experiments investigated whether the influence of these hormones on memory depends on the level of emotional arousal induced by the training experience. We investigated this issue in male Sprague-Dawley rats by examining the effects of immediate posttraining systemic injections of the glucocorticoid corticosterone on object recognition memory under two conditions that differed in their training-associated emotional arousal. In rats that were not previously habituated to the experimental context, corticosterone (0.3, 1.0, or 3.0 mg/kg, s.c.) administered immediately after a 3-min training trial enhanced 24-hr retention performance in an inverted-U shaped dose-response relationship. In contrast, corticosterone did not affect 24-hr retention of rats that received extensive prior habituation to the experimental context and, thus, had decreased novelty-induced emotional arousal during training. Additionally, immediate posttraining administration of corticosterone to nonhabituated rats, in doses that enhanced 24-hr retention, impaired object recognition performance at a 1-hr retention interval whereas corticosterone administered after training to well-habituated rats did not impair 1-hr retention. Thus, the present findings suggest that training-induced emotional arousal may be essential for glucocorticoid effects on object recognition memory.

  17. Contributions of Low and High Spatial Frequency Processing to Impaired Object Recognition Circuitry in Schizophrenia

    PubMed Central

    Calderone, Daniel J.; Hoptman, Matthew J.; Martínez, Antígona; Nair-Collins, Sangeeta; Mauro, Cristina J.; Bar, Moshe; Javitt, Daniel C.; Butler, Pamela D.

    2013-01-01

    Patients with schizophrenia exhibit cognitive and sensory impairment, and object recognition deficits have been linked to sensory deficits. The “frame and fill” model of object recognition posits that low spatial frequency (LSF) information rapidly reaches the prefrontal cortex (PFC) and creates a general shape of an object that feeds back to the ventral temporal cortex to assist object recognition. Visual dysfunction findings in schizophrenia suggest a preferential loss of LSF information. This study used functional magnetic resonance imaging (fMRI) and resting state functional connectivity (RSFC) to investigate the contribution of visual deficits to impaired object “framing” circuitry in schizophrenia. Participants were shown object stimuli that were intact or contained only LSF or high spatial frequency (HSF) information. For controls, fMRI revealed preferential activation to LSF information in precuneus, superior temporal, and medial and dorsolateral PFC areas, whereas patients showed a preference for HSF information or no preference. RSFC revealed a lack of connectivity between early visual areas and PFC for patients. These results demonstrate impaired processing of LSF information during object recognition in schizophrenia, with patients instead displaying increased processing of HSF information. This is consistent with findings of a preference for local over global visual information in schizophrenia. PMID:22735157

  18. How does the brain solve visual object recognition?

    PubMed Central

    Zoccolan, Davide; Rust, Nicole C.

    2012-01-01

    Mounting evidence suggests that “core object recognition,” the ability to rapidly recognize objects despite substantial appearance variation, is solved in the brain via a cascade of reflexive, largely feedforward computations that culminate in a powerful neuronal representation in the inferior temporal cortex. However, the algorithm that produces this solution remains little-understood. Here we review evidence ranging from individual neurons, to neuronal populations, to behavior, to computational models. We propose that understanding this algorithm will require using neuronal and psychophysical data to sift through many computational models, each based on building blocks of small, canonical sub-networks with a common functional goal. PMID:22325196

  19. Priming Contour-Deleted Images: Evidence for Immediate Representations in Visual Object Recognition.

    ERIC Educational Resources Information Center

    Biederman, Irving; Cooper, Eric E.

    1991-01-01

    Speed and accuracy of identification of pictures of objects are facilitated by prior viewing. Contributions of image features, convex or concave components, and object models in a repetition priming task were explored in 2 studies involving 96 college students. Results provide evidence of intermediate representations in visual object recognition.…

  20. Recurrent Convolutional Neural Networks: A Better Model of Biological Object Recognition.

    PubMed

    Spoerer, Courtney J; McClure, Patrick; Kriegeskorte, Nikolaus

    2017-01-01

    Feedforward neural networks provide the dominant model of how the brain performs visual object recognition. However, these networks lack the lateral and feedback connections, and the resulting recurrent neuronal dynamics, of the ventral visual pathway in the human and non-human primate brain. Here we investigate recurrent convolutional neural networks with bottom-up (B), lateral (L), and top-down (T) connections. Combining these types of connections yields four architectures (B, BT, BL, and BLT), which we systematically test and compare. We hypothesized that recurrent dynamics might improve recognition performance in the challenging scenario of partial occlusion. We introduce two novel occluded object recognition tasks to test the efficacy of the models, digit clutter (where multiple target digits occlude one another) and digit debris (where target digits are occluded by digit fragments). We find that recurrent neural networks outperform feedforward control models (approximately matched in parametric complexity) at recognizing objects, both in the absence of occlusion and in all occlusion conditions. Recurrent networks were also found to be more robust to the inclusion of additive Gaussian noise. Recurrent neural networks are better in two respects: (1) they are more neurobiologically realistic than their feedforward counterparts; (2) they are better in terms of their ability to recognize objects, especially under challenging conditions. This work shows that computer vision can benefit from using recurrent convolutional architectures and suggests that the ubiquitous recurrent connections in biological brains are essential for task performance.

  1. Biological object recognition in μ-radiography images

    NASA Astrophysics Data System (ADS)

    Prochazka, A.; Dammer, J.; Weyda, F.; Sopko, V.; Benes, J.; Zeman, J.; Jandejsek, I.

    2015-03-01

    This study presents an applicability of real-time microradiography to biological objects, namely to horse chestnut leafminer, Cameraria ohridella (Insecta: Lepidoptera, Gracillariidae) and following image processing focusing on image segmentation and object recognition. The microradiography of insects (such as horse chestnut leafminer) provides a non-invasive imaging that leaves the organisms alive. The imaging requires a high spatial resolution (micrometer scale) radiographic system. Our radiographic system consists of a micro-focus X-ray tube and two types of detectors. The first is a charge integrating detector (Hamamatsu flat panel), the second is a pixel semiconductor detector (Medipix2 detector). The latter allows detection of single quantum photon of ionizing radiation. We obtained numerous horse chestnuts leafminer pupae in several microradiography images easy recognizable in automatic mode using the image processing methods. We implemented an algorithm that is able to count a number of dead and alive pupae in images. The algorithm was based on two methods: 1) noise reduction using mathematical morphology filters, 2) Canny edge detection. The accuracy of the algorithm is higher for the Medipix2 (average recall for detection of alive pupae =0.99, average recall for detection of dead pupae =0.83), than for the flat panel (average recall for detection of alive pupae =0.99, average recall for detection of dead pupae =0.77). Therefore, we conclude that Medipix2 has lower noise and better displays contours (edges) of biological objects. Our method allows automatic selection and calculation of dead and alive chestnut leafminer pupae. It leads to faster monitoring of the population of one of the world's important insect pest.

  2. Using an Improved SIFT Algorithm and Fuzzy Closed-Loop Control Strategy for Object Recognition in Cluttered Scenes

    PubMed Central

    Nie, Haitao; Long, Kehui; Ma, Jun; Yue, Dan; Liu, Jinguo

    2015-01-01

    Partial occlusions, large pose variations, and extreme ambient illumination conditions generally cause the performance degradation of object recognition systems. Therefore, this paper presents a novel approach for fast and robust object recognition in cluttered scenes based on an improved scale invariant feature transform (SIFT) algorithm and a fuzzy closed-loop control method. First, a fast SIFT algorithm is proposed by classifying SIFT features into several clusters based on several attributes computed from the sub-orientation histogram (SOH), in the feature matching phase only features that share nearly the same corresponding attributes are compared. Second, a feature matching step is performed following a prioritized order based on the scale factor, which is calculated between the object image and the target object image, guaranteeing robust feature matching. Finally, a fuzzy closed-loop control strategy is applied to increase the accuracy of the object recognition and is essential for autonomous object manipulation process. Compared to the original SIFT algorithm for object recognition, the result of the proposed method shows that the number of SIFT features extracted from an object has a significant increase, and the computing speed of the object recognition processes increases by more than 40%. The experimental results confirmed that the proposed method performs effectively and accurately in cluttered scenes. PMID:25714094

  3. A knowledge-based object recognition system for applications in the space station

    NASA Technical Reports Server (NTRS)

    Dhawan, Atam P.

    1988-01-01

    A knowledge-based three-dimensional (3D) object recognition system is being developed. The system uses primitive-based hierarchical relational and structural matching for the recognition of 3D objects in the two-dimensional (2D) image for interpretation of the 3D scene. At present, the pre-processing, low-level preliminary segmentation, rule-based segmentation, and the feature extraction are completed. The data structure of the primitive viewing knowledge-base (PVKB) is also completed. Algorithms and programs based on attribute-trees matching for decomposing the segmented data into valid primitives were developed. The frame-based structural and relational descriptions of some objects were created and stored in a knowledge-base. This knowledge-base of the frame-based descriptions were developed on the MICROVAX-AI microcomputer in LISP environment. The simulated 3D scene of simple non-overlapping objects as well as real camera data of images of 3D objects of low-complexity have been successfully interpreted.

  4. Automation of the novel object recognition task for use in adolescent rats

    PubMed Central

    Silvers, Janelle M.; Harrod, Steven B.; Mactutus, Charles F.; Booze, Rosemarie M.

    2010-01-01

    The novel object recognition task is gaining popularity for its ability to test a complex behavior which relies on the integrity of memory and attention systems without placing undue stress upon the animal. While the task places few requirements upon the animal, it traditionally requires the experimenter to observe the test phase directly and record behavior. This approach can severely limit the number of subjects which can be tested in a reasonable period of time, as training and testing occur on the same day and span several hours. The current study was designed to test the feasibility of automation of this task for adolescent rats using standard activity chambers, with the goals of increased objectivity, flexibility, and throughput of subjects. PMID:17719091

  5. The research of edge extraction and target recognition based on inherent feature of objects

    NASA Astrophysics Data System (ADS)

    Xie, Yu-chan; Lin, Yu-chi; Huang, Yin-guo

    2008-03-01

    Current research on computer vision often needs specific techniques for particular problems. Little use has been made of high-level aspects of computer vision, such as three-dimensional (3D) object recognition, that are appropriate for large classes of problems and situations. In particular, high-level vision often focuses mainly on the extraction of symbolic descriptions, and pays little attention to the speed of processing. In order to extract and recognize target intelligently and rapidly, in this paper we developed a new 3D target recognition method based on inherent feature of objects in which cuboid was taken as model. On the basis of analysis cuboid nature contour and greyhound distributing characteristics, overall fuzzy evaluating technique was utilized to recognize and segment the target. Then Hough transform was used to extract and match model's main edges, we reconstruct aim edges by stereo technology in the end. There are three major contributions in this paper. Firstly, the corresponding relations between the parameters of cuboid model's straight edges lines in an image field and in the transform field were summed up. By those, the aimless computations and searches in Hough transform processing can be reduced greatly and the efficiency is improved. Secondly, as the priori knowledge about cuboids contour's geometry character known already, the intersections of the component extracted edges are taken, and assess the geometry of candidate edges matches based on the intersections, rather than the extracted edges. Therefore the outlines are enhanced and the noise is depressed. Finally, a 3-D target recognition method is proposed. Compared with other recognition methods, this new method has a quick response time and can be achieved with high-level computer vision. The method present here can be used widely in vision-guide techniques to strengthen its intelligence and generalization, which can also play an important role in object tracking, port AGV, robots

  6. RecceMan: an interactive recognition assistance for image-based reconnaissance: synergistic effects of human perception and computational methods for object recognition, identification, and infrastructure analysis

    NASA Astrophysics Data System (ADS)

    El Bekri, Nadia; Angele, Susanne; Ruckhäberle, Martin; Peinsipp-Byma, Elisabeth; Haelke, Bruno

    2015-10-01

    This paper introduces an interactive recognition assistance system for imaging reconnaissance. This system supports aerial image analysts on missions during two main tasks: Object recognition and infrastructure analysis. Object recognition concentrates on the classification of one single object. Infrastructure analysis deals with the description of the components of an infrastructure and the recognition of the infrastructure type (e.g. military airfield). Based on satellite or aerial images, aerial image analysts are able to extract single object features and thereby recognize different object types. It is one of the most challenging tasks in the imaging reconnaissance. Currently, there are no high potential ATR (automatic target recognition) applications available, as consequence the human observer cannot be replaced entirely. State-of-the-art ATR applications cannot assume in equal measure human perception and interpretation. Why is this still such a critical issue? First, cluttered and noisy images make it difficult to automatically extract, classify and identify object types. Second, due to the changed warfare and the rise of asymmetric threats it is nearly impossible to create an underlying data set containing all features, objects or infrastructure types. Many other reasons like environmental parameters or aspect angles compound the application of ATR supplementary. Due to the lack of suitable ATR procedures, the human factor is still important and so far irreplaceable. In order to use the potential benefits of the human perception and computational methods in a synergistic way, both are unified in an interactive assistance system. RecceMan® (Reconnaissance Manual) offers two different modes for aerial image analysts on missions: the object recognition mode and the infrastructure analysis mode. The aim of the object recognition mode is to recognize a certain object type based on the object features that originated from the image signatures. The

  7. Non-accidental properties, metric invariance, and encoding by neurons in a model of ventral stream visual object recognition, VisNet.

    PubMed

    Rolls, Edmund T; Mills, W Patrick C

    2018-05-01

    When objects transform into different views, some properties are maintained, such as whether the edges are convex or concave, and these non-accidental properties are likely to be important in view-invariant object recognition. The metric properties, such as the degree of curvature, may change with different views, and are less likely to be useful in object recognition. It is shown that in a model of invariant visual object recognition in the ventral visual stream, VisNet, non-accidental properties are encoded much more than metric properties by neurons. Moreover, it is shown how with the temporal trace rule training in VisNet, non-accidental properties of objects become encoded by neurons, and how metric properties are treated invariantly. We also show how VisNet can generalize between different objects if they have the same non-accidental property, because the metric properties are likely to overlap. VisNet is a 4-layer unsupervised model of visual object recognition trained by competitive learning that utilizes a temporal trace learning rule to implement the learning of invariance using views that occur close together in time. A second crucial property of this model of object recognition is, when neurons in the level corresponding to the inferior temporal visual cortex respond selectively to objects, whether neurons in the intermediate layers can respond to combinations of features that may be parts of two or more objects. In an investigation using the four sides of a square presented in every possible combination, it was shown that even though different layer 4 neurons are tuned to encode each feature or feature combination orthogonally, neurons in the intermediate layers can respond to features or feature combinations present is several objects. This property is an important part of the way in which high capacity can be achieved in the four-layer ventral visual cortical pathway. These findings concerning non-accidental properties and the use of neurons in

  8. Representations of Shape in Object Recognition and Long-Term Visual Memory

    DTIC Science & Technology

    1993-02-11

    in anything other than linguistic terms ( Biederman , 1987 , for example). STATUS 1. Viewpoint-Dependent Features in Object Representation Tarr and...is object- based orientation-independent representations sufficient for "basic-level" categorization ( Biederman , 1987 ; Corballis, 1988). Alternatively...space. REFERENCES Biederman , I. ( 1987 ). Recognition-by-components: A theory of human image understanding. Psychological Review, 94,115-147. Cooper, L

  9. Using Markov Chains and Multi-Objective Optimization for Energy-Efficient Context Recognition

    PubMed Central

    Janko, Vito

    2017-01-01

    The recognition of the user’s context with wearable sensing systems is a common problem in ubiquitous computing. However, the typically small battery of such systems often makes continuous recognition impractical. The strain on the battery can be reduced if the sensor setting is adapted to each context. We propose a method that efficiently finds near-optimal sensor settings for each context. It uses Markov chains to simulate the behavior of the system in different configurations and the multi-objective genetic algorithm to find a set of good non-dominated configurations. The method was evaluated on three real-life datasets and found good trade-offs between the system’s energy expenditure and the system’s accuracy. One of the solutions, for example, consumed five-times less energy than the default one, while sacrificing only two percentage points of accuracy. PMID:29286301

  10. Multiple degree of freedom object recognition using optical relational graph decision nets

    NASA Technical Reports Server (NTRS)

    Casasent, David P.; Lee, Andrew J.

    1988-01-01

    Multiple-degree-of-freedom object recognition concerns objects with no stable rest position with all scale, rotation, and aspect distortions possible. It is assumed that the objects are in a fairly benign background, so that feature extractors are usable. In-plane distortion invariance is provided by use of a polar-log coordinate transform feature space, and out-of-plane distortion invariance is provided by linear discriminant function design. Relational graph decision nets are considered for multiple-degree-of-freedom pattern recognition. The design of Fisher (1936) linear discriminant functions and synthetic discriminant function for use at the nodes of binary and multidecision nets is discussed. Case studies are detailed for two-class and multiclass problems. Simulation results demonstrate the robustness of the processors to quantization of the filter coefficients and to noise.

  11. Impaired recognition of faces and objects in dyslexia: Evidence for ventral stream dysfunction?

    PubMed

    Sigurdardottir, Heida Maria; Ívarsson, Eysteinn; Kristinsdóttir, Kristjana; Kristjánsson, Árni

    2015-09-01

    The objective of this study was to establish whether or not dyslexics are impaired at the recognition of faces and other complex nonword visual objects. This would be expected based on a meta-analysis revealing that children and adult dyslexics show functional abnormalities within the left fusiform gyrus, a brain region high up in the ventral visual stream, which is thought to support the recognition of words, faces, and other objects. 20 adult dyslexics (M = 29 years) and 20 matched typical readers (M = 29 years) participated in the study. One dyslexic-typical reader pair was excluded based on Adult Reading History Questionnaire scores and IS-FORM reading scores. Performance was measured on 3 high-level visual processing tasks: the Cambridge Face Memory Test, the Vanderbilt Holistic Face Processing Test, and the Vanderbilt Expertise Test. People with dyslexia are impaired in their recognition of faces and other visually complex objects. Their holistic processing of faces appears to be intact, suggesting that dyslexics may instead be specifically impaired at part-based processing of visual objects. The difficulty that people with dyslexia experience with reading might be the most salient manifestation of a more general high-level visual deficit. (c) 2015 APA, all rights reserved).

  12. Neural Representations that Support Invariant Object Recognition

    PubMed Central

    Goris, Robbe L. T.; Op de Beeck, Hans P.

    2008-01-01

    Neural mechanisms underlying invariant behaviour such as object recognition are not well understood. For brain regions critical for object recognition, such as inferior temporal cortex (ITC), there is now ample evidence indicating that single cells code for many stimulus aspects, implying that only a moderate degree of invariance is present. However, recent theoretical and empirical work seems to suggest that integrating responses of multiple non-invariant units may produce invariant representations at population level. We provide an explicit test for the hypothesis that a linear read-out mechanism of a pool of units resembling ITC neurons may achieve invariant performance in an identification task. A linear classifier was trained to decode a particular value in a 2-D stimulus space using as input the response pattern across a population of units. Only one dimension was relevant for the task, and the stimulus location on the irrelevant dimension (ID) was kept constant during training. In a series of identification tests, the stimulus location on the relevant dimension (RD) and ID was manipulated, yielding estimates for both the level of sensitivity and tolerance reached by the network. We studied the effects of several single-cell characteristics as well as population characteristics typically considered in the literature, but found little support for the hypothesis. While the classifier averages out effects of idiosyncratic tuning properties and inter-unit variability, its invariance is very much determined by the (hypothetical) ‘average’ neuron. Consequently, even at population level there exists a fundamental trade-off between selectivity and tolerance, and invariant behaviour does not emerge spontaneously. PMID:19242556

  13. Real Time Large Memory Optical Pattern Recognition.

    DTIC Science & Technology

    1984-06-01

    AD-Ri58 023 REAL TIME LARGE MEMORY OPTICAL PATTERN RECOGNITION(U) - h ARMY MISSILE COMMAND REDSTONE ARSENAL AL RESEARCH DIRECTORATE D A GREGORY JUN...TECHNICAL REPORT RR-84-9 Ln REAL TIME LARGE MEMORY OPTICAL PATTERN RECOGNITION Don A. Gregory Research Directorate US Army Missile Laboratory JUNE 1984 L...RR-84-9 , ___/_ _ __ _ __ _ __ _ __"__ _ 4. TITLE (and Subtitle) S. TYPE OF REPORT & PERIOD COVERED Real Time Large Memory Optical Pattern Technical

  14. A rat in the sewer: How mental imagery interacts with object recognition

    PubMed Central

    Hamburger, Kai

    2018-01-01

    The role of mental imagery has been puzzling researchers for more than two millennia. Both positive and negative effects of mental imagery on information processing have been discussed. The aim of this work was to examine how mental imagery affects object recognition and associative learning. Based on different perceptual and cognitive accounts we tested our imagery-induced interaction hypothesis in a series of two experiments. According to that, mental imagery could lead to (1) a superior performance in object recognition and associative learning if these objects are imagery-congruent (semantically) and to (2) an inferior performance if these objects are imagery-incongruent. In the first experiment, we used a static environment and tested associative learning. In the second experiment, subjects encoded object information in a dynamic environment by means of a virtual sewer system. Our results demonstrate that subjects who received a role adoption task (by means of guided mental imagery) performed better when imagery-congruent objects were used and worse when imagery-incongruent objects were used. We finally discuss our findings also with respect to alternative accounts and plead for a multi-methodological approach for future research in order to solve this issue. PMID:29590161

  15. A rat in the sewer: How mental imagery interacts with object recognition.

    PubMed

    Karimpur, Harun; Hamburger, Kai

    2018-01-01

    The role of mental imagery has been puzzling researchers for more than two millennia. Both positive and negative effects of mental imagery on information processing have been discussed. The aim of this work was to examine how mental imagery affects object recognition and associative learning. Based on different perceptual and cognitive accounts we tested our imagery-induced interaction hypothesis in a series of two experiments. According to that, mental imagery could lead to (1) a superior performance in object recognition and associative learning if these objects are imagery-congruent (semantically) and to (2) an inferior performance if these objects are imagery-incongruent. In the first experiment, we used a static environment and tested associative learning. In the second experiment, subjects encoded object information in a dynamic environment by means of a virtual sewer system. Our results demonstrate that subjects who received a role adoption task (by means of guided mental imagery) performed better when imagery-congruent objects were used and worse when imagery-incongruent objects were used. We finally discuss our findings also with respect to alternative accounts and plead for a multi-methodological approach for future research in order to solve this issue.

  16. Structured Kernel Dictionary Learning with Correlation Constraint for Object Recognition.

    PubMed

    Wang, Zhengjue; Wang, Yinghua; Liu, Hongwei; Zhang, Hao

    2017-06-21

    In this paper, we propose a new discriminative non-linear dictionary learning approach, called correlation constrained structured kernel KSVD, for object recognition. The objective function for dictionary learning contains a reconstructive term and a discriminative term. In the reconstructive term, signals are implicitly non-linearly mapped into a space, where a structured kernel dictionary, each sub-dictionary of which lies in the span of the mapped signals from the corresponding class, is established. In the discriminative term, by analyzing the classification mechanism, the correlation constraint is proposed in kernel form, constraining the correlations between different discriminative codes, and restricting the coefficient vectors to be transformed into a feature space, where the features are highly correlated inner-class and nearly independent between-classes. The objective function is optimized by the proposed structured kernel KSVD. During the classification stage, the specific form of the discriminative feature is needless to be known, while the inner product of the discriminative feature with kernel matrix embedded is available, and is suitable for a linear SVM classifier. Experimental results demonstrate that the proposed approach outperforms many state-of-the-art dictionary learning approaches for face, scene and synthetic aperture radar (SAR) vehicle target recognition.

  17. The Slow Developmental Time Course of Real-Time Spoken Word Recognition

    ERIC Educational Resources Information Center

    Rigler, Hannah; Farris-Trimble, Ashley; Greiner, Lea; Walker, Jessica; Tomblin, J. Bruce; McMurray, Bob

    2015-01-01

    This study investigated the developmental time course of spoken word recognition in older children using eye tracking to assess how the real-time processing dynamics of word recognition change over development. We found that 9-year-olds were slower to activate the target words and showed more early competition from competitor words than…

  18. A biologically plausible computational model for auditory object recognition.

    PubMed

    Larson, Eric; Billimoria, Cyrus P; Sen, Kamal

    2009-01-01

    Object recognition is a task of fundamental importance for sensory systems. Although this problem has been intensively investigated in the visual system, relatively little is known about the recognition of complex auditory objects. Recent work has shown that spike trains from individual sensory neurons can be used to discriminate between and recognize stimuli. Multiple groups have developed spike similarity or dissimilarity metrics to quantify the differences between spike trains. Using a nearest-neighbor approach the spike similarity metrics can be used to classify the stimuli into groups used to evoke the spike trains. The nearest prototype spike train to the tested spike train can then be used to identify the stimulus. However, how biological circuits might perform such computations remains unclear. Elucidating this question would facilitate the experimental search for such circuits in biological systems, as well as the design of artificial circuits that can perform such computations. Here we present a biologically plausible model for discrimination inspired by a spike distance metric using a network of integrate-and-fire model neurons coupled to a decision network. We then apply this model to the birdsong system in the context of song discrimination and recognition. We show that the model circuit is effective at recognizing individual songs, based on experimental input data from field L, the avian primary auditory cortex analog. We also compare the performance and robustness of this model to two alternative models of song discrimination: a model based on coincidence detection and a model based on firing rate.

  19. Crowded and Sparse Domains in Object Recognition: Consequences for Categorization and Naming

    ERIC Educational Resources Information Center

    Gale, Tim M.; Laws, Keith R.; Foley, Kerry

    2006-01-01

    Some models of object recognition propose that items from structurally crowded categories (e.g., living things) permit faster access to superordinate semantic information than structurally dissimilar categories (e.g., nonliving things), but slower access to individual object information when naming items. We present four experiments that utilize…

  20. Simple Learned Weighted Sums of Inferior Temporal Neuronal Firing Rates Accurately Predict Human Core Object Recognition Performance

    PubMed Central

    Hong, Ha; Solomon, Ethan A.; DiCarlo, James J.

    2015-01-01

    To go beyond qualitative models of the biological substrate of object recognition, we ask: can a single ventral stream neuronal linking hypothesis quantitatively account for core object recognition performance over a broad range of tasks? We measured human performance in 64 object recognition tests using thousands of challenging images that explore shape similarity and identity preserving object variation. We then used multielectrode arrays to measure neuronal population responses to those same images in visual areas V4 and inferior temporal (IT) cortex of monkeys and simulated V1 population responses. We tested leading candidate linking hypotheses and control hypotheses, each postulating how ventral stream neuronal responses underlie object recognition behavior. Specifically, for each hypothesis, we computed the predicted performance on the 64 tests and compared it with the measured pattern of human performance. All tested hypotheses based on low- and mid-level visually evoked activity (pixels, V1, and V4) were very poor predictors of the human behavioral pattern. However, simple learned weighted sums of distributed average IT firing rates exactly predicted the behavioral pattern. More elaborate linking hypotheses relying on IT trial-by-trial correlational structure, finer IT temporal codes, or ones that strictly respect the known spatial substructures of IT (“face patches”) did not improve predictive power. Although these results do not reject those more elaborate hypotheses, they suggest a simple, sufficient quantitative model: each object recognition task is learned from the spatially distributed mean firing rates (100 ms) of ∼60,000 IT neurons and is executed as a simple weighted sum of those firing rates. SIGNIFICANCE STATEMENT We sought to go beyond qualitative models of visual object recognition and determine whether a single neuronal linking hypothesis can quantitatively account for core object recognition behavior. To achieve this, we designed a

  1. Simple Learned Weighted Sums of Inferior Temporal Neuronal Firing Rates Accurately Predict Human Core Object Recognition Performance.

    PubMed

    Majaj, Najib J; Hong, Ha; Solomon, Ethan A; DiCarlo, James J

    2015-09-30

    To go beyond qualitative models of the biological substrate of object recognition, we ask: can a single ventral stream neuronal linking hypothesis quantitatively account for core object recognition performance over a broad range of tasks? We measured human performance in 64 object recognition tests using thousands of challenging images that explore shape similarity and identity preserving object variation. We then used multielectrode arrays to measure neuronal population responses to those same images in visual areas V4 and inferior temporal (IT) cortex of monkeys and simulated V1 population responses. We tested leading candidate linking hypotheses and control hypotheses, each postulating how ventral stream neuronal responses underlie object recognition behavior. Specifically, for each hypothesis, we computed the predicted performance on the 64 tests and compared it with the measured pattern of human performance. All tested hypotheses based on low- and mid-level visually evoked activity (pixels, V1, and V4) were very poor predictors of the human behavioral pattern. However, simple learned weighted sums of distributed average IT firing rates exactly predicted the behavioral pattern. More elaborate linking hypotheses relying on IT trial-by-trial correlational structure, finer IT temporal codes, or ones that strictly respect the known spatial substructures of IT ("face patches") did not improve predictive power. Although these results do not reject those more elaborate hypotheses, they suggest a simple, sufficient quantitative model: each object recognition task is learned from the spatially distributed mean firing rates (100 ms) of ∼60,000 IT neurons and is executed as a simple weighted sum of those firing rates. Significance statement: We sought to go beyond qualitative models of visual object recognition and determine whether a single neuronal linking hypothesis can quantitatively account for core object recognition behavior. To achieve this, we designed a

  2. HONTIOR - HIGHER-ORDER NEURAL NETWORK FOR TRANSFORMATION INVARIANT OBJECT RECOGNITION

    NASA Technical Reports Server (NTRS)

    Spirkovska, L.

    1994-01-01

    Neural networks have been applied in numerous fields, including transformation invariant object recognition, wherein an object is recognized despite changes in the object's position in the input field, size, or rotation. One of the more successful neural network methods used in invariant object recognition is the higher-order neural network (HONN) method. With a HONN, known relationships are exploited and the desired invariances are built directly into the architecture of the network, eliminating the need for the network to learn invariance to transformations. This results in a significant reduction in the training time required, since the network needs to be trained on only one view of each object, not on numerous transformed views. Moreover, one hundred percent accuracy is guaranteed for images characterized by the built-in distortions, providing noise is not introduced through pixelation. The program HONTIOR implements a third-order neural network having invariance to translation, scale, and in-plane rotation built directly into the architecture, Thus, for 2-D transformation invariance, the network needs only to be trained on just one view of each object. HONTIOR can also be used for 3-D transformation invariant object recognition by training the network only on a set of out-of-plane rotated views. Historically, the major drawback of HONNs has been that the size of the input field was limited to the memory required for the large number of interconnections in a fully connected network. HONTIOR solves this problem by coarse coding the input images (coding an image as a set of overlapping but offset coarser images). Using this scheme, large input fields (4096 x 4096 pixels) can easily be represented using very little virtual memory (30Mb). The HONTIOR distribution consists of three main programs. The first program contains the training and testing routines for a third-order neural network. The second program contains the same training and testing procedures as the

  3. Complementary Hemispheric Asymmetries in Object Naming and Recognition: A Voxel-Based Correlational Study

    ERIC Educational Resources Information Center

    Acres, K.; Taylor, K. I.; Moss, H. E.; Stamatakis, E. A.; Tyler, L. K.

    2009-01-01

    Cognitive neuroscientific research proposes complementary hemispheric asymmetries in naming and recognising visual objects, with a left temporal lobe advantage for object naming and a right temporal lobe advantage for object recognition. Specifically, it has been proposed that the left inferior temporal lobe plays a mediational role linking…

  4. Face Recognition Is Affected by Similarity in Spatial Frequency Range to a Greater Degree Than Within-Category Object Recognition

    ERIC Educational Resources Information Center

    Collin, Charles A.; Liu, Chang Hong; Troje, Nikolaus F.; McMullen, Patricia A.; Chaudhuri, Avi

    2004-01-01

    Previous studies have suggested that face identification is more sensitive to variations in spatial frequency content than object recognition, but none have compared how sensitive the 2 processes are to variations in spatial frequency overlap (SFO). The authors tested face and object matching accuracy under varying SFO conditions. Their results…

  5. Viewpoint dependence in the recognition of non-elongated familiar objects: testing the effects of symmetry, front-back axis, and familiarity.

    PubMed

    Niimi, Ryosuke; Yokosawa, Kazuhiko

    2009-01-01

    Visual recognition of three-dimensional (3-D) objects is relatively impaired for some particular views, called accidental views. For most familiar objects, the front and top views are considered to be accidental views. Previous studies have shown that foreshortening of the axes of elongation of objects in these views impairs recognition, but the influence of other possible factors is largely unknown. Using familiar objects without a salient axis of elongation, we found that a foreshortened symmetry plane of the object and low familiarity of the viewpoint accounted for the relatively worse recognition for front views and top views, independently of the effect of a foreshortened axis of elongation. We found no evidence that foreshortened front-back axes impaired recognition in front views. These results suggest that the viewpoint dependence of familiar object recognition is not a unitary phenomenon. The possible role of symmetry (either 2-D or 3-D) in familiar object recognition is also discussed.

  6. Examining Object Location and Object Recognition Memory in Mice

    PubMed Central

    Vogel-Ciernia, Annie; Wood, Marcelo A.

    2014-01-01

    Unit Introduction The ability to store and recall our life experiences defines a person's identity. Consequently, the loss of long-term memory is a particularly devastating part of a variety of cognitive disorders, diseases and injuries. There is a great need to develop therapeutics to treat memory disorders, and thus a variety of animal models and memory paradigms have been developed. Mouse models have been widely used both to study basic disease mechanisms and to evaluate potential drug targets for therapeutic development. The relative ease of genetic manipulation of Mus musculus has led to a wide variety of genetically altered mice that model cognitive disorders ranging from Alzheimer's disease to autism. Rodents, including mice, are particularly adept at encoding and remembering spatial relationships, and these long-term spatial memories are dependent on the medial temporal lobe of the brain. These brain regions are also some of the first and most heavily impacted in disorders of human memory including Alzheimer's disease. Consequently, some of the simplest and most commonly used tests of long-term memory in mice are those that examine memory for objects and spatial relationships. However, many of these tasks, such as Morris water maze and contextual fear conditioning, are dependent upon the encoding and retrieval of emotionally aversive and inherently stressful training events. While these types of memories are important, they do not reflect the typical day-to-day experiences or memories most commonly affected in human disease. In addition, stress hormone release alone can modulate memory and thus obscure or artificially enhance these types of tasks. To avoid these sorts of confounds, we and many others have utilized tasks testing animals’ memory for object location and novel object recognition. These tasks involve exploiting rodents’ innate preference for novelty, and are inherently not stressful. In this protocol we detail how memory for object location

  7. Rotation And Scale Invariant Object Recognition Using A Distributed Associative Memory

    NASA Astrophysics Data System (ADS)

    Wechsler, Harry; Zimmerman, George Lee

    1988-04-01

    This paper describes an approach to 2-dimensional object recognition. Complex-log conformal mapping is combined with a distributed associative memory to create a system which recognizes objects regardless of changes in rotation or scale. Recalled information from the memorized database is used to classify an object, reconstruct the memorized version of the object, and estimate the magnitude of changes in scale or rotation. The system response is resistant to moderate amounts of noise and occlusion. Several experiments, using real, gray scale images, are presented to show the feasibility of our approach.

  8. Learning Distance Functions for Exemplar-Based Object Recognition

    DTIC Science & Technology

    2007-08-08

    requires prior specific permission. Learning Distance Functions for Exemplar-Based Object Recognition by Andrea Lynn Frome B.S. ( Mary Washington...fantastic advisor and advocate when I was at Mary Washington College i and has since become a dear friend. Thank you, Dr. Bass, for continuing to stand...Antonio Torralba. 5 Chapter 1. Introduction 0 5 10 15 20 25 30 35 10 15 20 25 30 35 40 45 50 55 60 65 70 Number of training examples per class M ea n

  9. Enhancing Perception with Tactile Object Recognition in Adaptive Grippers for Human-Robot Interaction.

    PubMed

    Gandarias, Juan M; Gómez-de-Gabriel, Jesús M; García-Cerezo, Alfonso J

    2018-02-26

    The use of tactile perception can help first response robotic teams in disaster scenarios, where visibility conditions are often reduced due to the presence of dust, mud, or smoke, distinguishing human limbs from other objects with similar shapes. Here, the integration of the tactile sensor in adaptive grippers is evaluated, measuring the performance of an object recognition task based on deep convolutional neural networks (DCNNs) using a flexible sensor mounted in adaptive grippers. A total of 15 classes with 50 tactile images each were trained, including human body parts and common environment objects, in semi-rigid and flexible adaptive grippers based on the fin ray effect. The classifier was compared against the rigid configuration and a support vector machine classifier (SVM). Finally, a two-level output network has been proposed to provide both object-type recognition and human/non-human classification. Sensors in adaptive grippers have a higher number of non-null tactels (up to 37% more), with a lower mean of pressure values (up to 72% less) than when using a rigid sensor, with a softer grip, which is needed in physical human-robot interaction (pHRI). A semi-rigid implementation with 95.13% object recognition rate was chosen, even though the human/non-human classification had better results (98.78%) with a rigid sensor.

  10. Simulated Prosthetic Vision: The Benefits of Computer-Based Object Recognition and Localization.

    PubMed

    Macé, Marc J-M; Guivarch, Valérian; Denis, Grégoire; Jouffrais, Christophe

    2015-07-01

    Clinical trials with blind patients implanted with a visual neuroprosthesis showed that even the simplest tasks were difficult to perform with the limited vision restored with current implants. Simulated prosthetic vision (SPV) is a powerful tool to investigate the putative functions of the upcoming generations of visual neuroprostheses. Recent studies based on SPV showed that several generations of implants will be required before usable vision is restored. However, none of these studies relied on advanced image processing. High-level image processing could significantly reduce the amount of information required to perform visual tasks and help restore visuomotor behaviors, even with current low-resolution implants. In this study, we simulated a prosthetic vision device based on object localization in the scene. We evaluated the usability of this device for object recognition, localization, and reaching. We showed that a very low number of electrodes (e.g., nine) are sufficient to restore visually guided reaching movements with fair timing (10 s) and high accuracy. In addition, performance, both in terms of accuracy and speed, was comparable with 9 and 100 electrodes. Extraction of high level information (object recognition and localization) from video images could drastically enhance the usability of current visual neuroprosthesis. We suggest that this method-that is, localization of targets of interest in the scene-may restore various visuomotor behaviors. This method could prove functional on current low-resolution implants. The main limitation resides in the reliability of the vision algorithms, which are improving rapidly. Copyright © 2015 International Center for Artificial Organs and Transplantation and Wiley Periodicals, Inc.

  11. Visual object recognition for mobile tourist information systems

    NASA Astrophysics Data System (ADS)

    Paletta, Lucas; Fritz, Gerald; Seifert, Christin; Luley, Patrick; Almer, Alexander

    2005-03-01

    We describe a mobile vision system that is capable of automated object identification using images captured from a PDA or a camera phone. We present a solution for the enabling technology of outdoors vision based object recognition that will extend state-of-the-art location and context aware services towards object based awareness in urban environments. In the proposed application scenario, tourist pedestrians are equipped with GPS, W-LAN and a camera attached to a PDA or a camera phone. They are interested whether their field of view contains tourist sights that would point to more detailed information. Multimedia type data about related history, the architecture, or other related cultural context of historic or artistic relevance might be explored by a mobile user who is intending to learn within the urban environment. Learning from ambient cues is in this way achieved by pointing the device towards the urban sight, capturing an image, and consequently getting information about the object on site and within the focus of attention, i.e., the users current field of view.

  12. Learning Distance Functions for Exemplar-Based Object Recognition

    DTIC Science & Technology

    2007-01-01

    Learning Distance Functions for Exemplar-Based Object Recognition by Andrea Lynn Frome B.S. ( Mary Washington College) 1996 A dissertation submitted...advisor and advocate when I was at Mary Washington College i and has since become a dear friend. Thank you, Dr. Bass, for continuing to stand by my...Torralba. 5 Chapter 1. Introduction 0 5 10 15 20 25 30 35 10 15 20 25 30 35 40 45 50 55 60 65 70 Number of training examples per class M ea n re co

  13. Multivariate fMRI and Eye Tracking Reveal Differential Effects of Visual Interference on Recognition Memory Judgments for Objects and Scenes.

    PubMed

    O'Neil, Edward B; Watson, Hilary C; Dhillon, Sonya; Lobaugh, Nancy J; Lee, Andy C H

    2015-09-01

    Recent work has demonstrated that the perirhinal cortex (PRC) supports conjunctive object representations that aid object recognition memory following visual object interference. It is unclear, however, how these representations interact with other brain regions implicated in mnemonic retrieval and how congruent and incongruent interference influences the processing of targets and foils during object recognition. To address this, multivariate partial least squares was applied to fMRI data acquired during an interference match-to-sample task, in which participants made object or scene recognition judgments after object or scene interference. This revealed a pattern of activity sensitive to object recognition following congruent (i.e., object) interference that included PRC, prefrontal, and parietal regions. Moreover, functional connectivity analysis revealed a common pattern of PRC connectivity across interference and recognition conditions. Examination of eye movements during the same task in a separate study revealed that participants gazed more at targets than foils during correct object recognition decisions, regardless of interference congruency. By contrast, participants viewed foils more than targets for incorrect object memory judgments, but only after congruent interference. Our findings suggest that congruent interference makes object foils appear familiar and that a network of regions, including PRC, is recruited to overcome the effects of interference.

  14. The Dynamic Multisensory Engram: Neural Circuitry Underlying Crossmodal Object Recognition in Rats Changes with the Nature of Object Experience.

    PubMed

    Jacklin, Derek L; Cloke, Jacob M; Potvin, Alphonse; Garrett, Inara; Winters, Boyer D

    2016-01-27

    Rats, humans, and monkeys demonstrate robust crossmodal object recognition (CMOR), identifying objects across sensory modalities. We have shown that rats' performance of a spontaneous tactile-to-visual CMOR task requires functional integration of perirhinal (PRh) and posterior parietal (PPC) cortices, which seemingly provide visual and tactile object feature processing, respectively. However, research with primates has suggested that PRh is sufficient for multisensory object representation. We tested this hypothesis in rats using a modification of the CMOR task in which multimodal preexposure to the to-be-remembered objects significantly facilitates performance. In the original CMOR task, with no preexposure, reversible lesions of PRh or PPC produced patterns of impairment consistent with modality-specific contributions. Conversely, in the CMOR task with preexposure, PPC lesions had no effect, whereas PRh involvement was robust, proving necessary for phases of the task that did not require PRh activity when rats did not have preexposure; this pattern was supported by results from c-fos imaging. We suggest that multimodal preexposure alters the circuitry responsible for object recognition, in this case obviating the need for PPC contributions and expanding PRh involvement, consistent with the polymodal nature of PRh connections and results from primates indicating a key role for PRh in multisensory object representation. These findings have significant implications for our understanding of multisensory information processing, suggesting that the nature of an individual's past experience with an object strongly determines the brain circuitry involved in representing that object's multisensory features in memory. The ability to integrate information from multiple sensory modalities is crucial to the survival of organisms living in complex environments. Appropriate responses to behaviorally relevant objects are informed by integration of multisensory object features

  15. Emerging Object Representations in the Visual System Predict Reaction Times for Categorization

    PubMed Central

    Ritchie, J. Brendan; Tovar, David A.; Carlson, Thomas A.

    2015-01-01

    Recognizing an object takes just a fraction of a second, less than the blink of an eye. Applying multivariate pattern analysis, or “brain decoding”, methods to magnetoencephalography (MEG) data has allowed researchers to characterize, in high temporal resolution, the emerging representation of object categories that underlie our capacity for rapid recognition. Shortly after stimulus onset, object exemplars cluster by category in a high-dimensional activation space in the brain. In this emerging activation space, the decodability of exemplar category varies over time, reflecting the brain’s transformation of visual inputs into coherent category representations. How do these emerging representations relate to categorization behavior? Recently it has been proposed that the distance of an exemplar representation from a categorical boundary in an activation space is critical for perceptual decision-making, and that reaction times should therefore correlate with distance from the boundary. The predictions of this distance hypothesis have been born out in human inferior temporal cortex (IT), an area of the brain crucial for the representation of object categories. When viewed in the context of a time varying neural signal, the optimal time to “read out” category information is when category representations in the brain are most decodable. Here, we show that the distance from a decision boundary through activation space, as measured using MEG decoding methods, correlates with reaction times for visual categorization during the period of peak decodability. Our results suggest that the brain begins to read out information about exemplar category at the optimal time for use in choice behaviour, and support the hypothesis that the structure of the representation for objects in the visual system is partially constitutive of the decision process in recognition. PMID:26107634

  16. A rodent model for the study of invariant visual object recognition

    PubMed Central

    Zoccolan, Davide; Oertelt, Nadja; DiCarlo, James J.; Cox, David D.

    2009-01-01

    The human visual system is able to recognize objects despite tremendous variation in their appearance on the retina resulting from variation in view, size, lighting, etc. This ability—known as “invariant” object recognition—is central to visual perception, yet its computational underpinnings are poorly understood. Traditionally, nonhuman primates have been the animal model-of-choice for investigating the neuronal substrates of invariant recognition, because their visual systems closely mirror our own. Meanwhile, simpler and more accessible animal models such as rodents have been largely overlooked as possible models of higher-level visual functions, because their brains are often assumed to lack advanced visual processing machinery. As a result, little is known about rodents' ability to process complex visual stimuli in the face of real-world image variation. In the present work, we show that rats possess more advanced visual abilities than previously appreciated. Specifically, we trained pigmented rats to perform a visual task that required them to recognize objects despite substantial variation in their appearance, due to changes in size, view, and lighting. Critically, rats were able to spontaneously generalize to previously unseen transformations of learned objects. These results provide the first systematic evidence for invariant object recognition in rats and argue for an increased focus on rodents as models for studying high-level visual processing. PMID:19429704

  17. Infrared detection, recognition and identification of handheld objects

    NASA Astrophysics Data System (ADS)

    Adomeit, Uwe

    2012-10-01

    A main criterion for comparison and selection of thermal imagers for military applications is their nominal range performance. This nominal range performance is calculated for a defined task and standardized target and environmental conditions. The only standardization available to date is STANAG 4347. The target defined there is based on a main battle tank in front view. Because of modified military requirements, this target is no longer up-to-date. Today, different topics of interest are of interest, especially differentiation between friend and foe and identification of humans. There is no direct way to differentiate between friend and foe in asymmetric scenarios, but one clue can be that someone is carrying a weapon. This clue can be transformed in the observer tasks detection: a person is carrying or is not carrying an object, recognition: the object is a long / medium / short range weapon or civil equipment and identification: the object can be named (e. g. AK-47, M-4, G36, RPG7, Axe, Shovel etc.). These tasks can be assessed experimentally and from the results of such an assessment, a standard target for handheld objects may be derived. For a first assessment, a human carrying 13 different handheld objects in front of his chest was recorded at four different ranges with an IR-dual-band camera. From the recorded data, a perception experiment was prepared. It was conducted with 17 observers in a 13-alternative forced choice, unlimited observation time arrangement. The results of the test together with Minimum Temperature Difference Perceived measurements of the camera and temperature difference and critical dimension derived from the recorded imagery allowed defining a first standard target according to the above tasks. This standard target consist of 2.5 / 3.5 / 5 DRI line pairs on target, 0.24 m critical size and 1 K temperature difference. The values are preliminary and have to be refined in the future. Necessary are different aspect angles, different

  18. Dopamine D1 receptor stimulation modulates the formation and retrieval of novel object recognition memory: Role of the prelimbic cortex

    PubMed Central

    Pezze, Marie A.; Marshall, Hayley J.; Fone, Kevin C.F.; Cassaday, Helen J.

    2015-01-01

    Previous studies have shown that dopamine D1 receptor antagonists impair novel object recognition memory but the effects of dopamine D1 receptor stimulation remain to be determined. This study investigated the effects of the selective dopamine D1 receptor agonist SKF81297 on acquisition and retrieval in the novel object recognition task in male Wistar rats. SKF81297 (0.4 and 0.8 mg/kg s.c.) given 15 min before the sampling phase impaired novel object recognition evaluated 10 min or 24 h later. The same treatments also reduced novel object recognition memory tested 24 h after the sampling phase and when given 15 min before the choice session. These data indicate that D1 receptor stimulation modulates both the encoding and retrieval of object recognition memory. Microinfusion of SKF81297 (0.025 or 0.05 μg/side) into the prelimbic sub-region of the medial prefrontal cortex (mPFC) in this case 10 min before the sampling phase also impaired novel object recognition memory, suggesting that the mPFC is one important site mediating the effects of D1 receptor stimulation on visual recognition memory. PMID:26277743

  19. Image object recognition based on the Zernike moment and neural networks

    NASA Astrophysics Data System (ADS)

    Wan, Jianwei; Wang, Ling; Huang, Fukan; Zhou, Liangzhu

    1998-03-01

    This paper first give a comprehensive discussion about the concept of artificial neural network its research methods and the relations with information processing. On the basis of such a discussion, we expound the mathematical similarity of artificial neural network and information processing. Then, the paper presents a new method of image recognition based on invariant features and neural network by using image Zernike transform. The method not only has the invariant properties for rotation, shift and scale of image object, but also has good fault tolerance and robustness. Meanwhile, it is also compared with statistical classifier and invariant moments recognition method.

  20. Extraction of edge-based and region-based features for object recognition

    NASA Astrophysics Data System (ADS)

    Coutts, Benjamin; Ravi, Srinivas; Hu, Gongzhu; Shrikhande, Neelima

    1993-08-01

    One of the central problems of computer vision is object recognition. A catalogue of model objects is described as a set of features such as edges and surfaces. The same features are extracted from the scene and matched against the models for object recognition. Edges and surfaces extracted from the scenes are often noisy and imperfect. In this paper algorithms are described for improving low level edge and surface features. Existing edge extraction algorithms are applied to the intensity image to obtain edge features. Initial edges are traced by following directions of the current contour. These are improved by using corresponding depth and intensity information for decision making at branch points. Surface fitting routines are applied to the range image to obtain planar surface patches. An algorithm of region growing is developed that starts with a coarse segmentation and uses quadric surface fitting to iteratively merge adjacent regions into quadric surfaces based on approximate orthogonal distance regression. Surface information obtained is returned to the edge extraction routine to detect and remove fake edges. This process repeats until no more merging or edge improvement can take place. Both synthetic (with Gaussian noise) and real images containing multiple object scenes have been tested using the merging criteria. Results appeared quite encouraging.

  1. Activity and function recognition for moving and static objects in urban environments from wide-area persistent surveillance inputs

    NASA Astrophysics Data System (ADS)

    Levchuk, Georgiy; Bobick, Aaron; Jones, Eric

    2010-04-01

    In this paper, we describe results from experimental analysis of a model designed to recognize activities and functions of moving and static objects from low-resolution wide-area video inputs. Our model is based on representing the activities and functions using three variables: (i) time; (ii) space; and (iii) structures. The activity and function recognition is achieved by imposing lexical, syntactic, and semantic constraints on the lower-level event sequences. In the reported research, we have evaluated the utility and sensitivity of several algorithms derived from natural language processing and pattern recognition domains. We achieved high recognition accuracy for a wide range of activity and function types in the experiments using Electro-Optical (EO) imagery collected by Wide Area Airborne Surveillance (WAAS) platform.

  2. Exploring local regularities for 3D object recognition

    NASA Astrophysics Data System (ADS)

    Tian, Huaiwen; Qin, Shengfeng

    2016-11-01

    In order to find better simplicity measurements for 3D object recognition, a new set of local regularities is developed and tested in a stepwise 3D reconstruction method, including localized minimizing standard deviation of angles(L-MSDA), localized minimizing standard deviation of segment magnitudes(L-MSDSM), localized minimum standard deviation of areas of child faces (L-MSDAF), localized minimum sum of segment magnitudes of common edges (L-MSSM), and localized minimum sum of areas of child face (L-MSAF). Based on their effectiveness measurements in terms of form and size distortions, it is found that when two local regularities: L-MSDA and L-MSDSM are combined together, they can produce better performance. In addition, the best weightings for them to work together are identified as 10% for L-MSDSM and 90% for L-MSDA. The test results show that the combined usage of L-MSDA and L-MSDSM with identified weightings has a potential to be applied in other optimization based 3D recognition methods to improve their efficacy and robustness.

  3. How landmark suitability shapes recognition memory signals for objects in the medial temporal lobes.

    PubMed

    Martin, Chris B; Sullivan, Jacqueline A; Wright, Jessey; Köhler, Stefan

    2018-02-01

    A role of perirhinal cortex (PrC) in recognition memory for objects has been well established. Contributions of parahippocampal cortex (PhC) to this function, while documented, remain less well understood. Here, we used fMRI to examine whether the organization of item-based recognition memory signals across these two structures is shaped by object category, independent of any difference in representing episodic context. Guided by research suggesting that PhC plays a critical role in processing landmarks, we focused on three categories of objects that differ from each other in their landmark suitability as confirmed with behavioral ratings (buildings > trees > aircraft). Participants made item-based recognition-memory decisions for novel and previously studied objects from these categories, which were matched in accuracy. Multi-voxel pattern classification revealed category-specific item-recognition memory signals along the long axis of PrC and PhC, with no sharp functional boundaries between these structures. Memory signals for buildings were observed in the mid to posterior extent of PhC, signals for trees in anterior to posterior segments of PhC, and signals for aircraft in mid to posterior aspects of PrC and the anterior extent of PhC. Notably, item-based memory signals for the category with highest landmark suitability ratings were observed only in those posterior segments of PhC that also allowed for classification of landmark suitability of objects when memory status was held constant. These findings provide new evidence in support of the notion that item-based memory signals for objects are not limited to PrC, and that the organization of these signals along the longitudinal axis that crosses PrC and PhC can be captured with reference to landmark suitability. Copyright © 2017 Elsevier Inc. All rights reserved.

  4. Are Face and Object Recognition Independent? A Neurocomputational Modeling Exploration.

    PubMed

    Wang, Panqu; Gauthier, Isabel; Cottrell, Garrison

    2016-04-01

    Are face and object recognition abilities independent? Although it is commonly believed that they are, Gauthier et al. [Gauthier, I., McGugin, R. W., Richler, J. J., Herzmann, G., Speegle, M., & VanGulick, A. E. Experience moderates overlap between object and face recognition, suggesting a common ability. Journal of Vision, 14, 7, 2014] recently showed that these abilities become more correlated as experience with nonface categories increases. They argued that there is a single underlying visual ability, v, that is expressed in performance with both face and nonface categories as experience grows. Using the Cambridge Face Memory Test and the Vanderbilt Expertise Test, they showed that the shared variance between Cambridge Face Memory Test and Vanderbilt Expertise Test performance increases monotonically as experience increases. Here, we address why a shared resource across different visual domains does not lead to competition and to an inverse correlation in abilities? We explain this conundrum using our neurocomputational model of face and object processing ["The Model", TM, Cottrell, G. W., & Hsiao, J. H. Neurocomputational models of face processing. In A. J. Calder, G. Rhodes, M. Johnson, & J. Haxby (Eds.), The Oxford handbook of face perception. Oxford, UK: Oxford University Press, 2011]. We model the domain general ability v as the available computational resources (number of hidden units) in the mapping from input to label and experience as the frequency of individual exemplars in an object category appearing during network training. Our results show that, as in the behavioral data, the correlation between subordinate level face and object recognition accuracy increases as experience grows. We suggest that different domains do not compete for resources because the relevant features are shared between faces and objects. The essential power of experience is to generate a "spreading transform" for faces (separating them in representational space) that

  5. Higher-order neural network software for distortion invariant object recognition

    NASA Technical Reports Server (NTRS)

    Reid, Max B.; Spirkovska, Lilly

    1991-01-01

    The state-of-the-art in pattern recognition for such applications as automatic target recognition and industrial robotic vision relies on digital image processing. We present a higher-order neural network model and software which performs the complete feature extraction-pattern classification paradigm required for automatic pattern recognition. Using a third-order neural network, we demonstrate complete, 100 percent accurate invariance to distortions of scale, position, and in-plate rotation. In a higher-order neural network, feature extraction is built into the network, and does not have to be learned. Only the relatively simple classification step must be learned. This is key to achieving very rapid training. The training set is much smaller than with standard neural network software because the higher-order network only has to be shown one view of each object to be learned, not every possible view. The software and graphical user interface run on any Sun workstation. Results of the use of the neural software in autonomous robotic vision systems are presented. Such a system could have extensive application in robotic manufacturing.

  6. Novel object recognition ability in female mice following exposure to nanoparticle-rich diesel exhaust

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Win-Shwe, Tin-Tin, E-mail: tin.tin.win.shwe@nies.go.jp; Fujimaki, Hidekazu; Fujitani, Yuji

    2012-08-01

    Recently, our laboratory reported that exposure to nanoparticle-rich diesel exhaust (NRDE) for 3 months impaired hippocampus-dependent spatial learning ability and up-regulated the expressions of memory function-related genes in the hippocampus of female mice. However, whether NRDE affects the hippocampus-dependent non-spatial learning ability and the mechanism of NRDE-induced neurotoxicity was unknown. Female BALB/c mice were exposed to clean air, middle-dose NRDE (M-NRDE, 47 μg/m{sup 3}), high-dose NRDE (H-NRDE, 129 μg/m{sup 3}), or filtered H-NRDE (F-DE) for 3 months. We then investigated the effect of NRDE exposure on non-spatial learning ability and the expression of genes related to glutamate neurotransmission using amore » novel object recognition test and a real-time RT-PCR analysis, respectively. We also examined microglia marker Iba1 immunoreactivity in the hippocampus using immunohistochemical analyses. Mice exposed to H-NRDE or F-DE could not discriminate between familiar and novel objects. The control and M-NRDE-exposed groups showed a significantly increased discrimination index, compared to the H-NRDE-exposed group. Although no significant changes in the expression levels of the NMDA receptor subunits were observed, the expression of glutamate transporter EAAT4 was decreased and that of glutamic acid decarboxylase GAD65 was increased in the hippocampus of H-NRDE-exposed mice, compared with the expression levels in control mice. We also found that microglia activation was prominent in the hippocampal area of the H-NRDE-exposed mice, compared with the other groups. These results indicated that exposure to NRDE for 3 months impaired the novel object recognition ability. The present study suggests that genes related to glutamate metabolism may be involved in the NRDE-induced neurotoxicity observed in the present mouse model. -- Highlights: ► The effects of nanoparticle-induced neurotoxicity remain unclear. ► We investigated the effect of exposure

  7. Detailed 3D representations for object recognition and modeling.

    PubMed

    Zia, M Zeeshan; Stark, Michael; Schiele, Bernt; Schindler, Konrad

    2013-11-01

    Geometric 3D reasoning at the level of objects has received renewed attention recently in the context of visual scene understanding. The level of geometric detail, however, is typically limited to qualitative representations or coarse boxes. This is linked to the fact that today's object class detectors are tuned toward robust 2D matching rather than accurate 3D geometry, encouraged by bounding-box-based benchmarks such as Pascal VOC. In this paper, we revisit ideas from the early days of computer vision, namely, detailed, 3D geometric object class representations for recognition. These representations can recover geometrically far more accurate object hypotheses than just bounding boxes, including continuous estimates of object pose and 3D wireframes with relative 3D positions of object parts. In combination with robust techniques for shape description and inference, we outperform state-of-the-art results in monocular 3D pose estimation. In a series of experiments, we analyze our approach in detail and demonstrate novel applications enabled by such an object class representation, such as fine-grained categorization of cars and bicycles, according to their 3D geometry, and ultrawide baseline matching.

  8. Recognition of abstract objects via neural oscillators: interaction among topological organization, associative memory and gamma band synchronization.

    PubMed

    Ursino, Mauro; Magosso, Elisa; Cuppini, Cristiano

    2009-02-01

    Synchronization of neural activity in the gamma band is assumed to play a significant role not only in perceptual processing, but also in higher cognitive functions. Here, we propose a neural network of Wilson-Cowan oscillators to simulate recognition of abstract objects, each represented as a collection of four features. Features are ordered in topological maps of oscillators connected via excitatory lateral synapses, to implement a similarity principle. Experience on previous objects is stored in long-range synapses connecting the different topological maps, and trained via timing dependent Hebbian learning (previous knowledge principle). Finally, a downstream decision network detects the presence of a reliable object representation, when all features are oscillating in synchrony. Simulations performed giving various simultaneous objects to the network (from 1 to 4), with some missing and/or modified properties suggest that the network can reconstruct objects, and segment them from the other simultaneously present objects, even in case of deteriorated information, noise, and moderate correlation among the inputs (one common feature). The balance between sensitivity and specificity depends on the strength of the Hebbian learning. Achieving a correct reconstruction in all cases, however, requires ad hoc selection of the oscillation frequency. The model represents an attempt to investigate the interactions among topological maps, autoassociative memory, and gamma-band synchronization, for recognition of abstract objects.

  9. Tactile agnosia. Underlying impairment and implications for normal tactile object recognition.

    PubMed

    Reed, C L; Caselli, R J; Farah, M J

    1996-06-01

    In a series of experimental investigations of a subject with a unilateral impairment of tactile object recognition without impaired tactile sensation, several issues were addressed. First, is tactile agnosia secondary to a general impairment of spatial cognition? On tests of spatial ability, including those directed at the same spatial integration process assumed to be taxed by tactile object recognition, the subject performed well, implying a more specific impairment of high level, modality specific tactile perception. Secondly, within the realm of high level tactile perception, is there a distinction between the ability to derive shape ('what') and spatial ('where') information? Our testing showed an impairment confined to shape perception. Thirdly, what aspects of shape perception are impaired in tactile agnosia? Our results indicate that despite accurate encoding of metric length and normal manual exploration strategies, the ability tactually to perceive objects with the impaired hand, deteriorated as the complexity of shape increased. In addition, asymmetrical performance was not found for other body surfaces (e.g. her feet). Our results suggest that tactile shape perception can be disrupted independent of general spatial ability, tactile spatial ability, manual shape exploration, or even the precise perception of metric length in the tactile modality.

  10. BDNF Expression in Perirhinal Cortex is Associated with Exercise-Induced Improvement in Object Recognition Memory

    PubMed Central

    Hopkins, Michael E.; Bucci, David J.

    2010-01-01

    Physical exercise induces widespread neurobiological adaptations and improves learning and memory. Most research in this field has focused on hippocampus-based spatial tasks and changes in brain-derived neurotrophic factor (BDNF) as a putative substrate underlying exercise-induced cognitive improvements. Chronic exercise can also be anxiolytic and causes adaptive changes in stress reactivity. The present study employed a perirhinal cortex-dependent object recognition task as well as the elevated plus maze to directly test for interactions between the cognitive and anxiolytic effects of exercise in male Long Evans rats. Hippocampal and perirhinal cortex tissue was collected to determine whether the relationship between BDNF and cognitive performance extends to this non-spatial and non-hippocampal-dependent task. We also examined whether the cognitive improvements persisted once the exercise regimen was terminated. Our data indicate that 4 weeks of voluntary exercise every-other-day improved object recognition memory. Importantly, BDNF expression in the perirhinal cortex of exercising rats was strongly correlated with object recognition memory. Exercise also decreased anxiety-like behavior, however there was no evidence to support a relationship between anxiety-like behavior and performance on the novel object recognition task. There was a trend for a negative relationship between anxiety-like behavior and hippocampal BDNF. Neither the cognitive improvements nor the relationship between cognitive function and perirhinal BDNF levels persisted after 2 weeks of inactivity. These are the first data demonstrating that region-specific changes in BDNF protein levels are correlated with exercise-induced improvements in non-spatial memory, mediated by structures outside the hippocampus and are consistent with the theory that, with regard to object recognition, the anxiolytic and cognitive effects of exercise may be mediated through separable mechanisms. PMID:20601027

  11. Memory consolidation and expression of object recognition are susceptible to retroactive interference.

    PubMed

    Villar, María Eugenia; Martinez, María Cecilia; Lopes da Cunha, Pamela; Ballarini, Fabricio; Viola, Haydee

    2017-02-01

    With the aim of analyzing if object recognition long-term memory (OR-LTM) formation is susceptible to retroactive interference (RI), we submitted rats to sequential sample sessions using the same arena but changing the identity of a pair of objects placed in it. Separate groups of animals were tested in the arena in order to evaluate the LTM for these objects. Our results suggest that OR-LTM formation was retroactively interfered within a critical time window by the exploration of a new, but not familiar, object. This RI acted on the consolidation of the object explored in the first sample session because its OR-STM measured 3h after training was not affected, whereas the OR-LTM measured at 24h was impaired. This sample session also impaired the expression of OR memory when it took place before the test. Moreover, local inactivation of the dorsal Hippocampus (Hp) or the medial Prefrontal Cortex (mPFC) previous to the exploration of the second pair of objects impaired their consolidation restoring the LTM for the objects explored in the first session. This data suggests that both brain regions are involved in the processing of OR-memory and also that if those regions are engaged in another process before finishing the first consolidation process its LTM will be impaired by RI. Copyright © 2016 Elsevier Inc. All rights reserved.

  12. Repetition priming of face recognition in a serial choice reaction-time task.

    PubMed

    Roberts, T; Bruce, V

    1989-05-01

    Marshall & Walker (1987) found that pictorial stimuli yield visual priming that is disrupted by an unpredictable visual event in the response-stimulus interval. They argue that visual stimuli are represented in memory in the form of distinct visual and object codes. Bruce & Young (1986) propose similar pictorial, structural and semantic codes which mediate the recognition of faces, yet repetition priming results obtained with faces as stimuli (Bruce & Valentine, 1985), and with objects (Warren & Morton, 1982) are quite different from those of Marshall & Walker (1987), in the sense that recognition is facilitated by pictures presented 20 minutes earlier. The experiment reported here used different views of familiar and unfamiliar faces as stimuli in a serial choice reaction-time task and found that, with identical pictures, repetition priming survives and intervening item requiring a response, with both familiar and unfamiliar faces. Furthermore, with familiar faces such priming was present even when the view of the prime was different from the target. The theoretical implications of these results are discussed.

  13. Optical Pattern Recognition

    NASA Astrophysics Data System (ADS)

    Yu, Francis T. S.; Jutamulia, Suganda

    2008-10-01

    Contributors; Preface; 1. Pattern recognition with optics Francis T. S. Yu and Don A. Gregory; 2. Hybrid neural networks for nonlinear pattern recognition Taiwei Lu; 3. Wavelets, optics, and pattern recognition Yao Li and Yunglong Sheng; 4. Applications of the fractional Fourier transform to optical pattern recognition David Mendlovic, Zeev Zalesky and Haldum M. Oxaktas; 5. Optical implementation of mathematical morphology Tien-Hsin Chao; 6. Nonlinear optical correlators with improved discrimination capability for object location and recognition Leonid P. Yaroslavsky; 7. Distortion-invariant quadratic filters Gregory Gheen; 8. Composite filter synthesis as applied to pattern recognition Shizhou Yin and Guowen Lu; 9. Iterative procedures in electro-optical pattern recognition Joseph Shamir; 10. Optoelectronic hybrid system for three-dimensional object pattern recognition Guoguang Mu, Mingzhe Lu and Ying Sun; 11. Applications of photrefractive devices in optical pattern recognition Ziangyang Yang; 12. Optical pattern recognition with microlasers Eung-Gi Paek; 13. Optical properties and applications of bacteriorhodopsin Q. Wang Song and Yu-He Zhang; 14. Liquid-crystal spatial light modulators Aris Tanone and Suganda Jutamulia; 15. Representations of fully complex functions on real-time spatial light modulators Robert W. Cohn and Laurence G. Hassbrook; Index.

  14. Developmental Trajectories of Part-Based and Configural Object Recognition in Adolescence

    ERIC Educational Resources Information Center

    Juttner, Martin; Wakui, Elley; Petters, Dean; Kaur, Surinder; Davidoff, Jules

    2013-01-01

    Three experiments assessed the development of children's part and configural (part-relational) processing in object recognition during adolescence. In total, 312 school children aged 7-16 years and 80 adults were tested in 3-alternative forced choice (3-AFC) tasks. They judged the correct appearance of upright and inverted presented familiar…

  15. A Neuromorphic Architecture for Object Recognition and Motion Anticipation Using Burst-STDP

    PubMed Central

    Balduzzi, David; Tononi, Giulio

    2012-01-01

    In this work we investigate the possibilities offered by a minimal framework of artificial spiking neurons to be deployed in silico. Here we introduce a hierarchical network architecture of spiking neurons which learns to recognize moving objects in a visual environment and determine the correct motor output for each object. These tasks are learned through both supervised and unsupervised spike timing dependent plasticity (STDP). STDP is responsible for the strengthening (or weakening) of synapses in relation to pre- and post-synaptic spike times and has been described as a Hebbian paradigm taking place both in vitro and in vivo. We utilize a variation of STDP learning, called burst-STDP, which is based on the notion that, since spikes are expensive in terms of energy consumption, then strong bursting activity carries more information than single (sparse) spikes. Furthermore, this learning algorithm takes advantage of homeostatic renormalization, which has been hypothesized to promote memory consolidation during NREM sleep. Using this learning rule, we design a spiking neural network architecture capable of object recognition, motion detection, attention towards important objects, and motor control outputs. We demonstrate the abilities of our design in a simple environment with distractor objects, multiple objects moving concurrently, and in the presence of noise. Most importantly, we show how this neural network is capable of performing these tasks using a simple leaky-integrate-and-fire (LIF) neuron model with binary synapses, making it fully compatible with state-of-the-art digital neuromorphic hardware designs. As such, the building blocks and learning rules presented in this paper appear promising for scalable fully neuromorphic systems to be implemented in hardware chips. PMID:22615855

  16. Data-driven indexing mechanism for the recognition of polyhedral objects

    NASA Astrophysics Data System (ADS)

    McLean, Stewart; Horan, Peter; Caelli, Terry M.

    1992-02-01

    This paper is concerned with the problem of searching large model databases. To date, most object recognition systems have concentrated on the problem of matching using simple searching algorithms. This is quite acceptable when the number of object models is small. However, in the future, general purpose computer vision systems will be required to recognize hundreds or perhaps thousands of objects and, in such circumstances, efficient searching algorithms will be needed. The problem of searching a large model database is one which must be addressed if future computer vision systems are to be at all effective. In this paper we present a method we call data-driven feature-indexed hypothesis generation as one solution to the problem of searching large model databases.

  17. Fast and efficient indexing approach for object recognition

    NASA Astrophysics Data System (ADS)

    Hefnawy, Alaa; Mashali, Samia A.; Rashwan, Mohsen; Fikri, Magdi

    1999-08-01

    This paper introduces a fast and efficient indexing approach for both 2D and 3D model-based object recognition in the presence of rotation, translation, and scale variations of objects. The indexing entries are computed after preprocessing the data by Haar wavelet decomposition. The scheme is based on a unified image feature detection approach based on Zernike moments. A set of low level features, e.g. high precision edges, gray level corners, are estimated by a set of orthogonal Zernike moments, calculated locally around every image point. A high dimensional, highly descriptive indexing entries are then calculated based on the correlation of these local features and employed for fast access to the model database to generate hypotheses. A list of the most candidate models is then presented by evaluating the hypotheses. Experimental results are included to demonstrate the effectiveness of the proposed indexing approach.

  18. Impact of a voice recognition system on report cycle time and radiologist reading time

    NASA Astrophysics Data System (ADS)

    Melson, David L.; Brophy, Robert; Blaine, G. James; Jost, R. Gilbert; Brink, Gary S.

    1998-07-01

    Because of its exciting potential to improve clinical service, as well as reduce costs, a voice recognition system for radiological dictation was recently installed at our institution. This system will be clinically successful if it dramatically reduces radiology report turnaround time without substantially affecting radiologist dictation and editing time. This report summarizes an observer study currently under way in which radiologist reporting times using the traditional transcription system and the voice recognition system are compared. Four radiologists are observed interpreting portable intensive care unit (ICU) chest examinations at a workstation in the chest reading area. Data are recorded with the radiologists using the transcription system and using the voice recognition system. The measurements distinguish between time spent performing clerical tasks and time spent actually dictating the report. Editing time and the number of corrections made are recorded. Additionally, statistics are gathered to assess the voice recognition system's impact on the report cycle time -- the time from report dictation to availability of an edited and finalized report -- and the length of reports.

  19. The Role of Sensory-Motor Information in Object Recognition: Evidence from Category-Specific Visual Agnosia

    ERIC Educational Resources Information Center

    Wolk, D.A.; Coslett, H.B.; Glosser, G.

    2005-01-01

    The role of sensory-motor representations in object recognition was investigated in experiments involving AD, a patient with mild visual agnosia who was impaired in the recognition of visually presented living as compared to non-living entities. AD named visually presented items for which sensory-motor information was available significantly more…

  20. Somatostatin Signaling in Neuronal Cilia Is Criticalfor Object Recognition Memory

    PubMed Central

    Einstein, Emily B.; Patterson, Carlyn A.; Hon, Beverly J.; Regan, Kathleen A.; Reddi, Jyoti; Melnikoff, David E.; Mateer, Marcus J.; Schulz, Stefan; Johnson, Brian N.

    2010-01-01

    Most neurons possess a single, nonmotile cilium that projects out from the cell surface. These microtubule-based organelles are important in brain development and neurogenesis; however, their function in mature neurons is unknown. Cilia express a complement of proteins distinct from other neuronal compartments, one of which is the somatostatin receptor subtype SST3. We show here that SST3 is critical for object recognition memory in mice. sst3 knock-out mice are severely impaired in discriminating novel objects, whereas they retain normal memory for object location. Further, systemic injection of an SST3 antagonist (ACQ090) disrupts recall of familiar objects in wild-type mice. To examine mechanisms of SST3, we tested synaptic plasticity in CA1 hippocampus. Electrically evoked long-term potentiation (LTP) was normal in sst3 knock-out mice, while adenylyl cyclase/cAMP-mediated LTP was impaired. The SST3 antagonist also disrupted cAMP-mediated LTP. Basal cAMP levels in hippocampal lysate were reduced in sst3 knock-out mice compared with wild-type mice, while the forskolin-induced increase in cAMP levels was normal. The SST3 antagonist inhibited forskolin-stimulated cAMP increases, whereas the SST3 agonist L-796,778 increased basal cAMP levels in hippocampal slices but not hippocampal lysate. Our results show that somatostatin signaling in neuronal cilia is critical for recognition memory and suggest that the cAMP pathway is a conserved signaling motif in cilia. Neuronal cilia therefore represent a novel nonsynaptic compartment crucial for signaling involved in a specific form of synaptic plasticity and in novelty detection. PMID:20335466

  1. Face Memory and Object Recognition in Children with High-Functioning Autism or Asperger Syndrome and in Their Parents

    ERIC Educational Resources Information Center

    Kuusikko-Gauffin, Sanna; Jansson-Verkasalo, Eira; Carter, Alice; Pollock-Wurman, Rachel; Jussila, Katja; Mattila, Marja-Leena; Rahko, Jukka; Ebeling, Hanna; Pauls, David; Moilanen, Irma

    2011-01-01

    Children with Autism Spectrum Disorders (ASDs) have reported to have impairments in face, recognition and face memory, but intact object recognition and object memory. Potential abnormalities, in these fields at the family level of high-functioning children with ASD remains understudied despite, the ever-mounting evidence that ASDs are genetic and…

  2. Semantic attributes are encoded in human electrocorticographic signals during visual object recognition.

    PubMed

    Rupp, Kyle; Roos, Matthew; Milsap, Griffin; Caceres, Carlos; Ratto, Christopher; Chevillet, Mark; Crone, Nathan E; Wolmetz, Michael

    2017-03-01

    Non-invasive neuroimaging studies have shown that semantic category and attribute information are encoded in neural population activity. Electrocorticography (ECoG) offers several advantages over non-invasive approaches, but the degree to which semantic attribute information is encoded in ECoG responses is not known. We recorded ECoG while patients named objects from 12 semantic categories and then trained high-dimensional encoding models to map semantic attributes to spectral-temporal features of the task-related neural responses. Using these semantic attribute encoding models, untrained objects were decoded with accuracies comparable to whole-brain functional Magnetic Resonance Imaging (fMRI), and we observed that high-gamma activity (70-110Hz) at basal occipitotemporal electrodes was associated with specific semantic dimensions (manmade-animate, canonically large-small, and places-tools). Individual patient results were in close agreement with reports from other imaging modalities on the time course and functional organization of semantic processing along the ventral visual pathway during object recognition. The semantic attribute encoding model approach is critical for decoding objects absent from a training set, as well as for studying complex semantic encodings without artificially restricting stimuli to a small number of semantic categories. Copyright © 2017 The Authors. Published by Elsevier Inc. All rights reserved.

  3. Deep Neural Networks Rival the Representation of Primate IT Cortex for Core Visual Object Recognition

    PubMed Central

    Cadieu, Charles F.; Hong, Ha; Yamins, Daniel L. K.; Pinto, Nicolas; Ardila, Diego; Solomon, Ethan A.; Majaj, Najib J.; DiCarlo, James J.

    2014-01-01

    The primate visual system achieves remarkable visual object recognition performance even in brief presentations, and under changes to object exemplar, geometric transformations, and background variation (a.k.a. core visual object recognition). This remarkable performance is mediated by the representation formed in inferior temporal (IT) cortex. In parallel, recent advances in machine learning have led to ever higher performing models of object recognition using artificial deep neural networks (DNNs). It remains unclear, however, whether the representational performance of DNNs rivals that of the brain. To accurately produce such a comparison, a major difficulty has been a unifying metric that accounts for experimental limitations, such as the amount of noise, the number of neural recording sites, and the number of trials, and computational limitations, such as the complexity of the decoding classifier and the number of classifier training examples. In this work, we perform a direct comparison that corrects for these experimental limitations and computational considerations. As part of our methodology, we propose an extension of “kernel analysis” that measures the generalization accuracy as a function of representational complexity. Our evaluations show that, unlike previous bio-inspired models, the latest DNNs rival the representational performance of IT cortex on this visual object recognition task. Furthermore, we show that models that perform well on measures of representational performance also perform well on measures of representational similarity to IT, and on measures of predicting individual IT multi-unit responses. Whether these DNNs rely on computational mechanisms similar to the primate visual system is yet to be determined, but, unlike all previous bio-inspired models, that possibility cannot be ruled out merely on representational performance grounds. PMID:25521294

  4. Real-Time Hand Posture Recognition Using a Range Camera

    NASA Astrophysics Data System (ADS)

    Lahamy, Herve

    The basic goal of human computer interaction is to improve the interaction between users and computers by making computers more usable and receptive to the user's needs. Within this context, the use of hand postures in replacement of traditional devices such as keyboards, mice and joysticks is being explored by many researchers. The goal is to interpret human postures via mathematical algorithms. Hand posture recognition has gained popularity in recent years, and could become the future tool for humans to interact with computers or virtual environments. An exhaustive description of the frequently used methods available in literature for hand posture recognition is provided. It focuses on the different types of sensors and data used, the segmentation and tracking methods, the features used to represent the hand postures as well as the classifiers considered in the recognition process. Those methods are usually presented as highly robust with a recognition rate close to 100%. However, a couple of critical points necessary for a successful real-time hand posture recognition system require major improvement. Those points include the features used to represent the hand segment, the number of postures simultaneously recognizable, the invariance of the features with respect to rotation, translation and scale and also the behavior of the classifiers against non-perfect hand segments for example segments including part of the arm or missing part of the palm. A 3D time-of-flight camera named SR4000 has been chosen to develop a new methodology because of its capability to provide in real-time and at high frame rate 3D information on the scene imaged. This sensor has been described and evaluated for its capability for capturing in real-time a moving hand. A new recognition method that uses the 3D information provided by the range camera to recognize hand postures has been proposed. The different steps of this methodology including the segmentation, the tracking, the hand

  5. Behavioral methods for the study of the Ras-ERK pathway in memory formation and consolidation: passive avoidance and novel object recognition tests.

    PubMed

    d'Isa, Raffaele; Brambilla, Riccardo; Fasano, Stefania

    2014-01-01

    Memory is a high-level brain function that enables organisms to adapt their behavioral responses to the environment, hence increasing their probability of survival. The Ras-ERK pathway is a key molecular intracellular signalling cascade for memory consolidation. In this chapter we will describe two main one-trial behavioral tests commonly used in the field of memory research in order to assess the role of Ras-ERK signalling in long-term memory: passive avoidance and object recognition. Passive avoidance (PA) is a fear-motivated instrumental learning task, designed by Jarvik and Essman in 1960, in which animals learn to refrain from emitting a behavioral response that has previously been associated with a punishment. We will describe here the detailed protocol and show some examples of how PA can reveal impairments or enhancements in memory consolidation following loss or gain of function genetic manipulations of the Ras-ERK pathway. The phenotypes of global mutants as Ras-GRF1 KO, GENA53, and ERK1 KO mice, as well as of conditional region-specific mutants (striatal K-CREB mice), will be illustrated as examples. Novel object recognition (NOR), developed by Ennaceur and Delacour in 1988, is instead a more recent and highly ecological test, which relies on the natural tendency of rodents to spontaneously approach and explore novel objects, representing hence a useful non-stressful tool for the study of memory in animals without the employment of punishments or starvation/water restriction regimens. Careful indications will be given on how to select the positions for the novel object, in order to counterbalance for individual side preferences among mice during the training. Finally, the methods for calculating two learning indexes will be described. In addition to the classical discrimination index (DI) that measures the ability of an animal to discriminate between two different objects which are presented at the same time, we will describe the formula of a new index

  6. Mice deficient for striatal Vesicular Acetylcholine Transporter (VAChT) display impaired short-term but normal long-term object recognition memory.

    PubMed

    Palmer, Daniel; Creighton, Samantha; Prado, Vania F; Prado, Marco A M; Choleris, Elena; Winters, Boyer D

    2016-09-15

    Substantial evidence implicates Acetylcholine (ACh) in the acquisition of object memories. While most research has focused on the role of the cholinergic basal forebrain and its cortical targets, there are additional cholinergic networks that may contribute to object recognition. The striatum contains an independent cholinergic network comprised of interneurons. In the current study, we investigated the role of this cholinergic signalling in object recognition using mice deficient for Vesicular Acetylcholine Transporter (VAChT) within interneurons of the striatum. We tested whether these striatal VAChT(D2-Cre-flox/flox) mice would display normal short-term (5 or 15min retention delay) and long-term (3h retention delay) object recognition memory. In a home cage object recognition task, male and female VAChT(D2-Cre-flox/flox) mice were impaired selectively with a 15min retention delay. When tested on an object location task, VAChT(D2-Cre-flox/flox) mice displayed intact spatial memory. Finally, when object recognition was tested in a Y-shaped apparatus, designed to minimize the influence of spatial and contextual cues, only females displayed impaired recognition with a 5min retention delay, but when males were challenged with a 15min retention delay, they were also impaired; neither males nor females were impaired with the 3h delay. The pattern of results suggests that striatal cholinergic transmission plays a role in the short-term memory for object features, but not spatial location. Copyright © 2016 Elsevier B.V. All rights reserved.

  7. The Role of Fixation and Visual Attention in Object Recognition.

    DTIC Science & Technology

    1995-01-01

    computers", Technical Report, Aritificial Intelligence Lab, M.I. T., AI-Memo-915, June 1986. [29] D.P. Huttenlocher and S.Ullman, "Object Recognition Using...attention", Technical Report, Aritificial Intelligence Lab, M.I. T., AI-memo-770, Jan 1984. [35] E.Krotkov, K. Henriksen and R. Kories, "Stereo...MIT Artificial Intelligence Laboratory [ PCTBTBimON STATEMENT X \\ Afipioved tor puciic reieo*«* \\ »?*•;.., jDi*tiibutK» U»lisut»d* 19951004

  8. Large-memory real-time multichannel multiplexed pattern recognition

    NASA Technical Reports Server (NTRS)

    Gregory, D. A.; Liu, H. K.

    1984-01-01

    The principle and experimental design of a real-time multichannel multiplexed optical pattern recognition system via use of a 25-focus dichromated gelatin holographic lens (hololens) are described. Each of the 25 foci of the hololens may have a storage and matched filtering capability approaching that of a single-lens correlator. If the space-bandwidth product of an input image is limited, as is true in most practical cases, the 25-focus hololens system has 25 times the capability of a single lens. Experimental results have shown that the interfilter noise is not serious. The system has already demonstrated the storage and recognition of over 70 matched filters - which is a larger capacity than any optical pattern recognition system reported to date.

  9. Endomorphin-1 attenuates Aβ42 induced impairment of novel object and object location recognition tasks in mice.

    PubMed

    Zhang, Rui-san; Xu, Hong-jiao; Jiang, Jin-hong; Han, Ren-wen; Chang, Min; Peng, Ya-li; Wang, Yuan; Wang, Rui

    2015-12-10

    A growing body of evidence suggests that the agglomeration of amyloid-β (Aβ) may be a trigger for Alzheimer׳s disease (AD). Central infusion of Aβ42 can lead to memory impairment in mice. Inhibiting the aggregation of Aβ has been considered a therapeutic strategy for AD. Endomorphin-1 (EM-1), an endogenous agonist of μ-opioid receptors, has been shown to inhibit the aggregation of Aβ in vitro. In the present study, we investigated whether EM-1 could alleviate the memory-impairing effects of Aβ42 in mice using novel object recognition (NOR) and object location recognition (OLR) tasks. We showed that co-administration of EM-1 was able to ameliorate Aβ42-induced amnesia in the lateral ventricle and the hippocampus, and these effects could not be inhibited by naloxone, an antagonist of μ-opioid receptors. Infusion of EM-1 or naloxone separately into the lateral ventricle had no influence on memory in the tasks. These results suggested that EM-1 might be effective as a drug for AD preventative treatment by inhibiting Aβ aggregation directly as a molecular modifier. Copyright © 2015 Elsevier B.V. All rights reserved.

  10. Bidirectional Modulation of Recognition Memory

    PubMed Central

    Ho, Jonathan W.; Poeta, Devon L.; Jacobson, Tara K.; Zolnik, Timothy A.; Neske, Garrett T.; Connors, Barry W.

    2015-01-01

    Perirhinal cortex (PER) has a well established role in the familiarity-based recognition of individual items and objects. For example, animals and humans with perirhinal damage are unable to distinguish familiar from novel objects in recognition memory tasks. In the normal brain, perirhinal neurons respond to novelty and familiarity by increasing or decreasing firing rates. Recent work also implicates oscillatory activity in the low-beta and low-gamma frequency bands in sensory detection, perception, and recognition. Using optogenetic methods in a spontaneous object exploration (SOR) task, we altered recognition memory performance in rats. In the SOR task, normal rats preferentially explore novel images over familiar ones. We modulated exploratory behavior in this task by optically stimulating channelrhodopsin-expressing perirhinal neurons at various frequencies while rats looked at novel or familiar 2D images. Stimulation at 30–40 Hz during looking caused rats to treat a familiar image as if it were novel by increasing time looking at the image. Stimulation at 30–40 Hz was not effective in increasing exploration of novel images. Stimulation at 10–15 Hz caused animals to treat a novel image as familiar by decreasing time looking at the image, but did not affect looking times for images that were already familiar. We conclude that optical stimulation of PER at different frequencies can alter visual recognition memory bidirectionally. SIGNIFICANCE STATEMENT Recognition of novelty and familiarity are important for learning, memory, and decision making. Perirhinal cortex (PER) has a well established role in the familiarity-based recognition of individual items and objects, but how novelty and familiarity are encoded and transmitted in the brain is not known. Perirhinal neurons respond to novelty and familiarity by changing firing rates, but recent work suggests that brain oscillations may also be important for recognition. In this study, we showed that

  11. Voice reaction times with recognition for Commodore computers

    NASA Technical Reports Server (NTRS)

    Washburn, David A.; Putney, R. Thompson

    1990-01-01

    Hardware and software modifications are presented that allow for collection and recognition by a Commodore computer of spoken responses. Responses are timed with millisecond accuracy and automatically analyzed and scored. Accuracy data for this device from several experiments are presented. Potential applications and suggestions for improving recognition accuracy are also discussed.

  12. Marked Object Recognition Multitouch Screen Printed Touchpad for Interactive Applications

    PubMed Central

    Nunes, Jivago Serrado; Castro, Nelson; Pereira, Nélson; Correia, Vitor; Lanceros-Mendez, Senentxu

    2017-01-01

    The market for interactive platforms is rapidly growing, and touchscreens have been incorporated in an increasing number of devices. Thus, the area of smart objects and devices is strongly increasing by adding interactive touch and multimedia content, leading to new uses and capabilities. In this work, a flexible screen printed sensor matrix is fabricated based on silver ink in a polyethylene terephthalate (PET) substrate. Diamond shaped capacitive electrodes coupled with conventional capacitive reading electronics enables fabrication of a highly functional capacitive touchpad, and also allows for the identification of marked objects. For the latter, the capacitive signatures are identified by intersecting points and distances between them. Thus, this work demonstrates the applicability of a low cost method using royalty-free geometries and technologies for the development of flexible multitouch touchpads for the implementation of interactive and object recognition applications. PMID:29194414

  13. Marked Object Recognition Multitouch Screen Printed Touchpad for Interactive Applications.

    PubMed

    Nunes, Jivago Serrado; Castro, Nelson; Gonçalves, Sergio; Pereira, Nélson; Correia, Vitor; Lanceros-Mendez, Senentxu

    2017-12-01

    The market for interactive platforms is rapidly growing, and touchscreens have been incorporated in an increasing number of devices. Thus, the area of smart objects and devices is strongly increasing by adding interactive touch and multimedia content, leading to new uses and capabilities. In this work, a flexible screen printed sensor matrix is fabricated based on silver ink in a polyethylene terephthalate (PET) substrate. Diamond shaped capacitive electrodes coupled with conventional capacitive reading electronics enables fabrication of a highly functional capacitive touchpad, and also allows for the identification of marked objects. For the latter, the capacitive signatures are identified by intersecting points and distances between them. Thus, this work demonstrates the applicability of a low cost method using royalty-free geometries and technologies for the development of flexible multitouch touchpads for the implementation of interactive and object recognition applications.

  14. Hierachical Object Recognition Using Libraries of Parameterized Model Sub-Parts.

    DTIC Science & Technology

    1987-06-01

    SketchI Structure Hierarchy Constrained Search 20. AUISTR ACT (Ce.ntU..w se reveres. 01411 at 00 OW 4MI 9smtilp Me"h aindo" This thesis describes the... theseU hierarchies to achieve robust recognition based on effective organization and indexing schemes for model libraries. The goal of the system is to...with different relative scaling, rotation, or translation than in the models. The approach taken in this thesis is to develop an object shape

  15. Invariant object recognition based on the generalized discrete radon transform

    NASA Astrophysics Data System (ADS)

    Easley, Glenn R.; Colonna, Flavia

    2004-04-01

    We introduce a method for classifying objects based on special cases of the generalized discrete Radon transform. We adjust the transform and the corresponding ridgelet transform by means of circular shifting and a singular value decomposition (SVD) to obtain a translation, rotation and scaling invariant set of feature vectors. We then use a back-propagation neural network to classify the input feature vectors. We conclude with experimental results and compare these with other invariant recognition methods.

  16. Object-oriented recognition of high-resolution remote sensing image

    NASA Astrophysics Data System (ADS)

    Wang, Yongyan; Li, Haitao; Chen, Hong; Xu, Yuannan

    2016-01-01

    With the development of remote sensing imaging technology and the improvement of multi-source image's resolution in satellite visible light, multi-spectral and hyper spectral , the high resolution remote sensing image has been widely used in various fields, for example military field, surveying and mapping, geophysical prospecting, environment and so forth. In remote sensing image, the segmentation of ground targets, feature extraction and the technology of automatic recognition are the hotspot and difficulty in the research of modern information technology. This paper also presents an object-oriented remote sensing image scene classification method. The method is consist of vehicles typical objects classification generation, nonparametric density estimation theory, mean shift segmentation theory, multi-scale corner detection algorithm, local shape matching algorithm based on template. Remote sensing vehicles image classification software system is designed and implemented to meet the requirements .

  17. Get rich quick: the signal to respond procedure reveals the time course of semantic richness effects during visual word recognition.

    PubMed

    Hargreaves, Ian S; Pexman, Penny M

    2014-05-01

    According to several current frameworks, semantic processing involves an early influence of language-based information followed by later influences of object-based information (e.g., situated simulations; Santos, Chaigneau, Simmons, & Barsalou, 2011). In the present study we examined whether these predictions extend to the influence of semantic variables in visual word recognition. We investigated the time course of semantic richness effects in visual word recognition using a signal-to-respond (STR) paradigm fitted to a lexical decision (LDT) and a semantic categorization (SCT) task. We used linear mixed effects to examine the relative contributions of language-based (number of senses, ARC) and object-based (imageability, number of features, body-object interaction ratings) descriptions of semantic richness at four STR durations (75, 100, 200, and 400ms). Results showed an early influence of number of senses and ARC in the SCT. In both LDT and SCT, object-based effects were the last to influence participants' decision latencies. We interpret our results within a framework in which semantic processes are available to influence word recognition as a function of their availability over time, and of their relevance to task-specific demands. Copyright © 2014 Elsevier B.V. All rights reserved.

  18. Long-lasting effects of prenatal dietary choline availability on object recognition memory ability in adult rats.

    PubMed

    Moreno, Hayarelis C; de Brugada, Isabel; Carias, Diamela; Gallo, Milagros

    2013-11-01

    Choline is an essential nutrient required for early development. Previous studies have shown that prenatal choline availability influences adult memory abilities depending on the medial temporal lobe integrity. The relevance of prenatal choline availability on object recognition memory was assessed in adult Wistar rats. Three groups of pregnant Wistar rats were fed from E12 to E18 with choline-deficient (0 g/kg choline chloride), standard (1.1 g/kg choline chloride), or choline-supplemented (5 g/kg choline chloride) diets. The offspring was cross-fostered to rat dams fed a standard diet during pregnancy and tested at the age of 3 months in an object recognition memory task applying retention tests 24 and 48 hours after acquisition. Although no significant differences have been found in the performance of the three groups during the first retention test, the supplemented group exhibited improved memory compared with both the standard and the deficient group in the second retention test, 48 hours after acquisition. In addition, at the second retention test the deficient group did not differ from chance. Taken together, the results support the notion of a long-lasting beneficial effect of prenatal choline supplementation on object recognition memory which is evident when the rats reach adulthood. The results are discussed in terms of their relevance for improving the understanding of the cholinergic involvement in object recognition memory and the implications of the importance of maternal diet for lifelong cognitive abilities.

  19. An optical processor for object recognition and tracking

    NASA Technical Reports Server (NTRS)

    Sloan, J.; Udomkesmalee, S.

    1987-01-01

    The design and development of a miniaturized optical processor that performs real time image correlation are described. The optical correlator utilizes the Vander Lugt matched spatial filter technique. The correlation output, a focused beam of light, is imaged onto a CMOS photodetector array. In addition to performing target recognition, the device also tracks the target. The hardware, composed of optical and electro-optical components, occupies only 590 cu cm of volume. A complete correlator system would also include an input imaging lens. This optical processing system is compact, rugged, requires only 3.5 watts of operating power, and weighs less than 3 kg. It represents a major achievement in miniaturizing optical processors. When considered as a special-purpose processing unit, it is an attractive alternative to conventional digital image recognition processing. It is conceivable that the combined technology of both optical and ditital processing could result in a very advanced robot vision system.

  20. Effect of a Bluetooth-implemented hearing aid on speech recognition performance: subjective and objective measurement.

    PubMed

    Kim, Min-Beom; Chung, Won-Ho; Choi, Jeesun; Hong, Sung Hwa; Cho, Yang-Sun; Park, Gyuseok; Lee, Sangmin

    2014-06-01

    The object was to evaluate speech perception improvement through Bluetooth-implemented hearing aids in hearing-impaired adults. Thirty subjects with bilateral symmetric moderate sensorineural hearing loss participated in this study. A Bluetooth-implemented hearing aid was fitted unilaterally in all study subjects. Objective speech recognition score and subjective satisfaction were measured with a Bluetooth-implemented hearing aid to replace the acoustic connection from either a cellular phone or a loudspeaker system. In each system, participants were assigned to 4 conditions: wireless speech signal transmission into hearing aid (wireless mode) in quiet or noisy environment and conventional speech signal transmission using external microphone of hearing aid (conventional mode) in quiet or noisy environment. Also, participants completed questionnaires to investigate subjective satisfaction. Both cellular phone and loudspeaker system situation, participants showed improvements in sentence and word recognition scores with wireless mode compared to conventional mode in both quiet and noise conditions (P < .001). Participants also reported subjective improvements, including better sound quality, less noise interference, and better accuracy naturalness, when using the wireless mode (P < .001). Bluetooth-implemented hearing aids helped to improve subjective and objective speech recognition performances in quiet and noisy environments during the use of electronic audio devices.

  1. An Intelligent Systems Approach to Automated Object Recognition: A Preliminary Study

    USGS Publications Warehouse

    Maddox, Brian G.; Swadley, Casey L.

    2002-01-01

    Attempts at fully automated object recognition systems have met with varying levels of success over the years. However, none of the systems have achieved high enough accuracy rates to be run unattended. One of the reasons for this may be that they are designed from the computer's point of view and rely mainly on image-processing methods. A better solution to this problem may be to make use of modern advances in computational intelligence and distributed processing to try to mimic how the human brain is thought to recognize objects. As humans combine cognitive processes with detection techniques, such a system would combine traditional image-processing techniques with computer-based intelligence to determine the identity of various objects in a scene.

  2. Multispectral image analysis for object recognition and classification

    NASA Astrophysics Data System (ADS)

    Viau, C. R.; Payeur, P.; Cretu, A.-M.

    2016-05-01

    Computer and machine vision applications are used in numerous fields to analyze static and dynamic imagery in order to assist or automate decision-making processes. Advancements in sensor technologies now make it possible to capture and visualize imagery at various wavelengths (or bands) of the electromagnetic spectrum. Multispectral imaging has countless applications in various fields including (but not limited to) security, defense, space, medical, manufacturing and archeology. The development of advanced algorithms to process and extract salient information from the imagery is a critical component of the overall system performance. The fundamental objective of this research project was to investigate the benefits of combining imagery from the visual and thermal bands of the electromagnetic spectrum to improve the recognition rates and accuracy of commonly found objects in an office setting. A multispectral dataset (visual and thermal) was captured and features from the visual and thermal images were extracted and used to train support vector machine (SVM) classifiers. The SVM's class prediction ability was evaluated separately on the visual, thermal and multispectral testing datasets.

  3. Selective attention affects conceptual object priming and recognition: a study with young and older adults.

    PubMed

    Ballesteros, Soledad; Mayas, Julia

    2014-01-01

    In the present study, we investigated the effects of selective attention at encoding on conceptual object priming (Experiment 1) and old-new recognition memory (Experiment 2) tasks in young and older adults. The procedures of both experiments included encoding and memory test phases separated by a short delay. At encoding, the picture outlines of two familiar objects, one in blue and the other in green, were presented to the left and to the right of fixation. In Experiment 1, participants were instructed to attend to the picture outline of a certain color and to classify the object as natural or artificial. After a short delay, participants performed a natural/artificial speeded conceptual classification task with repeated attended, repeated unattended, and new pictures. In Experiment 2, participants at encoding memorized the attended pictures and classify them as natural or artificial. After the encoding phase, they performed an old-new recognition memory task. Consistent with previous findings with perceptual priming tasks, we found that conceptual object priming, like explicit memory, required attention at encoding. Significant priming was obtained in both age groups, but only for those pictures that were attended at encoding. Although older adults were slower than young adults, both groups showed facilitation for attended pictures. In line with previous studies, young adults had better recognition memory than older adults.

  4. Selective attention affects conceptual object priming and recognition: a study with young and older adults

    PubMed Central

    Ballesteros, Soledad; Mayas, Julia

    2015-01-01

    In the present study, we investigated the effects of selective attention at encoding on conceptual object priming (Experiment 1) and old–new recognition memory (Experiment 2) tasks in young and older adults. The procedures of both experiments included encoding and memory test phases separated by a short delay. At encoding, the picture outlines of two familiar objects, one in blue and the other in green, were presented to the left and to the right of fixation. In Experiment 1, participants were instructed to attend to the picture outline of a certain color and to classify the object as natural or artificial. After a short delay, participants performed a natural/artificial speeded conceptual classification task with repeated attended, repeated unattended, and new pictures. In Experiment 2, participants at encoding memorized the attended pictures and classify them as natural or artificial. After the encoding phase, they performed an old–new recognition memory task. Consistent with previous findings with perceptual priming tasks, we found that conceptual object priming, like explicit memory, required attention at encoding. Significant priming was obtained in both age groups, but only for those pictures that were attended at encoding. Although older adults were slower than young adults, both groups showed facilitation for attended pictures. In line with previous studies, young adults had better recognition memory than older adults. PMID:25628588

  5. Intracellular Zn(2+) signaling in the dentate gyrus is required for object recognition memory.

    PubMed

    Takeda, Atsushi; Tamano, Haruna; Ogawa, Taisuke; Takada, Shunsuke; Nakamura, Masatoshi; Fujii, Hiroaki; Ando, Masaki

    2014-11-01

    The role of perforant pathway-dentate granule cell synapses in cognitive behavior was examined focusing on synaptic Zn(2+) signaling in the dentate gyrus. Object recognition memory was transiently impaired when extracellular Zn(2+) levels were decreased by injection of clioquinol and N,N,N',N'-tetrakis-(2-pyridylmethyl) ethylendediamine. To pursue the effect of the loss and/or blockade of Zn(2+) signaling in dentate granule cells, ZnAF-2DA (100 pmol, 0.1 mM/1 µl), an intracellular Zn(2+) chelator, was locally injected into the dentate molecular layer of rats. ZnAF-2DA injection, which was estimated to chelate intracellular Zn(2+) signaling only in the dentate gyrus, affected object recognition memory 1 h after training without affecting intracellular Ca(2+) signaling in the dentate molecular layer. In vivo dentate gyrus long-term potentiation (LTP) was affected under the local perfusion of the recording region (the dentate granule cell layer) with 0.1 mM ZnAF-2DA, but not with 1-10 mM CaEDTA, an extracellular Zn(2+) chelator, suggesting that the blockade of intracellular Zn(2+) signaling in dentate granule cells affects dentate gyrus LTP. The present study demonstrates that intracellular Zn(2+) signaling in the dentate gyrus is required for object recognition memory, probably via dentate gyrus LTP expression. Copyright © 2014 Wiley Periodicals, Inc.

  6. Modified-hybrid optical neural network filter for multiple object recognition within cluttered scenes

    NASA Astrophysics Data System (ADS)

    Kypraios, Ioannis; Young, Rupert C. D.; Chatwin, Chris R.

    2009-08-01

    Motivated by the non-linear interpolation and generalization abilities of the hybrid optical neural network filter between the reference and non-reference images of the true-class object we designed the modifiedhybrid optical neural network filter. We applied an optical mask to the hybrid optical neural network's filter input. The mask was built with the constant weight connections of a randomly chosen image included in the training set. The resulted design of the modified-hybrid optical neural network filter is optimized for performing best in cluttered scenes of the true-class object. Due to the shift invariance properties inherited by its correlator unit the filter can accommodate multiple objects of the same class to be detected within an input cluttered image. Additionally, the architecture of the neural network unit of the general hybrid optical neural network filter allows the recognition of multiple objects of different classes within the input cluttered image by modifying the output layer of the unit. We test the modified-hybrid optical neural network filter for multiple objects of the same and of different classes' recognition within cluttered input images and video sequences of cluttered scenes. The filter is shown to exhibit with a single pass over the input data simultaneously out-of-plane rotation, shift invariance and good clutter tolerance. It is able to successfully detect and classify correctly the true-class objects within background clutter for which there has been no previous training.

  7. Towards Real-Time Speech Emotion Recognition for Affective E-Learning

    ERIC Educational Resources Information Center

    Bahreini, Kiavash; Nadolski, Rob; Westera, Wim

    2016-01-01

    This paper presents the voice emotion recognition part of the FILTWAM framework for real-time emotion recognition in affective e-learning settings. FILTWAM (Framework for Improving Learning Through Webcams And Microphones) intends to offer timely and appropriate online feedback based upon learner's vocal intonations and facial expressions in order…

  8. Biased figure-ground assignment affects conscious object recognition in spatial neglect.

    PubMed

    Eramudugolla, Ranmalee; Driver, Jon; Mattingley, Jason B

    2010-09-01

    Unilateral spatial neglect is a disorder of attention and spatial representation, in which early visual processes such as figure-ground segmentation have been assumed to be largely intact. There is evidence, however, that the spatial attention bias underlying neglect can bias the segmentation of a figural region from its background. Relatively few studies have explicitly examined the effect of spatial neglect on processing the figures that result from such scene segmentation. Here, we show that a neglect patient's bias in figure-ground segmentation directly influences his conscious recognition of these figures. By varying the relative salience of figural and background regions in static, two-dimensional displays, we show that competition between elements in such displays can modulate a neglect patient's ability to recognise parsed figures in a scene. The findings provide insight into the interaction between scene segmentation, explicit object recognition, and attention.

  9. Estradiol-induced object recognition memory consolidation is dependent on activation of mTOR signaling in the dorsal hippocampus

    PubMed Central

    Fortress, Ashley M.; Fan, Lu; Orr, Patrick T.; Zhao, Zaorui; Frick, Karyn M.

    2013-01-01

    The mammalian target of rapamycin (mTOR) signaling pathway is an important regulator of protein synthesis and is essential for various forms of hippocampal memory. Here, we asked whether the enhancement of object recognition memory consolidation produced by dorsal hippocampal infusion of 17β-estradiol (E2) is dependent on mTOR signaling in the dorsal hippocampus, and whether E2-induced mTOR signaling is dependent on dorsal hippocampal phosphatidylinositol 3-kinase (PI3K) and extracellular signal-regulated kinase (ERK) activation. We first demonstrated that the enhancement of object recognition induced by E2 was blocked by dorsal hippocampal inhibition of ERK, PI3K, or mTOR activation. We then showed that an increase in dorsal hippocampal ERK phosphorylation 5 min after intracerebroventricular (ICV) E2 infusion was also blocked by dorsal hippocampal infusion of the three cell signaling inhibitors. Next, we found that ICV infusion of E2 increased phosphorylation of the downstream mTOR targets S6K (Thr-421) and 4E-BP1 in the dorsal hippocampus 5 min after infusion, and that this phosphorylation was blocked by dorsal hippocampal infusion of inhibitors of ERK, PI3K, and mTOR. Collectively, these data demonstrate for the first time that activation of the dorsal hippocampal mTOR signaling pathway is necessary for E2 to enhance object recognition memory consolidation and that E2-induced mTOR activation is dependent on upstream activation of ERK and PI3K signaling. PMID:23422279

  10. Estradiol enhances object recognition memory in Swiss female mice by activating hippocampal estrogen receptor α.

    PubMed

    Pereira, Luciana M; Bastos, Cristiane P; de Souza, Jéssica M; Ribeiro, Fabíola M; Pereira, Grace S

    2014-10-01

    In rodents, 17β-estradiol (E2) enhances hippocampal function and improves performance in several memory tasks. Regarding the object recognition paradigm, E2 commonly act as a cognitive enhancer. However, the types of estrogen receptor (ER) involved, as well as the underlying molecular mechanisms are still under investigation. In the present study, we asked whether E2 enhances object recognition memory by activating ERα and/or ERβ in the hippocampus of Swiss female mice. First, we showed that immediately post-training intraperitoneal (i.p.) injection of E2 (0.2 mg/kg) allowed object recognition memory to persist 48 h in ovariectomized (OVX) Swiss female mice. This result indicates that Swiss female mice are sensitive to the promnesic effects of E2 and is in accordance with other studies, which used C57/BL6 female mice. To verify if the activation of hippocampal ERα or ERβ would be sufficient to improve object memory, we used PPT and DPN, which are selective ERα and ERβ agonists, respectively. We found that PPT, but not DPN, improved object memory in Swiss female mice. However, DPN was able to improve memory in C57/BL6 female mice, which is in accordance with other studies. Next, we tested if the E2 effect on improving object memory depends on ER activation in the hippocampus. Thus, we tested if the infusion of intra-hippocampal TPBM and PHTPP, selective antagonists of ERα and ERβ, respectively, would block the memory enhancement effect of E2. Our results showed that TPBM, but not PHTPP, blunted the promnesic effect of E2, strongly suggesting that in Swiss female mice, the ERα and not the ERβ is the receptor involved in the promnesic effect of E2. It was already demonstrated that E2, as well as PPT and DPN, increase the phospho-ERK2 level in the dorsal hippocampus of C57/BL6 mice. Here we observed that PPT increased phospho-ERK1, while DPN decreased phospho-ERK2 in the dorsal hippocampus of Swiss female mice subjected to the object recognition sample phase

  11. Super-recognition in development: A case study of an adolescent with extraordinary face recognition skills.

    PubMed

    Bennetts, Rachel J; Mole, Joseph; Bate, Sarah

    2017-09-01

    Face recognition abilities vary widely. While face recognition deficits have been reported in children, it is unclear whether superior face recognition skills can be encountered during development. This paper presents O.B., a 14-year-old female with extraordinary face recognition skills: a "super-recognizer" (SR). O.B. demonstrated exceptional face-processing skills across multiple tasks, with a level of performance that is comparable to adult SRs. Her superior abilities appear to be specific to face identity: She showed an exaggerated face inversion effect and her superior abilities did not extend to object processing or non-identity aspects of face recognition. Finally, an eye-movement task demonstrated that O.B. spent more time than controls examining the nose - a pattern previously reported in adult SRs. O.B. is therefore particularly skilled at extracting and using identity-specific facial cues, indicating that face and object recognition are dissociable during development, and that super recognition can be detected in adolescence.

  12. Short- and long-term effects of nicotine and the histone deacetylase inhibitor phenylbutyrate on novel object recognition in zebrafish.

    PubMed

    Faillace, M P; Pisera-Fuster, A; Medrano, M P; Bejarano, A C; Bernabeu, R O

    2017-03-01

    Zebrafish have a sophisticated color- and shape-sensitive visual system, so we examined color cue-based novel object recognition in zebrafish. We evaluated preference in the absence or presence of drugs that affect attention and memory retention in rodents: nicotine and the histone deacetylase inhibitor (HDACi) phenylbutyrate (PhB). The objective of this study was to evaluate whether nicotine and PhB affect innate preferences of zebrafish for familiar and novel objects after short- and long-retention intervals. We developed modified object recognition (OR) tasks using neutral novel and familiar objects in different colors. We also tested objects which differed with respect to the exploratory behavior they elicited from naïve zebrafish. Zebrafish showed an innate preference for exploring red or green objects rather than yellow or blue objects. Zebrafish were better at discriminating color changes than changes in object shape or size. Nicotine significantly enhanced or changed short-term innate novel object preference whereas PhB had similar effects when preference was assessed 24 h after training. Analysis of other zebrafish behaviors corroborated these results. Zebrafish were innately reluctant or prone to explore colored novel objects, so drug effects on innate preference for objects can be evaluated changing the color of objects with a simple geometry. Zebrafish exhibited recognition memory for novel objects with similar innate significance. Interestingly, nicotine and PhB significantly modified innate object preference.

  13. Cough Recognition Based on Mel Frequency Cepstral Coefficients and Dynamic Time Warping

    NASA Astrophysics Data System (ADS)

    Zhu, Chunmei; Liu, Baojun; Li, Ping

    Cough recognition provides important clinical information for the treatment of many respiratory diseases, but the assessment of cough frequency over a long period of time remains unsatisfied for either clinical or research purpose. In this paper, according to the advantage of dynamic time warping (DTW) and the characteristic of cough recognition, an attempt is made to adapt DTW as the recognition algorithm for cough recognition. The process of cough recognition based on mel frequency cepstral coefficients (MFCC) and DTW is introduced. Experiment results of testing samples from 3 subjects show that acceptable performances of cough recognition are obtained by DTW with a small training set.

  14. A Comparison of the Effects of Depth Rotation on Visual and Haptic Three-Dimensional Object Recognition

    ERIC Educational Resources Information Center

    Lawson, Rebecca

    2009-01-01

    A sequential matching task was used to compare how the difficulty of shape discrimination influences the achievement of object constancy for depth rotations across haptic and visual object recognition. Stimuli were nameable, 3-dimensional plastic models of familiar objects (e.g., bed, chair) and morphs midway between these endpoint shapes (e.g., a…

  15. The effect of Wi-Fi electromagnetic waves in unimodal and multimodal object recognition tasks in male rats.

    PubMed

    Hassanshahi, Amin; Shafeie, Seyed Ali; Fatemi, Iman; Hassanshahi, Elham; Allahtavakoli, Mohammad; Shabani, Mohammad; Roohbakhsh, Ali; Shamsizadeh, Ali

    2017-06-01

    Wireless internet (Wi-Fi) electromagnetic waves (2.45 GHz) have widespread usage almost everywhere, especially in our homes. Considering the recent reports about some hazardous effects of Wi-Fi signals on the nervous system, this study aimed to investigate the effect of 2.4 GHz Wi-Fi radiation on multisensory integration in rats. This experimental study was done on 80 male Wistar rats that were allocated into exposure and sham groups. Wi-Fi exposure to 2.4 GHz microwaves [in Service Set Identifier mode (23.6 dBm and 3% for power and duty cycle, respectively)] was done for 30 days (12 h/day). Cross-modal visual-tactile object recognition (CMOR) task was performed by four variations of spontaneous object recognition (SOR) test including standard SOR, tactile SOR, visual SOR, and CMOR tests. A discrimination ratio was calculated to assess the preference of animal to the novel object. The expression levels of M1 and GAT1 mRNA in the hippocampus were assessed by quantitative real-time RT-PCR. Results demonstrated that rats in Wi-Fi exposure groups could not discriminate significantly between the novel and familiar objects in any of the standard SOR, tactile SOR, visual SOR, and CMOR tests. The expression of M1 receptors increased following Wi-Fi exposure. In conclusion, results of this study showed that chronic exposure to Wi-Fi electromagnetic waves might impair both unimodal and cross-modal encoding of information.

  16. Real-time unconstrained object recognition: a processing pipeline based on the mammalian visual system.

    PubMed

    Aguilar, Mario; Peot, Mark A; Zhou, Jiangying; Simons, Stephen; Liao, Yuwei; Metwalli, Nader; Anderson, Mark B

    2012-03-01

    The mammalian visual system is still the gold standard for recognition accuracy, flexibility, efficiency, and speed. Ongoing advances in our understanding of function and mechanisms in the visual system can now be leveraged to pursue the design of computer vision architectures that will revolutionize the state of the art in computer vision.

  17. Anthropomorphic robot for recognition and drawing generalized object images

    NASA Astrophysics Data System (ADS)

    Ginzburg, Vera M.

    1998-10-01

    The process of recognition, for instance, understanding the text, written by different fonts, consists in the depriving of the individual attributes of the letters in the particular font. It is shown that such process, in Nature and technique, can be provided by the narrowing the spatial frequency of the object's image by its defocusing. In defocusing images remain only areas, so-called Informative Fragments (IFs), which all together form the generalized (stylized) image of many identical objects. It is shown that the variety of shapes of IFs is restricted and can be presented by `Geometrical alphabet'. The `letters' for this alphabet can be created using two basic `genetic' figures: a stripe and round spot. It is known from physiology that the special cells of visual cortex response to these particular figures. The prototype of such `genetic' alphabet has been made using Boolean algebra (Venn's diagrams). The algorithm for drawing the letter's (`genlet's') shape in this alphabet and generalized images of objects (for example, `sleeping cat'), are given. A scheme of an anthropomorphic robot is shown together with results of model computer experiment of the robot's action--`drawing' the generalized image.

  18. Incrementally learning objects by touch: online discriminative and generative models for tactile-based recognition.

    PubMed

    Soh, Harold; Demiris, Yiannis

    2014-01-01

    Human beings not only possess the remarkable ability to distinguish objects through tactile feedback but are further able to improve upon recognition competence through experience. In this work, we explore tactile-based object recognition with learners capable of incremental learning. Using the sparse online infinite Echo-State Gaussian process (OIESGP), we propose and compare two novel discriminative and generative tactile learners that produce probability distributions over objects during object grasping/palpation. To enable iterative improvement, our online methods incorporate training samples as they become available. We also describe incremental unsupervised learning mechanisms, based on novelty scores and extreme value theory, when teacher labels are not available. We present experimental results for both supervised and unsupervised learning tasks using the iCub humanoid, with tactile sensors on its five-fingered anthropomorphic hand, and 10 different object classes. Our classifiers perform comparably to state-of-the-art methods (C4.5 and SVM classifiers) and findings indicate that tactile signals are highly relevant for making accurate object classifications. We also show that accurate "early" classifications are possible using only 20-30 percent of the grasp sequence. For unsupervised learning, our methods generate high quality clusterings relative to the widely-used sequential k-means and self-organising map (SOM), and we present analyses into the differences between the approaches.

  19. Determinants of naming latencies, object comprehension times, and new norms for the Russian standardized set of the colorized version of the Snodgrass and Vanderwart pictures.

    PubMed

    Bonin, Patrick; Guillemard-Tsaparina, Diana; Méot, Alain

    2013-09-01

    We report object-naming and object recognition times collected from Russian native speakers for the colorized version of the Snodgrass and Vanderwart (Journal of Experimental Psychology: Human Learning and Memory 6:174-215, 1980) pictures (Rossion & Pourtois, Perception 33:217-236, 2004). New norms for image variability, body-object interaction [BOI], and subjective frequency collected in Russian, as well as new name agreement scores for the colorized pictures in French, are also reported. In both object-naming and object comprehension times, the name agreement, image agreement, and age-of-acquisition variables made significant independent contributions. Objective word frequency was reliable in object-naming latencies only. The variables of image variability, BOI, and subjective frequency were not significant in either object naming or object comprehension. Finally, imageability was reliable in both tasks. The new norms and object-naming and object recognition times are provided as supplemental materials.

  20. Developmental Changes in Visual Object Recognition between 18 and 24 Months of Age

    ERIC Educational Resources Information Center

    Pereira, Alfredo F.; Smith, Linda B.

    2009-01-01

    Two experiments examined developmental changes in children's visual recognition of common objects during the period of 18 to 24 months. Experiment 1 examined children's ability to recognize common category instances that presented three different kinds of information: (1) richly detailed and prototypical instances that presented both local and…

  1. Object Recognition in Mental Representations: Directions for Exploring Diagnostic Features through Visual Mental Imagery

    PubMed Central

    Roldan, Stephanie M.

    2017-01-01

    One of the fundamental goals of object recognition research is to understand how a cognitive representation produced from the output of filtered and transformed sensory information facilitates efficient viewer behavior. Given that mental imagery strongly resembles perceptual processes in both cortical regions and subjective visual qualities, it is reasonable to question whether mental imagery facilitates cognition in a manner similar to that of perceptual viewing: via the detection and recognition of distinguishing features. Categorizing the feature content of mental imagery holds potential as a reverse pathway by which to identify the components of a visual stimulus which are most critical for the creation and retrieval of a visual representation. This review will examine the likelihood that the information represented in visual mental imagery reflects distinctive object features thought to facilitate efficient object categorization and recognition during perceptual viewing. If it is the case that these representational features resemble their sensory counterparts in both spatial and semantic qualities, they may well be accessible through mental imagery as evaluated through current investigative techniques. In this review, methods applied to mental imagery research and their findings are reviewed and evaluated for their efficiency in accessing internal representations, and implications for identifying diagnostic features are discussed. An argument is made for the benefits of combining mental imagery assessment methods with diagnostic feature research to advance the understanding of visual perceptive processes, with suggestions for avenues of future investigation. PMID:28588538

  2. Object Recognition in Mental Representations: Directions for Exploring Diagnostic Features through Visual Mental Imagery.

    PubMed

    Roldan, Stephanie M

    2017-01-01

    One of the fundamental goals of object recognition research is to understand how a cognitive representation produced from the output of filtered and transformed sensory information facilitates efficient viewer behavior. Given that mental imagery strongly resembles perceptual processes in both cortical regions and subjective visual qualities, it is reasonable to question whether mental imagery facilitates cognition in a manner similar to that of perceptual viewing: via the detection and recognition of distinguishing features. Categorizing the feature content of mental imagery holds potential as a reverse pathway by which to identify the components of a visual stimulus which are most critical for the creation and retrieval of a visual representation. This review will examine the likelihood that the information represented in visual mental imagery reflects distinctive object features thought to facilitate efficient object categorization and recognition during perceptual viewing. If it is the case that these representational features resemble their sensory counterparts in both spatial and semantic qualities, they may well be accessible through mental imagery as evaluated through current investigative techniques. In this review, methods applied to mental imagery research and their findings are reviewed and evaluated for their efficiency in accessing internal representations, and implications for identifying diagnostic features are discussed. An argument is made for the benefits of combining mental imagery assessment methods with diagnostic feature research to advance the understanding of visual perceptive processes, with suggestions for avenues of future investigation.

  3. Better Object Recognition and Naming Outcome With MRI-Guided Stereotactic Laser Amygdalohippocampotomy for Temporal Lobe Epilepsy

    PubMed Central

    Drane, Daniel L.; Loring, David W.; Voets, Natalie L.; Price, Michele; Ojemann, Jeffrey G.; Willie, Jon T.; Saindane, Amit M.; Phatak, Vaishali; Ivanisevic, Mirjana; Millis, Scott; Helmers, Sandra L.; Miller, John W.; Meador, Kimford J.; Gross, Robert E.

    2015-01-01

    SUMMARY OBJECTIVES Temporal lobe epilepsy (TLE) patients experience significant deficits in category-related object recognition and naming following standard surgical approaches. These deficits may result from a decoupling of core processing modules (e.g., language, visual processing, semantic memory), due to “collateral damage” to temporal regions outside the hippocampus following open surgical approaches. We predicted stereotactic laser amygdalohippocampotomy (SLAH) would minimize such deficits because it preserves white matter pathways and neocortical regions critical for these cognitive processes. METHODS Tests of naming and recognition of common nouns (Boston Naming Test) and famous persons were compared with nonparametric analyses using exact tests between a group of nineteen patients with medically-intractable mesial TLE undergoing SLAH (10 dominant, 9 nondominant), and a comparable series of TLE patients undergoing standard surgical approaches (n=39) using a prospective, non-randomized, non-blinded, parallel group design. RESULTS Performance declines were significantly greater for the dominant TLE patients undergoing open resection versus SLAH for naming famous faces and common nouns (F=24.3, p<.0001, η2=.57, & F=11.2, p<.001, η2=.39, respectively), and for the nondominant TLE patients undergoing open resection versus SLAH for recognizing famous faces (F=3.9, p<.02, η2=.19). When examined on an individual subject basis, no SLAH patients experienced any performance declines on these measures. In contrast, 32 of the 39 undergoing standard surgical approaches declined on one or more measures for both object types (p<.001, Fisher’s exact test). Twenty-one of 22 left (dominant) TLE patients declined on one or both naming tasks after open resection, while 11 of 17 right (non-dominant) TLE patients declined on face recognition. SIGNIFICANCE Preliminary results suggest 1) naming and recognition functions can be spared in TLE patients undergoing SLAH, and 2

  4. Differential Roles for "Nr4a1" and "Nr4a2" in Object Location vs. Object Recognition Long-Term Memory

    ERIC Educational Resources Information Center

    McNulty, Susan E.; Barrett, Ruth M.; Vogel-Ciernia, Annie; Malvaez, Melissa; Hernandez, Nicole; Davatolhagh, M. Felicia; Matheos, Dina P.; Schiffman, Aaron; Wood, Marcelo A.

    2012-01-01

    "Nr4a1" and "Nr4a2" are transcription factors and immediate early genes belonging to the nuclear receptor Nr4a family. In this study, we examine their role in long-term memory formation for object location and object recognition. Using siRNA to block expression of either "Nr4a1" or "Nr4a2", we found that "Nr4a2" is necessary for both long-term…

  5. Perirhinal Cortex Resolves Feature Ambiguity in Configural Object Recognition and Perceptual Oddity Tasks

    ERIC Educational Resources Information Center

    Bartko, Susan J.; Winters, Boyer D.; Cowell, Rosemary A.; Saksida, Lisa M.; Bussey, Timothy J.

    2007-01-01

    The perirhinal cortex (PRh) has a well-established role in object recognition memory. More recent studies suggest that PRh is also important for two-choice visual discrimination tasks. Specifically, it has been suggested that PRh contains conjunctive representations that help resolve feature ambiguity, which occurs when a task cannot easily be…

  6. Coarse-coded higher-order neural networks for PSRI object recognition. [position, scale, and rotation invariant

    NASA Technical Reports Server (NTRS)

    Spirkovska, Lilly; Reid, Max B.

    1993-01-01

    A higher-order neural network (HONN) can be designed to be invariant to changes in scale, translation, and inplane rotation. Invariances are built directly into the architecture of a HONN and do not need to be learned. Consequently, fewer training passes and a smaller training set are required to learn to distinguish between objects. The size of the input field is limited, however, because of the memory required for the large number of interconnections in a fully connected HONN. By coarse coding the input image, the input field size can be increased to allow the larger input scenes required for practical object recognition problems. We describe a coarse coding technique and present simulation results illustrating its usefulness and its limitations. Our simulations show that a third-order neural network can be trained to distinguish between two objects in a 4096 x 4096 pixel input field independent of transformations in translation, in-plane rotation, and scale in less than ten passes through the training set. Furthermore, we empirically determine the limits of the coarse coding technique in the object recognition domain.

  7. Short-term blueberry-enriched antioxidant diet prevents and reverses object recognition memory loss in aged rats

    USDA-ARS?s Scientific Manuscript database

    Objective Previously, four months of a blueberry-enriched (BB) antioxidant diet prevented impaired object recognition memory in aged rats. Experiment 1 determined whether one and two-month BB diets would have a similar effect and whether the benefits would disappear promptly after terminating the d...

  8. Dentate gyrus supports slope recognition memory, shades of grey-context pattern separation and recognition memory, and CA3 supports pattern completion for object memory.

    PubMed

    Kesner, Raymond P; Kirk, Ryan A; Yu, Zhenghui; Polansky, Caitlin; Musso, Nick D

    2016-03-01

    In order to examine the role of the dorsal dentate gyrus (dDG) in slope (vertical space) recognition and possible pattern separation, various slope (vertical space) degrees were used in a novel exploratory paradigm to measure novelty detection for changes in slope (vertical space) recognition memory and slope memory pattern separation in Experiment 1. The results of the experiment indicate that control rats displayed a slope recognition memory function with a pattern separation process for slope memory that is dependent upon the magnitude of change in slope between study and test phases. In contrast, the dDG lesioned rats displayed an impairment in slope recognition memory, though because there was no significant interaction between the two groups and slope memory, a reliable pattern separation impairment for slope could not be firmly established in the DG lesioned rats. In Experiment 2, in order to determine whether, the dDG plays a role in shades of grey spatial context recognition and possible pattern separation, shades of grey were used in a novel exploratory paradigm to measure novelty detection for changes in the shades of grey context environment. The results of the experiment indicate that control rats displayed a shades of grey-context pattern separation effect across levels of separation of context (shades of grey). In contrast, the DG lesioned rats displayed a significant interaction between the two groups and levels of shades of grey suggesting impairment in a pattern separation function for levels of shades of grey. In Experiment 3 in order to determine whether the dorsal CA3 (dCA3) plays a role in object pattern completion, a new task requiring less training and using a choice that was based on choosing the correct set of objects on a two-choice discrimination task was used. The results indicated that control rats displayed a pattern completion function based on the availability of one, two, three or four cues. In contrast, the dCA3 lesioned rats

  9. The medial prefrontal cortex-lateral entorhinal cortex circuit is essential for episodic-like memory and associative object-recognition.

    PubMed

    Chao, Owen Y; Huston, Joseph P; Li, Jay-Shake; Wang, An-Li; de Souza Silva, Maria A

    2016-05-01

    The prefrontal cortex directly projects to the lateral entorhinal cortex (LEC), an important substrate for engaging item-associated information and relaying the information to the hippocampus. Here we ask to what extent the communication between the prefrontal cortex and LEC is critically involved in the processing of episodic-like memory. We applied a disconnection procedure to test whether the interaction between the medial prefrontal cortex (mPFC) and LEC is essential for the expression of recognition memory. It was found that male rats that received unilateral NMDA lesions of the mPFC and LEC in the same hemisphere, exhibited intact episodic-like (what-where-when) and object-recognition memories. When these lesions were placed in the opposite hemispheres (disconnection), episodic-like and associative memories for object identity, location and context were impaired. However, the disconnection did not impair the components of episodic memory, namely memory for novel object (what), object place (where) and temporal order (when), per se. Thus, the present findings suggest that the mPFC and LEC are a critical part of a neural circuit that underlies episodic-like and associative object-recognition memory. © 2015 Wiley Periodicals, Inc.

  10. Stereo Viewing Modulates Three-Dimensional Shape Processing During Object Recognition: A High-Density ERP Study

    PubMed Central

    2017-01-01

    The role of stereo disparity in the recognition of 3-dimensional (3D) object shape remains an unresolved issue for theoretical models of the human visual system. We examined this issue using high-density (128 channel) recordings of event-related potentials (ERPs). A recognition memory task was used in which observers were trained to recognize a subset of complex, multipart, 3D novel objects under conditions of either (bi-) monocular or stereo viewing. In a subsequent test phase they discriminated previously trained targets from untrained distractor objects that shared either local parts, 3D spatial configuration, or neither dimension, across both previously seen and novel viewpoints. The behavioral data showed a stereo advantage for target recognition at untrained viewpoints. ERPs showed early differential amplitude modulations to shape similarity defined by local part structure and global 3D spatial configuration. This occurred initially during an N1 component around 145–190 ms poststimulus onset, and then subsequently during an N2/P3 component around 260–385 ms poststimulus onset. For mono viewing, amplitude modulation during the N1 was greatest between targets and distracters with different local parts for trained views only. For stereo viewing, amplitude modulation during the N2/P3 was greatest between targets and distracters with different global 3D spatial configurations and generalized across trained and untrained views. The results show that image classification is modulated by stereo information about the local part, and global 3D spatial configuration of object shape. The findings challenge current theoretical models that do not attribute functional significance to stereo input during the computation of 3D object shape. PMID:29022728

  11. New neural-networks-based 3D object recognition system

    NASA Astrophysics Data System (ADS)

    Abolmaesumi, Purang; Jahed, M.

    1997-09-01

    Three-dimensional object recognition has always been one of the challenging fields in computer vision. In recent years, Ulman and Basri (1991) have proposed that this task can be done by using a database of 2-D views of the objects. The main problem in their proposed system is that the correspondent points should be known to interpolate the views. On the other hand, their system should have a supervisor to decide which class does the represented view belong to. In this paper, we propose a new momentum-Fourier descriptor that is invariant to scale, translation, and rotation. This descriptor provides the input feature vectors to our proposed system. By using the Dystal network, we show that the objects can be classified with over 95% precision. We have used this system to classify the objects like cube, cone, sphere, torus, and cylinder. Because of the nature of the Dystal network, this system reaches to its stable point by a single representation of the view to the system. This system can also classify the similar views to a single class (e.g., for the cube, the system generated 9 different classes for 50 different input views), which can be used to select an optimum database of training views. The system is also very flexible to the noise and deformed views.

  12. SigVox - A 3D feature matching algorithm for automatic street object recognition in mobile laser scanning point clouds

    NASA Astrophysics Data System (ADS)

    Wang, Jinhu; Lindenbergh, Roderik; Menenti, Massimo

    2017-06-01

    Urban road environments contain a variety of objects including different types of lamp poles and traffic signs. Its monitoring is traditionally conducted by visual inspection, which is time consuming and expensive. Mobile laser scanning (MLS) systems sample the road environment efficiently by acquiring large and accurate point clouds. This work proposes a methodology for urban road object recognition from MLS point clouds. The proposed method uses, for the first time, shape descriptors of complete objects to match repetitive objects in large point clouds. To do so, a novel 3D multi-scale shape descriptor is introduced, that is embedded in a workflow that efficiently and automatically identifies different types of lamp poles and traffic signs. The workflow starts by tiling the raw point clouds along the scanning trajectory and by identifying non-ground points. After voxelization of the non-ground points, connected voxels are clustered to form candidate objects. For automatic recognition of lamp poles and street signs, a 3D significant eigenvector based shape descriptor using voxels (SigVox) is introduced. The 3D SigVox descriptor is constructed by first subdividing the points with an octree into several levels. Next, significant eigenvectors of the points in each voxel are determined by principal component analysis (PCA) and mapped onto the appropriate triangle of a sphere approximating icosahedron. This step is repeated for different scales. By determining the similarity of 3D SigVox descriptors between candidate point clusters and training objects, street furniture is automatically identified. The feasibility and quality of the proposed method is verified on two point clouds obtained in opposite direction of a stretch of road of 4 km. 6 types of lamp pole and 4 types of road sign were selected as objects of interest. Ground truth validation showed that the overall accuracy of the ∼170 automatically recognized objects is approximately 95%. The results demonstrate

  13. Recognition of Simple 3D Geometrical Objects under Partial Occlusion

    NASA Astrophysics Data System (ADS)

    Barchunova, Alexandra; Sommer, Gerald

    In this paper we present a novel procedure for contour-based recognition of partially occluded three-dimensional objects. In our approach we use images of real and rendered objects whose contours have been deformed by a restricted change of the viewpoint. The preparatory part consists of contour extraction, preprocessing, local structure analysis and feature extraction. The main part deals with an extended construction and functionality of the classifier ensemble Adaptive Occlusion Classifier (AOC). It relies on a hierarchical fragmenting algorithm to perform a local structure analysis which is essential when dealing with occlusions. In the experimental part of this paper we present classification results for five classes of simple geometrical figures: prism, cylinder, half cylinder, a cube, and a bridge. We compare classification results for three classical feature extractors: Fourier descriptors, pseudo Zernike and Zernike moments.

  14. Constant Light Desynchronizes Olfactory versus Object and Visuospatial Recognition Memory Performance

    PubMed Central

    Tam, Shu K.E.; Hasan, Sibah; Brown, Laurence A.; Jagannath, Aarti; Hankins, Mark W.; Foster, Russell G.; Vyazovskiy, Vladyslav V.

    2017-01-01

    Circadian rhythms optimize physiology and behavior to the varying demands of the 24 h day. The master circadian clock is located in the suprachiasmatic nuclei (SCN) of the hypothalamus and it regulates circadian oscillators in tissues throughout the body to prevent internal desynchrony. Here, we demonstrate for the first time that, under standard 12 h:12 h light/dark (LD) cycles, object, visuospatial, and olfactory recognition performance in C57BL/6J mice is consistently better at midday relative to midnight. However, under repeated exposure to constant light (rLL), recognition performance becomes desynchronized, with object and visuospatial performance better at subjective midday and olfactory performance better at subjective midnight. This desynchrony in behavioral performance is mirrored by changes in expression of the canonical clock genes Period1 and Period2 (Per1 and Per2), as well as the immediate-early gene Fos in the SCN, dorsal hippocampus, and olfactory bulb. Under rLL, rhythmic Per1 and Fos expression is attenuated in the SCN. In contrast, hippocampal gene expression remains rhythmic, mirroring object and visuospatial performance. Strikingly, Per1 and Fos expression in the olfactory bulb is reversed, mirroring the inverted olfactory performance. Temporal desynchrony among these regions does not result in arrhythmicity because core body temperature and exploratory activity rhythms persist under rLL. Our data provide the first demonstration that abnormal lighting conditions can give rise to temporal desynchrony between autonomous circadian oscillators in different regions, with different consequences for performance across different sensory domains. Such a dispersed network of dissociable circadian oscillators may provide greater flexibility when faced with conflicting environmental signals. SIGNIFICANCE STATEMENT A master circadian clock in the suprachiasmatic nuclei (SCN) of the hypothalamus regulates physiology and behavior across the 24 h day by

  15. Joint object and action recognition via fusion of partially observable surveillance imagery data

    NASA Astrophysics Data System (ADS)

    Shirkhodaie, Amir; Chan, Alex L.

    2017-05-01

    Partially observable group activities (POGA) occurring in confined spaces are epitomized by their limited observability of the objects and actions involved. In many POGA scenarios, different objects are being used by human operators for the conduct of various operations. In this paper, we describe the ontology of such as POGA in the context of In-Vehicle Group Activity (IVGA) recognition. Initially, we describe the virtue of ontology modeling in the context of IVGA and show how such an ontology and a priori knowledge about the classes of in-vehicle activities can be fused for inference of human actions that consequentially leads to understanding of human activity inside the confined space of a vehicle. In this paper, we treat the problem of "action-object" as a duality problem. We postulate a correlation between observed human actions and the object that is being utilized within those actions, and conversely, if an object being handled is recognized, we may be able to expect a number of actions that are likely to be performed on that object. In this study, we use partially observable human postural sequences to recognition actions. Inspired by convolutional neural networks (CNNs) learning capability, we present an architecture design using a new CNN model to learn "action-object" perception from surveillance videos. In this study, we apply a sequential Deep Hidden Markov Model (DHMM) as a post-processor to CNN to decode realized observations into recognized actions and activities. To generate the needed imagery data set for the training and testing of these new methods, we use the IRIS virtual simulation software to generate high-fidelity and dynamic animated scenarios that depict in-vehicle group activities under different operational contexts. The results of our comparative investigation are discussed and presented in detail.

  16. Object location and object recognition memory impairments, motivation deficits and depression in a model of Gulf War illness.

    PubMed

    Hattiangady, Bharathi; Mishra, Vikas; Kodali, Maheedhar; Shuai, Bing; Rao, Xiolan; Shetty, Ashok K

    2014-01-01

    Memory and mood deficits are the enduring brain-related symptoms in Gulf War illness (GWI). Both animal model and epidemiological investigations have indicated that these impairments in a majority of GW veterans are linked to exposures to chemicals such as pyridostigmine bromide (PB, an antinerve gas drug), permethrin (PM, an insecticide) and DEET (a mosquito repellant) encountered during the Persian Gulf War-1. Our previous study in a rat model has shown that combined exposures to low doses of GWI-related (GWIR) chemicals PB, PM, and DEET with or without 5-min of restraint stress (a mild stress paradigm) causes hippocampus-dependent spatial memory dysfunction in a water maze test (WMT) and increased depressive-like behavior in a forced swim test (FST). In this study, using a larger cohort of rats exposed to GWIR-chemicals and stress, we investigated whether the memory deficiency identified earlier in a WMT is reproducible with an alternative and stress free hippocampus-dependent memory test such as the object location test (OLT). We also ascertained the possible co-existence of hippocampus-independent memory dysfunction using a novel object recognition test (NORT), and alterations in mood function with additional tests for motivation and depression. Our results provide new evidence that exposure to low doses of GWIR-chemicals and mild stress for 4 weeks causes deficits in hippocampus-dependent object location memory and perirhinal cortex-dependent novel object recognition memory. An open field test performed prior to other behavioral analyses revealed that memory impairments were not associated with increased anxiety or deficits in general motor ability. However, behavioral tests for mood function such as a voluntary physical exercise paradigm and a novelty suppressed feeding test (NSFT) demonstrated decreased motivation levels and depression. Thus, exposure to GWIR-chemicals and stress causes both hippocampus-dependent and hippocampus-independent memory

  17. Image Processing Strategies Based on a Visual Saliency Model for Object Recognition Under Simulated Prosthetic Vision.

    PubMed

    Wang, Jing; Li, Heng; Fu, Weizhen; Chen, Yao; Li, Liming; Lyu, Qing; Han, Tingting; Chai, Xinyu

    2016-01-01

    Retinal prostheses have the potential to restore partial vision. Object recognition in scenes of daily life is one of the essential tasks for implant wearers. Still limited by the low-resolution visual percepts provided by retinal prostheses, it is important to investigate and apply image processing methods to convey more useful visual information to the wearers. We proposed two image processing strategies based on Itti's visual saliency map, region of interest (ROI) extraction, and image segmentation. Itti's saliency model generated a saliency map from the original image, in which salient regions were grouped into ROI by the fuzzy c-means clustering. Then Grabcut generated a proto-object from the ROI labeled image which was recombined with background and enhanced in two ways--8-4 separated pixelization (8-4 SP) and background edge extraction (BEE). Results showed that both 8-4 SP and BEE had significantly higher recognition accuracy in comparison with direct pixelization (DP). Each saliency-based image processing strategy was subject to the performance of image segmentation. Under good and perfect segmentation conditions, BEE and 8-4 SP obtained noticeably higher recognition accuracy than DP, and under bad segmentation condition, only BEE boosted the performance. The application of saliency-based image processing strategies was verified to be beneficial to object recognition in daily scenes under simulated prosthetic vision. They are hoped to help the development of the image processing module for future retinal prostheses, and thus provide more benefit for the patients. Copyright © 2015 International Center for Artificial Organs and Transplantation and Wiley Periodicals, Inc.

  18. The Vasopressin 1b Receptor Antagonist A-988315 Blocks Stress Effects on the Retrieval of Object-Recognition Memory

    PubMed Central

    Barsegyan, Areg; Atsak, Piray; Hornberger, Wilfried B; Jacobson, Peer B; van Gaalen, Marcel M; Roozendaal, Benno

    2015-01-01

    Stress-induced activation of the hypothalamo–pituitary–adrenocortical (HPA) axis and high circulating glucocorticoid levels are well known to impair the retrieval of memory. Vasopressin can activate the HPA axis by stimulating vasopressin 1b (V1b) receptors located on the pituitary. In the present study, we investigated the effect of A-988315, a selective and highly potent non-peptidergic V1b-receptor antagonist with good pharmacokinetic properties, in blocking stress effects on HPA-axis activity and memory retrieval. To study cognitive performance, male Sprague-Dawley rats were trained on an object-discrimination task during which they could freely explore two identical objects. Memory for the objects and their location was tested 24 h later. A-988315 (20 or 60 mg/kg) or water was administered orally 90 min before retention testing, followed 60 min later by stress of footshock exposure. A-988315 dose-dependently dampened stress-induced increases in corticosterone plasma levels, but did not significantly alter HPA-axis activity of non-stressed control rats. Most importantly, A-988315 administration prevented stress-induced impairment of memory retrieval on both the object-recognition and the object-location tasks. A-988315 did not alter the retention of non-stressed rats and did not influence the total time spent exploring the objects or experimental context in either stressed or non-stressed rats. Thus, these findings indicate that direct antagonism of V1b receptors is an effective treatment to block stress-induced activation of the HPA axis and the consequent impairment of retrieval of different aspects of recognition memory. PMID:25669604

  19. Tactile recognition and localization using object models: the case of polyhedra on a plane.

    PubMed

    Gaston, P C; Lozano-Perez, T

    1984-03-01

    This paper discusses how data from multiple tactile sensors may be used to identify and locate one object, from among a set of known objects. We use only local information from sensors: 1) the position of contact points and 2) ranges of surface normals at the contact points. The recognition and localization process is structured as the development and pruning of a tree of consistent hypotheses about pairings between contact points and object surfaces. In this paper, we deal with polyhedral objects constrained to lie on a known plane, i.e., having three degrees of positioning freedom relative to the sensors. We illustrate the performance of the algorithm by simulation.

  20. The effect of scene context on episodic object recognition: parahippocampal cortex mediates memory encoding and retrieval success.

    PubMed

    Hayes, Scott M; Nadel, Lynn; Ryan, Lee

    2007-01-01

    Previous research has investigated intentional retrieval of contextual information and contextual influences on object identification and word recognition, yet few studies have investigated context effects in episodic memory for objects. To address this issue, unique objects embedded in a visually rich scene or on a white background were presented to participants. At test, objects were presented either in the original scene or on a white background. A series of behavioral studies with young adults demonstrated a context shift decrement (CSD)-decreased recognition performance when context is changed between encoding and retrieval. The CSD was not attenuated by encoding or retrieval manipulations, suggesting that binding of object and context may be automatic. A final experiment explored the neural correlates of the CSD, using functional Magnetic Resonance Imaging. Parahippocampal cortex (PHC) activation (right greater than left) during incidental encoding was associated with subsequent memory of objects in the context shift condition. Greater activity in right PHC was also observed during successful recognition of objects previously presented in a scene. Finally, a subset of regions activated during scene encoding, such as bilateral PHC, was reactivated when the object was presented on a white background at retrieval. Although participants were not required to intentionally retrieve contextual information, the results suggest that PHC may reinstate visual context to mediate successful episodic memory retrieval. The CSD is attributed to automatic and obligatory binding of object and context. The results suggest that PHC is important not only for processing of scene information, but also plays a role in successful episodic memory encoding and retrieval. These findings are consistent with the view that spatial information is stored in the hippocampal complex, one of the central tenets of Multiple Trace Theory. (c) 2007 Wiley-Liss, Inc.

  1. Chronic methylphenidate-effects over circadian cycle of young and adult rats submitted to open-field and object recognition tests.

    PubMed

    Gomes, Karin M; Souza, Renan P; Valvassori, Samira S; Réus, Gislaine Z; Inácio, Cecília G; Martins, Márcio R; Comim, Clarissa M; Quevedo, João

    2009-11-01

    In this study age-, circadian rhythm- and methylphenidate administration- effect on open field habituation and object recognition were analyzed. Young and adult male Wistar rats were treated with saline or methylphenidate 2.0 mg/kg for 28 days. Experiments were performed during the light and the dark cycle. Locomotor activity was significantly altered by circadian cycle and methylphenidate treatment during the training session and by drug treatment during the testing session. Exploratory activity was significantly modulated by age during the training session and by age and drug treatment during the testing session. Object recognition memory was altered by cycle at the training session; by age 1.5 h later and by cycle and age 24 h after the training session. These results show that methylphenidate treatment was the major modulator factor on open-field test while cycle and age had an important effect on object recognition experiment.

  2. Tactile Recognition and Localization Using Object Models: The Case of Polyhedra on a Plane.

    DTIC Science & Technology

    1983-03-01

    poor force resolution, but high spatial resolution. We feel that the viability of this recognition approach has important implications on the design of...of the touched object: 1. Surface point - On the basis of sensor readings, some points on the sensor can be identified as being in contact with...the sensor’s shape and location in space are known, one can determine the position of some point on the touched object, to within some uncertainty

  3. Cortical Thickness in Fusiform Face Area Predicts Face and Object Recognition Performance

    PubMed Central

    McGugin, Rankin W.; Van Gulick, Ana E.; Gauthier, Isabel

    2016-01-01

    The fusiform face area (FFA) is defined by its selectivity for faces. Several studies have shown that the response of FFA to non-face objects can predict behavioral performance for these objects. However, one possible account is that experts pay more attention to objects in their domain of expertise, driving signals up. Here we show an effect of expertise with non-face objects in FFA that cannot be explained by differential attention to objects of expertise. We explore the relationship between cortical thickness of FFA and face and object recognition using the Cambridge Face Memory Test and Vanderbilt Expertise Test, respectively. We measured cortical thickness in functionally-defined regions in a group of men who evidenced functional expertise effects for cars in FFA. Performance with faces and objects together accounted for approximately 40% of the variance in cortical thickness of several FFA patches. While subjects with a thicker FFA cortex performed better with vehicles, those with a thinner FFA cortex performed better with faces and living objects. The results point to a domain-general role of FFA in object perception and reveal an interesting double dissociation that does not contrast faces and objects, but rather living and non-living objects. PMID:26439272

  4. Using LabView for real-time monitoring and tracking of multiple biological objects

    NASA Astrophysics Data System (ADS)

    Nikolskyy, Aleksandr I.; Krasilenko, Vladimir G.; Bilynsky, Yosyp Y.; Starovier, Anzhelika

    2017-04-01

    Today real-time studying and tracking of movement dynamics of various biological objects is important and widely researched. Features of objects, conditions of their visualization and model parameters strongly influence the choice of optimal methods and algorithms for a specific task. Therefore, to automate the processes of adaptation of recognition tracking algorithms, several Labview project trackers are considered in the article. Projects allow changing templates for training and retraining the system quickly. They adapt to the speed of objects and statistical characteristics of noise in images. New functions of comparison of images or their features, descriptors and pre-processing methods will be discussed. The experiments carried out to test the trackers on real video files will be presented and analyzed.

  5. Object instance recognition using motion cues and instance specific appearance models

    NASA Astrophysics Data System (ADS)

    Schumann, Arne

    2014-03-01

    In this paper we present an object instance retrieval approach. The baseline approach consists of a pool of image features which are computed on the bounding boxes of a query object track and compared to a database of tracks in order to find additional appearances of the same object instance. We improve over this simple baseline approach in multiple ways: 1) we include motion cues to achieve improved robustness to viewpoint and rotation changes, 2) we include operator feedback to iteratively re-rank the resulting retrieval lists and 3) we use operator feedback and location constraints to train classifiers and learn an instance specific appearance model. We use these classifiers to further improve the retrieval results. The approach is evaluated on two popular public datasets for two different applications. We evaluate person re-identification on the CAVIAR shopping mall surveillance dataset and vehicle instance recognition on the VIVID aerial dataset and achieve significant improvements over our baseline results.

  6. Dopamine receptor activity participates in hippocampal synaptic plasticity associated with novel object recognition.

    PubMed

    Yang, Kechun; Broussard, John I; Levine, Amber T; Jenson, Daniel; Arenkiel, Benjamin R; Dani, John A

    2017-01-01

    Physiological and behavioral evidence supports that dopamine (DA) receptor signaling influences hippocampal function. While several recent studies examined how DA influences CA1 plasticity and learning, there are fewer studies investigating the influence of DA signaling to the dentate gyrus. The dentate gyrus receives convergent cortical input through the perforant path fiber tracts and has been conceptualized to detect novelty in spatial memory tasks. To test whether DA-receptor activity influences novelty-detection, we used a novel object recognition (NOR) task where mice remember previously presented objects as an indication of learning. Although DA innervation arises from other sources and the main DA signaling may be from those sources, our molecular approaches verified that midbrain dopaminergic fibers also sparsely innervate the dentate gyrus. During the NOR task, wild-type mice spent significantly more time investigating novel objects rather than previously observed objects. Dentate granule cells in slices cut from those mice showed an increased AMPA/NMDA-receptor current ratio indicative of potentiated synaptic transmission. Post-training injection of a D1-like receptor antagonist not only effectively blocked the preference for the novel objects, but also prevented the increased AMPA/NMDA ratio. Consistent with that finding, neither NOR learning nor the increase in the AMPA/NMDA ratio were observed in DA-receptor KO mice under the same experimental conditions. The results indicate that DA-receptor signaling contributes to the successful completion of the NOR task and to the associated synaptic plasticity of the dentate gyrus that likely contributes to the learning. © 2016 Federation of European Neuroscience Societies and John Wiley & Sons Ltd.

  7. Mechanisms and Neural Basis of Object and Pattern Recognition: A Study with Chess Experts

    ERIC Educational Resources Information Center

    Bilalic, Merim; Langner, Robert; Erb, Michael; Grodd, Wolfgang

    2010-01-01

    Comparing experts with novices offers unique insights into the functioning of cognition, based on the maximization of individual differences. Here we used this expertise approach to disentangle the mechanisms and neural basis behind two processes that contribute to everyday expertise: object and pattern recognition. We compared chess experts and…

  8. Implementation of a Peltier-based cooling device for localized deep cortical deactivation during in vivo object recognition testing

    NASA Astrophysics Data System (ADS)

    Marra, Kyle; Graham, Brett; Carouso, Samantha; Cox, David

    2012-02-01

    While the application of local cortical cooling has recently become a focus of neurological research, extended localized deactivation deep within brain structures is still unexplored. Using a wirelessly controlled thermoelectric (Peltier) device and water-based heat sink, we have achieved inactivating temperatures (<20 C) at greater depths (>8 mm) than previously reported. After implanting the device into Long Evans rats' basolateral amygdala (BLA), an inhibitory brain center that controls anxiety and fear, we ran an open field test during which anxiety-driven behavioral tendencies were observed to decrease during cooling, thus confirming the device's effect on behavior. Our device will next be implanted in the rats' temporal association cortex (TeA) and recordings from our signal-tracing multichannel microelectrodes will measure and compare activated and deactivated neuronal activity so as to isolate and study the TeA signals responsible for object recognition. Having already achieved a top performing computational face-recognition system, the lab will utilize this TeA activity data to generalize its computational efforts of face recognition to achieve general object recognition.

  9. OPTICAL INFORMATION PROCESSING: Synthesis of an object recognition system based on the profile of the envelope of a laser pulse in pulsed lidars

    NASA Astrophysics Data System (ADS)

    Buryi, E. V.

    1998-05-01

    The main problems in the synthesis of an object recognition system, based on the principles of operation of neuron networks, are considered. Advantages are demonstrated of a hierarchical structure of the recognition algorithm. The use of reading of the amplitude spectrum of signals as information tags is justified and a method is developed for determination of the dimensionality of the tag space. Methods are suggested for ensuring the stability of object recognition in the optical range. It is concluded that it should be possible to recognise perspectives of complex objects.

  10. Better object recognition and naming outcome with MRI-guided stereotactic laser amygdalohippocampotomy for temporal lobe epilepsy.

    PubMed

    Drane, Daniel L; Loring, David W; Voets, Natalie L; Price, Michele; Ojemann, Jeffrey G; Willie, Jon T; Saindane, Amit M; Phatak, Vaishali; Ivanisevic, Mirjana; Millis, Scott; Helmers, Sandra L; Miller, John W; Meador, Kimford J; Gross, Robert E

    2015-01-01

    Patients with temporal lobe epilepsy (TLE) experience significant deficits in category-related object recognition and naming following standard surgical approaches. These deficits may result from a decoupling of core processing modules (e.g., language, visual processing, and semantic memory), due to "collateral damage" to temporal regions outside the hippocampus following open surgical approaches. We predicted that stereotactic laser amygdalohippocampotomy (SLAH) would minimize such deficits because it preserves white matter pathways and neocortical regions that are critical for these cognitive processes. Tests of naming and recognition of common nouns (Boston Naming Test) and famous persons were compared with nonparametric analyses using exact tests between a group of 19 patients with medically intractable mesial TLE undergoing SLAH (10 dominant, 9 nondominant), and a comparable series of TLE patients undergoing standard surgical approaches (n=39) using a prospective, nonrandomized, nonblinded, parallel-group design. Performance declines were significantly greater for the patients with dominant TLE who were undergoing open resection versus SLAH for naming famous faces and common nouns (F=24.3, p<0.0001, η2=0.57, and F=11.2, p<0.001, η2=0.39, respectively), and for the patients with nondominant TLE undergoing open resection versus SLAH for recognizing famous faces (F=3.9, p<0.02, η2=0.19). When examined on an individual subject basis, no SLAH patients experienced any performance declines on these measures. In contrast, 32 of the 39 patients undergoing standard surgical approaches declined on one or more measures for both object types (p<0.001, Fisher's exact test). Twenty-one of 22 left (dominant) TLE patients declined on one or both naming tasks after open resection, while 11 of 17 right (nondominant) TLE patients declined on face recognition. Preliminary results suggest (1) naming and recognition functions can be spared in TLE patients undergoing SLAH, and (2

  11. Combined Dynamic Time Warping with Multiple Sensors for 3D Gesture Recognition

    PubMed Central

    2017-01-01

    Cyber-physical systems, which closely integrate physical systems and humans, can be applied to a wider range of applications through user movement analysis. In three-dimensional (3D) gesture recognition, multiple sensors are required to recognize various natural gestures. Several studies have been undertaken in the field of gesture recognition; however, gesture recognition was conducted based on data captured from various independent sensors, which rendered the capture and combination of real-time data complicated. In this study, a 3D gesture recognition method using combined information obtained from multiple sensors is proposed. The proposed method can robustly perform gesture recognition regardless of a user’s location and movement directions by providing viewpoint-weighted values and/or motion-weighted values. In the proposed method, the viewpoint-weighted dynamic time warping with multiple sensors has enhanced performance by preventing joint measurement errors and noise due to sensor measurement tolerance, which has resulted in the enhancement of recognition performance by comparing multiple joint sequences effectively. PMID:28817094

  12. Combined Dynamic Time Warping with Multiple Sensors for 3D Gesture Recognition.

    PubMed

    Choi, Hyo-Rim; Kim, TaeYong

    2017-08-17

    Cyber-physical systems, which closely integrate physical systems and humans, can be applied to a wider range of applications through user movement analysis. In three-dimensional (3D) gesture recognition, multiple sensors are required to recognize various natural gestures. Several studies have been undertaken in the field of gesture recognition; however, gesture recognition was conducted based on data captured from various independent sensors, which rendered the capture and combination of real-time data complicated. In this study, a 3D gesture recognition method using combined information obtained from multiple sensors is proposed. The proposed method can robustly perform gesture recognition regardless of a user's location and movement directions by providing viewpoint-weighted values and/or motion-weighted values. In the proposed method, the viewpoint-weighted dynamic time warping with multiple sensors has enhanced performance by preventing joint measurement errors and noise due to sensor measurement tolerance, which has resulted in the enhancement of recognition performance by comparing multiple joint sequences effectively.

  13. Logarithmic r-θ mapping for hybrid optical neural network filter for multiple objects recognition within cluttered scenes

    NASA Astrophysics Data System (ADS)

    Kypraios, Ioannis; Young, Rupert C. D.; Chatwin, Chris R.; Birch, Phil M.

    2009-04-01

    θThe window unit in the design of the complex logarithmic r-θ mapping for hybrid optical neural network filter can allow multiple objects of the same class to be detected within the input image. Additionally, the architecture of the neural network unit of the complex logarithmic r-θ mapping for hybrid optical neural network filter becomes attractive for accommodating the recognition of multiple objects of different classes within the input image by modifying the output layer of the unit. We test the overall filter for multiple objects of the same and of different classes' recognition within cluttered input images and video sequences of cluttered scenes. Logarithmic r-θ mapping for hybrid optical neural network filter is shown to exhibit with a single pass over the input data simultaneously in-plane rotation, out-of-plane rotation, scale, log r-θ map translation and shift invariance, and good clutter tolerance by recognizing correctly the different objects within the cluttered scenes. We record in our results additional extracted information from the cluttered scenes about the objects' relative position, scale and in-plane rotation.

  14. Depth rotation and mirror-image reflection reduce affective preference as well as recognition memory for pictures of novel objects.

    PubMed

    Lawson, Rebecca

    2004-10-01

    In two experiments, the identification of novel 3-D objects was worse for depth-rotated and mirror-reflected views, compared with the study view in an implicit affective preference memory task, as well as in an explicit recognition memory task. In Experiment 1, recognition was worse and preference was lower when depth-rotated views of an object were paired with an unstudied object relative to trials when the study view of that object was shown. There was a similar trend for mirror-reflected views. In Experiment 2, the study view of an object was both recognized and preferred above chance when it was paired with either depth-rotated or mirror-reflected views of that object. These results suggest that view-sensitive representations of objects mediate performance in implicit, as well as explicit, memory tasks. The findings do not support the claim that separate episodic and structural description representations underlie performance in implicit and explicit memory tasks, respectively.

  15. Short-term blueberry-enriched diet prevents and reverses object recognition memory loss in aging rats.

    PubMed

    Malin, David H; Lee, David R; Goyarzu, Pilar; Chang, Yu-Hsuan; Ennis, Lalanya J; Beckett, Elizabeth; Shukitt-Hale, Barbara; Joseph, James A

    2011-03-01

    Previously, 4 mo of a blueberry-enriched (BB) antioxidant diet prevented impaired object recognition memory in aging rats. Experiment 1 determined whether 1- and 2-mo BB diets would have a similar effect and whether the benefits would disappear promptly after terminating the diets. Experiment 2 determined whether a 1-mo BB diet could subsequently reverse existing object memory impairment in aging rats. In experiment 1, Fischer-344 rats were maintained on an appropriate control diet or on 1 or 2 mo of the BB diet before testing object memory at 19 mo postnatally. In experiment 2, rats were tested for object recognition memory at 19 mo and again at 20 mo after 1 mo of maintenance on a 2% BB or control diet. In experiment 1, the control group performed no better than chance, whereas the 1- and 2-mo BB diet groups performed similarly and significantly better than controls. The 2-mo BB-diet group, but not the 1-mo group, maintained its performance over a subsequent month on a standard laboratory diet. In experiment 2, the 19-mo-old rats performed near chance. At 20 mo of age, the rats subsequently maintained on the BB diet significantly increased their object memory scores, whereas the control diet group exhibited a non-significant decline. The change in object memory scores differed significantly between the two diet groups. These results suggest that a considerable degree of age-related object memory decline can be prevented and reversed by brief maintenance on BB diets. Copyright © 2011 Elsevier Inc. All rights reserved.

  16. Ignorance- versus Evidence-Based Decision Making: A Decision Time Analysis of the Recognition Heuristic

    ERIC Educational Resources Information Center

    Hilbig, Benjamin E.; Pohl, Rudiger F.

    2009-01-01

    According to part of the adaptive toolbox notion of decision making known as the recognition heuristic (RH), the decision process in comparative judgments--and its duration--is determined by whether recognition discriminates between objects. By contrast, some recently proposed alternative models predict that choices largely depend on the amount of…

  17. Unsupervised and self-mapping category formation and semantic object recognition for mobile robot vision used in an actual environment

    NASA Astrophysics Data System (ADS)

    Madokoro, H.; Tsukada, M.; Sato, K.

    2013-07-01

    This paper presents an unsupervised learning-based object category formation and recognition method for mobile robot vision. Our method has the following features: detection of feature points and description of features using a scale-invariant feature transform (SIFT), selection of target feature points using one class support vector machines (OC-SVMs), generation of visual words using self-organizing maps (SOMs), formation of labels using adaptive resonance theory 2 (ART-2), and creation and classification of categories on a category map of counter propagation networks (CPNs) for visualizing spatial relations between categories. Classification results of dynamic images using time-series images obtained using two different-size robots and according to movements respectively demonstrate that our method can visualize spatial relations of categories while maintaining time-series characteristics. Moreover, we emphasize the effectiveness of our method for category formation of appearance changes of objects.

  18. Category Specificity in Normal Episodic Learning: Applications to Object Recognition and Category-Specific Agnosia

    ERIC Educational Resources Information Center

    Bukach, Cindy M.; Bub, Daniel N.; Masson, Michael E. J.; Lindsay, D. Stephen

    2004-01-01

    Studies of patients with category-specific agnosia (CSA) have given rise to multiple theories of object recognition, most of which assume the existence of a stable, abstract semantic memory system. We applied an episodic view of memory to questions raised by CSA in a series of studies examining normal observers' recall of newly learned attributes…

  19. Pattern recognition with composite correlation filters designed with multi-object combinatorial optimization

    DOE PAGES

    Awwal, Abdul; Diaz-Ramirez, Victor H.; Cuevas, Andres; ...

    2014-10-23

    Composite correlation filters are used for solving a wide variety of pattern recognition problems. These filters are given by a combination of several training templates chosen by a designer in an ad hoc manner. In this work, we present a new approach for the design of composite filters based on multi-objective combinatorial optimization. Given a vast search space of training templates, an iterative algorithm is used to synthesize a filter with an optimized performance in terms of several competing criteria. Furthermore, by employing a suggested binary-search procedure a filter bank with a minimum number of filters can be constructed, formore » a prespecified trade-off of performance metrics. Computer simulation results obtained with the proposed method in recognizing geometrically distorted versions of a target in cluttered and noisy scenes are discussed and compared in terms of recognition performance and complexity with existing state-of-the-art filters.« less

  20. Pattern recognition with composite correlation filters designed with multi-object combinatorial optimization

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Awwal, Abdul; Diaz-Ramirez, Victor H.; Cuevas, Andres

    Composite correlation filters are used for solving a wide variety of pattern recognition problems. These filters are given by a combination of several training templates chosen by a designer in an ad hoc manner. In this work, we present a new approach for the design of composite filters based on multi-objective combinatorial optimization. Given a vast search space of training templates, an iterative algorithm is used to synthesize a filter with an optimized performance in terms of several competing criteria. Furthermore, by employing a suggested binary-search procedure a filter bank with a minimum number of filters can be constructed, formore » a prespecified trade-off of performance metrics. Computer simulation results obtained with the proposed method in recognizing geometrically distorted versions of a target in cluttered and noisy scenes are discussed and compared in terms of recognition performance and complexity with existing state-of-the-art filters.« less

  1. Aging and solid shape recognition: Vision and haptics.

    PubMed

    Norman, J Farley; Cheeseman, Jacob R; Adkins, Olivia C; Cox, Andrea G; Rogers, Connor E; Dowell, Catherine J; Baxter, Michael W; Norman, Hideko F; Reyes, Cecia M

    2015-10-01

    The ability of 114 younger and older adults to recognize naturally-shaped objects was evaluated in three experiments. The participants viewed or haptically explored six randomly-chosen bell peppers (Capsicum annuum) in a study session and were later required to judge whether each of twelve bell peppers was "old" (previously presented during the study session) or "new" (not presented during the study session). When recognition memory was tested immediately after study, the younger adults' (Experiment 1) performance for vision and haptics was identical when the individual study objects were presented once. Vision became superior to haptics, however, when the individual study objects were presented multiple times. When 10- and 20-min delays (Experiment 2) were inserted in between study and test sessions, no significant differences occurred between vision and haptics: recognition performance in both modalities was comparable. When the recognition performance of older adults was evaluated (Experiment 3), a negative effect of age was found for visual shape recognition (younger adults' overall recognition performance was 60% higher). There was no age effect, however, for haptic shape recognition. The results of the present experiments indicate that the visual recognition of natural object shape is different from haptic recognition in multiple ways: visual shape recognition can be superior to that of haptics and is affected by aging, while haptic shape recognition is less accurate and unaffected by aging. Copyright © 2015 Elsevier Ltd. All rights reserved.

  2. On the Time Course of Vocal Emotion Recognition

    PubMed Central

    Pell, Marc D.; Kotz, Sonja A.

    2011-01-01

    How quickly do listeners recognize emotions from a speaker's voice, and does the time course for recognition vary by emotion type? To address these questions, we adapted the auditory gating paradigm to estimate how much vocal information is needed for listeners to categorize five basic emotions (anger, disgust, fear, sadness, happiness) and neutral utterances produced by male and female speakers of English. Semantically-anomalous pseudo-utterances (e.g., The rivix jolled the silling) conveying each emotion were divided into seven gate intervals according to the number of syllables that listeners heard from sentence onset. Participants (n = 48) judged the emotional meaning of stimuli presented at each gate duration interval, in a successive, blocked presentation format. Analyses looked at how recognition of each emotion evolves as an utterance unfolds and estimated the “identification point” for each emotion. Results showed that anger, sadness, fear, and neutral expressions are recognized more accurately at short gate intervals than happiness, and particularly disgust; however, as speech unfolds, recognition of happiness improves significantly towards the end of the utterance (and fear is recognized more accurately than other emotions). When the gate associated with the emotion identification point of each stimulus was calculated, data indicated that fear (M = 517 ms), sadness (M = 576 ms), and neutral (M = 510 ms) expressions were identified from shorter acoustic events than the other emotions. These data reveal differences in the underlying time course for conscious recognition of basic emotions from vocal expressions, which should be accounted for in studies of emotional speech processing. PMID:22087275

  3. Erythropoietin improves object placement recognition memory in a time dependent manner in both, uninjured animals and fimbria-fornix-lesioned male rats.

    PubMed

    Almaguer-Melian, W; Mercerón-Martinez, D; Delgado-Ocaña, S; Alberti-Amador, E; Gonzalez-Gómez, R; Bergado, Jorge A

    2018-04-01

    An increasing number of reports sustain a possible role of erythropoietin (EPO) as neuroprotective agent. In two previous articles we have evaluated EPO as plasticity promoting agent, and to contribute the restoration of brain function affected by acquired damage. We have shown that EPO is able to induce an increased synaptic efficacy in vivo along with a plasticity promoting effect. In the Morris water maze EPO administration to fimbria-fornix lesioned male rats induces a significant improvement of their spatial memory, affected by the lesion. Singularly, EPO was only effective when administered shortly after training (10 min) but not after several hours (5 h), suggesting a specific EPO effect on time dependent plasticity process. In the present paper we have expanded this line of evidence using a low stress paradigm of object placement recognition in lesioned and healthy male rats. The memory trace in this model is short-lasting; animals could recognize the change in object position when tested 24 h after, but not 48 or 72 h after the acquisition session. EPO administration 10 min after acquisition significantly prolongs retention to, at least, 72 h in healthy rats. No effect was seen if EPO was administered 5 h after training, suggesting a specific EPO modulatory effect on the consolidation process. Remarkably, early EPO treatment to fimbria fornix lesioned animals reverts the memory deficit caused by the lesion. An increased expression of the plasticity related gene arc, was also confirmed in the hippocampus and the prefrontal cortex, that is likely to be involved in the behavioral improvement observed. Copyright © 2018 Elsevier Inc. All rights reserved.

  4. Real-Time Pattern Recognition - An Industrial Example

    NASA Astrophysics Data System (ADS)

    Fitton, Gary M.

    1981-11-01

    Rapid advancements in cost effective sensors and micro computers are now making practical the on-line implementation of pattern recognition based systems for a variety of industrial applications requiring high processing speeds. One major application area for real time pattern recognition is in the sorting of packaged/cartoned goods at high speed for automated warehousing and return goods cataloging. While there are many OCR and bar code readers available to perform these functions, it is often impractical to use such codes (package too small, adverse esthetics, poor print quality) and an approach which recognizes an item by its graphic content alone is desirable. This paper describes a specific application within the tobacco industry, that of sorting returned cigarette goods by brand and size.

  5. Divergent short- and long-term effects of acute stress in object recognition memory are mediated by endogenous opioid system activation.

    PubMed

    Nava-Mesa, Mauricio O; Lamprea, Marisol R; Múnera, Alejandro

    2013-11-01

    Acute stress induces short-term object recognition memory impairment and elicits endogenous opioid system activation. The aim of this study was thus to evaluate whether opiate system activation mediates the acute stress-induced object recognition memory changes. Adult male Wistar rats were trained in an object recognition task designed to test both short- and long-term memory. Subjects were randomly assigned to receive an intraperitoneal injection of saline, 1 mg/kg naltrexone or 3 mg/kg naltrexone, four and a half hours before the sample trial. Five minutes after the injection, half the subjects were submitted to movement restraint during four hours while the other half remained in their home cages. Non-stressed subjects receiving saline (control) performed adequately during the short-term memory test, while stressed subjects receiving saline displayed impaired performance. Naltrexone prevented such deleterious effect, in spite of the fact that it had no intrinsic effect on short-term object recognition memory. Stressed subjects receiving saline and non-stressed subjects receiving naltrexone performed adequately during the long-term memory test; however, control subjects as well as stressed subjects receiving a high dose of naltrexone performed poorly. Control subjects' dissociated performance during both memory tests suggests that the short-term memory test induced a retroactive interference effect mediated through light opioid system activation; such effect was prevented either by low dose naltrexone administration or by strongly activating the opioid system through acute stress. Both short-term memory retrieval impairment and long-term memory improvement observed in stressed subjects may have been mediated through strong opioid system activation, since they were prevented by high dose naltrexone administration. Therefore, the activation of the opioid system plays a dual modulating role in object recognition memory. Copyright © 2013 Elsevier Inc. All rights

  6. Heterozygous Che-1 KO mice show deficiencies in object recognition memory persistence.

    PubMed

    Zalcman, Gisela; Corbi, Nicoletta; Di Certo, Maria Grazia; Mattei, Elisabetta; Federman, Noel; Romano, Arturo

    2016-10-06

    Transcriptional regulation is a key process in the formation of long-term memories. Che-1 is a protein involved in the regulation of gene transcription that has recently been proved to bind the transcription factor NF-κB, which is known to be involved in many memory-related molecular events. This evidence prompted us to investigate the putative role of Che-1 in memory processes. For this study we newly generated a line of Che-1(+/-) heterozygous mice. Che-1 homozygous KO mouse is lethal during development, but Che-1(+/-) heterozygous mouse is normal in its general anatomical and physiological characteristics. We analyzed the behavioral characteristic and memory performance of Che-1(+/-) mice in two NF-κB dependent types of memory. We found that Che-1(+/-) mice show similar locomotor activity and thigmotactic behavior than wild type (WT) mice in an open field. In a similar way, no differences were found in anxiety-like behavior between Che-1(+/-) and WT mice in an elevated plus maze as well as in fear response in a contextual fear conditioning (CFC) and object exploration in a novel object recognition (NOR) task. No differences were found between WT and Che-1(+/-) mice performance in CFC training and when tested at 24h or 7days after training. Similar performance was found between groups in NOR task, both in training and 24h testing performance. However, we found that object recognition memory persistence at 7days was impaired in Che-1(+/-) heterozygous mice. This is the first evidence showing that Che-1 is involved in memory processes. Copyright © 2016 Elsevier Ireland Ltd. All rights reserved.

  7. Age and measurement time-of-day effects on speech recognition in noise.

    PubMed

    Veneman, Carrie E; Gordon-Salant, Sandra; Matthews, Lois J; Dubno, Judy R

    2013-01-01

    The purpose of this study was to determine the effect of measurement time of day on speech recognition in noise and the extent to which time-of-day effects differ with age. Older adults tend to have more difficulty understanding speech in noise than younger adults, even when hearing is normal. Two possible contributors to this age difference in speech recognition may be measurement time of day and inhibition. Most younger adults are "evening-type," showing peak circadian arousal in the evening, whereas most older adults are "morning-type," with circadian arousal peaking in the morning. Tasks that require inhibition of irrelevant information have been shown to be affected by measurement time of day, with maximum performance attained at one's peak time of day. The authors hypothesized that a change in inhibition will be associated with measurement time of day and therefore affect speech recognition in noise, with better performance in the morning for older adults and in the evening for younger adults. Fifteen younger evening-type adults (20-28 years) and 15 older morning-type adults with normal hearing (66-78 years) listened to the Hearing in Noise Test (HINT) and the Quick Speech in Noise (QuickSIN) test in the morning and evening (peak and off-peak times). Time of day preference was assessed using the Morningness-Eveningness Questionnaire. Sentences and noise were presented binaurally through insert earphones. During morning and evening sessions, participants solved word-association problems within the visual-distraction task (VDT), which was used as an estimate of inhibition. After each session, participants rated perceived mental demand of the tasks using a revised version of the NASA Task Load Index. Younger adults performed significantly better on the speech-in-noise tasks and rated themselves as requiring significantly less mental demand when tested at their peak (evening) than off-peak (morning) time of day. In contrast, time-of-day effects were not observed

  8. Real-Time (Vision-Based) Road Sign Recognition Using an Artificial Neural Network.

    PubMed

    Islam, Kh Tohidul; Raj, Ram Gopal

    2017-04-13

    Road sign recognition is a driver support function that can be used to notify and warn the driver by showing the restrictions that may be effective on the current stretch of road. Examples for such regulations are 'traffic light ahead' or 'pedestrian crossing' indications. The present investigation targets the recognition of Malaysian road and traffic signs in real-time. Real-time video is taken by a digital camera from a moving vehicle and real world road signs are then extracted using vision-only information. The system is based on two stages, one performs the detection and another one is for recognition. In the first stage, a hybrid color segmentation algorithm has been developed and tested. In the second stage, an introduced robust custom feature extraction method is used for the first time in a road sign recognition approach. Finally, a multilayer artificial neural network (ANN) has been created to recognize and interpret various road signs. It is robust because it has been tested on both standard and non-standard road signs with significant recognition accuracy. This proposed system achieved an average of 99.90% accuracy with 99.90% of sensitivity, 99.90% of specificity, 99.90% of f-measure, and 0.001 of false positive rate (FPR) with 0.3 s computational time. This low FPR can increase the system stability and dependability in real-time applications.

  9. Real-Time (Vision-Based) Road Sign Recognition Using an Artificial Neural Network

    PubMed Central

    Islam, Kh Tohidul; Raj, Ram Gopal

    2017-01-01

    Road sign recognition is a driver support function that can be used to notify and warn the driver by showing the restrictions that may be effective on the current stretch of road. Examples for such regulations are ‘traffic light ahead’ or ‘pedestrian crossing’ indications. The present investigation targets the recognition of Malaysian road and traffic signs in real-time. Real-time video is taken by a digital camera from a moving vehicle and real world road signs are then extracted using vision-only information. The system is based on two stages, one performs the detection and another one is for recognition. In the first stage, a hybrid color segmentation algorithm has been developed and tested. In the second stage, an introduced robust custom feature extraction method is used for the first time in a road sign recognition approach. Finally, a multilayer artificial neural network (ANN) has been created to recognize and interpret various road signs. It is robust because it has been tested on both standard and non-standard road signs with significant recognition accuracy. This proposed system achieved an average of 99.90% accuracy with 99.90% of sensitivity, 99.90% of specificity, 99.90% of f-measure, and 0.001 of false positive rate (FPR) with 0.3 s computational time. This low FPR can increase the system stability and dependability in real-time applications. PMID:28406471

  10. HOTS: A Hierarchy of Event-Based Time-Surfaces for Pattern Recognition.

    PubMed

    Lagorce, Xavier; Orchard, Garrick; Galluppi, Francesco; Shi, Bertram E; Benosman, Ryad B

    2017-07-01

    This paper describes novel event-based spatio-temporal features called time-surfaces and how they can be used to create a hierarchical event-based pattern recognition architecture. Unlike existing hierarchical architectures for pattern recognition, the presented model relies on a time oriented approach to extract spatio-temporal features from the asynchronously acquired dynamics of a visual scene. These dynamics are acquired using biologically inspired frameless asynchronous event-driven vision sensors. Similarly to cortical structures, subsequent layers in our hierarchy extract increasingly abstract features using increasingly large spatio-temporal windows. The central concept is to use the rich temporal information provided by events to create contexts in the form of time-surfaces which represent the recent temporal activity within a local spatial neighborhood. We demonstrate that this concept can robustly be used at all stages of an event-based hierarchical model. First layer feature units operate on groups of pixels, while subsequent layer feature units operate on the output of lower level feature units. We report results on a previously published 36 class character recognition task and a four class canonical dynamic card pip task, achieving near 100 percent accuracy on each. We introduce a new seven class moving face recognition task, achieving 79 percent accuracy.This paper describes novel event-based spatio-temporal features called time-surfaces and how they can be used to create a hierarchical event-based pattern recognition architecture. Unlike existing hierarchical architectures for pattern recognition, the presented model relies on a time oriented approach to extract spatio-temporal features from the asynchronously acquired dynamics of a visual scene. These dynamics are acquired using biologically inspired frameless asynchronous event-driven vision sensors. Similarly to cortical structures, subsequent layers in our hierarchy extract increasingly abstract

  11. Object recognition using deep convolutional neural networks with complete transfer and partial frozen layers

    NASA Astrophysics Data System (ADS)

    Kruithof, Maarten C.; Bouma, Henri; Fischer, Noëlle M.; Schutte, Klamer

    2016-10-01

    Object recognition is important to understand the content of video and allow flexible querying in a large number of cameras, especially for security applications. Recent benchmarks show that deep convolutional neural networks are excellent approaches for object recognition. This paper describes an approach of domain transfer, where features learned from a large annotated dataset are transferred to a target domain where less annotated examples are available as is typical for the security and defense domain. Many of these networks trained on natural images appear to learn features similar to Gabor filters and color blobs in the first layer. These first-layer features appear to be generic for many datasets and tasks while the last layer is specific. In this paper, we study the effect of copying all layers and fine-tuning a variable number. We performed an experiment with a Caffe-based network on 1000 ImageNet classes that are randomly divided in two equal subgroups for the transfer from one to the other. We copy all layers and vary the number of layers that is fine-tuned and the size of the target dataset. We performed additional experiments with the Keras platform on CIFAR-10 dataset to validate general applicability. We show with both platforms and both datasets that the accuracy on the target dataset improves when more target data is used. When the target dataset is large, it is beneficial to freeze only a few layers. For a large target dataset, the network without transfer learning performs better than the transfer network, especially if many layers are frozen. When the target dataset is small, it is beneficial to transfer (and freeze) many layers. For a small target dataset, the transfer network boosts generalization and it performs much better than the network without transfer learning. Learning time can be reduced by freezing many layers in a network.

  12. A Scientific Workflow Platform for Generic and Scalable Object Recognition on Medical Images

    NASA Astrophysics Data System (ADS)

    Möller, Manuel; Tuot, Christopher; Sintek, Michael

    In the research project THESEUS MEDICO we aim at a system combining medical image information with semantic background knowledge from ontologies to give clinicians fully cross-modal access to biomedical image repositories. Therefore joint efforts have to be made in more than one dimension: Object detection processes have to be specified in which an abstraction is performed starting from low-level image features across landmark detection utilizing abstract domain knowledge up to high-level object recognition. We propose a system based on a client-server extension of the scientific workflow platform Kepler that assists the collaboration of medical experts and computer scientists during development and parameter learning.

  13. Visual object recognition for automatic micropropagation of plants

    NASA Astrophysics Data System (ADS)

    Brendel, Thorsten; Schwanke, Joerg; Jensch, Peter F.

    1994-11-01

    Micropropagation of plants is done by cutting juvenile plants and placing them into special container-boxes with nutrient-solution where the pieces can grow up and be cut again several times. To produce high amounts of biomass it is necessary to do plant micropropagation by a robotic system. In this paper we describe parts of the vision system that recognizes plants and their particular cutting points. Therefore, it is necessary to extract elements of the plants and relations between these elements (for example root, stem, leaf). Different species vary in their morphological appearance, variation is also immanent in plants of the same species. Therefore, we introduce several morphological classes of plants from that we expect same recognition methods.

  14. Improving human object recognition performance using video enhancement techniques

    NASA Astrophysics Data System (ADS)

    Whitman, Lucy S.; Lewis, Colin; Oakley, John P.

    2004-12-01

    Atmospheric scattering causes significant degradation in the quality of video images, particularly when imaging over long distances. The principle problem is the reduction in contrast due to scattered light. It is known that when the scattering particles are not too large compared with the imaging wavelength (i.e. Mie scattering) then high spatial resolution information may be contained within a low-contrast image. Unfortunately this information is not easily perceived by a human observer, particularly when using a standard video monitor. A secondary problem is the difficulty of achieving a sharp focus since automatic focus techniques tend to fail in such conditions. Recently several commercial colour video processing systems have become available. These systems use various techniques to improve image quality in low contrast conditions whilst retaining colour content. These systems produce improvements in subjective image quality in some situations, particularly in conditions of haze and light fog. There is also some evidence that video enhancement leads to improved ATR performance when used as a pre-processing stage. Psychological literature indicates that low contrast levels generally lead to a reduction in the performance of human observers in carrying out simple visual tasks. The aim of this paper is to present the results of an empirical study on object recognition in adverse viewing conditions. The chosen visual task was vehicle number plate recognition at long ranges (500 m and beyond). Two different commercial video enhancement systems are evaluated using the same protocol. The results show an increase in effective range with some differences between the different enhancement systems.

  15. The Role of Competitive Inhibition and Top-Down Feedback in Binding during Object Recognition

    PubMed Central

    Wyatte, Dean; Herd, Seth; Mingus, Brian; O’Reilly, Randall

    2012-01-01

    How does the brain bind together visual features that are processed concurrently by different neurons into a unified percept suitable for processes such as object recognition? Here, we describe how simple, commonly accepted principles of neural processing can interact over time to solve the brain’s binding problem. We focus on mechanisms of neural inhibition and top-down feedback. Specifically, we describe how inhibition creates competition among neural populations that code different features, effectively suppressing irrelevant information, and thus minimizing illusory conjunctions. Top-down feedback contributes to binding in a similar manner, but by reinforcing relevant features. Together, inhibition and top-down feedback contribute to a competitive environment that ensures only the most appropriate features are bound together. We demonstrate this overall proposal using a biologically realistic neural model of vision that processes features across a hierarchy of interconnected brain areas. Finally, we argue that temporal synchrony plays only a limited role in binding – it does not simultaneously bind multiple objects, but does aid in creating additional contrast between relevant and irrelevant features. Thus, our overall theory constitutes a solution to the binding problem that relies only on simple neural principles without any binding-specific processes. PMID:22719733

  16. Effect of Time Delay on Recognition Memory for Pictures: The Modulatory Role of Emotion

    PubMed Central

    Wang, Bo

    2014-01-01

    This study investigated the modulatory role of emotion in the effect of time delay on recognition memory for pictures. Participants viewed neutral, positive and negative pictures, and took a recognition memory test 5 minutes, 24 hours, or 1 week after learning. The findings are: 1) For neutral, positive and negative pictures, overall recognition accuracy in the 5-min delay did not significantly differ from that in the 24-h delay. For neutral and positive pictures, overall recognition accuracy in the 1-week delay was lower than in the 24-h delay; for negative pictures, overall recognition in the 24-h and 1-week delay did not significantly differ. Therefore negative emotion modulates the effect of time delay on recognition memory, maintaining retention of overall recognition accuracy only within a certain frame of time. 2) For the three types of pictures, recollection and familiarity in the 5-min delay did not significantly differ from that in the 24-h and the 1-week delay. Thus emotion does not appear to modulate the effect of time delay on recollection and familiarity. However, recollection in the 24-h delay was higher than in the 1-week delay, whereas familiarity in the 24-h delay was lower than in the 1-week delay. PMID:24971457

  17. Scene recognition following locomotion around a scene.

    PubMed

    Motes, Michael A; Finlay, Cory A; Kozhevnikov, Maria

    2006-01-01

    Effects of locomotion on scene-recognition reaction time (RT) and accuracy were studied. In experiment 1, observers memorized an 11-object scene and made scene-recognition judgments on subsequently presented scenes from the encoded view or different views (ie scenes were rotated or observers moved around the scene, both from 40 degrees to 360 degrees). In experiment 2, observers viewed different 5-object scenes on each trial and made scene-recognition judgments from the encoded view or after moving around the scene, from 36 degrees to 180 degrees. Across experiments, scene-recognition RT increased (in experiment 2 accuracy decreased) with angular distance between encoded and judged views, regardless of how the viewpoint changes occurred. The findings raise questions about conditions in which locomotion produces spatially updated representations of scenes.

  18. Physical Exercise During Adolescence Versus Adulthood: Differential Effects on Object Recognition Memory and BDNF Levels

    PubMed Central

    Hopkins, Michael E.; Nitecki, Roni; Bucci, David J.

    2011-01-01

    It is well established that physical exercise can enhance hippocampal-dependent forms of learning and memory in laboratory animals, commensurate with increases in hippocampal neural plasticity (BDNF mRNA/protein, neurogenesis, LTP). However, very little is known about the effects of exercise on other, non-spatial forms of learning and memory. In addition, there has been little investigation of the duration of the effects of exercise on behavior or plasticity. Likewise, few studies have compared the effects of exercising during adulthood versus adolescence. This is particularly important since exercise may capitalize on the peak of neural plasticity observed during adolescence, resulting in a different pattern of behavioral and neurobiological effects. The present study addressed these gaps in the literature by comparing the effects of 4 weeks of voluntary exercise (wheel running) during adulthood or adolescence on novel object recognition and BDNF levels in the perirhinal cortex (PER) and hippocampus (HP). Exercising during adulthood improved object recognition memory when rats were tested immediately after 4 weeks of exercise, an effect that was accompanied by increased BDNF levels in PER and HP. When rats were tested again 2 weeks after exercise ended, the effects of exercise on recognition memory and BDNF levels were no longer present. Exercising during adolescence had a very different pattern of effects. First, both exercising and non-exercising rats could discriminate between novel and familiar objects immediately after the exercise regimen ended; furthermore there was no group difference in BDNF levels. Two or four weeks later, however, rats that had previously exercised as adolescents could still discriminate between novel and familiar objects, while non-exercising rats could not. Moreover, the formerly exercising rats exhibited higher levels of BDNF in PER compared to HP, while the reverse was true in the non-exercising rats. These findings reveal a novel

  19. The role of line junctions in object recognition: The case of reading musical notation.

    PubMed

    Wong, Yetta Kwailing; Wong, Alan C-N

    2018-04-30

    Previous work has shown that line junctions are informative features for visual perception of objects, letters, and words. However, the sources of such sensitivity and their generalizability to other object categories are largely unclear. We addressed these questions by studying perceptual expertise in reading musical notation, a domain in which individuals with different levels of expertise are readily available. We observed that removing line junctions created by the contact between musical notes and staff lines selectively impaired recognition performance in experts and intermediate readers, but not in novices. The degree of performance impairment was predicted by individual fluency in reading musical notation. Our findings suggest that line junctions provide diagnostic information about object identity across various categories, including musical notation. However, human sensitivity to line junctions does not readily transfer from familiar to unfamiliar object categories, and has to be acquired through perceptual experience with the specific objects.

  20. Toward open set recognition.

    PubMed

    Scheirer, Walter J; de Rezende Rocha, Anderson; Sapkota, Archana; Boult, Terrance E

    2013-07-01

    To date, almost all experimental evaluations of machine learning-based recognition algorithms in computer vision have taken the form of "closed set" recognition, whereby all testing classes are known at training time. A more realistic scenario for vision applications is "open set" recognition, where incomplete knowledge of the world is present at training time, and unknown classes can be submitted to an algorithm during testing. This paper explores the nature of open set recognition and formalizes its definition as a constrained minimization problem. The open set recognition problem is not well addressed by existing algorithms because it requires strong generalization. As a step toward a solution, we introduce a novel "1-vs-set machine," which sculpts a decision space from the marginal distances of a 1-class or binary SVM with a linear kernel. This methodology applies to several different applications in computer vision where open set recognition is a challenging problem, including object recognition and face verification. We consider both in this work, with large scale cross-dataset experiments performed over the Caltech 256 and ImageNet sets, as well as face matching experiments performed over the Labeled Faces in the Wild set. The experiments highlight the effectiveness of machines adapted for open set evaluation compared to existing 1-class and binary SVMs for the same tasks.

  1. Deficits in object-in-place but not relative recency performance in the APPswe/PS1dE9 mouse model of Alzheimer's disease: Implications for object recognition.

    PubMed

    Bonardi, Charlotte; Pardon, Marie-Christine; Armstrong, Paul

    2016-10-15

    Performance was examined on three variants of the spontaneous object recognition (SOR) task, in 5-month old APPswe/PS1dE9 mice and wild-type littermate controls. A deficit was observed in an object-in-place (OIP) task, in which mice are preexposed to four different objects in specific locations, and then at test two of the objects swap locations (Experiment 2). Typically more exploration is seen of the objects which have switched location, which is taken as evidence of a retrieval-generated priming mechanism. However, no significant transgenic deficit was found in a relative recency (RR) task (Experiment 1), in which mice are exposed to two different objects in two separate sample phases, and then tested with both objects. Typically more exploration of the first-presented object is observed, which is taken as evidence of a self-generated priming mechanism. Nor was there any impairment in the simplest variant, the spontaneous object recognition (SOR) task, in which mice are preexposed to one object and then tested with the familiar and a novel object. This was true regardless of whether the sample-test interval was 5min (Experiment 1) or 24h (Experiments 1 and 2). It is argued that SOR performance depends on retrieval-generated priming as well as self-generated priming, and our preliminary evidence suggests that the retrieval-generated priming process is especially impaired in these young transgenic animals. Copyright © 2016 Elsevier B.V. All rights reserved.

  2. Recognition of 3-D symmetric objects from range images in automated assembly tasks

    NASA Technical Reports Server (NTRS)

    Alvertos, Nicolas; Dcunha, Ivan

    1990-01-01

    A new technique is presented for the three dimensional recognition of symmetric objects from range images. Beginning from the implicit representation of quadrics, a set of ten coefficients is determined for symmetric objects like spheres, cones, cylinders, ellipsoids, and parallelepipeds. Instead of using these ten coefficients trying to fit them to smooth surfaces (patches) based on the traditional way of determining curvatures, a new approach based on two dimensional geometry is used. For each symmetric object, a unique set of two dimensional curves is obtained from the various angles at which the object is intersected with a plane. Using the same ten coefficients obtained earlier and based on the discriminant method, each of these curves is classified as a parabola, circle, ellipse, or hyperbola. Each symmetric object is found to possess a unique set of these two dimensional curves whereby it can be differentiated from the others. It is shown that instead of using the three dimensional discriminant which involves evaluation of the rank of its matrix, it is sufficient to use the two dimensional discriminant which only requires three arithmetic operations.

  3. Detection and recognition of targets by using signal polarization properties

    NASA Astrophysics Data System (ADS)

    Ponomaryov, Volodymyr I.; Peralta-Fabi, Ricardo; Popov, Anatoly V.; Babakov, Mikhail F.

    1999-08-01

    The quality of radar target recognition can be enhanced by exploiting its polarization signatures. A specialized X-band polarimetric radar was used for target recognition in experimental investigations. The following polarization characteristics connected to the object geometrical properties were investigated: the amplitudes of the polarization matrix elements; an anisotropy coefficient; depolarization coefficient; asymmetry coefficient; the energy of a backscattering signal; object shape factor. A large quantity of polarimetric radar data was measured and processed to form a database of different object and different weather conditions. The histograms of polarization signatures were approximated by a Nakagami distribution, then used for real- time target recognition. The Neyman-Pearson criterion was used for the target detection, and the criterion of the maximum of a posterior probability was used for recognition problem. Some results of experimental verification of pattern recognition and detection of objects with different electrophysical and geometrical characteristics urban in clutter are presented in this paper.

  4. Investigation of Time Series Representations and Similarity Measures for Structural Damage Pattern Recognition

    PubMed Central

    Swartz, R. Andrew

    2013-01-01

    This paper investigates the time series representation methods and similarity measures for sensor data feature extraction and structural damage pattern recognition. Both model-based time series representation and dimensionality reduction methods are studied to compare the effectiveness of feature extraction for damage pattern recognition. The evaluation of feature extraction methods is performed by examining the separation of feature vectors among different damage patterns and the pattern recognition success rate. In addition, the impact of similarity measures on the pattern recognition success rate and the metrics for damage localization are also investigated. The test data used in this study are from the System Identification to Monitor Civil Engineering Structures (SIMCES) Z24 Bridge damage detection tests, a rigorous instrumentation campaign that recorded the dynamic performance of a concrete box-girder bridge under progressively increasing damage scenarios. A number of progressive damage test case datasets and damage test data with different damage modalities are used. The simulation results show that both time series representation methods and similarity measures have significant impact on the pattern recognition success rate. PMID:24191136

  5. Using Constant Time Delay to Teach Braille Word Recognition

    ERIC Educational Resources Information Center

    Hooper, Jonathan; Ivy, Sarah; Hatton, Deborah

    2014-01-01

    Introduction: Constant time delay has been identified as an evidence-based practice to teach print sight words and picture recognition (Browder, Ahlbrim-Delzell, Spooner, Mims, & Baker, 2009). For the study presented here, we tested the effectiveness of constant time delay to teach new braille words. Methods: A single-subject multiple baseline…

  6. Neuritin reverses deficits in murine novel object associative recognition memory caused by exposure to extremely low-frequency (50 Hz) electromagnetic fields.

    PubMed

    Zhao, Qian-Ru; Lu, Jun-Mei; Yao, Jin-Jing; Zhang, Zheng-Yu; Ling, Chen; Mei, Yan-Ai

    2015-07-03

    Animal studies have shown that electromagnetic field exposure may interfere with the activity of brain cells, thereby generating behavioral and cognitive disturbances. However, the underlying mechanisms and possible preventions are still unknown. In this study, we used a mouse model to examine the effects of exposure to extremely low-frequency (50 Hz) electromagnetic fields (ELF MFs) on a recognition memory task and morphological changes of hippocampal neurons. The data showed that ELF MFs exposure (1 mT, 12 h/day) induced a time-dependent deficit in novel object associative recognition memory and also decreased hippocampal dendritic spine density. This effect was observed without corresponding changes in spontaneous locomotor activity and was transient, which has only been seen after exposing mice to ELF MFs for 7-10 days. The over-expression of hippocampal neuritin, an activity-dependent neurotrophic factor, using an adeno-associated virus (AAV) vector significantly increased the neuritin level and dendritic spine density. This increase was paralleled with ELF MFs exposure-induced deficits in recognition memory and reductions of dendritic spine density. Collectively, our study provides evidence for the association between ELF MFs exposure, impairment of recognition memory, and resulting changes in hippocampal dendritic spine density. Neuritin prevented this ELF MFs-exposure-induced effect by increasing the hippocampal spine density.

  7. Intra-pulse modulation recognition using short-time ramanujan Fourier transform spectrogram

    NASA Astrophysics Data System (ADS)

    Ma, Xiurong; Liu, Dan; Shan, Yunlong

    2017-12-01

    Intra-pulse modulation recognition under negative signal-to-noise ratio (SNR) environment is a research challenge. This article presents a robust algorithm for the recognition of 5 types of radar signals with large variation range in the signal parameters in low SNR using the combination of the Short-time Ramanujan Fourier transform (ST-RFT) and pseudo-Zernike moments invariant features. The ST-RFT provides the time-frequency distribution features for 5 modulations. The pseudo-Zernike moments provide invariance properties that are able to recognize different modulation schemes on different parameter variation conditions from the ST-RFT spectrograms. Simulation results demonstrate that the proposed algorithm achieves the probability of successful recognition (PSR) of over 90% when SNR is above -5 dB with large variation range in the signal parameters: carrier frequency (CF) for all considered signals, hop size (HS) for frequency shift keying (FSK) signals, and the time-bandwidth product for Linear Frequency Modulation (LFM) signals.

  8. Short exposure to a diet rich in both fat and sugar or sugar alone impairs place, but not object recognition memory in rats.

    PubMed

    Beilharz, Jessica E; Maniam, Jayanthi; Morris, Margaret J

    2014-03-01

    High energy diets have been shown to impair cognition however, the rapidity of these effects, and the dietary component/s responsible are currently unclear. We conducted two experiments in rats to examine the effects of short-term exposure to a diet rich in sugar and fat or rich in sugar on object (perirhinal-dependent) and place (hippocampal-dependent) recognition memory, and the role of inflammatory mediators in these responses. In Experiment 1, rats fed a cafeteria style diet containing chow supplemented with lard, cakes, biscuits, and a 10% sucrose solution performed worse on the place, but not the object recognition task, than chow fed control rats when tested after 5, 11, and 20 days. In Experiment 2, rats fed the cafeteria style diet either with or without sucrose and rats fed chow supplemented with sucrose also performed worse on the place, but not the object recognition task when tested after 5, 11, and 20 days. Rats fed the cafeteria diets consumed five times more energy than control rats and exhibited increased plasma leptin, insulin and triglyceride concentrations; these were not affected in the sucrose only rats. Rats exposed to sucrose exhibited both increased hippocampal inflammation (TNF-α and IL-1β mRNA) and oxidative stress, as indicated by an upregulation of NRF1 mRNA compared to control rats. In contrast, these markers were not significantly elevated in rats that received the cafeteria diet without added sucrose. Hippocampal BDNF and neuritin mRNA were similar across all groups. These results show that relatively short exposures to diets rich in both fat and sugar or rich in sugar, impair hippocampal-dependent place recognition memory prior to the emergence of weight differences, and suggest a role for oxidative stress and neuroinflammation in this impairment. Crown Copyright © 2013. Published by Elsevier Inc. All rights reserved.

  9. Recognition-induced forgetting is not due to category-based set size.

    PubMed

    Maxcey, Ashleigh M

    2016-01-01

    What are the consequences of accessing a visual long-term memory representation? Previous work has shown that accessing a long-term memory representation via retrieval improves memory for the targeted item and hurts memory for related items, a phenomenon called retrieval-induced forgetting. Recently we found a similar forgetting phenomenon with recognition of visual objects. Recognition-induced forgetting occurs when practice recognizing an object during a two-alternative forced-choice task, from a group of objects learned at the same time, leads to worse memory for objects from that group that were not practiced. An alternative explanation of this effect is that category-based set size is inducing forgetting, not recognition practice as claimed by some researchers. This alternative explanation is possible because during recognition practice subjects make old-new judgments in a two-alternative forced-choice task, and are thus exposed to more objects from practiced categories, potentially inducing forgetting due to set-size. Herein I pitted the category-based set size hypothesis against the recognition-induced forgetting hypothesis. To this end, I parametrically manipulated the amount of practice objects received in the recognition-induced forgetting paradigm. If forgetting is due to category-based set size, then the magnitude of forgetting of related objects will increase as the number of practice trials increases. If forgetting is recognition induced, the set size of exemplars from any given category should not be predictive of memory for practiced objects. Consistent with this latter hypothesis, additional practice systematically improved memory for practiced objects, but did not systematically affect forgetting of related objects. These results firmly establish that recognition practice induces forgetting of related memories. Future directions and important real-world applications of using recognition to access our visual memories of previously encountered

  10. A Dynamic Time Warping Approach to Real-Time Activity Recognition for Food Preparation

    NASA Astrophysics Data System (ADS)

    Pham, Cuong; Plötz, Thomas; Olivier, Patrick

    We present a dynamic time warping based activity recognition system for the analysis of low-level food preparation activities. Accelerometers embedded into kitchen utensils provide continuous sensor data streams while people are using them for cooking. The recognition framework analyzes frames of contiguous sensor readings in real-time with low latency. It thereby adapts to the idiosyncrasies of utensil use by automatically maintaining a template database. We demonstrate the effectiveness of the classification approach by a number of real-world practical experiments on a publically available dataset. The adaptive system shows superior performance compared to a static recognizer. Furthermore, we demonstrate the generalization capabilities of the system by gradually reducing the amount of training samples. The system achieves excellent classification results even if only a small number of training samples is available, which is especially relevant for real-world scenarios.

  11. Object recognition and localization from 3D point clouds by maximum-likelihood estimation

    NASA Astrophysics Data System (ADS)

    Dantanarayana, Harshana G.; Huntley, Jonathan M.

    2017-08-01

    We present an algorithm based on maximum-likelihood analysis for the automated recognition of objects, and estimation of their pose, from 3D point clouds. Surfaces segmented from depth images are used as the features, unlike `interest point'-based algorithms which normally discard such data. Compared to the 6D Hough transform, it has negligible memory requirements, and is computationally efficient compared to iterative closest point algorithms. The same method is applicable to both the initial recognition/pose estimation problem as well as subsequent pose refinement through appropriate choice of the dispersion of the probability density functions. This single unified approach therefore avoids the usual requirement for different algorithms for these two tasks. In addition to the theoretical description, a simple 2 degrees of freedom (d.f.) example is given, followed by a full 6 d.f. analysis of 3D point cloud data from a cluttered scene acquired by a projected fringe-based scanner, which demonstrated an RMS alignment error as low as 0.3 mm.

  12. Direction of Magnetoencephalography Sources Associated with Feedback and Feedforward Contributions in a Visual Object Recognition Task

    PubMed Central

    Ahlfors, Seppo P.; Jones, Stephanie R.; Ahveninen, Jyrki; Hämäläinen, Matti S.; Belliveau, John W.; Bar, Moshe

    2014-01-01

    Identifying inter-area communication in terms of the hierarchical organization of functional brain areas is of considerable interest in human neuroimaging. Previous studies have suggested that the direction of magneto- and electroencephalography (MEG, EEG) source currents depends on the layer-specific input patterns into a cortical area. We examined the direction in MEG source currents in a visual object recognition experiment in which there were specific expectations of activation in the fusiform region being driven by either feedforward or feedback inputs. The source for the early non-specific visual evoked response, presumably corresponding to feedforward driven activity, pointed outward, i.e., away from the white matter. In contrast, the source for the later, object-recognition related signals, expected to be driven by feedback inputs, pointed inward, toward the white matter. Associating specific features of the MEG/EEG source waveforms to feedforward and feedback inputs could provide unique information about the activation patterns within hierarchically organized cortical areas. PMID:25445356

  13. GPU Accelerated Symmetry Transform for Object Saliency

    DTIC Science & Technology

    significantly reduced execution time for VGA-sized images. The symmetry transform has potential for use in finding salient regions containing objects, and also may be applicable to stable object keypoints for recognition .

  14. Neuropeptide S enhances memory and mitigates memory impairment induced by MK801, scopolamine or Aβ₁₋₄₂ in mice novel object and object location recognition tasks.

    PubMed

    Han, Ren-Wen; Zhang, Rui-San; Xu, Hong-Jiao; Chang, Min; Peng, Ya-Li; Wang, Rui

    2013-07-01

    Neuropeptide S (NPS), the endogenous ligand of NPSR, has been shown to promote arousal and anxiolytic-like effects. According to the predominant distribution of NPSR in brain tissues associated with learning and memory, NPS has been reported to modulate cognitive function in rodents. Here, we investigated the role of NPS in memory formation, and determined whether NPS could mitigate memory impairment induced by selective N-methyl-D-aspartate receptor antagonist MK801, muscarinic cholinergic receptor antagonist scopolamine or Aβ₁₋₄₂ in mice, using novel object and object location recognition tasks. Intracerebroventricular (i.c.v.) injection of 1 nmol NPS 5 min after training not only facilitated object recognition memory formation, but also prolonged memory retention in both tasks. The improvement of object recognition memory induced by NPS could be blocked by the selective NPSR antagonist SHA 68, indicating pharmacological specificity. Then, we found that i.c.v. injection of NPS reversed memory disruption induced by MK801, scopolamine or Aβ₁₋₄₂ in both tasks. In summary, our results indicate that NPS facilitates memory formation and prolongs the retention of memory through activation of the NPSR, and mitigates amnesia induced by blockage of glutamatergic or cholinergic system or by Aβ₁₋₄₂, suggesting that NPS/NPSR system may be a new target for enhancing memory and treating amnesia. Copyright © 2013 Elsevier Ltd. All rights reserved.

  15. Effects of ethanol, Δ(9)-tetrahydrocannabinol, or their combination on object recognition memory and object preference in adolescent and adult male rats.

    PubMed

    Swartzwelder, Nicholas A; Risher, M Louise; Abdelwahab, Sabri H; D'Abo, Anouska; Rezvani, Amir H; Levin, Edward D; Wilson, Wilkie A; Swartzwelder, H Scott; Acheson, Shawn K

    2012-10-03

    Recent advances have been made in our understanding of the deleterious effects of both ethanol and THC on adolescent behavior and brain development. However, very little is known about the combined effects of EtOH+THC during adolescence, a time in which these drugs are often used together. The purpose of this experiment was to: (1) determine whether EtOH and/or THC induced greater working memory impairment in adolescent than adult male rats using the novel object recognition (NOR) task and (2) determine whether the EtOH+THC combination would produce a more potent additive effect in adolescents than adults when compared to these drugs alone. NOR was performed with a 24h delay under each of the four drug conditions: vehicle; 1.5g/kg ethanol; 1.0mg/kg THC; and 1.5g/kg EtOH+1.0mg/kg THC, at 72h intervals. The results show that there was an age effect on working memory in NOR after the EtOH+THC challenge. Specifically, adolescent animals showed a preference for the familiar object whereas adults showed no preference for the novel or familiar object, the latter being characteristic of a classic working memory deficit. These effects were not dependent on changes in exploration across session, global activity across drug condition, or total object exploration. These novel findings clearly indicate that further understanding of this age-drug interaction is crucial to elucidating the influence that adolescent EtOH+THC use may have on repeated drug use and abuse later in life. Published by Elsevier Ireland Ltd.

  16. Neural Dynamics of Object-Based Multifocal Visual Spatial Attention and Priming: Object Cueing, Useful-Field-of-View, and Crowding

    ERIC Educational Resources Information Center

    Foley, Nicholas C.; Grossberg, Stephen; Mingolla, Ennio

    2012-01-01

    How are spatial and object attention coordinated to achieve rapid object learning and recognition during eye movement search? How do prefrontal priming and parietal spatial mechanisms interact to determine the reaction time costs of intra-object attention shifts, inter-object attention shifts, and shifts between visible objects and covertly cued…

  17. What can neuromorphic event-driven precise timing add to spike-based pattern recognition?

    PubMed

    Akolkar, Himanshu; Meyer, Cedric; Clady, Zavier; Marre, Olivier; Bartolozzi, Chiara; Panzeri, Stefano; Benosman, Ryad

    2015-03-01

    This letter introduces a study to precisely measure what an increase in spike timing precision can add to spike-driven pattern recognition algorithms. The concept of generating spikes from images by converting gray levels into spike timings is currently at the basis of almost every spike-based modeling of biological visual systems. The use of images naturally leads to generating incorrect artificial and redundant spike timings and, more important, also contradicts biological findings indicating that visual processing is massively parallel, asynchronous with high temporal resolution. A new concept for acquiring visual information through pixel-individual asynchronous level-crossing sampling has been proposed in a recent generation of asynchronous neuromorphic visual sensors. Unlike conventional cameras, these sensors acquire data not at fixed points in time for the entire array but at fixed amplitude changes of their input, resulting optimally sparse in space and time-pixel individually and precisely timed only if new, (previously unknown) information is available (event based). This letter uses the high temporal resolution spiking output of neuromorphic event-based visual sensors to show that lowering time precision degrades performance on several recognition tasks specifically when reaching the conventional range of machine vision acquisition frequencies (30-60 Hz). The use of information theory to characterize separability between classes for each temporal resolution shows that high temporal acquisition provides up to 70% more information that conventional spikes generated from frame-based acquisition as used in standard artificial vision, thus drastically increasing the separability between classes of objects. Experiments on real data show that the amount of information loss is correlated with temporal precision. Our information-theoretic study highlights the potentials of neuromorphic asynchronous visual sensors for both practical applications and theoretical

  18. Visual object recognition and tracking

    NASA Technical Reports Server (NTRS)

    Chang, Chu-Yin (Inventor); English, James D. (Inventor); Tardella, Neil M. (Inventor)

    2010-01-01

    This invention describes a method for identifying and tracking an object from two-dimensional data pictorially representing said object by an object-tracking system through processing said two-dimensional data using at least one tracker-identifier belonging to the object-tracking system for providing an output signal containing: a) a type of the object, and/or b) a position or an orientation of the object in three-dimensions, and/or c) an articulation or a shape change of said object in said three dimensions.

  19. Timing of presentation and nature of stimuli determine retroactive interference with social recognition memory in mice.

    PubMed

    Perna, Judith Camats; Wotjak, Carsten T; Stork, Oliver; Engelmann, Mario

    2015-05-01

    The present study was designed to further investigate the nature of stimuli and the timing of their presentation, which can induce retroactive interference with social recognition memory in mice. In accordance with our previous observations, confrontation with an unfamiliar conspecific juvenile 3h and 6h, but not 22 h, after the initial learning session resulted in retroactive interference. The same effect was observed with the exposure to both enantiomers of the monomolecular odour carvone, and with a novel object. Exposure to a loud tone (12 KHz, 90 dB) caused retroactive interference at 6h, but not 3h and 22 h, after sampling. Our data show that retroactive interference of social recognition memory can be induced by exposing the experimental subjects to the defined stimuli presented <22 h after learning in their home cage. The distinct interference triggered by the tone presentation at 6h after sampling may be linked to the intrinsic aversiveness of the loud tone and suggests that at this time point memory consolidation is particularly sensitive to stress. Copyright © 2015 Elsevier Inc. All rights reserved.

  20. Representation of 3-Dimenstional Objects by the Rat Perirhinal Cortex

    PubMed Central

    Burke, S.N.; Maurer, A.P.; Hartzell, A.L.; Nematollahi, S.; Uprety, A.; Wallace, J.L.; Barnes, C.A.

    2012-01-01

    The perirhinal cortex (PRC) is known to play an important role in object recognition. Little is known, however, regarding the activity of PRC neurons during the presentation of stimuli that are commonly used for recognition memory tasks in rodents, that is, 3-dimensional objects. Rats in the present study were exposed to 3-dimensional objects while they traversed a circular track for food reward. Under some behavioral conditions the track contained novel objects, familiar objects, or no objects. Approximately 38% of PRC neurons demonstrated ‘object fields’ (a selective increase in firing at the location of one or more objects). Although the rats spent more time exploring the objects when they were novel compared to familiar, indicating successful recognition memory, the proportion of object fields and the firing rates of PRC neurons were not affected by the rats’ previous experience with the objects. Together these data indicate that the activity of PRC cells is powerfully affected by the presence of objects while animals navigate through an environment, but under these conditions, the firing patterns are not altered by the relative novelty of objects during successful object recognition. PMID:22987680

  1. Object-Place Recognition Learning Triggers Rapid Induction of Plasticity-Related Immediate Early Genes and Synaptic Proteins in the Rat Dentate Gyrus

    PubMed Central

    Soulé, Jonathan; Penke, Zsuzsa; Kanhema, Tambudzai; Alme, Maria Nordheim; Laroche, Serge; Bramham, Clive R.

    2008-01-01

    Long-term recognition memory requires protein synthesis, but little is known about the coordinate regulation of specific genes. Here, we examined expression of the plasticity-associated immediate early genes (Arc, Zif268, and Narp) in the dentate gyrus following long-term object-place recognition learning in rats. RT-PCR analysis from dentate gyrus tissue collected shortly after training did not reveal learning-specific changes in Arc mRNA expression. In situ hybridization and immunohistochemistry were therefore used to assess possible sparse effects on gene expression. Learning about objects increased the density of granule cells expressing Arc, and to a lesser extent Narp, specifically in the dorsal blade of the dentate gyrus, while Zif268 expression was elevated across both blades. Thus, object-place recognition triggers rapid, blade-specific upregulation of plasticity-associated immediate early genes. Furthermore, Western blot analysis of dentate gyrus homogenates demonstrated concomitant upregulation of three postsynaptic density proteins (Arc, PSD-95, and α-CaMKII) with key roles in long-term synaptic plasticity and long-term memory. PMID:19190776

  2. Real-time traffic sign recognition based on a general purpose GPU and deep-learning.

    PubMed

    Lim, Kwangyong; Hong, Yongwon; Choi, Yeongwoo; Byun, Hyeran

    2017-01-01

    We present a General Purpose Graphics Processing Unit (GPGPU) based real-time traffic sign detection and recognition method that is robust against illumination changes. There have been many approaches to traffic sign recognition in various research fields; however, previous approaches faced several limitations when under low illumination or wide variance of light conditions. To overcome these drawbacks and improve processing speeds, we propose a method that 1) is robust against illumination changes, 2) uses GPGPU-based real-time traffic sign detection, and 3) performs region detecting and recognition using a hierarchical model. This method produces stable results in low illumination environments. Both detection and hierarchical recognition are performed in real-time, and the proposed method achieves 0.97 F1-score on our collective dataset, which uses the Vienna convention traffic rules (Germany and South Korea).

  3. Real-time traffic sign recognition based on a general purpose GPU and deep-learning

    PubMed Central

    Hong, Yongwon; Choi, Yeongwoo; Byun, Hyeran

    2017-01-01

    We present a General Purpose Graphics Processing Unit (GPGPU) based real-time traffic sign detection and recognition method that is robust against illumination changes. There have been many approaches to traffic sign recognition in various research fields; however, previous approaches faced several limitations when under low illumination or wide variance of light conditions. To overcome these drawbacks and improve processing speeds, we propose a method that 1) is robust against illumination changes, 2) uses GPGPU-based real-time traffic sign detection, and 3) performs region detecting and recognition using a hierarchical model. This method produces stable results in low illumination environments. Both detection and hierarchical recognition are performed in real-time, and the proposed method achieves 0.97 F1-score on our collective dataset, which uses the Vienna convention traffic rules (Germany and South Korea). PMID:28264011

  4. Excess influx of Zn(2+) into dentate granule cells affects object recognition memory via attenuated LTP.

    PubMed

    Suzuki, Miki; Fujise, Yuki; Tsuchiya, Yuka; Tamano, Haruna; Takeda, Atsushi

    2015-08-01

    The influx of extracellular Zn(2+) into dentate granule cells is nonessential for dentate gyrus long-term potentiation (LTP) and the physiological significance of extracellular Zn(2+) dynamics is unknown in the dentate gyrus. Excess increase in extracellular Zn(2+) in the hippocampal CA1, which is induced with excitation of zincergic neurons, induces memory deficit via excess influx of Zn(2+) into CA1 pyramidal cells. In the present study, it was examined whether extracellular Zn(2+) induces object recognition memory deficit via excess influx of Zn(2+) into dentate granule cells. KCl (100 mM, 2 µl) was locally injected into the dentate gyrus. The increase in intracellular Zn(2+) in dentate granule cells induced with high K(+) was blocked by co-injection of CaEDTA and CNQX, an extracellular Zn(2+) chelator and an AMPA receptor antagonist, respectively, suggesting that high K(+) increases the influx of Zn(2+) into dentate granule cells via AMPA receptor activation. Dentate gyrus LTP induction was attenuated 1 h after KCl injection into the dentate gyrus and also attenuated when KCl was injected 5 min after the induction. Memory deficit was induced when training of object recognition test was performed 1 h after KCl injection into the dentate gyrus and also induced when KCl was injected 5 min after the training. High K(+)-induced impairments of LTP and memory were rescued by co-injection of CaEDTA. These results indicate that excess influx of Zn(2+) into dentate granule cells via AMPA receptor activation affects object recognition memory via attenuated LTP induction. Even in the dentate gyrus where is scarcely innervated by zincergic neurons, it is likely that extracellular Zn(2+) homeostasis is strictly regulated for cognition. Copyright © 2015 Elsevier Ltd. All rights reserved.

  5. STDP-based spiking deep convolutional neural networks for object recognition.

    PubMed

    Kheradpisheh, Saeed Reza; Ganjtabesh, Mohammad; Thorpe, Simon J; Masquelier, Timothée

    2018-03-01

    Previous studies have shown that spike-timing-dependent plasticity (STDP) can be used in spiking neural networks (SNN) to extract visual features of low or intermediate complexity in an unsupervised manner. These studies, however, used relatively shallow architectures, and only one layer was trainable. Another line of research has demonstrated - using rate-based neural networks trained with back-propagation - that having many layers increases the recognition robustness, an approach known as deep learning. We thus designed a deep SNN, comprising several convolutional (trainable with STDP) and pooling layers. We used a temporal coding scheme where the most strongly activated neurons fire first, and less activated neurons fire later or not at all. The network was exposed to natural images. Thanks to STDP, neurons progressively learned features corresponding to prototypical patterns that were both salient and frequent. Only a few tens of examples per category were required and no label was needed. After learning, the complexity of the extracted features increased along the hierarchy, from edge detectors in the first layer to object prototypes in the last layer. Coding was very sparse, with only a few thousands spikes per image, and in some cases the object category could be reasonably well inferred from the activity of a single higher-order neuron. More generally, the activity of a few hundreds of such neurons contained robust category information, as demonstrated using a classifier on Caltech 101, ETH-80, and MNIST databases. We also demonstrate the superiority of STDP over other unsupervised techniques such as random crops (HMAX) or auto-encoders. Taken together, our results suggest that the combination of STDP with latency coding may be a key to understanding the way that the primate visual system learns, its remarkable processing speed and its low energy consumption. These mechanisms are also interesting for artificial vision systems, particularly for hardware

  6. Emerging technologies with potential for objectively evaluating speech recognition skills.

    PubMed

    Rawool, Vishakha Waman

    2016-01-01

    Work-related exposure to noise and other ototoxins can cause damage to the cochlea, synapses between the inner hair cells, the auditory nerve fibers, and higher auditory pathways, leading to difficulties in recognizing speech. Procedures designed to determine speech recognition scores (SRS) in an objective manner can be helpful in disability compensation cases where the worker claims to have poor speech perception due to exposure to noise or ototoxins. Such measures can also be helpful in determining SRS in individuals who cannot provide reliable responses to speech stimuli, including patients with Alzheimer's disease, traumatic brain injuries, and infants with and without hearing loss. Cost-effective neural monitoring hardware and software is being rapidly refined due to the high demand for neurogaming (games involving the use of brain-computer interfaces), health, and other applications. More specifically, two related advances in neuro-technology include relative ease in recording neural activity and availability of sophisticated analysing techniques. These techniques are reviewed in the current article and their applications for developing objective SRS procedures are proposed. Issues related to neuroaudioethics (ethics related to collection of neural data evoked by auditory stimuli including speech) and neurosecurity (preservation of a person's neural mechanisms and free will) are also discussed.

  7. Combining high-speed SVM learning with CNN feature encoding for real-time target recognition in high-definition video for ISR missions

    NASA Astrophysics Data System (ADS)

    Kroll, Christine; von der Werth, Monika; Leuck, Holger; Stahl, Christoph; Schertler, Klaus

    2017-05-01

    For Intelligence, Surveillance, Reconnaissance (ISR) missions of manned and unmanned air systems typical electrooptical payloads provide high-definition video data which has to be exploited with respect to relevant ground targets in real-time by automatic/assisted target recognition software. Airbus Defence and Space is developing required technologies for real-time sensor exploitation since years and has combined the latest advances of Deep Convolutional Neural Networks (CNN) with a proprietary high-speed Support Vector Machine (SVM) learning method into a powerful object recognition system with impressive results on relevant high-definition video scenes compared to conventional target recognition approaches. This paper describes the principal requirements for real-time target recognition in high-definition video for ISR missions and the Airbus approach of combining an invariant feature extraction using pre-trained CNNs and the high-speed training and classification ability of a novel frequency-domain SVM training method. The frequency-domain approach allows for a highly optimized implementation for General Purpose Computation on a Graphics Processing Unit (GPGPU) and also an efficient training of large training samples. The selected CNN which is pre-trained only once on domain-extrinsic data reveals a highly invariant feature extraction. This allows for a significantly reduced adaptation and training of the target recognition method for new target classes and mission scenarios. A comprehensive training and test dataset was defined and prepared using relevant high-definition airborne video sequences. The assessment concept is explained and performance results are given using the established precision-recall diagrams, average precision and runtime figures on representative test data. A comparison to legacy target recognition approaches shows the impressive performance increase by the proposed CNN+SVM machine-learning approach and the capability of real-time high

  8. Genetic Mapping in Mice Reveals the Involvement of Pcdh9 in Long-Term Social and Object Recognition and Sensorimotor Development.

    PubMed

    Bruining, Hilgo; Matsui, Asuka; Oguro-Ando, Asami; Kahn, René S; Van't Spijker, Heleen M; Akkermans, Guus; Stiedl, Oliver; van Engeland, Herman; Koopmans, Bastijn; van Lith, Hein A; Oppelaar, Hugo; Tieland, Liselotte; Nonkes, Lourens J; Yagi, Takeshi; Kaneko, Ryosuke; Burbach, J Peter H; Yamamoto, Nobuhiko; Kas, Martien J

    2015-10-01

    Quantitative genetic analysis of basic mouse behaviors is a powerful tool to identify novel genetic phenotypes contributing to neurobehavioral disorders. Here, we analyzed genetic contributions to single-trial, long-term social and nonsocial recognition and subsequently studied the functional impact of an identified candidate gene on behavioral development. Genetic mapping of single-trial social recognition was performed in chromosome substitution strains, a sophisticated tool for detecting quantitative trait loci (QTL) of complex traits. Follow-up occurred by generating and testing knockout (KO) mice of a selected QTL candidate gene. Functional characterization of these mice was performed through behavioral and neurological assessments across developmental stages and analyses of gene expression and brain morphology. Chromosome substitution strain 14 mapping studies revealed an overlapping QTL related to long-term social and object recognition harboring Pcdh9, a cell-adhesion gene previously associated with autism spectrum disorder. Specific long-term social and object recognition deficits were confirmed in homozygous (KO) Pcdh9-deficient mice, while heterozygous mice only showed long-term social recognition impairment. The recognition deficits in KO mice were not associated with alterations in perception, multi-trial discrimination learning, sociability, behavioral flexibility, or fear memory. Rather, KO mice showed additional impairments in sensorimotor development reflected by early touch-evoked biting, rotarod performance, and sensory gating deficits. This profile emerged with structural changes in deep layers of sensory cortices, where Pcdh9 is selectively expressed. This behavior-to-gene study implicates Pcdh9 in cognitive functions required for long-term social and nonsocial recognition. This role is supported by the involvement of Pcdh9 in sensory cortex development and sensorimotor phenotypes. Copyright © 2015 Society of Biological Psychiatry. Published

  9. Real-Time Reconfigurable Adaptive Speech Recognition Command and Control Apparatus and Method

    NASA Technical Reports Server (NTRS)

    Salazar, George A. (Inventor); Haynes, Dena S. (Inventor); Sommers, Marc J. (Inventor)

    1998-01-01

    An adaptive speech recognition and control system and method for controlling various mechanisms and systems in response to spoken instructions and in which spoken commands are effective to direct the system into appropriate memory nodes, and to respective appropriate memory templates corresponding to the voiced command is discussed. Spoken commands from any of a group of operators for which the system is trained may be identified, and voice templates are updated as required in response to changes in pronunciation and voice characteristics over time of any of the operators for which the system is trained. Provisions are made for both near-real-time retraining of the system with respect to individual terms which are determined not be positively identified, and for an overall system training and updating process in which recognition of each command and vocabulary term is checked, and in which the memory templates are retrained if necessary for respective commands or vocabulary terms with respect to an operator currently using the system. In one embodiment, the system includes input circuitry connected to a microphone and including signal processing and control sections for sensing the level of vocabulary recognition over a given period and, if recognition performance falls below a given level, processing audio-derived signals for enhancing recognition performance of the system.

  10. Real time biometric surveillance with gait recognition

    NASA Astrophysics Data System (ADS)

    Mohapatra, Subasish; Swain, Anisha; Das, Manaswini; Mohanty, Subhadarshini

    2018-04-01

    Bio metric surveillance has become indispensable for every system in the recent years. The contribution of bio metric authentication, identification, and screening purposes are widely used in various domains for preventing unauthorized access. A large amount of data needs to be updated, segregated and safeguarded from malicious software and misuse. Bio metrics is the intrinsic characteristics of each individual. Recently fingerprints, iris, passwords, unique keys, and cards are commonly used for authentication purposes. These methods have various issues related to security and confidentiality. These systems are not yet automated to provide the safety and security. The gait recognition system is the alternative for overcoming the drawbacks of the recent bio metric based authentication systems. Gait recognition is newer as it hasn't been implemented in the real-world scenario so far. This is an un-intrusive system that requires no knowledge or co-operation of the subject. Gait is a unique behavioral characteristic of every human being which is hard to imitate. The walking style of an individual teamed with the orientation of joints in the skeletal structure and inclinations between them imparts the unique characteristic. A person can alter one's own external appearance but not skeletal structure. These are real-time, automatic systems that can even process low-resolution images and video frames. In this paper, we have proposed a gait recognition system and compared the performance with conventional bio metric identification systems.

  11. Crowded and sparse domains in object recognition: consequences for categorization and naming.

    PubMed

    Gale, Tim M; Laws, Keith R; Foley, Kerry

    2006-03-01

    Some models of object recognition propose that items from structurally crowded categories (e.g., living things) permit faster access to superordinate semantic information than structurally dissimilar categories (e.g., nonliving things), but slower access to individual object information when naming items. We present four experiments that utilize the same matched stimuli: two examine superordinate categorization and two examine picture naming. Experiments 1 and 2 required participants to sort pictures into their appropriate superordinate categories and both revealed faster categorization for living than nonliving things. Nonetheless, the living thing superiority disappeared when the atypical categories of body parts and musical instruments were excluded. Experiment 3 examined naming latency and found no difference between living and nonliving things. This finding was replicated in Experiment 4 where the same items were presented in different formats (e.g., color and line-drawn versions). Taken as a whole, these experiments show that the ease with which people categorize items maps strongly onto the ease with which they name them.

  12. Reader error, object recognition, and visual search

    NASA Astrophysics Data System (ADS)

    Kundel, Harold L.

    2004-05-01

    Small abnormalities such as hairline fractures, lung nodules and breast tumors are missed by competent radiologists with sufficient frequency to make them a matter of concern to the medical community; not only because they lead to litigation but also because they delay patient care. It is very easy to attribute misses to incompetence or inattention. To do so may be placing an unjustified stigma on the radiologists involved and may allow other radiologists to continue a false optimism that it can never happen to them. This review presents some of the fundamentals of visual system function that are relevant to understanding the search for and the recognition of small targets embedded in complicated but meaningful backgrounds like chests and mammograms. It presents a model for visual search that postulates a pre-attentive global analysis of the retinal image followed by foveal checking fixations and eventually discovery scanning. The model will be used to differentiate errors of search, recognition and decision making. The implications for computer aided diagnosis and for functional workstation design are discussed.

  13. Conversion of short-term to long-term memory in the novel object recognition paradigm.

    PubMed

    Moore, Shannon J; Deshpande, Kaivalya; Stinnett, Gwen S; Seasholtz, Audrey F; Murphy, Geoffrey G

    2013-10-01

    It is well-known that stress can significantly impact learning; however, whether this effect facilitates or impairs the resultant memory depends on the characteristics of the stressor. Investigation of these dynamics can be confounded by the role of the stressor in motivating performance in a task. Positing a cohesive model of the effect of stress on learning and memory necessitates elucidating the consequences of stressful stimuli independently from task-specific functions. Therefore, the goal of this study was to examine the effect of manipulating a task-independent stressor (elevated light level) on short-term and long-term memory in the novel object recognition paradigm. Short-term memory was elicited in both low light and high light conditions, but long-term memory specifically required high light conditions during the acquisition phase (familiarization trial) and was independent of the light level during retrieval (test trial). Additionally, long-term memory appeared to be independent of stress-mediated glucocorticoid release, as both low and high light produced similar levels of plasma corticosterone, which further did not correlate with subsequent memory performance. Finally, both short-term and long-term memory showed no savings between repeated experiments suggesting that this novel object recognition paradigm may be useful for longitudinal studies, particularly when investigating treatments to stabilize or enhance weak memories in neurodegenerative diseases or during age-related cognitive decline. Copyright © 2013 Elsevier Inc. All rights reserved.

  14. Conversion of short-term to long-term memory in the novel object recognition paradigm

    PubMed Central

    Moore, Shannon J.; Deshpande, Kaivalya; Stinnett, Gwen S.; Seasholtz, Audrey F.; Murphy, Geoffrey G.

    2013-01-01

    It is well-known that stress can significantly impact learning; however, whether this effect facilitates or impairs the resultant memory depends on the characteristics of the stressor. Investigation of these dynamics can be confounded by the role of the stressor in motivating performance in a task. Positing a cohesive model of the effect of stress on learning and memory necessitates elucidating the consequences of stressful stimuli independently from task-specific functions. Therefore, the goal of this study was to examine the effect of manipulating a task-independent stressor (elevated light level) on short-term and long-term memory in the novel object recognition paradigm. Short-term memory was elicited in both low light and high light conditions, but long-term memory specifically required high light conditions during the acquisition phase (familiarization trial) and was independent of the light level during retrieval (test trial). Additionally, long-term memory appeared to be independent of stress-mediated glucocorticoid release, as both low and high light produced similar levels of plasma corticosterone, which further did not correlate with subsequent memory performance. Finally, both short-term and long-term memory showed no savings between repeated experiments suggesting that this novel object recognition paradigm may be useful for longitudinal studies, particularly when investigating treatments to stabilize or enhance weak memories in neurodegenerative diseases or during age-related cognitive decline. PMID:23835143

  15. Recognition of partially occluded threat objects using the annealed Hopefield network

    NASA Technical Reports Server (NTRS)

    Kim, Jung H.; Yoon, Sung H.; Park, Eui H.; Ntuen, Celestine A.

    1992-01-01

    Recognition of partially occluded objects has been an important issue to airport security because occlusion causes significant problems in identifying and locating objects during baggage inspection. The neural network approach is suitable for the problems in the sense that the inherent parallelism of neural networks pursues many hypotheses in parallel resulting in high computation rates. Moreover, they provide a greater degree of robustness or fault tolerance than conventional computers. The annealed Hopfield network which is derived from the mean field annealing (MFA) has been developed to find global solutions of a nonlinear system. In the study, it has been proven that the system temperature of MFA is equivalent to the gain of the sigmoid function of a Hopfield network. In our early work, we developed the hybrid Hopfield network (HHN) for fast and reliable matching. However, HHN doesn't guarantee global solutions and yields false matching under heavily occluded conditions because HHN is dependent on initial states by its nature. In this paper, we present the annealed Hopfield network (AHN) for occluded object matching problems. In AHN, the mean field theory is applied to the hybird Hopfield network in order to improve computational complexity of the annealed Hopfield network and provide reliable matching under heavily occluded conditions. AHN is slower than HHN. However, AHN provides near global solutions without initial restrictions and provides less false matching than HHN. In conclusion, a new algorithm based upon a neural network approach was developed to demonstrate the feasibility of the automated inspection of threat objects from x-ray images. The robustness of the algorithm is proved by identifying occluded target objects with large tolerance of their features.

  16. Spoken Word Recognition in Toddlers Who Use Cochlear Implants

    PubMed Central

    Grieco-Calub, Tina M.; Saffran, Jenny R.; Litovsky, Ruth Y.

    2010-01-01

    Purpose The purpose of this study was to assess the time course of spoken word recognition in 2-year-old children who use cochlear implants (CIs) in quiet and in the presence of speech competitors. Method Children who use CIs and age-matched peers with normal acoustic hearing listened to familiar auditory labels, in quiet or in the presence of speech competitors, while their eye movements to target objects were digitally recorded. Word recognition performance was quantified by measuring each child’s reaction time (i.e., the latency between the spoken auditory label and the first look at the target object) and accuracy (i.e., the amount of time that children looked at target objects within 367 ms to 2,000 ms after the label onset). Results Children with CIs were less accurate and took longer to fixate target objects than did age-matched children without hearing loss. Both groups of children showed reduced performance in the presence of the speech competitors, although many children continued to recognize labels at above-chance levels. Conclusion The results suggest that the unique auditory experience of young CI users slows the time course of spoken word recognition abilities. In addition, real-world listening environments may slow language processing in young language learners, regardless of their hearing status. PMID:19951921

  17. Recognition memory in tree shrew (Tupaia belangeri) after repeated familiarization sessions.

    PubMed

    Khani, Abbas; Rainer, Gregor

    2012-07-01

    Recognition memories are formed during perceptual experience and allow subsequent recognition of previously encountered objects as well as their distinction from novel objects. As a consequence, novel objects are generally explored longer than familiar objects by many species. This novelty preference has been documented in rodents using the novel object recognition (NOR) test, as well is in primates including humans using preferential looking time paradigms. Here, we examine novelty preference using the NOR task in tree shrew, a small animal species that is considered to be an intermediary between rodents and primates. Our paradigm consisted of three phases: arena familiarization, object familiarization sessions with two identical objects in the arena and finally a test session following a 24-h retention period with a familiar and a novel object in the arena. We employed two different object familiarization durations: one and three sessions on consecutive days. After three object familiarization sessions, tree shrews exhibited robust preference for novel objects on the test day. This was accompanied by significant reduction in familiar object exploration time, occurring largely between the first and second day of object familiarization. By contrast, tree shrews did not show a significant preference for the novel object after a one-session object familiarization. Nonetheless, they spent significantly less time exploring the familiar object on the test day compared to the object familiarization day, indicating that they did maintain a memory trace for the familiar object. Our study revealed different time courses for familiar object habituation and emergence of novelty preference, suggesting that novelty preference is dependent on well-consolidated memory of the competing familiar object. Taken together, our results demonstrate robust novelty preference of tree shrews, in general similarity to previous findings in rodents and primates. Copyright © 2012 Elsevier B

  18. Fast and flexible 3D object recognition solutions for machine vision applications

    NASA Astrophysics Data System (ADS)

    Effenberger, Ira; Kühnle, Jens; Verl, Alexander

    2013-03-01

    In automation and handling engineering, supplying work pieces between different stages along the production process chain is of special interest. Often the parts are stored unordered in bins or lattice boxes and hence have to be separated and ordered for feeding purposes. An alternative to complex and spacious mechanical systems such as bowl feeders or conveyor belts, which are typically adapted to the parts' geometry, is using a robot to grip the work pieces out of a bin or from a belt. Such applications are in need of reliable and precise computer-aided object detection and localization systems. For a restricted range of parts, there exists a variety of 2D image processing algorithms that solve the recognition problem. However, these methods are often not well suited for the localization of randomly stored parts. In this paper we present a fast and flexible 3D object recognizer that localizes objects by identifying primitive features within the objects. Since technical work pieces typically consist to a substantial degree of geometric primitives such as planes, cylinders and cones, such features usually carry enough information in order to determine the position of the entire object. Our algorithms use 3D best-fitting combined with an intelligent data pre-processing step. The capability and performance of this approach is shown by applying the algorithms to real data sets of different industrial test parts in a prototypical bin picking demonstration system.

  19. A fast 3-D object recognition algorithm for the vision system of a special-purpose dexterous manipulator

    NASA Technical Reports Server (NTRS)

    Hung, Stephen H. Y.

    1989-01-01

    A fast 3-D object recognition algorithm that can be used as a quick-look subsystem to the vision system for the Special-Purpose Dexterous Manipulator (SPDM) is described. Global features that can be easily computed from range data are used to characterize the images of a viewer-centered model of an object. This algorithm will speed up the processing by eliminating the low level processing whenever possible. It may identify the object, reject a set of bad data in the early stage, or create a better environment for a more powerful algorithm to carry the work further.

  20. Under what conditions is recognition spared relative to recall after selective hippocampal damage in humans?

    PubMed

    Holdstock, J S; Mayes, A R; Roberts, N; Cezayirli, E; Isaac, C L; O'Reilly, R C; Norman, K A

    2002-01-01

    The claim that recognition memory is spared relative to recall after focal hippocampal damage has been disputed in the literature. We examined this claim by investigating object and object-location recall and recognition memory in a patient, YR, who has adult-onset selective hippocampal damage. Our aim was to identify the conditions under which recognition was spared relative to recall in this patient. She showed unimpaired forced-choice object recognition but clearly impaired recall, even when her control subjects found the object recognition task to be numerically harder than the object recall task. However, on two other recognition tests, YR's performance was not relatively spared. First, she was clearly impaired at an equivalently difficult yes/no object recognition task, but only when targets and foils were very similar. Second, YR was clearly impaired at forced-choice recognition of object-location associations. This impairment was also unrelated to difficulty because this task was no more difficult than the forced-choice object recognition task for control subjects. The clear impairment of yes/no, but not of forced-choice, object recognition after focal hippocampal damage, when targets and foils are very similar, is predicted by the neural network-based Complementary Learning Systems model of recognition. This model postulates that recognition is mediated by hippocampally dependent recollection and cortically dependent familiarity; thus hippocampal damage should not impair item familiarity. The model postulates that familiarity is ineffective when very similar targets and foils are shown one at a time and subjects have to identify which items are old (yes/no recognition). In contrast, familiarity is effective in discriminating which of similar targets and foils, seen together, is old (forced-choice recognition). Independent evidence from the remember/know procedure also indicates that YR's familiarity is normal. The Complementary Learning Systems model can

  1. Qualitative Differences in the Representation of Spatial Relations for Different Object Classes

    ERIC Educational Resources Information Center

    Cooper, Eric E.; Brooks, Brian E.

    2004-01-01

    Two experiments investigated whether the representations used for animal, produce, and object recognition code spatial relations in a similar manner. Experiment 1 tested the effects of planar rotation on the recognition of animals and nonanimal objects. Response times for recognizing animals followed an inverted U-shaped function, whereas those…

  2. Neuritin reverses deficits in murine novel object associative recognition memory caused by exposure to extremely low-frequency (50 Hz) electromagnetic fields

    PubMed Central

    Zhao, Qian-Ru; Lu, Jun-Mei; Yao, Jin-Jing; Zhang, Zheng-Yu; Ling, Chen; Mei, Yan-Ai

    2015-01-01

    Animal studies have shown that electromagnetic field exposure may interfere with the activity of brain cells, thereby generating behavioral and cognitive disturbances. However, the underlying mechanisms and possible preventions are still unknown. In this study, we used a mouse model to examine the effects of exposure to extremely low-frequency (50 Hz) electromagnetic fields (ELF MFs) on a recognition memory task and morphological changes of hippocampal neurons. The data showed that ELF MFs exposure (1 mT, 12 h/day) induced a time-dependent deficit in novel object associative recognition memory and also decreased hippocampal dendritic spine density. This effect was observed without corresponding changes in spontaneous locomotor activity and was transient, which has only been seen after exposing mice to ELF MFs for 7-10 days. The over-expression of hippocampal neuritin, an activity-dependent neurotrophic factor, using an adeno-associated virus (AAV) vector significantly increased the neuritin level and dendritic spine density. This increase was paralleled with ELF MFs exposure-induced deficits in recognition memory and reductions of dendritic spine density. Collectively, our study provides evidence for the association between ELF MFs exposure, impairment of recognition memory, and resulting changes in hippocampal dendritic spine density. Neuritin prevented this ELF MFs-exposure-induced effect by increasing the hippocampal spine density. PMID:26138388

  3. Real-time Human Activity Recognition

    NASA Astrophysics Data System (ADS)

    Albukhary, N.; Mustafah, Y. M.

    2017-11-01

    The traditional Closed-circuit Television (CCTV) system requires human to monitor the CCTV for 24/7 which is inefficient and costly. Therefore, there’s a need for a system which can recognize human activity effectively in real-time. This paper concentrates on recognizing simple activity such as walking, running, sitting, standing and landing by using image processing techniques. Firstly, object detection is done by using background subtraction to detect moving object. Then, object tracking and object classification are constructed so that different person can be differentiated by using feature detection. Geometrical attributes of tracked object, which are centroid and aspect ratio of identified tracked are manipulated so that simple activity can be detected.

  4. A Temporally Distinct Role for Group I and Group II Metabotropic Glutamate Receptors in Object Recognition Memory

    ERIC Educational Resources Information Center

    Brown, Malcolm Watson; Warburton, Elizabeth Clea; Barker, Gareth Robert Isaac; Bashir, Zafar Iqbal

    2006-01-01

    Recognition memory, involving the ability to discriminate between a novel and familiar object, depends on the integrity of the perirhinal cortex (PRH). Glutamate, the main excitatory neurotransmitter in the cortex, is essential for many types of memory processes. Of the subtypes of glutamate receptor, metabotropic receptors (mGluRs) have received…

  5. Operator for object recognition and scene analysis by estimation of set occupancy with noisy and incomplete data sets

    NASA Astrophysics Data System (ADS)

    Rees, S. J.; Jones, Bryan F.

    1992-11-01

    Once feature extraction has occurred in a processed image, the recognition problem becomes one of defining a set of features which maps sufficiently well onto one of the defined shape/object models to permit a claimed recognition. This process is usually handled by aggregating features until a large enough weighting is obtained to claim membership, or an adequate number of located features are matched to the reference set. A requirement has existed for an operator or measure capable of a more direct assessment of membership/occupancy between feature sets, particularly where the feature sets may be defective representations. Such feature set errors may be caused by noise, by overlapping of objects, and by partial obscuration of features. These problems occur at the point of acquisition: repairing the data would then assume a priori knowledge of the solution. The technique described in this paper offers a set theoretical measure for partial occupancy defined in terms of the set of minimum additions to permit full occupancy and the set of locations of occupancy if such additions are made. As is shown, this technique permits recognition of partial feature sets with quantifiable degrees of uncertainty. A solution to the problems of obscuration and overlapping is therefore available.

  6. On-chip learning of hyper-spectral data for real time target recognition

    NASA Technical Reports Server (NTRS)

    Duong, T. A.; Daud, T.; Thakoor, A.

    2000-01-01

    As the focus of our present paper, we have used the cascade error projection (CEP) learning algorithm (shown to be hardware-implementable) with on-chip learning (OCL) scheme to obtain three orders of magnitude speed-up in target recognition compared to software-based learning schemes. Thus, it is shown, real time learning as well as data processing for target recognition can be achieved.

  7. Evidence for the Activation of Sensorimotor Information during Visual Word Recognition: The Body-Object Interaction Effect

    ERIC Educational Resources Information Center

    Siakaluk, Paul D.; Pexman, Penny M.; Aguilera, Laura; Owen, William J.; Sears, Christopher R.

    2008-01-01

    We examined the effects of sensorimotor experience in two visual word recognition tasks. Body-object interaction (BOI) ratings were collected for a large set of words. These ratings assess perceptions of the ease with which a human body can physically interact with a word's referent. A set of high BOI words (e.g., "mask") and a set of low BOI…

  8. Intracerebroventricular oxytocin administration in rats enhances object recognition and increases expression of neurotrophins, microtubule-associated protein 2, and synapsin I.

    PubMed

    Havranek, Tomas; Zatkova, Martina; Lestanova, Zuzana; Bacova, Zuzana; Mravec, Boris; Hodosy, Julius; Strbak, Vladimir; Bakos, Jan

    2015-06-01

    Brain oxytocin regulates a variety of social and affiliative behaviors and affects also learning and memory. However, mechanisms of its action at the level of neuronal circuits are not fully understood. The present study tests the hypothesis that molecular factors required for memory formation and synaptic plasticity, including brain-derived neurotrophic factor, neural growth factor, nestin, microtubule-associated protein 2 (MAP2), and synapsin I, are enhanced by central administration of oxytocin. We also investigated whether oxytocin enhances object recognition and acts as anxiolytic agent. Therefore, male Wistar rats were infused continuously with oxytocin (20 ng/µl) via an osmotic minipump into the lateral cerebral ventricle for 7 days; controls were infused with vehicle. The object recognition test, open field test, and elevated plus maze test were performed on the sixth, seventh, and eighth days from starting the infusion. No significant effects of oxytocin on anxious-like behavior were observed. The object recognition test showed that oxytocin-treated rats significantly preferred unknown objects. Oxytocin treatment significantly increased gene expression and protein levels of neurotrophins, MAP2, and synapsin I in the hippocampus. No changes were observed in nestin expression. Our results provide the first direct evidence implicating oxytocin as a regulator of brain plasticity at the level of changes of neuronal growth factors, cytoskeletal proteins, and behavior. The data support assumption that oxytocin is important for short-term hippocampus-dependent memory. © 2015 Wiley Periodicals, Inc.

  9. Optical Pattern Recognition for Missile Guidance.

    DTIC Science & Technology

    1982-11-15

    directed to novel pattern recognition algo- rithms (that allow pattern recognition and object classification in the face of various geometrical and...I wats EF5 = 50) p.j/t’ni 2 (for btith image pat tern recognitio itas a preproicessing oiperatiton. Ini devices). TIhe rt’ad light intensity (0.33t mW...electrodes on its large faces . This Priz light modulator and the motivation for its devel- SLM is known as the Prom (Pockels real-time optical opment. In Sec

  10. Human action recognition based on kinematic similarity in real time

    PubMed Central

    Chen, Longting; Luo, Ailing; Zhang, Sicong

    2017-01-01

    Human action recognition using 3D pose data has gained a growing interest in the field of computer robotic interfaces and pattern recognition since the availability of hardware to capture human pose. In this paper, we propose a fast, simple, and powerful method of human action recognition based on human kinematic similarity. The key to this method is that the action descriptor consists of joints position, angular velocity and angular acceleration, which can meet the different individual sizes and eliminate the complex normalization. The angular parameters of joints within a short sliding time window (approximately 5 frames) around the current frame are used to express each pose frame of human action sequence. Moreover, three modified KNN (k-nearest-neighbors algorithm) classifiers are employed in our method: one for achieving the confidence of every frame in the training step, one for estimating the frame label of each descriptor, and one for classifying actions. Additional estimating of the frame’s time label makes it possible to address single input frames. This approach can be used on difficult, unsegmented sequences. The proposed method is efficient and can be run in real time. The research shows that many public datasets are irregularly segmented, and a simple method is provided to regularize the datasets. The approach is tested on some challenging datasets such as MSR-Action3D, MSRDailyActivity3D, and UTD-MHAD. The results indicate our method achieves a higher accuracy. PMID:29073131

  11. Neural dynamics of object-based multifocal visual spatial attention and priming: Object cueing, useful-field-of-view, and crowding

    PubMed Central

    Foley, Nicholas C.; Grossberg, Stephen; Mingolla, Ennio

    2015-01-01

    How are spatial and object attention coordinated to achieve rapid object learning and recognition during eye movement search? How do prefrontal priming and parietal spatial mechanisms interact to determine the reaction time costs of intra-object attention shifts, inter-object attention shifts, and shifts between visible objects and covertly cued locations? What factors underlie individual differences in the timing and frequency of such attentional shifts? How do transient and sustained spatial attentional mechanisms work and interact? How can volition, mediated via the basal ganglia, influence the span of spatial attention? A neural model is developed of how spatial attention in the where cortical stream coordinates view-invariant object category learning in the what cortical stream under free viewing conditions. The model simulates psychological data about the dynamics of covert attention priming and switching requiring multifocal attention without eye movements. The model predicts how “attentional shrouds” are formed when surface representations in cortical area V4 resonate with spatial attention in posterior parietal cortex (PPC) and prefrontal cortex (PFC), while shrouds compete among themselves for dominance. Winning shrouds support invariant object category learning, and active surface-shroud resonances support conscious surface perception and recognition. Attentive competition between multiple objects and cues simulates reaction-time data from the two-object cueing paradigm. The relative strength of sustained surface-driven and fast-transient motion-driven spatial attention controls individual differences in reaction time for invalid cues. Competition between surface-driven attentional shrouds controls individual differences in detection rate of peripheral targets in useful-field-of-view tasks. The model proposes how the strength of competition can be mediated, though learning or momentary changes in volition, by the basal ganglia. A new explanation of

  12. Neural dynamics of object-based multifocal visual spatial attention and priming: object cueing, useful-field-of-view, and crowding.

    PubMed

    Foley, Nicholas C; Grossberg, Stephen; Mingolla, Ennio

    2012-08-01

    How are spatial and object attention coordinated to achieve rapid object learning and recognition during eye movement search? How do prefrontal priming and parietal spatial mechanisms interact to determine the reaction time costs of intra-object attention shifts, inter-object attention shifts, and shifts between visible objects and covertly cued locations? What factors underlie individual differences in the timing and frequency of such attentional shifts? How do transient and sustained spatial attentional mechanisms work and interact? How can volition, mediated via the basal ganglia, influence the span of spatial attention? A neural model is developed of how spatial attention in the where cortical stream coordinates view-invariant object category learning in the what cortical stream under free viewing conditions. The model simulates psychological data about the dynamics of covert attention priming and switching requiring multifocal attention without eye movements. The model predicts how "attentional shrouds" are formed when surface representations in cortical area V4 resonate with spatial attention in posterior parietal cortex (PPC) and prefrontal cortex (PFC), while shrouds compete among themselves for dominance. Winning shrouds support invariant object category learning, and active surface-shroud resonances support conscious surface perception and recognition. Attentive competition between multiple objects and cues simulates reaction-time data from the two-object cueing paradigm. The relative strength of sustained surface-driven and fast-transient motion-driven spatial attention controls individual differences in reaction time for invalid cues. Competition between surface-driven attentional shrouds controls individual differences in detection rate of peripheral targets in useful-field-of-view tasks. The model proposes how the strength of competition can be mediated, though learning or momentary changes in volition, by the basal ganglia. A new explanation of

  13. Appearance-based face recognition and light-fields.

    PubMed

    Gross, Ralph; Matthews, Iain; Baker, Simon

    2004-04-01

    Arguably the most important decision to be made when developing an object recognition algorithm is selecting the scene measurements or features on which to base the algorithm. In appearance-based object recognition, the features are chosen to be the pixel intensity values in an image of the object. These pixel intensities correspond directly to the radiance of light emitted from the object along certain rays in space. The set of all such radiance values over all possible rays is known as the plenoptic function or light-field. In this paper, we develop a theory of appearance-based object recognition from light-fields. This theory leads directly to an algorithm for face recognition across pose that uses as many images of the face as are available, from one upwards. All of the pixels, whichever image they come from, are treated equally and used to estimate the (eigen) light-field of the object. The eigen light-field is then used as the set of features on which to base recognition, analogously to how the pixel intensities are used in appearance-based face and object recognition.

  14. False recall and recognition of brand names increases over time.

    PubMed

    Sherman, Susan M

    2013-01-01

    Using the Deese-Roediger-McDermott (DRM) paradigm, participants are presented with lists of associated words (e.g., bed, awake, night). Subsequently, they reliably have false memories for related but nonpresented words (e.g., SLEEP). Previous research has found that false memories can be created for brand names (e.g., Morrisons, Sainsbury's, Waitrose, and TESCO). The present study investigates the effect of a week's delay on false memories for brand names. Participants were presented with lists of brand names followed by a distractor task. In two between-subjects experiments, participants completed a free recall task or a recognition task either immediately or a week later. In two within-subjects experiments, participants completed a free recall task or a recognition task both immediately and a week later. Correct recall for presented list items decreased over time, whereas false recall for nonpresented lure items increased. For recognition, raw scores revealed an increase in false memory across time reflected in an increase in Remember responses. Analysis of Pr scores revealed that false memory for lures stayed constant over a week, but with an increase in Remember responses in the between-subjects experiment and a trend in the same direction in the within-subjects experiment. Implications for theories of false memory are discussed.

  15. Effects of heavy particle irradiation and diet on object recognition memory in rats

    NASA Astrophysics Data System (ADS)

    Rabin, Bernard M.; Carrihill-Knoll, Kirsty; Hinchman, Marie; Shukitt-Hale, Barbara; Joseph, James A.; Foster, Brian C.

    2009-04-01

    On long-duration missions to other planets astronauts will be exposed to types and doses of radiation that are not experienced in low earth orbit. Previous research using a ground-based model for exposure to cosmic rays has shown that exposure to heavy particles, such as 56Fe, disrupts spatial learning and memory measured using the Morris water maze. Maintaining rats on diets containing antioxidant phytochemicals for 2 weeks prior to irradiation ameliorated this deficit. The present experiments were designed to determine: (1) the generality of the particle-induced disruption of memory by examining the effects of exposure to 56Fe particles on object recognition memory; and (2) whether maintaining rats on these antioxidant diets for 2 weeks prior to irradiation would also ameliorate any potential deficit. The results showed that exposure to low doses of 56Fe particles does disrupt recognition memory and that maintaining rats on antioxidant diets containing blueberry and strawberry extract for only 2 weeks was effective in ameliorating the disruptive effects of irradiation. The results are discussed in terms of the mechanisms by which exposure to these particles may produce effects on neurocognitive performance.

  16. Character Recognition Method by Time-Frequency Analyses Using Writing Pressure

    NASA Astrophysics Data System (ADS)

    Watanabe, Tatsuhito; Katsura, Seiichiro

    With the development of information and communication technology, personal verification becomes more and more important. In the future ubiquitous society, the development of terminals handling personal information requires the personal verification technology. The signature is one of the personal verification methods; however, the number of characters is limited in the case of the signature and therefore false signature is used easily. Thus, personal identification is difficult from handwriting. This paper proposes a “haptic pen” that extracts the writing pressure, and shows a character recognition method by time-frequency analyses. Although the figures of characters written by different amanuenses are similar, the differences appear in the time-frequency domain. As a result, it is possible to use the proposed character recognition for personal identification more exactly. The experimental results showed the viability of the proposed method.

  17. Altered object exploration but not temporal order memory retrieval in an object recognition test following treatment of rats with the group II metabotropic glutamate receptor agonist LY379268.

    PubMed

    Lins, Brittney R; Ballendine, Stephanie A; Howland, John G

    2014-02-07

    Temporal order memory refers to the ability to distinguish past experiences in the order that they occurred. Temporal order memory for objects is often tested in rodents using spontaneous object recognition paradigms. The circuitry mediating memory in these tests is distributed and involves ionotropic glutamate receptors in the perirhinal cortex and medial prefrontal cortex. It is unknown what role, if any, metabotropic glutamate receptors have in temporal order memory for objects. The present experiment examined the role of metabotropic glutamate receptors in temporal memory retrieval using the group II metabotropic glutamate receptor selective agonist LY379268. Rats were trained on a temporal memory test with three phases: two sample phases (60 min between them) in which rats explored two novel objects and a test phase (60 min after the second sample phase) which included a copy of each object previously encountered. Under these conditions, we confirmed that rats showed a significant exploratory preference for the object presented during the first sample phase. In a second experiment, we found that LY379268 (0.3, 1.0, or 3.0mg/kg; i.p.; 30 min before the test phase) had no effect on temporal memory retrieval but dose-dependently reduced time spent exploring the objects. Our results show that enhancing mGluR2 activity under conditions when TM is intact does not influence memory retrieval. Copyright © 2013 Elsevier Ireland Ltd. All rights reserved.

  18. Real-time image restoration for iris recognition systems.

    PubMed

    Kang, Byung Jun; Park, Kang Ryoung

    2007-12-01

    In the field of biometrics, it has been reported that iris recognition techniques have shown high levels of accuracy because unique patterns of the human iris, which has very many degrees of freedom, are used. However, because conventional iris cameras have small depth-of-field (DOF) areas, input iris images can easily be blurred, which can lead to lower recognition performance, since iris patterns are transformed by the blurring caused by optical defocusing. To overcome these problems, an autofocusing camera can be used. However, this inevitably increases the cost, size, and complexity of the system. Therefore, we propose a new real-time iris image-restoration method, which can increase the camera's DOF without requiring any additional hardware. This paper presents five novelties as compared to previous works: 1) by excluding eyelash and eyelid regions, it is possible to obtain more accurate focus scores from input iris images; 2) the parameter of the point spread function (PSF) can be estimated in terms of camera optics and measured focus scores; therefore, parameter estimation is more accurate than it has been in previous research; 3) because the PSF parameter can be obtained by using a predetermined equation, iris image restoration can be done in real-time; 4) by using a constrained least square (CLS) restoration filter that considers noise, performance can be greatly enhanced; and 5) restoration accuracy can also be enhanced by estimating the weight value of the noise-regularization term of the CLS filter according to the amount of image blurring. Experimental results showed that iris recognition errors when using the proposed restoration method were greatly reduced as compared to those results achieved without restoration or those achieved using previous iris-restoration methods.

  19. Learned Non-Rigid Object Motion is a View-Invariant Cue to Recognizing Novel Objects

    PubMed Central

    Chuang, Lewis L.; Vuong, Quoc C.; Bülthoff, Heinrich H.

    2012-01-01

    There is evidence that observers use learned object motion to recognize objects. For instance, studies have shown that reversing the learned direction in which a rigid object rotated in depth impaired recognition accuracy. This motion reversal can be achieved by playing animation sequences of moving objects in reverse frame order. In the current study, we used this sequence-reversal manipulation to investigate whether observers encode the motion of dynamic objects in visual memory, and whether such dynamic representations are encoded in a way that is dependent on the viewing conditions. Participants first learned dynamic novel objects, presented as animation sequences. Following learning, they were then tested on their ability to recognize these learned objects when their animation sequence was shown in the same sequence order as during learning or in the reverse sequence order. In Experiment 1, we found that non-rigid motion contributed to recognition performance; that is, sequence-reversal decreased sensitivity across different tasks. In subsequent experiments, we tested the recognition of non-rigidly deforming (Experiment 2) and rigidly rotating (Experiment 3) objects across novel viewpoints. Recognition performance was affected by viewpoint changes for both experiments. Learned non-rigid motion continued to contribute to recognition performance and this benefit was the same across all viewpoint changes. By comparison, learned rigid motion did not contribute to recognition performance. These results suggest that non-rigid motion provides a source of information for recognizing dynamic objects, which is not affected by changes to viewpoint. PMID:22661939

  20. Administration of the Phosphodiesterase Type 4 Inhibitor Rolipram into the Amygdala at a Specific Time Interval after Learning Increases Recognition Memory Persistence

    ERIC Educational Resources Information Center

    Werenicz, Aline; Christoff, Raissa R.; Blank, Martina; Jobim, Paulo F. C.; Pedroso, Thiago R.; Reolon, Gustavo K.; Schroder, Nadja; Roesler, Rafael

    2012-01-01

    Here we show that administration of the phosphodiesterase type 4 (PDE4) inhibitor rolipram into the basolateral complex of the amygdala (BLA) at a specific time interval after training enhances memory consolidation and induces memory persistence for novel object recognition (NOR) in rats. Intra-BLA infusion of rolipram immediately, 1.5 h, or 6 h…

  1. The time course of individual face recognition: A pattern analysis of ERP signals.

    PubMed

    Nemrodov, Dan; Niemeier, Matthias; Mok, Jenkin Ngo Yin; Nestor, Adrian

    2016-05-15

    An extensive body of work documents the time course of neural face processing in the human visual cortex. However, the majority of this work has focused on specific temporal landmarks, such as N170 and N250 components, derived through univariate analyses of EEG data. Here, we take on a broader evaluation of ERP signals related to individual face recognition as we attempt to move beyond the leading theoretical and methodological framework through the application of pattern analysis to ERP data. Specifically, we investigate the spatiotemporal profile of identity recognition across variation in emotional expression. To this end, we apply pattern classification to ERP signals both in time, for any single electrode, and in space, across multiple electrodes. Our results confirm the significance of traditional ERP components in face processing. At the same time though, they support the idea that the temporal profile of face recognition is incompletely described by such components. First, we show that signals associated with different facial identities can be discriminated from each other outside the scope of these components, as early as 70ms following stimulus presentation. Next, electrodes associated with traditional ERP components as well as, critically, those not associated with such components are shown to contribute information to stimulus discriminability. And last, the levels of ERP-based pattern discrimination are found to correlate with recognition accuracy across subjects confirming the relevance of these methods for bridging brain and behavior data. Altogether, the current results shed new light on the fine-grained time course of neural face processing and showcase the value of novel methods for pattern analysis to investigating fundamental aspects of visual recognition. Copyright © 2016 Elsevier Inc. All rights reserved.

  2. Role of the Anterior Cingulate Cortex in the Retrieval of Novel Object Recognition Memory after a Long Delay

    ERIC Educational Resources Information Center

    Pezze, Marie A.; Marshall, Hayley J.; Fone, Kevin C. F.; Cassaday, Helen J.

    2017-01-01

    Previous in vivo electrophysiological studies suggest that the anterior cingulate cortex (ACgx) is an important substrate of novel object recognition (NOR) memory. However, intervention studies are needed to confirm this conclusion and permanent lesion studies cannot distinguish effects on encoding and retrieval. The interval between encoding and…

  3. A comparison of moving object detection methods for real-time moving object detection

    NASA Astrophysics Data System (ADS)

    Roshan, Aditya; Zhang, Yun

    2014-06-01

    Moving object detection has a wide variety of applications from traffic monitoring, site monitoring, automatic theft identification, face detection to military surveillance. Many methods have been developed across the globe for moving object detection, but it is very difficult to find one which can work globally in all situations and with different types of videos. The purpose of this paper is to evaluate existing moving object detection methods which can be implemented in software on a desktop or laptop, for real time object detection. There are several moving object detection methods noted in the literature, but few of them are suitable for real time moving object detection. Most of the methods which provide for real time movement are further limited by the number of objects and the scene complexity. This paper evaluates the four most commonly used moving object detection methods as background subtraction technique, Gaussian mixture model, wavelet based and optical flow based methods. The work is based on evaluation of these four moving object detection methods using two (2) different sets of cameras and two (2) different scenes. The moving object detection methods have been implemented using MatLab and results are compared based on completeness of detected objects, noise, light change sensitivity, processing time etc. After comparison, it is observed that optical flow based method took least processing time and successfully detected boundary of moving objects which also implies that it can be implemented for real-time moving object detection.

  4. Towards Wearable A-Mode Ultrasound Sensing for Real-Time Finger Motion Recognition.

    PubMed

    Yang, Xingchen; Sun, Xueli; Zhou, Dalin; Li, Yuefeng; Liu, Honghai

    2018-06-01

    It is evident that surface electromyography (sEMG) based human-machine interfaces (HMI) have inherent difficulty in predicting dexterous musculoskeletal movements such as finger motions. This paper is an attempt to investigate a plausible alternative to sEMG, ultrasound-driven HMI, for dexterous motion recognition due to its characteristic of detecting morphological changes of deep muscles and tendons. A multi-channel A-mode ultrasound lightweight device is adopted to evaluate the performance of finger motion recognition; an experiment is designed for both widely acceptable offline and online algorithms with eight able-bodied subjects employed. The experiment result presents that the offline recognition accuracy is up to 98.83% ± 0.79%. The real-time motion completion rate is 95.4% ± 8.7% and online motion selection time is 0.243 ± 0.127 s. The outcomes confirm the feasibility of A-mode ultrasound based wearable HMI and its prosperous applications in prosthetic devices, virtual reality, and remote manipulation.

  5. Chronic cannabidiol treatment improves social and object recognition in double transgenic APPswe/PS1∆E9 mice.

    PubMed

    Cheng, David; Low, Jac Kee; Logge, Warren; Garner, Brett; Karl, Tim

    2014-08-01

    Patients suffering from Alzheimer's disease (AD) exhibit a decline in cognitive abilities including an inability to recognise familiar faces. Hallmark pathological changes in AD include the aggregation of amyloid-β (Aβ), tau protein hyperphosphorylation as well as pronounced neurodegeneration, neuroinflammation, neurotoxicity and oxidative damage. The non-psychoactive phytocannabinoid cannabidiol (CBD) exerts neuroprotective, anti-oxidant and anti-inflammatory effects and promotes neurogenesis. CBD also reverses Aβ-induced spatial memory deficits in rodents. Thus we determined the therapeutic-like effects of chronic CBD treatment (20 mg/kg, daily intraperitoneal injections for 3 weeks) on the APPswe/PS1∆E9 (APPxPS1) transgenic mouse model for AD in a number of cognitive tests, including the social preference test, the novel object recognition task and the fear conditioning paradigm. We also analysed the impact of CBD on anxiety behaviours in the elevated plus maze. Vehicle-treated APPxPS1 mice demonstrated impairments in social recognition and novel object recognition compared to wild type-like mice. Chronic CBD treatment reversed these cognitive deficits in APPxPS1 mice without affecting anxiety-related behaviours. This is the first study to investigate the effect of chronic CBD treatment on cognition in an AD transgenic mouse model. Our findings suggest that CBD may have therapeutic potential for specific cognitive impairments associated with AD.

  6. Risperidone reverses the spatial object recognition impairment and hippocampal BDNF-TrkB signalling system alterations induced by acute MK-801 treatment

    PubMed Central

    Chen, Guangdong; Lin, Xiaodong; Li, Gongying; Jiang, Diego; Lib, Zhiruo; Jiang, Ronghuan; Zhuo, Chuanjun

    2017-01-01

    The aim of the present study was to investigate the effects of a commonly-used atypical antipsychotic, risperidone, on alterations in spatial learning and in the hippocampal brain-derived neurotrophic factor (BDNF)-tyrosine receptor kinase B (TrkB) signalling system caused by acute dizocilpine maleate (MK-801) treatment. In experiment 1, adult male Sprague-Dawley rats subjected to acute treatment of either low-dose MK801 (0.1 mg/kg) or normal saline (vehicle) were tested for spatial object recognition and hippocampal expression levels of BDNF, TrkB and the phophorylation of TrkB (p-TrkB). We found that compared to the vehicle, MK-801 treatment impaired spatial object recognition of animals and downregulated the expression levels of p-TrkB. In experiment 2, MK-801- or vehicle-treated animals were further injected with risperidone (0.1 mg/kg) or vehicle before behavioural testing and sacrifice. Of note, we found that risperidone successfully reversed the deleterious effects of MK-801 on spatial object recognition and upregulated the hippocampal BDNF-TrkB signalling system. Collectively, the findings suggest that cognitive deficits from acute N-methyl-D-aspartate receptor blockade may be associated with the hypofunction of hippocampal BDNF-TrkB signalling system and that risperidone was able to reverse these alterations. PMID:28451387

  7. Modafinil improves methamphetamine-induced object recognition deficits and restores prefrontal cortex ERK signaling in mice.

    PubMed

    González, Betina; Raineri, Mariana; Cadet, Jean Lud; García-Rill, Edgar; Urbano, Francisco J; Bisagno, Veronica

    2014-12-01

    Chronic use of methamphetamine (METH) leads to long-lasting cognitive dysfunction in humans and in animal models. Modafinil is a wake-promoting compound approved for the treatment of sleeping disorders. It is also prescribed off label to treat METH dependence. In the present study, we investigated whether modafinil could improve cognitive deficits induced by sub-chronic METH treatment in mice by measuring visual retention in a Novel Object Recognition (NOR) task. After sub-chronic METH treatment (1 mg/kg, once a day for 7 days), mice performed the NOR task, which consisted of habituation to the object recognition arena (5 min a day, 3 consecutive days), training session (2 equal objects, 10 min, day 4), and a retention session (1 novel object, 5 min, day 5). One hour before the training session, mice were given a single dose of modafinil (30 or 90 mg/kg). METH-treated mice showed impairments in visual memory retention, evidenced by equal preference of familiar and novel objects during the retention session. The lower dose of modafinil (30 mg/kg) had no effect on visual retention scores in METH-treated mice, while the higher dose (90 mg/kg) rescued visual memory retention to control values. We also measured extracellular signal-regulated kinase (ERK) phosphorylation in medial prefrontal cortex (mPFC), hippocampus, and nucleus accumbens (NAc) of METH- and vehicle-treated mice that received modafinil 1 h before exposure to novel objects in the training session, compared to mice placed in the arena without objects. Elevated ERK phosphorylation was found in the mPFC of vehicle-treated mice, but not in METH-treated mice, exposed to objects. The lower dose of modafinil had no effect on ERK phosphorylation in METH-treated mice, while 90 mg/kg modafinil treatment restored the ERK phosphorylation induced by novelty in METH-treated mice to values comparable to controls. We found neither a novelty nor treatment effect on ERK phosphorylation in hippocampus or NAc of vehicle

  8. Remembering the object you fear: brain potentials during recognition of spiders in spider-fearful individuals.

    PubMed

    Michalowski, Jaroslaw M; Weymar, Mathias; Hamm, Alfons O

    2014-01-01

    In the present study we investigated long-term memory for unpleasant, neutral and spider pictures in 15 spider-fearful and 15 non-fearful control individuals using behavioral and electrophysiological measures. During the initial (incidental) encoding, pictures were passively viewed in three separate blocks and were subsequently rated for valence and arousal. A recognition memory task was performed one week later in which old and new unpleasant, neutral and spider pictures were presented. Replicating previous results, we found enhanced memory performance and higher confidence ratings for unpleasant when compared to neutral materials in both animal fearful individuals and controls. When compared to controls high animal fearful individuals also showed a tendency towards better memory accuracy and significantly higher confidence during recognition of spider pictures, suggesting that memory of objects prompting specific fear is also facilitated in fearful individuals. In line, spider-fearful but not control participants responded with larger ERP positivity for correctly recognized old when compared to correctly rejected new spider pictures, thus showing the same effects in the neural signature of emotional memory for feared objects that were already discovered for other emotional materials. The increased fear memory for phobic materials observed in the present study in spider-fearful individuals might result in an enhanced fear response and reinforce negative beliefs aggravating anxiety symptomatology and hindering recovery.

  9. Young Children's Self-Generated Object Views and Object Recognition

    ERIC Educational Resources Information Center

    James, Karin H.; Jones, Susan S.; Smith, Linda B.; Swain, Shelley N.

    2014-01-01

    Two important and related developments in children between 18 and 24 months of age are the rapid expansion of object name vocabularies and the emergence of an ability to recognize objects from sparse representations of their geometric shapes. In the same period, children also begin to show a preference for planar views (i.e., views of objects held…

  10. Evidence for the activation of sensorimotor information during visual word recognition: the body-object interaction effect.

    PubMed

    Siakaluk, Paul D; Pexman, Penny M; Aguilera, Laura; Owen, William J; Sears, Christopher R

    2008-01-01

    We examined the effects of sensorimotor experience in two visual word recognition tasks. Body-object interaction (BOI) ratings were collected for a large set of words. These ratings assess perceptions of the ease with which a human body can physically interact with a word's referent. A set of high BOI words (e.g., mask) and a set of low BOI words (e.g., ship) were created, matched on imageability and concreteness. Facilitatory BOI effects were observed in lexical decision and phonological lexical decision tasks: responses were faster for high BOI words than for low BOI words. We discuss how our findings may be accounted for by (a) semantic feedback within the visual word recognition system, and (b) an embodied view of cognition (e.g., Barsalou's perceptual symbol systems theory), which proposes that semantic knowledge is grounded in sensorimotor interactions with the environment.

  11. Automatic Target Recognition Based on Cross-Plot

    PubMed Central

    Wong, Kelvin Kian Loong; Abbott, Derek

    2011-01-01

    Automatic target recognition that relies on rapid feature extraction of real-time target from photo-realistic imaging will enable efficient identification of target patterns. To achieve this objective, Cross-plots of binary patterns are explored as potential signatures for the observed target by high-speed capture of the crucial spatial features using minimal computational resources. Target recognition was implemented based on the proposed pattern recognition concept and tested rigorously for its precision and recall performance. We conclude that Cross-plotting is able to produce a digital fingerprint of a target that correlates efficiently and effectively to signatures of patterns having its identity in a target repository. PMID:21980508

  12. Involvement of hippocampal NMDA receptors in encoding and consolidation, but not retrieval, processes of spontaneous object location memory in rats.

    PubMed

    Yamada, Kazuo; Arai, Misaki; Suenaga, Toshiko; Ichitani, Yukio

    2017-07-28

    The hippocampus is thought to be involved in object location recognition memory, yet the contribution of hippocampal NMDA receptors to the memory processes, such as encoding, retention and retrieval, is unknown. First, we confirmed that hippocampal infusion of a competitive NMDA receptor antagonist, AP5 (2-amino-5-phosphonopentanoic acid, 20-40nmol), impaired performance of spontaneous object location recognition test but not that of novel object recognition test in Wistar rats. Next, the effects of hippocampal AP5 treatment on each process of object location recognition memory were examined with three different injection times using a 120min delay-interposed test: 15min before the sample phase (Time I), immediately after the sample phase (Time II), and 15min before the test phase (Time III). The blockade of hippocampal NMDA receptors before and immediately after the sample phase, but not before the test phase, markedly impaired performance of object location recognition test, suggesting that hippocampal NMDA receptors play an important role in encoding and consolidation/retention, but not retrieval, of spontaneous object location memory. Copyright © 2017 Elsevier B.V. All rights reserved.

  13. Automated speech recognition for time recording in out-of-hospital emergency medicine-an experimental approach.

    PubMed

    Gröschel, J; Philipp, F; Skonetzki, St; Genzwürker, H; Wetter, Th; Ellinger, K

    2004-02-01

    Precise documentation of medical treatment in emergency medical missions and for resuscitation is essential from a medical, legal and quality assurance point of view [Anästhesiologie und Intensivmedizin, 41 (2000) 737]. All conventional methods of time recording are either too inaccurate or elaborate for routine application. Automated speech recognition may offer a solution. A special erase programme for the documentation of all time events was developed. Standard speech recognition software (IBM ViaVoice 7.0) was adapted and installed on two different computer systems. One was a stationary PC (500MHz Pentium III, 128MB RAM, Soundblaster PCI 128 Soundcard, Win NT 4.0), the other was a mobile pen-PC that had already proven its value during emergency missions [Der Notarzt 16, p. 177] (Fujitsu Stylistic 2300, 230Mhz MMX Processor, 160MB RAM, embedded soundcard ESS 1879 chipset, Win98 2nd ed.). On both computers two different microphones were tested. One was a standard headset that came with the recognition software, the other was a small microphone (Lavalier-Kondensatormikrofon EM 116 from Vivanco), that could be attached to the operators collar. Seven women and 15 men spoke a text with 29 phrases to be recognised. Two emergency physicians tested the system in a simulated emergency setting using the collar microphone and the pen-PC with an analogue wireless connection. Overall recognition was best for the PC with a headset (89%) followed by the pen-PC with a headset (85%), the PC with a microphone (84%) and the pen-PC with a microphone (80%). Nevertheless, the difference was not statistically significant. Recognition became significantly worse (89.5% versus 82.3%, P<0.0001 ) when numbers had to be recognised. The gender of speaker and the number of words in a sentence had no influence. Average recognition in the simulated emergency setting was 75%. At no time did false recognition appear. Time recording with automated speech recognition seems to be possible in

  14. The Effect of Inversion on 3- to 5-Year-Olds' Recognition of Face and Nonface Visual Objects

    ERIC Educational Resources Information Center

    Picozzi, Marta; Cassia, Viola Macchi; Turati, Chiara; Vescovo, Elena

    2009-01-01

    This study compared the effect of stimulus inversion on 3- to 5-year-olds' recognition of faces and two nonface object categories matched with faces for a number of attributes: shoes (Experiment 1) and frontal images of cars (Experiments 2 and 3). The inversion effect was present for faces but not shoes at 3 years of age (Experiment 1). Analogous…

  15. A Longitudinal Study of Cognitive Representation in Symbolic Play, Self-recognition, and Object Permanence during the Second Year.

    ERIC Educational Resources Information Center

    Chapman, Michael

    1987-01-01

    Explores development of cognitive representation in 20 infants 12 to 24 months of age with regard to (l) their understanding of agency in symbolic play (agent use), (2) recognition of their own mirror image, and (3) object permanence. Results were generally consistent with developmental sequences predicted by Fischer's Skill Theory for agent use…

  16. Estradiol-Induced Object Recognition Memory Consolidation Is Dependent on Activation of mTOR Signaling in the Dorsal Hippocampus

    ERIC Educational Resources Information Center

    Fortress, Ashley M.; Fan, Lu; Orr, Patrick T.; Zhao, Zaorui; Frick, Karyn M.

    2013-01-01

    The mammalian target of rapamycin (mTOR) signaling pathway is an important regulator of protein synthesis and is essential for various forms of hippocampal memory. Here, we asked whether the enhancement of object recognition memory consolidation produced by dorsal hippocampal infusion of 17[Beta]-estradiol (E[subscript 2]) is dependent on mTOR…

  17. Effect of 1.8 GHz radiofrequency electromagnetic radiation on novel object associative recognition memory in mice

    PubMed Central

    Wang, Kai; Lu, Jun-Mei; Xing, Zhen-He; Zhao, Qian-Ru; Hu, Lin-Qi; Xue, Lei; Zhang, Jie; Mei, Yan-Ai

    2017-01-01

    Mounting evidence suggests that exposure to radiofrequency electromagnetic radiation (RF-EMR) can influence learning and memory in rodents. In this study, we examined the effects of single exposure to 1.8 GHz RF-EMR for 30 min on subsequent recognition memory in mice, using the novel object recognition task (NORT). RF-EMR exposure at an intensity of >2.2 W/kg specific absorption rate (SAR) power density induced a significant density-dependent increase in NORT index with no corresponding changes in spontaneous locomotor activity. RF-EMR exposure increased dendritic-spine density and length in hippocampal and prefrontal cortical neurons, as shown by Golgi staining. Whole-cell recordings in acute hippocampal and medial prefrontal cortical slices showed that RF-EMR exposure significantly altered the resting membrane potential and action potential frequency, and reduced the action potential half-width, threshold, and onset delay in pyramidal neurons. These results demonstrate that exposure to 1.8 GHz RF-EMR for 30 min can significantly increase recognition memory in mice, and can change dendritic-spine morphology and neuronal excitability in the hippocampus and prefrontal cortex. The SAR in this study (3.3 W/kg) was outside the range encountered in normal daily life, and its relevance as a potential therapeutic approach for disorders associated with recognition memory deficits remains to be clarified. PMID:28303965

  18. The short and long term effects of docetaxel chemotherapy on rodent object recognition and spatial reference memory.

    PubMed

    Fardell, Joanna E; Vardy, Janette; Johnston, Ian N

    2013-10-17

    Previous animal studies have examined the potential for cytostatic drugs to induce learning and memory deficits in laboratory animals but, to date, there is no pre-clinical evidence that taxanes have the potential to cause cognitive impairment. Therefore our aim was to explore the short- and long-term cognitive effects of different dosing schedules of the taxane docetaxel (DTX) on laboratory rodents. Healthy male hooded Wistar rats were treated with DTX (6 mg/kg, 10mg/kg) or physiological saline (control), once a week for 3 weeks (Experiment 1) or once only (10mg/kg; Experiment 2). Cognitive function was assessed using the novel object recognition (NOR) task and spatial water maze (WM) task 1 to 3 weeks after treatment and again 4 months after treatment. Shortly after DTX treatment, rats perform poorly on NOR regardless of treatment regimen. Treatment with a single injection of 10mg/kg DTX does not appear to induce sustained deficits in object recognition or peripheral neuropathy. Overall these findings show that treatment with the taxane DTX in the absence of cancer and other anti-cancer treatments causes cognitive impairment in healthy rodents. Copyright © 2013 Elsevier Inc. All rights reserved.

  19. Behavior analysis of video object in complicated background

    NASA Astrophysics Data System (ADS)

    Zhao, Wenting; Wang, Shigang; Liang, Chao; Wu, Wei; Lu, Yang

    2016-10-01

    This paper aims to achieve robust behavior recognition of video object in complicated background. Features of the video object are described and modeled according to the depth information of three-dimensional video. Multi-dimensional eigen vector are constructed and used to process high-dimensional data. Stable object tracing in complex scenes can be achieved with multi-feature based behavior analysis, so as to obtain the motion trail. Subsequently, effective behavior recognition of video object is obtained according to the decision criteria. What's more, the real-time of algorithms and accuracy of analysis are both improved greatly. The theory and method on the behavior analysis of video object in reality scenes put forward by this project have broad application prospect and important practical significance in the security, terrorism, military and many other fields.

  20. Learning Models and Real-Time Speech Recognition.

    ERIC Educational Resources Information Center

    Danforth, Douglas G.; And Others

    This report describes the construction and testing of two "psychological" learning models for the purpose of computer recognition of human speech over the telephone. One of the two models was found to be superior in all tests. A regression analysis yielded a 92.3% recognition rate for 14 subjects ranging in age from 6 to 13 years. Tests…

  1. Time-Elastic Generative Model for Acceleration Time Series in Human Activity Recognition.

    PubMed

    Munoz-Organero, Mario; Ruiz-Blazquez, Ramona

    2017-02-08

    Body-worn sensors in general and accelerometers in particular have been widely used in order to detect human movements and activities. The execution of each type of movement by each particular individual generates sequences of time series of sensed data from which specific movement related patterns can be assessed. Several machine learning algorithms have been used over windowed segments of sensed data in order to detect such patterns in activity recognition based on intermediate features (either hand-crafted or automatically learned from data). The underlying assumption is that the computed features will capture statistical differences that can properly classify different movements and activities after a training phase based on sensed data. In order to achieve high accuracy and recall rates (and guarantee the generalization of the system to new users), the training data have to contain enough information to characterize all possible ways of executing the activity or movement to be detected. This could imply large amounts of data and a complex and time-consuming training phase, which has been shown to be even more relevant when automatically learning the optimal features to be used. In this paper, we present a novel generative model that is able to generate sequences of time series for characterizing a particular movement based on the time elasticity properties of the sensed data. The model is used to train a stack of auto-encoders in order to learn the particular features able to detect human movements. The results of movement detection using a newly generated database with information on five users performing six different movements are presented. The generalization of results using an existing database is also presented in the paper. The results show that the proposed mechanism is able to obtain acceptable recognition rates ( F = 0.77) even in the case of using different people executing a different sequence of movements and using different hardware.

  2. An observational study of implicit motor imagery using laterality recognition of the hand after stroke.

    PubMed

    Amesz, Sarah; Tessari, Alessia; Ottoboni, Giovanni; Marsden, Jon

    2016-01-01

    To explore the relationship between laterality recognition after stroke and impairments in attention, 3D object rotation and functional ability. Observational cross-sectional study. Acute care teaching hospital. Thirty-two acute and sub-acute people with stroke and 36 healthy, age-matched controls. Laterality recognition, attention and mental rotation of objects. Within the stroke group, the relationship between laterality recognition and functional ability, neglect, hemianopia and dyspraxia were further explored. People with stroke were significantly less accurate (69% vs 80%) and showed delayed reaction times (3.0 vs 1.9 seconds) when determining the laterality of a pictured hand. Deficits either in accuracy or reaction times were seen in 53% of people with stroke. The accuracy of laterality recognition was associated with reduced functional ability (R(2) = 0.21), less accurate mental rotation of objects (R(2) = 0.20) and dyspraxia (p = 0.03). Implicit motor imagery is affected in a significant number of patients after stroke with these deficits related to lesions to the motor networks as well as other deficits seen after stroke. This research provides new insights into how laterality recognition is related to a number of other deficits after stroke, including the mental rotation of 3D objects, attention and dyspraxia. Further research is required to determine if treatment programmes can improve deficits in laterality recognition and impact functional outcomes after stroke.

  3. Conducting spoken word recognition research online: Validation and a new timing method.

    PubMed

    Slote, Joseph; Strand, Julia F

    2016-06-01

    Models of spoken word recognition typically make predictions that are then tested in the laboratory against the word recognition scores of human subjects (e.g., Luce & Pisoni Ear and Hearing, 19, 1-36, 1998). Unfortunately, laboratory collection of large sets of word recognition data can be costly and time-consuming. Due to the numerous advantages of online research in speed, cost, and participant diversity, some labs have begun to explore the use of online platforms such as Amazon's Mechanical Turk (AMT) to source participation and collect data (Buhrmester, Kwang, & Gosling Perspectives on Psychological Science, 6, 3-5, 2011). Many classic findings in cognitive psychology have been successfully replicated online, including the Stroop effect, task-switching costs, and Simon and flanker interference (Crump, McDonnell, & Gureckis PLoS ONE, 8, e57410, 2013). However, tasks requiring auditory stimulus delivery have not typically made use of AMT. In the present study, we evaluated the use of AMT for collecting spoken word identification and auditory lexical decision data. Although online users were faster and less accurate than participants in the lab, the results revealed strong correlations between the online and laboratory measures for both word identification accuracy and lexical decision speed. In addition, the scores obtained in the lab and online were equivalently correlated with factors that have been well established to predict word recognition, including word frequency and phonological neighborhood density. We also present and analyze a method for precise auditory reaction timing that is novel to behavioral research. Taken together, these findings suggest that AMT can be a viable alternative to the traditional laboratory setting as a source of participation for some spoken word recognition research.

  4. Dissociating electrophysiological correlates of subjective, objective, and correct memory in investigating the emotion-induced recognition bias.

    PubMed

    Windmann, Sabine; Hill, Holger

    2014-10-01

    Performance on tasks requiring discrimination of at least two stimuli can be viewed either from an objective perspective (referring to actual stimulus differences), or from a subjective perspective (corresponding to participant's responses). Using event-related potentials recorded during an old/new recognition memory test involving emotionally laden and neutral words studied either blockwise or randomly intermixed, we show here how the objective perspective (old versus new items) yields late effects of blockwise emotional item presentation at parietal sites that the subjective perspective fails to find, whereas the subjective perspective ("old" versus "new" responses) is more sensitive to early effects of emotion at anterior sites than the objective perspective. Our results demonstrate the potential advantage of dissociating the subjective and the objective perspective onto task performance (in addition to analyzing trials with correct responses), especially for investigations of illusions and information processing biases, in behavioral and cognitive neuroscience studies. Copyright © 2014 Elsevier Inc. All rights reserved.

  5. Automatic anatomy recognition via multiobject oriented active shape models.

    PubMed

    Chen, Xinjian; Udupa, Jayaram K; Alavi, Abass; Torigian, Drew A

    2010-12-01

    This paper studies the feasibility of developing an automatic anatomy recognition (AAR) system in clinical radiology and demonstrates its operation on clinical 2D images. The anatomy recognition method described here consists of two main components: (a) multiobject generalization of OASM and (b) object recognition strategies. The OASM algorithm is generalized to multiple objects by including a model for each object and assigning a cost structure specific to each object in the spirit of live wire. The delineation of multiobject boundaries is done in MOASM via a three level dynamic programming algorithm, wherein the first level is at pixel level which aims to find optimal oriented boundary segments between successive landmarks, the second level is at landmark level which aims to find optimal location for the landmarks, and the third level is at the object level which aims to find optimal arrangement of object boundaries over all objects. The object recognition strategy attempts to find that pose vector (consisting of translation, rotation, and scale component) for the multiobject model that yields the smallest total boundary cost for all objects. The delineation and recognition accuracies were evaluated separately utilizing routine clinical chest CT, abdominal CT, and foot MRI data sets. The delineation accuracy was evaluated in terms of true and false positive volume fractions (TPVF and FPVF). The recognition accuracy was assessed (1) in terms of the size of the space of the pose vectors for the model assembly that yielded high delineation accuracy, (2) as a function of the number of objects and objects' distribution and size in the model, (3) in terms of the interdependence between delineation and recognition, and (4) in terms of the closeness of the optimum recognition result to the global optimum. When multiple objects are included in the model, the delineation accuracy in terms of TPVF can be improved to 97%-98% with a low FPVF of 0.1%-0.2%. Typically, a

  6. Real-time color/shape-based traffic signs acquisition and recognition system

    NASA Astrophysics Data System (ADS)

    Saponara, Sergio

    2013-02-01

    A real-time system is proposed to acquire from an automotive fish-eye CMOS camera the traffic signs, and provide their automatic recognition on the vehicle network. Differently from the state-of-the-art, in this work color-detection is addressed exploiting the HSI color space which is robust to lighting changes. Hence the first stage of the processing system implements fish-eye correction and RGB to HSI transformation. After color-based detection a noise deletion step is implemented and then, for the classification, a template-based correlation method is adopted to identify potential traffic signs, of different shapes, from acquired images. Starting from a segmented-image a matching with templates of the searched signs is carried out using a distance transform. These templates are organized hierarchically to reduce the number of operations and hence easing real-time processing for several types of traffic signs. Finally, for the recognition of the specific traffic sign, a technique based on extraction of signs characteristics and thresholding is adopted. Implemented on DSP platform the system recognizes traffic signs in less than 150 ms at a distance of about 15 meters from 640x480-pixel acquired images. Tests carried out with hundreds of images show a detection and recognition rate of about 93%.

  7. Magnet hospital recognition in hospital systems over time.

    PubMed

    Lasater, Karen B; Richards, Michael R; Dandapani, Nikila B; Burns, Lawton R; McHugh, Matthew D

    2017-06-13

    Magnet hospitals are recognized for nursing excellence and high-value patient outcomes, yet little is known about which and when hospitals pursue Magnet recognition. Concurrently, hospital systems are becoming a more prominent feature of the U.S. health care landscape. The aim of the study was to examine Magnet adoption among hospital systems over time. Using American Hospital Association surveys (1998-2012), we characterized the proportion of Magnet hospitals belonging to systems. We used hospital level fixed-effects regressions to capture changes in a given system hospital's Magnet status over time in relation to a variety of conditions, including prior Magnet adoption by system affiliates and nonaffiliates in local and geographically distant markets and whether these relationships varied by degree of system centralization. The proportion of Magnet hospitals belonging to a system is increasing. Prior Magnet adoption by a hospital within the local market was associated with an increased likelihood of a given system hospital becoming Magnet, but the effect was larger if there was prior adoption by affiliates (7.4% higher likelihood) versus nonaffiliates (2.7% higher likelihood). Prior adoption by affiliates and nonaffiliates in geographically distant markets had a lesser effect. Hospitals belonging to centralized systems were more reactive to Magnet adoption of nonaffiliate hospitals as compared with those in decentralized systems. Hospital systems take an organizational perspective toward Magnet adoption, whereby more system affiliates achieve Magnet recognition over time. The findings are relevant to health care and nursing administrators and policymakers interested in the diffusion of an empirically supported organizational innovation associated with quality outcomes, particularly in a time of increasing hospital consolidation and system expansion. We identify factors associated with Magnet adoption across system hospitals and demonstrate the importance of

  8. Influence of time and length size feature selections for human activity sequences recognition.

    PubMed

    Fang, Hongqing; Chen, Long; Srinivasan, Raghavendiran

    2014-01-01

    In this paper, Viterbi algorithm based on a hidden Markov model is applied to recognize activity sequences from observed sensors events. Alternative features selections of time feature values of sensors events and activity length size feature values are tested, respectively, and then the results of activity sequences recognition performances of Viterbi algorithm are evaluated. The results show that the selection of larger time feature values of sensor events and/or smaller activity length size feature values will generate relatively better results on the activity sequences recognition performances. © 2013 ISA Published by ISA All rights reserved.

  9. Representational dynamics of object recognition: Feedforward and feedback information flows.

    PubMed

    Goddard, Erin; Carlson, Thomas A; Dermody, Nadene; Woolgar, Alexandra

    2016-03-01

    Object perception involves a range of visual and cognitive processes, and is known to include both a feedfoward flow of information from early visual cortical areas to higher cortical areas, along with feedback from areas such as prefrontal cortex. Previous studies have found that low and high spatial frequency information regarding object identity may be processed over different timescales. Here we used the high temporal resolution of magnetoencephalography (MEG) combined with multivariate pattern analysis to measure information specifically related to object identity in peri-frontal and peri-occipital areas. Using stimuli closely matched in their low-level visual content, we found that activity in peri-occipital cortex could be used to decode object identity from ~80ms post stimulus onset, and activity in peri-frontal cortex could also be used to decode object identity from a later time (~265ms post stimulus onset). Low spatial frequency information related to object identity was present in the MEG signal at an earlier time than high spatial frequency information for peri-occipital cortex, but not for peri-frontal cortex. We additionally used Granger causality analysis to compare feedforward and feedback influences on representational content, and found evidence of both an early feedfoward flow and later feedback flow of information related to object identity. We discuss our findings in relation to existing theories of object processing and propose how the methods we use here could be used to address further questions of the neural substrates underlying object perception. Copyright © 2016 Elsevier Inc. All rights reserved.

  10. Metacognitive Influences on Study Time Allocation in an Associative Recognition Task: An Analysis of Adult Age Differences

    PubMed Central

    Hines, Jarrod C.; Touron, Dayna R.; Hertzog, Christopher

    2009-01-01

    The current study evaluated a metacognitive account of study time allocation, which argues that metacognitive monitoring of recognition test accuracy and latency influences subsequent strategic control and regulation. We examined judgments of learning (JOLs), recognition test confidence judgments (CJs), and subjective response time (RT) judgments by younger and older adults in an associative recognition task involving two study-test phases, with self-paced study in phase 2. Multilevel regression analyses assessed the degree to which age and metacognitive variables predicted phase 2 study time independent of actual test accuracy and RT. Outcomes supported the metacognitive account – JOLs and CJs predicted study time independent of recognition accuracy. For older adults with errant RT judgments, subjective retrieval fluency influenced response confidence as well as (mediated through confidence) subsequent study time allocation. Older adults studied items longer which had been assigned lower CJs, suggesting no age deficit in using memory monitoring to control learning. PMID:19485662

  11. Self-Recognition in Autistic Children.

    ERIC Educational Resources Information Center

    Dawson, Geraldine; McKissick, Fawn Celeste

    1984-01-01

    Fifteen autistic children (four to six years old) were assessed for visual self-recognition ability, as well as for object permanence and gestural imitation. It was found that 13 of 15 autistic children showed evidence of self-recognition. Consistent relationships were suggested between self-cognition and object permanence but not between…

  12. Genetic specificity of face recognition.

    PubMed

    Shakeshaft, Nicholas G; Plomin, Robert

    2015-10-13

    Specific cognitive abilities in diverse domains are typically found to be highly heritable and substantially correlated with general cognitive ability (g), both phenotypically and genetically. Recent twin studies have found the ability to memorize and recognize faces to be an exception, being similarly heritable but phenotypically substantially uncorrelated both with g and with general object recognition. However, the genetic relationships between face recognition and other abilities (the extent to which they share a common genetic etiology) cannot be determined from phenotypic associations. In this, to our knowledge, first study of the genetic associations between face recognition and other domains, 2,000 18- and 19-year-old United Kingdom twins completed tests assessing their face recognition, object recognition, and general cognitive abilities. Results confirmed the substantial heritability of face recognition (61%), and multivariate genetic analyses found that most of this genetic influence is unique and not shared with other cognitive abilities.

  13. Robust Real-Time and Rotation-Invariant American Sign Language Alphabet Recognition Using Range Camera

    NASA Astrophysics Data System (ADS)

    Lahamy, H.; Lichti, D.

    2012-07-01

    The automatic interpretation of human gestures can be used for a natural interaction with computers without the use of mechanical devices such as keyboards and mice. The recognition of hand postures have been studied for many years. However, most of the literature in this area has considered 2D images which cannot provide a full description of the hand gestures. In addition, a rotation-invariant identification remains an unsolved problem even with the use of 2D images. The objective of the current study is to design a rotation-invariant recognition process while using a 3D signature for classifying hand postures. An heuristic and voxelbased signature has been designed and implemented. The tracking of the hand motion is achieved with the Kalman filter. A unique training image per posture is used in the supervised classification. The designed recognition process and the tracking procedure have been successfully evaluated. This study has demonstrated the efficiency of the proposed rotation invariant 3D hand posture signature which leads to 98.24% recognition rate after testing 12723 samples of 12 gestures taken from the alphabet of the American Sign Language.

  14. 36 CFR 218.10 - Objection time periods and process.

    Code of Federal Regulations, 2012 CFR

    2012-07-01

    ... 36 Parks, Forests, and Public Property 2 2012-07-01 2012-07-01 false Objection time periods and... Objection time periods and process. (a) Time to file an objection. Written objections, including any... of objectors to ensure that their objection is received in a timely manner. (b) Computation of time...

  15. Beyond perceptual expertise: revisiting the neural substrates of expert object recognition

    PubMed Central

    Harel, Assaf; Kravitz, Dwight; Baker, Chris I.

    2013-01-01

    Real-world expertise provides a valuable opportunity to understand how experience shapes human behavior and neural function. In the visual domain, the study of expert object recognition, such as in car enthusiasts or bird watchers, has produced a large, growing, and often-controversial literature. Here, we synthesize this literature, focusing primarily on results from functional brain imaging, and propose an interactive framework that incorporates the impact of high-level factors, such as attention and conceptual knowledge, in supporting expertise. This framework contrasts with the perceptual view of object expertise that has concentrated largely on stimulus-driven processing in visual cortex. One prominent version of this perceptual account has almost exclusively focused on the relation of expertise to face processing and, in terms of the neural substrates, has centered on face-selective cortical regions such as the Fusiform Face Area (FFA). We discuss the limitations of this face-centric approach as well as the more general perceptual view, and highlight that expert related activity is: (i) found throughout visual cortex, not just FFA, with a strong relationship between neural response and behavioral expertise even in the earliest stages of visual processing, (ii) found outside visual cortex in areas such as parietal and prefrontal cortices, and (iii) modulated by the attentional engagement of the observer suggesting that it is neither automatic nor driven solely by stimulus properties. These findings strongly support a framework in which object expertise emerges from extensive interactions within and between the visual system and other cognitive systems, resulting in widespread, distributed patterns of expertise-related activity across the entire cortex. PMID:24409134

  16. Modeling Fan Effects on the Time Course of Associative Recognition

    ERIC Educational Resources Information Center

    Schneider, Darryl W.; Anderson, John R.

    2012-01-01

    We investigated the time course of associative recognition using the response signal procedure, whereby a stimulus is presented and followed after a variable lag by a signal indicating that an immediate response is required. More specifically, we examined the effects of associative fan (the number of associations that an item has with other items…

  17. Comparison Analysis of Recognition Algorithms of Forest-Cover Objects on Hyperspectral Air-Borne and Space-Borne Images

    NASA Astrophysics Data System (ADS)

    Kozoderov, V. V.; Kondranin, T. V.; Dmitriev, E. V.

    2017-12-01

    The basic model for the recognition of natural and anthropogenic objects using their spectral and textural features is described in the problem of hyperspectral air-borne and space-borne imagery processing. The model is based on improvements of the Bayesian classifier that is a computational procedure of statistical decision making in machine-learning methods of pattern recognition. The principal component method is implemented to decompose the hyperspectral measurements on the basis of empirical orthogonal functions. Application examples are shown of various modifications of the Bayesian classifier and Support Vector Machine method. Examples are provided of comparing these classifiers and a metrical classifier that operates on finding the minimal Euclidean distance between different points and sets in the multidimensional feature space. A comparison is also carried out with the " K-weighted neighbors" method that is close to the nonparametric Bayesian classifier.

  18. Two speed factors of visual recognition independently correlated with fluid intelligence.

    PubMed

    Tachibana, Ryosuke; Namba, Yuri; Noguchi, Yasuki

    2014-01-01

    Growing evidence indicates a moderate but significant relationship between processing speed in visuo-cognitive tasks and general intelligence. On the other hand, findings from neuroscience proposed that the primate visual system consists of two major pathways, the ventral pathway for objects recognition and the dorsal pathway for spatial processing and attentive analysis. Previous studies seeking for visuo-cognitive factors of human intelligence indicated a significant correlation between fluid intelligence and the inspection time (IT), an index for a speed of object recognition performed in the ventral pathway. We thus presently examined a possibility that neural processing speed in the dorsal pathway also represented a factor of intelligence. Specifically, we used the mental rotation (MR) task, a popular psychometric measure for mental speed of spatial processing in the dorsal pathway. We found that the speed of MR was significantly correlated with intelligence scores, while it had no correlation with one's IT (recognition speed of visual objects). Our results support the new possibility that intelligence could be explained by two types of mental speed, one related to object recognition (IT) and another for manipulation of mental images (MR).

  19. Two Speed Factors of Visual Recognition Independently Correlated with Fluid Intelligence

    PubMed Central

    Tachibana, Ryosuke; Namba, Yuri; Noguchi, Yasuki

    2014-01-01

    Growing evidence indicates a moderate but significant relationship between processing speed in visuo-cognitive tasks and general intelligence. On the other hand, findings from neuroscience proposed that the primate visual system consists of two major pathways, the ventral pathway for objects recognition and the dorsal pathway for spatial processing and attentive analysis. Previous studies seeking for visuo-cognitive factors of human intelligence indicated a significant correlation between fluid intelligence and the inspection time (IT), an index for a speed of object recognition performed in the ventral pathway. We thus presently examined a possibility that neural processing speed in the dorsal pathway also represented a factor of intelligence. Specifically, we used the mental rotation (MR) task, a popular psychometric measure for mental speed of spatial processing in the dorsal pathway. We found that the speed of MR was significantly correlated with intelligence scores, while it had no correlation with one’s IT (recognition speed of visual objects). Our results support the new possibility that intelligence could be explained by two types of mental speed, one related to object recognition (IT) and another for manipulation of mental images (MR). PMID:24825574

  20. Automatic anatomy recognition on CT images with pathology

    NASA Astrophysics Data System (ADS)

    Huang, Lidong; Udupa, Jayaram K.; Tong, Yubing; Odhner, Dewey; Torigian, Drew A.

    2016-03-01

    Body-wide anatomy recognition on CT images with pathology becomes crucial for quantifying body-wide disease burden. This, however, is a challenging problem because various diseases result in various abnormalities of objects such as shape and intensity patterns. We previously developed an automatic anatomy recognition (AAR) system [1] whose applicability was demonstrated on near normal diagnostic CT images in different body regions on 35 organs. The aim of this paper is to investigate strategies for adapting the previous AAR system to diagnostic CT images of patients with various pathologies as a first step toward automated body-wide disease quantification. The AAR approach consists of three main steps - model building, object recognition, and object delineation. In this paper, within the broader AAR framework, we describe a new strategy for object recognition to handle abnormal images. In the model building stage an optimal threshold interval is learned from near-normal training images for each object. This threshold is optimally tuned to the pathological manifestation of the object in the test image. Recognition is performed following a hierarchical representation of the objects. Experimental results for the abdominal body region based on 50 near-normal images used for model building and 20 abnormal images used for object recognition show that object localization accuracy within 2 voxels for liver and spleen and 3 voxels for kidney can be achieved with the new strategy.

  1. A validated set of tool pictures with matched objects and non-objects for laterality research.

    PubMed

    Verma, Ark; Brysbaert, Marc

    2015-01-01

    Neuropsychological and neuroimaging research has established that knowledge related to tool use and tool recognition is lateralized to the left cerebral hemisphere. Recently, behavioural studies with the visual half-field technique have confirmed the lateralization. A limitation of this research was that different sets of stimuli had to be used for the comparison of tools to other objects and objects to non-objects. Therefore, we developed a new set of stimuli containing matched triplets of tools, other objects and non-objects. With the new stimulus set, we successfully replicated the findings of no visual field advantage for objects in an object recognition task combined with a significant right visual field advantage for tools in a tool recognition task. The set of stimuli is available as supplemental data to this article.

  2. Speech Recognition as a Transcription Aid: A Randomized Comparison With Standard Transcription

    PubMed Central

    Mohr, David N.; Turner, David W.; Pond, Gregory R.; Kamath, Joseph S.; De Vos, Cathy B.; Carpenter, Paul C.

    2003-01-01

    Objective. Speech recognition promises to reduce information entry costs for clinical information systems. It is most likely to be accepted across an organization if physicians can dictate without concerning themselves with real-time recognition and editing; assistants can then edit and process the computer-generated document. Our objective was to evaluate the use of speech-recognition technology in a randomized controlled trial using our institutional infrastructure. Design. Clinical note dictations from physicians in two specialty divisions were randomized to either a standard transcription process or a speech-recognition process. Secretaries and transcriptionists also were assigned randomly to each of these processes. Measurements. The duration of each dictation was measured. The amount of time spent processing a dictation to yield a finished document also was measured. Secretarial and transcriptionist productivity, defined as hours of secretary work per minute of dictation processed, was determined for speech recognition and standard transcription. Results. Secretaries in the endocrinology division were 87.3% (confidence interval, 83.3%, 92.3%) as productive with the speech-recognition technology as implemented in this study as they were using standard transcription. Psychiatry transcriptionists and secretaries were similarly less productive. Author, secretary, and type of clinical note were significant (p < 0.05) predictors of productivity. Conclusion. When implemented in an organization with an existing document-processing infrastructure (which included training and interfaces of the speech-recognition editor with the existing document entry application), speech recognition did not improve the productivity of secretaries or transcriptionists. PMID:12509359

  3. Genetic specificity of face recognition

    PubMed Central

    Shakeshaft, Nicholas G.; Plomin, Robert

    2015-01-01

    Specific cognitive abilities in diverse domains are typically found to be highly heritable and substantially correlated with general cognitive ability (g), both phenotypically and genetically. Recent twin studies have found the ability to memorize and recognize faces to be an exception, being similarly heritable but phenotypically substantially uncorrelated both with g and with general object recognition. However, the genetic relationships between face recognition and other abilities (the extent to which they share a common genetic etiology) cannot be determined from phenotypic associations. In this, to our knowledge, first study of the genetic associations between face recognition and other domains, 2,000 18- and 19-year-old United Kingdom twins completed tests assessing their face recognition, object recognition, and general cognitive abilities. Results confirmed the substantial heritability of face recognition (61%), and multivariate genetic analyses found that most of this genetic influence is unique and not shared with other cognitive abilities. PMID:26417086

  4. Oxytocin Reduces Face Processing Time but Leaves Recognition Accuracy and Eye-Gaze Unaffected.

    PubMed

    Hubble, Kelly; Daughters, Katie; Manstead, Antony S R; Rees, Aled; Thapar, Anita; van Goozen, Stephanie H M

    2017-01-01

    Previous studies have found that oxytocin (OXT) can improve the recognition of emotional facial expressions; it has been proposed that this effect is mediated by an increase in attention to the eye-region of faces. Nevertheless, evidence in support of this claim is inconsistent, and few studies have directly tested the effect of oxytocin on emotion recognition via altered eye-gaze Methods: In a double-blind, within-subjects, randomized control experiment, 40 healthy male participants received 24 IU intranasal OXT and placebo in two identical experimental sessions separated by a 2-week interval. Visual attention to the eye-region was assessed on both occasions while participants completed a static facial emotion recognition task using medium intensity facial expressions. Although OXT had no effect on emotion recognition accuracy, recognition performance was improved because face processing was faster across emotions under the influence of OXT. This effect was marginally significant (p<.06). Consistent with a previous study using dynamic stimuli, OXT had no effect on eye-gaze patterns when viewing static emotional faces and this was not related to recognition accuracy or face processing time. These findings suggest that OXT-induced enhanced facial emotion recognition is not necessarily mediated by an increase in attention to the eye-region of faces, as previously assumed. We discuss several methodological issues which may explain discrepant findings and suggest the effect of OXT on visual attention may differ depending on task requirements. (JINS, 2017, 23, 23-33).

  5. 36 CFR 218.10 - Objection time periods and process.

    Code of Federal Regulations, 2010 CFR

    2010-07-01

    ... calendar day (11:59 p.m. in the time zone of the receiving office) for objections filed by electronic means... 36 Parks, Forests, and Public Property 2 2010-07-01 2010-07-01 false Objection time periods and... Objection time periods and process. (a) Time to file an objection. Written objections, including any...

  6. Effect of Context and Hearing Loss on Time-Gated Word Recognition in Children

    PubMed Central

    Lewis, Dawna E.; Kopun, Judy; McCreery, Ryan; Brennan, Marc; Nishi, Kanae; Cordrey, Evan; Stelmachowicz, Pat; Moeller, Mary Pat

    2016-01-01

    Objectives The purpose of this study was to examine word recognition in children who are hard of hearing (CHH) and children with normal hearing (CNH) in response to time-gated words presented in high- vs. low-predictability sentences (HP, LP), where semantic cues were manipulated. Findings inform our understanding of how CHH combine cognitive-linguistic and acoustic-phonetic cues to support spoken word recognition. It was hypothesized that both groups of children would be able to make use of linguistic cues provided by HP sentences to support word recognition. CHH were expected to require greater acoustic information (more gates) than CNH to correctly identify words in the LP condition. In addition, it was hypothesized that error patterns would differ across groups. Design Sixteen CHH with mild-to-moderate hearing loss and 16 age-matched CNH participated (5–12 yrs). Test stimuli included 15 LP and 15 HP age-appropriate sentences. The final word of each sentence was divided into segments and recombined with the sentence frame to create series of sentences in which the final word was progressively longer by the gated increments. Stimuli were presented monaurally through headphones and children were asked to identify the target word at each successive gate. They also were asked to rate their confidence in their word choice using a 5- or 3-point scale. For CHH, the signals were processed through a hearing aid simulator. Standardized language measures were used to assess the contribution of linguistic skills. Results Analysis of language measures revealed that the CNH and CHH performed within the average range on language abilities. Both groups correctly recognized a significantly higher percentage of words in the HP condition than in the LP condition. Although CHH performed comparably to CNH in terms of successfully recognizing the majority of words, differences were observed in the amount of acoustic-phonetic information needed to achieve accurate word recognition

  7. Age, environment, object recognition and morphological diversity of GFAP-immunolabeled astrocytes.

    PubMed

    Diniz, Daniel Guerreiro; de Oliveira, Marcus Augusto; de Lima, Camila Mendes; Fôro, César Augusto Raiol; Sosthenes, Marcia Consentino Kronka; Bento-Torres, João; da Costa Vasconcelos, Pedro Fernando; Anthony, Daniel Clive; Diniz, Cristovam Wanderley Picanço

    2016-10-10

    Few studies have explored the glial response to a standard environment and how the response may be associated with age-related cognitive decline in learning and memory. Here we investigated aging and environmental influences on hippocampal-dependent tasks and on the morphology of an unbiased selected population of astrocytes from the molecular layer of dentate gyrus, which is the main target of perforant pathway. Six and twenty-month-old female, albino Swiss mice were housed, from weaning, in a standard or enriched environment, including running wheels for exercise and tested for object recognition and contextual memories. Young adult and aged subjects, independent of environment, were able to distinguish familiar from novel objects. All experimental groups, except aged mice from standard environment, distinguish stationary from displaced objects. Young adult but not aged mice, independent of environment, were able to distinguish older from recent objects. Only young mice from an enriched environment were able to distinguish novel from familiar contexts. Unbiased selected astrocytes from the molecular layer of the dentate gyrus were reconstructed in three-dimensions and classified using hierarchical cluster analysis of bimodal or multimodal morphological features. We found two morphological phenotypes of astrocytes and we designated type I the astrocytes that exhibited significantly higher values of morphological complexity as compared with type II. Complexity = [Sum of the terminal orders + Number of terminals] × [Total branch length/Number of primary branches]. On average, type I morphological complexity seems to be much more sensitive to age and environmental influences than that of type II. Indeed, aging and environmental impoverishment interact and reduce the morphological complexity of type I astrocytes at a point that they could not be distinguished anymore from type II. We suggest these two types of astrocytes may have different physiological roles

  8. Deep Networks Can Resemble Human Feed-forward Vision in Invariant Object Recognition

    PubMed Central

    Kheradpisheh, Saeed Reza; Ghodrati, Masoud; Ganjtabesh, Mohammad; Masquelier, Timothée

    2016-01-01

    Deep convolutional neural networks (DCNNs) have attracted much attention recently, and have shown to be able to recognize thousands of object categories in natural image databases. Their architecture is somewhat similar to that of the human visual system: both use restricted receptive fields, and a hierarchy of layers which progressively extract more and more abstracted features. Yet it is unknown whether DCNNs match human performance at the task of view-invariant object recognition, whether they make similar errors and use similar representations for this task, and whether the answers depend on the magnitude of the viewpoint variations. To investigate these issues, we benchmarked eight state-of-the-art DCNNs, the HMAX model, and a baseline shallow model and compared their results to those of humans with backward masking. Unlike in all previous DCNN studies, we carefully controlled the magnitude of the viewpoint variations to demonstrate that shallow nets can outperform deep nets and humans when variations are weak. When facing larger variations, however, more layers were needed to match human performance and error distributions, and to have representations that are consistent with human behavior. A very deep net with 18 layers even outperformed humans at the highest variation level, using the most human-like representations. PMID:27601096

  9. Face recognition increases during saccade preparation.

    PubMed

    Lin, Hai; Rizak, Joshua D; Ma, Yuan-ye; Yang, Shang-chuan; Chen, Lin; Hu, Xin-tian

    2014-01-01

    Face perception is integral to human perception system as it underlies social interactions. Saccadic eye movements are frequently made to bring interesting visual information, such as faces, onto the fovea for detailed processing. Just before eye movement onset, the processing of some basic features, such as the orientation, of an object improves at the saccade landing point. Interestingly, there is also evidence that indicates faces are processed in early visual processing stages similar to basic features. However, it is not known whether this early enhancement of processing includes face recognition. In this study, three experiments were performed to map the timing of face presentation to the beginning of the eye movement in order to evaluate pre-saccadic face recognition. Faces were found to be similarly processed as simple objects immediately prior to saccadic movements. Starting ∼ 120 ms before a saccade to a target face, independent of whether or not the face was surrounded by other faces, the face recognition gradually improved and the critical spacing of the crowding decreased as saccade onset was approaching. These results suggest that an upcoming saccade prepares the visual system for new information about faces at the saccade landing site and may reduce the background in a crowd to target the intended face. This indicates an important role of pre-saccadic eye movement signals in human face recognition.

  10. The Struggle for Identity in Today's Schools: Cultural Recognition in a Time of Increasing Diversity

    ERIC Educational Resources Information Center

    Jenlink, Patrick M., Ed; Townes, Faye Hicks, Ed

    2009-01-01

    The "Struggle for Identity in Today's Schools" examines cultural recognition and the struggle for identity in America's schools. In particular, the contributing authors focus on the recognition and misrecognition as antagonistic cultural forces that work to shape, and at times distort identity. What surfaces throughout the chapters are two lessons…

  11. Detection of hidden objects using a real-time 3-D millimeter-wave imaging system

    NASA Astrophysics Data System (ADS)

    Rozban, Daniel; Aharon, Avihai; Levanon, Assaf; Abramovich, Amir; Yitzhaky, Yitzhak; Kopeika, N. S.

    2014-10-01

    Millimeter (mm)and sub-mm wavelengths or terahertz (THz) band have several properties that motivate their use in imaging for security applications such as recognition of hidden objects, dangerous materials, aerosols, imaging through walls as in hostage situations, and also in bad weather conditions. There is no known ionization hazard for biological tissue, and atmospheric degradation of THz radiation is relatively low for practical imaging distances. We recently developed a new technology for the detection of THz radiation. This technology is based on very inexpensive plasma neon indicator lamps, also known as Glow Discharge Detector (GDD), that can be used as very sensitive THz radiation detectors. Using them, we designed and constructed a Focal Plane Array (FPA) and obtained recognizable2-dimensional THz images of both dielectric and metallic objects. Using THz wave it is shown here that even concealed weapons made of dielectric material can be detected. An example is an image of a knife concealed inside a leather bag and also under heavy clothing. Three-dimensional imaging using radar methods can enhance those images since it can allow the isolation of the concealed objects from the body and environmental clutter such as nearby furniture or other people. The GDDs enable direct heterodyning between the electric field of the target signal and the reference signal eliminating the requirement for expensive mixers, sources, and Low Noise Amplifiers (LNAs).We expanded the ability of the FPA so that we are able to obtain recognizable 2-dimensional THz images in real time. We show here that the THz detection of objects in three dimensions, using FMCW principles is also applicable in real time. This imaging system is also shown here to be capable of imaging objects from distances allowing standoff detection of suspicious objects and humans from large distances.

  12. Line fitting based feature extraction for object recognition

    NASA Astrophysics Data System (ADS)

    Li, Bing

    2014-06-01

    Image feature extraction plays a significant role in image based pattern applications. In this paper, we propose a new approach to generate hierarchical features. This new approach applies line fitting to adaptively divide regions based upon the amount of information and creates line fitting features for each subsequent region. It overcomes the feature wasting drawback of the wavelet based approach and demonstrates high performance in real applications. For gray scale images, we propose a diffusion equation approach to map information-rich pixels (pixels near edges and ridge pixels) into high values, and pixels in homogeneous regions into small values near zero that form energy map images. After the energy map images are generated, we propose a line fitting approach to divide regions recursively and create features for each region simultaneously. This new feature extraction approach is similar to wavelet based hierarchical feature extraction in which high layer features represent global characteristics and low layer features represent local characteristics. However, the new approach uses line fitting to adaptively focus on information-rich regions so that we avoid the feature waste problems of the wavelet approach in homogeneous regions. Finally, the experiments for handwriting word recognition show that the new method provides higher performance than the regular handwriting word recognition approach.

  13. Time-Elastic Generative Model for Acceleration Time Series in Human Activity Recognition

    PubMed Central

    Munoz-Organero, Mario; Ruiz-Blazquez, Ramona

    2017-01-01

    Body-worn sensors in general and accelerometers in particular have been widely used in order to detect human movements and activities. The execution of each type of movement by each particular individual generates sequences of time series of sensed data from which specific movement related patterns can be assessed. Several machine learning algorithms have been used over windowed segments of sensed data in order to detect such patterns in activity recognition based on intermediate features (either hand-crafted or automatically learned from data). The underlying assumption is that the computed features will capture statistical differences that can properly classify different movements and activities after a training phase based on sensed data. In order to achieve high accuracy and recall rates (and guarantee the generalization of the system to new users), the training data have to contain enough information to characterize all possible ways of executing the activity or movement to be detected. This could imply large amounts of data and a complex and time-consuming training phase, which has been shown to be even more relevant when automatically learning the optimal features to be used. In this paper, we present a novel generative model that is able to generate sequences of time series for characterizing a particular movement based on the time elasticity properties of the sensed data. The model is used to train a stack of auto-encoders in order to learn the particular features able to detect human movements. The results of movement detection using a newly generated database with information on five users performing six different movements are presented. The generalization of results using an existing database is also presented in the paper. The results show that the proposed mechanism is able to obtain acceptable recognition rates (F = 0.77) even in the case of using different people executing a different sequence of movements and using different hardware. PMID

  14. Combining point context and dynamic time warping for online gesture recognition

    NASA Astrophysics Data System (ADS)

    Mao, Xia; Li, Chen

    2017-05-01

    Previous gesture recognition methods usually focused on recognizing gestures after the entire gesture sequences were obtained. However, in many practical applications, a system has to identify gestures before they end to give instant feedback. We present an online gesture recognition approach that can realize early recognition of unfinished gestures with low latency. First, a curvature buffer-based point context (CBPC) descriptor is proposed to extract the shape feature of a gesture trajectory. The CBPC descriptor is a complete descriptor with a simple computation, and thus has its superiority in online scenarios. Then, we introduce an online windowed dynamic time warping algorithm to realize online matching between the ongoing gesture and the template gestures. In the algorithm, computational complexity is effectively decreased by adding a sliding window to the accumulative distance matrix. Lastly, the experiments are conducted on the Australian sign language data set and the Kinect hand gesture (KHG) data set. Results show that the proposed method outperforms other state-of-the-art methods especially when gesture information is incomplete.

  15. Object Recognition in Flight: How Do Bees Distinguish between 3D Shapes?

    PubMed Central

    Werner, Annette; Stürzl, Wolfgang; Zanker, Johannes

    2016-01-01

    Honeybees (Apis mellifera) discriminate multiple object features such as colour, pattern and 2D shape, but it remains unknown whether and how bees recover three-dimensional shape. Here we show that bees can recognize objects by their three-dimensional form, whereby they employ an active strategy to uncover the depth profiles. We trained individual, free flying honeybees to collect sugar water from small three-dimensional objects made of styrofoam (sphere, cylinder, cuboids) or folded paper (convex, concave, planar) and found that bees can easily discriminate between these stimuli. We also tested possible strategies employed by the bees to uncover the depth profiles. For the card stimuli, we excluded overall shape and pictorial features (shading, texture gradients) as cues for discrimination. Lacking sufficient stereo vision, bees are known to use speed gradients in optic flow to detect edges; could the bees apply this strategy also to recover the fine details of a surface depth profile? Analysing the bees’ flight tracks in front of the stimuli revealed specific combinations of flight maneuvers (lateral translations in combination with yaw rotations), which are particularly suitable to extract depth cues from motion parallax. We modelled the generated optic flow and found characteristic patterns of angular displacement corresponding to the depth profiles of our stimuli: optic flow patterns from pure translations successfully recovered depth relations from the magnitude of angular displacements, additional rotation provided robust depth information based on the direction of the displacements; thus, the bees flight maneuvers may reflect an optimized visuo-motor strategy to extract depth structure from motion signals. The robustness and simplicity of this strategy offers an efficient solution for 3D-object-recognition without stereo vision, and could be employed by other flying insects, or mobile robots. PMID:26886006

  16. Object Recognition in Flight: How Do Bees Distinguish between 3D Shapes?

    PubMed

    Werner, Annette; Stürzl, Wolfgang; Zanker, Johannes

    2016-01-01

    Honeybees (Apis mellifera) discriminate multiple object features such as colour, pattern and 2D shape, but it remains unknown whether and how bees recover three-dimensional shape. Here we show that bees can recognize objects by their three-dimensional form, whereby they employ an active strategy to uncover the depth profiles. We trained individual, free flying honeybees to collect sugar water from small three-dimensional objects made of styrofoam (sphere, cylinder, cuboids) or folded paper (convex, concave, planar) and found that bees can easily discriminate between these stimuli. We also tested possible strategies employed by the bees to uncover the depth profiles. For the card stimuli, we excluded overall shape and pictorial features (shading, texture gradients) as cues for discrimination. Lacking sufficient stereo vision, bees are known to use speed gradients in optic flow to detect edges; could the bees apply this strategy also to recover the fine details of a surface depth profile? Analysing the bees' flight tracks in front of the stimuli revealed specific combinations of flight maneuvers (lateral translations in combination with yaw rotations), which are particularly suitable to extract depth cues from motion parallax. We modelled the generated optic flow and found characteristic patterns of angular displacement corresponding to the depth profiles of our stimuli: optic flow patterns from pure translations successfully recovered depth relations from the magnitude of angular displacements, additional rotation provided robust depth information based on the direction of the displacements; thus, the bees flight maneuvers may reflect an optimized visuo-motor strategy to extract depth structure from motion signals. The robustness and simplicity of this strategy offers an efficient solution for 3D-object-recognition without stereo vision, and could be employed by other flying insects, or mobile robots.

  17. Surviving blind decomposition: A distributional analysis of the time-course of complex word recognition.

    PubMed

    Schmidtke, Daniel; Matsuki, Kazunaga; Kuperman, Victor

    2017-11-01

    The current study addresses a discrepancy in the psycholinguistic literature about the chronology of information processing during the visual recognition of morphologically complex words. Form-then-meaning accounts of complex word recognition claim that morphemes are processed as units of form prior to any influence of their meanings, whereas form-and-meaning models posit that recognition of complex word forms involves the simultaneous access of morphological and semantic information. The study reported here addresses this theoretical discrepancy by applying a nonparametric distributional technique of survival analysis (Reingold & Sheridan, 2014) to 2 behavioral measures of complex word processing. Across 7 experiments reported here, this technique is employed to estimate the point in time at which orthographic, morphological, and semantic variables exert their earliest discernible influence on lexical decision RTs and eye movement fixation durations. Contrary to form-then-meaning predictions, Experiments 1-4 reveal that surface frequency is the earliest lexical variable to exert a demonstrable influence on lexical decision RTs for English and Dutch derived words (e.g., badness ; bad + ness ), English pseudoderived words (e.g., wander ; wand + er ) and morphologically simple control words (e.g., ballad ; ball + ad ). Furthermore, for derived word processing across lexical decision and eye-tracking paradigms (Experiments 1-2; 5-7), semantic effects emerge early in the time-course of word recognition, and their effects either precede or emerge simultaneously with morphological effects. These results are not consistent with the premises of the form-then-meaning view of complex word recognition, but are convergent with a form-and-meaning account of complex word recognition. (PsycINFO Database Record (c) 2017 APA, all rights reserved).

  18. Hippocampal Activation of Rac1 Regulates the Forgetting of Object Recognition Memory.

    PubMed

    Liu, Yunlong; Du, Shuwen; Lv, Li; Lei, Bo; Shi, Wei; Tang, Yikai; Wang, Lianzhang; Zhong, Yi

    2016-09-12

    Forgetting is a universal feature for most types of memories. The best-defined and extensively characterized behaviors that depict forgetting are natural memory decay and interference-based forgetting [1, 2]. Molecular mechanisms underlying the active forgetting remain to be determined for memories in vertebrates. Recent progress has begun to unravel such mechanisms underlying the active forgetting [3-11] that is induced through the behavior-dependent activation of intracellular signaling pathways. In Drosophila, training-induced activation of the small G protein Rac1 mediates natural memory decay and interference-based forgetting of aversive conditioning memory [3]. In mice, the activation of photoactivable-Rac1 in recently potentiated spines in a motor learning task erases the motor memory [12]. These lines of evidence prompted us to investigate a role for Rac1 in time-based natural memory decay and interference-based forgetting in mice. The inhibition of Rac1 activity in hippocampal neurons through targeted expression of a dominant-negative Rac1 form extended object recognition memory from less than 72 hr to over 72 hr, whereas Rac1 activation accelerated memory decay within 24 hr. Interference-induced forgetting of this memory was correlated with Rac1 activation and was completely blocked by inhibition of Rac1 activity. Electrophysiological recordings of long-term potentiation provided independent evidence that further supported a role for Rac1 activation in forgetting. Thus, Rac1-dependent forgetting is evolutionarily conserved from invertebrates to vertebrates. Copyright © 2016 Elsevier Ltd. All rights reserved.

  19. Working Memory Load Affects Processing Time in Spoken Word Recognition: Evidence from Eye-Movements

    PubMed Central

    Hadar, Britt; Skrzypek, Joshua E.; Wingfield, Arthur; Ben-David, Boaz M.

    2016-01-01

    In daily life, speech perception is usually accompanied by other tasks that tap into working memory capacity. However, the role of working memory on speech processing is not clear. The goal of this study was to examine how working memory load affects the timeline for spoken word recognition in ideal listening conditions. We used the “visual world” eye-tracking paradigm. The task consisted of spoken instructions referring to one of four objects depicted on a computer monitor (e.g., “point at the candle”). Half of the trials presented a phonological competitor to the target word that either overlapped in the initial syllable (onset) or at the last syllable (offset). Eye movements captured listeners' ability to differentiate the target noun from its depicted phonological competitor (e.g., candy or sandal). We manipulated working memory load by using a digit pre-load task, where participants had to retain either one (low-load) or four (high-load) spoken digits for the duration of a spoken word recognition trial. The data show that the high-load condition delayed real-time target discrimination. Specifically, a four-digit load was sufficient to delay the point of discrimination between the spoken target word and its phonological competitor. Our results emphasize the important role working memory plays in speech perception, even when performed by young adults in ideal listening conditions. PMID:27242424

  20. Newborn chickens generate invariant object representations at the onset of visual object experience

    PubMed Central

    Wood, Justin N.

    2013-01-01

    To recognize objects quickly and accurately, mature visual systems build invariant object representations that generalize across a range of novel viewing conditions (e.g., changes in viewpoint). To date, however, the origins of this core cognitive ability have not yet been established. To examine how invariant object recognition develops in a newborn visual system, I raised chickens from birth for 2 weeks within controlled-rearing chambers. These chambers provided complete control over all visual object experiences. In the first week of life, subjects’ visual object experience was limited to a single virtual object rotating through a 60° viewpoint range. In the second week of life, I examined whether subjects could recognize that virtual object from novel viewpoints. Newborn chickens were able to generate viewpoint-invariant representations that supported object recognition across large, novel, and complex changes in the object’s appearance. Thus, newborn visual systems can begin building invariant object representations at the onset of visual object experience. These abstract representations can be generated from sparse data, in this case from a visual world containing a single virtual object seen from a limited range of viewpoints. This study shows that powerful, robust, and invariant object recognition machinery is an inherent feature of the newborn brain. PMID:23918372

  1. Episodic Short-Term Recognition Requires Encoding into Visual Working Memory: Evidence from Probe Recognition after Letter Report

    PubMed Central

    Poth, Christian H.; Schneider, Werner X.

    2016-01-01

    Human vision is organized in discrete processing episodes (e.g., eye fixations or task-steps). Object information must be transmitted across episodes to enable episodic short-term recognition: recognizing whether a current object has been seen in a previous episode. We ask whether episodic short-term recognition presupposes that objects have been encoded into capacity-limited visual working memory (VWM), which retains visual information for report. Alternatively, it could rely on the activation of visual features or categories that occurs before encoding into VWM. We assessed the dependence of episodic short-term recognition on VWM by a new paradigm combining letter report and probe recognition. Participants viewed displays of 10 letters and reported as many as possible after a retention interval (whole report). Next, participants viewed a probe letter and indicated whether it had been one of the 10 letters (probe recognition). In Experiment 1, probe recognition was more accurate for letters that had been encoded into VWM (reported letters) compared with non-encoded letters (non-reported letters). Interestingly, those letters that participants reported in their whole report had been near to one another within the letter displays. This suggests that the encoding into VWM proceeded in a spatially clustered manner. In Experiment 2, participants reported only one of 10 letters (partial report) and probes either referred to this letter, to letters that had been near to it, or far from it. Probe recognition was more accurate for near than for far letters, although none of these letters had to be reported. These findings indicate that episodic short-term recognition is constrained to a small number of simultaneously presented objects that have been encoded into VWM. PMID:27713722

  2. Episodic Short-Term Recognition Requires Encoding into Visual Working Memory: Evidence from Probe Recognition after Letter Report.

    PubMed

    Poth, Christian H; Schneider, Werner X

    2016-01-01

    Human vision is organized in discrete processing episodes (e.g., eye fixations or task-steps). Object information must be transmitted across episodes to enable episodic short-term recognition: recognizing whether a current object has been seen in a previous episode. We ask whether episodic short-term recognition presupposes that objects have been encoded into capacity-limited visual working memory (VWM), which retains visual information for report. Alternatively, it could rely on the activation of visual features or categories that occurs before encoding into VWM. We assessed the dependence of episodic short-term recognition on VWM by a new paradigm combining letter report and probe recognition. Participants viewed displays of 10 letters and reported as many as possible after a retention interval (whole report). Next, participants viewed a probe letter and indicated whether it had been one of the 10 letters (probe recognition). In Experiment 1, probe recognition was more accurate for letters that had been encoded into VWM (reported letters) compared with non-encoded letters (non-reported letters). Interestingly, those letters that participants reported in their whole report had been near to one another within the letter displays. This suggests that the encoding into VWM proceeded in a spatially clustered manner. In Experiment 2, participants reported only one of 10 letters (partial report) and probes either referred to this letter, to letters that had been near to it, or far from it. Probe recognition was more accurate for near than for far letters, although none of these letters had to be reported. These findings indicate that episodic short-term recognition is constrained to a small number of simultaneously presented objects that have been encoded into VWM.

  3. Acute fasting inhibits central caspase-1 activity reducing anxiety-like behavior and increasing novel object and object location recognition.

    PubMed

    Towers, Albert E; Oelschlager, Maci L; Patel, Jay; Gainey, Stephen J; McCusker, Robert H; Freund, Gregory G

    2017-06-01

    Inflammation within the central nervous system (CNS) is frequently comorbid with anxiety. Importantly, the pro-inflammatory cytokine most commonly associated with anxiety is IL-1β. The bioavailability and activity of IL-1β are regulated by caspase-1-dependent proteolysis vis-a-vis the inflammasome. Thus, interventions regulating the activation or activity of caspase-1 should reduce anxiety especially in states that foster IL-1β maturation. Male C57BL/6j, C57BL/6j mice treated with the capase-1 inhibitor biotin-YVAD-cmk, caspase-1 knockout (KO) mice and IL-1R1 KO mice were fasted for 24h or allowed ad libitum access to food. Immediately after fasting, caspase-1 activity was measured in brain region homogenates while activated caspase-1 was localized in the brain by immunohistochemistry. Mouse anxiety-like behavior and cognition were tested using the elevated zero maze and novel object/object location tasks, respectively. A 24h fast in mice reduced the activity of caspase-1 in whole brain and in the prefrontal cortex, amygdala, hippocampus, and hypothalamus by 35%, 25%, 40%, 40%, and 40% respectively. A 24h fast also reduced anxiety-like behavior by 40% and increased novel object and object location recognition by 21% and 31%, respectively. IL-1β protein, however, was not reduced in the brain by fasting. ICV administration of YVAD decreased caspase-1 activity in the prefrontal cortex and amygdala by 55%, respectively leading to a 64% reduction in anxiety like behavior. Importantly, when caspase-1 KO or IL1-R1 KO mice are fasted, no fasting-dependent reduction in anxiety-like behavior was observed. Results indicate that fasting decrease anxiety-like behavior and improves memory by a mechanism tied to reducing caspase-1 activity throughout the brain. Copyright © 2017 Elsevier Inc. All rights reserved.

  4. Acute fasting inhibits central caspase-1 activity reducing anxiety-like behavior and increasing novel object and object location recognition

    PubMed Central

    Towers, Albert E; Oelschlager, Maci L.; Patel, Jay; Gainey, Stephen J.; McCusker, Robert; Freund, Gregory G.

    2017-01-01

    Background Inflammation within the central nervous system (CNS) is frequently comorbid with anxiety. Importantly, the pro-inflammatory cytokine most commonly associated with anxiety is IL-1β. The bioavailability and activity of IL-1β is regulated by caspase-1-dependent proteolysis vis-a-vis the inflammasome. Thus, interventions regulating the activation or activity of caspase-1 should reduce anxiety especially in states that foster IL-1β maturation. Methods Male C57BL/6j, C57BL/6j mice treated with the capase-1 inhibitor biotin-YVAD-cmk, caspase-1 knockout (KO) mice and IL-1R1 KO mice were fasted for 24 hours or allowed ad libitum access to food. Immediately after fasting, caspase-1 activity was measured in brain region homogenates while activated caspase-1 was localized in the brain by immunohistochemistry. Mouse anxiety-like behavior and cognition were tested using the elevated zero maze and novel object/object location tasks, respectively. Results A 24 h fast in mice reduced the activity of caspase-1 in whole brain and in the prefrontal cortex, amygdala, hippocampus, and hypothalamus by 35%, 25%, 40%, 40%, and 40% respectively. A 24 h fast also reduced anxiety-like behavior by 40% and increased novel object and object location recognition by 21% and 31%, respectively. IL-1β protein, however, was not reduced in the brain by fasting. ICV administration of YVAD decreased caspase-1 activity in the prefrontal cortex and amygdala by 55%, respectively leading to a 64% reduction in anxiety like behavior. Importantly, when caspase-1 KO or IL1-R1 KO mice are fasted, no fasting-dependent reduction in anxiety-like behavior was observed. Conclusions Results indicate that fasting decrease anxiety-like behavior and improves memory by a mechanism tied to reducing caspase-1 activity throughout the brain. PMID:28521881

  5. Kisspeptin-13 enhances memory and mitigates memory impairment induced by Aβ1-42 in mice novel object and object location recognition tasks.

    PubMed

    Jiang, J H; He, Z; Peng, Y L; Jin, W D; Wang, Z; Han, R W; Chang, M; Wang, R

    2015-09-01

    Kisspeptin (KP), the endogenous ligand of GPR54, is a recently discovered neuropeptide shown to be involved in regulating reproductive system, anxiety-related behavior, locomotion, food intake, and suppression of metastasis across a range of cancers. KP is transcribed within the hippocampus, and GPR54 has been found in the amygdala and hippocampus, suggesting that KP might be involved in mediating learning and memory. However, the role of KP in cognition was largely unclear. Here, we investigated the role of KP-13, one of the endogenous active isoforms, in memory processes, and determined whether KP-13 could mitigate memory impairment induced by Aβ1-42 in mice, using novel object recognition (NOR) and object location recognition (OLR) tasks. Intracerebroventricular (i.c.v.) infusion of KP-13 (2μg) immediately after training not only facilitated memory formation, but also prolonged memory retention in both tasks. The memory-improving effects of KP-13 could be blocked by the GPR54 receptor antagonist, kisspeptin-234 (234), and GnRH receptors antagonist, Cetrorelix, suggesting pharmacological specificity. Then the memory-enhancing effects were also presented after infusion of KP-13 into the hippocampus. Moreover, we found that i.c.v. injection of KP-13 was able to reverse the memory impairment induced by Aβ1-42, which was inhibited by 234. To sum up, the results of our work indicate that KP-13 could facilitate memory formation and prolong memory retention through activation of the GPR54 and GnRH receptors, and suppress memory-impairing effect of Aβ1-42 through activation of the GPR54, suggesting that KP-13 may be a potential drug for enhancing memory and treating Alzheimer's disease. Copyright © 2015 Elsevier Inc. All rights reserved.

  6. The writer independent online handwriting recognition system frog on hand and cluster generative statistical dynamic time warping.

    PubMed

    Bahlmann, Claus; Burkhardt, Hans

    2004-03-01

    In this paper, we give a comprehensive description of our writer-independent online handwriting recognition system frog on hand. The focus of this work concerns the presentation of the classification/training approach, which we call cluster generative statistical dynamic time warping (CSDTW). CSDTW is a general, scalable, HMM-based method for variable-sized, sequential data that holistically combines cluster analysis and statistical sequence modeling. It can handle general classification problems that rely on this sequential type of data, e.g., speech recognition, genome processing, robotics, etc. Contrary to previous attempts, clustering and statistical sequence modeling are embedded in a single feature space and use a closely related distance measure. We show character recognition experiments of frog on hand using CSDTW on the UNIPEN online handwriting database. The recognition accuracy is significantly higher than reported results of other handwriting recognition systems. Finally, we describe the real-time implementation of frog on hand on a Linux Compaq iPAQ embedded device.

  7. Shape and texture fused recognition of flying targets

    NASA Astrophysics Data System (ADS)

    Kovács, Levente; Utasi, Ákos; Kovács, Andrea; Szirányi, Tamás

    2011-06-01

    This paper presents visual detection and recognition of flying targets (e.g. planes, missiles) based on automatically extracted shape and object texture information, for application areas like alerting, recognition and tracking. Targets are extracted based on robust background modeling and a novel contour extraction approach, and object recognition is done by comparisons to shape and texture based query results on a previously gathered real life object dataset. Application areas involve passive defense scenarios, including automatic object detection and tracking with cheap commodity hardware components (CPU, camera and GPS).

  8. Online graphic symbol recognition using neural network and ARG matching

    NASA Astrophysics Data System (ADS)

    Yang, Bing; Li, Changhua; Xie, Weixing

    2001-09-01

    This paper proposes a novel method for on-line recognition of line-based graphic symbol. The input strokes are usually warped into a cursive form due to the sundry drawing style, and classifying them is very difficult. To deal with this, an ART-2 neural network is used to classify the input strokes. It has the advantages of high recognition rate, less recognition time and forming classes in a self-organized manner. The symbol recognition is achieved by an Attribute Relational Graph (ARG) matching algorithm. The ARG is very efficient for representing complex objects, but computation cost is very high. To over come this, we suggest a fast graph matching algorithm using symbol structure information. The experimental results show that the proposed method is effective for recognition of symbols with hierarchical structure.

  9. Stereo vision with distance and gradient recognition

    NASA Astrophysics Data System (ADS)

    Kim, Soo-Hyun; Kang, Suk-Bum; Yang, Tae-Kyu

    2007-12-01

    Robot vision technology is needed for the stable walking, object recognition and the movement to the target spot. By some sensors which use infrared rays and ultrasonic, robot can overcome the urgent state or dangerous time. But stereo vision of three dimensional space would make robot have powerful artificial intelligence. In this paper we consider about the stereo vision for stable and correct movement of a biped robot. When a robot confront with an inclination plane or steps, particular algorithms are needed to go on without failure. This study developed the recognition algorithm of distance and gradient of environment by stereo matching process.

  10. Poka Yoke system based on image analysis and object recognition

    NASA Astrophysics Data System (ADS)

    Belu, N.; Ionescu, L. M.; Misztal, A.; Mazăre, A.

    2015-11-01

    Poka Yoke is a method of quality management which is related to prevent faults from arising during production processes. It deals with “fail-sating” or “mistake-proofing”. The Poka-yoke concept was generated and developed by Shigeo Shingo for the Toyota Production System. Poka Yoke is used in many fields, especially in monitoring production processes. In many cases, identifying faults in a production process involves a higher cost than necessary cost of disposal. Usually, poke yoke solutions are based on multiple sensors that identify some nonconformities. This means the presence of different equipment (mechanical, electronic) on production line. As a consequence, coupled with the fact that the method itself is an invasive, affecting the production process, would increase its price diagnostics. The bulky machines are the means by which a Poka Yoke system can be implemented become more sophisticated. In this paper we propose a solution for the Poka Yoke system based on image analysis and identification of faults. The solution consists of a module for image acquisition, mid-level processing and an object recognition module using associative memory (Hopfield network type). All are integrated into an embedded system with AD (Analog to Digital) converter and Zync 7000 (22 nm technology).

  11. Intarsia-sensorized band and textrodes for real-time myoelectric pattern recognition.

    PubMed

    Brown, Shannon; Ortiz-Catalan, Max; Petersson, Joel; Rodby, Kristian; Seoane, Fernando

    2016-08-01

    Surface Electromyography (sEMG) has applications in prosthetics, diagnostics and neuromuscular rehabilitation. Self-adhesive Ag/AgCl are the electrodes preferentially used to capture sEMG in short-term studies, however their long-term application is limited. In this study we designed and evaluated a fully integrated smart textile band with electrical connecting tracks knitted with intarsia techniques and knitted textile electrodes. Real-time myoelectric pattern recognition for motor volition and signal-to-noise ratio (SNR) were used to compare its sensing performance versus the conventional Ag-AgCl electrodes. After a comprehending measurement and performance comparison of the sEMG recordings, no significant differences were found between the textile and the Ag-AgCl electrodes in SNR and prediction accuracy obtained from pattern recognition classifiers.

  12. Object recognition through turbulence with a modified plenoptic camera

    NASA Astrophysics Data System (ADS)

    Wu, Chensheng; Ko, Jonathan; Davis, Christopher

    2015-03-01

    Atmospheric turbulence adds accumulated distortion to images obtained by cameras and surveillance systems. When the turbulence grows stronger or when the object is further away from the observer, increasing the recording device resolution helps little to improve the quality of the image. Many sophisticated methods to correct the distorted images have been invented, such as using a known feature on or near the target object to perform a deconvolution process, or use of adaptive optics. However, most of the methods depend heavily on the object's location, and optical ray propagation through the turbulence is not directly considered. Alternatively, selecting a lucky image over many frames provides a feasible solution, but at the cost of time. In our work, we propose an innovative approach to improving image quality through turbulence by making use of a modified plenoptic camera. This type of camera adds a micro-lens array to a traditional high-resolution camera to form a semi-camera array that records duplicate copies of the object as well as "superimposed" turbulence at slightly different angles. By performing several steps of image reconstruction, turbulence effects will be suppressed to reveal more details of the object independently (without finding references near the object). Meanwhile, the redundant information obtained by the plenoptic camera raises the possibility of performing lucky image algorithmic analysis with fewer frames, which is more efficient. In our work, the details of our modified plenoptic cameras and image processing algorithms will be introduced. The proposed method can be applied to coherently illuminated object as well as incoherently illuminated objects. Our result shows that the turbulence effect can be effectively suppressed by the plenoptic camera in the hardware layer and a reconstructed "lucky image" can help the viewer identify the object even when a "lucky image" by ordinary cameras is not achievable.

  13. Ciproxifan, an H3 receptor antagonist, improves short-term recognition memory impaired by isoflurane anesthesia.

    PubMed

    Ding, Fang; Zheng, Limin; Liu, Min; Chen, Rongfa; Leung, L Stan; Luo, Tao

    2016-08-01

    Exposure to volatile anesthetics has been reported to cause temporary or sustained impairments in learning and memory in pre-clinical studies. The selective antagonists of the histamine H3 receptors (H3R) are considered to be a promising group of novel therapeutic agents for the treatment of cognitive disorders. The aim of this study was to evaluate the effect of H3R antagonist ciproxifan on isoflurane-induced deficits in an object recognition task. Adult C57BL/6 J mice were exposed to isoflurane (1.3 %) or vehicle gas for 2 h. The object recognition tests were carried at 24 h or 7 days after exposure to anesthesia to exploit the tendency of mice to prefer exploring novel objects in an environment when a familiar object is also present. During the training phase, two identical objects were placed in two defined sites of the chamber. During the test phase, performed 1 or 24 h after the training phase, one of the objects was replaced by a new object with a different shape. The time spent exploring each object was recorded. A robust deficit in object recognition memory occurred 1 day after exposure to isoflurane anesthesia. Isoflurane-treated mice spent significantly less time exploring a novel object at 1 h but not at 24 h after the training phase. The deficit in short-term memory was reversed by the administration of ciproxifan 30 min before behavioral training. Isoflurane exposure induces reversible deficits in object recognition memory. Ciproxifan appears to be a potential therapeutic agent for improving post-anesthesia cognitive memory performance.

  14. Intelligent data processing of an ultrasonic sensor system for pattern recognition improvements

    NASA Astrophysics Data System (ADS)

    Na, Seung You; Park, Min-Sang; Hwang, Won-Gul; Kee, Chang-Doo

    1999-05-01

    Though conventional time-of-flight ultrasonic sensor systems are popular due to the advantages of low cost and simplicity, the usage of the sensors is rather narrowly restricted within object detection and distance readings. There is a strong need to enlarge the amount of environmental information for mobile applications to provide intelligent autonomy. Wide sectors of such neighboring object recognition problems can be satisfactorily handled with coarse vision data such as sonar maps instead of accurate laser or optic measurements. For the usage of object pattern recognition, ultrasonic senors have inherent shortcomings of poor directionality and specularity which result in low spatial resolution and indistinctiveness of object patterns. To resolve these problems an array of increased number of sensor elements has been used for large objects. In this paper we propose a method of sensor array system with improved recognition capability using electronic circuits accompanying the sensor array and neuro-fuzzy processing of data fusion. The circuit changes transmitter output voltages of array elements in several steps. Relying upon the known sensor characteristics, a set of different return signals from neighboring senors is manipulated to provide an enhanced pattern recognition in the aspects of inclination angle, size and shift as well as distance of objects. The results show improved resolution of the measurements for smaller targets.

  15. Recognition of disturbances with specified morphology in time series. Part 1: Spikes on magnetograms of the worldwide INTERMAGNET network

    NASA Astrophysics Data System (ADS)

    Bogoutdinov, Sh. R.; Gvishiani, A. D.; Agayan, S. M.; Solovyev, A. A.; Kin, E.

    2010-11-01

    The International Real-time Magnetic Observatory Network (INTERMAGNET) is the world's biggest international network of ground-based observatories, providing geomagnetic data almost in real time (within 72 hours of collection) [Kerridge, 2001]. The observation data are rapidly transferred by the observatories participating in the program to regional Geomagnetic Information Nodes (GINs), which carry out a global exchange of data and process the results. The observations of the main (core) magnetic field of the Earth and its study are one of the key problems of geophysics. The INTERMAGNET system is the basis of monitoring the state of the Earth's magnetic field; therefore, the information provided by the system is required to be very reliable. Despite the rigid high-quality standard of the recording devices, they are subject to external effects that affect the quality of the records. Therefore, an objective and formalized recognition with the subsequent remedy of the anomalies (artifacts) that occur on the records is an important task. Expanding on the ideas of Agayan [Agayan et al., 2005] and Gvishiani [Gvishiani et al., 2008a; 2008b], this paper suggests a new algorithm of automatic recognition of anomalies with specified morphology, capable of identifying both physically- and anthropogenically-derived spikes on the magnetograms. The algorithm is constructed using fuzzy logic and, as such, is highly adaptive and universal. The developed algorithmic system formalizes the work of the expert-interpreter in terms of artificial intelligence. This ensures identical processing of large data arrays, almost unattainable manually. Besides the algorithm, the paper also reports on the application of the developed algorithmic system for identifying spikes at the INTERMAGNET observatories. The main achievement of the work is the creation of an algorithm permitting the almost unmanned extraction of spike-free (definitive) magnetograms from preliminary records. This automated

  16. Willughbeia cochinchinensis prevents scopolamine-induced deficits in memory, spatial learning, and object recognition in rodents.

    PubMed

    Can, Mao Van; Tran, Anh Hai; Pham, Dam Minh; Dinh, Bao Quoc; Le, Quan Van; Nguyen, Ba Van; Nguyen, Mai Thanh Thi; Nguyen, Hai Xuan; Nguyen, Nhan Trung; Nishijo, Hisao

    2018-03-25

    Willughbeia cochinchinensis (WC) has been used in Vietnamese traditional medicine for the treatment of dementia as well as diarrhea, heartburn, and cutaneous abscess and as a diuretic. Alzheimer's disease (AD) is one of the most prevalent diseases in elderly individuals. Acetylcholinesterase (AChE) and butyrylcholinesterase (BChE) inhibitors have been widely used to treat patients with AD. In the present study, we investigated anti-AChE and anti-BChE activities of a natural product, WC, for its potential applications in therapies to prevent/treat dementia. First, compounds extracted from WC were tested for their AChE and BChE inhibitory activities in vitro. Second, in vivo behavioral experiments were performed to investigate the effects of WC at doses of 100, 150, and 200mg/kg on scopolamine (1.5mg/kg)-induced memory and cognitive deficits in mice. The behavior of mice treated with and without WC and/or scopolamine was tested using the Y-maze, Morris water maze, and novel object recognition task. The results of the in vitro assay demonstrated anti-AChE and anti-BChE activities of the compounds extracted from WC. The results of behavioral experiments showed that the administration of WC prevented 1) scopolamine-induced decrease in spontaneous alternation (%) behavior in the Y-maze, 2) scopolamine-induced deficits in spatial learning and memory in the Morris water maze, and 3) scopolamine-induced deficits in novel object recognition. These results indicate that WC prevents cognitive and memory deficits induced by scopolamine injection. Our findings suggest that WC may represent a novel candidate for the treatment of memory and cognitive deficits in humans with dementia. Copyright © 2017. Published by Elsevier B.V.

  17. Disconnection of the Perirhinal and Postrhinal Cortices Impairs Recognition of Objects in Context But Not Contextual Fear Conditioning.

    PubMed

    Heimer-McGinn, Victoria R; Poeta, Devon L; Aghi, Krishan; Udawatta, Methma; Burwell, Rebecca D

    2017-05-03

    The perirhinal cortex (PER) is known to process object information, whereas the rodent postrhinal cortex (POR), homolog to the parahippocampal cortex in primates, is thought to process spatial information. A number of studies, however, provide evidence that both areas are involved in processing contextual information. In this study, we tested the hypothesis that the rat POR relies on object information received from the PER to form complex representations of context. Using three fear-conditioning (FC) paradigms (signaled, unsignaled, and renewal) and two context-guided object recognition tasks (with 3D and 2D objects), we examined the effects of crossed excitotoxic lesions to the POR and the contralateral PER. Performance of rats with crossed lesions was compared with that of rats with ipsilateral POR plus PER lesions and sham-operated rats. We found that rats with contralateral PER-POR lesions were impaired in object-context recognition but not in contextual FC. Therefore, interaction between the POR and PER is necessary for context-guided exploratory behavior but not for associating fear with context. Our results provide evidence for the hypothesis that the POR relies on object and pattern information from the PER to encode representations of context. The association of fear with a context, however, may be supported by alternate cortical and/or subcortical pathways when PER-POR interaction is not available. Our results suggest that contextual FC may represent a special case of context-guided behavior. SIGNIFICANCE STATEMENT Representations of context are important for perception, memory, decision making, and other cognitive processes. Moreover, there is extensive evidence that the use of contextual representations to guide appropriate behavior is disrupted in neuropsychiatric and neurological disorders including developmental disorders, schizophrenia, affective disorders, and Alzheimer's disease. Many of these disorders are accompanied by changes in

  18. Eye-movement assessment of the time course in facial expression recognition: Neurophysiological implications.

    PubMed

    Calvo, Manuel G; Nummenmaa, Lauri

    2009-12-01

    Happy, surprised, disgusted, angry, sad, fearful, and neutral faces were presented extrafoveally, with fixations on faces allowed or not. The faces were preceded by a cue word that designated the face to be saccaded in a two-alternative forced-choice discrimination task (2AFC; Experiments 1 and 2), or were followed by a probe word for recognition (Experiment 3). Eye tracking was used to decompose the recognition process into stages. Relative to the other expressions, happy faces (1) were identified faster (as early as 160 msec from stimulus onset) in extrafoveal vision, as revealed by shorter saccade latencies in the 2AFC task; (2) required less encoding effort, as indexed by shorter first fixations and dwell times; and (3) required less decision-making effort, as indicated by fewer refixations on the face after the recognition probe was presented. This reveals a happy-face identification advantage both prior to and during overt attentional processing. The results are discussed in relation to prior neurophysiological findings on latencies in facial expression recognition.

  19. Study and response time for the visual recognition of 'similarity' and identity

    NASA Technical Reports Server (NTRS)

    Derks, P. L.; Bauer, T. M.

    1974-01-01

    Four subjects compared successively presented pairs of line patterns for a match between any lines in the pattern (similarity) and for a match between all lines (identity). The encoding or study times for pattern recognition from immediate memory and the latency in responses to comparison stimuli were examined. Qualitative differences within and between subjects were most evident in study times.

  20. A method of 3D object recognition and localization in a cloud of points

    NASA Astrophysics Data System (ADS)

    Bielicki, Jerzy; Sitnik, Robert

    2013-12-01

    The proposed method given in this article is prepared for analysis of data in the form of cloud of points directly from 3D measurements. It is designed for use in the end-user applications that can directly be integrated with 3D scanning software. The method utilizes locally calculated feature vectors (FVs) in point cloud data. Recognition is based on comparison of the analyzed scene with reference object library. A global descriptor in the form of a set of spatially distributed FVs is created for each reference model. During the detection process, correlation of subsets of reference FVs with FVs calculated in the scene is computed. Features utilized in the algorithm are based on parameters, which qualitatively estimate mean and Gaussian curvatures. Replacement of differentiation with averaging in the curvatures estimation makes the algorithm more resistant to discontinuities and poor quality of the input data. Utilization of the FV subsets allows to detect partially occluded and cluttered objects in the scene, while additional spatial information maintains false positive rate at a reasonably low level.