A framework for the recognition of 3D faces and expressions
NASA Astrophysics Data System (ADS)
Li, Chao; Barreto, Armando
2006-04-01
Face recognition technology has been a focus both in academia and industry for the last couple of years because of its wide potential applications and its importance to meet the security needs of today's world. Most of the systems developed are based on 2D face recognition technology, which uses pictures for data processing. With the development of 3D imaging technology, 3D face recognition emerges as an alternative to overcome the difficulties inherent with 2D face recognition, i.e. sensitivity to illumination conditions and orientation positioning of the subject. But 3D face recognition still needs to tackle the problem of deformation of facial geometry that results from the expression changes of a subject. To deal with this issue, a 3D face recognition framework is proposed in this paper. It is composed of three subsystems: an expression recognition system, a system for the identification of faces with expression, and neutral face recognition system. A system for the recognition of faces with one type of expression (happiness) and neutral faces was implemented and tested on a database of 30 subjects. The results proved the feasibility of this framework.
Online handwritten mathematical expression recognition
NASA Astrophysics Data System (ADS)
Büyükbayrak, Hakan; Yanikoglu, Berrin; Erçil, Aytül
2007-01-01
We describe a system for recognizing online, handwritten mathematical expressions. The system is designed with a user-interface for writing scientific articles, supporting the recognition of basic mathematical expressions as well as integrals, summations, matrices etc. A feed-forward neural network recognizes symbols which are assumed to be single-stroke and a recursive algorithm parses the expression by combining neural network output and the structure of the expression. Preliminary results show that writer-dependent recognition rates are very high (99.8%) while writer-independent symbol recognition rates are lower (75%). The interface associated with the proposed system integrates the built-in recognition capabilities of the Microsoft's Tablet PC API for recognizing textual input and supports conversion of hand-drawn figures into PNG format. This enables the user to enter text, mathematics and draw figures in a single interface. After recognition, all output is combined into one LATEX code and compiled into a PDF file.
iFER: facial expression recognition using automatically selected geometric eye and eyebrow features
NASA Astrophysics Data System (ADS)
Oztel, Ismail; Yolcu, Gozde; Oz, Cemil; Kazan, Serap; Bunyak, Filiz
2018-03-01
Facial expressions have an important role in interpersonal communications and estimation of emotional states or intentions. Automatic recognition of facial expressions has led to many practical applications and became one of the important topics in computer vision. We present a facial expression recognition system that relies on geometry-based features extracted from eye and eyebrow regions of the face. The proposed system detects keypoints on frontal face images and forms a feature set using geometric relationships among groups of detected keypoints. Obtained feature set is refined and reduced using the sequential forward selection (SFS) algorithm and fed to a support vector machine classifier to recognize five facial expression classes. The proposed system, iFER (eye-eyebrow only facial expression recognition), is robust to lower face occlusions that may be caused by beards, mustaches, scarves, etc. and lower face motion during speech production. Preliminary experiments on benchmark datasets produced promising results outperforming previous facial expression recognition studies using partial face features, and comparable results to studies using whole face information, only slightly lower by ˜ 2.5 % compared to the best whole face facial recognition system while using only ˜ 1 / 3 of the facial region.
NASA Astrophysics Data System (ADS)
Petpairote, Chayanut; Madarasmi, Suthep; Chamnongthai, Kosin
2018-01-01
The practical identification of individuals using facial recognition techniques requires the matching of faces with specific expressions to faces from a neutral face database. A method for facial recognition under varied expressions against neutral face samples of individuals via recognition of expression warping and the use of a virtual expression-face database is proposed. In this method, facial expressions are recognized and the input expression faces are classified into facial expression groups. To aid facial recognition, the virtual expression-face database is sorted into average facial-expression shapes and by coarse- and fine-featured facial textures. Wrinkle information is also employed in classification by using a process of masking to adjust input faces to match the expression-face database. We evaluate the performance of the proposed method using the CMU multi-PIE, Cohn-Kanade, and AR expression-face databases, and we find that it provides significantly improved results in terms of face recognition accuracy compared to conventional methods and is acceptable for facial recognition under expression variation.
Dynamic facial expression recognition based on geometric and texture features
NASA Astrophysics Data System (ADS)
Li, Ming; Wang, Zengfu
2018-04-01
Recently, dynamic facial expression recognition in videos has attracted growing attention. In this paper, we propose a novel dynamic facial expression recognition method by using geometric and texture features. In our system, the facial landmark movements and texture variations upon pairwise images are used to perform the dynamic facial expression recognition tasks. For one facial expression sequence, pairwise images are created between the first frame and each of its subsequent frames. Integration of both geometric and texture features further enhances the representation of the facial expressions. Finally, Support Vector Machine is used for facial expression recognition. Experiments conducted on the extended Cohn-Kanade database show that our proposed method can achieve a competitive performance with other methods.
Support vector machine for automatic pain recognition
NASA Astrophysics Data System (ADS)
Monwar, Md Maruf; Rezaei, Siamak
2009-02-01
Facial expressions are a key index of emotion and the interpretation of such expressions of emotion is critical to everyday social functioning. In this paper, we present an efficient video analysis technique for recognition of a specific expression, pain, from human faces. We employ an automatic face detector which detects face from the stored video frame using skin color modeling technique. For pain recognition, location and shape features of the detected faces are computed. These features are then used as inputs to a support vector machine (SVM) for classification. We compare the results with neural network based and eigenimage based automatic pain recognition systems. The experiment results indicate that using support vector machine as classifier can certainly improve the performance of automatic pain recognition system.
Facial recognition in education system
NASA Astrophysics Data System (ADS)
Krithika, L. B.; Venkatesh, K.; Rathore, S.; Kumar, M. Harish
2017-11-01
Human beings exploit emotions comprehensively for conveying messages and their resolution. Emotion detection and face recognition can provide an interface between the individuals and technologies. The most successful applications of recognition analysis are recognition of faces. Many different techniques have been used to recognize the facial expressions and emotion detection handle varying poses. In this paper, we approach an efficient method to recognize the facial expressions to track face points and distances. This can automatically identify observer face movements and face expression in image. This can capture different aspects of emotion and facial expressions.
Bologna, Matteo; Berardelli, Isabella; Paparella, Giulia; Marsili, Luca; Ricciardi, Lucia; Fabbrini, Giovanni; Berardelli, Alfredo
2016-01-01
Altered emotional processing, including reduced emotion facial expression and defective emotion recognition, has been reported in patients with Parkinson's disease (PD). However, few studies have objectively investigated facial expression abnormalities in PD using neurophysiological techniques. It is not known whether altered facial expression and recognition in PD are related. To investigate possible deficits in facial emotion expression and emotion recognition and their relationship, if any, in patients with PD. Eighteen patients with PD and 16 healthy controls were enrolled in this study. Facial expressions of emotion were recorded using a 3D optoelectronic system and analyzed using the facial action coding system. Possible deficits in emotion recognition were assessed using the Ekman test. Participants were assessed in one experimental session. Possible relationship between the kinematic variables of facial emotion expression, the Ekman test scores, and clinical and demographic data in patients were evaluated using the Spearman's test and multiple regression analysis. The facial expression of all six basic emotions had slower velocity and lower amplitude in patients in comparison to healthy controls (all P s < 0.05). Patients also yielded worse Ekman global score and disgust, sadness, and fear sub-scores than healthy controls (all P s < 0.001). Altered facial expression kinematics and emotion recognition deficits were unrelated in patients (all P s > 0.05). Finally, no relationship emerged between kinematic variables of facial emotion expression, the Ekman test scores, and clinical and demographic data in patients (all P s > 0.05). The results in this study provide further evidence of altered emotional processing in PD. The lack of any correlation between altered facial emotion expression kinematics and emotion recognition deficits in patients suggests that these abnormalities are mediated by separate pathophysiological mechanisms.
Multi-layer sparse representation for weighted LBP-patches based facial expression recognition.
Jia, Qi; Gao, Xinkai; Guo, He; Luo, Zhongxuan; Wang, Yi
2015-03-19
In this paper, a novel facial expression recognition method based on sparse representation is proposed. Most contemporary facial expression recognition systems suffer from limited ability to handle image nuisances such as low resolution and noise. Especially for low intensity expression, most of the existing training methods have quite low recognition rates. Motivated by sparse representation, the problem can be solved by finding sparse coefficients of the test image by the whole training set. Deriving an effective facial representation from original face images is a vital step for successful facial expression recognition. We evaluate facial representation based on weighted local binary patterns, and Fisher separation criterion is used to calculate the weighs of patches. A multi-layer sparse representation framework is proposed for multi-intensity facial expression recognition, especially for low-intensity expressions and noisy expressions in reality, which is a critical problem but seldom addressed in the existing works. To this end, several experiments based on low-resolution and multi-intensity expressions are carried out. Promising results on publicly available databases demonstrate the potential of the proposed approach.
Support vector machine-based facial-expression recognition method combining shape and appearance
NASA Astrophysics Data System (ADS)
Han, Eun Jung; Kang, Byung Jun; Park, Kang Ryoung; Lee, Sangyoun
2010-11-01
Facial expression recognition can be widely used for various applications, such as emotion-based human-machine interaction, intelligent robot interfaces, face recognition robust to expression variation, etc. Previous studies have been classified as either shape- or appearance-based recognition. The shape-based method has the disadvantage that the individual variance of facial feature points exists irrespective of similar expressions, which can cause a reduction of the recognition accuracy. The appearance-based method has a limitation in that the textural information of the face is very sensitive to variations in illumination. To overcome these problems, a new facial-expression recognition method is proposed, which combines both shape and appearance information, based on the support vector machine (SVM). This research is novel in the following three ways as compared to previous works. First, the facial feature points are automatically detected by using an active appearance model. From these, the shape-based recognition is performed by using the ratios between the facial feature points based on the facial-action coding system. Second, the SVM, which is trained to recognize the same and different expression classes, is proposed to combine two matching scores obtained from the shape- and appearance-based recognitions. Finally, a single SVM is trained to discriminate four different expressions, such as neutral, a smile, anger, and a scream. By determining the expression of the input facial image whose SVM output is at a minimum, the accuracy of the expression recognition is much enhanced. The experimental results showed that the recognition accuracy of the proposed method was better than previous researches and other fusion methods.
Improvement of emotional healthcare system with stress detection from ECG signal.
Tivatansakul, S; Ohkura, M
2015-01-01
Our emotional healthcare system is designed to cope with users' negative emotions in daily life. To make the system more intelligent, we integrated emotion recognition by facial expression to provide appropriate services based on user's current emotional state. Our emotion recognition by facial expression has confusion issue to recognize some positive, neutral and negative emotions that make the emotional healthcare system provide a relaxation service even though users don't have negative emotions. Therefore, to increase the effectiveness of the system to provide the relaxation service, we integrate stress detection from ECG signal. The stress detection might be able to address the confusion issue of emotion recognition by facial expression to provide the service. Indeed, our results show that integration of stress detection increases the effectiveness and efficiency of the emotional healthcare system to provide services.
Herba, Catherine; Phillips, Mary
2004-10-01
Intact emotion processing is critical for normal emotional development. Recent advances in neuroimaging have facilitated the examination of brain development, and have allowed for the exploration of the relationships between the development of emotion processing abilities, and that of associated neural systems. A literature review was performed of published studies examining the development of emotion expression recognition in normal children and psychiatric populations, and of the development of neural systems important for emotion processing. Few studies have explored the development of emotion expression recognition throughout childhood and adolescence. Behavioural studies suggest continued development throughout childhood and adolescence (reflected by accuracy scores and speed of processing), which varies according to the category of emotion displayed. Factors such as sex, socio-economic status, and verbal ability may also affect this development. Functional neuroimaging studies in adults highlight the role of the amygdala in emotion processing. Results of the few neuroimaging studies in children have focused on the role of the amygdala in the recognition of fearful expressions. Although results are inconsistent, they provide evidence throughout childhood and adolescence for the continued development of and sex differences in amygdalar function in response to fearful expressions. Studies exploring emotion expression recognition in psychiatric populations of children and adolescents suggest deficits that are specific to the type of disorder and to the emotion displayed. Results from behavioural and neuroimaging studies indicate continued development of emotion expression recognition and neural regions important for this process throughout childhood and adolescence. Methodological inconsistencies and disparate findings make any conclusion difficult, however. Further studies are required examining the relationship between the development of emotion expression recognition and that of underlying neural systems, in particular subcortical and prefrontal cortical structures. These will inform understanding of the neural bases of normal and abnormal emotional development, and aid the development of earlier interventions for children and adolescents with psychiatric disorders.
Hierarchical Recognition Scheme for Human Facial Expression Recognition Systems
Siddiqi, Muhammad Hameed; Lee, Sungyoung; Lee, Young-Koo; Khan, Adil Mehmood; Truc, Phan Tran Ho
2013-01-01
Over the last decade, human facial expressions recognition (FER) has emerged as an important research area. Several factors make FER a challenging research problem. These include varying light conditions in training and test images; need for automatic and accurate face detection before feature extraction; and high similarity among different expressions that makes it difficult to distinguish these expressions with a high accuracy. This work implements a hierarchical linear discriminant analysis-based facial expressions recognition (HL-FER) system to tackle these problems. Unlike the previous systems, the HL-FER uses a pre-processing step to eliminate light effects, incorporates a new automatic face detection scheme, employs methods to extract both global and local features, and utilizes a HL-FER to overcome the problem of high similarity among different expressions. Unlike most of the previous works that were evaluated using a single dataset, the performance of the HL-FER is assessed using three publicly available datasets under three different experimental settings: n-fold cross validation based on subjects for each dataset separately; n-fold cross validation rule based on datasets; and, finally, a last set of experiments to assess the effectiveness of each module of the HL-FER separately. Weighted average recognition accuracy of 98.7% across three different datasets, using three classifiers, indicates the success of employing the HL-FER for human FER. PMID:24316568
Zhu, Bi; Chen, Chuansheng; Moyzis, Robert K; Dong, Qi; Chen, Chunhui; He, Qinghua; Stern, Hal S; Li, He; Li, Jin; Li, Jun; Lessard, Jared; Lin, Chongde
2012-01-01
This study investigated the relation between genetic variations in the dopamine system and facial expression recognition. A sample of Chinese college students (n = 478) was given a facial expression recognition task. Subjects were genotyped for 98 loci [96 single-nucleotide polymorphisms (SNPs) and 2 variable number tandem repeats] in 16 genes involved in the dopamine neurotransmitter system, including its 4 subsystems: synthesis (TH, DDC, and DBH), degradation/transport (COMT,MAOA,MAOB, and SLC6A3), receptors (DRD1,DRD2,DRD3,DRD4, and DRD5), and modulation (NTS,NTSR1,NTSR2, and NLN). To quantify the total contributions of the dopamine system to emotion recognition, we used a series of multiple regression models. Permutation analyses were performed to assess the posterior probabilities of obtaining such results. Among the 78 loci that were included in the final analyses (after excluding 12 SNPs that were in high linkage disequilibrium and 8 that were not in Hardy-Weinberg equilibrium), 1 (for fear), 3 (for sadness), 5 (for anger), 13 (for surprise), and 15 (for disgust) loci exhibited main effects on the recognition of facial expressions. Genetic variations in the dopamine system accounted for 3% for fear, 6% for sadness, 7% for anger, 10% for surprise, and 18% for disgust, with the latter surviving a stringent permutation test. Genetic variations in the dopamine system (especially the dopamine synthesis and modulation subsystems) made significant contributions to individual differences in the recognition of disgust faces. Copyright © 2012 S. Karger AG, Basel.
Proposal of Self-Learning and Recognition System of Facial Expression
NASA Astrophysics Data System (ADS)
Ogawa, Yukihiro; Kato, Kunihito; Yamamoto, Kazuhiko
We describe realization of more complicated function by using the information acquired from some equipped unripe functions. The self-learning and recognition system of the human facial expression, which achieved under the natural relation between human and robot, are proposed. The robot with this system can understand human facial expressions and behave according to their facial expressions after the completion of learning process. The system modelled after the process that a baby learns his/her parents’ facial expressions. Equipping the robot with a camera the system can get face images and equipping the CdS sensors on the robot’s head the robot can get the information of human action. Using the information of these sensors, the robot can get feature of each facial expression. After self-learning is completed, when a person changed his facial expression in front of the robot, the robot operates actions under the relevant facial expression.
NASA Astrophysics Data System (ADS)
Harit, Aditya; Joshi, J. C., Col; Gupta, K. K.
2018-03-01
The paper proposed an automatic facial emotion recognition algorithm which comprises of two main components: feature extraction and expression recognition. The algorithm uses a Gabor filter bank on fiducial points to find the facial expression features. The resulting magnitudes of Gabor transforms, along with 14 chosen FAPs (Facial Animation Parameters), compose the feature space. There are two stages: the training phase and the recognition phase. Firstly, for the present 6 different emotions, the system classifies all training expressions in 6 different classes (one for each emotion) in the training stage. In the recognition phase, it recognizes the emotion by applying the Gabor bank to a face image, then finds the fiducial points, and then feeds it to the trained neural architecture.
Decreased acetylcholine release delays the consolidation of object recognition memory.
De Jaeger, Xavier; Cammarota, Martín; Prado, Marco A M; Izquierdo, Iván; Prado, Vania F; Pereira, Grace S
2013-02-01
Acetylcholine (ACh) is important for different cognitive functions such as learning, memory and attention. The release of ACh depends on its vesicular loading by the vesicular acetylcholine transporter (VAChT). It has been demonstrated that VAChT expression can modulate object recognition memory. However, the role of VAChT expression on object recognition memory persistence still remains to be understood. To address this question we used distinct mouse lines with reduced expression of VAChT, as well as pharmacological manipulations of the cholinergic system. We showed that reduction of cholinergic tone impairs object recognition memory measured at 24h. Surprisingly, object recognition memory, measured at 4 days after training, was impaired by substantial, but not moderate, reduction in VAChT expression. Our results suggest that levels of acetylcholine release strongly modulate object recognition memory consolidation and appear to be of particular importance for memory persistence 4 days after training. Copyright © 2012 Elsevier B.V. All rights reserved.
Facial Expression Recognition using Multiclass Ensemble Least-Square Support Vector Machine
NASA Astrophysics Data System (ADS)
Lawi, Armin; Sya'Rani Machrizzandi, M.
2018-03-01
Facial expression is one of behavior characteristics of human-being. The use of biometrics technology system with facial expression characteristics makes it possible to recognize a person’s mood or emotion. The basic components of facial expression analysis system are face detection, face image extraction, facial classification and facial expressions recognition. This paper uses Principal Component Analysis (PCA) algorithm to extract facial features with expression parameters, i.e., happy, sad, neutral, angry, fear, and disgusted. Then Multiclass Ensemble Least-Squares Support Vector Machine (MELS-SVM) is used for the classification process of facial expression. The result of MELS-SVM model obtained from our 185 different expression images of 10 persons showed high accuracy level of 99.998% using RBF kernel.
More Pronounced Deficits in Facial Emotion Recognition for Schizophrenia than Bipolar Disorder
Goghari, Vina M; Sponheim, Scott R
2012-01-01
Schizophrenia and bipolar disorder are typically separated in diagnostic systems. Behavioural, cognitive, and brain abnormalities associated with each disorder nonetheless overlap. We evaluated the diagnostic specificity of facial emotion recognition deficits in schizophrenia and bipolar disorder to determine whether select aspects of emotion recognition differed for the two disorders. The investigation used an experimental task that included the same facial images in an emotion recognition condition and an age recognition condition (to control for processes associated with general face recognition) in 27 schizophrenia patients, 16 bipolar I patients, and 30 controls. Schizophrenia and bipolar patients exhibited both shared and distinct aspects of facial emotion recognition deficits. Schizophrenia patients had deficits in recognizing angry facial expressions compared to healthy controls and bipolar patients. Compared to control participants, both schizophrenia and bipolar patients were more likely to mislabel facial expressions of anger as fear. Given that schizophrenia patients exhibited a deficit in emotion recognition for angry faces, which did not appear due to generalized perceptual and cognitive dysfunction, improving recognition of threat-related expression may be an important intervention target to improve social functioning in schizophrenia. PMID:23218816
Emotion-independent face recognition
NASA Astrophysics Data System (ADS)
De Silva, Liyanage C.; Esther, Kho G. P.
2000-12-01
Current face recognition techniques tend to work well when recognizing faces under small variations in lighting, facial expression and pose, but deteriorate under more extreme conditions. In this paper, a face recognition system to recognize faces of known individuals, despite variations in facial expression due to different emotions, is developed. The eigenface approach is used for feature extraction. Classification methods include Euclidean distance, back propagation neural network and generalized regression neural network. These methods yield 100% recognition accuracy when the training database is representative, containing one image representing the peak expression for each emotion of each person apart from the neutral expression. The feature vectors used for comparison in the Euclidean distance method and for training the neural network must be all the feature vectors of the training set. These results are obtained for a face database consisting of only four persons.
2004-05-01
Army Soldier System Command: http://www.natick.armv.mil Role Name Facial Recognition Program Manager, Army Technical Lead Mark Chandler...security force with a facial recognition system. Mike Holloran, technology officer with the 6 Fleet, directed LCDR Hoa Ho and CAPT(s) Todd Morgan to...USN 6th Fleet was accomplished with the admiral expressing his support for continuing the evaluation of the a facial recognition system. This went
NASA Astrophysics Data System (ADS)
Wan, Qianwen; Panetta, Karen; Agaian, Sos
2017-05-01
Autonomous facial recognition system is widely used in real-life applications, such as homeland border security, law enforcement identification and authentication, and video-based surveillance analysis. Issues like low image quality, non-uniform illumination as well as variations in poses and facial expressions can impair the performance of recognition systems. To address the non-uniform illumination challenge, we present a novel robust autonomous facial recognition system inspired by the human visual system based, so called, logarithmical image visualization technique. In this paper, the proposed method, for the first time, utilizes the logarithmical image visualization technique coupled with the local binary pattern to perform discriminative feature extraction for facial recognition system. The Yale database, the Yale-B database and the ATT database are used for computer simulation accuracy and efficiency testing. The extensive computer simulation demonstrates the method's efficiency, accuracy, and robustness of illumination invariance for facial recognition.
Utterance independent bimodal emotion recognition in spontaneous communication
NASA Astrophysics Data System (ADS)
Tao, Jianhua; Pan, Shifeng; Yang, Minghao; Li, Ya; Mu, Kaihui; Che, Jianfeng
2011-12-01
Emotion expressions sometimes are mixed with the utterance expression in spontaneous face-to-face communication, which makes difficulties for emotion recognition. This article introduces the methods of reducing the utterance influences in visual parameters for the audio-visual-based emotion recognition. The audio and visual channels are first combined under a Multistream Hidden Markov Model (MHMM). Then, the utterance reduction is finished by finding the residual between the real visual parameters and the outputs of the utterance related visual parameters. This article introduces the Fused Hidden Markov Model Inversion method which is trained in the neutral expressed audio-visual corpus to solve the problem. To reduce the computing complexity the inversion model is further simplified to a Gaussian Mixture Model (GMM) mapping. Compared with traditional bimodal emotion recognition methods (e.g., SVM, CART, Boosting), the utterance reduction method can give better results of emotion recognition. The experiments also show the effectiveness of our emotion recognition system when it was used in a live environment.
On Assisting a Visual-Facial Affect Recognition System with Keyboard-Stroke Pattern Information
NASA Astrophysics Data System (ADS)
Stathopoulou, I.-O.; Alepis, E.; Tsihrintzis, G. A.; Virvou, M.
Towards realizing a multimodal affect recognition system, we are considering the advantages of assisting a visual-facial expression recognition system with keyboard-stroke pattern information. Our work is based on the assumption that the visual-facial and keyboard modalities are complementary to each other and that their combination can significantly improve the accuracy in affective user models. Specifically, we present and discuss the development and evaluation process of two corresponding affect recognition subsystems, with emphasis on the recognition of 6 basic emotional states, namely happiness, sadness, surprise, anger and disgust as well as the emotion-less state which we refer to as neutral. We find that emotion recognition by the visual-facial modality can be aided greatly by keyboard-stroke pattern information and the combination of the two modalities can lead to better results towards building a multimodal affect recognition system.
An audiovisual emotion recognition system
NASA Astrophysics Data System (ADS)
Han, Yi; Wang, Guoyin; Yang, Yong; He, Kun
2007-12-01
Human emotions could be expressed by many bio-symbols. Speech and facial expression are two of them. They are both regarded as emotional information which is playing an important role in human-computer interaction. Based on our previous studies on emotion recognition, an audiovisual emotion recognition system is developed and represented in this paper. The system is designed for real-time practice, and is guaranteed by some integrated modules. These modules include speech enhancement for eliminating noises, rapid face detection for locating face from background image, example based shape learning for facial feature alignment, and optical flow based tracking algorithm for facial feature tracking. It is known that irrelevant features and high dimensionality of the data can hurt the performance of classifier. Rough set-based feature selection is a good method for dimension reduction. So 13 speech features out of 37 ones and 10 facial features out of 33 ones are selected to represent emotional information, and 52 audiovisual features are selected due to the synchronization when speech and video fused together. The experiment results have demonstrated that this system performs well in real-time practice and has high recognition rate. Our results also show that the work in multimodules fused recognition will become the trend of emotion recognition in the future.
Wang, Jing; Cui, Xun; Yang, Le; Zhang, Zhe; Lv, Liping; Wang, Haoyuan; Zhao, Zhenmin; Guan, Ningzi; Dong, Lichun; Chen, Rachel
2017-07-01
Artificial control of bio-functions through regulating gene expression is one of the most important and attractive technologies to build novel living systems that are useful in the areas of chemical synthesis, nanotechnology, pharmacology, cell biology. Here, we present a novel real-time control system of gene regulation that includes an enhancement element by introducing duplex DNA aptamers upstream promoter and a repression element by introducing a RNA aptamer upstream ribosome binding site. With the presence of ligands corresponding to the DNA aptamers, the expression of the target gene can be potentially enhanced at the transcriptional level by strengthening the recognition capability of RNAP to the recognition region and speeding up the separation efficiency of the unwinding region due to the induced DNA bubble around the thrombin-bound aptamers; while with the presence of RNA aptamer ligand, the gene expression can be repressed at the translational level by weakening the recognition capability of ribosome to RBS due to the shielding of RBS by the formed aptamer-ligand complex upstream RBS. The effectiveness and potential utility of the developed gene regulation system were demonstrated by regulating the expression of ecaA gene in the cell-free systems. The realistic metabolic engineering application of the system has also tested by regulating the expression of mgtC gene and thrombin cDNA in Escherichia coli JD1021 for controlling metabolic flux and improving thrombin production, verifying that the real-time control system of gene regulation is able to realize the dynamic regulation of gene expression with potential applications in bacterial physiology studies and metabolic engineering. Copyright © 2017. Published by Elsevier Inc.
Image ratio features for facial expression recognition application.
Song, Mingli; Tao, Dacheng; Liu, Zicheng; Li, Xuelong; Zhou, Mengchu
2010-06-01
Video-based facial expression recognition is a challenging problem in computer vision and human-computer interaction. To target this problem, texture features have been extracted and widely used, because they can capture image intensity changes raised by skin deformation. However, existing texture features encounter problems with albedo and lighting variations. To solve both problems, we propose a new texture feature called image ratio features. Compared with previously proposed texture features, e.g., high gradient component features, image ratio features are more robust to albedo and lighting variations. In addition, to further improve facial expression recognition accuracy based on image ratio features, we combine image ratio features with facial animation parameters (FAPs), which describe the geometric motions of facial feature points. The performance evaluation is based on the Carnegie Mellon University Cohn-Kanade database, our own database, and the Japanese Female Facial Expression database. Experimental results show that the proposed image ratio feature is more robust to albedo and lighting variations, and the combination of image ratio features and FAPs outperforms each feature alone. In addition, we study asymmetric facial expressions based on our own facial expression database and demonstrate the superior performance of our combined expression recognition system.
In-the-wild facial expression recognition in extreme poses
NASA Astrophysics Data System (ADS)
Yang, Fei; Zhang, Qian; Zheng, Chi; Qiu, Guoping
2018-04-01
In the computer research area, facial expression recognition is a hot research problem. Recent years, the research has moved from the lab environment to in-the-wild circumstances. It is challenging, especially under extreme poses. But current expression detection systems are trying to avoid the pose effects and gain the general applicable ability. In this work, we solve the problem in the opposite approach. We consider the head poses and detect the expressions within special head poses. Our work includes two parts: detect the head pose and group it into one pre-defined head pose class; do facial expression recognize within each pose class. Our experiments show that the recognition results with pose class grouping are much better than that of direct recognition without considering poses. We combine the hand-crafted features, SIFT, LBP and geometric feature, with deep learning feature as the representation of the expressions. The handcrafted features are added into the deep learning framework along with the high level deep learning features. As a comparison, we implement SVM and random forest to as the prediction models. To train and test our methodology, we labeled the face dataset with 6 basic expressions.
[Neurological disease and facial recognition].
Kawamura, Mitsuru; Sugimoto, Azusa; Kobayakawa, Mutsutaka; Tsuruya, Natsuko
2012-07-01
To discuss the neurological basis of facial recognition, we present our case reports of impaired recognition and a review of previous literature. First, we present a case of infarction and discuss prosopagnosia, which has had a large impact on face recognition research. From a study of patient symptoms, we assume that prosopagnosia may be caused by unilateral right occipitotemporal lesion and right cerebral dominance of facial recognition. Further, circumscribed lesion and degenerative disease may also cause progressive prosopagnosia. Apperceptive prosopagnosia is observed in patients with posterior cortical atrophy (PCA), pathologically considered as Alzheimer's disease, and associative prosopagnosia in frontotemporal lobar degeneration (FTLD). Second, we discuss face recognition as part of communication. Patients with Parkinson disease show social cognitive impairments, such as difficulty in facial expression recognition and deficits in theory of mind as detected by the reading the mind in the eyes test. Pathological and functional imaging studies indicate that social cognitive impairment in Parkinson disease is possibly related to damages in the amygdalae and surrounding limbic system. The social cognitive deficits can be observed in the early stages of Parkinson disease, and even in the prodromal stage, for example, patients with rapid eye movement (REM) sleep behavior disorder (RBD) show impairment in facial expression recognition. Further, patients with myotonic dystrophy type 1 (DM 1), which is a multisystem disease that mainly affects the muscles, show social cognitive impairment similar to that of Parkinson disease. Our previous study showed that facial expression recognition impairment of DM 1 patients is associated with lesion in the amygdalae and insulae. Our study results indicate that behaviors and personality traits in DM 1 patients, which are revealed by social cognitive impairment, are attributable to dysfunction of the limbic system.
Kinnavane, L; Amin, E; Horne, M; Aggleton, J P
2014-01-01
The present study examined immediate-early gene expression in the perirhinal cortex of rats with hippocampal lesions. The goal was to test those models of recognition memory which assume that the perirhinal cortex can function independently of the hippocampus. The c-fos gene was targeted, as its expression in the perirhinal cortex is strongly associated with recognition memory. Four groups of rats were examined. Rats with hippocampal lesions and their surgical controls were given either a recognition memory task (novel vs. familiar objects) or a relative recency task (objects with differing degrees of familiarity). Perirhinal Fos expression in the hippocampal-lesioned groups correlated with both recognition and recency performance. The hippocampal lesions, however, had no apparent effect on overall levels of perirhinal or entorhinal cortex c-fos expression in response to novel objects, with only restricted effects being seen in the recency condition. Network analyses showed that whereas the patterns of parahippocampal interactions were differentially affected by novel or familiar objects, these correlated networks were not altered by hippocampal lesions. Additional analyses in control rats revealed two modes of correlated medial temporal activation. Novel stimuli recruited the pathway from the lateral entorhinal cortex (cortical layer II or III) to hippocampal field CA3, and thence to CA1. Familiar stimuli recruited the direct pathway from the lateral entorhinal cortex (principally layer III) to CA1. The present findings not only reveal the independence from the hippocampus of some perirhinal systems associated with recognition memory, but also show how novel stimuli engage hippocampal subfields in qualitatively different ways from familiar stimuli. PMID:25264133
Computerized system for recognition of autism on the basis of gene expression microarray data.
Latkowski, Tomasz; Osowski, Stanislaw
2015-01-01
The aim of this paper is to provide a means to recognize a case of autism using gene expression microarrays. The crucial task is to discover the most important genes which are strictly associated with autism. The paper presents an application of different methods of gene selection, to select the most representative input attributes for an ensemble of classifiers. The set of classifiers is responsible for distinguishing autism data from the reference class. Simultaneous application of a few gene selection methods enables analysis of the ill-conditioned gene expression matrix from different points of view. The results of selection combined with a genetic algorithm and SVM classifier have shown increased accuracy of autism recognition. Early recognition of autism is extremely important for treatment of children and increases the probability of their recovery and return to normal social communication. The results of this research can find practical application in early recognition of autism on the basis of gene expression microarray analysis. Copyright © 2014 Elsevier Ltd. All rights reserved.
Self-organized Evaluation of Dynamic Hand Gestures for Sign Language Recognition
NASA Astrophysics Data System (ADS)
Buciu, Ioan; Pitas, Ioannis
Two main theories exist with respect to face encoding and representation in the human visual system (HVS). The first one refers to the dense (holistic) representation of the face, where faces have "holon"-like appearance. The second one claims that a more appropriate face representation is given by a sparse code, where only a small fraction of the neural cells corresponding to face encoding is activated. Theoretical and experimental evidence suggest that the HVS performs face analysis (encoding, storing, face recognition, facial expression recognition) in a structured and hierarchical way, where both representations have their own contribution and goal. According to neuropsychological experiments, it seems that encoding for face recognition, relies on holistic image representation, while a sparse image representation is used for facial expression analysis and classification. From the computer vision perspective, the techniques developed for automatic face and facial expression recognition fall into the same two representation types. Like in Neuroscience, the techniques which perform better for face recognition yield a holistic image representation, while those techniques suitable for facial expression recognition use a sparse or local image representation. The proposed mathematical models of image formation and encoding try to simulate the efficient storing, organization and coding of data in the human cortex. This is equivalent with embedding constraints in the model design regarding dimensionality reduction, redundant information minimization, mutual information minimization, non-negativity constraints, class information, etc. The presented techniques are applied as a feature extraction step followed by a classification method, which also heavily influences the recognition results.
Daini, Roberta; Comparetti, Chiara M.; Ricciardelli, Paola
2014-01-01
Neuropsychological and neuroimaging studies have shown that facial recognition and emotional expressions are dissociable. However, it is unknown if a single system supports the processing of emotional and non-emotional facial expressions. We aimed to understand if individuals with impairment in face recognition from birth (congenital prosopagnosia, CP) can use non-emotional facial expressions to recognize a face as an already seen one, and thus, process this facial dimension independently from features (which are impaired in CP), and basic emotional expressions. To this end, we carried out a behavioral study in which we compared the performance of 6 CP individuals to that of typical development individuals, using upright and inverted faces. Four avatar faces with a neutral expression were presented in the initial phase. The target faces presented in the recognition phase, in which a recognition task was requested (2AFC paradigm), could be identical (neutral) to those of the initial phase or present biologically plausible changes to features, non-emotional expressions, or emotional expressions. After this task, a second task was performed, in which the participants had to detect whether or not the recognized face exactly matched the study face or showed any difference. The results confirmed the CPs' impairment in the configural processing of the invariant aspects of the face, but also showed a spared configural processing of non-emotional facial expression (task 1). Interestingly and unlike the non-emotional expressions, the configural processing of emotional expressions was compromised in CPs and did not improve their change detection ability (task 2). These new results have theoretical implications for face perception models since they suggest that, at least in CPs, non-emotional expressions are processed configurally, can be dissociated from other facial dimensions, and may serve as a compensatory strategy to achieve face recognition. PMID:25520643
Daini, Roberta; Comparetti, Chiara M; Ricciardelli, Paola
2014-01-01
Neuropsychological and neuroimaging studies have shown that facial recognition and emotional expressions are dissociable. However, it is unknown if a single system supports the processing of emotional and non-emotional facial expressions. We aimed to understand if individuals with impairment in face recognition from birth (congenital prosopagnosia, CP) can use non-emotional facial expressions to recognize a face as an already seen one, and thus, process this facial dimension independently from features (which are impaired in CP), and basic emotional expressions. To this end, we carried out a behavioral study in which we compared the performance of 6 CP individuals to that of typical development individuals, using upright and inverted faces. Four avatar faces with a neutral expression were presented in the initial phase. The target faces presented in the recognition phase, in which a recognition task was requested (2AFC paradigm), could be identical (neutral) to those of the initial phase or present biologically plausible changes to features, non-emotional expressions, or emotional expressions. After this task, a second task was performed, in which the participants had to detect whether or not the recognized face exactly matched the study face or showed any difference. The results confirmed the CPs' impairment in the configural processing of the invariant aspects of the face, but also showed a spared configural processing of non-emotional facial expression (task 1). Interestingly and unlike the non-emotional expressions, the configural processing of emotional expressions was compromised in CPs and did not improve their change detection ability (task 2). These new results have theoretical implications for face perception models since they suggest that, at least in CPs, non-emotional expressions are processed configurally, can be dissociated from other facial dimensions, and may serve as a compensatory strategy to achieve face recognition.
Cognitive penetrability and emotion recognition in human facial expressions
Marchi, Francesco
2015-01-01
Do our background beliefs, desires, and mental images influence our perceptual experience of the emotions of others? In this paper, we will address the possibility of cognitive penetration (CP) of perceptual experience in the domain of social cognition. In particular, we focus on emotion recognition based on the visual experience of facial expressions. After introducing the current debate on CP, we review examples of perceptual adaptation for facial expressions of emotion. This evidence supports the idea that facial expressions are perceptually processed as wholes. That is, the perceptual system integrates lower-level facial features, such as eyebrow orientation, mouth angle etc., into facial compounds. We then present additional experimental evidence showing that in some cases, emotion recognition on the basis of facial expression is sensitive to and modified by the background knowledge of the subject. We argue that such sensitivity is best explained as a difference in the visual experience of the facial expression, not just as a modification of the judgment based on this experience. The difference in experience is characterized as the result of the interference of background knowledge with the perceptual integration process for faces. Thus, according to the best explanation, we have to accept CP in some cases of emotion recognition. Finally, we discuss a recently proposed mechanism for CP in the face-based recognition of emotion. PMID:26150796
Ipser, Jonathan C; Terburg, David; Syal, Supriya; Phillips, Nicole; Solms, Mark; Panksepp, Jaak; Malcolm-Smith, Susan; Thomas, Kevin; Stein, Dan J; van Honk, Jack
2013-01-01
In rodents, the endogenous opioid system has been implicated in emotion regulation, and in the reduction of fear in particular. In humans, while there is evidence that the opioid antagonist naloxone acutely enhances the acquisition of conditioned fear, there are no corresponding data on the effect of opioid agonists in moderating responses to fear. We investigated whether a single 0.2mg administration of the mu-opioid agonist buprenorphine would decrease fear sensitivity with an emotion-recognition paradigm. Healthy human subjects participated in a randomized placebo-controlled within-subject design, in which they performed a dynamic emotion recognition task 120min after administration of buprenorphine and placebo. In the recognition task, basic emotional expressions were morphed between their full expression and neutral in 2% steps, and presented as dynamic video-clips with final frames of different emotional intensity for each trial, which allows for a fine-grained measurement of emotion sensitivity. Additionally, visual analog scales were used to investigate acute effects of buprenorphine on mood. Compared to placebo, buprenorphine resulted in a significant reduction in the sensitivity for recognizing fearful facial expressions exclusively. Our data demonstrate, for the first time in humans, that acute up-regulation of the opioid system reduces fear recognition sensitivity. Moreover, the absence of an effect of buprenorphine on mood provides evidence of a direct influence of opioids upon the core fear system in the human brain. Copyright © 2012 Elsevier Ltd. All rights reserved.
ERIC Educational Resources Information Center
Wigmore, Angela; Hunter, Gordon; Pflugel, Eckhard; Denholm-Price, James; Binelli, Vincent
2009-01-01
Speech technology--especially automatic speech recognition--has now advanced to a level where it can be of great benefit both to able-bodied people and those with various disabilities. In this paper we describe an application "TalkMaths" which, using the output from a commonly-used conventional automatic speech recognition system,…
Astrocytic expression of HIV-1 Nef impairs spatial and recognition memory
Chompre, Gladys; Cruz, Emmanuel; Maldonado, Lucianette; Rivera-Amill, Vanessa; Porter, James T.; Noel, Richard J.
2012-01-01
Despite the widespread use of antiretroviral therapy that effectively limits viral replication, memory impairment remains a dilemma for HIV infected people. In the CNS, HIV infection of astrocytes leads to the production of the HIV-1 Nef protein without viral replication. Post mortem studies have found Nef expression in hippocampal astrocytes of people with HIV associated dementia suggesting that astrocytic Nef may contribute to HIV associated cognitive impairment even when viral replication is suppressed. To test whether astrocytic expression of Nef is sufficient to induce cognitive deficits, we examined the effect of implanting primary rat astrocytes expressing Nef into the hippocampus on spatial and recognition memory. Rats implanted unilaterally with astrocytes expressing Nef showed impaired novel location and novel object recognition in comparison with controls implanted with astrocytes expressing green fluorescent protein (GFP). This impairment was correlated with an increase in chemokine ligand 2 (CCL2) expression and the infiltration of peripheral macrophages into the hippocampus at the site of injection. Furthermore, the Nef exposed rats exhibited a bilateral loss of CA3 neurons. These results suggest that Nef protein expressed by the implanted astrocytes activates the immune system leading to neuronal damage and spatial and recognition memory deficits. Therefore, the continued expression of Nef by astrocytes in the absence of viral replication has the potential to contribute to HIV associated cognitive impairment. PMID:22926191
Human and animal sounds influence recognition of body language.
Van den Stock, Jan; Grèzes, Julie; de Gelder, Beatrice
2008-11-25
In naturalistic settings emotional events have multiple correlates and are simultaneously perceived by several sensory systems. Recent studies have shown that recognition of facial expressions is biased towards the emotion expressed by a simultaneously presented emotional expression in the voice even if attention is directed to the face only. So far, no study examined whether this phenomenon also applies to whole body expressions, although there is no obvious reason why this crossmodal influence would be specific for faces. Here we investigated whether perception of emotions expressed in whole body movements is influenced by affective information provided by human and by animal vocalizations. Participants were instructed to attend to the action displayed by the body and to categorize the expressed emotion. The results indicate that recognition of body language is biased towards the emotion expressed by the simultaneously presented auditory information, whether it consist of human or of animal sounds. Our results show that a crossmodal influence from auditory to visual emotional information obtains for whole body video images with the facial expression blanked and includes human as well as animal sounds.
Automatic Facial Expression Recognition and Operator Functional State
NASA Technical Reports Server (NTRS)
Blanson, Nina
2012-01-01
The prevalence of human error in safety-critical occupations remains a major challenge to mission success despite increasing automation in control processes. Although various methods have been proposed to prevent incidences of human error, none of these have been developed to employ the detection and regulation of Operator Functional State (OFS), or the optimal condition of the operator while performing a task, in work environments due to drawbacks such as obtrusiveness and impracticality. A video-based system with the ability to infer an individual's emotional state from facial feature patterning mitigates some of the problems associated with other methods of detecting OFS, like obtrusiveness and impracticality in integration with the mission environment. This paper explores the utility of facial expression recognition as a technology for inferring OFS by first expounding on the intricacies of OFS and the scientific background behind emotion and its relationship with an individual's state. Then, descriptions of the feedback loop and the emotion protocols proposed for the facial recognition program are explained. A basic version of the facial expression recognition program uses Haar classifiers and OpenCV libraries to automatically locate key facial landmarks during a live video stream. Various methods of creating facial expression recognition software are reviewed to guide future extensions of the program. The paper concludes with an examination of the steps necessary in the research of emotion and recommendations for the creation of an automatic facial expression recognition program for use in real-time, safety-critical missions
Automatic Facial Expression Recognition and Operator Functional State
NASA Technical Reports Server (NTRS)
Blanson, Nina
2011-01-01
The prevalence of human error in safety-critical occupations remains a major challenge to mission success despite increasing automation in control processes. Although various methods have been proposed to prevent incidences of human error, none of these have been developed to employ the detection and regulation of Operator Functional State (OFS), or the optimal condition of the operator while performing a task, in work environments due to drawbacks such as obtrusiveness and impracticality. A video-based system with the ability to infer an individual's emotional state from facial feature patterning mitigates some of the problems associated with other methods of detecting OFS, like obtrusiveness and impracticality in integration with the mission environment. This paper explores the utility of facial expression recognition as a technology for inferring OFS by first expounding on the intricacies of OFS and the scientific background behind emotion and its relationship with an individual's state. Then, descriptions of the feedback loop and the emotion protocols proposed for the facial recognition program are explained. A basic version of the facial expression recognition program uses Haar classifiers and OpenCV libraries to automatically locate key facial landmarks during a live video stream. Various methods of creating facial expression recognition software are reviewed to guide future extensions of the program. The paper concludes with an examination of the steps necessary in the research of emotion and recommendations for the creation of an automatic facial expression recognition program for use in real-time, safety-critical missions.
Cross-domain expression recognition based on sparse coding and transfer learning
NASA Astrophysics Data System (ADS)
Yang, Yong; Zhang, Weiyi; Huang, Yong
2017-05-01
Traditional facial expression recognition methods usually assume that the training set and the test set are independent and identically distributed. However, in actual expression recognition applications, the conditions of independent and identical distribution are hardly satisfied for the training set and test set because of the difference of light, shade, race and so on. In order to solve this problem and improve the performance of expression recognition in the actual applications, a novel method based on transfer learning and sparse coding is applied to facial expression recognition. First of all, a common primitive model, that is, the dictionary is learnt. Then, based on the idea of transfer learning, the learned primitive pattern is transferred to facial expression and the corresponding feature representation is obtained by sparse coding. The experimental results in CK +, JAFFE and NVIE database shows that the transfer learning based on sparse coding method can effectively improve the expression recognition rate in the cross-domain expression recognition task and is suitable for the practical facial expression recognition applications.
[Prosopagnosia and facial expression recognition].
Koyama, Shinichi
2014-04-01
This paper reviews clinical neuropsychological studies that have indicated that the recognition of a person's identity and the recognition of facial expressions are processed by different cortical and subcortical areas of the brain. The fusiform gyrus, especially the right fusiform gyrus, plays an important role in the recognition of identity. The superior temporal sulcus, amygdala, and medial frontal cortex play important roles in facial-expression recognition. Both facial recognition and facial-expression recognition are highly intellectual processes that involve several regions of the brain.
Effects of exposure to facial expression variation in face learning and recognition.
Liu, Chang Hong; Chen, Wenfeng; Ward, James
2015-11-01
Facial expression is a major source of image variation in face images. Linking numerous expressions to the same face can be a huge challenge for face learning and recognition. It remains largely unknown what level of exposure to this image variation is critical for expression-invariant face recognition. We examined this issue in a recognition memory task, where the number of facial expressions of each face being exposed during a training session was manipulated. Faces were either trained with multiple expressions or a single expression, and they were later tested in either the same or different expressions. We found that recognition performance after learning three emotional expressions had no improvement over learning a single emotional expression (Experiments 1 and 2). However, learning three emotional expressions improved recognition compared to learning a single neutral expression (Experiment 3). These findings reveal both the limitation and the benefit of multiple exposures to variations of emotional expression in achieving expression-invariant face recognition. The transfer of expression training to a new type of expression is likely to depend on a relatively extensive level of training and a certain degree of variation across the types of expressions.
Pan, Ning; Wu, Gui-Hua; Zhang, Ling; Zhao, Ya-Fen; Guan, Han; Xu, Cai-Juan; Jing, Jin; Jin, Yu
2017-03-01
To investigate the features of intelligence development, facial expression recognition ability, and the association between them in children with autism spectrum disorder (ASD). A total of 27 ASD children aged 6-16 years (ASD group, full intelligence quotient >70) and age- and gender-matched normally developed children (control group) were enrolled. Wechsler Intelligence Scale for Children Fourth Edition and Chinese Static Facial Expression Photos were used for intelligence evaluation and facial expression recognition test. Compared with the control group, the ASD group had significantly lower scores of full intelligence quotient, verbal comprehension index, perceptual reasoning index (PRI), processing speed index(PSI), and working memory index (WMI) (P<0.05). The ASD group also had a significantly lower overall accuracy rate of facial expression recognition and significantly lower accuracy rates of the recognition of happy, angry, sad, and frightened expressions than the control group (P<0.05). In the ASD group, the overall accuracy rate of facial expression recognition and the accuracy rates of the recognition of happy and frightened expressions were positively correlated with PRI (r=0.415, 0.455, and 0.393 respectively; P<0.05). The accuracy rate of the recognition of angry expression was positively correlated with WMI (r=0.397; P<0.05). ASD children have delayed intelligence development compared with normally developed children and impaired expression recognition ability. Perceptual reasoning and working memory abilities are positively correlated with expression recognition ability, which suggests that insufficient perceptual reasoning and working memory abilities may be important factors affecting facial expression recognition ability in ASD children.
Bihippocampal damage with emotional dysfunction: impaired auditory recognition of fear.
Ghika-Schmid, F; Ghika, J; Vuilleumier, P; Assal, G; Vuadens, P; Scherer, K; Maeder, P; Uske, A; Bogousslavsky, J
1997-01-01
A right-handed man developed a sudden transient, amnestic syndrome associated with bilateral hemorrhage of the hippocampi, probably due to Urbach-Wiethe disease. In the 3rd month, despite significant hippocampal structural damage on imaging, only a milder degree of retrograde and anterograde amnesia persisted on detailed neuropsychological examination. On systematic testing of recognition of facial and vocal expression of emotion, we found an impairment of the vocal perception of fear, but not that of other emotions, such as joy, sadness and anger. Such selective impairment of fear perception was not present in the recognition of facial expression of emotion. Thus emotional perception varies according to the different aspects of emotions and the different modality of presentation (faces versus voices). This is consistent with the idea that there may be multiple emotion systems. The study of emotional perception in this unique case of bilateral involvement of hippocampus suggests that this structure may play a critical role in the recognition of fear in vocal expression, possibly dissociated from that of other emotions and from that of fear in facial expression. In regard of recent data suggesting that the amygdala is playing a role in the recognition of fear in the auditory as well as in the visual modality this could suggest that the hippocampus may be part of the auditory pathway of fear recognition.
Pose-variant facial expression recognition using an embedded image system
NASA Astrophysics Data System (ADS)
Song, Kai-Tai; Han, Meng-Ju; Chang, Shuo-Hung
2008-12-01
In recent years, one of the most attractive research areas in human-robot interaction is automated facial expression recognition. Through recognizing the facial expression, a pet robot can interact with human in a more natural manner. In this study, we focus on the facial pose-variant problem. A novel method is proposed in this paper to recognize pose-variant facial expressions. After locating the face position in an image frame, the active appearance model (AAM) is applied to track facial features. Fourteen feature points are extracted to represent the variation of facial expressions. The distance between feature points are defined as the feature values. These feature values are sent to a support vector machine (SVM) for facial expression determination. The pose-variant facial expression is classified into happiness, neutral, sadness, surprise or anger. Furthermore, in order to evaluate the performance for practical applications, this study also built a low resolution database (160x120 pixels) using a CMOS image sensor. Experimental results show that the recognition rate is 84% with the self-built database.
NASA Astrophysics Data System (ADS)
Cui, Chen; Asari, Vijayan K.
2014-03-01
Biometric features such as fingerprints, iris patterns, and face features help to identify people and restrict access to secure areas by performing advanced pattern analysis and matching. Face recognition is one of the most promising biometric methodologies for human identification in a non-cooperative security environment. However, the recognition results obtained by face recognition systems are a affected by several variations that may happen to the patterns in an unrestricted environment. As a result, several algorithms have been developed for extracting different facial features for face recognition. Due to the various possible challenges of data captured at different lighting conditions, viewing angles, facial expressions, and partial occlusions in natural environmental conditions, automatic facial recognition still remains as a difficult issue that needs to be resolved. In this paper, we propose a novel approach to tackling some of these issues by analyzing the local textural descriptions for facial feature representation. The textural information is extracted by an enhanced local binary pattern (ELBP) description of all the local regions of the face. The relationship of each pixel with respect to its neighborhood is extracted and employed to calculate the new representation. ELBP reconstructs a much better textural feature extraction vector from an original gray level image in different lighting conditions. The dimensionality of the texture image is reduced by principal component analysis performed on each local face region. Each low dimensional vector representing a local region is now weighted based on the significance of the sub-region. The weight of each sub-region is determined by employing the local variance estimate of the respective region, which represents the significance of the region. The final facial textural feature vector is obtained by concatenating the reduced dimensional weight sets of all the modules (sub-regions) of the face image. Experiments conducted on various popular face databases show promising performance of the proposed algorithm in varying lighting, expression, and partial occlusion conditions. Four databases were used for testing the performance of the proposed system: Yale Face database, Extended Yale Face database B, Japanese Female Facial Expression database, and CMU AMP Facial Expression database. The experimental results in all four databases show the effectiveness of the proposed system. Also, the computation cost is lower because of the simplified calculation steps. Research work is progressing to investigate the effectiveness of the proposed face recognition method on pose-varying conditions as well. It is envisaged that a multilane approach of trained frameworks at different pose bins and an appropriate voting strategy would lead to a good recognition rate in such situation.
Facial Emotion Recognition: A Survey and Real-World User Experiences in Mixed Reality
Mehta, Dhwani; Siddiqui, Mohammad Faridul Haque
2018-01-01
Extensive possibilities of applications have made emotion recognition ineluctable and challenging in the field of computer science. The use of non-verbal cues such as gestures, body movement, and facial expressions convey the feeling and the feedback to the user. This discipline of Human–Computer Interaction places reliance on the algorithmic robustness and the sensitivity of the sensor to ameliorate the recognition. Sensors play a significant role in accurate detection by providing a very high-quality input, hence increasing the efficiency and the reliability of the system. Automatic recognition of human emotions would help in teaching social intelligence in the machines. This paper presents a brief study of the various approaches and the techniques of emotion recognition. The survey covers a succinct review of the databases that are considered as data sets for algorithms detecting the emotions by facial expressions. Later, mixed reality device Microsoft HoloLens (MHL) is introduced for observing emotion recognition in Augmented Reality (AR). A brief introduction of its sensors, their application in emotion recognition and some preliminary results of emotion recognition using MHL are presented. The paper then concludes by comparing results of emotion recognition by the MHL and a regular webcam. PMID:29389845
Facial Emotion Recognition: A Survey and Real-World User Experiences in Mixed Reality.
Mehta, Dhwani; Siddiqui, Mohammad Faridul Haque; Javaid, Ahmad Y
2018-02-01
Extensive possibilities of applications have made emotion recognition ineluctable and challenging in the field of computer science. The use of non-verbal cues such as gestures, body movement, and facial expressions convey the feeling and the feedback to the user. This discipline of Human-Computer Interaction places reliance on the algorithmic robustness and the sensitivity of the sensor to ameliorate the recognition. Sensors play a significant role in accurate detection by providing a very high-quality input, hence increasing the efficiency and the reliability of the system. Automatic recognition of human emotions would help in teaching social intelligence in the machines. This paper presents a brief study of the various approaches and the techniques of emotion recognition. The survey covers a succinct review of the databases that are considered as data sets for algorithms detecting the emotions by facial expressions. Later, mixed reality device Microsoft HoloLens (MHL) is introduced for observing emotion recognition in Augmented Reality (AR). A brief introduction of its sensors, their application in emotion recognition and some preliminary results of emotion recognition using MHL are presented. The paper then concludes by comparing results of emotion recognition by the MHL and a regular webcam.
Foundations for a syntatic pattern recognition system for genomic DNA sequences
DOE Office of Scientific and Technical Information (OSTI.GOV)
Searles, D.B.
1993-03-01
The goal of the proposed work is the creation of a software system that will perform sophisticated pattern recognition and related functions at a level of abstraction and with expressive power beyond current general-purpose pattern-matching systems for biological sequences; and with a more uniform language, environment, and graphical user interface, and with greater flexibility, extensibility, embeddability, and ability to incorporate other algorithms, than current special-purpose analytic software.
Amparyup, Piti; Sutthangkul, Jantiwan; Charoensapsri, Walaiporn; Tassanakajon, Anchalee
2012-01-01
The prophenoloxidase (proPO) system is activated upon recognition of pathogens by pattern recognition proteins (PRPs), including a lipopolysaccharide- and β-1,3-glucan-binding protein (LGBP). However, shrimp LGBPs that are involved in the proPO system have yet to be clarified. Here, we focus on characterizing the role of a Penaeus monodon LGBP (PmLGBP) in the proPO system. We found that PmLGBP transcripts are expressed primarily in the hemocytes and are increased at 24 h after pathogenic bacterium Vibrio harveyi challenge. The binding studies carried out using ELISA indicated that recombinant (r)PmLGBP binds to β-1,3-glucan and LPS with a dissociation constant of 6.86 × 10−7 m and 3.55 × 10−7 m, respectively. Furthermore, we found that rPmLGBP could enhance the phenoloxidase (PO) activity of hemocyte suspensions in the presence of LPS or β-1,3-glucan. Using dsRNA interference-mediated gene silencing assay, we further demonstrated that knockdown of PmLGBP in shrimp in vivo significantly decreased the PmLGBP transcript level but had no effect on the expression of the other immune genes tested, including shrimp antimicrobial peptides (AMPs). However, suppression of proPO expression down-regulated PmLGBP, proPO-activating enzyme (PmPPAE2), and AMPs (penaeidin and crustin). Such PmLGBP down-regulated shrimp showed significantly decreased total PO activity. We conclude that PmLGBP functions as a pattern recognition protein for LPS and β-1,3-glucan in the shrimp proPO activating system. PMID:22235126
Fusion of Visible and Thermal Descriptors Using Genetic Algorithms for Face Recognition Systems.
Hermosilla, Gabriel; Gallardo, Francisco; Farias, Gonzalo; San Martin, Cesar
2015-07-23
The aim of this article is to present a new face recognition system based on the fusion of visible and thermal features obtained from the most current local matching descriptors by maximizing face recognition rates through the use of genetic algorithms. The article considers a comparison of the performance of the proposed fusion methodology against five current face recognition methods and classic fusion techniques used commonly in the literature. These were selected by considering their performance in face recognition. The five local matching methods and the proposed fusion methodology are evaluated using the standard visible/thermal database, the Equinox database, along with a new database, the PUCV-VTF, designed for visible-thermal studies in face recognition and described for the first time in this work. The latter is created considering visible and thermal image sensors with different real-world conditions, such as variations in illumination, facial expression, pose, occlusion, etc. The main conclusions of this article are that two variants of the proposed fusion methodology surpass current face recognition methods and the classic fusion techniques reported in the literature, attaining recognition rates of over 97% and 99% for the Equinox and PUCV-VTF databases, respectively. The fusion methodology is very robust to illumination and expression changes, as it combines thermal and visible information efficiently by using genetic algorithms, thus allowing it to choose optimal face areas where one spectrum is more representative than the other.
Fusion of Visible and Thermal Descriptors Using Genetic Algorithms for Face Recognition Systems
Hermosilla, Gabriel; Gallardo, Francisco; Farias, Gonzalo; San Martin, Cesar
2015-01-01
The aim of this article is to present a new face recognition system based on the fusion of visible and thermal features obtained from the most current local matching descriptors by maximizing face recognition rates through the use of genetic algorithms. The article considers a comparison of the performance of the proposed fusion methodology against five current face recognition methods and classic fusion techniques used commonly in the literature. These were selected by considering their performance in face recognition. The five local matching methods and the proposed fusion methodology are evaluated using the standard visible/thermal database, the Equinox database, along with a new database, the PUCV-VTF, designed for visible-thermal studies in face recognition and described for the first time in this work. The latter is created considering visible and thermal image sensors with different real-world conditions, such as variations in illumination, facial expression, pose, occlusion, etc. The main conclusions of this article are that two variants of the proposed fusion methodology surpass current face recognition methods and the classic fusion techniques reported in the literature, attaining recognition rates of over 97% and 99% for the Equinox and PUCV-VTF databases, respectively. The fusion methodology is very robust to illumination and expression changes, as it combines thermal and visible information efficiently by using genetic algorithms, thus allowing it to choose optimal face areas where one spectrum is more representative than the other. PMID:26213932
NASA Astrophysics Data System (ADS)
Medjkoune, Sofiane; Mouchère, Harold; Petitrenaud, Simon; Viard-Gaudin, Christian
2013-01-01
The work reported in this paper concerns the problem of mathematical expressions recognition. This task is known to be a very hard one. We propose to alleviate the difficulties by taking into account two complementary modalities. The modalities referred to are handwriting and audio ones. To combine the signals coming from both modalities, various fusion methods are explored. Performances evaluated on the HAMEX dataset show a significant improvement compared to a single modality (handwriting) based system.
Recognition of chemical entities: combining dictionary-based and grammar-based approaches.
Akhondi, Saber A; Hettne, Kristina M; van der Horst, Eelke; van Mulligen, Erik M; Kors, Jan A
2015-01-01
The past decade has seen an upsurge in the number of publications in chemistry. The ever-swelling volume of available documents makes it increasingly hard to extract relevant new information from such unstructured texts. The BioCreative CHEMDNER challenge invites the development of systems for the automatic recognition of chemicals in text (CEM task) and for ranking the recognized compounds at the document level (CDI task). We investigated an ensemble approach where dictionary-based named entity recognition is used along with grammar-based recognizers to extract compounds from text. We assessed the performance of ten different commercial and publicly available lexical resources using an open source indexing system (Peregrine), in combination with three different chemical compound recognizers and a set of regular expressions to recognize chemical database identifiers. The effect of different stop-word lists, case-sensitivity matching, and use of chunking information was also investigated. We focused on lexical resources that provide chemical structure information. To rank the different compounds found in a text, we used a term confidence score based on the normalized ratio of the term frequencies in chemical and non-chemical journals. The use of stop-word lists greatly improved the performance of the dictionary-based recognition, but there was no additional benefit from using chunking information. A combination of ChEBI and HMDB as lexical resources, the LeadMine tool for grammar-based recognition, and the regular expressions, outperformed any of the individual systems. On the test set, the F-scores were 77.8% (recall 71.2%, precision 85.8%) for the CEM task and 77.6% (recall 71.7%, precision 84.6%) for the CDI task. Missed terms were mainly due to tokenization issues, poor recognition of formulas, and term conjunctions. We developed an ensemble system that combines dictionary-based and grammar-based approaches for chemical named entity recognition, outperforming any of the individual systems that we considered. The system is able to provide structure information for most of the compounds that are found. Improved tokenization and better recognition of specific entity types is likely to further improve system performance.
Recognition of chemical entities: combining dictionary-based and grammar-based approaches
2015-01-01
Background The past decade has seen an upsurge in the number of publications in chemistry. The ever-swelling volume of available documents makes it increasingly hard to extract relevant new information from such unstructured texts. The BioCreative CHEMDNER challenge invites the development of systems for the automatic recognition of chemicals in text (CEM task) and for ranking the recognized compounds at the document level (CDI task). We investigated an ensemble approach where dictionary-based named entity recognition is used along with grammar-based recognizers to extract compounds from text. We assessed the performance of ten different commercial and publicly available lexical resources using an open source indexing system (Peregrine), in combination with three different chemical compound recognizers and a set of regular expressions to recognize chemical database identifiers. The effect of different stop-word lists, case-sensitivity matching, and use of chunking information was also investigated. We focused on lexical resources that provide chemical structure information. To rank the different compounds found in a text, we used a term confidence score based on the normalized ratio of the term frequencies in chemical and non-chemical journals. Results The use of stop-word lists greatly improved the performance of the dictionary-based recognition, but there was no additional benefit from using chunking information. A combination of ChEBI and HMDB as lexical resources, the LeadMine tool for grammar-based recognition, and the regular expressions, outperformed any of the individual systems. On the test set, the F-scores were 77.8% (recall 71.2%, precision 85.8%) for the CEM task and 77.6% (recall 71.7%, precision 84.6%) for the CDI task. Missed terms were mainly due to tokenization issues, poor recognition of formulas, and term conjunctions. Conclusions We developed an ensemble system that combines dictionary-based and grammar-based approaches for chemical named entity recognition, outperforming any of the individual systems that we considered. The system is able to provide structure information for most of the compounds that are found. Improved tokenization and better recognition of specific entity types is likely to further improve system performance. PMID:25810767
[Emotional facial expression recognition impairment in Parkinson disease].
Lachenal-Chevallet, Karine; Bediou, Benoit; Bouvard, Martine; Thobois, Stéphane; Broussolle, Emmanuel; Vighetto, Alain; Krolak-Salmon, Pierre
2006-03-01
some behavioral disturbances observed in Parkinson's disease (PD) could be related to impaired recognition of various social messages particularly emotional facial expressions. facial expression recognition was assessed using morphed faces (five emotions: happiness, fear, anger, disgust, neutral), and compared to gender recognition and general cognitive assessment in 12 patients with Parkinson's disease and 14 controls subjects. facial expression recognition was impaired among patients, whereas gender recognitions, visuo-perceptive capacities and total efficiency were preserved. Post hoc analyses disclosed a deficit for fear and disgust recognition compared to control subjects. the impairment of emotional facial expression recognition in PD appears independent of other cognitive deficits. This impairment may be related to the dopaminergic depletion in basal ganglia and limbic brain regions. They could take a part in psycho-behavioral disorders and particularly in communication disorders observed in Parkinson's disease patients.
Facial expression recognition based on improved deep belief networks
NASA Astrophysics Data System (ADS)
Wu, Yao; Qiu, Weigen
2017-08-01
In order to improve the robustness of facial expression recognition, a method of face expression recognition based on Local Binary Pattern (LBP) combined with improved deep belief networks (DBNs) is proposed. This method uses LBP to extract the feature, and then uses the improved deep belief networks as the detector and classifier to extract the LBP feature. The combination of LBP and improved deep belief networks is realized in facial expression recognition. In the JAFFE (Japanese Female Facial Expression) database on the recognition rate has improved significantly.
Kitada, Ryo; Okamoto, Yuko; Sasaki, Akihiro T.; Kochiyama, Takanori; Miyahara, Motohide; Lederman, Susan J.; Sadato, Norihiro
2012-01-01
Face perception is critical for social communication. Given its fundamental importance in the course of evolution, the innate neural mechanisms can anticipate the computations necessary for representing faces. However, the effect of visual deprivation on the formation of neural mechanisms that underlie face perception is largely unknown. We previously showed that sighted individuals can recognize basic facial expressions by haptics surprisingly well. Moreover, the inferior frontal gyrus (IFG) and posterior superior temporal sulcus (pSTS) in the sighted subjects are involved in haptic and visual recognition of facial expressions. Here, we conducted both psychophysical and functional magnetic-resonance imaging (fMRI) experiments to determine the nature of the neural representation that subserves the recognition of basic facial expressions in early blind individuals. In a psychophysical experiment, both early blind and sighted subjects haptically identified basic facial expressions at levels well above chance. In the subsequent fMRI experiment, both groups haptically identified facial expressions and shoe types (control). The sighted subjects then completed the same task visually. Within brain regions activated by the visual and haptic identification of facial expressions (relative to that of shoes) in the sighted group, corresponding haptic identification in the early blind activated regions in the inferior frontal and middle temporal gyri. These results suggest that the neural system that underlies the recognition of basic facial expressions develops supramodally even in the absence of early visual experience. PMID:23372547
Kitada, Ryo; Okamoto, Yuko; Sasaki, Akihiro T; Kochiyama, Takanori; Miyahara, Motohide; Lederman, Susan J; Sadato, Norihiro
2013-01-01
Face perception is critical for social communication. Given its fundamental importance in the course of evolution, the innate neural mechanisms can anticipate the computations necessary for representing faces. However, the effect of visual deprivation on the formation of neural mechanisms that underlie face perception is largely unknown. We previously showed that sighted individuals can recognize basic facial expressions by haptics surprisingly well. Moreover, the inferior frontal gyrus (IFG) and posterior superior temporal sulcus (pSTS) in the sighted subjects are involved in haptic and visual recognition of facial expressions. Here, we conducted both psychophysical and functional magnetic-resonance imaging (fMRI) experiments to determine the nature of the neural representation that subserves the recognition of basic facial expressions in early blind individuals. In a psychophysical experiment, both early blind and sighted subjects haptically identified basic facial expressions at levels well above chance. In the subsequent fMRI experiment, both groups haptically identified facial expressions and shoe types (control). The sighted subjects then completed the same task visually. Within brain regions activated by the visual and haptic identification of facial expressions (relative to that of shoes) in the sighted group, corresponding haptic identification in the early blind activated regions in the inferior frontal and middle temporal gyri. These results suggest that the neural system that underlies the recognition of basic facial expressions develops supramodally even in the absence of early visual experience.
Impact of severity of drug use on discrete emotions recognition in polysubstance abusers.
Fernández-Serrano, María José; Lozano, Oscar; Pérez-García, Miguel; Verdejo-García, Antonio
2010-06-01
Neuropsychological studies support the association between severity of drug intake and alterations in specific cognitive domains and neural systems, but there is disproportionately less research on the neuropsychology of emotional alterations associated with addiction. One of the key aspects of adaptive emotional functioning potentially relevant to addiction progression and treatment is the ability to recognize basic emotions in the faces of others. Therefore, the aims of this study were: (i) to examine facial emotion recognition in abstinent polysubstance abusers, and (ii) to explore the association between patterns of quantity and duration of use of several drugs co-abused (including alcohol, cannabis, cocaine, heroin and MDMA) and the ability to identify discrete facial emotional expressions portraying basic emotions. We compared accuracy of emotion recognition of facial expressions portraying six basic emotions (measured with the Ekman Faces Test) between polysubstance abusers (PSA, n=65) and non-drug using comparison individuals (NDCI, n=30), and used regression models to explore the association between quantity and duration of use of the different drugs co-abused and indices of recognition of each of the six emotions, while controlling for relevant socio-demographic and affect-related confounders. Results showed: (i) that PSA had significantly poorer recognition than NDCI for facial expressions of anger, disgust, fear and sadness; (ii) that measures of quantity and duration of drugs used significantly predicted poorer discrete emotions recognition: quantity of cocaine use predicted poorer anger recognition, and duration of cocaine use predicted both poorer anger and fear recognition. Severity of cocaine use also significantly predicted overall recognition accuracy. Copyright (c) 2010 Elsevier Ireland Ltd. All rights reserved.
Ponomarev, S A; Berendeeva, T A; Kalinin, S A; Muranova, A V
The system of signaling pattern recognition receptors was studied in 8 cosmonauts aged 35 to 56 years before and after (R+) long-duration missions to the International space station. Peripheral blood samples were analyzed for the content of monocytes and granulocytes that express the signaling pattern recognition Toll- like (TLR) receptors localized as on cell surface (TLR1, TLR2, TLR4, TLR5, TLR6), so inside cells (TLR3, TLR8, TLR9). In parallel, serum concentrations of TLR2 (HSP60) and TLR4 ligands (HSP70, HMGB1) were measured. The results of investigations showed growth of HSP60, HSP70 and HMGB1 concentrations on R+1. In the;majority of cosmonauts increases in endogenous ligands were followed by growth in the number of both monocytes and granulocytes that express TLR2 1 TLR4. This consistency gives ground to assume that changes in the system of signaling pattern recognition receptors can stem .from the predominantly endogenous ligands' response to the effects of long-duration space flight on human organism.
Nomi, Jason S; Rhodes, Matthew G; Cleary, Anne M
2013-01-01
This study examined how participants' predictions of future memory performance are influenced by emotional facial expressions. Participants made judgements of learning (JOLs) predicting the likelihood that they would correctly identify a face displaying a happy, angry, or neutral emotional expression in a future two-alternative forced-choice recognition test of identity (i.e., recognition that a person's face was seen before). JOLs were higher for studied faces with happy and angry emotional expressions than for neutral faces. However, neutral test faces with studied neutral expressions had significantly higher identity recognition rates than neutral test faces studied with happy or angry expressions. Thus, these data are the first to demonstrate that people believe happy and angry emotional expressions will lead to better identity recognition in the future relative to neutral expressions. This occurred despite the fact that neutral expressions elicited better identity recognition than happy and angry expressions. These findings contribute to the growing literature examining the interaction of cognition and emotion.
Impact of Childhood Maltreatment on the Recognition of Facial Expressions of Emotions.
Ardizzi, Martina; Martini, Francesca; Umiltà, Maria Alessandra; Evangelista, Valentina; Ravera, Roberto; Gallese, Vittorio
2015-01-01
The development of the explicit recognition of facial expressions of emotions can be affected by childhood maltreatment experiences. A previous study demonstrated the existence of an explicit recognition bias for angry facial expressions among a population of adolescent Sierra Leonean street-boys exposed to high levels of maltreatment. In the present study, the recognition bias for angry facial expressions was investigated in a younger population of street-children and age-matched controls. Participants performed a forced-choice facial expressions recognition task. Recognition bias was measured as participants' tendency to over-attribute anger label to other negative facial expressions. Participants' heart rate was assessed and related to their behavioral performance, as index of their stress-related physiological responses. Results demonstrated the presence of a recognition bias for angry facial expressions among street-children, also pinpointing a similar, although significantly less pronounced, tendency among controls. Participants' performance was controlled for age, cognitive and educational levels and for naming skills. None of these variables influenced the recognition bias for angry facial expressions. Differently, a significant effect of heart rate on participants' tendency to use anger label was evidenced. Taken together, these results suggest that childhood exposure to maltreatment experiences amplifies children's "pre-existing bias" for anger labeling in forced-choice emotion recognition task. Moreover, they strengthen the thesis according to which the recognition bias for angry facial expressions is a manifestation of a functional adaptive mechanism that tunes victim's perceptive and attentive focus on salient environmental social stimuli.
Impact of Childhood Maltreatment on the Recognition of Facial Expressions of Emotions
Ardizzi, Martina; Martini, Francesca; Umiltà, Maria Alessandra; Evangelista, Valentina; Ravera, Roberto; Gallese, Vittorio
2015-01-01
The development of the explicit recognition of facial expressions of emotions can be affected by childhood maltreatment experiences. A previous study demonstrated the existence of an explicit recognition bias for angry facial expressions among a population of adolescent Sierra Leonean street-boys exposed to high levels of maltreatment. In the present study, the recognition bias for angry facial expressions was investigated in a younger population of street-children and age-matched controls. Participants performed a forced-choice facial expressions recognition task. Recognition bias was measured as participants’ tendency to over-attribute anger label to other negative facial expressions. Participants’ heart rate was assessed and related to their behavioral performance, as index of their stress-related physiological responses. Results demonstrated the presence of a recognition bias for angry facial expressions among street-children, also pinpointing a similar, although significantly less pronounced, tendency among controls. Participants’ performance was controlled for age, cognitive and educational levels and for naming skills. None of these variables influenced the recognition bias for angry facial expressions. Differently, a significant effect of heart rate on participants’ tendency to use anger label was evidenced. Taken together, these results suggest that childhood exposure to maltreatment experiences amplifies children’s “pre-existing bias” for anger labeling in forced-choice emotion recognition task. Moreover, they strengthen the thesis according to which the recognition bias for angry facial expressions is a manifestation of a functional adaptive mechanism that tunes victim’s perceptive and attentive focus on salient environmental social stimuli. PMID:26509890
Art critic: Multisignal vision and speech interaction system in a gaming context.
Reale, Michael J; Liu, Peng; Yin, Lijun; Canavan, Shaun
2013-12-01
True immersion of a player within a game can only occur when the world simulated looks and behaves as close to reality as possible. This implies that the game must correctly read and understand, among other things, the player's focus, attitude toward the objects/persons in focus, gestures, and speech. In this paper, we proposed a novel system that integrates eye gaze estimation, head pose estimation, facial expression recognition, speech recognition, and text-to-speech components for use in real-time games. Both the eye gaze and head pose components utilize underlying 3-D models, and our novel head pose estimation algorithm uniquely combines scene flow with a generic head model. The facial expression recognition module uses the local binary patterns with three orthogonal planes approach on the 2-D shape index domain rather than the pixel domain, resulting in improved classification. Our system has also been extended to use a pan-tilt-zoom camera driven by the Kinect, allowing us to track a moving player. A test game, Art Critic, is also presented, which not only demonstrates the utility of our system but also provides a template for player/non-player character (NPC) interaction in a gaming context. The player alters his/her view of the 3-D world using head pose, looks at paintings/NPCs using eye gaze, and makes an evaluation based on the player's expression and speech. The NPC artist will respond with facial expression and synthetic speech based on its personality. Both qualitative and quantitative evaluations of the system are performed to illustrate the system's effectiveness.
Effects of GABAB receptors in the insula on recognition memory observed with intellicage.
Wu, Nan; Wang, Feng; Jin, Zhe; Zhang, Zhen; Wang, Lian-Kun; Zhang, Chun; Sun, Tao
2017-04-17
Insular function has gradually become a topic of intense study in cognitive research. Recognition memory is a commonly studied type of memory in memory research. GABA B R has been shown to be closely related to memory formation. In the present study, we used intellicage, which is a new intelligent behavioural test system, and a bilateral drug microinjection technique to inject into the bilateral insula, to examine the relationship between GABA B R and recognition memory. Male Sprague-Dawley rats were randomly divided into control, Sham, Nacl, baclofen and CGP35348 groups. Different testing procedures were employed using intellicage to detect changes in rat recognition memory. The expression of GABA B R (GB1, GB2) in the insula of rats was determined by immunofluorescence and western blotting at the protein level. In addition, the expression of GABA B R (GB 1 , GB 2 ) was detected by RT-PCR at the mRNA level. The results of the intellicage test showed that recognition memory was impaired in terms of position learning, punitive learning and punitive reversal learning by using baclofen and CGP35348. In position reversal learning, no significant differences were found in terms of cognitive memory ability between the control groups and the CGP and baclofen groups. Immunofluorescence data showed GABA B R (GB1, GB2) expression in the insula, while data from RT-PCR and western blot analysis demonstrated that the relative expression of GB1 and GB2 was significantly increased in the baclofen group compared with the control groups. In the CGP35348 group, the expression of GB1 and GB2 was significantly decreased, but there was no significant difference in GB1 or GB2 expression in the control groups. GABA B R expression in the insula plays an important role in the formation of recognition memory in rats.
Employing Textual and Facial Emotion Recognition to Design an Affective Tutoring System
ERIC Educational Resources Information Center
Lin, Hao-Chiang Koong; Wang, Cheng-Hung; Chao, Ching-Ju; Chien, Ming-Kuan
2012-01-01
Emotional expression in Artificial Intelligence has gained lots of attention in recent years, people applied its affective computing not only in enhancing and realizing the interaction between computers and human, it also makes computer more humane. In this study, emotional expressions were applied into intelligent tutoring system, where learners'…
Facial expression recognition based on improved local ternary pattern and stacked auto-encoder
NASA Astrophysics Data System (ADS)
Wu, Yao; Qiu, Weigen
2017-08-01
In order to enhance the robustness of facial expression recognition, we propose a method of facial expression recognition based on improved Local Ternary Pattern (LTP) combined with Stacked Auto-Encoder (SAE). This method uses the improved LTP extraction feature, and then uses the improved depth belief network as the detector and classifier to extract the LTP feature. The combination of LTP and improved deep belief network is realized in facial expression recognition. The recognition rate on CK+ databases has improved significantly.
NASA Astrophysics Data System (ADS)
Zhao, Lei; Wang, Zengcai; Wang, Xiaojin; Qi, Yazhou; Liu, Qing; Zhang, Guoxin
2016-09-01
Human fatigue is an important cause of traffic accidents. To improve the safety of transportation, we propose, in this paper, a framework for fatigue expression recognition using image-based facial dynamic multi-information and a bimodal deep neural network. First, the landmark of face region and the texture of eye region, which complement each other in fatigue expression recognition, are extracted from facial image sequences captured by a single camera. Then, two stacked autoencoder neural networks are trained for landmark and texture, respectively. Finally, the two trained neural networks are combined by learning a joint layer on top of them to construct a bimodal deep neural network. The model can be used to extract a unified representation that fuses landmark and texture modalities together and classify fatigue expressions accurately. The proposed system is tested on a human fatigue dataset obtained from an actual driving environment. The experimental results demonstrate that the proposed method performs stably and robustly, and that the average accuracy achieves 96.2%.
Fooprateepsiri, Rerkchai; Kurutach, Werasak
2014-03-01
Face authentication is a biometric classification method that verifies the identity of a user based on image of their face. Accuracy of the authentication is reduced when the pose, illumination and expression of the training face images are different than the testing image. The methods in this paper are designed to improve the accuracy of a features-based face recognition system when the pose between the input images and training images are different. First, an efficient 2D-to-3D integrated face reconstruction approach is introduced to reconstruct a personalized 3D face model from a single frontal face image with neutral expression and normal illumination. Second, realistic virtual faces with different poses are synthesized based on the personalized 3D face to characterize the face subspace. Finally, face recognition is conducted based on these representative virtual faces. Compared with other related works, this framework has the following advantages: (1) only one single frontal face is required for face recognition, which avoids the burdensome enrollment work; and (2) the synthesized face samples provide the capability to conduct recognition under difficult conditions like complex pose, illumination and expression. From the experimental results, we conclude that the proposed method improves the accuracy of face recognition by varying the pose, illumination and expression. Copyright © 2014 Elsevier Ireland Ltd. All rights reserved.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Searles, D.B.
1993-03-01
The goal of the proposed work is the creation of a software system that will perform sophisticated pattern recognition and related functions at a level of abstraction and with expressive power beyond current general-purpose pattern-matching systems for biological sequences; and with a more uniform language, environment, and graphical user interface, and with greater flexibility, extensibility, embeddability, and ability to incorporate other algorithms, than current special-purpose analytic software.
Research of Face Recognition with Fisher Linear Discriminant
NASA Astrophysics Data System (ADS)
Rahim, R.; Afriliansyah, T.; Winata, H.; Nofriansyah, D.; Ratnadewi; Aryza, S.
2018-01-01
Face identification systems are developing rapidly, and these developments drive the advancement of biometric-based identification systems that have high accuracy. However, to develop a good face recognition system and to have high accuracy is something that’s hard to find. Human faces have diverse expressions and attribute changes such as eyeglasses, mustache, beard and others. Fisher Linear Discriminant (FLD) is a class-specific method that distinguishes facial image images into classes and also creates distance between classes and intra classes so as to produce better classification.
Corneanu, Ciprian Adrian; Simon, Marc Oliu; Cohn, Jeffrey F; Guerrero, Sergio Escalera
2016-08-01
Facial expressions are an important way through which humans interact socially. Building a system capable of automatically recognizing facial expressions from images and video has been an intense field of study in recent years. Interpreting such expressions remains challenging and much research is needed about the way they relate to human affect. This paper presents a general overview of automatic RGB, 3D, thermal and multimodal facial expression analysis. We define a new taxonomy for the field, encompassing all steps from face detection to facial expression recognition, and describe and classify the state of the art methods accordingly. We also present the important datasets and the bench-marking of most influential methods. We conclude with a general discussion about trends, important questions and future lines of research.
de la Rosa, Stephan; Fademrecht, Laura; Bülthoff, Heinrich H; Giese, Martin A; Curio, Cristóbal
2018-06-01
Motor-based theories of facial expression recognition propose that the visual perception of facial expression is aided by sensorimotor processes that are also used for the production of the same expression. Accordingly, sensorimotor and visual processes should provide congruent emotional information about a facial expression. Here, we report evidence that challenges this view. Specifically, the repeated execution of facial expressions has the opposite effect on the recognition of a subsequent facial expression than the repeated viewing of facial expressions. Moreover, the findings of the motor condition, but not of the visual condition, were correlated with a nonsensory condition in which participants imagined an emotional situation. These results can be well accounted for by the idea that facial expression recognition is not always mediated by motor processes but can also be recognized on visual information alone.
False recognition of facial expressions of emotion: causes and implications.
Fernández-Dols, José-Miguel; Carrera, Pilar; Barchard, Kimberly A; Gacitua, Marta
2008-08-01
This article examines the importance of semantic processes in the recognition of emotional expressions, through a series of three studies on false recognition. The first study found a high frequency of false recognition of prototypical expressions of emotion when participants viewed slides and video clips of nonprototypical fearful and happy expressions. The second study tested whether semantic processes caused false recognition. The authors found that participants made significantly higher error rates when asked to detect expressions that corresponded to semantic labels than when asked to detect visual stimuli. Finally, given that previous research reported that false memories are less prevalent in younger children, the third study tested whether false recognition of prototypical expressions increased with age. The authors found that 67% of eight- to nine-year-old children reported nonpresent prototypical expressions of fear in a fearful context, but only 40% of 6- to 7-year-old children did so. Taken together, these three studies demonstrate the importance of semantic processes in the detection and categorization of prototypical emotional expressions.
Fukushiro, Masato; Takeuchi, Takahiro; Takeuchi, Yuki; Hur, Sung-Pyo; Sugama, Nozomi; Takemura, Akihiro; Kubo, Yoko; Okano, Keiko; Okano, Toshiyuki
2011-01-01
Lunar cycle-associated physiology has been found in a wide variety of organisms. Recent study has revealed that mRNA levels of Cryptochrome (Cry), one of the circadian clock genes, were significantly higher on a full moon night than on a new moon night in coral, implying the involvement of a photoreception system in the lunar-synchronized spawning. To better establish the generalities surrounding such a mechanism and explore the underlying molecular mechanism, we focused on the relationship between lunar phase, Cry gene expression, and the spawning behavior in a lunar-synchronized spawner, the goldlined spinefoot (Siganus guttatus), and we identified two kinds of Cry genes in this animal. Their mRNA levels showed lunar cycle-dependent expression in the medial part of the brain (mesencephalon and diencephalon) peaking at the first quarter moon. Since this lunar phase coincided with the reproductive phase of the goldlined spinefoot, Cry gene expression was considered a state variable in the lunar phase recognition system. Based on the expression profiles of SgCrys together with the moonlight's pattern of timing and duration during its nightly lunar cycle, we have further speculated on a model of lunar phase recognition for reproductive control in the goldlined spinefoot, which integrates both moonlight and circadian signals in a manner similar to photoperiodic response. PMID:22163321
Fukushiro, Masato; Takeuchi, Takahiro; Takeuchi, Yuki; Hur, Sung-Pyo; Sugama, Nozomi; Takemura, Akihiro; Kubo, Yoko; Okano, Keiko; Okano, Toshiyuki
2011-01-01
Lunar cycle-associated physiology has been found in a wide variety of organisms. Recent study has revealed that mRNA levels of Cryptochrome (Cry), one of the circadian clock genes, were significantly higher on a full moon night than on a new moon night in coral, implying the involvement of a photoreception system in the lunar-synchronized spawning. To better establish the generalities surrounding such a mechanism and explore the underlying molecular mechanism, we focused on the relationship between lunar phase, Cry gene expression, and the spawning behavior in a lunar-synchronized spawner, the goldlined spinefoot (Siganus guttatus), and we identified two kinds of Cry genes in this animal. Their mRNA levels showed lunar cycle-dependent expression in the medial part of the brain (mesencephalon and diencephalon) peaking at the first quarter moon. Since this lunar phase coincided with the reproductive phase of the goldlined spinefoot, Cry gene expression was considered a state variable in the lunar phase recognition system. Based on the expression profiles of SgCrys together with the moonlight's pattern of timing and duration during its nightly lunar cycle, we have further speculated on a model of lunar phase recognition for reproductive control in the goldlined spinefoot, which integrates both moonlight and circadian signals in a manner similar to photoperiodic response.
Perceptual and affective mechanisms in facial expression recognition: An integrative review.
Calvo, Manuel G; Nummenmaa, Lauri
2016-09-01
Facial expressions of emotion involve a physical component of morphological changes in a face and an affective component conveying information about the expresser's internal feelings. It remains unresolved how much recognition and discrimination of expressions rely on the perception of morphological patterns or the processing of affective content. This review of research on the role of visual and emotional factors in expression recognition reached three major conclusions. First, behavioral, neurophysiological, and computational measures indicate that basic expressions are reliably recognized and discriminated from one another, albeit the effect may be inflated by the use of prototypical expression stimuli and forced-choice responses. Second, affective content along the dimensions of valence and arousal is extracted early from facial expressions, although this coarse affective representation contributes minimally to categorical recognition of specific expressions. Third, the physical configuration and visual saliency of facial features contribute significantly to expression recognition, with "emotionless" computational models being able to reproduce some of the basic phenomena demonstrated in human observers. We conclude that facial expression recognition, as it has been investigated in conventional laboratory tasks, depends to a greater extent on perceptual than affective information and mechanisms.
Temporal identity in axonal target layer recognition.
Petrovic, Milan; Hummel, Thomas
2008-12-11
The segregation of axon and dendrite projections into distinct synaptic layers is a fundamental principle of nervous system organization and the structural basis for information processing in the brain. Layer-specific recognition molecules that allow projecting neurons to stabilize transient contacts and initiate synaptogenesis have been identified. However, most of the neuronal cell-surface molecules critical for layer organization are expressed broadly in the developing nervous system, raising the question of how these so-called permissive adhesion molecules support synaptic specificity. Here we show that the temporal expression dynamics of the zinc-finger protein sequoia is the major determinant of Drosophila photoreceptor connectivity into distinct synaptic layers. Neighbouring R8 and R7 photoreceptors show consecutive peaks of elevated sequoia expression, which correspond to their sequential target-layer innervation. Loss of sequoia in R7 leads to a projection switch into the R8 recipient layer, whereas a prolonged expression in R8 induces a redirection of their axons into the R7 layer. The sequoia-induced axon targeting is mediated through the ubiquitously expressed Cadherin-N cell adhesion molecule. Our data support a model in which recognition specificity during synaptic layer formation is generated through a temporally restricted axonal competence to respond to broadly expressed adhesion molecules. Because developing neurons innervating the same target area often project in a distinct, birth-order-dependent sequence, temporal identity seems to contain crucial information in generating not only cell type diversity during neuronal division but also connection diversity of projecting neurons.
Conformal Predictions in Multimedia Pattern Recognition
ERIC Educational Resources Information Center
Nallure Balasubramanian, Vineeth
2010-01-01
The fields of pattern recognition and machine learning are on a fundamental quest to design systems that can learn the way humans do. One important aspect of human intelligence that has so far not been given sufficient attention is the capability of humans to express when they are certain about a decision, or when they are not. Machine learning…
Biometrics Foundation Documents
2009-01-01
a digital form. The quality of the sensor used has a significant impact on the recognition results. Example “sensors” could be digital cameras...Difficult to control sensor and channel variances that significantly impact capabilities Not sufficiently distinctive for identification over large...expressions, hairstyle, glasses, hats, makeup, etc. have on face recognition systems? Minor variances , such as those mentioned, will have a moderate
ERIC Educational Resources Information Center
Bekele, Esubalew; Crittendon, Julie; Zheng, Zhi; Swanson, Amy; Weitlauf, Amy; Warren, Zachary; Sarkar, Nilanjan
2014-01-01
Teenagers with autism spectrum disorder (ASD) and age-matched controls participated in a dynamic facial affect recognition task within a virtual reality (VR) environment. Participants identified the emotion of a facial expression displayed at varied levels of intensity by a computer generated avatar. The system assessed performance (i.e.,…
NASA Astrophysics Data System (ADS)
Bushel, Pierre R.; Bennett, Lee; Hamadeh, Hisham; Green, James; Ableson, Alan; Misener, Steve; Paules, Richard; Afshari, Cynthia
2002-06-01
We present an analysis of pattern recognition procedures used to predict the classes of samples exposed to pharmacologic agents by comparing gene expression patterns from samples treated with two classes of compounds. Rat liver mRNA samples following exposure for 24 hours with phenobarbital or peroxisome proliferators were analyzed using a 1700 rat cDNA microarray platform. Sets of genes that were consistently differentially expressed in the rat liver samples following treatment were stored in the MicroArray Project System (MAPS) database. MAPS identified 238 genes in common that possessed a low probability (P < 0.01) of being randomly detected as differentially expressed at the 95% confidence level. Hierarchical cluster analysis on the 238 genes clustered specific gene expression profiles that separated samples based on exposure to a particular class of compound.
Chen, Guangdong; Lin, Xiaodong; Li, Gongying; Jiang, Diego; Lib, Zhiruo; Jiang, Ronghuan; Zhuo, Chuanjun
2017-01-01
The aim of the present study was to investigate the effects of a commonly-used atypical antipsychotic, risperidone, on alterations in spatial learning and in the hippocampal brain-derived neurotrophic factor (BDNF)-tyrosine receptor kinase B (TrkB) signalling system caused by acute dizocilpine maleate (MK-801) treatment. In experiment 1, adult male Sprague-Dawley rats subjected to acute treatment of either low-dose MK801 (0.1 mg/kg) or normal saline (vehicle) were tested for spatial object recognition and hippocampal expression levels of BDNF, TrkB and the phophorylation of TrkB (p-TrkB). We found that compared to the vehicle, MK-801 treatment impaired spatial object recognition of animals and downregulated the expression levels of p-TrkB. In experiment 2, MK-801- or vehicle-treated animals were further injected with risperidone (0.1 mg/kg) or vehicle before behavioural testing and sacrifice. Of note, we found that risperidone successfully reversed the deleterious effects of MK-801 on spatial object recognition and upregulated the hippocampal BDNF-TrkB signalling system. Collectively, the findings suggest that cognitive deficits from acute N-methyl-D-aspartate receptor blockade may be associated with the hypofunction of hippocampal BDNF-TrkB signalling system and that risperidone was able to reverse these alterations. PMID:28451387
Impaired recognition of happy facial expressions in bipolar disorder.
Lawlor-Savage, Linette; Sponheim, Scott R; Goghari, Vina M
2014-08-01
The ability to accurately judge facial expressions is important in social interactions. Individuals with bipolar disorder have been found to be impaired in emotion recognition; however, the specifics of the impairment are unclear. This study investigated whether facial emotion recognition difficulties in bipolar disorder reflect general cognitive, or emotion-specific, impairments. Impairment in the recognition of particular emotions and the role of processing speed in facial emotion recognition were also investigated. Clinically stable bipolar patients (n = 17) and healthy controls (n = 50) judged five facial expressions in two presentation types, time-limited and self-paced. An age recognition condition was used as an experimental control. Bipolar patients' overall facial recognition ability was unimpaired. However, patients' specific ability to judge happy expressions under time constraints was impaired. Findings suggest a deficit in happy emotion recognition impacted by processing speed. Given the limited sample size, further investigation with a larger patient sample is warranted.
Han, Ren-Wen; Xu, Hong-Jiao; Zhang, Rui-San; Wang, Pei; Chang, Min; Peng, Ya-Li; Deng, Ke-Yu; Wang, Rui
2014-01-01
The noradrenergic activity in the basolateral amygdala (BLA) was reported to be involved in the regulation of object recognition memory. As the BLA expresses high density of receptors for Neuropeptide S (NPS), we investigated whether the BLA is involved in mediating NPS's effects on object recognition memory consolidation and whether such effects require noradrenergic activity. Intracerebroventricular infusion of NPS (1nmol) post training facilitated 24-h memory in a mouse novel object recognition task. The memory-enhancing effect of NPS could be blocked by the β-adrenoceptor antagonist propranolol. Furthermore, post-training intra-BLA infusions of NPS (0.5nmol/side) improved 24-h memory for objects, which was impaired by co-administration of propranolol (0.5μg/side). Taken together, these results indicate that NPS interacts with the BLA noradrenergic system in improving object recognition memory during consolidation. Copyright © 2013 Elsevier Inc. All rights reserved.
When Early Experiences Build a Wall to Others’ Emotions: An Electrophysiological and Autonomic Study
Ardizzi, Martina; Martini, Francesca; Umiltà, Maria Alessandra; Sestito, Mariateresa; Ravera, Roberto; Gallese, Vittorio
2013-01-01
Facial expression of emotions is a powerful vehicle for communicating information about others’ emotional states and it normally induces facial mimicry in the observers. The aim of this study was to investigate if early aversive experiences could interfere with emotion recognition, facial mimicry, and with the autonomic regulation of social behaviors. We conducted a facial emotion recognition task in a group of “street-boys” and in an age-matched control group. We recorded facial electromyography (EMG), a marker of facial mimicry, and respiratory sinus arrhythmia (RSA), an index of the recruitment of autonomic system promoting social behaviors and predisposition, in response to the observation of facial expressions of emotions. Results showed an over-attribution of anger, and reduced EMG responses during the observation of both positive and negative expressions only among street-boys. Street-boys also showed lower RSA after observation of facial expressions and ineffective RSA suppression during presentation of non-threatening expressions. Our findings suggest that early aversive experiences alter not only emotion recognition but also facial mimicry of emotions. These deficits affect the autonomic regulation of social behaviors inducing lower social predisposition after the visualization of facial expressions and an ineffective recruitment of defensive behavior in response to non-threatening expressions. PMID:23593374
Doi, Hirokazu; Fujisawa, Takashi X; Kanai, Chieko; Ohta, Haruhisa; Yokoi, Hideki; Iwanami, Akira; Kato, Nobumasa; Shinohara, Kazuyuki
2013-09-01
This study investigated the ability of adults with Asperger syndrome to recognize emotional categories of facial expressions and emotional prosodies with graded emotional intensities. The individuals with Asperger syndrome showed poorer recognition performance for angry and sad expressions from both facial and vocal information. The group difference in facial expression recognition was prominent for stimuli with low or intermediate emotional intensities. In contrast to this, the individuals with Asperger syndrome exhibited lower recognition accuracy than typically-developed controls mainly for emotional prosody with high emotional intensity. In facial expression recognition, Asperger and control groups showed an inversion effect for all categories. The magnitude of this effect was less in the Asperger group for angry and sad expressions, presumably attributable to reduced recruitment of the configural mode of face processing. The individuals with Asperger syndrome outperformed the control participants in recognizing inverted sad expressions, indicating enhanced processing of local facial information representing sad emotion. These results suggest that the adults with Asperger syndrome rely on modality-specific strategies in emotion recognition from facial expression and prosodic information.
Fu, Si-Yao; Yang, Guo-Sheng; Kuai, Xin-Kai
2012-01-01
In this paper, we present a quantitative, highly structured cortex-simulated model, which can be simply described as feedforward, hierarchical simulation of ventral stream of visual cortex using biologically plausible, computationally convenient spiking neural network system. The motivation comes directly from recent pioneering works on detailed functional decomposition analysis of the feedforward pathway of the ventral stream of visual cortex and developments on artificial spiking neural networks (SNNs). By combining the logical structure of the cortical hierarchy and computing power of the spiking neuron model, a practical framework has been presented. As a proof of principle, we demonstrate our system on several facial expression recognition tasks. The proposed cortical-like feedforward hierarchy framework has the merit of capability of dealing with complicated pattern recognition problems, suggesting that, by combining the cognitive models with modern neurocomputational approaches, the neurosystematic approach to the study of cortex-like mechanism has the potential to extend our knowledge of brain mechanisms underlying the cognitive analysis and to advance theoretical models of how we recognize face or, more specifically, perceive other people's facial expression in a rich, dynamic, and complex environment, providing a new starting point for improved models of visual cortex-like mechanism. PMID:23193391
Fu, Si-Yao; Yang, Guo-Sheng; Kuai, Xin-Kai
2012-01-01
In this paper, we present a quantitative, highly structured cortex-simulated model, which can be simply described as feedforward, hierarchical simulation of ventral stream of visual cortex using biologically plausible, computationally convenient spiking neural network system. The motivation comes directly from recent pioneering works on detailed functional decomposition analysis of the feedforward pathway of the ventral stream of visual cortex and developments on artificial spiking neural networks (SNNs). By combining the logical structure of the cortical hierarchy and computing power of the spiking neuron model, a practical framework has been presented. As a proof of principle, we demonstrate our system on several facial expression recognition tasks. The proposed cortical-like feedforward hierarchy framework has the merit of capability of dealing with complicated pattern recognition problems, suggesting that, by combining the cognitive models with modern neurocomputational approaches, the neurosystematic approach to the study of cortex-like mechanism has the potential to extend our knowledge of brain mechanisms underlying the cognitive analysis and to advance theoretical models of how we recognize face or, more specifically, perceive other people's facial expression in a rich, dynamic, and complex environment, providing a new starting point for improved models of visual cortex-like mechanism.
Neural network face recognition using wavelets
NASA Astrophysics Data System (ADS)
Karunaratne, Passant V.; Jouny, Ismail I.
1997-04-01
The recognition of human faces is a phenomenon that has been mastered by the human visual system and that has been researched extensively in the domain of computer neural networks and image processing. This research is involved in the study of neural networks and wavelet image processing techniques in the application of human face recognition. The objective of the system is to acquire a digitized still image of a human face, carry out pre-processing on the image as required, an then, given a prior database of images of possible individuals, be able to recognize the individual in the image. The pre-processing segment of the system includes several procedures, namely image compression, denoising, and feature extraction. The image processing is carried out using Daubechies wavelets. Once the images have been passed through the wavelet-based image processor they can be efficiently analyzed by means of a neural network. A back- propagation neural network is used for the recognition segment of the system. The main constraints of the system is with regard to the characteristics of the images being processed. The system should be able to carry out effective recognition of the human faces irrespective of the individual's facial-expression, presence of extraneous objects such as head-gear or spectacles, and face/head orientation. A potential application of this face recognition system would be as a secondary verification method in an automated teller machine.
Eye-Gaze Analysis of Facial Emotion Recognition and Expression in Adolescents with ASD.
Wieckowski, Andrea Trubanova; White, Susan W
2017-01-01
Impaired emotion recognition and expression in individuals with autism spectrum disorder (ASD) may contribute to observed social impairment. The aim of this study was to examine the role of visual attention directed toward nonsocial aspects of a scene as a possible mechanism underlying recognition and expressive ability deficiency in ASD. One recognition and two expression tasks were administered. Recognition was assessed in force-choice paradigm, and expression was assessed during scripted and free-choice response (in response to emotional stimuli) tasks in youth with ASD (n = 20) and an age-matched sample of typically developing youth (n = 20). During stimulus presentation prior to response in each task, participants' eye gaze was tracked. Youth with ASD were less accurate at identifying disgust and sadness in the recognition task. They fixated less to the eye region of stimuli showing surprise. A group difference was found during the free-choice response task, such that those with ASD expressed emotion less clearly but not during the scripted task. Results suggest altered eye gaze to the mouth region but not the eye region as a candidate mechanism for decreased ability to recognize or express emotion. Findings inform our understanding of the association between social attention and emotion recognition and expression deficits.
Facial expression system on video using widrow hoff
NASA Astrophysics Data System (ADS)
Jannah, M.; Zarlis, M.; Mawengkang, H.
2018-03-01
Facial expressions recognition is one of interesting research. This research contains human feeling to computer application Such as the interaction between human and computer, data compression, facial animation and facial detection from the video. The purpose of this research is to create facial expression system that captures image from the video camera. The system in this research uses Widrow-Hoff learning method in training and testing image with Adaptive Linear Neuron (ADALINE) approach. The system performance is evaluated by two parameters, detection rate and false positive rate. The system accuracy depends on good technique and face position that trained and tested.
The nonverbal expression of pride: evidence for cross-cultural recognition.
Tracy, Jessica L; Robins, Richard W
2008-03-01
The present research tests whether recognition for the nonverbal expression of pride generalizes across cultures. Study 1 provided the first evidence for cross-cultural recognition of pride, demonstrating that the expression generalizes across Italy and the United States. Study 2 found that the pride expression generalizes beyond Western cultures; individuals from a preliterate, highly isolated tribe in Burkina Faso, West Africa, reliably recognized pride, regardless of whether it was displayed by African or American targets. These Burkinabe participants were unlikely to have learned the pride expression through cross-cultural transmission, so their recognition suggests that pride may be a human universal. Studies 3 and 4 used drawn figures to systematically manipulate the ethnicity and gender of targets showing the expression, and demonstrated that pride recognition generalizes across male and female targets of African, Asian, and Caucasian descent. Discussion focuses on the implications of the findings for the universality of the pride expression.
VLSI Microsystem for Rapid Bioinformatic Pattern Recognition
NASA Technical Reports Server (NTRS)
Fang, Wai-Chi; Lue, Jaw-Chyng
2009-01-01
A system comprising very-large-scale integrated (VLSI) circuits is being developed as a means of bioinformatics-oriented analysis and recognition of patterns of fluorescence generated in a microarray in an advanced, highly miniaturized, portable genetic-expression-assay instrument. Such an instrument implements an on-chip combination of polymerase chain reactions and electrochemical transduction for amplification and detection of deoxyribonucleic acid (DNA).
Trinkler, Iris; Cleret de Langavant, Laurent; Bachoud-Lévi, Anne-Catherine
2013-02-01
Patients with Huntington's disease (HD), a neurodegenerative disorder that causes major motor impairments, also show cognitive and emotional deficits. While their deficit in recognising emotions has been explored in depth, little is known about their ability to express emotions and understand their feelings. If these faculties were impaired, patients might not only mis-read emotion expressions in others but their own emotions might be mis-interpreted by others as well, or thirdly, they might have difficulties understanding and describing their feelings. We compared the performance of recognition and expression of facial emotions in 13 HD patients with mild motor impairments but without significant bucco-facial abnormalities, and 13 controls matched for age and education. Emotion recognition was investigated in a forced-choice recognition test (FCR), and emotion expression by filming participants while they mimed the six basic emotional facial expressions (anger, disgust, fear, surprise, sadness and joy) to the experimenter. The films were then segmented into 60 stimuli per participant and four external raters performed a FCR on this material. Further, we tested understanding of feelings in self (alexithymia) and others (empathy) using questionnaires. Both recognition and expression were impaired across different emotions in HD compared to controls and recognition and expression scores were correlated. By contrast, alexithymia and empathy scores were very similar in HD and controls. This might suggest that emotion deficits in HD might be tied to the expression itself. Because similar emotion recognition-expression deficits are also found in Parkinson's Disease and vascular lesions of the striatum, our results further confirm the importance of the striatum for emotion recognition and expression, while access to the meaning of feelings relies on a different brain network, and is spared in HD. Copyright © 2011 Elsevier Ltd. All rights reserved.
ERIC Educational Resources Information Center
Sheaffer, Beverly L.; Golden, Jeannie A.; Averett, Paige
2009-01-01
The ability to recognize facial expressions of emotion is integral in social interaction. Although the importance of facial expression recognition is reflected in increased research interest as well as in popular culture, clinicians may know little about this topic. The purpose of this article is to discuss facial expression recognition literature…
NASA Astrophysics Data System (ADS)
Sato, Ayuko; Iwasaki, Akiko
2004-11-01
Pattern recognition by Toll-like receptors (TLRs) is known to be important for the induction of dendritic cell (DC) maturation. DCs, in turn, are critically important in the initiation of T cell responses. However, most viruses do not infect DCs. This recognition system poses a biological problem in ensuring that most viral infections be detected by pattern recognition receptors. Furthermore, it is unknown what, if any, is the contribution of TLRs expressed by cells that are infected by a virus, versus TLRs expressed by DCs, in the initiation of antiviral adaptive immunity. Here we address these issues using a physiologically relevant model of mucosal infection with herpes simplex virus type 2. We demonstrate that innate immune recognition of viral infection occurs in two distinct stages, one at the level of the infected epithelial cells and the other at the level of the noninfected DCs. Importantly, both TLR-mediated recognition events are required for the induction of effector T cells. Our results demonstrate that virally infected tissues instruct DCs to initiate the appropriate class of effector T cell responses and reveal the critical importance of the stromal cells in detecting infectious agents through their own pattern recognition receptors. mucosal immunity | pattern recognition | viral infection
Guillaume, Fabrice; Guillem, François; Tiberghien, Guy; Stip, Emmanuel
2012-09-01
The objective was to investigate the electrophysiological (ERP) correlates of mismatched expression on face recognition in schizophrenia. Expression-change effects and associated ERPs were explored in patients with schizophrenia (n = 20) and paired comparison participants (n = 20) on a long-term face-recognition task. A facial-expression change decreased discriminability for patients with schizophrenia than for healthy participants. The patients' recognition deficit was accompanied by the absence of the midfrontal FN400 and late parietal ERP old/new effects in the mismatched-expression condition. By contrast, preserved midfrontal FN400 and late parietal ERP old/new effects were found in both groups in the unchanged-expression condition. Thus, the preserved parietal old/new effect previously observed in schizophrenia was no longer found here in the situation in which expression changes took place between the study and recognition phases. These findings suggest that, when they are not supposed to take the change of expression into account, the recognition deficit observed here in patients with schizophrenia resulted from an impairment in the mechanisms underlying the emergence, assessment, or utilization of familiarity--as indexed by the ERP old/new effects. In these natural conditions, the impact of the expression change on the implementation of retrieval processes offers new insight into schizophrenia-linked deficits in face recognition, with substantial phenomenological differences with respect to the emergence of familiarity.
Facial Emotion Recognition and Expression in Parkinson's Disease: An Emotional Mirror Mechanism?
Ricciardi, Lucia; Visco-Comandini, Federica; Erro, Roberto; Morgante, Francesca; Bologna, Matteo; Fasano, Alfonso; Ricciardi, Diego; Edwards, Mark J; Kilner, James
2017-01-01
Parkinson's disease (PD) patients have impairment of facial expressivity (hypomimia) and difficulties in interpreting the emotional facial expressions produced by others, especially for aversive emotions. We aimed to evaluate the ability to produce facial emotional expressions and to recognize facial emotional expressions produced by others in a group of PD patients and a group of healthy participants in order to explore the relationship between these two abilities and any differences between the two groups of participants. Twenty non-demented, non-depressed PD patients and twenty healthy participants (HC) matched for demographic characteristics were studied. The ability of recognizing emotional facial expressions was assessed with the Ekman 60-faces test (Emotion recognition task). Participants were video-recorded while posing facial expressions of 6 primary emotions (happiness, sadness, surprise, disgust, fear and anger). The most expressive pictures for each emotion were derived from the videos. Ten healthy raters were asked to look at the pictures displayed on a computer-screen in pseudo-random fashion and to identify the emotional label in a six-forced-choice response format (Emotion expressivity task). Reaction time (RT) and accuracy of responses were recorded. At the end of each trial the participant was asked to rate his/her confidence in his/her perceived accuracy of response. For emotion recognition, PD reported lower score than HC for Ekman total score (p<0.001), and for single emotions sub-scores happiness, fear, anger, sadness (p<0.01) and surprise (p = 0.02). In the facial emotion expressivity task, PD and HC significantly differed in the total score (p = 0.05) and in the sub-scores for happiness, sadness, anger (all p<0.001). RT and the level of confidence showed significant differences between PD and HC for the same emotions. There was a significant positive correlation between the emotion facial recognition and expressivity in both groups; the correlation was even stronger when ranking emotions from the best recognized to the worst (R = 0.75, p = 0.004). PD patients showed difficulties in recognizing emotional facial expressions produced by others and in posing facial emotional expressions compared to healthy subjects. The linear correlation between recognition and expression in both experimental groups suggests that the two mechanisms share a common system, which could be deteriorated in patients with PD. These results open new clinical and rehabilitation perspectives.
Clark, G F; Dell, A; Morris, H R; Patankar, M S; Easton, R L
2001-01-01
We have previously suggested that the human fetus is protected during human development by a system of both soluble and cell surface associated glycoconjugates that utilize their carbohydrate sequences as functional groups to enable them to evoke tolerance. The proposed model has been referred to as the human fetoembryonic defense system hypothesis (hu-FEDS). In this paradigm, it has previously been proposed that similar oligosaccharides are used to mediate crucial recognition events required during both human sperm-egg binding and immune-inflammatory cell interactions. This vertical integration suggested to us that the sperm-egg binding itself is related to universal recognition events that occur between immune and inflammatory cells, except that in this case recognition of 'species' rather than recognition of 'self' is being manifested. In this paper, we have designated this component of hu-FEDS as the species recognition system (SRS). We propose that the SRS is an integral component of the hu-FEDS used to enable sperm-egg recognition and protection of the gametes from potential immune responses. Recent structural data indicates that the glycan sequences implicated in mediating murine gamete recognition are also expressed on CD45 in activated murine T lymphocytes and cytotoxic T lymphocytes. This overlap supports our contention that there is an overlap between the immune and gamete recognition systems. Therefore the hu-FEDS paradigm may be a subset of a larger model that also applies to other placental mammals. We therefore propose that the hu-FEDS model for protection should in the future be referred to as the eutherian fetoembryonic defense system hypothesis (eu-FEDS) to account for this extension. The possibility exists that the SRS component of eu-FEDS could predate eutherians and extend to all sexually reproducing organisms. Future investigation of the interactions between the immune and gamete recognition system will be required to determine the degree of overlap. Copyright 2001 S. Karger AG, Basel
Wingenbach, Tanja S H; Ashwin, Chris; Brosnan, Mark
2018-01-01
There has been much research on sex differences in the ability to recognise facial expressions of emotions, with results generally showing a female advantage in reading emotional expressions from the face. However, most of the research to date has used static images and/or 'extreme' examples of facial expressions. Therefore, little is known about how expression intensity and dynamic stimuli might affect the commonly reported female advantage in facial emotion recognition. The current study investigated sex differences in accuracy of response (Hu; unbiased hit rates) and response latencies for emotion recognition using short video stimuli (1sec) of 10 different facial emotion expressions (anger, disgust, fear, sadness, surprise, happiness, contempt, pride, embarrassment, neutral) across three variations in the intensity of the emotional expression (low, intermediate, high) in an adolescent and adult sample (N = 111; 51 male, 60 female) aged between 16 and 45 (M = 22.2, SD = 5.7). Overall, females showed more accurate facial emotion recognition compared to males and were faster in correctly recognising facial emotions. The female advantage in reading expressions from the faces of others was unaffected by expression intensity levels and emotion categories used in the study. The effects were specific to recognition of emotions, as males and females did not differ in the recognition of neutral faces. Together, the results showed a robust sex difference favouring females in facial emotion recognition using video stimuli of a wide range of emotions and expression intensity variations.
Sex differences in facial emotion recognition across varying expression intensity levels from videos
2018-01-01
There has been much research on sex differences in the ability to recognise facial expressions of emotions, with results generally showing a female advantage in reading emotional expressions from the face. However, most of the research to date has used static images and/or ‘extreme’ examples of facial expressions. Therefore, little is known about how expression intensity and dynamic stimuli might affect the commonly reported female advantage in facial emotion recognition. The current study investigated sex differences in accuracy of response (Hu; unbiased hit rates) and response latencies for emotion recognition using short video stimuli (1sec) of 10 different facial emotion expressions (anger, disgust, fear, sadness, surprise, happiness, contempt, pride, embarrassment, neutral) across three variations in the intensity of the emotional expression (low, intermediate, high) in an adolescent and adult sample (N = 111; 51 male, 60 female) aged between 16 and 45 (M = 22.2, SD = 5.7). Overall, females showed more accurate facial emotion recognition compared to males and were faster in correctly recognising facial emotions. The female advantage in reading expressions from the faces of others was unaffected by expression intensity levels and emotion categories used in the study. The effects were specific to recognition of emotions, as males and females did not differ in the recognition of neutral faces. Together, the results showed a robust sex difference favouring females in facial emotion recognition using video stimuli of a wide range of emotions and expression intensity variations. PMID:29293674
A Multimodal Emotion Detection System during Human-Robot Interaction
Alonso-Martín, Fernando; Malfaz, María; Sequeira, João; Gorostiza, Javier F.; Salichs, Miguel A.
2013-01-01
In this paper, a multimodal user-emotion detection system for social robots is presented. This system is intended to be used during human–robot interaction, and it is integrated as part of the overall interaction system of the robot: the Robotics Dialog System (RDS). Two modes are used to detect emotions: the voice and face expression analysis. In order to analyze the voice of the user, a new component has been developed: Gender and Emotion Voice Analysis (GEVA), which is written using the Chuck language. For emotion detection in facial expressions, the system, Gender and Emotion Facial Analysis (GEFA), has been also developed. This last system integrates two third-party solutions: Sophisticated High-speed Object Recognition Engine (SHORE) and Computer Expression Recognition Toolbox (CERT). Once these new components (GEVA and GEFA) give their results, a decision rule is applied in order to combine the information given by both of them. The result of this rule, the detected emotion, is integrated into the dialog system through communicative acts. Hence, each communicative act gives, among other things, the detected emotion of the user to the RDS so it can adapt its strategy in order to get a greater satisfaction degree during the human–robot dialog. Each of the new components, GEVA and GEFA, can also be used individually. Moreover, they are integrated with the robotic control platform ROS (Robot Operating System). Several experiments with real users were performed to determine the accuracy of each component and to set the final decision rule. The results obtained from applying this decision rule in these experiments show a high success rate in automatic user emotion recognition, improving the results given by the two information channels (audio and visual) separately. PMID:24240598
Textual emotion recognition for enhancing enterprise computing
NASA Astrophysics Data System (ADS)
Quan, Changqin; Ren, Fuji
2016-05-01
The growing interest in affective computing (AC) brings a lot of valuable research topics that can meet different application demands in enterprise systems. The present study explores a sub area of AC techniques - textual emotion recognition for enhancing enterprise computing. Multi-label emotion recognition in text is able to provide a more comprehensive understanding of emotions than single label emotion recognition. A representation of 'emotion state in text' is proposed to encompass the multidimensional emotions in text. It ensures the description in a formal way of the configurations of basic emotions as well as of the relations between them. Our method allows recognition of the emotions for the words bear indirect emotions, emotion ambiguity and multiple emotions. We further investigate the effect of word order for emotional expression by comparing the performances of bag-of-words model and sequence model for multi-label sentence emotion recognition. The experiments show that the classification results under sequence model are better than under bag-of-words model. And homogeneous Markov model showed promising results of multi-label sentence emotion recognition. This emotion recognition system is able to provide a convenient way to acquire valuable emotion information and to improve enterprise competitive ability in many aspects.
Quantitative Expression and Immunogenicity of MAGE-3 and -6 in Upper Aerodigestive Tract Cancer
Andrade Filho, Pedro A.; López-Albaitero, Andrés; Xi, Liqiang; Gooding, William; Godfrey, Tony; Ferris, Robert L.
2009-01-01
The MAGE antigens are frequently expressed cancer vaccine targets. However, quantitative analysis of MAGE expression in upper aero-digestive tract (UADT) tumor cells and its association with T cell recognition has not been performed, hindering the selection of appropriate candidates for MAGE specific immunotherapy. Using quantitative RT-PCR (QRT-PCR), we evaluated the expression of MAGE-3/6 in 65 UADT cancers, 48 normal samples from tumor matched sites and 7 HLA-A*0201+squamous cell carcinoma of the head and neck (SCCHN) cell lines. Expression results were confirmed using western blot. HLA-A*0201:MAGE-3(271–279) specific cytotoxic T lymphocytes (MAGE-CTL) from SCCHN patients and healthy donors showed that MAGE-3/6 expression was highly associated with CTL recognition in vitro. Based on MAGE-3/6 expression we could identify 31 (47%) of the 65 UADT tumors which appeared to express MAGE-3/6 at levels that correlated with efficient CTL recognition. To confirm that the level of MAGE-3 expression was responsible for CTL recognition, two MAGE-3/6 mRNAhigh SCCHN cell lines, PCI-13 and PCI-30, were subjected to MAGE-3/6 specific knockdown. RNAi–transfected cells showed that MAGE expression, and MAGE-CTL recognition, were significantly reduced. Furthermore, treatment of cells expressing low MAGE-3/6 mRNA with a demethylating agent, 5-aza-2'-deoxycytidine (DAC), increased the expression of MAGE-3/6 and CTL recognition. Thus, using QRT-PCR UADT cancers frequently express MAGE-3/6 at levels sufficient for CTL recognition, supporting the use of a QRT-PCR based assay for the selection of candidates likely to respond to MAGE-3/6 immunotherapy. Demethylating agents could increase the number of patients amenable for targeting epigenetically modified tumor antigens in vaccine trials. PMID:19610063
Quantitative expression and immunogenicity of MAGE-3 and -6 in upper aerodigestive tract cancer.
Filho, Pedro A Andrade; López-Albaitero, Andrés; Xi, Liqiang; Gooding, William; Godfrey, Tony; Ferris, Robert L
2009-10-15
The MAGE antigens are frequently expressed cancer vaccine targets. However, quantitative analysis of MAGE expression in upper aerodigestive tract (UADT) tumor cells and its association with T-cell recognition has not been performed, hindering the selection of appropriate candidates for MAGE-specific immunotherapy. Using quantitative RT-PCR (QRT-PCR), we evaluated the expression of MAGE-3/6 in 65 UADT cancers, 48 normal samples from tumor matched sites and 7 HLA-A*0201+ squamous cell carcinoma of the head and neck (SCCHN) cell lines. Expression results were confirmed using Western blot. HLA-A*0201:MAGE-3- (271-279) specific cytotoxic T lymphocytes (MAGE-CTL) from SCCHN patients and healthy donors showed that MAGE-3/6 expression was highly associated with CTL recognition in vitro. On the basis of the MAGE-3/6 expression, we could identify 31 (47%) of the 65 UADT tumors, which appeared to express MAGE-3/6 at levels that correlated with efficient CTL recognition. To confirm that the level of MAGE-3 expression was responsible for CTL recognition, 2 MAGE-3/6 mRNA(high) SCCHN cell lines, PCI-13 and PCI-30, were subjected to MAGE-3/6-specific knockdown. RNAi-transfected cells showed that MAGE expression and MAGE-CTL recognition were significantly reduced. Furthermore, treatment of cells expressing low MAGE-3/6 mRNA with a demethylating agent, 5-aza-2'-deoxycytidine (DAC), increased the expression of MAGE-3/6 and CTL recognition. Thus, using QRT-PCR UADT cancers frequently express MAGE-3/6 at levels sufficient for CTL recognition, supporting the use of a QRT-PCR-based assay for the selection of candidates likely to respond to MAGE-3/6 immunotherapy. Demethylating agents could increase the number of patients amenable for targeting epigenetically modified tumor antigens in vaccine trials.
Progesterone Impairs Social Recognition in Male Rats
Auger, Catherine J.
2012-01-01
The influence of progesterone in the brain and on the behavior of females is fairly well understood. However, less is known about the effect of progesterone in the male system. In male rats, receptors for progesterone are present in virtually all vasopressin (AVP) immunoreactive cells in the bed nucleus of the stria terminalis (BST) and the medial amygdala (MeA). This colocalization functions to regulate AVP expression, as progesterone and/or progestin receptors (PR)s suppresses AVP expression in these same extrahypothalamic regions in the brain. These data suggest that progesterone may influence AVP-dependant behavior. While AVP is implicated in numerous behavioral and physiological functions in rodents, AVP appears essential for social recognition of conspecifics. Therefore, we examined the effects of progesterone on social recognition. We report that progesterone plays an important role in modulating social recognition in the male brain, as progesterone treatment lead to a significant impairment of social recognition in male rats. Moreover, progesterone appears to act on PRs to impair social recognition, as progesterone impairment of social recognition is blocked by a PR antagonist, RU-486. Social recognition is also impaired by a specific progestin agonist, R5020. Interestingly, we show that progesterone does not interfere with either general memory or olfactory processes, suggesting that progesterone seems critically important to social recognition memory. These data provide strong evidence that physiological levels of progesterone can have an important impact on social behavior in male rats. PMID:22366506
Progesterone impairs social recognition in male rats.
Bychowski, Meaghan E; Auger, Catherine J
2012-04-01
The influence of progesterone in the brain and on the behavior of females is fairly well understood. However, less is known about the effect of progesterone in the male system. In male rats, receptors for progesterone are present in virtually all vasopressin (AVP) immunoreactive cells in the bed nucleus of the stria terminalis (BST) and the medial amygdala (MeA). This colocalization functions to regulate AVP expression, as progesterone and/or progestin receptors (PR)s suppress AVP expression in these same extrahypothalamic regions in the brain. These data suggest that progesterone may influence AVP-dependent behavior. While AVP is implicated in numerous behavioral and physiological functions in rodents, AVP appears essential for social recognition of conspecifics. Therefore, we examined the effects of progesterone on social recognition. We report that progesterone plays an important role in modulating social recognition in the male brain, as progesterone treatment leads to a significant impairment of social recognition in male rats. Moreover, progesterone appears to act on PRs to impair social recognition, as progesterone impairment of social recognition is blocked by a PR antagonist, RU-486. Social recognition is also impaired by a specific progestin agonist, R5020. Interestingly, we show that progesterone does not interfere with either general memory or olfactory processes, suggesting that progesterone seems critically important to social recognition memory. These data provide strong evidence that physiological levels of progesterone can have an important impact on social behavior in male rats. Copyright © 2012 Elsevier Inc. All rights reserved.
Ecdysone triggered PGRP-LC expression controls Drosophila innate immunity.
Rus, Florentina; Flatt, Thomas; Tong, Mei; Aggarwal, Kamna; Okuda, Kendi; Kleino, Anni; Yates, Elisabeth; Tatar, Marc; Silverman, Neal
2013-05-29
Throughout the animal kingdom, steroid hormones have been implicated in the defense against microbial infection, but how these systemic signals control immunity is unclear. Here, we show that the steroid hormone ecdysone controls the expression of the pattern recognition receptor PGRP-LC in Drosophila, thereby tightly regulating innate immune recognition and defense against bacterial infection. We identify a group of steroid-regulated transcription factors as well as two GATA transcription factors that act as repressors and activators of the immune response and are required for the proper hormonal control of PGRP-LC expression. Together, our results demonstrate that Drosophila use complex mechanisms to modulate innate immune responses, and identify a transcriptional hierarchy that integrates steroid signalling and immunity in animals.
Drapeau, Joanie; Gosselin, Nathalie; Peretz, Isabelle; McKerral, Michelle
2017-01-01
To assess emotion recognition from dynamic facial, vocal and musical expressions in sub-groups of adults with traumatic brain injuries (TBI) of different severities and identify possible common underlying mechanisms across domains. Forty-one adults participated in this study: 10 with moderate-severe TBI, nine with complicated mild TBI, 11 with uncomplicated mild TBI and 11 healthy controls, who were administered experimental (emotional recognition, valence-arousal) and control tasks (emotional and structural discrimination) for each domain. Recognition of fearful faces was significantly impaired in moderate-severe and in complicated mild TBI sub-groups, as compared to those with uncomplicated mild TBI and controls. Effect sizes were medium-large. Participants with lower GCS scores performed more poorly when recognizing fearful dynamic facial expressions. Emotion recognition from auditory domains was preserved following TBI, irrespective of severity. All groups performed equally on control tasks, indicating no perceptual disorders. Although emotional recognition from vocal and musical expressions was preserved, no correlation was found across auditory domains. This preliminary study may contribute to improving comprehension of emotional recognition following TBI. Future studies of larger samples could usefully include measures of functional impacts of recognition deficits for fearful facial expressions. These could help refine interventions for emotional recognition following a brain injury.
Dual Temporal Scale Convolutional Neural Network for Micro-Expression Recognition.
Peng, Min; Wang, Chongyang; Chen, Tong; Liu, Guangyuan; Fu, Xiaolan
2017-01-01
Facial micro-expression is a brief involuntary facial movement and can reveal the genuine emotion that people try to conceal. Traditional methods of spontaneous micro-expression recognition rely excessively on sophisticated hand-crafted feature design and the recognition rate is not high enough for its practical application. In this paper, we proposed a Dual Temporal Scale Convolutional Neural Network (DTSCNN) for spontaneous micro-expressions recognition. The DTSCNN is a two-stream network. Different of stream of DTSCNN is used to adapt to different frame rate of micro-expression video clips. Each stream of DSTCNN consists of independent shallow network for avoiding the overfitting problem. Meanwhile, we fed the networks with optical-flow sequences to ensure that the shallow networks can further acquire higher-level features. Experimental results on spontaneous micro-expression databases (CASME I/II) showed that our method can achieve a recognition rate almost 10% higher than what some state-of-the-art method can achieve.
Anodal tDCS targeting the right orbitofrontal cortex enhances facial expression recognition
Murphy, Jillian M.; Ridley, Nicole J.; Vercammen, Ans
2015-01-01
The orbitofrontal cortex (OFC) has been implicated in the capacity to accurately recognise facial expressions. The aim of the current study was to determine if anodal transcranial direct current stimulation (tDCS) targeting the right OFC in healthy adults would enhance facial expression recognition, compared with a sham condition. Across two counterbalanced sessions of tDCS (i.e. anodal and sham), 20 undergraduate participants (18 female) completed a facial expression labelling task comprising angry, disgusted, fearful, happy, sad and neutral expressions, and a control (social judgement) task comprising the same expressions. Responses on the labelling task were scored for accuracy, median reaction time and overall efficiency (i.e. combined accuracy and reaction time). Anodal tDCS targeting the right OFC enhanced facial expression recognition, reflected in greater efficiency and speed of recognition across emotions, relative to the sham condition. In contrast, there was no effect of tDCS to responses on the control task. This is the first study to demonstrate that anodal tDCS targeting the right OFC boosts facial expression recognition. This finding provides a solid foundation for future research to examine the efficacy of this technique as a means to treat facial expression recognition deficits, particularly in individuals with OFC damage or dysfunction. PMID:25971602
Development and validation of an Argentine set of facial expressions of emotion.
Vaiman, Marcelo; Wagner, Mónica Anna; Caicedo, Estefanía; Pereno, Germán Leandro
2017-02-01
Pictures of facial expressions of emotion are used in a wide range of experiments. The last decade has seen an increase in the number of studies presenting local sets of emotion stimuli. However, only a few existing sets contain pictures of Latin Americans, despite the growing attention emotion research is receiving in this region. Here we present the development and validation of the Universidad Nacional de Cordoba, Expresiones de Emociones Faciales (UNCEEF), a Facial Action Coding System (FACS)-verified set of pictures of Argentineans expressing the six basic emotions, plus neutral expressions. FACS scores, recognition rates, Hu scores, and discrimination indices are reported. Evidence of convergent validity was obtained using the Pictures of Facial Affect in an Argentine sample. However, recognition accuracy was greater for UNCEEF. The importance of local sets of emotion pictures is discussed.
RNA-binding proteins regulate the expression of the immune activating ligand MICB
Nachmani, Daphna; Gutschner, Tony; Reches, Adi
2014-01-01
The recognition of stress-induced ligands by the activating receptor NKG2D expressed on cytotoxic lymphocytes is crucial for the prevention and containment of various diseases and is also one of the best-studied examples of how danger is sensed by the immune system. Still, however, the mechanisms leading to the expression of the NKG2D ligands are far from being completely understood. Here, we use an unbiased and systematic RNA pull-down approach combined with mass spectrometry to identify six RNA-binding proteins (RBPs) that bind and regulate the expression of MICB, one of the major stress-induced ligands of NKG2D. We further demonstrate that at least two of the identified RBPs function during genotoxic stress. Our data provide insights into stress recognition and hopefully open new therapeutic venues. PMID:24924487
Maki, Yohko; Yoshida, Hiroshi; Yamaguchi, Tomoharu; Yamaguchi, Haruyasu
2013-01-01
Positivity recognition bias has been reported for facial expression as well as memory and visual stimuli in aged individuals, whereas emotional facial recognition in Alzheimer disease (AD) patients is controversial, with possible involvement of confounding factors such as deficits in spatial processing of non-emotional facial features and in verbal processing to express emotions. Thus, we examined whether recognition of positive facial expressions was preserved in AD patients, by adapting a new method that eliminated the influences of these confounding factors. Sensitivity of six basic facial expressions (happiness, sadness, surprise, anger, disgust, and fear) was evaluated in 12 outpatients with mild AD, 17 aged normal controls (ANC), and 25 young normal controls (YNC). To eliminate the factors related to non-emotional facial features, averaged faces were prepared as stimuli. To eliminate the factors related to verbal processing, the participants were required to match the images of stimulus and answer, avoiding the use of verbal labels. In recognition of happiness, there was no difference in sensitivity between YNC and ANC, and between ANC and AD patients. AD patients were less sensitive than ANC in recognition of sadness, surprise, and anger. ANC were less sensitive than YNC in recognition of surprise, anger, and disgust. Within the AD patient group, sensitivity of happiness was significantly higher than those of the other five expressions. In AD patient, recognition of happiness was relatively preserved; recognition of happiness was most sensitive and was preserved against the influences of age and disease.
EMOTION RECOGNITION OF VIRTUAL AGENTS FACIAL EXPRESSIONS: THE EFFECTS OF AGE AND EMOTION INTENSITY
Beer, Jenay M.; Fisk, Arthur D.; Rogers, Wendy A.
2014-01-01
People make determinations about the social characteristics of an agent (e.g., robot or virtual agent) by interpreting social cues displayed by the agent, such as facial expressions. Although a considerable amount of research has been conducted investigating age-related differences in emotion recognition of human faces (e.g., Sullivan, & Ruffman, 2004), the effect of age on emotion identification of virtual agent facial expressions has been largely unexplored. Age-related differences in emotion recognition of facial expressions are an important factor to consider in the design of agents that may assist older adults in a recreational or healthcare setting. The purpose of the current research was to investigate whether age-related differences in facial emotion recognition can extend to emotion-expressive virtual agents. Younger and older adults performed a recognition task with a virtual agent expressing six basic emotions. Larger age-related differences were expected for virtual agents displaying negative emotions, such as anger, sadness, and fear. In fact, the results indicated that older adults showed a decrease in emotion recognition accuracy for a virtual agent's emotions of anger, fear, and happiness. PMID:25552896
Impaired recognition of body expressions in the behavioral variant of frontotemporal dementia.
Van den Stock, Jan; De Winter, François-Laurent; de Gelder, Beatrice; Rangarajan, Janaki Raman; Cypers, Gert; Maes, Frederik; Sunaert, Stefan; Goffin, Karolien; Vandenberghe, Rik; Vandenbulcke, Mathieu
2015-08-01
Progressive deterioration of social cognition and emotion processing are core symptoms of the behavioral variant of frontotemporal dementia (bvFTD). Here we investigate whether bvFTD is also associated with impaired recognition of static (Experiment 1) and dynamic (Experiment 2) bodily expressions. In addition, we compared body expression processing with processing of static (Experiment 3) and dynamic (Experiment 4) facial expressions, as well as with face identity processing (Experiment 5). The results reveal that bvFTD is associated with impaired recognition of static and dynamic bodily and facial expressions, while identity processing was intact. No differential impairments were observed regarding motion (static vs. dynamic) or category (body vs. face). Within the bvFTD group, we observed a significant partial correlation between body and face expression recognition, when controlling for performance on the identity task. Voxel-Based Morphometry (VBM) analysis revealed that body emotion recognition was positively associated with gray matter volume in a region of the inferior frontal gyrus (pars orbitalis/triangularis). The results are in line with a supramodal emotion recognition deficit in bvFTD. Copyright © 2015 Elsevier Ltd. All rights reserved.
3D facial expression recognition using maximum relevance minimum redundancy geometrical features
NASA Astrophysics Data System (ADS)
Rabiu, Habibu; Saripan, M. Iqbal; Mashohor, Syamsiah; Marhaban, Mohd Hamiruce
2012-12-01
In recent years, facial expression recognition (FER) has become an attractive research area, which besides the fundamental challenges, it poses, finds application in areas, such as human-computer interaction, clinical psychology, lie detection, pain assessment, and neurology. Generally the approaches to FER consist of three main steps: face detection, feature extraction and expression recognition. The recognition accuracy of FER hinges immensely on the relevance of the selected features in representing the target expressions. In this article, we present a person and gender independent 3D facial expression recognition method, using maximum relevance minimum redundancy geometrical features. The aim is to detect a compact set of features that sufficiently represents the most discriminative features between the target classes. Multi-class one-against-one SVM classifier was employed to recognize the seven facial expressions; neutral, happy, sad, angry, fear, disgust, and surprise. The average recognition accuracy of 92.2% was recorded. Furthermore, inter database homogeneity was investigated between two independent databases the BU-3DFE and UPM-3DFE the results showed a strong homogeneity between the two databases.
A wearable device for emotional recognition using facial expression and physiological response.
Jangho Kwon; Da-Hye Kim; Wanjoo Park; Laehyun Kim
2016-08-01
This paper introduces a glasses-typed wearable system to detect user's emotions using facial expression and physiological responses. The system is designed to acquire facial expression through a built-in camera and physiological responses such as photoplethysmogram (PPG) and electrodermal activity (EDA) in unobtrusive way. We used video clips for induced emotions to test the system suitability in the experiment. The results showed a few meaningful properties that associate emotions with facial expressions and physiological responses captured by the developed wearable device. We expect that this wearable system with a built-in camera and physiological sensors may be a good solution to monitor user's emotional state in daily life.
Developing an automated speech-recognition telephone diabetes intervention.
Goldman, Roberta E; Sanchez-Hernandez, Maya; Ross-Degnan, Dennis; Piette, John D; Trinacty, Connie Mah; Simon, Steven R
2008-08-01
Many patients do not receive guideline-recommended care for diabetes and other chronic conditions. Automated speech-recognition telephone outreach to supplement in-person physician-patient communication may enhance patient care for chronic illness. We conducted this study to inform the development of an automated telephone outreach intervention for improving diabetes care among members of a large, not-for-profit health plan. In-depth telephone interviews with qualitative analysis. participants Individuals with diabetes (n=36) enrolled in a large regional health plan in the USA. Main outcome measure Patients' opinions about automated speech-recognition telephone technology. Patients who were recently diagnosed with diabetes and some with diabetes for a decade or more expressed basic informational needs. While most would prefer to speak with a live person rather than a computer-recorded voice, many felt that the automated system could successfully supplement the information they receive from their physicians and could serve as an integral part of their care. Patients suggested that such a system could provide specific dietary advice, information about diabetes and its self-care, a call-in menu of information topics, reminders about laboratory test results and appointments, tracking of personal laboratory results and feedback about their self-monitoring. While some patients expressed negative attitudes toward automated speech recognition telephone systems generally, most felt that a variety of functions of such a system could be beneficial to their diabetes care. In-depth interviews resulted in substantive input from health plan members for the design of an automated telephone outreach system to supplement in-person physician-patient communication in this population.
Zangara, Andrea; Blair, R J R; Curran, H Valerie
2002-08-01
Accumulating evidence from neuropsychological and neuroimaging research suggests that facial expressions are processed by at least partially separable neurocognitive systems. Recent evidence implies that the processing of different facial expressions may also be dissociable pharmacologically by GABAergic and noradrenergic compounds, although no study has directly compared the two types of drugs. The present study therefore directly compared the effects of a benzodiazepine with those of a beta-adrenergic blocker on the ability to recognise emotional expressions. A double-blind, independent group design was used with 45 volunteers to compare the effects of diazepam (15 mg) and metoprolol (50 mg) with matched placebo. Participants were presented with morphed facial expression stimuli and asked to identify which of the six basic emotions (sadness, happiness, anger, disgust, fear and surprise) were portrayed. Control measures of mood, pulse rate and word recall were also taken. Diazepam selectively impaired participants' ability to recognise expressions of both anger and fear but not other emotional expressions. Errors were mainly mistaking fear for surprise and disgust for anger. Metoprolol did not significantly affect facial expression recognition. These findings are interpreted as providing further support for the suggestion that there are dissociable systems responsible for processing emotional expressions. The results may have implications for understanding why 'paradoxical' aggression is sometimes elicited by benzodiazepines and for extending our psychological understanding of the anxiolytic effects of these drugs.
Micro-Expression Recognition Using Color Spaces.
Wang, Su-Jing; Yan, Wen-Jing; Li, Xiaobai; Zhao, Guoying; Zhou, Chun-Guang; Fu, Xiaolan; Yang, Minghao; Tao, Jianhua
2015-12-01
Micro-expressions are brief involuntary facial expressions that reveal genuine emotions and, thus, help detect lies. Because of their many promising applications, they have attracted the attention of researchers from various fields. Recent research reveals that two perceptual color spaces (CIELab and CIELuv) provide useful information for expression recognition. This paper is an extended version of our International Conference on Pattern Recognition paper, in which we propose a novel color space model, tensor independent color space (TICS), to help recognize micro-expressions. In this paper, we further show that CIELab and CIELuv are also helpful in recognizing micro-expressions, and we indicate why these three color spaces achieve better performance. A micro-expression color video clip is treated as a fourth-order tensor, i.e., a four-dimension array. The first two dimensions are the spatial information, the third is the temporal information, and the fourth is the color information. We transform the fourth dimension from RGB into TICS, in which the color components are as independent as possible. The combination of dynamic texture and independent color components achieves a higher accuracy than does that of RGB. In addition, we define a set of regions of interests (ROIs) based on the facial action coding system and calculated the dynamic texture histograms for each ROI. Experiments are conducted on two micro-expression databases, CASME and CASME 2, and the results show that the performances for TICS, CIELab, and CIELuv are better than those for RGB or gray.
Facial Emotion Recognition in Bipolar Disorder and Healthy Aging.
Altamura, Mario; Padalino, Flavia A; Stella, Eleonora; Balzotti, Angela; Bellomo, Antonello; Palumbo, Rocco; Di Domenico, Alberto; Mammarella, Nicola; Fairfield, Beth
2016-03-01
Emotional face recognition is impaired in bipolar disorder, but it is not clear whether this is specific for the illness. Here, we investigated how aging and bipolar disorder influence dynamic emotional face recognition. Twenty older adults, 16 bipolar patients, and 20 control subjects performed a dynamic affective facial recognition task and a subsequent rating task. Participants pressed a key as soon as they were able to discriminate whether the neutral face was assuming a happy or angry facial expression and then rated the intensity of each facial expression. Results showed that older adults recognized happy expressions faster, whereas bipolar patients recognized angry expressions faster. Furthermore, both groups rated emotional faces more intensely than did the control subjects. This study is one of the first to compare how aging and clinical conditions influence emotional facial recognition and underlines the need to consider the role of specific and common factors in emotional face recognition.
Comparison of emotion recognition from facial expression and music.
Gaspar, Tina; Labor, Marina; Jurić, Iva; Dumancić, Dijana; Ilakovac, Vesna; Heffer, Marija
2011-01-01
The recognition of basic emotions in everyday communication involves interpretation of different visual and auditory clues. The ability to recognize emotions is not clearly determined as their presentation is usually very short (micro expressions), whereas the recognition itself does not have to be a conscious process. We assumed that the recognition from facial expressions is selected over the recognition of emotions communicated through music. In order to compare the success rate in recognizing emotions presented as facial expressions or in classical music works we conducted a survey which included 90 elementary school and 87 high school students from Osijek (Croatia). The participants had to match 8 photographs of different emotions expressed on the face and 8 pieces of classical music works with 8 offered emotions. The recognition of emotions expressed through classical music pieces was significantly less successful than the recognition of emotional facial expressions. The high school students were significantly better at recognizing facial emotions than the elementary school students, whereas girls were better than boys. The success rate in recognizing emotions from music pieces was associated with higher grades in mathematics. Basic emotions are far better recognized if presented on human faces than in music, possibly because the understanding of facial emotions is one of the oldest communication skills in human society. Female advantage in emotion recognition was selected due to the necessity of their communication with the newborns during early development. The proficiency in recognizing emotional content of music and mathematical skills probably share some general cognitive skills like attention, memory and motivation. Music pieces were differently processed in brain than facial expressions and consequently, probably differently evaluated as relevant emotional clues.
Matsugu, Masakazu; Mori, Katsuhiko; Mitari, Yusuke; Kaneda, Yuji
2003-01-01
Reliable detection of ordinary facial expressions (e.g. smile) despite the variability among individuals as well as face appearance is an important step toward the realization of perceptual user interface with autonomous perception of persons. We describe a rule-based algorithm for robust facial expression recognition combined with robust face detection using a convolutional neural network. In this study, we address the problem of subject independence as well as translation, rotation, and scale invariance in the recognition of facial expression. The result shows reliable detection of smiles with recognition rate of 97.6% for 5600 still images of more than 10 subjects. The proposed algorithm demonstrated the ability to discriminate smiling from talking based on the saliency score obtained from voting visual cues. To the best of our knowledge, it is the first facial expression recognition model with the property of subject independence combined with robustness to variability in facial appearance.
Younger and Older Users’ Recognition of Virtual Agent Facial Expressions
Beer, Jenay M.; Smarr, Cory-Ann; Fisk, Arthur D.; Rogers, Wendy A.
2015-01-01
As technology advances, robots and virtual agents will be introduced into the home and healthcare settings to assist individuals, both young and old, with everyday living tasks. Understanding how users recognize an agent’s social cues is therefore imperative, especially in social interactions. Facial expression, in particular, is one of the most common non-verbal cues used to display and communicate emotion in on-screen agents (Cassell, Sullivan, Prevost, & Churchill, 2000). Age is important to consider because age-related differences in emotion recognition of human facial expression have been supported (Ruffman et al., 2008), with older adults showing a deficit for recognition of negative facial expressions. Previous work has shown that younger adults can effectively recognize facial emotions displayed by agents (Bartneck & Reichenbach, 2005; Courgeon et al. 2009; 2011; Breazeal, 2003); however, little research has compared in-depth younger and older adults’ ability to label a virtual agent’s facial emotions, an import consideration because social agents will be required to interact with users of varying ages. If such age-related differences exist for recognition of virtual agent facial expressions, we aim to understand if those age-related differences are influenced by the intensity of the emotion, dynamic formation of emotion (i.e., a neutral expression developing into an expression of emotion through motion), or the type of virtual character differing by human-likeness. Study 1 investigated the relationship between age-related differences, the implication of dynamic formation of emotion, and the role of emotion intensity in emotion recognition of the facial expressions of a virtual agent (iCat). Study 2 examined age-related differences in recognition expressed by three types of virtual characters differing by human-likeness (non-humanoid iCat, synthetic human, and human). Study 2 also investigated the role of configural and featural processing as a possible explanation for age-related differences in emotion recognition. First, our findings show age-related differences in the recognition of emotions expressed by a virtual agent, with older adults showing lower recognition for the emotions of anger, disgust, fear, happiness, sadness, and neutral. These age-related difference might be explained by older adults having difficulty discriminating similarity in configural arrangement of facial features for certain emotions; for example, older adults often mislabeled the similar emotions of fear as surprise. Second, our results did not provide evidence for the dynamic formation improving emotion recognition; but, in general, the intensity of the emotion improved recognition. Lastly, we learned that emotion recognition, for older and younger adults, differed by character type, from best to worst: human, synthetic human, and then iCat. Our findings provide guidance for design, as well as the development of a framework of age-related differences in emotion recognition. PMID:25705105
Facial Emotion Recognition and Expression in Parkinson’s Disease: An Emotional Mirror Mechanism?
Ricciardi, Lucia; Visco-Comandini, Federica; Erro, Roberto; Morgante, Francesca; Bologna, Matteo; Fasano, Alfonso; Ricciardi, Diego; Edwards, Mark J.; Kilner, James
2017-01-01
Background and aim Parkinson’s disease (PD) patients have impairment of facial expressivity (hypomimia) and difficulties in interpreting the emotional facial expressions produced by others, especially for aversive emotions. We aimed to evaluate the ability to produce facial emotional expressions and to recognize facial emotional expressions produced by others in a group of PD patients and a group of healthy participants in order to explore the relationship between these two abilities and any differences between the two groups of participants. Methods Twenty non-demented, non-depressed PD patients and twenty healthy participants (HC) matched for demographic characteristics were studied. The ability of recognizing emotional facial expressions was assessed with the Ekman 60-faces test (Emotion recognition task). Participants were video-recorded while posing facial expressions of 6 primary emotions (happiness, sadness, surprise, disgust, fear and anger). The most expressive pictures for each emotion were derived from the videos. Ten healthy raters were asked to look at the pictures displayed on a computer-screen in pseudo-random fashion and to identify the emotional label in a six-forced-choice response format (Emotion expressivity task). Reaction time (RT) and accuracy of responses were recorded. At the end of each trial the participant was asked to rate his/her confidence in his/her perceived accuracy of response. Results For emotion recognition, PD reported lower score than HC for Ekman total score (p<0.001), and for single emotions sub-scores happiness, fear, anger, sadness (p<0.01) and surprise (p = 0.02). In the facial emotion expressivity task, PD and HC significantly differed in the total score (p = 0.05) and in the sub-scores for happiness, sadness, anger (all p<0.001). RT and the level of confidence showed significant differences between PD and HC for the same emotions. There was a significant positive correlation between the emotion facial recognition and expressivity in both groups; the correlation was even stronger when ranking emotions from the best recognized to the worst (R = 0.75, p = 0.004). Conclusions PD patients showed difficulties in recognizing emotional facial expressions produced by others and in posing facial emotional expressions compared to healthy subjects. The linear correlation between recognition and expression in both experimental groups suggests that the two mechanisms share a common system, which could be deteriorated in patients with PD. These results open new clinical and rehabilitation perspectives. PMID:28068393
Dual Temporal Scale Convolutional Neural Network for Micro-Expression Recognition
Peng, Min; Wang, Chongyang; Chen, Tong; Liu, Guangyuan; Fu, Xiaolan
2017-01-01
Facial micro-expression is a brief involuntary facial movement and can reveal the genuine emotion that people try to conceal. Traditional methods of spontaneous micro-expression recognition rely excessively on sophisticated hand-crafted feature design and the recognition rate is not high enough for its practical application. In this paper, we proposed a Dual Temporal Scale Convolutional Neural Network (DTSCNN) for spontaneous micro-expressions recognition. The DTSCNN is a two-stream network. Different of stream of DTSCNN is used to adapt to different frame rate of micro-expression video clips. Each stream of DSTCNN consists of independent shallow network for avoiding the overfitting problem. Meanwhile, we fed the networks with optical-flow sequences to ensure that the shallow networks can further acquire higher-level features. Experimental results on spontaneous micro-expression databases (CASME I/II) showed that our method can achieve a recognition rate almost 10% higher than what some state-of-the-art method can achieve. PMID:29081753
Affective State Level Recognition in Naturalistic Facial and Vocal Expressions.
Meng, Hongying; Bianchi-Berthouze, Nadia
2014-03-01
Naturalistic affective expressions change at a rate much slower than the typical rate at which video or audio is recorded. This increases the probability that consecutive recorded instants of expressions represent the same affective content. In this paper, we exploit such a relationship to improve the recognition performance of continuous naturalistic affective expressions. Using datasets of naturalistic affective expressions (AVEC 2011 audio and video dataset, PAINFUL video dataset) continuously labeled over time and over different dimensions, we analyze the transitions between levels of those dimensions (e.g., transitions in pain intensity level). We use an information theory approach to show that the transitions occur very slowly and hence suggest modeling them as first-order Markov models. The dimension levels are considered to be the hidden states in the Hidden Markov Model (HMM) framework. Their discrete transition and emission matrices are trained by using the labels provided with the training set. The recognition problem is converted into a best path-finding problem to obtain the best hidden states sequence in HMMs. This is a key difference from previous use of HMMs as classifiers. Modeling of the transitions between dimension levels is integrated in a multistage approach, where the first level performs a mapping between the affective expression features and a soft decision value (e.g., an affective dimension level), and further classification stages are modeled as HMMs that refine that mapping by taking into account the temporal relationships between the output decision labels. The experimental results for each of the unimodal datasets show overall performance to be significantly above that of a standard classification system that does not take into account temporal relationships. In particular, the results on the AVEC 2011 audio dataset outperform all other systems presented at the international competition.
Ardizzi, Martina; Evangelista, Valentina; Ferroni, Francesca; Umiltà, Maria A.; Ravera, Roberto; Gallese, Vittorio
2017-01-01
One of the crucial features defining basic emotions and their prototypical facial expressions is their value for survival. Childhood traumatic experiences affect the effective recognition of facial expressions of negative emotions, normally allowing the recruitment of adequate behavioral responses to environmental threats. Specifically, anger becomes an extraordinarily salient stimulus unbalancing victims’ recognition of negative emotions. Despite the plethora of studies on this topic, to date, it is not clear whether this phenomenon reflects an overall response tendency toward anger recognition or a selective proneness to the salience of specific facial expressive cues of anger after trauma exposure. To address this issue, a group of underage Sierra Leonean Ebola virus disease survivors (mean age 15.40 years, SE 0.35; years of schooling 8.8 years, SE 0.46; 14 males) and a control group (mean age 14.55, SE 0.30; years of schooling 8.07 years, SE 0.30, 15 males) performed a forced-choice chimeric facial expressions recognition task. The chimeric facial expressions were obtained pairing upper and lower half faces of two different negative emotions (selected from anger, fear and sadness for a total of six different combinations). Overall, results showed that upper facial expressive cues were more salient than lower facial expressive cues. This priority was lost among Ebola virus disease survivors for the chimeric facial expressions of anger. In this case, differently from controls, Ebola virus disease survivors recognized anger regardless of the upper or lower position of the facial expressive cues of this emotion. The present results demonstrate that victims’ performance in the recognition of the facial expression of anger does not reflect an overall response tendency toward anger recognition, but rather the specific greater salience of facial expressive cues of anger. Furthermore, the present results show that traumatic experiences deeply modify the perceptual analysis of philogenetically old behavioral patterns like the facial expressions of emotions. PMID:28690565
Ardizzi, Martina; Evangelista, Valentina; Ferroni, Francesca; Umiltà, Maria A; Ravera, Roberto; Gallese, Vittorio
2017-01-01
One of the crucial features defining basic emotions and their prototypical facial expressions is their value for survival. Childhood traumatic experiences affect the effective recognition of facial expressions of negative emotions, normally allowing the recruitment of adequate behavioral responses to environmental threats. Specifically, anger becomes an extraordinarily salient stimulus unbalancing victims' recognition of negative emotions. Despite the plethora of studies on this topic, to date, it is not clear whether this phenomenon reflects an overall response tendency toward anger recognition or a selective proneness to the salience of specific facial expressive cues of anger after trauma exposure. To address this issue, a group of underage Sierra Leonean Ebola virus disease survivors (mean age 15.40 years, SE 0.35; years of schooling 8.8 years, SE 0.46; 14 males) and a control group (mean age 14.55, SE 0.30; years of schooling 8.07 years, SE 0.30, 15 males) performed a forced-choice chimeric facial expressions recognition task. The chimeric facial expressions were obtained pairing upper and lower half faces of two different negative emotions (selected from anger, fear and sadness for a total of six different combinations). Overall, results showed that upper facial expressive cues were more salient than lower facial expressive cues. This priority was lost among Ebola virus disease survivors for the chimeric facial expressions of anger. In this case, differently from controls, Ebola virus disease survivors recognized anger regardless of the upper or lower position of the facial expressive cues of this emotion. The present results demonstrate that victims' performance in the recognition of the facial expression of anger does not reflect an overall response tendency toward anger recognition, but rather the specific greater salience of facial expressive cues of anger. Furthermore, the present results show that traumatic experiences deeply modify the perceptual analysis of philogenetically old behavioral patterns like the facial expressions of emotions.
Domain Regeneration for Cross-Database Micro-Expression Recognition
NASA Astrophysics Data System (ADS)
Zong, Yuan; Zheng, Wenming; Huang, Xiaohua; Shi, Jingang; Cui, Zhen; Zhao, Guoying
2018-05-01
In this paper, we investigate the cross-database micro-expression recognition problem, where the training and testing samples are from two different micro-expression databases. Under this setting, the training and testing samples would have different feature distributions and hence the performance of most existing micro-expression recognition methods may decrease greatly. To solve this problem, we propose a simple yet effective method called Target Sample Re-Generator (TSRG) in this paper. By using TSRG, we are able to re-generate the samples from target micro-expression database and the re-generated target samples would share same or similar feature distributions with the original source samples. For this reason, we can then use the classifier learned based on the labeled source samples to accurately predict the micro-expression categories of the unlabeled target samples. To evaluate the performance of the proposed TSRG method, extensive cross-database micro-expression recognition experiments designed based on SMIC and CASME II databases are conducted. Compared with recent state-of-the-art cross-database emotion recognition methods, the proposed TSRG achieves more promising results.
Mapping correspondence between facial mimicry and emotion recognition in healthy subjects.
Ponari, Marta; Conson, Massimiliano; D'Amico, Nunzia Pina; Grossi, Dario; Trojano, Luigi
2012-12-01
We aimed at verifying the hypothesis that facial mimicry is causally and selectively involved in emotion recognition. For this purpose, in Experiment 1, we explored the effect of tonic contraction of muscles in upper or lower half of participants' face on their ability to recognize emotional facial expressions. We found that the "lower" manipulation specifically impaired recognition of happiness and disgust, the "upper" manipulation impaired recognition of anger, while both manipulations affected recognition of fear; recognition of surprise and sadness were not affected by either blocking manipulations. In Experiment 2, we verified whether emotion recognition is hampered by stimuli in which an upper or lower half-face showing an emotional expression is combined with a neutral half-face. We found that the neutral lower half-face interfered with recognition of happiness and disgust, whereas the neutral upper half impaired recognition of anger; recognition of fear and sadness was impaired by both manipulations, whereas recognition of surprise was not affected by either manipulation. Taken together, the present findings support simulation models of emotion recognition and provide insight into the role of mimicry in comprehension of others' emotional facial expressions. PsycINFO Database Record (c) 2012 APA, all rights reserved.
Hoffmann, Holger; Kessler, Henrik; Eppel, Tobias; Rukavina, Stefanie; Traue, Harald C
2010-11-01
Two experiments were conducted in order to investigate the effect of expression intensity on gender differences in the recognition of facial emotions. The first experiment compared recognition accuracy between female and male participants when emotional faces were shown with full-blown (100% emotional content) or subtle expressiveness (50%). In a second experiment more finely grained analyses were applied in order to measure recognition accuracy as a function of expression intensity (40%-100%). The results show that although women were more accurate than men in recognizing subtle facial displays of emotion, there was no difference between male and female participants when recognizing highly expressive stimuli. Copyright © 2010 Elsevier B.V. All rights reserved.
The look of fear and anger: facial maturity modulates recognition of fearful and angry expressions.
Sacco, Donald F; Hugenberg, Kurt
2009-02-01
The current series of studies provide converging evidence that facial expressions of fear and anger may have co-evolved to mimic mature and babyish faces in order to enhance their communicative signal. In Studies 1 and 2, fearful and angry facial expressions were manipulated to have enhanced babyish features (larger eyes) or enhanced mature features (smaller eyes) and in the context of a speeded categorization task in Study 1 and a visual noise paradigm in Study 2, results indicated that larger eyes facilitated the recognition of fearful facial expressions, while smaller eyes facilitated the recognition of angry facial expressions. Study 3 manipulated facial roundness, a stable structure that does not vary systematically with expressions, and found that congruency between maturity and expression (narrow face-anger; round face-fear) facilitated expression recognition accuracy. Results are discussed as representing a broad co-evolutionary relationship between facial maturity and fearful and angry facial expressions. (c) 2009 APA, all rights reserved
Transfer between pose and expression training in face recognition.
Chen, Wenfeng; Liu, Chang Hong
2009-02-01
Prior research has shown that recognition of unfamiliar faces is susceptible to image variations due to pose and expression changes. However, little is known about how these variations on a new face are learnt and handled. We aimed to investigate whether exposures to one type of variation facilitate recognition in the untrained variation. In Experiment 1, faces were trained in multiple or single pose but were tested with a new expression. In Experiment 2, faces were trained in multiple or single expression but were tested in a new pose. We found that higher level of exposure to pose information facilitated recognition of the trained face in a new expression. However, multiple-expression training failed to transfer to a new pose. The findings suggest that generalisation of pose training may be extended to different types of variation whereas generalisation of expression training is largely confined within the trained type of variation.
Emotional Faces in Context: Age Differences in Recognition Accuracy and Scanning Patterns
Noh, Soo Rim; Isaacowitz, Derek M.
2014-01-01
While age-related declines in facial expression recognition are well documented, previous research relied mostly on isolated faces devoid of context. We investigated the effects of context on age differences in recognition of facial emotions and in visual scanning patterns of emotional faces. While their eye movements were monitored, younger and older participants viewed facial expressions (i.e., anger, disgust) in contexts that were emotionally congruent, incongruent, or neutral to the facial expression to be identified. Both age groups had highest recognition rates of facial expressions in the congruent context, followed by the neutral context, and recognition rates in the incongruent context were worst. These context effects were more pronounced for older adults. Compared to younger adults, older adults exhibited a greater benefit from congruent contextual information, regardless of facial expression. Context also influenced the pattern of visual scanning characteristics of emotional faces in a similar manner across age groups. In addition, older adults initially attended more to context overall. Our data highlight the importance of considering the role of context in understanding emotion recognition in adulthood. PMID:23163713
Tardif, Carole; Lainé, France; Rodriguez, Mélissa; Gepner, Bruno
2007-09-01
This study examined the effects of slowing down presentation of facial expressions and their corresponding vocal sounds on facial expression recognition and facial and/or vocal imitation in children with autism. Twelve autistic children and twenty-four normal control children were presented with emotional and non-emotional facial expressions on CD-Rom, under audio or silent conditions, and under dynamic visual conditions (slowly, very slowly, at normal speed) plus a static control. Overall, children with autism showed lower performance in expression recognition and more induced facial-vocal imitation than controls. In the autistic group, facial expression recognition and induced facial-vocal imitation were significantly enhanced in slow conditions. Findings may give new perspectives for understanding and intervention for verbal and emotional perceptive and communicative impairments in autistic populations.
ERIC Educational Resources Information Center
Golouboff, Nathalie; Fiori, Nicole; Delalande, Olivier; Fohlen, Martine; Dellatolas, Georges; Jambaque, Isabelle
2008-01-01
The amygdala has been implicated in the recognition of facial emotions, especially fearful expressions, in adults with early-onset right temporal lobe epilepsy (TLE). The present study investigates the recognition of facial emotions in children and adolescents, 8-16 years old, with epilepsy. Twenty-nine subjects had TLE (13 right, 16 left) and…
Mapping the emotional face. How individual face parts contribute to successful emotion recognition.
Wegrzyn, Martin; Vogt, Maria; Kireclioglu, Berna; Schneider, Julia; Kissler, Johanna
2017-01-01
Which facial features allow human observers to successfully recognize expressions of emotion? While the eyes and mouth have been frequently shown to be of high importance, research on facial action units has made more precise predictions about the areas involved in displaying each emotion. The present research investigated on a fine-grained level, which physical features are most relied on when decoding facial expressions. In the experiment, individual faces expressing the basic emotions according to Ekman were hidden behind a mask of 48 tiles, which was sequentially uncovered. Participants were instructed to stop the sequence as soon as they recognized the facial expression and assign it the correct label. For each part of the face, its contribution to successful recognition was computed, allowing to visualize the importance of different face areas for each expression. Overall, observers were mostly relying on the eye and mouth regions when successfully recognizing an emotion. Furthermore, the difference in the importance of eyes and mouth allowed to group the expressions in a continuous space, ranging from sadness and fear (reliance on the eyes) to disgust and happiness (mouth). The face parts with highest diagnostic value for expression identification were typically located in areas corresponding to action units from the facial action coding system. A similarity analysis of the usefulness of different face parts for expression recognition demonstrated that faces cluster according to the emotion they express, rather than by low-level physical features. Also, expressions relying more on the eyes or mouth region were in close proximity in the constructed similarity space. These analyses help to better understand how human observers process expressions of emotion, by delineating the mapping from facial features to psychological representation.
Mapping the emotional face. How individual face parts contribute to successful emotion recognition
Wegrzyn, Martin; Vogt, Maria; Kireclioglu, Berna; Schneider, Julia; Kissler, Johanna
2017-01-01
Which facial features allow human observers to successfully recognize expressions of emotion? While the eyes and mouth have been frequently shown to be of high importance, research on facial action units has made more precise predictions about the areas involved in displaying each emotion. The present research investigated on a fine-grained level, which physical features are most relied on when decoding facial expressions. In the experiment, individual faces expressing the basic emotions according to Ekman were hidden behind a mask of 48 tiles, which was sequentially uncovered. Participants were instructed to stop the sequence as soon as they recognized the facial expression and assign it the correct label. For each part of the face, its contribution to successful recognition was computed, allowing to visualize the importance of different face areas for each expression. Overall, observers were mostly relying on the eye and mouth regions when successfully recognizing an emotion. Furthermore, the difference in the importance of eyes and mouth allowed to group the expressions in a continuous space, ranging from sadness and fear (reliance on the eyes) to disgust and happiness (mouth). The face parts with highest diagnostic value for expression identification were typically located in areas corresponding to action units from the facial action coding system. A similarity analysis of the usefulness of different face parts for expression recognition demonstrated that faces cluster according to the emotion they express, rather than by low-level physical features. Also, expressions relying more on the eyes or mouth region were in close proximity in the constructed similarity space. These analyses help to better understand how human observers process expressions of emotion, by delineating the mapping from facial features to psychological representation. PMID:28493921
Novel dynamic Bayesian networks for facial action element recognition and understanding
NASA Astrophysics Data System (ADS)
Zhao, Wei; Park, Jeong-Seon; Choi, Dong-You; Lee, Sang-Woong
2011-12-01
In daily life, language is an important tool of communication between people. Besides language, facial action can also provide a great amount of information. Therefore, facial action recognition has become a popular research topic in the field of human-computer interaction (HCI). However, facial action recognition is quite a challenging task due to its complexity. In a literal sense, there are thousands of facial muscular movements, many of which have very subtle differences. Moreover, muscular movements always occur simultaneously when the pose is changed. To address this problem, we first build a fully automatic facial points detection system based on a local Gabor filter bank and principal component analysis. Then, novel dynamic Bayesian networks are proposed to perform facial action recognition using the junction tree algorithm over a limited number of feature points. In order to evaluate the proposed method, we have used the Korean face database for model training. For testing, we used the CUbiC FacePix, facial expressions and emotion database, Japanese female facial expression database, and our own database. Our experimental results clearly demonstrate the feasibility of the proposed approach.
Recognition of face identity and emotion in expressive specific language impairment.
Merkenschlager, A; Amorosa, H; Kiefl, H; Martinius, J
2012-01-01
To study face and emotion recognition in children with mostly expressive specific language impairment (SLI-E). A test movie to study perception and recognition of faces and mimic-gestural expression was applied to 24 children diagnosed as suffering from SLI-E and an age-matched control group of normally developing children. Compared to a normal control group, the SLI-E children scored significantly worse in both the face and expression recognition tasks with a preponderant effect on emotion recognition. The performance of the SLI-E group could not be explained by reduced attention during the test session. We conclude that SLI-E is associated with a deficiency in decoding non-verbal emotional facial and gestural information, which might lead to profound and persistent problems in social interaction and development. Copyright © 2012 S. Karger AG, Basel.
A Rapid and Quantitative Recombinase Activity Assay
USDA-ARS?s Scientific Manuscript database
We present here a comparison between the recombinase systems FLP-FRT and Cre-loxP. A transient excision based dual luciferase expression assay is used for its rapid and repeatable nature. The detection system was designed within an intron to remove the remaining recombinase recognition site and no...
Pham, Tuyen Danh; Park, Young Ho; Nguyen, Dat Tien; Kwon, Seung Yong; Park, Kang Ryoung
2015-01-01
Biometrics is a technology that enables an individual person to be identified based on human physiological and behavioral characteristics. Among biometrics technologies, face recognition has been widely used because of its advantages in terms of convenience and non-contact operation. However, its performance is affected by factors such as variation in the illumination, facial expression, and head pose. Therefore, fingerprint and iris recognitions are preferred alternatives. However, the performance of the former can be adversely affected by the skin condition, including scarring and dryness. In addition, the latter has the disadvantages of high cost, large system size, and inconvenience to the user, who has to align their eyes with the iris camera. In an attempt to overcome these problems, finger-vein recognition has been vigorously researched, but an analysis of its accuracies according to various factors has not received much attention. Therefore, we propose a nonintrusive finger-vein recognition system using a near infrared (NIR) image sensor and analyze its accuracies considering various factors. The experimental results obtained with three databases showed that our system can be operated in real applications with high accuracy; and the dissimilarity of the finger-veins of different people is larger than that of the finger types and hands. PMID:26184214
Pham, Tuyen Danh; Park, Young Ho; Nguyen, Dat Tien; Kwon, Seung Yong; Park, Kang Ryoung
2015-07-13
Biometrics is a technology that enables an individual person to be identified based on human physiological and behavioral characteristics. Among biometrics technologies, face recognition has been widely used because of its advantages in terms of convenience and non-contact operation. However, its performance is affected by factors such as variation in the illumination, facial expression, and head pose. Therefore, fingerprint and iris recognitions are preferred alternatives. However, the performance of the former can be adversely affected by the skin condition, including scarring and dryness. In addition, the latter has the disadvantages of high cost, large system size, and inconvenience to the user, who has to align their eyes with the iris camera. In an attempt to overcome these problems, finger-vein recognition has been vigorously researched, but an analysis of its accuracies according to various factors has not received much attention. Therefore, we propose a nonintrusive finger-vein recognition system using a near infrared (NIR) image sensor and analyze its accuracies considering various factors. The experimental results obtained with three databases showed that our system can be operated in real applications with high accuracy; and the dissimilarity of the finger-veins of different people is larger than that of the finger types and hands.
Identifying and detecting facial expressions of emotion in peripheral vision.
Smith, Fraser W; Rossit, Stephanie
2018-01-01
Facial expressions of emotion are signals of high biological value. Whilst recognition of facial expressions has been much studied in central vision, the ability to perceive these signals in peripheral vision has only seen limited research to date, despite the potential adaptive advantages of such perception. In the present experiment, we investigate facial expression recognition and detection performance for each of the basic emotions (plus neutral) at up to 30 degrees of eccentricity. We demonstrate, as expected, a decrease in recognition and detection performance with increasing eccentricity, with happiness and surprised being the best recognized expressions in peripheral vision. In detection however, while happiness and surprised are still well detected, fear is also a well detected expression. We show that fear is a better detected than recognized expression. Our results demonstrate that task constraints shape the perception of expression in peripheral vision and provide novel evidence that detection and recognition rely on partially separate underlying mechanisms, with the latter more dependent on the higher spatial frequency content of the face stimulus.
Identifying and detecting facial expressions of emotion in peripheral vision
Rossit, Stephanie
2018-01-01
Facial expressions of emotion are signals of high biological value. Whilst recognition of facial expressions has been much studied in central vision, the ability to perceive these signals in peripheral vision has only seen limited research to date, despite the potential adaptive advantages of such perception. In the present experiment, we investigate facial expression recognition and detection performance for each of the basic emotions (plus neutral) at up to 30 degrees of eccentricity. We demonstrate, as expected, a decrease in recognition and detection performance with increasing eccentricity, with happiness and surprised being the best recognized expressions in peripheral vision. In detection however, while happiness and surprised are still well detected, fear is also a well detected expression. We show that fear is a better detected than recognized expression. Our results demonstrate that task constraints shape the perception of expression in peripheral vision and provide novel evidence that detection and recognition rely on partially separate underlying mechanisms, with the latter more dependent on the higher spatial frequency content of the face stimulus. PMID:29847562
Four-Channel Biosignal Analysis and Feature Extraction for Automatic Emotion Recognition
NASA Astrophysics Data System (ADS)
Kim, Jonghwa; André, Elisabeth
This paper investigates the potential of physiological signals as a reliable channel for automatic recognition of user's emotial state. For the emotion recognition, little attention has been paid so far to physiological signals compared to audio-visual emotion channels such as facial expression or speech. All essential stages of automatic recognition system using biosignals are discussed, from recording physiological dataset up to feature-based multiclass classification. Four-channel biosensors are used to measure electromyogram, electrocardiogram, skin conductivity and respiration changes. A wide range of physiological features from various analysis domains, including time/frequency, entropy, geometric analysis, subband spectra, multiscale entropy, etc., is proposed in order to search the best emotion-relevant features and to correlate them with emotional states. The best features extracted are specified in detail and their effectiveness is proven by emotion recognition results.
[Developmental change in facial recognition by premature infants during infancy].
Konishi, Yukihiko; Kusaka, Takashi; Nishida, Tomoko; Isobe, Kenichi; Itoh, Susumu
2014-09-01
Premature infants are thought to be at increased risk for developmental disorders. We evaluated facial recognition by premature infants during early infancy, as this ability has been reported to be impaired commonly in developmentally disabled children. In premature infants and full-term infants at the age of 4 months (4 corrected months for premature infants), visual behaviors while performing facial recognition tasks were determined and analyzed using an eye-tracking system (Tobii T60 manufactured by Tobii Technologics, Sweden). Both types of infants had a preference towards normal facial expressions; however, no preference towards the upper face was observed in premature infants. Our study suggests that facial recognition ability in premature infants may develop differently from that in full-term infants.
Static facial expression recognition with convolution neural networks
NASA Astrophysics Data System (ADS)
Zhang, Feng; Chen, Zhong; Ouyang, Chao; Zhang, Yifei
2018-03-01
Facial expression recognition is a currently active research topic in the fields of computer vision, pattern recognition and artificial intelligence. In this paper, we have developed a convolutional neural networks (CNN) for classifying human emotions from static facial expression into one of the seven facial emotion categories. We pre-train our CNN model on the combined FER2013 dataset formed by train, validation and test set and fine-tune on the extended Cohn-Kanade database. In order to reduce the overfitting of the models, we utilized different techniques including dropout and batch normalization in addition to data augmentation. According to the experimental result, our CNN model has excellent classification performance and robustness for facial expression recognition.
Thonse, Umesh; Behere, Rishikesh V; Praharaj, Samir Kumar; Sharma, Podila Sathya Venkata Narasimha
2018-06-01
Facial emotion recognition deficits have been consistently demonstrated in patients with severe mental disorders. Expressed emotion is found to be an important predictor of relapse. However, the relationship between facial emotion recognition abilities and expressed emotions and its influence on socio-occupational functioning in schizophrenia versus bipolar disorder has not been studied. In this study we examined 91 patients with schizophrenia and 71 with bipolar disorder for psychopathology, socio occupational functioning and emotion recognition abilities. Primary caregivers of 62 patients with schizophrenia and 49 with bipolar disorder were assessed on Family Attitude Questionnaire to assess their expressed emotions. Patients of schizophrenia and bipolar disorder performed similarly on the emotion recognition task. Patients with schizophrenia group experienced higher critical comments and had a poorer socio-occupational functioning as compared to patients with bipolar disorder. Poorer socio-occupational functioning in patients with schizophrenia was significantly associated with greater dissatisfaction in their caregivers. In patients with bipolar disorder, poorer emotion recognition scores significantly correlated with poorer adaptive living skills and greater hostility and dissatisfaction in their caregivers. The findings of our study suggest that emotion recognition abilities in patients with bipolar disorder are associated with negative expressed emotions leading to problems in adaptive living skills. Copyright © 2018 Elsevier B.V. All rights reserved.
Activation of Supraoptic Oxytocin Neurons by Secretin Facilitates Social Recognition.
Takayanagi, Yuki; Yoshida, Masahide; Takashima, Akihide; Takanami, Keiko; Yoshida, Shoma; Nishimori, Katsuhiko; Nishijima, Ichiko; Sakamoto, Hirotaka; Yamagata, Takanori; Onaka, Tatsushi
2017-02-01
Social recognition underlies social behavior in animals, and patients with psychiatric disorders associated with social deficits show abnormalities in social recognition. Oxytocin is implicated in social behavior and has received attention as an effective treatment for sociobehavioral deficits. Secretin receptor-deficient mice show deficits in social behavior. The relationship between oxytocin and secretin concerning social behavior remains to be determined. Expression of c-Fos in oxytocin neurons and release of oxytocin from their dendrites after secretin application were investigated. Social recognition was examined after intracerebroventricular or local injection of secretin, oxytocin, or an oxytocin receptor antagonist in rats, oxytocin receptor-deficient mice, and secretin receptor-deficient mice. Electron and light microscopic immunohistochemical analysis was also performed to determine whether oxytocin neurons extend their dendrites into the medial amygdala. Supraoptic oxytocin neurons expressed the secretin receptor. Secretin activated supraoptic oxytocin neurons and facilitated oxytocin release from dendrites. Secretin increased acquisition of social recognition in an oxytocin receptor-dependent manner. Local application of secretin into the supraoptic nucleus facilitated social recognition, and this facilitation was blocked by an oxytocin receptor antagonist injected into, but not outside of, the medial amygdala. In the medial amygdala, dendrite-like thick oxytocin processes were found to extend from the supraoptic nucleus. Furthermore, oxytocin treatment restored deficits of social recognition in secretin receptor-deficient mice. The results of our study demonstrate that secretin-induced dendritic oxytocin release from supraoptic neurons enhances social recognition. The newly defined secretin-oxytocin system may lead to a possible treatment for social deficits. Copyright © 2016 Society of Biological Psychiatry. Published by Elsevier Inc. All rights reserved.
Face to face: blocking facial mimicry can selectively impair recognition of emotional expressions.
Oberman, Lindsay M; Winkielman, Piotr; Ramachandran, Vilayanur S
2007-01-01
People spontaneously mimic a variety of behaviors, including emotional facial expressions. Embodied cognition theories suggest that mimicry reflects internal simulation of perceived emotion in order to facilitate its understanding. If so, blocking facial mimicry should impair recognition of expressions, especially of emotions that are simulated using facial musculature. The current research tested this hypothesis using four expressions (happy, disgust, fear, and sad) and two mimicry-interfering manipulations (1) biting on a pen and (2) chewing gum, as well as two control conditions. Experiment 1 used electromyography over cheek, mouth, and nose regions. The bite manipulation consistently activated assessed muscles, whereas the chew manipulation activated muscles only intermittently. Further, expressing happiness generated most facial action. Experiment 2 found that the bite manipulation interfered most with recognition of happiness. These findings suggest that facial mimicry differentially contributes to recognition of specific facial expressions, thus allowing for more refined predictions from embodied cognition theories.
[Recognition of facial expression of emotions in Parkinson's disease: a theoretical review].
Alonso-Recio, L; Serrano-Rodriguez, J M; Carvajal-Molina, F; Loeches-Alonso, A; Martin-Plasencia, P
2012-04-16
Emotional facial expression is a basic guide during social interaction and, therefore, alterations in their expression or recognition are important limitations for communication. To examine facial expression recognition abilities and their possible impairment in Parkinson's disease. First, we review the studies on this topic which have not found entirely similar results. Second, we analyze the factors that may explain these discrepancies and, in particular, as third objective, we consider the relationship between emotional recognition problems and cognitive impairment associated with the disease. Finally, we propose alternatives strategies for the development of studies that could clarify the state of these abilities in Parkinson's disease. Most studies suggest deficits in facial expression recognition, especially in those with negative emotional content. However, it is possible that these alterations are related to those that also appear in the course of the disease in other perceptual and executive processes. To advance in this issue, we consider necessary to design emotional recognition studies implicating differentially the executive or visuospatial processes, and/or contrasting cognitive abilities with facial expressions and non emotional stimuli. The precision of the status of these abilities, as well as increase our knowledge of the functional consequences of the characteristic brain damage in the disease, may indicate if we should pay special attention in their rehabilitation inside the programs implemented.
Lodder, Gerine M A; Scholte, Ron H J; Goossens, Luc; Engels, Rutger C M E; Verhagen, Maaike
2016-02-01
Based on the belongingness regulation theory (Gardner et al., 2005, Pers. Soc. Psychol. Bull., 31, 1549), this study focuses on the relationship between loneliness and social monitoring. Specifically, we examined whether loneliness relates to performance on three emotion recognition tasks and whether lonely individuals show increased gazing towards their conversation partner's faces in a real-life conversation. Study 1 examined 170 college students (Mage = 19.26; SD = 1.21) who completed an emotion recognition task with dynamic stimuli (morph task) and a micro(-emotion) expression recognition task. Study 2 examined 130 college students (Mage = 19.33; SD = 2.00) who completed the Reading the Mind in the Eyes Test and who had a conversation with an unfamiliar peer while their gaze direction was videotaped. In both studies, loneliness was measured using the UCLA Loneliness Scale version 3 (Russell, 1996, J. Pers. Assess., 66, 20). The results showed that loneliness was unrelated to emotion recognition on all emotion recognition tasks, but that it was related to increased gaze towards their conversation partner's faces. Implications for the belongingness regulation system of lonely individuals are discussed. © 2015 The British Psychological Society.
Dissociation between recognition and detection advantage for facial expressions: a meta-analysis.
Nummenmaa, Lauri; Calvo, Manuel G
2015-04-01
Happy facial expressions are recognized faster and more accurately than other expressions in categorization tasks, whereas detection in visual search tasks is widely believed to be faster for angry than happy faces. We used meta-analytic techniques for resolving this categorization versus detection advantage discrepancy for positive versus negative facial expressions. Effect sizes were computed on the basis of the r statistic for a total of 34 recognition studies with 3,561 participants and 37 visual search studies with 2,455 participants, yielding a total of 41 effect sizes for recognition accuracy, 25 for recognition speed, and 125 for visual search speed. Random effects meta-analysis was conducted to estimate effect sizes at population level. For recognition tasks, an advantage in recognition accuracy and speed for happy expressions was found for all stimulus types. In contrast, for visual search tasks, moderator analysis revealed that a happy face detection advantage was restricted to photographic faces, whereas a clear angry face advantage was found for schematic and "smiley" faces. Robust detection advantage for nonhappy faces was observed even when stimulus emotionality was distorted by inversion or rearrangement of the facial features, suggesting that visual features primarily drive the search. We conclude that the recognition advantage for happy faces is a genuine phenomenon related to processing of facial expression category and affective valence. In contrast, detection advantages toward either happy (photographic stimuli) or nonhappy (schematic) faces is contingent on visual stimulus features rather than facial expression, and may not involve categorical or affective processing. (c) 2015 APA, all rights reserved).
Contributions of feature shapes and surface cues to the recognition of facial expressions.
Sormaz, Mladen; Young, Andrew W; Andrews, Timothy J
2016-10-01
Theoretical accounts of face processing often emphasise feature shapes as the primary visual cue to the recognition of facial expressions. However, changes in facial expression also affect the surface properties of the face. In this study, we investigated whether this surface information can also be used in the recognition of facial expression. First, participants identified facial expressions (fear, anger, disgust, sadness, happiness) from images that were manipulated such that they varied mainly in shape or mainly in surface properties. We found that the categorization of facial expression is possible in either type of image, but that different expressions are relatively dependent on surface or shape properties. Next, we investigated the relative contributions of shape and surface information to the categorization of facial expressions. This employed a complementary method that involved combining the surface properties of one expression with the shape properties from a different expression. Our results showed that the categorization of facial expressions in these hybrid images was equally dependent on the surface and shape properties of the image. Together, these findings provide a direct demonstration that both feature shape and surface information make significant contributions to the recognition of facial expressions. Copyright © 2016 Elsevier Ltd. All rights reserved.
NASA Astrophysics Data System (ADS)
Zhang, Jinmai; Luo, Huajie; Liu, Hao; Ye, Wei; Luo, Ray; Chen, Hai-Feng
2016-04-01
Histone modification plays a key role in gene regulation and gene expression. TRIM24 as a histone reader can recognize histone modification. However the specific recognition mechanism between TRIM24 and histone modification is unsolved. Here, systems biology method of dynamics correlation network based on molecular dynamics simulation was used to answer the question. Our network analysis shows that the dynamics correlation network of H3K23ac is distinctly different from that of wild type and other modifications. A hypothesis of “synergistic modification induced recognition” is then proposed to link histone modification and TRIM24 binding. These observations were further confirmed from community analysis of networks with mutation and network perturbation. Finally, a possible recognition pathway is also identified based on the shortest path search for H3K23ac. Significant difference of recognition pathway was found among different systems due to methylation and acetylation modifications. The analysis presented here and other studies show that the dynamic network-based analysis might be a useful general strategy to study the biology of protein post-translational modification and associated recognition.
Muecas: A Multi-Sensor Robotic Head for Affective Human Robot Interaction and Imitation
Cid, Felipe; Moreno, Jose; Bustos, Pablo; Núñez, Pedro
2014-01-01
This paper presents a multi-sensor humanoid robotic head for human robot interaction. The design of the robotic head, Muecas, is based on ongoing research on the mechanisms of perception and imitation of human expressions and emotions. These mechanisms allow direct interaction between the robot and its human companion through the different natural language modalities: speech, body language and facial expressions. The robotic head has 12 degrees of freedom, in a human-like configuration, including eyes, eyebrows, mouth and neck, and has been designed and built entirely by IADeX (Engineering, Automation and Design of Extremadura) and RoboLab. A detailed description of its kinematics is provided along with the design of the most complex controllers. Muecas can be directly controlled by FACS (Facial Action Coding System), the de facto standard for facial expression recognition and synthesis. This feature facilitates its use by third party platforms and encourages the development of imitation and of goal-based systems. Imitation systems learn from the user, while goal-based ones use planning techniques to drive the user towards a final desired state. To show the flexibility and reliability of the robotic head, the paper presents a software architecture that is able to detect, recognize, classify and generate facial expressions in real time using FACS. This system has been implemented using the robotics framework, RoboComp, which provides hardware-independent access to the sensors in the head. Finally, the paper presents experimental results showing the real-time functioning of the whole system, including recognition and imitation of human facial expressions. PMID:24787636
Pombo, Marina A; Zheng, Yi; Fernandez-Pozo, Noe; Dunham, Diane M; Fei, Zhangjun; Martin, Gregory B
2014-01-01
Plants have two related immune systems to defend themselves against pathogen attack. Initially,pattern-triggered immunity is activated upon recognition of microbe-associated molecular patterns by pattern recognition receptors. Pathogenic bacteria deliver effector proteins into the plant cell that interfere with this immune response and promote disease. However, some plants express resistance proteins that detect the presence of specific effectors leading to a robust defense response referred to as effector-triggered immunity. The interaction of tomato with Pseudomonas syringae pv. tomato is an established model system for understanding the molecular basis of these plant immune responses. We apply high-throughput RNA sequencing to this pathosystem to identify genes whose expression changes specifically during pattern-triggered or effector-triggered immunity. We then develop reporter genes for each of these responses that will enable characterization of the host response to the large collection of P. s. pv. tomato strains that express different combinations of effectors. Virus-induced gene silencing of 30 of the effector-triggered immunity-specific genes identifies Epk1 which encodes a predicted protein kinase from a family previously unknown to be involved in immunity. Knocked-down expression of Epk1 compromises effector-triggered immunity triggered by three bacterial effectors but not by effectors from non-bacterial pathogens. Epistasis experiments indicate that Epk1 acts upstream of effector-triggered immunity-associated MAP kinase signaling. Using RNA-seq technology we identify genes involved in specific immune responses. A functional genomics screen led to the discovery of Epk1, a novel predicted protein kinase required for plant defense activation upon recognition of three different bacterial effectors.
Not on the Face Alone: Perception of Contextualized Face Expressions in Huntington's Disease
ERIC Educational Resources Information Center
Aviezer, Hillel; Bentin, Shlomo; Hassin, Ran R.; Meschino, Wendy S.; Kennedy, Jeanne; Grewal, Sonya; Esmail, Sherali; Cohen, Sharon; Moscovitch, Morris
2009-01-01
Numerous studies have demonstrated that Huntington's disease mutation-carriers have deficient explicit recognition of isolated facial expressions. There are no studies, however, which have investigated the recognition of facial expressions embedded within an emotional body and scene context. Real life facial expressions are typically embedded in…
Wang, Yamin; Fu, Xiaolan; Johnston, Robert A.; Yan, Zheng
2013-01-01
Using Garner’s speeded classification task existing studies demonstrated an asymmetric interference in the recognition of facial identity and facial expression. It seems that expression is hard to interfere with identity recognition. However, discriminability of identity and expression, a potential confounding variable, had not been carefully examined in existing studies. In current work, we manipulated discriminability of identity and expression by matching facial shape (long or round) in identity and matching mouth (opened or closed) in facial expression. Garner interference was found either from identity to expression (Experiment 1) or from expression to identity (Experiment 2). Interference was also found in both directions (Experiment 3) or in neither direction (Experiment 4). The results support that Garner interference tends to occur under condition of low discriminability of relevant dimension regardless of facial property. Our findings indicate that Garner interference is not necessarily related to interdependent processing in recognition of facial identity and expression. The findings also suggest that discriminability as a mediating factor should be carefully controlled in future research. PMID:24391609
Age-related differences in emotion recognition ability: a cross-sectional study.
Mill, Aire; Allik, Jüri; Realo, Anu; Valk, Raivo
2009-10-01
Experimental studies indicate that recognition of emotions, particularly negative emotions, decreases with age. However, there is no consensus at which age the decrease in emotion recognition begins, how selective this is to negative emotions, and whether this applies to both facial and vocal expression. In the current cross-sectional study, 607 participants ranging in age from 18 to 84 years (mean age = 32.6 +/- 14.9 years) were asked to recognize emotions expressed either facially or vocally. In general, older participants were found to be less accurate at recognizing emotions, with the most distinctive age difference pertaining to a certain group of negative emotions. Both modalities revealed an age-related decline in the recognition of sadness and -- to a lesser degree -- anger, starting at about 30 years of age. Although age-related differences in the recognition of expression of emotion were not mediated by personality traits, 2 of the Big 5 traits, openness and conscientiousness, made an independent contribution to emotion-recognition performance. Implications of age-related differences in facial and vocal emotion expression and early onset of the selective decrease in emotion recognition are discussed in terms of previous findings and relevant theoretical models.
Mathematical formula recognition using graph grammar
NASA Astrophysics Data System (ADS)
Lavirotte, Stephane; Pottier, Loic
1998-04-01
This paper describes current results of Ofr, a system for extracting and understanding mathematical expressions in documents. Such a tool could be really useful to be able to re-use knowledge in scientific books which are not available in electronic form. We currently also study use of this system for direct input of formulas with a graphical tablet for computer algebra system softwares. Existing solutions for mathematical recognition have problems to analyze 2D expressions like vectors and matrices. This is because they often try to use extended classical grammar to analyze formulas, relatively to baseline. But a lot of mathematical notations do not respect rules for such a parsing and that is the reason why they fail to extend text parsing technic. We investigate graph grammar and graph rewriting as a solution to recognize 2D mathematical notations. Graph grammar provide a powerful formalism to describe structural manipulations of multi-dimensional data. The main two problems to solve are ambiguities between rules of grammar and construction of graph.
Ehrhardt, Anja; Xu, Hui; Huang, Zan; Engler, Jeffrey A; Kay, Mark A
2005-05-01
In this study we performed a head-to-head comparison of the integrase phiC31 derived from a Streptomyces phage and the Sleeping Beauty (SB) transposase, a member of the TC1/mariner superfamily of transposable elements. Mouse liver was cotransfused with a vector containing our most robust human coagulation factor IX expression cassette and the appropriate recombinase recognition site and either a phiC31- or a SB transposase-expressing vector. To analyze transgene persistence and to prove somatic integration in vivo we induced cell cycling of mouse hepatocytes and found that the transgene expression levels dropped by only 16 to 21% and 56 to 66% in mice that received phiC31 and SB, respectively. Notably, no difference in the toxicity profile was detected in mice treated with either recombinase. Moreover we observed that with the integrase-mediated gene transfer, transgene expression levels were dependent on the remaining noncoding vector sequences, which also integrate into the host genome. Further analyses of a hot spot of integration after phiC31-mediated integration revealed small chromosomal deletions at the target site and that the recombination process was not dependent on the orientation in which the phiC31 recognition site attached to the pseudo-recognition sites in the host genome. Coupled together with ongoing improvements in both systems this study suggests that both nonviral vector systems will have important roles in achieving stable gene transfer in vivo.
Influence of Emotional Facial Expressions on 3-5-Year-Olds' Face Recognition
ERIC Educational Resources Information Center
Freitag, Claudia; Schwarzer, Gudrun
2011-01-01
Three experiments examined 3- and 5-year-olds' recognition of faces in constant and varied emotional expressions. Children were asked to identify repeatedly presented target faces, distinguishing them from distractor faces, during an immediate recognition test and during delayed assessments after 10 min and one week. Emotional facial expression…
Kadak, Muhammed Tayyib; Demirel, Omer Faruk; Yavuz, Mesut; Demir, Türkay
2014-07-01
Research findings debate about features of broad autism phenotype. In this study, we tested whether parents of children with autism have problems recognizing emotional facial expression and the contribution of such an impairment to the broad phenotype of autism. Seventy-two parents of children with autistic spectrum disorder and 38 parents of control group participated in the study. Broad autism features was measured with Autism Quotient (AQ). Recognition of Emotional Face Expression Test was assessed with the Emotion Recognition Test, consisting a set of photographs from Ekman & Friesen's. In a two-tailed analysis of variance of AQ, there was a significant difference for social skills (F(1, 106)=6.095; p<.05). Analyses of variance revealed significant difference in the recognition of happy, surprised and neutral expressions (F(1, 106)=4.068, p=.046; F(1, 106)=4.068, p=.046; F(1, 106)=6.064, p=.016). According to our findings, social impairment could be considered a characteristic feature of BAP. ASD parents had difficulty recognizing neutral expressions, suggesting that ASD parents may have impaired recognition of ambiguous expressions as do autistic children. Copyright © 2014 Elsevier Inc. All rights reserved.
NASA Astrophysics Data System (ADS)
Iqtait, M.; Mohamad, F. S.; Mamat, M.
2018-03-01
Biometric is a pattern recognition system which is used for automatic recognition of persons based on characteristics and features of an individual. Face recognition with high recognition rate is still a challenging task and usually accomplished in three phases consisting of face detection, feature extraction, and expression classification. Precise and strong location of trait point is a complicated and difficult issue in face recognition. Cootes proposed a Multi Resolution Active Shape Models (ASM) algorithm, which could extract specified shape accurately and efficiently. Furthermore, as the improvement of ASM, Active Appearance Models algorithm (AAM) is proposed to extracts both shape and texture of specified object simultaneously. In this paper we give more details about the two algorithms and give the results of experiments, testing their performance on one dataset of faces. We found that the ASM is faster and gains more accurate trait point location than the AAM, but the AAM gains a better match to the texture.
Fenske, Sabrina; Lis, Stefanie; Liebke, Lisa; Niedtfeld, Inga; Kirsch, Peter; Mier, Daniela
2015-01-01
Borderline Personality Disorder (BPD) is characterized by severe deficits in social interactions, which might be linked to deficits in emotion recognition. Research on emotion recognition abilities in BPD revealed heterogeneous results, ranging from deficits to heightened sensitivity. The most stable findings point to an impairment in the evaluation of neutral facial expressions as neutral, as well as to a negative bias in emotion recognition; that is the tendency to attribute negative emotions to neutral expressions, or in a broader sense to report a more negative emotion category than depicted. However, it remains unclear which contextual factors influence the occurrence of this negative bias. Previous studies suggest that priming by preceding emotional information and also constrained processing time might augment the emotion recognition deficit in BPD. To test these assumptions, 32 female BPD patients and 31 healthy females, matched for age and education, participated in an emotion recognition study, in which every facial expression was preceded by either a positive, neutral or negative scene. Furthermore, time constraints for processing were varied by presenting the facial expressions with short (100 ms) or long duration (up to 3000 ms) in two separate blocks. BPD patients showed a significant deficit in emotion recognition for neutral and positive facial expression, associated with a significant negative bias. In BPD patients, this emotion recognition deficit was differentially affected by preceding emotional information and time constraints, with a greater influence of emotional information during long face presentations and a greater influence of neutral information during short face presentations. Our results are in line with previous findings supporting the existence of a negative bias in emotion recognition in BPD patients, and provide further insights into biased social perceptions in BPD patients.
The recognition of emotional expression in prosopagnosia: decoding whole and part faces.
Stephan, Blossom Christa Maree; Breen, Nora; Caine, Diana
2006-11-01
Prosopagnosia is currently viewed within the constraints of two competing theories of face recognition, one highlighting the analysis of features, the other focusing on configural processing of the whole face. This study investigated the role of feature analysis versus whole face configural processing in the recognition of facial expression. A prosopagnosic patient, SC made expression decisions from whole and incomplete (eyes-only and mouth-only) faces where features had been obscured. SC was impaired at recognizing some (e.g., anger, sadness, and fear), but not all (e.g., happiness) emotional expressions from the whole face. Analyses of his performance on incomplete faces indicated that his recognition of some expressions actually improved relative to his performance on the whole face condition. We argue that in SC interference from damaged configural processes seem to override an intact ability to utilize part-based or local feature cues.
Chowdhury, E H; Akaike, Toshihiro
2005-05-20
The DNA delivery to mammalian cells is an essential tool for analyzing gene structure, regulation, and function. The approach holds great promise for the further development of gene therapy techniques and DNA vaccination strategies to treat and control diseases. Here, we report on the establishment of a cell-specific gene delivery and expression system by physical adsorption of a cell-recognition molecule on the nano-crystal surface of carbonate apatite. As a model, DNA/nano-particles were successfully coated with asialofetuin to facilitate uptake by hepatocyte-derived cell lines through the asialoglycoprotein receptor (ASGPr) and albumin to prevent non-specific interactions of the particles with cell-surface. The resulting composite particles with dual surface properties could accelerate DNA uptake and enhance expression to a notable extent. Nano-particles coated with transferrin in the same manner dramatically enhanced transgene expression in the corresponding receptor-bearing cells and thus our newly developed strategy represents a universal phenomenon for anchoring a bio-recognition macromolecule on the apatite crystal surface for targeted gene delivery, having immediate applications in basic research laboratories and great promise for gene therapy. (c) 2005 Wiley Periodicals, Inc.
Labuschagne, Izelle; Jones, Rebecca; Callaghan, Jenny; Whitehead, Daisy; Dumas, Eve M; Say, Miranda J; Hart, Ellen P; Justo, Damian; Coleman, Allison; Dar Santos, Rachelle C; Frost, Chris; Craufurd, David; Tabrizi, Sarah J; Stout, Julie C
2013-05-15
Facial emotion recognition impairments have been reported in Huntington's disease (HD). However, the nature of the impairments across the spectrum of HD remains unclear. We report on emotion recognition data from 344 participants comprising premanifest HD (PreHD) and early HD patients, and controls. In a test of recognition of facial emotions, we examined responses to six basic emotional expressions and neutral expressions. In addition, and within the early HD sample, we tested for differences on emotion recognition performance between those 'on' vs. 'off' neuroleptic or selective serotonin reuptake inhibitor (SSRI) medications. The PreHD groups showed significant (p<0.05) impaired recognition, compared to controls, on fearful, angry and surprised faces; whereas the early HD groups were significantly impaired across all emotions including neutral expressions. In early HD, neuroleptic use was associated with worse facial emotion recognition, whereas SSRI use was associated with better facial emotion recognition. The findings suggest that emotion recognition impairments exist across the HD spectrum, but are relatively more widespread in manifest HD than in the premanifest period. Commonly prescribed medications to treat HD-related symptoms also appear to affect emotion recognition. These findings have important implications for interpersonal communication and medication usage in HD. Copyright © 2012 Elsevier Ireland Ltd. All rights reserved.
Wingenbach, Tanja S. H.; Brosnan, Mark; Pfaltz, Monique C.; Plichta, Michael M.; Ashwin, Chris
2018-01-01
According to embodied cognition accounts, viewing others’ facial emotion can elicit the respective emotion representation in observers which entails simulations of sensory, motor, and contextual experiences. In line with that, published research found viewing others’ facial emotion to elicit automatic matched facial muscle activation, which was further found to facilitate emotion recognition. Perhaps making congruent facial muscle activity explicit produces an even greater recognition advantage. If there is conflicting sensory information, i.e., incongruent facial muscle activity, this might impede recognition. The effects of actively manipulating facial muscle activity on facial emotion recognition from videos were investigated across three experimental conditions: (a) explicit imitation of viewed facial emotional expressions (stimulus-congruent condition), (b) pen-holding with the lips (stimulus-incongruent condition), and (c) passive viewing (control condition). It was hypothesised that (1) experimental condition (a) and (b) result in greater facial muscle activity than (c), (2) experimental condition (a) increases emotion recognition accuracy from others’ faces compared to (c), (3) experimental condition (b) lowers recognition accuracy for expressions with a salient facial feature in the lower, but not the upper face area, compared to (c). Participants (42 males, 42 females) underwent a facial emotion recognition experiment (ADFES-BIV) while electromyography (EMG) was recorded from five facial muscle sites. The experimental conditions’ order was counter-balanced. Pen-holding caused stimulus-incongruent facial muscle activity for expressions with facial feature saliency in the lower face region, which reduced recognition of lower face region emotions. Explicit imitation caused stimulus-congruent facial muscle activity without modulating recognition. Methodological implications are discussed. PMID:29928240
Wingenbach, Tanja S H; Brosnan, Mark; Pfaltz, Monique C; Plichta, Michael M; Ashwin, Chris
2018-01-01
According to embodied cognition accounts, viewing others' facial emotion can elicit the respective emotion representation in observers which entails simulations of sensory, motor, and contextual experiences. In line with that, published research found viewing others' facial emotion to elicit automatic matched facial muscle activation, which was further found to facilitate emotion recognition. Perhaps making congruent facial muscle activity explicit produces an even greater recognition advantage. If there is conflicting sensory information, i.e., incongruent facial muscle activity, this might impede recognition. The effects of actively manipulating facial muscle activity on facial emotion recognition from videos were investigated across three experimental conditions: (a) explicit imitation of viewed facial emotional expressions (stimulus-congruent condition), (b) pen-holding with the lips (stimulus-incongruent condition), and (c) passive viewing (control condition). It was hypothesised that (1) experimental condition (a) and (b) result in greater facial muscle activity than (c), (2) experimental condition (a) increases emotion recognition accuracy from others' faces compared to (c), (3) experimental condition (b) lowers recognition accuracy for expressions with a salient facial feature in the lower, but not the upper face area, compared to (c). Participants (42 males, 42 females) underwent a facial emotion recognition experiment (ADFES-BIV) while electromyography (EMG) was recorded from five facial muscle sites. The experimental conditions' order was counter-balanced. Pen-holding caused stimulus-incongruent facial muscle activity for expressions with facial feature saliency in the lower face region, which reduced recognition of lower face region emotions. Explicit imitation caused stimulus-congruent facial muscle activity without modulating recognition. Methodological implications are discussed.
Dissociable roles of internal feelings and face recognition ability in facial expression decoding.
Zhang, Lin; Song, Yiying; Liu, Ling; Liu, Jia
2016-05-15
The problem of emotion recognition has been tackled by researchers in both affective computing and cognitive neuroscience. While affective computing relies on analyzing visual features from facial expressions, it has been proposed that humans recognize emotions by internally simulating the emotional states conveyed by others' expressions, in addition to perceptual analysis of facial features. Here we investigated whether and how our internal feelings contributed to the ability to decode facial expressions. In two independent large samples of participants, we observed that individuals who generally experienced richer internal feelings exhibited a higher ability to decode facial expressions, and the contribution of internal feelings was independent of face recognition ability. Further, using voxel-based morphometry, we found that the gray matter volume (GMV) of bilateral superior temporal sulcus (STS) and the right inferior parietal lobule was associated with facial expression decoding through the mediating effect of internal feelings, while the GMV of bilateral STS, precuneus, and the right central opercular cortex contributed to facial expression decoding through the mediating effect of face recognition ability. In addition, the clusters in bilateral STS involved in the two components were neighboring yet separate. Our results may provide clues about the mechanism by which internal feelings, in addition to face recognition ability, serve as an important instrument for humans in facial expression decoding. Copyright © 2016 Elsevier Inc. All rights reserved.
Gaze Dynamics in the Recognition of Facial Expressions of Emotion.
Barabanschikov, Vladimir A
2015-01-01
We studied preferably fixated parts and features of human face in the process of recognition of facial expressions of emotion. Photographs of facial expressions were used. Participants were to categorize these as basic emotions; during this process, eye movements were registered. It was found that variation in the intensity of an expression is mirrored in accuracy of emotion recognition; it was also reflected by several indices of oculomotor function: duration of inspection of certain areas of the face, its upper and bottom or right parts, right and left sides; location, number and duration of fixations, viewing trajectory. In particular, for low-intensity expressions, right side of the face was found to be attended predominantly (right-side dominance); the right-side dominance effect, was, however, absent for expressions of high intensity. For both low- and high-intensity expressions, upper face part was predominantly fixated, though with greater fixation of high-intensity expressions. The majority of trials (70%), in line with findings in previous studies, revealed a V-shaped pattern of inspection trajectory. No relationship, between accuracy of recognition of emotional expressions, was found, though, with either location and duration of fixations or pattern of gaze directedness in the face. © The Author(s) 2015.
Assessment of Emotional Experience and Emotional Recognition in Complicated Grief
Fernández-Alcántara, Manuel; Cruz-Quintana, Francisco; Pérez-Marfil, M. N.; Catena-Martínez, Andrés; Pérez-García, Miguel; Turnbull, Oliver H.
2016-01-01
There is substantial evidence of bias in the processing of emotion in people with complicated grief (CG). Previous studies have tended to assess the expression of emotion in CG, but other aspects of emotion (mainly emotion recognition, and the subjective aspects of emotion) have not been addressed, despite their importance for practicing clinicians. A quasi-experimental design with two matched groups (Complicated Grief, N = 24 and Non-Complicated Grief, N = 20) was carried out. The Facial Expression of Emotion Test (emotion recognition), a set of pictures from the International Affective Picture System (subjective experience of emotion) and the Symptom Checklist 90 Revised (psychopathology) were employed. The CG group showed lower scores on the dimension of valence for specific conditions on the IAPS, related to the subjective experience of emotion. In addition, they presented higher values of psychopathology. In contrast, statistically significant results were not found for the recognition of emotion. In conclusion, from a neuropsychological point of view, the subjective aspects of emotion and psychopathology seem central in explaining the experience of those with CG. These results are clinically significant for psychotherapists and psychoanalysts working in the field of grief and loss. PMID:26903928
ERIC Educational Resources Information Center
Gross, Thomas F.
2008-01-01
The recognition of facial immaturity and emotional expression by children with autism, language disorders, mental retardation, and non-disabled controls was studied in two experiments. Children identified immaturity and expression in upright and inverted faces. The autism group identified fewer immature faces and expressions than control (Exp. 1 &…
Violent Media Consumption and the Recognition of Dynamic Facial Expressions
ERIC Educational Resources Information Center
Kirsh, Steven J.; Mounts, Jeffrey R. W.; Olczak, Paul V.
2006-01-01
This study assessed the speed of recognition of facial emotional expressions (happy and angry) as a function of violent media consumption. Color photos of calm facial expressions morphed to either an angry or a happy facial expression. Participants were asked to make a speeded identification of the emotion (happiness or anger) during the morph.…
Facial expressions recognition with an emotion expressive robotic head
NASA Astrophysics Data System (ADS)
Doroftei, I.; Adascalitei, F.; Lefeber, D.; Vanderborght, B.; Doroftei, I. A.
2016-08-01
The purpose of this study is to present the preliminary steps in facial expressions recognition with a new version of an expressive social robotic head. So, in a first phase, our main goal was to reach a minimum level of emotional expressiveness in order to obtain nonverbal communication between the robot and human by building six basic facial expressions. To evaluate the facial expressions, the robot was used in some preliminary user studies, among children and adults.
Emotional facial recognition in proactive and reactive violent offenders.
Philipp-Wiegmann, Florence; Rösler, Michael; Retz-Junginger, Petra; Retz, Wolfgang
2017-10-01
The purpose of this study is to analyse individual differences in the ability of emotional facial recognition in violent offenders, who were characterised as either reactive or proactive in relation to their offending. In accordance with findings of our previous study, we expected higher impairments in facial recognition in reactive than proactive violent offenders. To assess the ability to recognize facial expressions, the computer-based Facial Emotional Expression Labeling Test (FEEL) was performed. Group allocation of reactive und proactive violent offenders and assessment of psychopathic traits were performed by an independent forensic expert using rating scales (PROREA, PCL-SV). Compared to proactive violent offenders and controls, the performance of emotion recognition in the reactive offender group was significantly lower, both in total and especially in recognition of negative emotions such as anxiety (d = -1.29), sadness (d = -1.54), and disgust (d = -1.11). Furthermore, reactive violent offenders showed a tendency to interpret non-anger emotions as anger. In contrast, proactive violent offenders performed as well as controls. General and specific deficits in reactive violent offenders are in line with the results of our previous study and correspond to predictions of the Integrated Emotion System (IES, 7) and the hostile attribution processes (21). Due to the different error pattern in the FEEL test, the theoretical distinction between proactive and reactive aggression can be supported based on emotion recognition, even though aggression itself is always a heterogeneous act rather than a distinct one-dimensional concept.
Clipperton-Allen, Amy E.; Lee, Anna W.; Reyes, Anny; Devidze, Nino; Phan, Anna; Pfaff, Donald W.; Choleris, Elena
2012-01-01
Inter- and intra-species differences in social behavior and recognition-related hormones and receptors suggest that different distribution and/or expression patterns may relate to social recognition. We used qRT-PCR to investigate naturally occurring differences in expression of estrogen receptor-alpha (ERα), ER-beta (ERβ), progesterone receptor (PR), oxytocin (OT) and receptor, and vasopressin (AVP) and receptors in proestrous female mice. Following four 5 min exposures to the same two conspecifics, one was replaced with a novel mouse in the final trial (T5). Gene expression was examined in mice showing high (85–100%) and low (40–60%) social recognition scores (i.e., preferential novel mouse investigation in T5) in eight socially-relevant brain regions. Results supported OT and AVP involvement in social recognition, and suggest that in the medial preoptic area, increased OT and AVP mRNA, together with ERα and ERβ gene activation, relate to improved social recognition. Initial social investigation correlated with ERs, PR and OTR in the dorsolateral septum, suggesting that these receptors may modulate social interest without affecting social recognition. Finally, increased lateral amygdala gene activation in the LR mice may be associated with general learning impairments, while decreased lateral amygdala activity may indicate more efficient cognitive mechanisms in the HR mice. PMID:22079582
Specific Impairments in the Recognition of Emotional Facial Expressions in Parkinson’s Disease
Clark, Uraina S.; Neargarder, Sandy; Cronin-Golomb, Alice
2008-01-01
Studies investigating the ability to recognize emotional facial expressions in non-demented individuals with Parkinson’s disease (PD) have yielded equivocal findings. A possible reason for this variability may lie in the confounding of emotion recognition with cognitive task requirements, a confound arising from the lack of a control condition using non-emotional stimuli. The present study examined emotional facial expression recognition abilities in 20 non-demented patients with PD and 23 control participants relative to their performances on a non-emotional landscape categorization test with comparable task requirements. We found that PD participants were normal on the control task but exhibited selective impairments in the recognition of facial emotion, specifically for anger (driven by those with right hemisphere pathology) and surprise (driven by those with left hemisphere pathology), even when controlling for depression level. Male but not female PD participants further displayed specific deficits in the recognition of fearful expressions. We suggest that the neural substrates that may subserve these impairments include the ventral striatum, amygdala, and prefrontal cortices. Finally, we observed that in PD participants, deficiencies in facial emotion recognition correlated with higher levels of interpersonal distress, which calls attention to the significant psychosocial impact that facial emotion recognition impairments may have on individuals with PD. PMID:18485422
Aviezer, Hillel; Hassin, Ran. R.; Perry, Anat; Dudarev, Veronica; Bentin, Shlomo
2012-01-01
The current study examined the nature of deficits in emotion recognition from facial expressions in case LG, an individual with a rare form of developmental visual agnosia (DVA). LG presents with profoundly impaired recognition of facial expressions, yet the underlying nature of his deficit remains unknown. During typical face processing, normal sighted individuals extract information about expressed emotions from face regions with activity diagnostic for specific emotion categories. Given LG’s impairment, we sought to shed light on his emotion perception by examining if priming facial expressions with diagnostic emotional face components would facilitate his recognition of the emotion expressed by the face. LG and control participants matched isolated face components with components appearing in a subsequently presented full-face and then categorized the face’s emotion. Critically, the matched components were from regions which were diagnostic or non-diagnostic of the emotion portrayed by the full face. In experiment 1, when the full faces were briefly presented (150 ms), LG’s performance was strongly influenced by the diagnosticity of the components: His emotion recognition was boosted within normal limits when diagnostic components were used and was obliterated when non-diagnostic components were used. By contrast, in experiment 2, when the face-exposure duration was extended (2000 ms), the beneficial effect of the diagnostic matching was diminished as was the detrimental effect of the non-diagnostic matching. These data highlight the impact of diagnostic facial features in normal expression recognition and suggest that impaired emotion recognition in DVA results from deficient visual integration across diagnostic face components. PMID:22349446
Interactive object recognition assistance: an approach to recognition starting from target objects
NASA Astrophysics Data System (ADS)
Geisler, Juergen; Littfass, Michael
1999-07-01
Recognition of target objects in remotely sensed imagery required detailed knowledge about the target object domain as well as about mapping properties of the sensing system. The art of object recognition is to combine both worlds appropriately and to provide models of target appearance with respect to sensor characteristics. Common approaches to support interactive object recognition are either driven from the sensor point of view and address the problem of displaying images in a manner adequate to the sensing system. Or they focus on target objects and provide exhaustive encyclopedic information about this domain. Our paper discusses an approach to assist interactive object recognition based on knowledge about target objects and taking into account the significance of object features with respect to characteristics of the sensed imagery, e.g. spatial and spectral resolution. An `interactive recognition assistant' takes the image analyst through the interpretation process by indicating step-by-step the respectively most significant features of objects in an actual set of candidates. The significance of object features is expressed by pregenerated trees of significance, and by the dynamic computation of decision relevance for every feature at each step of the recognition process. In the context of this approach we discuss the question of modeling and storing the multisensorial/multispectral appearances of target objects and object classes as well as the problem of an adequate dynamic human-machine-interface that takes into account various mental models of human image interpretation.
Dissociation between facial and bodily expressions in emotion recognition: A case study.
Leiva, Samanta; Margulis, Laura; Micciulli, Andrea; Ferreres, Aldo
2017-12-21
Existing single-case studies have reported deficit in recognizing basic emotions through facial expression and unaffected performance with body expressions, but not the opposite pattern. The aim of this paper is to present a case study with impaired emotion recognition through body expressions and intact performance with facial expressions. In this single-case study we assessed a 30-year-old patient with autism spectrum disorder, without intellectual disability, and a healthy control group (n = 30) with four tasks of basic and complex emotion recognition through face and body movements, and two non-emotional control tasks. To analyze the dissociation between facial and body expressions, we used Crawford and Garthwaite's operational criteria, and we compared the patient and the control group performance with a modified one-tailed t-test designed specifically for single-case studies. There were no statistically significant differences between the patient's and the control group's performances on the non-emotional body movement task or the facial perception task. For both kinds of emotions (basic and complex) when the patient's performance was compared to the control group's, statistically significant differences were only observed for the recognition of body expressions. There were no significant differences between the patient's and the control group's correct answers for emotional facial stimuli. Our results showed a profile of impaired emotion recognition through body expressions and intact performance with facial expressions. This is the first case study that describes the existence of this kind of dissociation pattern between facial and body expressions of basic and complex emotions.
Felisberti, Fatima; Terry, Philip
2015-09-01
The study compared alcohol's effects on the recognition of briefly displayed facial expressions of emotion (so-called microexpressions) with expressions presented for a longer period. Using a repeated-measures design, we tested 18 participants three times (counterbalanced), after (i) a placebo drink, (ii) a low-to-moderate dose of alcohol (0.17 g/kg women; 0.20 g/kg men) and (iii) a moderate-to-high dose of alcohol (0.52 g/kg women; 0.60 g/kg men). On each session, participants were presented with stimuli representing six emotions (happiness, sadness, anger, fear, disgust and contempt) overlaid on a generic avatar in a six-alternative forced-choice paradigm. A neutral expression (1 s) preceded and followed a target expression presented for 200 ms (microexpressions) or 400 ms. Participants mouse clicked the correct answer. The recognition of disgust was significantly better after the high dose of alcohol than after the low dose or placebo drinks at both durations of stimulus presentation. A similar profile of effects was found for the recognition of contempt. There were no effects on response latencies. Alcohol can increase sensitivity to expressions of disgust and contempt. Such effects are not dependent on stimulus duration up to 400 ms and may reflect contextual modulation of alcohol's effects on emotion recognition. Copyright © 2015 John Wiley & Sons, Ltd.
Social appraisal influences recognition of emotions.
Mumenthaler, Christian; Sander, David
2012-06-01
The notion of social appraisal emphasizes the importance of a social dimension in appraisal theories of emotion by proposing that the way an individual appraises an event is influenced by the way other individuals appraise and feel about the same event. This study directly tested this proposal by asking participants to recognize dynamic facial expressions of emotion (fear, happiness, or anger in Experiment 1; fear, happiness, anger, or neutral in Experiment 2) in a target face presented at the center of a screen while a contextual face, which appeared simultaneously in the periphery of the screen, expressed an emotion (fear, happiness, anger) or not (neutral) and either looked at the target face or not. We manipulated gaze direction to be able to distinguish between a mere contextual effect (gaze away from both the target face and the participant) and a specific social appraisal effect (gaze toward the target face). Results of both experiments provided evidence for a social appraisal effect in emotion recognition, which differed from the mere effect of contextual information: Whereas facial expressions were identical in both conditions, the direction of the gaze of the contextual face influenced emotion recognition. Social appraisal facilitated the recognition of anger, happiness, and fear when the contextual face expressed the same emotion. This facilitation was stronger than the mere contextual effect. Social appraisal also allowed better recognition of fear when the contextual face expressed anger and better recognition of anger when the contextual face expressed fear. 2012 APA, all rights reserved
Reyes, B Nicole; Segal, Shira C; Moulson, Margaret C
2018-01-01
Emotion recognition is important for social interaction and communication, yet previous research has identified a cross-cultural emotion recognition deficit: Recognition is less accurate for emotions expressed by individuals from a cultural group different than one's own. The current study examined whether social categorization based on race, in the absence of cultural differences, influences emotion recognition in a diverse context. South Asian and White Canadians in the Greater Toronto Area completed an emotion recognition task that required them to identify the seven basic emotional expressions when posed by members of the same two groups, allowing us to tease apart the contributions of culture and social group membership. Contrary to our hypothesis, there was no mutual in-group advantage in emotion recognition: Participants were not more accurate at recognizing emotions posed by their respective racial in-groups. Both groups were more accurate at recognizing expressions when posed by South Asian faces, and White participants were more accurate overall compared to South Asian participants. These results suggest that in a diverse environment, categorization based on race alone does not lead to the creation of social out-groups in a way that negatively impacts emotion recognition.
An investigation of the effect of race-based social categorization on adults’ recognition of emotion
Reyes, B. Nicole; Segal, Shira C.
2018-01-01
Emotion recognition is important for social interaction and communication, yet previous research has identified a cross-cultural emotion recognition deficit: Recognition is less accurate for emotions expressed by individuals from a cultural group different than one’s own. The current study examined whether social categorization based on race, in the absence of cultural differences, influences emotion recognition in a diverse context. South Asian and White Canadians in the Greater Toronto Area completed an emotion recognition task that required them to identify the seven basic emotional expressions when posed by members of the same two groups, allowing us to tease apart the contributions of culture and social group membership. Contrary to our hypothesis, there was no mutual in-group advantage in emotion recognition: Participants were not more accurate at recognizing emotions posed by their respective racial in-groups. Both groups were more accurate at recognizing expressions when posed by South Asian faces, and White participants were more accurate overall compared to South Asian participants. These results suggest that in a diverse environment, categorization based on race alone does not lead to the creation of social out-groups in a way that negatively impacts emotion recognition. PMID:29474367
Face recognition via edge-based Gabor feature representation for plastic surgery-altered images
NASA Astrophysics Data System (ADS)
Chude-Olisah, Chollette C.; Sulong, Ghazali; Chude-Okonkwo, Uche A. K.; Hashim, Siti Z. M.
2014-12-01
Plastic surgery procedures on the face introduce skin texture variations between images of the same person (intra-subject), thereby making the task of face recognition more difficult than in normal scenario. Usually, in contemporary face recognition systems, the original gray-level face image is used as input to the Gabor descriptor, which translates to encoding some texture properties of the face image. The texture-encoding process significantly degrades the performance of such systems in the case of plastic surgery due to the presence of surgically induced intra-subject variations. Based on the proposition that the shape of significant facial components such as eyes, nose, eyebrow, and mouth remains unchanged after plastic surgery, this paper employs an edge-based Gabor feature representation approach for the recognition of surgically altered face images. We use the edge information, which is dependent on the shapes of the significant facial components, to address the plastic surgery-induced texture variation problems. To ensure that the significant facial components represent useful edge information with little or no false edges, a simple illumination normalization technique is proposed for preprocessing. Gabor wavelet is applied to the edge image to accentuate on the uniqueness of the significant facial components for discriminating among different subjects. The performance of the proposed method is evaluated on the Georgia Tech (GT) and the Labeled Faces in the Wild (LFW) databases with illumination and expression problems, and the plastic surgery database with texture changes. Results show that the proposed edge-based Gabor feature representation approach is robust against plastic surgery-induced face variations amidst expression and illumination problems and outperforms the existing plastic surgery face recognition methods reported in the literature.
A unified probabilistic framework for spontaneous facial action modeling and understanding.
Tong, Yan; Chen, Jixu; Ji, Qiang
2010-02-01
Facial expression is a natural and powerful means of human communication. Recognizing spontaneous facial actions, however, is very challenging due to subtle facial deformation, frequent head movements, and ambiguous and uncertain facial motion measurements. Because of these challenges, current research in facial expression recognition is limited to posed expressions and often in frontal view. A spontaneous facial expression is characterized by rigid head movements and nonrigid facial muscular movements. More importantly, it is the coherent and consistent spatiotemporal interactions among rigid and nonrigid facial motions that produce a meaningful facial expression. Recognizing this fact, we introduce a unified probabilistic facial action model based on the Dynamic Bayesian network (DBN) to simultaneously and coherently represent rigid and nonrigid facial motions, their spatiotemporal dependencies, and their image measurements. Advanced machine learning methods are introduced to learn the model based on both training data and subjective prior knowledge. Given the model and the measurements of facial motions, facial action recognition is accomplished through probabilistic inference by systematically integrating visual measurements with the facial action model. Experiments show that compared to the state-of-the-art techniques, the proposed system yields significant improvements in recognizing both rigid and nonrigid facial motions, especially for spontaneous facial expressions.
Flores-Balter, Gabriela; Cordova-Jadue, Héctor; Chiti-Morales, Alessandra; Lespay, Carolyne; Espina-Marchant, Pablo; Falcon, Romina; Grinspun, Noemi; Sanchez, Jessica; Bustamante, Diego; Morales, Paola; Herrera-Marschitz, Mario; Valdés, José L
2016-10-15
Perinatal asphyxia (PA) is associated with long-term neuronal damage and cognitive deficits in adulthood, such as learning and memory disabilities. After PA, specific brain regions are compromised, including neocortex, hippocampus, basal ganglia, and ascending neuromodulatory pathways, such as dopamine system, explaining some of the cognitive disabilities. We hypothesize that other neuromodulatory systems, such as histamine system from the tuberomammillary nucleus (TMN), which widely project to telencephalon, shown to be relevant for learning and memory, may be compromised by PA. We investigated here the effect of PA on (i) Density and neuronal activity of TMN neurons by double immunoreactivity for adenosine deaminase (ADA) and c-Fos, as marker for histaminergic neurons and neuronal activity respectively. (ii) Expression of the histamine-synthesizing enzyme, histidine decarboxylase (HDC) by western blot and (iii) thioperamide an H3 histamine receptor antagonist, on an object recognition memory task. Asphyxia-exposed rats showed a decrease of ADA density and c-Fos activity in TMN, and decrease of HDC expression in hypothalamus. Asphyxia-exposed rats also showed a low performance in object recognition memory compared to caesarean-delivered controls, which was reverted in a dose-dependent manner by the H3 antagonist thioperamide (5-10mg/kg, i.p.). The present results show that the histaminergic neuronal system of the TMN is involved in the long-term effects induced by PA, affecting learning and memory. Copyright © 2016 Elsevier B.V. All rights reserved.
Leppänen, J M; Niehaus, D J H; Koen, L; Du Toit, E; Schoeman, R; Emsley, R
2006-06-01
Schizophrenia is associated with a deficit in the recognition of negative emotions from facial expressions. The present study examined the universality of this finding by studying facial expression recognition in African Xhosa population. Forty-four Xhosa patients with schizophrenia and forty healthy controls were tested with a computerized task requiring rapid perceptual discrimination of matched positive (i.e. happy), negative (i.e. angry), and neutral faces. Patients were equally accurate as controls in recognizing happy faces but showed a marked impairment in recognition of angry faces. The impairment was particularly pronounced for high-intensity (open-mouth) angry faces. Patients also exhibited more false happy and angry responses to neutral faces than controls. No correlation between level of education or illness duration and emotion recognition was found but the deficit in the recognition of negative emotions was more pronounced in familial compared to non-familial cases of schizophrenia. These findings suggest that the deficit in the recognition of negative facial expressions may constitute a universal neurocognitive marker of schizophrenia.
Age, gender, and puberty influence the development of facial emotion recognition.
Lawrence, Kate; Campbell, Ruth; Skuse, David
2015-01-01
Our ability to differentiate between simple facial expressions of emotion develops between infancy and early adulthood, yet few studies have explored the developmental trajectory of emotion recognition using a single methodology across a wide age-range. We investigated the development of emotion recognition abilities through childhood and adolescence, testing the hypothesis that children's ability to recognize simple emotions is modulated by chronological age, pubertal stage and gender. In order to establish norms, we assessed 478 children aged 6-16 years, using the Ekman-Friesen Pictures of Facial Affect. We then modeled these cross-sectional data in terms of competence in accurate recognition of the six emotions studied, when the positive correlation between emotion recognition and IQ was controlled. Significant linear trends were seen in children's ability to recognize facial expressions of happiness, surprise, fear, and disgust; there was improvement with increasing age. In contrast, for sad and angry expressions there is little or no change in accuracy over the age range 6-16 years; near-adult levels of competence are established by middle-childhood. In a sampled subset, pubertal status influenced the ability to recognize facial expressions of disgust and anger; there was an increase in competence from mid to late puberty, which occurred independently of age. A small female advantage was found in the recognition of some facial expressions. The normative data provided in this study will aid clinicians and researchers in assessing the emotion recognition abilities of children and will facilitate the identification of abnormalities in a skill that is often impaired in neurodevelopmental disorders. If emotion recognition abilities are a good model with which to understand adolescent development, then these results could have implications for the education, mental health provision and legal treatment of teenagers.
Age, gender, and puberty influence the development of facial emotion recognition
Lawrence, Kate; Campbell, Ruth; Skuse, David
2015-01-01
Our ability to differentiate between simple facial expressions of emotion develops between infancy and early adulthood, yet few studies have explored the developmental trajectory of emotion recognition using a single methodology across a wide age-range. We investigated the development of emotion recognition abilities through childhood and adolescence, testing the hypothesis that children’s ability to recognize simple emotions is modulated by chronological age, pubertal stage and gender. In order to establish norms, we assessed 478 children aged 6–16 years, using the Ekman-Friesen Pictures of Facial Affect. We then modeled these cross-sectional data in terms of competence in accurate recognition of the six emotions studied, when the positive correlation between emotion recognition and IQ was controlled. Significant linear trends were seen in children’s ability to recognize facial expressions of happiness, surprise, fear, and disgust; there was improvement with increasing age. In contrast, for sad and angry expressions there is little or no change in accuracy over the age range 6–16 years; near-adult levels of competence are established by middle-childhood. In a sampled subset, pubertal status influenced the ability to recognize facial expressions of disgust and anger; there was an increase in competence from mid to late puberty, which occurred independently of age. A small female advantage was found in the recognition of some facial expressions. The normative data provided in this study will aid clinicians and researchers in assessing the emotion recognition abilities of children and will facilitate the identification of abnormalities in a skill that is often impaired in neurodevelopmental disorders. If emotion recognition abilities are a good model with which to understand adolescent development, then these results could have implications for the education, mental health provision and legal treatment of teenagers. PMID:26136697
Yu, Shao Hua; Zhu, Jun Peng; Xu, You; Zheng, Lei Lei; Chai, Hao; He, Wei; Liu, Wei Bo; Li, Hui Chun; Wang, Wei
2012-12-01
To study the contribution of executive function to abnormal recognition of facial expressions of emotion in schizophrenia patients. Abnormal recognition of facial expressions of emotion was assayed according to Japanese and Caucasian facial expressions of emotion (JACFEE), Wisconsin card sorting test (WCST), positive and negative symptom scale, and Hamilton anxiety and depression scale, respectively, in 88 paranoid schizophrenia patients and 75 healthy volunteers. Patients scored higher on the Positive and Negative Symptom Scale and the Hamilton Anxiety and Depression Scales, displayed lower JACFEE recognition accuracies and poorer WCST performances. The JACFEE recognition accuracy of contempt and disgust was negatively correlated with the negative symptom scale score while the recognition accuracy of fear was positively with the positive symptom scale score and the recognition accuracy of surprise was negatively with the general psychopathology score in patients. Moreover, the WCST could predict the JACFEE recognition accuracy of contempt, disgust, and sadness in patients, and the perseverative errors negatively predicted the recognition accuracy of sadness in healthy volunteers. The JACFEE recognition accuracy of sadness could predict the WCST categories in paranoid schizophrenia patients. Recognition accuracy of social-/moral emotions, such as contempt, disgust and sadness is related to the executive function in paranoid schizophrenia patients, especially when regarding sadness. Copyright © 2012 The Editorial Board of Biomedical and Environmental Sciences. Published by Elsevier B.V. All rights reserved.
ERIC Educational Resources Information Center
Haettig, Jakob; Sun, Yanjun; Wood, Marcelo A.; Xu, Xiangmin
2013-01-01
The allatostatin receptor (AlstR)/ligand inactivation system enables potent regulation of neuronal circuit activity. To examine how different cell types participate in memory formation, we have used this system through Cre-directed, cell-type specific expression in mouse hippocampal CA1 in vivo and examined functional effects of inactivation of…
Single-Molecule View of Small RNA-Guided Target Search and Recognition.
Globyte, Viktorija; Kim, Sung Hyun; Joo, Chirlmin
2018-05-20
Most everyday processes in life involve a necessity for an entity to locate its target. On a cellular level, many proteins have to find their target to perform their function. From gene-expression regulation to DNA repair to host defense, numerous nucleic acid-interacting proteins use distinct target search mechanisms. Several proteins achieve that with the help of short RNA strands known as guides. This review focuses on single-molecule advances studying the target search and recognition mechanism of Argonaute and CRISPR (clustered regularly interspaced short palindromic repeats) systems. We discuss different steps involved in search and recognition, from the initial complex prearrangement into the target-search competent state to the final proofreading steps. We focus on target search mechanisms that range from weak interactions, to one- and three-dimensional diffusion, to conformational proofreading. We compare the mechanisms of Argonaute and CRISPR with a well-studied target search system, RecA.
Multimedia Content Development as a Facial Expression Datasets for Recognition of Human Emotions
NASA Astrophysics Data System (ADS)
Mamonto, N. E.; Maulana, H.; Liliana, D. Y.; Basaruddin, T.
2018-02-01
Datasets that have been developed before contain facial expression from foreign people. The development of multimedia content aims to answer the problems experienced by the research team and other researchers who will conduct similar research. The method used in the development of multimedia content as facial expression datasets for human emotion recognition is the Villamil-Molina version of the multimedia development method. Multimedia content developed with 10 subjects or talents with each talent performing 3 shots with each capturing talent having to demonstrate 19 facial expressions. After the process of editing and rendering, tests are carried out with the conclusion that the multimedia content can be used as a facial expression dataset for recognition of human emotions.
ERIC Educational Resources Information Center
Doi, Hirokazu; Fujisawa, Takashi X.; Kanai, Chieko; Ohta, Haruhisa; Yokoi, Hideki; Iwanami, Akira; Kato, Nobumasa; Shinohara, Kazuyuki
2013-01-01
This study investigated the ability of adults with Asperger syndrome to recognize emotional categories of facial expressions and emotional prosodies with graded emotional intensities. The individuals with Asperger syndrome showed poorer recognition performance for angry and sad expressions from both facial and vocal information. The group…
USDA-ARS?s Scientific Manuscript database
Macrophages express various pathogen-recognition receptors (PRRs) which recognize pathogen-associated molecular patterns (PAMPs) and activate genes responsible for host defense. The aim of this study was to characterize two porcine macrophage cell lines (Cdelta+ and Cdelta–) for the expression of P...
Hwang, Wonjun; Wang, Haitao; Kim, Hyunwoo; Kee, Seok-Cheol; Kim, Junmo
2011-04-01
The authors present a robust face recognition system for large-scale data sets taken under uncontrolled illumination variations. The proposed face recognition system consists of a novel illumination-insensitive preprocessing method, a hybrid Fourier-based facial feature extraction, and a score fusion scheme. First, in the preprocessing stage, a face image is transformed into an illumination-insensitive image, called an "integral normalized gradient image," by normalizing and integrating the smoothed gradients of a facial image. Then, for feature extraction of complementary classifiers, multiple face models based upon hybrid Fourier features are applied. The hybrid Fourier features are extracted from different Fourier domains in different frequency bandwidths, and then each feature is individually classified by linear discriminant analysis. In addition, multiple face models are generated by plural normalized face images that have different eye distances. Finally, to combine scores from multiple complementary classifiers, a log likelihood ratio-based score fusion scheme is applied. The proposed system using the face recognition grand challenge (FRGC) experimental protocols is evaluated; FRGC is a large available data set. Experimental results on the FRGC version 2.0 data sets have shown that the proposed method shows an average of 81.49% verification rate on 2-D face images under various environmental variations such as illumination changes, expression changes, and time elapses.
Stereotypes and prejudice affect the recognition of emotional body postures.
Bijlstra, Gijsbert; Holland, Rob W; Dotsch, Ron; Wigboldus, Daniel H J
2018-03-26
Most research on emotion recognition focuses on facial expressions. However, people communicate emotional information through bodily cues as well. Prior research on facial expressions has demonstrated that emotion recognition is modulated by top-down processes. Here, we tested whether this top-down modulation generalizes to the recognition of emotions from body postures. We report three studies demonstrating that stereotypes and prejudice about men and women may affect how fast people classify various emotional body postures. Our results suggest that gender cues activate gender associations, which affect the recognition of emotions from body postures in a top-down fashion. (PsycINFO Database Record (c) 2018 APA, all rights reserved).
ERIC Educational Resources Information Center
Tardif, Carole; Laine, France; Rodriguez, Melissa; Gepner, Bruno
2007-01-01
This study examined the effects of slowing down presentation of facial expressions and their corresponding vocal sounds on facial expression recognition and facial and/or vocal imitation in children with autism. Twelve autistic children and twenty-four normal control children were presented with emotional and non-emotional facial expressions on…
Conclusiveness of natural languages and recognition of images
DOE Office of Scientific and Technical Information (OSTI.GOV)
Wojcik, Z.M.
1983-01-01
The conclusiveness is investigated using recognition processes and one-one correspondence between expressions of a natural language and graphs representing events. The graphs, as conceived in psycholinguistics, are obtained as a result of perception processes. It is possible to generate and process the graphs automatically, using computers and then to convert the resulting graphs into expressions of a natural language. Correctness and conclusiveness of the graphs and sentences are investigated using the fundamental condition for events representation processes. Some consequences of the conclusiveness are discussed, e.g. undecidability of arithmetic, human brain assymetry, correctness of statistical calculations and operations research. It ismore » suggested that the group theory should be imposed on mathematical models of any real system. Proof of the fundamental condition is also presented. 14 references.« less
The recognition of facial emotion expressions in Parkinson's disease.
Assogna, Francesca; Pontieri, Francesco E; Caltagirone, Carlo; Spalletta, Gianfranco
2008-11-01
A limited number of studies in Parkinson's Disease (PD) suggest a disturbance of recognition of facial emotion expressions. In particular, disgust recognition impairment has been reported in unmedicated and medicated PD patients. However, the results are rather inconclusive in the definition of the degree and the selectivity of emotion recognition impairment, and an associated impairment of almost all basic facial emotions in PD is also described. Few studies have investigated the relationship with neuropsychiatric and neuropsychological symptoms with mainly negative results. This inconsistency may be due to many different problems, such as emotion assessment, perception deficit, cognitive impairment, behavioral symptoms, illness severity and antiparkinsonian therapy. Here we review the clinical characteristics and neural structures involved in the recognition of specific facial emotion expressions, and the plausible role of dopamine transmission and dopamine replacement therapy in these processes. It is clear that future studies should be directed to clarify all these issues.
Clipperton-Allen, Amy E; Lee, Anna W; Reyes, Anny; Devidze, Nino; Phan, Anna; Pfaff, Donald W; Choleris, Elena
2012-02-28
Inter- and intra-species differences in social behavior and recognition-related hormones and receptors suggest that different distribution and/or expression patterns may relate to social recognition. We used qRT-PCR to investigate naturally occurring differences in expression of estrogen receptor-alpha (ERα), ER-beta (ERβ), progesterone receptor (PR), oxytocin (OT) and receptor, and vasopressin (AVP) and receptors in proestrous female mice. Following four 5 min exposures to the same two conspecifics, one was replaced with a novel mouse in the final trial (T5). Gene expression was examined in mice showing high (85-100%) and low (40-60%) social recognition scores (i.e., preferential novel mouse investigation in T5) in eight socially-relevant brain regions. Results supported OT and AVP involvement in social recognition, and suggest that in the medial preoptic area, increased OT and AVP mRNA, together with ERα and ERβ gene activation, relate to improved social recognition. Initial social investigation correlated with ERs, PR and OTR in the dorsolateral septum, suggesting that these receptors may modulate social interest without affecting social recognition. Finally, increased lateral amygdala gene activation in the LR mice may be associated with general learning impairments, while decreased lateral amygdala activity may indicate more efficient cognitive mechanisms in the HR mice. Copyright © 2011 Elsevier Inc. All rights reserved.
NASA Astrophysics Data System (ADS)
Han, Sheng; Xi, Shi-qiong; Geng, Wei-dong
2017-11-01
In order to solve the problem of low recognition rate of traditional feature extraction operators under low-resolution images, a novel algorithm of expression recognition is proposed, named central oblique average center-symmetric local binary pattern (CS-LBP) with adaptive threshold (ATCS-LBP). Firstly, the features of face images can be extracted by the proposed operator after pretreatment. Secondly, the obtained feature image is divided into blocks. Thirdly, the histogram of each block is computed independently and all histograms can be connected serially to create a final feature vector. Finally, expression classification is achieved by using support vector machine (SVM) classifier. Experimental results on Japanese female facial expression (JAFFE) database show that the proposed algorithm can achieve a recognition rate of 81.9% when the resolution is as low as 16×16, which is much better than that of the traditional feature extraction operators.
Aviezer, Hillel; Hassin, Ran R; Perry, Anat; Dudarev, Veronica; Bentin, Shlomo
2012-04-01
The current study examined the nature of deficits in emotion recognition from facial expressions in case LG, an individual with a rare form of developmental visual agnosia (DVA). LG presents with profoundly impaired recognition of facial expressions, yet the underlying nature of his deficit remains unknown. During typical face processing, normal sighted individuals extract information about expressed emotions from face regions with activity diagnostic for specific emotion categories. Given LG's impairment, we sought to shed light on his emotion perception by examining if priming facial expressions with diagnostic emotional face components would facilitate his recognition of the emotion expressed by the face. LG and control participants matched isolated face components with components appearing in a subsequently presented full-face and then categorized the face's emotion. Critically, the matched components were from regions which were diagnostic or non-diagnostic of the emotion portrayed by the full face. In experiment 1, when the full faces were briefly presented (150 ms), LG's performance was strongly influenced by the diagnosticity of the components: his emotion recognition was boosted within normal limits when diagnostic components were used and was obliterated when non-diagnostic components were used. By contrast, in experiment 2, when the face-exposure duration was extended (2000 ms), the beneficial effect of the diagnostic matching was diminished as was the detrimental effect of the non-diagnostic matching. These data highlight the impact of diagnostic facial features in normal expression recognition and suggest that impaired emotion recognition in DVA results from deficient visual integration across diagnostic face components. Copyright © 2012 Elsevier Ltd. All rights reserved.
Orienting to face expression during encoding improves men's recognition of own gender faces.
Fulton, Erika K; Bulluck, Megan; Hertzog, Christopher
2015-10-01
It is unclear why women have superior episodic memory of faces, but the benefit may be partially the result of women engaging in superior processing of facial expressions. Therefore, we hypothesized that orienting instructions to attend to facial expression at encoding would significantly improve men's memory of faces and possibly reduce gender differences. We directed 203 college students (122 women) to study 120 faces under instructions to orient to either the person's gender or their emotional expression. They later took a recognition test of these faces by either judging whether they had previously studied the same person or that person with the exact same expression; the latter test evaluated recollection of specific facial details. Orienting to facial expressions during encoding significantly improved men's recognition of own-gender faces and eliminated the advantage that women had for male faces under gender orienting instructions. Although gender differences in spontaneous strategy use when orienting to faces cannot fully account for gender differences in face recognition, orienting men to facial expression during encoding is one way to significantly improve their episodic memory for male faces. Copyright © 2015 Elsevier B.V. All rights reserved.
[The monodigital recognition test (MRT)--a sensitivity-specific variant of Moberg's pick-up test].
Clemens, R
1979-01-01
The Test for Mono-digital Recognition (MRT) is a procedure specifically designed for the measurement of gnostic performance. The new test was derived from MOBERG's Pick-up Test with the aim of eliminating its shortcomings. Additional injuries do not restrict the applicability of the MRT and the test is not confined to the skin area supplied by the median nerve. The MRT expresses the degree of gnostic disturbances by means of a point system. For testing a patient only two to four minutes are needed.
Hu, T H; Wan, L; Liu, T A; Wang, M W; Chen, T; Wang, Y H
2017-12-01
Deep learning and neural network models have been new research directions and hot issues in the fields of machine learning and artificial intelligence in recent years. Deep learning has made a breakthrough in the applications of image and speech recognitions, and also has been extensively used in the fields of face recognition and information retrieval because of its special superiority. Bone X-ray images express different variations in black-white-gray gradations, which have image features of black and white contrasts and level differences. Based on these advantages of deep learning in image recognition, we combine it with the research of bone age assessment to provide basic datum for constructing a forensic automatic system of bone age assessment. This paper reviews the basic concept and network architectures of deep learning, and describes its recent research progress on image recognition in different research fields at home and abroad, and explores its advantages and application prospects in bone age assessment. Copyright© by the Editorial Department of Journal of Forensic Medicine.
Tell, Dina; Davidson, Denise; Camras, Linda A.
2014-01-01
Eye gaze direction and expression intensity effects on emotion recognition in children with autism disorder and typically developing children were investigated. Children with autism disorder and typically developing children identified happy and angry expressions equally well. Children with autism disorder, however, were less accurate in identifying fear expressions across intensities and eye gaze directions. Children with autism disorder rated expressions with direct eyes, and 50% expressions, as more intense than typically developing children. A trend was also found for sad expressions, as children with autism disorder were less accurate in recognizing sadness at 100% intensity with direct eyes than typically developing children. Although the present research showed that children with autism disorder are sensitive to eye gaze direction, impairments in the recognition of fear, and possibly sadness, exist. Furthermore, children with autism disorder and typically developing children perceive the intensity of emotional expressions differently. PMID:24804098
Role of pattern recognition receptors of the neurovascular unit in inflamm-aging.
Wilhelm, Imola; Nyúl-Tóth, Ádám; Kozma, Mihály; Farkas, Attila E; Krizbai, István A
2017-11-01
Aging is associated with chronic inflammation partly mediated by increased levels of damage-associated molecular patterns, which activate pattern recognition receptors (PRRs) of the innate immune system. Furthermore, many aging-related disorders are associated with inflammation. PRRs, such as Toll-like receptors (TLRs) and nucleotide-binding oligomerization domain-like receptors (NLRs), are expressed not only in cells of the innate immune system but also in other cells, including cells of the neurovascular unit and cerebral vasculature forming the blood-brain barrier. In this review, we summarize our present knowledge about the relationship between activation of PRRs expressed by cells of the neurovascular unit-blood-brain barrier, chronic inflammation, and aging-related pathologies of the brain. The most important damage-associated molecular pattern-sensing PRRs in the brain are TLR2, TLR4, and NLR family pyrin domain-containing protein-1 and pyrin domain-containing protein-3, which are activated during physiological and pathological aging in microglia, neurons, astrocytes, and possibly endothelial cells and pericytes. Copyright © 2017 the American Physiological Society.
Automatic decoding of facial movements reveals deceptive pain expressions
Bartlett, Marian Stewart; Littlewort, Gwen C.; Frank, Mark G.; Lee, Kang
2014-01-01
Summary In highly social species such as humans, faces have evolved to convey rich information for social interaction, including expressions of emotions and pain [1–3]. Two motor pathways control facial movement [4–7]. A subcortical extrapyramidal motor system drives spontaneous facial expressions of felt emotions. A cortical pyramidal motor system controls voluntary facial expressions. The pyramidal system enables humans to simulate facial expressions of emotions not actually experienced. Their simulation is so successful that they can deceive most observers [8–11]. Machine vision may, however, be able to distinguish deceptive from genuine facial signals by identifying the subtle differences between pyramidally and extrapyramidally driven movements. Here we show that human observers could not discriminate real from faked expressions of pain better than chance, and after training, improved accuracy to a modest 55%. However a computer vision system that automatically measures facial movements and performs pattern recognition on those movements attained 85% accuracy. The machine system’s superiority is attributable to its ability to differentiate the dynamics of genuine from faked expressions. Thus by revealing the dynamics of facial action through machine vision systems, our approach has the potential to elucidate behavioral fingerprints of neural control systems involved in emotional signaling. PMID:24656830
[Neural mechanisms of facial recognition].
Nagai, Chiyoko
2007-01-01
We review recent researches in neural mechanisms of facial recognition in the light of three aspects: facial discrimination and identification, recognition of facial expressions, and face perception in itself. First, it has been demonstrated that the fusiform gyrus has a main role of facial discrimination and identification. However, whether the FFA (fusiform face area) is really a special area for facial processing or not is controversial; some researchers insist that the FFA is related to 'becoming an expert' for some kinds of visual objects, including faces. Neural mechanisms of prosopagnosia would be deeply concerned to this issue. Second, the amygdala seems to be very concerned to recognition of facial expressions, especially fear. The amygdala, connected with the superior temporal sulcus and the orbitofrontal cortex, appears to operate the cortical function. The amygdala and the superior temporal sulcus are related to gaze recognition, which explains why a patient with bilateral amygdala damage could not recognize only a fear expression; the information from eyes is necessary for fear recognition. Finally, even a newborn infant can recognize a face as a face, which is congruent with the innate hypothesis of facial recognition. Some researchers speculate that the neural basis of such face perception is the subcortical network, comprised of the amygdala, the superior colliculus, and the pulvinar. This network would relate to covert recognition that prosopagnosic patients have.
Richter, Anni; Barman, Adriana; Wüstenberg, Torsten; Soch, Joram; Schanze, Denny; Deibele, Anna; Behnisch, Gusalija; Assmann, Anne; Klein, Marieke; Zenker, Martin; Seidenbecher, Constanze; Schott, Björn H.
2017-01-01
Dopamine is critically important in the neural manifestation of motivated behavior, and alterations in the human dopaminergic system have been implicated in the etiology of motivation-related psychiatric disorders, most prominently addiction. Patients with chronic addiction exhibit reduced dopamine D2 receptor (DRD2) availability in the striatum, and the DRD2 TaqIA (rs1800497) and C957T (rs6277) genetic polymorphisms have previously been linked to individual differences in striatal dopamine metabolism and clinical risk for alcohol and nicotine dependence. Here, we investigated the hypothesis that the variants of these polymorphisms would show increased reward-related memory formation, which has previously been shown to jointly engage the mesolimbic dopaminergic system and the hippocampus, as a potential intermediate phenotype for addiction memory. To this end, we performed functional magnetic resonance imaging (fMRI) in 62 young, healthy individuals genotyped for DRD2 TaqIA and C957T variants. Participants performed an incentive delay task, followed by a recognition memory task 24 h later. We observed effects of both genotypes on the overall recognition performance with carriers of low-expressing variants, namely TaqIA A1 carriers and C957T C homozygotes, showing better performance than the other genotype groups. In addition to the better memory performance, C957T C homozygotes also exhibited a response bias for cues predicting monetary reward. At the neural level, the C957T polymorphism was associated with a genotype-related modulation of right hippocampal and striatal fMRI responses predictive of subsequent recognition confidence for reward-predicting items. Our results indicate that genetic variations associated with DRD2 expression affect explicit memory, specifically for rewarded stimuli. We suggest that the relatively better memory for rewarded stimuli in carriers of low-expressing DRD2 variants may reflect an intermediate phenotype of addiction memory. PMID:28507526
Win-Shwe, Tin-Tin; Fujitani, Yuji; Kyi-Tha-Thu, Chaw; Furuyama, Akiko; Michikawa, Takehiro; Tsukahara, Shinji; Nitta, Hiroshi; Hirano, Seishiro
2014-01-01
Epidemiological studies have reported an increased risk of cardiopulmonary and lung cancer mortality associated with increasing exposure to air pollution. Ambient particulate matter consists of primary particles emitted directly from diesel engine vehicles and secondary organic aerosols (SOAs) are formed by oxidative reaction of the ultrafine particle components of diesel exhaust (DE) in the atmosphere. However, little is known about the relationship between exposure to SOA and central nervous system functions. Recently, we have reported that an acute single intranasal instillation of SOA may induce inflammatory response in lung, but not in brain of adult mice. To clarify the whole body exposure effects of SOA on central nervous system functions, we first created inhalation chambers for diesel exhaust origin secondary organic aerosols (DE-SOAs) produced by oxidation of diesel exhaust particles caused by adding ozone. Male BALB/c mice were exposed to clean air (control), DE and DE-SOA in inhalation chambers for one or three months (5 h/day, 5 days/week) and were examined for memory function using a novel object recognition test and for memory function-related gene expressions in the hippocampus by real-time RT-PCR. Moreover, female mice exposed to DE-SOA for one month were mated and maternal behaviors and the related gene expressions in the hypothalamus examined. Novel object recognition ability and N-methyl-d-aspartate (NMDA) receptor expression in the hippocampus were affected in male mice exposed to DE-SOA. Furthermore, a tendency to decrease maternal performance and significantly decreased expression levels of estrogen receptor (ER)-α, and oxytocin receptor were found in DE-SOA exposed dams compared with the control. This is the first study of this type and our results suggest that the constituents of DE-SOA may be associated with memory function and maternal performance based on the impaired gene expressions in the hippocampus and hypothalamus, respectively. PMID:25361045
ERIC Educational Resources Information Center
Evers, Kris; Steyaert, Jean; Noens, Ilse; Wagemans, Johan
2015-01-01
Emotion labelling was evaluated in two matched samples of 6-14-year old children with and without an autism spectrum disorder (ASD; N = 45 and N = 50, resp.), using six dynamic facial expressions. The Emotion Recognition Task proved to be valuable demonstrating subtle emotion recognition difficulties in ASD, as we showed a general poorer emotion…
Facial recognition using multisensor images based on localized kernel eigen spaces.
Gundimada, Satyanadh; Asari, Vijayan K
2009-06-01
A feature selection technique along with an information fusion procedure for improving the recognition accuracy of a visual and thermal image-based facial recognition system is presented in this paper. A novel modular kernel eigenspaces approach is developed and implemented on the phase congruency feature maps extracted from the visual and thermal images individually. Smaller sub-regions from a predefined neighborhood within the phase congruency images of the training samples are merged to obtain a large set of features. These features are then projected into higher dimensional spaces using kernel methods. The proposed localized nonlinear feature selection procedure helps to overcome the bottlenecks of illumination variations, partial occlusions, expression variations and variations due to temperature changes that affect the visual and thermal face recognition techniques. AR and Equinox databases are used for experimentation and evaluation of the proposed technique. The proposed feature selection procedure has greatly improved the recognition accuracy for both the visual and thermal images when compared to conventional techniques. Also, a decision level fusion methodology is presented which along with the feature selection procedure has outperformed various other face recognition techniques in terms of recognition accuracy.
Exploring Cultural Differences in the Recognition of the Self-Conscious Emotions.
Chung, Joanne M; Robins, Richard W
2015-01-01
Recent research suggests that the self-conscious emotions of embarrassment, shame, and pride have distinct, nonverbal expressions that can be recognized in the United States at above-chance levels. However, few studies have examined the recognition of these emotions in other cultures, and little research has been conducted in Asia. Consequently the cross-cultural generalizability of self-conscious emotions has not been firmly established. Additionally, there is no research that examines cultural variability in the recognition of the self-conscious emotions. Cultural values and exposure to Western culture have been identified as contributors to variability in recognition rates for the basic emotions; we sought to examine this for the self-conscious emotions using the University of California, Davis Set of Emotion Expressions (UCDSEE). The present research examined recognition of the self-conscious emotion expressions in South Korean college students and found that recognition rates were very high for pride, low but above chance for shame, and near zero for embarrassment. To examine what might be underlying the recognition rates we found in South Korea, recognition of self-conscious emotions and several cultural values were examined in a U.S. college student sample of European Americans, Asian Americans, and Asian-born individuals. Emotion recognition rates were generally similar between the European Americans and Asian Americans, and higher than emotion recognition rates for Asian-born individuals. These differences were not explained by cultural values in an interpretable manner, suggesting that exposure to Western culture is a more important mediator than values.
Exploring Cultural Differences in the Recognition of the Self-Conscious Emotions
Chung, Joanne M.; Robins, Richard W.
2015-01-01
Recent research suggests that the self-conscious emotions of embarrassment, shame, and pride have distinct, nonverbal expressions that can be recognized in the United States at above-chance levels. However, few studies have examined the recognition of these emotions in other cultures, and little research has been conducted in Asia. Consequently the cross-cultural generalizability of self-conscious emotions has not been firmly established. Additionally, there is no research that examines cultural variability in the recognition of the self-conscious emotions. Cultural values and exposure to Western culture have been identified as contributors to variability in recognition rates for the basic emotions; we sought to examine this for the self-conscious emotions using the University of California, Davis Set of Emotion Expressions (UCDSEE). The present research examined recognition of the self-conscious emotion expressions in South Korean college students and found that recognition rates were very high for pride, low but above chance for shame, and near zero for embarrassment. To examine what might be underlying the recognition rates we found in South Korea, recognition of self-conscious emotions and several cultural values were examined in a U.S. college student sample of European Americans, Asian Americans, and Asian-born individuals. Emotion recognition rates were generally similar between the European Americans and Asian Americans, and higher than emotion recognition rates for Asian-born individuals. These differences were not explained by cultural values in an interpretable manner, suggesting that exposure to Western culture is a more important mediator than values. PMID:26309215
Estrogenic involvement in social learning, social recognition and pathogen avoidance.
Choleris, Elena; Clipperton-Allen, Amy E; Phan, Anna; Valsecchi, Paola; Kavaliers, Martin
2012-04-01
Sociality comes with specific cognitive skills that allow the proper processing of information about others (social recognition), as well as of information originating from others (social learning). Because sociality and social interactions can also facilitate the spread of infection among individuals the ability to recognize and avoid pathogen threat is also essential. We review here various studies primarily from the rodent literature supporting estrogenic involvement in the regulation of social recognition, social learning (socially acquired food preferences and mate choice copying) and the recognition and avoidance of infected and potentially infected individuals. We consider both genomic and rapid estrogenic effects involving estrogen receptors α and β, and G-protein coupled estrogen receptor 1, along with their interactions with neuropeptide systems in the processing of social stimuli and the regulation and expression of these various socially relevant behaviors. Copyright © 2012 Elsevier Inc. All rights reserved.
The automaticity of emotion recognition.
Tracy, Jessica L; Robins, Richard W
2008-02-01
Evolutionary accounts of emotion typically assume that humans evolved to quickly and efficiently recognize emotion expressions because these expressions convey fitness-enhancing messages. The present research tested this assumption in 2 studies. Specifically, the authors examined (a) how quickly perceivers could recognize expressions of anger, contempt, disgust, embarrassment, fear, happiness, pride, sadness, shame, and surprise; (b) whether accuracy is improved when perceivers deliberate about each expression's meaning (vs. respond as quickly as possible); and (c) whether accurate recognition can occur under cognitive load. Across both studies, perceivers quickly and efficiently (i.e., under cognitive load) recognized most emotion expressions, including the self-conscious emotions of pride, embarrassment, and shame. Deliberation improved accuracy in some cases, but these improvements were relatively small. Discussion focuses on the implications of these findings for the cognitive processes underlying emotion recognition.
Fashioning the Face: Sensorimotor Simulation Contributes to Facial Expression Recognition.
Wood, Adrienne; Rychlowska, Magdalena; Korb, Sebastian; Niedenthal, Paula
2016-03-01
When we observe a facial expression of emotion, we often mimic it. This automatic mimicry reflects underlying sensorimotor simulation that supports accurate emotion recognition. Why this is so is becoming more obvious: emotions are patterns of expressive, behavioral, physiological, and subjective feeling responses. Activation of one component can therefore automatically activate other components. When people simulate a perceived facial expression, they partially activate the corresponding emotional state in themselves, which provides a basis for inferring the underlying emotion of the expresser. We integrate recent evidence in favor of a role for sensorimotor simulation in emotion recognition. We then connect this account to a domain-general understanding of how sensory information from multiple modalities is integrated to generate perceptual predictions in the brain. Copyright © 2016 Elsevier Ltd. All rights reserved.
Weighted Feature Gaussian Kernel SVM for Emotion Recognition
Jia, Qingxuan
2016-01-01
Emotion recognition with weighted feature based on facial expression is a challenging research topic and has attracted great attention in the past few years. This paper presents a novel method, utilizing subregion recognition rate to weight kernel function. First, we divide the facial expression image into some uniform subregions and calculate corresponding recognition rate and weight. Then, we get a weighted feature Gaussian kernel function and construct a classifier based on Support Vector Machine (SVM). At last, the experimental results suggest that the approach based on weighted feature Gaussian kernel function has good performance on the correct rate in emotion recognition. The experiments on the extended Cohn-Kanade (CK+) dataset show that our method has achieved encouraging recognition results compared to the state-of-the-art methods. PMID:27807443
NASA Astrophysics Data System (ADS)
Benitez-Garcia, Gibran; Nakamura, Tomoaki; Kaneko, Masahide
2017-01-01
Darwin was the first one to assert that facial expressions are innate and universal, which are recognized across all cultures. However, recent some cross-cultural studies have questioned this assumed universality. Therefore, this paper presents an analysis of the differences between Western and East-Asian faces of the six basic expressions (anger, disgust, fear, happiness, sadness and surprise) focused on three individual facial regions of eyes-eyebrows, nose and mouth. The analysis is conducted by applying PCA for two feature extraction methods: appearance-based by using the pixel intensities of facial parts, and geometric-based by handling 125 feature points from the face. Both methods are evaluated using 4 standard databases for both racial groups and the results are compared with a cross-cultural human study applied to 20 participants. Our analysis reveals that differences between Westerns and East-Asians exist mainly on the regions of eyes-eyebrows and mouth for expressions of fear and disgust respectively. This work presents important findings for a better design of automatic facial expression recognition systems based on the difference between two racial groups.
Effects of facial emotion recognition remediation on visual scanning of novel face stimuli.
Marsh, Pamela J; Luckett, Gemma; Russell, Tamara; Coltheart, Max; Green, Melissa J
2012-11-01
Previous research shows that emotion recognition in schizophrenia can be improved with targeted remediation that draws attention to important facial features (eyes, nose, mouth). Moreover, the effects of training have been shown to last for up to one month after training. The aim of this study was to investigate whether improved emotion recognition of novel faces is associated with concomitant changes in visual scanning of these same novel facial expressions. Thirty-nine participants with schizophrenia received emotion recognition training using Ekman's Micro-Expression Training Tool (METT), with emotion recognition and visual scanpath (VSP) recordings to face stimuli collected simultaneously. Baseline ratings of interpersonal and cognitive functioning were also collected from all participants. Post-METT training, participants showed changes in foveal attention to the features of facial expressions of emotion not used in METT training, which were generally consistent with the information about important features from the METT. In particular, there were changes in how participants looked at the features of facial expressions of emotion surprise, disgust, fear, happiness, and neutral, demonstrating that improved emotion recognition is paralleled by changes in the way participants with schizophrenia viewed novel facial expressions of emotion. However, there were overall decreases in foveal attention to sad and neutral faces that indicate more intensive instruction might be needed for these faces during training. Most importantly, the evidence shows that participant gender may affect training outcomes. Copyright © 2012 Elsevier B.V. All rights reserved.
Endoribonuclease-Based Two-Component Repressor Systems for Tight Gene Expression Control in Plants
Liang, Yan; Richardson, Sarah; Yan, Jingwei; ...
2017-01-17
Tight control and multifactorial regulation of gene expression are important challenges in genetic engineering and are critical for the development of regulatory circuits. In meeting these challenges we will facilitate transgene expression regulation and support the fine-tuning of metabolic pathways to avoid the accumulation of undesired intermediates. By employing the endoribonuclease Csy4 and its recognition sequence from Pseudomonas aeruginosa and manipulating 5'UTR of mRNA, we developed a two-component expression–repression system to tightly control synthesis of transgene products. We demonstrated that this regulatory device was functional in monocotyledonous and dicotyledonous plant species, and showed that it can be used to repressmore » transgene expression by >400-fold and to synchronize transgene repression. In addition to tissue-specific transgene repression, this system offers stimuli-dependent expression control. Here, we identified 54 orthologous systems from various bacteria, using a bioinformatics approach and then validated in planta the activity for a few of those systems, demonstrating the potential diversity of such a two-component repressor system.« less
Endoribonuclease-Based Two-Component Repressor Systems for Tight Gene Expression Control in Plants
DOE Office of Scientific and Technical Information (OSTI.GOV)
Liang, Yan; Richardson, Sarah; Yan, Jingwei
Tight control and multifactorial regulation of gene expression are important challenges in genetic engineering and are critical for the development of regulatory circuits. In meeting these challenges we will facilitate transgene expression regulation and support the fine-tuning of metabolic pathways to avoid the accumulation of undesired intermediates. By employing the endoribonuclease Csy4 and its recognition sequence from Pseudomonas aeruginosa and manipulating 5'UTR of mRNA, we developed a two-component expression–repression system to tightly control synthesis of transgene products. We demonstrated that this regulatory device was functional in monocotyledonous and dicotyledonous plant species, and showed that it can be used to repressmore » transgene expression by >400-fold and to synchronize transgene repression. In addition to tissue-specific transgene repression, this system offers stimuli-dependent expression control. Here, we identified 54 orthologous systems from various bacteria, using a bioinformatics approach and then validated in planta the activity for a few of those systems, demonstrating the potential diversity of such a two-component repressor system.« less
Toll-like receptors participate in Naegleria fowleri recognition.
Martínez-Castillo, Moisés; Santos-Argumedo, Leopoldo; Galván-Moroyoqui, José Manuel; Serrano-Luna, Jesús; Shibayama, Mineko
2018-01-01
Naegleria fowleri is a protozoan that invades the central nervous system and causes primary amoebic meningoencephalitis. It has been reported that N. fowleri induces an important inflammatory response during the infection. In the present study, we evaluated the roles of Toll-like receptors in the recognition of N. fowleri trophozoites by human mucoepithelial cells, analyzing the expression and production of innate immune response mediators. After amoebic interactions with NCI-H292 cells, the expression and production levels of IL-8, TNF-α, IL-1β, and human beta defensin-2 were evaluated by RT-PCR, ELISA, immunofluorescence, and dot blot assays, respectively. To determine whether the canonical signaling pathways were engaged, we used different inhibitors, namely, IMG-2005 for MyD88 and BAY 11-7085 for the nuclear factor NFkB. Our results showed that the expression and production of the pro-inflammatory cytokines and beta defensin-2 were induced by N. fowleri mainly through the canonical TLR4 pathway in a time-dependent manner.
Huang, Charles Lung-Cheng; Hsiao, Sigmund; Hwu, Hai-Gwo; Howng, Shen-Long
2012-12-30
The Chinese Facial Emotion Recognition Database (CFERD), a computer-generated three-dimensional (3D) paradigm, was developed to measure the recognition of facial emotional expressions at different intensities. The stimuli consisted of 3D colour photographic images of six basic facial emotional expressions (happiness, sadness, disgust, fear, anger and surprise) and neutral faces of the Chinese. The purpose of the present study is to describe the development and validation of CFERD with nonclinical healthy participants (N=100; 50 men; age ranging between 18 and 50 years), and to generate normative data set. The results showed that the sensitivity index d' [d'=Z(hit rate)-Z(false alarm rate), where function Z(p), p∈[0,1
Quantifying facial expression signal and intensity use during development.
Rodger, Helen; Lao, Junpeng; Caldara, Roberto
2018-06-12
Behavioral studies investigating facial expression recognition during development have applied various methods to establish by which age emotional expressions can be recognized. Most commonly, these methods employ static images of expressions at their highest intensity (apex) or morphed expressions of different intensities, but they have not previously been compared. Our aim was to (a) quantify the intensity and signal use for recognition of six emotional expressions from early childhood to adulthood and (b) compare both measures and assess their functional relationship to better understand the use of different measures across development. Using a psychophysical approach, we isolated the quantity of signal necessary to recognize an emotional expression at full intensity and the quantity of expression intensity (using neutral expression image morphs of varying intensities) necessary for each observer to recognize the six basic emotions while maintaining performance at 75%. Both measures revealed that fear and happiness were the most difficult and easiest expressions to recognize across age groups, respectively, a pattern already stable during early childhood. The quantity of signal and intensity needed to recognize sad, angry, disgust, and surprise expressions decreased with age. Using a Bayesian update procedure, we then reconstructed the response profiles for both measures. This analysis revealed that intensity and signal processing are similar only during adulthood and, therefore, cannot be straightforwardly compared during development. Altogether, our findings offer novel methodological and theoretical insights and tools for the investigation of the developing affective system. Copyright © 2018 Elsevier Inc. All rights reserved.
McKinney, J D
1989-01-01
Molecular/theoretical modeling studies have revealed that thyroid hormones and toxic chlorinated aromatic hydrocarbons of environmental significance (for which dioxin or TCDD is the prototype) have similar structural properties that could be important in molecular recognition in biochemical systems. These molecular properties include a somewhat rigid, sterically accessible and polarizable aromatic ring and size-limited, hydrophobic lateral substituents, usually contained in opposite adjoining rings of a diphenyl compound. These molecular properties define the primary binding groups thought to be important in molecular recognition of both types of structures in biochemical systems. Similar molecular reactivities are supported by the demonstration of effective specific binding of thyroid hormones and chlorinated aromatic hydrocarbons with four different proteins, enzymes, or receptor preparations that are known or suspected to be involved in the expression of thyroid hormone activity. These binding interactions represent both aromatic-aromatic (stacking) and molecular cleft-type recognition processes. A multiple protein or multifunctional receptor-ligand binding mechanism model is proposed as a way of visualizing the details and possible role of both the stacking and cleft type molecular recognition factors in the expression of biological activity. The model suggests a means by which hormone-responsive effector-linked sites (possible protein-protein-DNA complexes) can maintain highly structurally specific control of hormone action. Finally, the model also provides a theoretical basis for the design and conduct of further biological experimentation on the molecular mechanism(s) of action of toxic chlorinated aromatic hydrocarbons and thyroid hormones. Images FIGURE 3. A FIGURE 3. B FIGURE 3. C FIGURE 3. D PMID:2551666
A New Method of Facial Expression Recognition Based on SPE Plus SVM
NASA Astrophysics Data System (ADS)
Ying, Zilu; Huang, Mingwei; Wang, Zhen; Wang, Zhewei
A novel method of facial expression recognition (FER) is presented, which uses stochastic proximity embedding (SPE) for data dimension reduction, and support vector machine (SVM) for expression classification. The proposed algorithm is applied to Japanese Female Facial Expression (JAFFE) database for FER, better performance is obtained compared with some traditional algorithms, such as PCA and LDA etc.. The result have further proved the effectiveness of the proposed algorithm.
NASA Astrophysics Data System (ADS)
Duman, M.; Pfleger, M.; Zhu, R.; Rankl, C.; Chtcheglova, L. A.; Neundlinger, I.; Bozna, B. L.; Mayer, B.; Salio, M.; Shepherd, D.; Polzella, P.; Moertelmaier, M.; Kada, G.; Ebner, A.; Dieudonne, M.; Schütz, G. J.; Cerundolo, V.; Kienberger, F.; Hinterdorfer, P.
2010-03-01
The combination of fluorescence microscopy and atomic force microscopy has a great potential in single-molecule-detection applications, overcoming many of the limitations coming from each individual technique. Here we present a new platform of combined fluorescence and simultaneous topography and recognition imaging (TREC) for improved localization of cellular receptors. Green fluorescent protein (GFP) labeled human sodium-glucose cotransporter (hSGLT1) expressed Chinese Hamster Ovary (CHO) cells and endothelial cells (MyEnd) from mouse myocardium stained with phalloidin-rhodamine were used as cell systems to study AFM topography and fluorescence microscopy on the same surface area. Topographical AFM images revealed membrane features such as lamellipodia, cytoskeleton fibers, F-actin filaments and small globular structures with heights ranging from 20 to 30 nm. Combined fluorescence and TREC imaging was applied to detect density, distribution and localization of YFP-labeled CD1d molecules on α-galactosylceramide (αGalCer)-loaded THP1 cells. While the expression level, distribution and localization of CD1d molecules on THP1 cells were detected with fluorescence microscopy, the nanoscale distribution of binding sites was investigated with molecular recognition imaging by using a chemically modified AFM tip. Using TREC on the inverted light microscope, the recognition sites of cell receptors were detected in recognition images with domain sizes ranging from ~ 25 to ~ 160 nm, with the smaller domains corresponding to a single CD1d molecule.
Duman, M; Pfleger, M; Zhu, R; Rankl, C; Chtcheglova, L A; Neundlinger, I; Bozna, B L; Mayer, B; Salio, M; Shepherd, D; Polzella, P; Moertelmaier, M; Kada, G; Ebner, A; Dieudonne, M; Schütz, G J; Cerundolo, V; Kienberger, F; Hinterdorfer, P
2010-03-19
The combination of fluorescence microscopy and atomic force microscopy has a great potential in single-molecule-detection applications, overcoming many of the limitations coming from each individual technique. Here we present a new platform of combined fluorescence and simultaneous topography and recognition imaging (TREC) for improved localization of cellular receptors. Green fluorescent protein (GFP) labeled human sodium-glucose cotransporter (hSGLT1) expressed Chinese Hamster Ovary (CHO) cells and endothelial cells (MyEnd) from mouse myocardium stained with phalloidin-rhodamine were used as cell systems to study AFM topography and fluorescence microscopy on the same surface area. Topographical AFM images revealed membrane features such as lamellipodia, cytoskeleton fibers, F-actin filaments and small globular structures with heights ranging from 20 to 30 nm. Combined fluorescence and TREC imaging was applied to detect density, distribution and localization of YFP-labeled CD1d molecules on alpha-galactosylceramide (alphaGalCer)-loaded THP1 cells. While the expression level, distribution and localization of CD1d molecules on THP1 cells were detected with fluorescence microscopy, the nanoscale distribution of binding sites was investigated with molecular recognition imaging by using a chemically modified AFM tip. Using TREC on the inverted light microscope, the recognition sites of cell receptors were detected in recognition images with domain sizes ranging from approximately 25 to approximately 160 nm, with the smaller domains corresponding to a single CD1d molecule.
Individual differences in the recognition of facial expressions: an event-related potentials study.
Tamamiya, Yoshiyuki; Hiraki, Kazuo
2013-01-01
Previous studies have shown that early posterior components of event-related potentials (ERPs) are modulated by facial expressions. The goal of the current study was to investigate individual differences in the recognition of facial expressions by examining the relationship between ERP components and the discrimination of facial expressions. Pictures of 3 facial expressions (angry, happy, and neutral) were presented to 36 young adults during ERP recording. Participants were asked to respond with a button press as soon as they recognized the expression depicted. A multiple regression analysis, where ERP components were set as predictor variables, assessed hits and reaction times in response to the facial expressions as dependent variables. The N170 amplitudes significantly predicted for accuracy of angry and happy expressions, and the N170 latencies were predictive for accuracy of neutral expressions. The P2 amplitudes significantly predicted reaction time. The P2 latencies significantly predicted reaction times only for neutral faces. These results suggest that individual differences in the recognition of facial expressions emerge from early components in visual processing.
Recognition and Posing of Emotional Expressions by Abused Children and Their Mothers.
ERIC Educational Resources Information Center
Camras, Linda A.; And Others
1988-01-01
A total of 20 abused and 20 nonabused pairs of children of three-seven years and their mothers participated in a facial expression posing task and a facial expression recognition task. Findings suggest that abused children may not observe as often as nonabused children do the easily interpreted voluntary displays of emotion by their mothers. (RH)
ERIC Educational Resources Information Center
Brenna, Viola; Proietti, Valentina; Montirosso, Rosario; Turati, Chiara
2013-01-01
The current study examined whether and how the presence of a positive or a negative emotional expression may affect the face recognition process at 3 months of age. Using a familiarization procedure, Experiment 1 demonstrated that positive (i.e., happiness), but not negative (i.e., fear and anger) facial expressions facilitate infants' ability to…
Automatic integration of social information in emotion recognition.
Mumenthaler, Christian; Sander, David
2015-04-01
This study investigated the automaticity of the influence of social inference on emotion recognition. Participants were asked to recognize dynamic facial expressions of emotion (fear or anger in Experiment 1 and blends of fear and surprise or of anger and disgust in Experiment 2) in a target face presented at the center of a screen while a subliminal contextual face appearing in the periphery expressed an emotion (fear or anger) or not (neutral) and either looked at the target face or not. Results of Experiment 1 revealed that recognition of the target emotion of fear was improved when a subliminal angry contextual face gazed toward-rather than away from-the fearful face. We replicated this effect in Experiment 2, in which facial expression blends of fear and surprise were more often and more rapidly categorized as expressing fear when the subliminal contextual face expressed anger and gazed toward-rather than away from-the target face. With the contextual face appearing for 30 ms in total, including only 10 ms of emotion expression, and being immediately masked, our data provide the first evidence that social influence on emotion recognition can occur automatically. (c) 2015 APA, all rights reserved).
Jürgens, Rebecca; Grass, Annika; Drolet, Matthis; Fischer, Julia
Both in the performative arts and in emotion research, professional actors are assumed to be capable of delivering emotions comparable to spontaneous emotional expressions. This study examines the effects of acting training on vocal emotion depiction and recognition. We predicted that professional actors express emotions in a more realistic fashion than non-professional actors. However, professional acting training may lead to a particular speech pattern; this might account for vocal expressions by actors that are less comparable to authentic samples than the ones by non-professional actors. We compared 80 emotional speech tokens from radio interviews with 80 re-enactments by professional and inexperienced actors, respectively. We analyzed recognition accuracies for emotion and authenticity ratings and compared the acoustic structure of the speech tokens. Both play-acted conditions yielded similar recognition accuracies and possessed more variable pitch contours than the spontaneous recordings. However, professional actors exhibited signs of different articulation patterns compared to non-trained speakers. Our results indicate that for emotion research, emotional expressions by professional actors are not better suited than those from non-actors.
Soulé, Jonathan; Penke, Zsuzsa; Kanhema, Tambudzai; Alme, Maria Nordheim; Laroche, Serge; Bramham, Clive R.
2008-01-01
Long-term recognition memory requires protein synthesis, but little is known about the coordinate regulation of specific genes. Here, we examined expression of the plasticity-associated immediate early genes (Arc, Zif268, and Narp) in the dentate gyrus following long-term object-place recognition learning in rats. RT-PCR analysis from dentate gyrus tissue collected shortly after training did not reveal learning-specific changes in Arc mRNA expression. In situ hybridization and immunohistochemistry were therefore used to assess possible sparse effects on gene expression. Learning about objects increased the density of granule cells expressing Arc, and to a lesser extent Narp, specifically in the dorsal blade of the dentate gyrus, while Zif268 expression was elevated across both blades. Thus, object-place recognition triggers rapid, blade-specific upregulation of plasticity-associated immediate early genes. Furthermore, Western blot analysis of dentate gyrus homogenates demonstrated concomitant upregulation of three postsynaptic density proteins (Arc, PSD-95, and α-CaMKII) with key roles in long-term synaptic plasticity and long-term memory. PMID:19190776
Kobayashi, Kazuya; Matsuura, Kenji
2017-01-01
Insects protect themselves from microbial infections through innate immune responses, including pathogen recognition, phagocytosis, the activation of proteolytic cascades, and the synthesis of antimicrobial peptides. Termites, eusocial insects inhabiting microbe-rich wood, live in closely-related family groups that are susceptible to shared pathogen infections. To resist pathogenic infection, termite families have evolved diverse immune adaptations at both individual and societal levels, and a strategy of trade-offs between reproduction and immunity has been suggested. Although termite immune-inducible genes have been identified, few studies have investigated the differential expression of these genes between reproductive and neuter castes, and between sexes in each caste. In this study, we compared the expression levels of immune-related genes among castes, sexes, and ages in a Japanese subterranean termite, Reticulitermes speratus. Using RNA-seq, we found 197 immune-related genes, including 40 pattern recognition proteins, 97 signalling proteins, 60 effectors. Among these genes, 174 showed differential expression among castes. Comparing expression levels between males and females in each caste, we found sexually dimorphic expression of immune-related genes not only in reproductive castes, but also in neuter castes. Moreover, we identified age-related differential expression of 162 genes in male and/or female reproductives. In addition, although R. speratus is known to use the antibacterial peptide C-type lysozyme as an egg recognition pheromone, we determined that R. speratus has not only C-type, but also P-type and I-type lysozymes, as well as other termite species. Our transcriptomic analyses revealed immune response plasticity among all castes, and sex-biased expression of immune genes even in neuter castes, suggesting a sexual division of labor in the immune system of R. speratus. This study heightens the understanding of the evolution of antimicrobial strategies in eusocial insects, and of sexual roles in insect societies as a whole. PMID:28410430
Mitaka, Yuki; Kobayashi, Kazuya; Matsuura, Kenji
2017-01-01
Insects protect themselves from microbial infections through innate immune responses, including pathogen recognition, phagocytosis, the activation of proteolytic cascades, and the synthesis of antimicrobial peptides. Termites, eusocial insects inhabiting microbe-rich wood, live in closely-related family groups that are susceptible to shared pathogen infections. To resist pathogenic infection, termite families have evolved diverse immune adaptations at both individual and societal levels, and a strategy of trade-offs between reproduction and immunity has been suggested. Although termite immune-inducible genes have been identified, few studies have investigated the differential expression of these genes between reproductive and neuter castes, and between sexes in each caste. In this study, we compared the expression levels of immune-related genes among castes, sexes, and ages in a Japanese subterranean termite, Reticulitermes speratus. Using RNA-seq, we found 197 immune-related genes, including 40 pattern recognition proteins, 97 signalling proteins, 60 effectors. Among these genes, 174 showed differential expression among castes. Comparing expression levels between males and females in each caste, we found sexually dimorphic expression of immune-related genes not only in reproductive castes, but also in neuter castes. Moreover, we identified age-related differential expression of 162 genes in male and/or female reproductives. In addition, although R. speratus is known to use the antibacterial peptide C-type lysozyme as an egg recognition pheromone, we determined that R. speratus has not only C-type, but also P-type and I-type lysozymes, as well as other termite species. Our transcriptomic analyses revealed immune response plasticity among all castes, and sex-biased expression of immune genes even in neuter castes, suggesting a sexual division of labor in the immune system of R. speratus. This study heightens the understanding of the evolution of antimicrobial strategies in eusocial insects, and of sexual roles in insect societies as a whole.
Calvo, Manuel G; Nummenmaa, Lauri
2009-12-01
Happy, surprised, disgusted, angry, sad, fearful, and neutral faces were presented extrafoveally, with fixations on faces allowed or not. The faces were preceded by a cue word that designated the face to be saccaded in a two-alternative forced-choice discrimination task (2AFC; Experiments 1 and 2), or were followed by a probe word for recognition (Experiment 3). Eye tracking was used to decompose the recognition process into stages. Relative to the other expressions, happy faces (1) were identified faster (as early as 160 msec from stimulus onset) in extrafoveal vision, as revealed by shorter saccade latencies in the 2AFC task; (2) required less encoding effort, as indexed by shorter first fixations and dwell times; and (3) required less decision-making effort, as indicated by fewer refixations on the face after the recognition probe was presented. This reveals a happy-face identification advantage both prior to and during overt attentional processing. The results are discussed in relation to prior neurophysiological findings on latencies in facial expression recognition.
Food-Induced Emotional Resonance Improves Emotion Recognition.
Pandolfi, Elisa; Sacripante, Riccardo; Cardini, Flavia
2016-01-01
The effect of food substances on emotional states has been widely investigated, showing, for example, that eating chocolate is able to reduce negative mood. Here, for the first time, we have shown that the consumption of specific food substances is not only able to induce particular emotional states, but more importantly, to facilitate recognition of corresponding emotional facial expressions in others. Participants were asked to perform an emotion recognition task before and after eating either a piece of chocolate or a small amount of fish sauce-which we expected to induce happiness or disgust, respectively. Our results showed that being in a specific emotional state improves recognition of the corresponding emotional facial expression. Indeed, eating chocolate improved recognition of happy faces, while disgusted expressions were more readily recognized after eating fish sauce. In line with the embodied account of emotion understanding, we suggest that people are better at inferring the emotional state of others when their own emotional state resonates with the observed one.
Food-Induced Emotional Resonance Improves Emotion Recognition
Pandolfi, Elisa; Sacripante, Riccardo; Cardini, Flavia
2016-01-01
The effect of food substances on emotional states has been widely investigated, showing, for example, that eating chocolate is able to reduce negative mood. Here, for the first time, we have shown that the consumption of specific food substances is not only able to induce particular emotional states, but more importantly, to facilitate recognition of corresponding emotional facial expressions in others. Participants were asked to perform an emotion recognition task before and after eating either a piece of chocolate or a small amount of fish sauce—which we expected to induce happiness or disgust, respectively. Our results showed that being in a specific emotional state improves recognition of the corresponding emotional facial expression. Indeed, eating chocolate improved recognition of happy faces, while disgusted expressions were more readily recognized after eating fish sauce. In line with the embodied account of emotion understanding, we suggest that people are better at inferring the emotional state of others when their own emotional state resonates with the observed one. PMID:27973559
Effects of the duration of expressions on the recognition of microexpressions*
Shen, Xun-bing; Wu, Qi; Fu, Xiao-lan
2012-01-01
Objective: The purpose of this study was to investigate the effects of the duration of expressions on the recognition of microexpressions, which are closely related to deception. Methods: In two experiments, participants were briefly (from 20 to 300 ms) shown one of six basic expressions and then were asked to identify the expression. Results: The results showed that the participants’ performance in recognition of microexpressions increased with the duration of the expressions, reaching a turning point at 200 ms before levelling off. The results also indicated that practice could improve the participants’ performance. Conclusions: The results of this study suggest that the proper upper limit of the duration of microexpressions might be around 1/5 of a second and confirmed that the ability to recognize microexpressions can be enhanced with practice. PMID:22374615
2010-03-01
allows the programmer to use the English language in an expressive manor while still maintaining the logical structure of a programming language ( Pressman ...and Choudhury Tanzeem. 2000. Face Recognition for Smart Environments, IEEE Computer, pp. 50–55. Pressman , Roger. 2010. Software Engineering A
Model of Emotional Expressions in Movements
ERIC Educational Resources Information Center
Rozaliev, Vladimir L.; Orlova, Yulia A.
2013-01-01
This paper presents a new approach to automated identification of human emotions based on analysis of body movements, a recognition of gestures and poses. Methodology, models and automated system for emotion identification are considered. To characterize the person emotions in the model, body movements are described with linguistic variables and a…
Wang, Xu; Song, Yiying; Zhen, Zonglei; Liu, Jia
2016-05-01
Face perception is essential for daily and social activities. Neuroimaging studies have revealed a distributed face network (FN) consisting of multiple regions that exhibit preferential responses to invariant or changeable facial information. However, our understanding about how these regions work collaboratively to facilitate facial information processing is limited. Here, we focused on changeable facial information processing, and investigated how the functional integration of the FN is related to the performance of facial expression recognition. To do so, we first defined the FN as voxels that responded more strongly to faces than objects, and then used a voxel-based global brain connectivity method based on resting-state fMRI to characterize the within-network connectivity (WNC) of each voxel in the FN. By relating the WNC and performance in the "Reading the Mind in the Eyes" Test across participants, we found that individuals with stronger WNC in the right posterior superior temporal sulcus (rpSTS) were better at recognizing facial expressions. Further, the resting-state functional connectivity (FC) between the rpSTS and right occipital face area (rOFA), early visual cortex (EVC), and bilateral STS were positively correlated with the ability of facial expression recognition, and the FCs of EVC-pSTS and OFA-pSTS contributed independently to facial expression recognition. In short, our study highlights the behavioral significance of intrinsic functional integration of the FN in facial expression processing, and provides evidence for the hub-like role of the rpSTS for facial expression recognition. Hum Brain Mapp 37:1930-1940, 2016. © 2016 Wiley Periodicals, Inc. © 2016 Wiley Periodicals, Inc.
Face recognition using facial expression: a novel approach
NASA Astrophysics Data System (ADS)
Singh, Deepak Kumar; Gupta, Priya; Tiwary, U. S.
2008-04-01
Facial expressions are undoubtedly the most effective nonverbal communication. The face has always been the equation of a person's identity. The face draws the demarcation line between identity and extinction. Each line on the face adds an attribute to the identity. These lines become prominent when we experience an emotion and these lines do not change completely with age. In this paper we have proposed a new technique for face recognition which focuses on the facial expressions of the subject to identify his face. This is a grey area on which not much light has been thrown earlier. According to earlier researches it is difficult to alter the natural expression. So our technique will be beneficial for identifying occluded or intentionally disguised faces. The test results of the experiments conducted prove that this technique will give a new direction in the field of face recognition. This technique will provide a strong base to the area of face recognition and will be used as the core method for critical defense security related issues.
Fink, Elian; de Rosnay, Marc; Wierda, Marlies; Koot, Hans M; Begeer, Sander
2014-09-01
The empirical literature has presented inconsistent evidence for deficits in the recognition of basic emotion expressions in children with autism spectrum disorders (ASD), which may be due to the focus on research with relatively small sample sizes. Additionally, it is proposed that although children with ASD may correctly identify emotion expression they rely on more deliberate, more time-consuming strategies in order to accurately recognize emotion expressions when compared to typically developing children. In the current study, we examine both emotion recognition accuracy and response time in a large sample of children, and explore the moderating influence of verbal ability on these findings. The sample consisted of 86 children with ASD (M age = 10.65) and 114 typically developing children (M age = 10.32) between 7 and 13 years of age. All children completed a pre-test (emotion word-word matching), and test phase consisting of basic emotion recognition, whereby they were required to match a target emotion expression to the correct emotion word; accuracy and response time were recorded. Verbal IQ was controlled for in the analyses. We found no evidence of a systematic deficit in emotion recognition accuracy or response time for children with ASD, controlling for verbal ability. However, when controlling for children's accuracy in word-word matching, children with ASD had significantly lower emotion recognition accuracy when compared to typically developing children. The findings suggest that the social impairments observed in children with ASD are not the result of marked deficits in basic emotion recognition accuracy or longer response times. However, children with ASD may be relying on other perceptual skills (such as advanced word-word matching) to complete emotion recognition tasks at a similar level as typically developing children.
Postsurgical Disfigurement Influences Disgust Recognition: A Case-Control Study.
Lisan, Quentin; George, Nathalie; Hans, Stephane; Laccourreye, Ollivier; Lemogne, Cédric
Little is known about how emotion recognition may be modified in individuals prone to elicit disgust. We sought to determine if subjects with total laryngectomy would present a modified recognition of facial expressions of disgust. A total of 29 patients presenting with a history of advanced-stage laryngeal cancer were recruited, 17 being surgically treated (total laryngectomy) and 12 treated with chemoradiation therapy only. Based on a validated set of images of facial expressions of fear, disgust, surprise, happiness, sadness and anger displayed by 6 actors, we presented participants with expressions of each emotion at 5 levels of increasing intensity and measured their ability to recognize these emotions. Participants with (vs without) laryngectomy showed a higher threshold for the recognition of disgust (3.2. vs 2.7 images needed before emotion recognition, p = 0.03) and a lower success rate of correct recognition (75.5% vs 88.9%, p = 0.03). Subjects presenting with an aesthetic impairment of the head and neck showed poorer performance in disgust recognition when compared with those without disfigurement. These findings might relate either to some perceptual adaptation, habituation phenomenon, or to some higher-level processes related to emotion regulation strategies. Copyright © 2018 Academy of Consultation-Liaison Psychiatry. Published by Elsevier Inc. All rights reserved.
Facial recognition: a cognitive study of elderly dementia patients and normal older adults.
Zandi, T; Cooper, M; Garrison, L
1992-01-01
Dementia patients' and normal elderlies' recognition of familiar, ordinary emotional and facial expressions was tested. In three conditions subjects were required to name the emotions depicted in pictures and to produce them while presented with the verbal labels of the expressions. The dementia patients' best performance occurred when they had access to the verbal labels while viewing the pictures. The major deficiency in facial recognition was found to be dysnomia related. Findings of this study suggest that the connection between the gnostic units of expression and the gnostic units of verbal labeling is not impaired significantly among the dementia patients.
Towards Multimodal Emotion Recognition in E-Learning Environments
ERIC Educational Resources Information Center
Bahreini, Kiavash; Nadolski, Rob; Westera, Wim
2016-01-01
This paper presents a framework (FILTWAM (Framework for Improving Learning Through Webcams And Microphones)) for real-time emotion recognition in e-learning by using webcams. FILTWAM offers timely and relevant feedback based upon learner's facial expressions and verbalizations. FILTWAM's facial expression software module has been developed and…
A self-organized learning strategy for object recognition by an embedded line of attraction
NASA Astrophysics Data System (ADS)
Seow, Ming-Jung; Alex, Ann T.; Asari, Vijayan K.
2012-04-01
For humans, a picture is worth a thousand words, but to a machine, it is just a seemingly random array of numbers. Although machines are very fast and efficient, they are vastly inferior to humans for everyday information processing. Algorithms that mimic the way the human brain computes and learns may be the solution. In this paper we present a theoretical model based on the observation that images of similar visual perceptions reside in a complex manifold in an image space. The perceived features are often highly structured and hidden in a complex set of relationships or high-dimensional abstractions. To model the pattern manifold, we present a novel learning algorithm using a recurrent neural network. The brain memorizes information using a dynamical system made of interconnected neurons. Retrieval of information is accomplished in an associative sense. It starts from an arbitrary state that might be an encoded representation of a visual image and converges to another state that is stable. The stable state is what the brain remembers. In designing a recurrent neural network, it is usually of prime importance to guarantee the convergence in the dynamics of the network. We propose to modify this picture: if the brain remembers by converging to the state representing familiar patterns, it should also diverge from such states when presented with an unknown encoded representation of a visual image belonging to a different category. That is, the identification of an instability mode is an indication that a presented pattern is far away from any stored pattern and therefore cannot be associated with current memories. These properties can be used to circumvent the plasticity-stability dilemma by using the fluctuating mode as an indicator to create new states. We capture this behavior using a novel neural architecture and learning algorithm, in which the system performs self-organization utilizing a stability mode and an instability mode for the dynamical system. Based on this observation we developed a self- organizing line attractor, which is capable of generating new lines in the feature space to learn unrecognized patterns. Experiments performed on UMIST pose database and CMU face expression variant database for face recognition have shown that the proposed nonlinear line attractor is able to successfully identify the individuals and it provided better recognition rate when compared to the state of the art face recognition techniques. Experiments on FRGC version 2 database has also provided excellent recognition rate in images captured in complex lighting environments. Experiments performed on the Japanese female face expression database and Essex Grimace database using the self organizing line attractor have also shown successful expression invariant face recognition. These results show that the proposed model is able to create nonlinear manifolds in a multidimensional feature space to distinguish complex patterns.
Training facial expression production in children on the autism spectrum.
Gordon, Iris; Pierce, Matthew D; Bartlett, Marian S; Tanaka, James W
2014-10-01
Children with autism spectrum disorder (ASD) show deficits in their ability to produce facial expressions. In this study, a group of children with ASD and IQ-matched, typically developing (TD) children were trained to produce "happy" and "angry" expressions with the FaceMaze computer game. FaceMaze uses an automated computer recognition system that analyzes the child's facial expression in real time. Before and after playing the Angry and Happy versions of FaceMaze, children posed "happy" and "angry" expressions. Naïve raters judged the post-FaceMaze "happy" and "angry" expressions of the ASD group as higher in quality than their pre-FaceMaze productions. Moreover, the post-game expressions of the ASD group were rated as equal in quality as the expressions of the TD group.
Embodied emotion impairment in Huntington's Disease.
Trinkler, Iris; Devignevielle, Sévérine; Achaibou, Amal; Ligneul, Romain V; Brugières, Pierre; Cleret de Langavant, Laurent; De Gelder, Beatrice; Scahill, Rachael; Schwartz, Sophie; Bachoud-Lévi, Anne-Catherine
2017-07-01
Theories of embodied cognition suggest that perceiving an emotion involves somatovisceral and motoric re-experiencing. Here we suggest taking such an embodied stance when looking at emotion processing deficits in patients with Huntington's Disease (HD), a neurodegenerative motor disorder. The literature on these patients' emotion recognition deficit has recently been enriched by some reports of impaired emotion expression. The goal of the study was to find out if expression deficits might be linked to a more motoric level of impairment. We used electromyography (EMG) to compare voluntary emotion expression from words to emotion imitation from static face images, and spontaneous emotion mimicry in 28 HD patients and 24 matched controls. For the latter two imitation conditions, an underlying emotion understanding is not imperative (even though performance might be helped by it). EMG measures were compared to emotion recognition and to the capacity to identify and describe emotions using alexithymia questionnaires. Alexithymia questionnaires tap into the more somato-visceral or interoceptive aspects of emotion perception. Furthermore, we correlated patients' expression and recognition scores to cerebral grey matter volume using voxel-based morphometry (VBM). EMG results replicated impaired voluntary emotion expression in HD. Critically, voluntary imitation and spontaneous mimicry were equally impaired and correlated with impaired recognition. By contrast, alexithymia scores were normal, suggesting that emotion representations on the level of internal experience might be spared. Recognition correlated with brain volume in the caudate as well as in areas previously associated with shared action representations, namely somatosensory, posterior parietal, posterior superior temporal sulcus (pSTS) and subcentral sulcus. Together, these findings indicate that in these patients emotion deficits might be tied to the "motoric level" of emotion expression. Such a double-sided recognition and expression impairment may have important consequences, interrupting empathy in nonverbal communication both ways (understanding and being understood), independently of intact internal experience of emotion. Copyright © 2017 Elsevier Ltd. All rights reserved.
Jin, Xing-Kun; Li, Shuang; Guo, Xiao-Nv; Cheng, Lin; Wu, Min-Hao; Tan, Shang-Jian; Zhu, You-Ting; Yu, Ai-Qing; Li, Wei-Wei; Wang, Qun
2013-12-01
The first step of host fighting against pathogens is that pattern recognition receptors recognized pathogen-associated molecular patterns. However, the specificity of recognition within the innate immune molecular of invertebrates remains largely unknown. In the present study, we investigated how invertebrate pattern recognition receptor (PRR) C-type lectins might be involved in the antimicrobial response in crustacean. Based on our previously obtained completed coding regions of EsLecA and EsLecG in Eriocheir sinensis, the recombinant EsLectin proteins were produced via prokaryotic expression system and affinity chromatography. Subsequently, both rEsLecA and rEsLecG were discovered to have wide spectrum binding activities towards microorganisms, and their microbial-binding was calcium-independent. Moreover, the binding activities of both rEsLecA and rEsLecG induced the aggregation against microbial pathogens. Both microorganism growth inhibitory activities assays and antibacterial activities assays revealed their capabilities of suppressing microorganisms growth and directly killing microorganisms respectively. Furthermore, the encapsulation assays signified that both rEsLecA and rEsLecG could stimulate the cellular encapsulation in vitro. Collectively, data presented here demonstrated the successful expression and purification of two C-type lectins proteins in the Chinese mitten crab, and their critical role in the innate immune system of an invertebrate. Copyright © 2013 Elsevier Ltd. All rights reserved.
Structural constraints determine the glycosylation of HIV-1 envelope trimers
Pritchard, Laura K.; Vasiljevic, Snezana; Ozorowski, Gabriel; Seabright, Gemma E.; Cupo, Albert; Ringe, Rajesh; Kim, Helen J.; Sanders, Rogier W.; Doores, Katie J.; Burton, Dennis R.; Wilson, Ian A.; Ward, Andrew B.; Moore, John P.; Crispin, Max
2015-01-01
A highly glycosylated, trimeric envelope glycoprotein (Env) mediates HIV-1 cell entry. The high density and heterogeneity of the glycans shield Env from recognition by the immune system but, paradoxically, many potent broadly neutralizing antibodies (bNAbs) recognize epitopes involving this glycan shield. To better understand Env glycosylation and its role in bNAb recognition, we characterized a soluble, cleaved recombinant trimer (BG505 SOSIP.664) that is a close structural and antigenic mimic of native Env. Large, unprocessed oligomannose-type structures (Man8-9GlcNAc2) are notably prevalent on the gp120 components of the trimer, irrespective of the mammalian cell expression system or the bNAb used for affinity-purification. In contrast, gp41 subunits carry more highly processed glycans. The glycans on uncleaved, non-native oligomeric gp140 proteins are also highly processed. A homogeneous, oligomannose-dominated glycan profile is therefore a hallmark of a native Env conformation and a potential Achilles’ heel that can be exploited for bNAb recognition and vaccine design. PMID:26051934
Muñoz, Pablo C; Aspé, Mauricio A; Contreras, Luis S; Palacios, Adrián G
2010-01-01
Object recognition memory allows discrimination between novel and familiar objects. This kind of memory consists of two components: recollection, which depends on the hippocampus, and familiarity, which depends on the perirhinal cortex (Pcx). The importance of brain-derived neurotrophic factor (BDNF) for recognition memory has already been recognized. Recent evidence suggests that DNA methylation regulates the expression of BDNF and memory. Behavioral and molecular approaches were used to understand the potential contribution of DNA methylation to recognition memory. To that end, rats were tested for their ability to distinguish novel from familiar objects by using a spontaneous object recognition task. Furthermore, the level of DNA methylation was estimated after trials with a methyl-sensitive PCR. We found a significant correlation between performance on the novel object task and the expression of BDNF, negatively in hippocampal slices and positively in perirhinal cortical slices. By contrast, methylation of DNA in CpG island 1 in the promoter of exon 1 in BDNF only correlated in hippocampal slices, but not in the Pxc cortical slices from trained animals. These results suggest that DNA methylation may be involved in the regulation of the BDNF gene during recognition memory, at least in the hippocampus.
Deficits in facial affect recognition among antisocial populations: a meta-analysis.
Marsh, Abigail A; Blair, R J R
2008-01-01
Individuals with disorders marked by antisocial behavior frequently show deficits in recognizing displays of facial affect. Antisociality may be associated with specific deficits in identifying fearful expressions, which would implicate dysfunction in neural structures that subserve fearful expression processing. A meta-analysis of 20 studies was conducted to assess: (a) if antisocial populations show any consistent deficits in recognizing six emotional expressions; (b) beyond any generalized impairment, whether specific fear recognition deficits are apparent; and (c) if deficits in fear recognition are a function of task difficulty. Results show a robust link between antisocial behavior and specific deficits in recognizing fearful expressions. This impairment cannot be attributed solely to task difficulty. These results suggest dysfunction among antisocial individuals in specified neural substrates, namely the amygdala, involved in processing fearful facial affect.
Development of emotional facial recognition in late childhood and adolescence.
Thomas, Laura A; De Bellis, Michael D; Graham, Reiko; LaBar, Kevin S
2007-09-01
The ability to interpret emotions in facial expressions is crucial for social functioning across the lifespan. Facial expression recognition develops rapidly during infancy and improves with age during the preschool years. However, the developmental trajectory from late childhood to adulthood is less clear. We tested older children, adolescents and adults on a two-alternative forced-choice discrimination task using morphed faces that varied in emotional content. Actors appeared to pose expressions that changed incrementally along three progressions: neutral-to-fear, neutral-to-anger, and fear-to-anger. Across all three morph types, adults displayed more sensitivity to subtle changes in emotional expression than children and adolescents. Fear morphs and fear-to-anger blends showed a linear developmental trajectory, whereas anger morphs showed a quadratic trend, increasing sharply from adolescents to adults. The results provide evidence for late developmental changes in emotional expression recognition with some specificity in the time course for distinct emotions.
The Automaticity of Emotional Face-Context Integration
Aviezer, Hillel; Dudarev, Veronica; Bentin, Shlomo; Hassin, Ran R.
2011-01-01
Recent studies have demonstrated that context can dramatically influence the recognition of basic facial expressions, yet the nature of this phenomenon is largely unknown. In the present paper we begin to characterize the underlying process of face-context integration. Specifically, we examine whether it is a relatively controlled or automatic process. In Experiment 1 participants were motivated and instructed to avoid using the context while categorizing contextualized facial expression, or they were led to believe that the context was irrelevant. Nevertheless, they were unable to disregard the context, which exerted a strong effect on their emotion recognition. In Experiment 2, participants categorized contextualized facial expressions while engaged in a concurrent working memory task. Despite the load, the context exerted a strong influence on their recognition of facial expressions. These results suggest that facial expressions and their body contexts are integrated in an unintentional, uncontrollable, and relatively effortless manner. PMID:21707150
A Genetic Approach to Promoter Recognition during Trans Induction of Viral Gene Expression
NASA Astrophysics Data System (ADS)
Coen, Donald M.; Weinheimer, Steven P.; McKnight, Steven L.
1986-10-01
Viral infection of mammalian cells entails the regulated induction of viral gene expression. The induction of many viral genes, including the herpes simplex virus gene encoding thymidine kinase (tk), depends on viral regulatory proteins that act in trans. Because recognition of the tk promoter by cellular transcription factors is well understood, its trans induction by viral regulatory proteins may serve as a useful model for the regulation of eukaryotic gene expression. A comprehensive set of mutations was therefore introduced into the chromosome of herpes simplex virus at the tk promoter to directly analyze the effects of promoter mutations on tk transcription. The promoter domains required for efficient tk expression under conditions of trans induction corresponded to those important for recognition by cellular transcription factors. Thus, trans induction of tk expression may be catalyzed initially by the interaction of viral regulatory proteins with cellular transcription factors.
Development of Emotional Facial Recognition in Late Childhood and Adolescence
ERIC Educational Resources Information Center
Thomas, Laura A.; De Bellis, Michael D.; Graham, Reiko; Labar, Kevin S.
2007-01-01
The ability to interpret emotions in facial expressions is crucial for social functioning across the lifespan. Facial expression recognition develops rapidly during infancy and improves with age during the preschool years. However, the developmental trajectory from late childhood to adulthood is less clear. We tested older children, adolescents…
Teaching Emotion Recognition Skills to Children with Autism
ERIC Educational Resources Information Center
Ryan, Christian; Charragain, Caitriona Ni
2010-01-01
Autism is associated with difficulty interacting with others and an impaired ability to recognize facial expressions of emotion. Previous teaching programmes have not addressed weak central coherence. Emotion recognition training focused on components of facial expressions. The training was administered in small groups ranging from 4 to 7…
Recognition of emotion with temporal lobe epilepsy and asymmetrical amygdala damage.
Fowler, Helen L; Baker, Gus A; Tipples, Jason; Hare, Dougal J; Keller, Simon; Chadwick, David W; Young, Andrew W
2006-08-01
Impairments in emotion recognition occur when there is bilateral damage to the amygdala. In this study, ability to recognize auditory and visual expressions of emotion was investigated in people with asymmetrical amygdala damage (AAD) and temporal lobe epilepsy (TLE). Recognition of five emotions was tested across three participant groups: those with right AAD and TLE, those with left AAD and TLE, and a comparison group. Four tasks were administered: recognition of emotion from facial expressions, sentences describing emotion-laden situations, nonverbal sounds, and prosody. Accuracy scores for each task and emotion were analysed, and no consistent overall effect of AAD on emotion recognition was found. However, some individual participants with AAD were significantly impaired at recognizing emotions, in both auditory and visual domains. The findings indicate that a minority of individuals with AAD have impairments in emotion recognition, but no evidence of specific impairments (e.g., visual or auditory) was found.
2015-05-28
recognition is simpler and requires less computational resources compared to other inputs such as facial expressions . The Berlin database of Emotional ...Processing Magazine, IEEE, vol. 18, no. 1, pp. 32– 80, 2001. [15] K. R. Scherer, T. Johnstone, and G. Klasmeyer, “Vocal expression of emotion ...Network for Real-Time Speech- Emotion Recognition 5a. CONTRACT NUMBER IN-HOUSE 5b. GRANT NUMBER 5c. PROGRAM ELEMENT NUMBER 62788F 6. AUTHOR(S) Q
ɣδ T cell receptor ligands and modes of antigen recognition
Champagne, Eric
2011-01-01
T lymphocytes expressing the γδ-type of T cell receptors for antigens contribute to all aspects of immune responses, including defenses against viruses, bacteria, parasites and tumors, allergy and autoimmunity. Multiple subsets have been individualized in humans as well as in mice and they appear to recognize in a TCR-dependent manner antigens as diverse as small non-peptidic molecules, soluble or membrane-anchored polypeptides and molecules related to MHC antigens on cell surfaces, implying diverse modes of antigen recognition. We review here the γδ TCR ligands which have been identified along the years and their characteristics, with emphasis on a few systems which have been extensively studied such as human γδ T cells responding to phosphoantigens or murine γδ T cells activated by allogeneic MHC antigens. We discuss a speculative model of antigen recognition involving simultaneous TCR recognition of MHC-like and non-MHC ligands which could fit with most available data and shares many similarities with the classical model of MHC-restricted antigen recognition for peptides or lipids by T cells subsets with αβ-type TCRs. PMID:21298486
γδ T cell receptor ligands and modes of antigen recognition.
Champagne, Eric
2011-04-01
T lymphocytes expressing the γδ-type of T cell receptors (TCRs) for antigens contribute to all aspects of immune responses, including defenses against viruses, bacteria, parasites and tumors, allergy and autoimmunity. Multiple subsets have been individualized in humans as well as in mice and they appear to recognize in a TCR-dependent manner antigens as diverse as small non-peptidic molecules, soluble or membrane-anchored polypeptides and molecules related to MHC antigens on cell surfaces, implying diverse modes of antigen recognition. We review here the γδ TCR ligands which have been identified along the years and their characteristics, with emphasis on a few systems which have been extensively studied such as human γδ T cells responding to phosphoantigens or murine γδ T cells activated by allogeneic MHC antigens. We discuss a speculative model of antigen recognition involving simultaneous TCR recognition of MHC-like and non-MHC ligands which could fit with most available data and shares many similarities with the classical model of MHC-restricted antigen recognition for peptides or lipids by T cells subsets with αβ-type TCRs.
Hills, Peter J; Hill, Dominic M
2017-07-12
Sad individuals perform more accurately at face identity recognition (Hills, Werno, & Lewis, 2011), possibly because they scan more of the face during encoding. During expression identification tasks, sad individuals do not fixate on the eyes as much as happier individuals (Wu, Pu, Allen, & Pauli, 2012). Fixating on features other than the eyes leads to a reduced own-ethnicity bias (Hills & Lewis, 2006). This background indicates that sad individuals would not view the eyes as much as happy individuals and this would result in improved expression recognition and a reduced own-ethnicity bias. This prediction was tested using an expression identification task, with eye tracking. We demonstrate that sad-induced participants show enhanced expression recognition and a reduced own-ethnicity bias than happy-induced participants due to scanning more facial features. We conclude that mood affects eye movements and face encoding by causing a wider sampling strategy and deeper encoding of facial features diagnostic for expression identification.
A multiplexable TALE-based binary expression system for in vivo cellular interaction studies.
Toegel, Markus; Azzam, Ghows; Lee, Eunice Y; Knapp, David J H F; Tan, Ying; Fa, Ming; Fulga, Tudor A
2017-11-21
Binary expression systems have revolutionised genetic research by enabling delivery of loss-of-function and gain-of-function transgenes with precise spatial-temporal resolution in vivo. However, at present, each existing platform relies on a defined exogenous transcription activator capable of binding a unique recognition sequence. Consequently, none of these technologies alone can be used to simultaneously target different tissues or cell types in the same organism. Here, we report a modular system based on programmable transcription activator-like effector (TALE) proteins, which enables parallel expression of multiple transgenes in spatially distinct tissues in vivo. Using endogenous enhancers coupled to TALE drivers, we demonstrate multiplexed orthogonal activation of several transgenes carrying cognate variable activating sequences (VAS) in distinct neighbouring cell types of the Drosophila central nervous system. Since the number of combinatorial TALE-VAS pairs is virtually unlimited, this platform provides an experimental framework for highly complex genetic manipulation studies in vivo.
Infrared and visible fusion face recognition based on NSCT domain
NASA Astrophysics Data System (ADS)
Xie, Zhihua; Zhang, Shuai; Liu, Guodong; Xiong, Jinquan
2018-01-01
Visible face recognition systems, being vulnerable to illumination, expression, and pose, can not achieve robust performance in unconstrained situations. Meanwhile, near infrared face images, being light- independent, can avoid or limit the drawbacks of face recognition in visible light, but its main challenges are low resolution and signal noise ratio (SNR). Therefore, near infrared and visible fusion face recognition has become an important direction in the field of unconstrained face recognition research. In this paper, a novel fusion algorithm in non-subsampled contourlet transform (NSCT) domain is proposed for Infrared and visible face fusion recognition. Firstly, NSCT is used respectively to process the infrared and visible face images, which exploits the image information at multiple scales, orientations, and frequency bands. Then, to exploit the effective discriminant feature and balance the power of high-low frequency band of NSCT coefficients, the local Gabor binary pattern (LGBP) and Local Binary Pattern (LBP) are applied respectively in different frequency parts to obtain the robust representation of infrared and visible face images. Finally, the score-level fusion is used to fuse the all the features for final classification. The visible and near infrared face recognition is tested on HITSZ Lab2 visible and near infrared face database. Experiments results show that the proposed method extracts the complementary features of near-infrared and visible-light images and improves the robustness of unconstrained face recognition.
Candra, Henry; Yuwono, Mitchell; Rifai Chai; Nguyen, Hung T; Su, Steven
2016-08-01
Psychotherapy requires appropriate recognition of patient's facial-emotion expression to provide proper treatment in psychotherapy session. To address the needs this paper proposed a facial emotion recognition system using Combination of Viola-Jones detector together with a feature descriptor we term Edge-Histogram of Oriented Gradients (E-HOG). The performance of the proposed method is compared with various feature sources including the face, the eyes, the mouth, as well as both the eyes and the mouth. Seven classes of basic emotions have been successfully identified with 96.4% accuracy using Multi-class Support Vector Machine (SVM). The proposed descriptor E-HOG is much leaner to compute compared to traditional HOG as shown by a significant improvement in processing time as high as 1833.33% (p-value = 2.43E-17) with a slight reduction in accuracy of only 1.17% (p-value = 0.0016).
Convolutional neural networks with balanced batches for facial expressions recognition
NASA Astrophysics Data System (ADS)
Battini Sönmez, Elena; Cangelosi, Angelo
2017-03-01
This paper considers the issue of fully automatic emotion classification on 2D faces. In spite of the great effort done in recent years, traditional machine learning approaches based on hand-crafted feature extraction followed by the classification stage failed to develop a real-time automatic facial expression recognition system. The proposed architecture uses Convolutional Neural Networks (CNN), which are built as a collection of interconnected processing elements to simulate the brain of human beings. The basic idea of CNNs is to learn a hierarchical representation of the input data, which results in a better classification performance. In this work we present a block-based CNN algorithm, which uses noise, as data augmentation technique, and builds batches with a balanced number of samples per class. The proposed architecture is a very simple yet powerful CNN, which can yield state-of-the-art accuracy on the very competitive benchmark algorithm of the Extended Cohn Kanade database.
Pizzolla, Angela; Smith, Jeffery M; Brooks, Andrew G; Reading, Patrick C
2017-04-01
Influenza remains a major global health issue and the effectiveness of current vaccines and antiviral drugs is limited by the continual evolution of influenza viruses. Therefore, identifying novel prophylactic or therapeutic treatments that induce appropriate innate immune responses to protect against influenza infection would represent an important advance in efforts to limit the impact of influenza. Cellular pattern recognition receptors (PRRs) recognize conserved structures expressed by pathogens to trigger intracellular signaling cascades, promoting expression of proinflammatory molecules and innate immunity. Therefore, a number of approaches have been developed to target specific PRRs in an effort to stimulate innate immunity and reduce disease in a variety of settings, including during influenza infections. Herein, we discuss progress in immunomodulation strategies designed to target cell-associated PRRs of the innate immune system, thereby, modifying innate responses to IAV infection and/or augmenting immune responses to influenza vaccines. © Society for Leukocyte Biology.
ERIC Educational Resources Information Center
Herba, Catherine; Phillips, Mary
2004-01-01
Background: Intact emotion processing is critical for normal emotional development. Recent advances in neuroimaging have facilitated the examination of brain development, and have allowed for the exploration of the relationships between the development of emotion processing abilities, and that of associated neural systems. Methods: A literature…
USDA-ARS?s Scientific Manuscript database
The antigen recognition by the host immune system is a complex biochemical process that requires a battery of enzymes. Cathepsins are one of the enzyme superfamilies involving in antigenic degradations. We observed the up-regulation of cathepsin H and L transcripts during the early stage of Edward...
The within-subjects design in the study of facial expressions.
Yik, Michelle; Widen, Sherri C; Russell, James A
2013-01-01
The common within-subjects design of studies on the recognition of emotion from facial expressions allows the judgement of one face to be influenced by previous faces, thus introducing the potential for artefacts. The present study (N=344) showed that the canonical "disgust face" was judged as disgusted, provided that the preceding set of faces included "anger expressions", but was judged as angry when the preceding set of faces excluded anger but instead included persons who looked sad or about to be sick. Chinese observers showed lower recognition of the "disgust face" than did American observers. Chinese observers also showed lower recognition of the "fear face" when responding in Chinese than in English.
Impaired recognition of facial emotions from low-spatial frequencies in Asperger syndrome.
Kätsyri, Jari; Saalasti, Satu; Tiippana, Kaisa; von Wendt, Lennart; Sams, Mikko
2008-01-01
The theory of 'weak central coherence' [Happe, F., & Frith, U. (2006). The weak coherence account: Detail-focused cognitive style in autism spectrum disorders. Journal of Autism and Developmental Disorders, 36(1), 5-25] implies that persons with autism spectrum disorders (ASDs) have a perceptual bias for local but not for global stimulus features. The recognition of emotional facial expressions representing various different levels of detail has not been studied previously in ASDs. We analyzed the recognition of four basic emotional facial expressions (anger, disgust, fear and happiness) from low-spatial frequencies (overall global shapes without local features) in adults with an ASD. A group of 20 participants with Asperger syndrome (AS) was compared to a group of non-autistic age- and sex-matched controls. Emotion recognition was tested from static and dynamic facial expressions whose spatial frequency contents had been manipulated by low-pass filtering at two levels. The two groups recognized emotions similarly from non-filtered faces and from dynamic vs. static facial expressions. In contrast, the participants with AS were less accurate than controls in recognizing facial emotions from very low-spatial frequencies. The results suggest intact recognition of basic facial emotions and dynamic facial information, but impaired visual processing of global features in ASDs.
Sleep deprivation impairs the accurate recognition of human emotions.
van der Helm, Els; Gujar, Ninad; Walker, Matthew P
2010-03-01
Investigate the impact of sleep deprivation on the ability to recognize the intensity of human facial emotions. Randomized total sleep-deprivation or sleep-rested conditions, involving between-group and within-group repeated measures analysis. Experimental laboratory study. Thirty-seven healthy participants, (21 females) aged 18-25 y, were randomly assigned to the sleep control (SC: n = 17) or total sleep deprivation group (TSD: n = 20). Participants performed an emotional face recognition task, in which they evaluated 3 different affective face categories: Sad, Happy, and Angry, each ranging in a gradient from neutral to increasingly emotional. In the TSD group, the task was performed once under conditions of sleep deprivation, and twice under sleep-rested conditions following different durations of sleep recovery. In the SC group, the task was performed twice under sleep-rested conditions, controlling for repeatability. In the TSD group, when sleep-deprived, there was a marked and significant blunting in the recognition of Angry and Happy affective expressions in the moderate (but not extreme) emotional intensity range; differences that were most reliable and significant in female participants. No change in the recognition of Sad expressions was observed. These recognition deficits were, however, ameliorated following one night of recovery sleep. No changes in task performance were observed in the SC group. Sleep deprivation selectively impairs the accurate judgment of human facial emotions, especially threat relevant (Anger) and reward relevant (Happy) categories, an effect observed most significantly in females. Such findings suggest that sleep loss impairs discrete affective neural systems, disrupting the identification of salient affective social cues.
RNA recognition by human TLR8 can lead to autoimmune inflammation
Gong, Mei; Cepika, Alma-Martina; Xu, Zhaohui; Tripodo, Claudio; Bennett, Lynda; Crain, Chad; Quartier, Pierre; Cush, John J.; Pascual, Virginia; Coffman, Robert L.; Barrat, Franck J.
2013-01-01
Studies on the role of the RNA receptor TLR8 in inflammation have been limited by its different function in human versus rodents. We have generated multiple lines of transgenic mice expressing different levels of human TLR8. The high copy number chimeras were unable to pass germline; developed severe inflammation targeting the pancreas, salivary glands, and joints; and the severity of the specific phenotypes closely correlated with the huTLR8 expression levels. Mice with relatively low expression levels survived and bred successfully but had increased susceptibility to collagen-induced arthritis, and the levels of huTLR8 correlated with proinflammatory cytokines in the joints of the animals. At the cellular level, huTLR8 signaling exerted a DC-intrinsic effect leading to up-regulation of co-stimulatory molecules and subsequent T cell activation. A pathogenic role for TLR8 in human diseases was suggested by its increased expression in patients with systemic arthritis and the correlation of TLR8 expression with the elevation of IL-1β levels and disease status. We found that the consequence of self-recognition via TLR8 results in a constellation of diseases, strikingly distinct from those related to TLR7 signaling, and points to specific inflammatory diseases that may benefit from inhibition of TLR8 in humans. PMID:24277153
Early effects of duloxetine on emotion recognition in healthy volunteers
Bamford, Susan; Penton-Voak, Ian; Pinkney, Verity; Baldwin, David S; Munafò, Marcus R; Garner, Matthew
2015-01-01
The serotonin-noradrenaline reuptake inhibitor (SNRI) duloxetine is an effective treatment for major depression and generalised anxiety disorder. Neuropsychological models of antidepressant drug action suggest therapeutic effects might be mediated by the early correction of maladaptive biases in emotion processing, including the recognition of emotional expressions. Sub-chronic administration of duloxetine (for two weeks) produces adaptive changes in neural circuitry implicated in emotion processing; however, its effects on emotional expression recognition are unknown. Forty healthy participants were randomised to receive either 14 days of duloxetine (60 mg/day, titrated from 30 mg after three days) or matched placebo (with sham titration) in a double-blind, between-groups, repeated-measures design. On day 0 and day 14 participants completed a computerised emotional expression recognition task that measured sensitivity to the six primary emotions. Thirty-eight participants (19 per group) completed their course of tablets and were included in the analysis. Results provide evidence that duloxetine, compared to placebo, may reduce the accurate recognition of sadness. Drug effects were driven by changes in participants’ ability to correctly detect subtle expressions of sadness, with greater change observed in the placebo relative to the duloxetine group. These effects occurred in the absence of changes in mood. Our preliminary findings require replication, but complement recent evidence that sadness recognition is a therapeutic target in major depression, and a mechanism through which SNRIs could resolve negative biases in emotion processing to achieve therapeutic effects. PMID:25759400
Emotion Recognition in Face and Body Motion in Bulimia Nervosa.
Dapelo, Marcela Marin; Surguladze, Simon; Morris, Robin; Tchanturia, Kate
2017-11-01
Social cognition has been studied extensively in anorexia nervosa (AN), but there are few studies in bulimia nervosa (BN). This study investigated the ability of people with BN to recognise emotions in ambiguous facial expressions and in body movement. Participants were 26 women with BN, who were compared with 35 with AN, and 42 healthy controls. Participants completed an emotion recognition task by using faces portraying blended emotions, along with a body emotion recognition task by using videos of point-light walkers. The results indicated that BN participants exhibited difficulties recognising disgust in less-ambiguous facial expressions, and a tendency to interpret non-angry faces as anger, compared with healthy controls. These difficulties were similar to those found in AN. There were no significant differences amongst the groups in body motion emotion recognition. The findings suggest that difficulties with disgust and anger recognition in facial expressions may be shared transdiagnostically in people with eating disorders. Copyright © 2017 John Wiley & Sons, Ltd and Eating Disorders Association. Copyright © 2017 John Wiley & Sons, Ltd and Eating Disorders Association.
Encoding conditions affect recognition of vocally expressed emotions across cultures.
Jürgens, Rebecca; Drolet, Matthis; Pirow, Ralph; Scheiner, Elisabeth; Fischer, Julia
2013-01-01
Although the expression of emotions in humans is considered to be largely universal, cultural effects contribute to both emotion expression and recognition. To disentangle the interplay between these factors, play-acted and authentic (non-instructed) vocal expressions of emotions were used, on the assumption that cultural effects may contribute differentially to the recognition of staged and spontaneous emotions. Speech tokens depicting four emotions (anger, sadness, joy, fear) were obtained from German radio archives and re-enacted by professional actors, and presented to 120 participants from Germany, Romania, and Indonesia. Participants in all three countries were poor at distinguishing between play-acted and spontaneous emotional utterances (58.73% correct on average with only marginal cultural differences). Nevertheless, authenticity influenced emotion recognition: across cultures, anger was recognized more accurately when play-acted (z = 15.06, p < 0.001) and sadness when authentic (z = 6.63, p < 0.001), replicating previous findings from German populations. German subjects revealed a slight advantage in recognizing emotions, indicating a moderate in-group advantage. There was no difference between Romanian and Indonesian subjects in the overall emotion recognition. Differential cultural effects became particularly apparent in terms of differential biases in emotion attribution. While all participants labeled play-acted expressions as anger more frequently than expected, German participants exhibited a further bias toward choosing anger for spontaneous stimuli. In contrast to the German sample, Romanian and Indonesian participants were biased toward choosing sadness. These results support the view that emotion recognition rests on a complex interaction of human universals and cultural specificities. Whether and in which way the observed biases are linked to cultural differences in self-construal remains an issue for further investigation.
Yalcin-Siedentopf, Nursen; Hoertnagl, Christine M; Biedermann, Falko; Baumgartner, Susanne; Deisenhammer, Eberhard A; Hausmann, Armand; Kaufmann, Alexandra; Kemmler, Georg; Mühlbacher, Moritz; Rauch, Anna-Sophia; Fleischhacker, W Wolfgang; Hofer, Alex
2014-02-01
Both schizophrenia and bipolar disorder (BD) have consistently been associated with deficits in facial affect recognition (FAR). These impairments have been related to various aspects of social competence and functioning and are relatively stable over time. However, individuals in remission may outperform patients experiencing an acute phase of the disorders. The present study directly contrasted FAR in symptomatically remitted patients with schizophrenia or BD and healthy volunteers and investigated its relationship with patients' outcomes. Compared to healthy control subjects, schizophrenia patients were impaired in the recognition of angry, disgusted, sad and happy facial expressions, while BD patients showed deficits only in the recognition of disgusted and happy facial expressions. When directly comparing the two patient groups individuals suffering from BD outperformed those with schizophrenia in the recognition of expressions depicting anger. There was no significant association between affect recognition abilities and symptomatic or psychosocial outcomes in schizophrenia patients. Among BD patients, relatively higher depression scores were associated with impairments in both the identification of happy faces and psychosocial functioning. Overall, our findings indicate that during periods of symptomatic remission the recognition of facial affect may be less impaired in patients with BD than in those suffering from schizophrenia. However, in the psychosocial context BD patients seem to be more sensitive to residual symptomatology. Copyright © 2013 Elsevier B.V. All rights reserved.
Does Facial Expression Recognition Provide a Toehold for the Development of Emotion Understanding?
ERIC Educational Resources Information Center
Strand, Paul S.; Downs, Andrew; Barbosa-Leiker, Celestina
2016-01-01
The authors explored predictions from basic emotion theory (BET) that facial emotion expression recognition skills are insular with respect to their own development, and yet foundational to the development of emotional perspective-taking skills. Participants included 417 preschool children for whom estimates of these 2 emotion understanding…
Alvarez-Vallina, L; Yañez, R; Blanco, B; Gil, M; Russell, S J
2000-04-01
Adoptive therapy with autologous T cells expressing chimeric T-cell receptors (chTCRs) is of potential interest for the treatment of malignancy. To limit possible T-cell-mediated damage to normal tissues that weakly express the targeted tumor antigen (Ag), we have tested a strategy for the suppression of target cell recognition by engineered T cells. Jurkat T cells were transduced with an anti-hapten chTCR tinder the control of a tetracycline-suppressible promoter and were shown to respond to Ag-positive (hapten-coated) but not to Ag-negative target cells. The engineered T cells were then reacted with hapten-coated target cells at different effector to target cell ratios before and after exposure to tetracycline. When the engineered T cells were treated with tetracycline, expression of the chTCR was greatly decreased and recognition of the hapten-coated target cells was completely suppressed. Tetracycline-mediated suppression of target cell recognition by engineered T cells may be a useful strategy to limit the toxicity of the approach to cancer gene therapy.
Quantifying facial expression recognition across viewing conditions.
Goren, Deborah; Wilson, Hugh R
2006-04-01
Facial expressions are key to social interactions and to assessment of potential danger in various situations. Therefore, our brains must be able to recognize facial expressions when they are transformed in biologically plausible ways. We used synthetic happy, sad, angry and fearful faces to determine the amount of geometric change required to recognize these emotions during brief presentations. Five-alternative forced choice conditions involving central viewing, peripheral viewing and inversion were used to study recognition among the four emotions. Two-alternative forced choice was used to study affect discrimination when spatial frequency information in the stimulus was modified. The results show an emotion and task-dependent pattern of detection. Facial expressions presented with low peak frequencies are much harder to discriminate from neutral than faces defined by either mid or high peak frequencies. Peripheral presentation of faces also makes recognition much more difficult, except for happy faces. Differences between fearful detection and recognition tasks are probably due to common confusions with sadness when recognizing fear from among other emotions. These findings further support the idea that these emotions are processed separately from each other.
Quest Hierarchy for Hyperspectral Face Recognition
2011-03-01
numerous face recognition algorithms available, several very good literature surveys are available that include Abate [29], Samal [110], Kong [18], Zou...Perception, Japan (January 1994). [110] Samal , Ashok and P. Iyengar, Automatic Recognition and Analysis of Human Faces and Facial Expressions: A Survey
Recognizing Action Units for Facial Expression Analysis
Tian, Ying-li; Kanade, Takeo; Cohn, Jeffrey F.
2010-01-01
Most automatic expression analysis systems attempt to recognize a small set of prototypic expressions, such as happiness, anger, surprise, and fear. Such prototypic expressions, however, occur rather infrequently. Human emotions and intentions are more often communicated by changes in one or a few discrete facial features. In this paper, we develop an Automatic Face Analysis (AFA) system to analyze facial expressions based on both permanent facial features (brows, eyes, mouth) and transient facial features (deepening of facial furrows) in a nearly frontal-view face image sequence. The AFA system recognizes fine-grained changes in facial expression into action units (AUs) of the Facial Action Coding System (FACS), instead of a few prototypic expressions. Multistate face and facial component models are proposed for tracking and modeling the various facial features, including lips, eyes, brows, cheeks, and furrows. During tracking, detailed parametric descriptions of the facial features are extracted. With these parameters as the inputs, a group of action units (neutral expression, six upper face AUs and 10 lower face AUs) are recognized whether they occur alone or in combinations. The system has achieved average recognition rates of 96.4 percent (95.4 percent if neutral expressions are excluded) for upper face AUs and 96.7 percent (95.6 percent with neutral expressions excluded) for lower face AUs. The generalizability of the system has been tested by using independent image databases collected and FACS-coded for ground-truth by different research teams. PMID:25210210
Kaminitz, Ayelet; Barzilay, Ran; Segal, Hadar; Taler, Michal; Offen, Daniel; Gil-Ad, Irit; Mechoulam, Raphael; Weizman, Abraham
2014-01-01
OBJECTIVES. Disrupted in schizophrenia 1 (DISC1) is considered the most prominent candidate gene for schizophrenia. In this study, we aimed to characterize behavioural and brain biochemical traits in a mouse expressing a dominant negative DISC1mutant (DN-DISC1). DN-DISC1 mice underwent behavioural tests to evaluate object recognition, social preference and social novelty seeking. ELISA was conducted on brain tissue to evaluate BDNF levels. Western blot was employed to measure BDNF receptor (TrkB) and cannabinoid receptor CB1. The mutant DISC1 mice displayed deficits in preference to social novelty while both social preference and object recognition were intact. Biochemical analysis of prefrontal cortex and hippocampus revealed a modest reduction in cortical TrkB protein levels of male mice while no differences in BDNF levels were observed. We found sex dependent differences in the expression of cannabinoid-1 receptors. We describe novel behavioural and biochemical abnormalities in the DN-DISC1 mouse model of schizophrenia. The data shows for the first time a possible link between DISC1 mutation and the cannabinoid system.
Facial emotion recognition and borderline personality pathology.
Meehan, Kevin B; De Panfilis, Chiara; Cain, Nicole M; Antonucci, Camilla; Soliani, Antonio; Clarkin, John F; Sambataro, Fabio
2017-09-01
The impact of borderline personality pathology on facial emotion recognition has been in dispute; with impaired, comparable, and enhanced accuracy found in high borderline personality groups. Discrepancies are likely driven by variations in facial emotion recognition tasks across studies (stimuli type/intensity) and heterogeneity in borderline personality pathology. This study evaluates facial emotion recognition for neutral and negative emotions (fear/sadness/disgust/anger) presented at varying intensities. Effortful control was evaluated as a moderator of facial emotion recognition in borderline personality. Non-clinical multicultural undergraduates (n = 132) completed a morphed facial emotion recognition task of neutral and negative emotional expressions across different intensities (100% Neutral; 25%/50%/75% Emotion) and self-reported borderline personality features and effortful control. Greater borderline personality features related to decreased accuracy in detecting neutral faces, but increased accuracy in detecting negative emotion faces, particularly at low-intensity thresholds. This pattern was moderated by effortful control; for individuals with low but not high effortful control, greater borderline personality features related to misattributions of emotion to neutral expressions, and enhanced detection of low-intensity emotional expressions. Individuals with high borderline personality features may therefore exhibit a bias toward detecting negative emotions that are not or barely present; however, good self-regulatory skills may protect against this potential social-cognitive vulnerability. Copyright © 2017 Elsevier Ireland Ltd. All rights reserved.
Neurotrophins play differential roles in short and long-term recognition memory.
Callaghan, Charlotte K; Kelly, Aine M
2013-09-01
The neurotrophin family of proteins are believed to mediate various forms of synaptic plasticity in the adult brain. Here we have assessed the roles of these proteins in object recognition memory in the rat, using icv infusions of function-blocking antibodies or the tyrosine kinase antagonist, tyrphostin AG879, to block Trk receptors. We report that tyrphostin AG879 impairs both short-term and long-term recognition memory, indicating a requirement for Trk receptor activation in both processes. The effect of inhibition of each of the neurotrophins with activity-blocking neutralising antibodies was also tested. Treatment with anti-BDNF, anti-NGF or anti-NT4 had no effect on short-term memory, but blocked long-term recognition memory. Treatment with anti-NT3 had no effect on either process. We also assessed changes in expression of neurotrophins and their respective receptors in the hippocampus, dentate gyrus and perirhinal cortex over a 24 h period following training in the object recognition task. We observed time-dependent changes in expression of the Trk receptors and their ligands in the dentate gyrus and perirhinal cortex. The data are consistent with a pivotal role for neurotrophic factors in the expression of recognition memory. Copyright © 2013 Elsevier Inc. All rights reserved.
Platt, Bradley; Kamboj, Sunjeev; Morgan, Celia J A; Curran, H Valerie
2010-11-01
While heavy cannabis-users seem to show various cognitive impairments, it remains unclear whether they also experience significant deficits in affective functioning. Evidence of such deficits may contribute to our understanding of the interpersonal difficulties in cannabis-users, and the link between cannabis-use and psychological disorders (Moore et al., 2007). Emotion recognition performance of heavy cannabis-users and non-using controls was compared. A measure of emotion recognition was used in which participants identified facial expressions as they changed from neutral (open-mouth) to gradually more intense expressions of sadness, neutral, anger or happiness (open or closed mouth). Reaction times and accuracy were recorded as the facial expressions changed. Participants also completed measures of 'theory of mind,' depression and impulsivity. Cannabis-users were significantly slower than controls at identifying all three emotional expressions. There was no difference between groups in identifying facial expressions changing from open-mouth neutral expressions to closed-mouth neutral expressions suggesting that differences in emotion recognition were not due to a general slowing of reaction times. Cannabis-users were also significantly more liberal in their response criterion for recognising sadness. Heavy cannabis-use may be associated with affect recognition deficits. In particular, a greater intensity of emotion expression was required before identification of positive and negative emotions. This was found using stimuli which simulated dynamic changes in emotion expression, and in turn, suggests that cannabis-users may experience generalised problems in decoding basic emotions during social interactions. The implications of these findings are discussed for vulnerability to psychological and interpersonal difficulties in cannabis-users. Copyright © 2010 Elsevier Ireland Ltd. All rights reserved.
Baran Tatar, Zeynep; Yargıç, İlhan; Oflaz, Serap; Büyükgök, Deniz
2015-01-01
Interpersonal relationship disorders in adults with Attention Deficit Hyperactivity Disorder (ADHD) can be associated with the impairment of non-verbal communication. The purpose of our study was to compare the emotion recognition, facial recognition and neuropsychological assessments of adult ADHD patients with those of healthy controls, and to thus determine the effect of neuropsychological data on the recognition of emotional expressions. This study, which was based on a case-control model, was conducted with patients diagnosed with ADHD according to the DSM-IV-TR, being followed and monitored at the adult ADHD clinic of the Psychiatry Department of the Istanbul University Istanbul Medical Faculty Hospital. The study group consisted of 40 adults (27.5% female) between the ages of 20-65 (mean age 25.96 ± 6.07; education level: 15.02±2.34 years) diagnosed with ADHD, and 40 controls who were matched/similar with the study group with respect to age, gender, and education level. In the ADHD group, 14 (35%) of the patients had concomitant diseases. Pictures of Facial Affect, the Benton Face Recognition Test, and the Continuous Performance Test were used to respectively evaluate emotion recognition, facial recognition, and attention deficit and impulsivity of the patients. It was determined that, in comparison to the control group, the ADHD group made more mistakes in recognizing all types of emotional expressions and neutral expressions. The ADHD group also demonstrated more cognitive mistakes. Facial recognition was similar in both groups. It was determined that impulsivity had a significant effect on facial recognition. The social relationship disorders observed in ADHD can be affected by emotion recognition processes. In future studies, it may be possible to investigate the effects that early psychopharmacological and psychotherapeutic interventions administered for the main symptoms of ADHD have on the impairment of emotion recognition.
Speaker emotion recognition: from classical classifiers to deep neural networks
NASA Astrophysics Data System (ADS)
Mezghani, Eya; Charfeddine, Maha; Nicolas, Henri; Ben Amar, Chokri
2018-04-01
Speaker emotion recognition is considered among the most challenging tasks in recent years. In fact, automatic systems for security, medicine or education can be improved when considering the speech affective state. In this paper, a twofold approach for speech emotion classification is proposed. At the first side, a relevant set of features is adopted, and then at the second one, numerous supervised training techniques, involving classic methods as well as deep learning, are experimented. Experimental results indicate that deep architecture can improve classification performance on two affective databases, the Berlin Dataset of Emotional Speech and the SAVEE Dataset Surrey Audio-Visual Expressed Emotion.
Riordan, Alexander J; Schaler, Ari W; Fried, Jenny; Paine, Tracie A; Thornton, Janice E
2018-05-01
The cognitive symptoms of schizophrenia are poorly understood and difficult to treat. Estrogens may mitigate these symptoms via unknown mechanisms. To examine these mechanisms, we tested whether increasing estradiol (E) or decreasing luteinizing hormone (LH) could mitigate short-term episodic memory loss in a phencyclidine (PCP) model of schizophrenia. We then assessed whether changes in cortical or hippocampal GABA may underlie these effects. Female rats were ovariectomized and injected subchronically with PCP. To modulate E and LH, animals received estradiol capsules or Antide injections. Short-term episodic memory was assessed using the novel object recognition task (NORT). Brain expression of GAD67 was analyzed via western blot, and parvalbumin-containing cells were counted using immunohistochemistry. Some rats received hippocampal infusions of a GABA A agonist, GABA A antagonist, or GAD inhibitor before behavioral testing. We found that PCP reduced hippocampal GAD67 and abolished recognition memory. Antide restored hippocampal GAD67 and rescued recognition memory in PCP-treated animals. Estradiol prevented PCP's amnesic effect in NORT but failed to restore hippocampal GAD67. PCP did not cause significant differences in number of parvalbumin-expressing cells or cortical expression of GAD67. Hippocampal infusions of a GABA A agonist restored recognition memory in PCP-treated rats. Blocking hippocampal GAD or GABA A receptors in ovx animals reproduced recognition memory loss similar to PCP and inhibited estradiol's protection of recognition memory in PCP-treated animals. In summary, decreasing LH or increasing E can lessen short-term episodic memory loss, as measured by novel object recognition, in a PCP model of schizophrenia. Alterations in hippocampal GABA may contribute to both PCP's effects on recognition memory and the hormones' ability to prevent or reverse them. Copyright © 2018 Elsevier Ltd. All rights reserved.
Facial Expression Recognition: Can Preschoolers with Cochlear Implants and Hearing Aids Catch It?
ERIC Educational Resources Information Center
Wang, Yifang; Su, Yanjie; Fang, Ping; Zhou, Qingxia
2011-01-01
Tager-Flusberg and Sullivan (2000) presented a cognitive model of theory of mind (ToM), in which they thought ToM included two components--a social-perceptual component and a social-cognitive component. Facial expression recognition (FER) is an ability tapping the social-perceptual component. Previous findings suggested that normal hearing…
ERIC Educational Resources Information Center
Fairchild, Graeme; Van Goozen, Stephanie H. M.; Calder, Andrew J.; Stollery, Sarah J.; Goodyer, Ian M.
2009-01-01
Background: We examined whether conduct disorder (CD) is associated with deficits in facial expression recognition and, if so, whether these deficits are specific to the early-onset form of CD, which emerges in childhood. The findings could potentially inform the developmental taxonomic theory of antisocial behaviour, which suggests that…
ERIC Educational Resources Information Center
Tell, Dina; Davidson, Denise
2015-01-01
In this research, the emotion recognition abilities of children with autism spectrum disorder and typically developing children were compared. When facial expressions and situational cues of emotion were congruent, accuracy in recognizing emotions was good for both children with autism spectrum disorder and typically developing children. When…
Orientation and Affective Expression Effects on Face Recognition in Williams Syndrome and Autism
ERIC Educational Resources Information Center
Rose, Fredric E.; Lincoln, Alan J.; Lai, Zona; Ene, Michaela; Searcy, Yvonne M.; Bellugi, Ursula
2007-01-01
We sought to clarify the nature of the face processing strength commonly observed in individuals with Williams syndrome (WS) by comparing the face recognition ability of persons with WS to that of persons with autism and to healthy controls under three conditions: Upright faces with neutral expressions, upright faces with varying affective…
Sawyer, Alyssa C P; Williamson, Paul; Young, Robyn L
2012-04-01
Research has shown that individuals with Autism Spectrum Disorders (ASD) have difficulties recognising emotions from facial expressions. Since eye contact is important for accurate emotion recognition, and individuals with ASD tend to avoid eye contact, this tendency for gaze aversion has been proposed as an explanation for the emotion recognition deficit. This explanation was investigated using a newly developed emotion and mental state recognition task. Individuals with Asperger's Syndrome were less accurate at recognising emotions and mental states, but did not show evidence of gaze avoidance compared to individuals without Asperger's Syndrome. This suggests that the way individuals with Asperger's Syndrome look at faces cannot account for the difficulty they have recognising expressions.
Alfimova, M V; Golimbet, V E; Korovaitseva, G I; Lezheiko, T V; Abramova, L I; Aksenova, E V; Bolgov, M I
2014-01-01
The 5-HTTLPR SLC6A4 and catechol-o-methyltransferase (COMT) Val158Met polymorphisms are reported to be associated with processing of facial expressions in general population. Impaired recognition of facial expressions that is characteristic of schizophrenia negatively impacts on the social adaptation of the patients. To search for molecular mechanisms of this deficit, we studied main and epistatic effects of 5-HTTLPR and Val158Met polymorphisms on the facial emotion recognition in patients with schizophrenia (n=299) and healthy controls (n=232). The 5-HTTLPR polymorphism was associated with the emotion recognition in patients. The ll-homozygotes recognized facial emotions significantly better compared to those with an s-allele (F=8.00; p=0.005). Although the recognition of facial emotions was correlated with negative symptoms, verbal learning and trait anxiety, these variables did not significantly modified the association. In both groups, no effect of the COMT on the recognition of facial emotions was found.
Automatic recognition of emotions from facial expressions
NASA Astrophysics Data System (ADS)
Xue, Henry; Gertner, Izidor
2014-06-01
In the human-computer interaction (HCI) process it is desirable to have an artificial intelligent (AI) system that can identify and categorize human emotions from facial expressions. Such systems can be used in security, in entertainment industries, and also to study visual perception, social interactions and disorders (e.g. schizophrenia and autism). In this work we survey and compare the performance of different feature extraction algorithms and classification schemes. We introduce a faster feature extraction method that resizes and applies a set of filters to the data images without sacrificing the accuracy. In addition, we have enhanced SVM to multiple dimensions while retaining the high accuracy rate of SVM. The algorithms were tested using the Japanese Female Facial Expression (JAFFE) Database and the Database of Faces (AT&T Faces).
Mapping the impairment in decoding static facial expressions of emotion in prosopagnosia.
Fiset, Daniel; Blais, Caroline; Royer, Jessica; Richoz, Anne-Raphaëlle; Dugas, Gabrielle; Caldara, Roberto
2017-08-01
Acquired prosopagnosia is characterized by a deficit in face recognition due to diverse brain lesions, but interestingly most prosopagnosic patients suffering from posterior lesions use the mouth instead of the eyes for face identification. Whether this bias is present for the recognition of facial expressions of emotion has not yet been addressed. We tested PS, a pure case of acquired prosopagnosia with bilateral occipitotemporal lesions anatomically sparing the regions dedicated for facial expression recognition. PS used mostly the mouth to recognize facial expressions even when the eye area was the most diagnostic. Moreover, PS directed most of her fixations towards the mouth. Her impairment was still largely present when she was instructed to look at the eyes, or when she was forced to look at them. Control participants showed a performance comparable to PS when only the lower part of the face was available. These observations suggest that the deficits observed in PS with static images are not solely attentional, but are rooted at the level of facial information use. This study corroborates neuroimaging findings suggesting that the Occipital Face Area might play a critical role in extracting facial features that are integrated for both face identification and facial expression recognition in static images. © The Author (2017). Published by Oxford University Press.
Niedtfeld, Inga; Defiebre, Nadine; Regenbogen, Christina; Mier, Daniela; Fenske, Sabrina; Kirsch, Peter; Lis, Stefanie; Schmahl, Christian
2017-04-01
Previous research has revealed alterations and deficits in facial emotion recognition in patients with borderline personality disorder (BPD). During interpersonal communication in daily life, social signals such as speech content, variation in prosody, and facial expression need to be considered simultaneously. We hypothesized that deficits in higher level integration of social stimuli contribute to difficulties in emotion recognition in BPD, and heightened arousal might explain this effect. Thirty-one patients with BPD and thirty-one healthy controls were asked to identify emotions in short video clips, which were designed to represent different combinations of the three communication channels: facial expression, speech content, and prosody. Skin conductance was recorded as a measure of sympathetic arousal, while controlling for state dissociation. Patients with BPD showed lower mean accuracy scores than healthy control subjects in all conditions comprising emotional facial expressions. This was true for the condition with facial expression only, and for the combination of all three communication channels. Electrodermal responses were enhanced in BPD only in response to auditory stimuli. In line with the major body of facial emotion recognition studies, we conclude that deficits in the interpretation of facial expressions lead to the difficulties observed in multimodal emotion processing in BPD.
Seeing Life through Positive-Tinted Glasses: Color–Meaning Associations
Gil, Sandrine; Le Bigot, Ludovic
2014-01-01
There is a growing body of literature to show that color can convey information, owing to its emotionally meaningful associations. Most research so far has focused on negative hue–meaning associations (e.g., red) with the exception of the positive aspects associated with green. We therefore set out to investigate the positive associations of two colors (i.e., green and pink), using an emotional facial expression recognition task in which colors provided the emotional contextual information for the face processing. In two experiments, green and pink backgrounds enhanced happy face recognition and impaired sad face recognition, compared with a control color (gray). Our findings therefore suggest that because green and pink both convey positive information, they facilitate the processing of emotionally congruent facial expressions (i.e., faces expressing happiness) and interfere with that of incongruent facial expressions (i.e., faces expressing sadness). Data also revealed a positive association for white. Results are discussed within the theoretical framework of emotional cue processing and color meaning. PMID:25098167
Seeing life through positive-tinted glasses: color-meaning associations.
Gil, Sandrine; Le Bigot, Ludovic
2014-01-01
There is a growing body of literature to show that color can convey information, owing to its emotionally meaningful associations. Most research so far has focused on negative hue-meaning associations (e.g., red) with the exception of the positive aspects associated with green. We therefore set out to investigate the positive associations of two colors (i.e., green and pink), using an emotional facial expression recognition task in which colors provided the emotional contextual information for the face processing. In two experiments, green and pink backgrounds enhanced happy face recognition and impaired sad face recognition, compared with a control color (gray). Our findings therefore suggest that because green and pink both convey positive information, they facilitate the processing of emotionally congruent facial expressions (i.e., faces expressing happiness) and interfere with that of incongruent facial expressions (i.e., faces expressing sadness). Data also revealed a positive association for white. Results are discussed within the theoretical framework of emotional cue processing and color meaning.
Facial expression recognition based on weber local descriptor and sparse representation
NASA Astrophysics Data System (ADS)
Ouyang, Yan
2018-03-01
Automatic facial expression recognition has been one of the research hotspots in the area of computer vision for nearly ten years. During the decade, many state-of-the-art methods have been proposed which perform very high accurate rate based on the face images without any interference. Nowadays, many researchers begin to challenge the task of classifying the facial expression images with corruptions and occlusions and the Sparse Representation based Classification framework has been wildly used because it can robust to the corruptions and occlusions. Therefore, this paper proposed a novel facial expression recognition method based on Weber local descriptor (WLD) and Sparse representation. The method includes three parts: firstly the face images are divided into many local patches, and then the WLD histograms of each patch are extracted, finally all the WLD histograms features are composed into a vector and combined with SRC to classify the facial expressions. The experiment results on the Cohn-Kanade database show that the proposed method is robust to occlusions and corruptions.
ERIC Educational Resources Information Center
Osguthorpe, Russell T.; Li Chang, Linda
1988-01-01
A computerized symbol processor system using an Apple IIe computer and a Power Pad graphics tablet was tested with 22 nonspeaking, multiply disabled students. The students were taught to express themselves independently in writing, and they did significantly better than control students on measures of language comprehension and symbol recognition.…
ERIC Educational Resources Information Center
Chalmers, Denise
2011-01-01
For more than 20 years there have been growing and widely expressed concerns that teaching is not sufficiently rewarded and recognized in universities, particularly in comparison to research. Individuals, institutions and governments have each responded in different ways to promote changes in institutional systems and practices. Two of the major…
USDA-ARS?s Scientific Manuscript database
Induction of innate immune pathways is critical for early host defense but there is limited understanding of how teleost fish recognize pathogen molecules and activate these pathways. In mammals, cells of the innate immune system detect pathogenic molecular structures using pattern recognition rece...
Traffic sign recognition based on deep convolutional neural network
NASA Astrophysics Data System (ADS)
Yin, Shi-hao; Deng, Ji-cai; Zhang, Da-wei; Du, Jing-yuan
2017-11-01
Traffic sign recognition (TSR) is an important component of automated driving systems. It is a rather challenging task to design a high-performance classifier for the TSR system. In this paper, we propose a new method for TSR system based on deep convolutional neural network. In order to enhance the expression of the network, a novel structure (dubbed block-layer below) which combines network-in-network and residual connection is designed. Our network has 10 layers with parameters (block-layer seen as a single layer): the first seven are alternate convolutional layers and block-layers, and the remaining three are fully-connected layers. We train our TSR network on the German traffic sign recognition benchmark (GTSRB) dataset. To reduce overfitting, we perform data augmentation on the training images and employ a regularization method named "dropout". The activation function we employ in our network adopts scaled exponential linear units (SELUs), which can induce self-normalizing properties. To speed up the training, we use an efficient GPU to accelerate the convolutional operation. On the test dataset of GTSRB, we achieve the accuracy rate of 99.67%, exceeding the state-of-the-art results.
Wolf, Richard C; Pujara, Maia; Baskaya, Mustafa K; Koenigs, Michael
2016-09-01
Facial emotion recognition is a critical aspect of human communication. Since abnormalities in facial emotion recognition are associated with social and affective impairment in a variety of psychiatric and neurological conditions, identifying the neural substrates and psychological processes underlying facial emotion recognition will help advance basic and translational research on social-affective function. Ventromedial prefrontal cortex (vmPFC) has recently been implicated in deploying visual attention to the eyes of emotional faces, although there is mixed evidence regarding the importance of this brain region for recognition accuracy. In the present study of neurological patients with vmPFC damage, we used an emotion recognition task with morphed facial expressions of varying intensities to determine (1) whether vmPFC is essential for emotion recognition accuracy, and (2) whether instructed attention to the eyes of faces would be sufficient to improve any accuracy deficits. We found that vmPFC lesion patients are impaired, relative to neurologically healthy adults, at recognizing moderate intensity expressions of anger and that recognition accuracy can be improved by providing instructions of where to fixate. These results suggest that vmPFC may be important for the recognition of facial emotion through a role in guiding visual attention to emotionally salient regions of faces. Copyright © 2016 Elsevier Ltd. All rights reserved.
Does vigilance to pain make individuals experts in facial recognition of pain?
Baum, Corinna; Kappesser, Judith; Schneider, Raphaela; Lautenbacher, Stefan
2013-01-01
It is well known that individual factors are important in the facial recognition of pain. However, it is unclear whether vigilance to pain as a pain-related attentional mechanism is among these relevant factors. Vigilance to pain may have two different effects on the recognition of facial pain expressions: pain-vigilant individuals may detect pain faces better but overinclude other facial displays, misinterpreting them as expressing pain; or they may be true experts in discriminating between pain and other facial expressions. The present study aimed to test these two hypotheses. Furthermore, pain vigilance was assumed to be a distinct predictor, the impact of which on recognition cannot be completely replaced by related concepts such as pain catastrophizing and fear of pain. Photographs of neutral, happy, angry and pain facial expressions were presented to 40 healthy participants, who were asked to classify them into the appropriate emotion categories and provide a confidence rating for each classification. Additionally, potential predictors of the discrimination performance for pain and anger faces - pain vigilance, pain-related catastrophizing, fear of pain--were assessed using self-report questionnaires. Pain-vigilant participants classified pain faces more accurately and did not misclassify anger as pain faces more frequently. However, vigilance to pain was not related to the confidence of recognition ratings. Pain catastrophizing and fear of pain did not account for the recognition performance. Moderate pain vigilance, as assessed in the present study, appears to be associated with appropriate detection of pain-related cues and not necessarily with the overinclusion of other negative cues.
Does vigilance to pain make individuals experts in facial recognition of pain?
Baum, Corinna; Kappesser, Judith; Schneider, Raphaela; Lautenbacher, Stefan
2013-01-01
BACKGROUND: It is well known that individual factors are important in the facial recognition of pain. However, it is unclear whether vigilance to pain as a pain-related attentional mechanism is among these relevant factors. OBJECTIVES: Vigilance to pain may have two different effects on the recognition of facial pain expressions: pain-vigilant individuals may detect pain faces better but overinclude other facial displays, misinterpreting them as expressing pain; or they may be true experts in discriminating between pain and other facial expressions. The present study aimed to test these two hypotheses. Furthermore, pain vigilance was assumed to be a distinct predictor, the impact of which on recognition cannot be completely replaced by related concepts such as pain catastrophizing and fear of pain. METHODS: Photographs of neutral, happy, angry and pain facial expressions were presented to 40 healthy participants, who were asked to classify them into the appropriate emotion categories and provide a confidence rating for each classification. Additionally, potential predictors of the discrimination performance for pain and anger faces – pain vigilance, pain-related catastrophizing, fear of pain – were assessed using self-report questionnaires. RESULTS: Pain-vigilant participants classified pain faces more accurately and did not misclassify anger as pain faces more frequently. However, vigilance to pain was not related to the confidence of recognition ratings. Pain catastrophizing and fear of pain did not account for the recognition performance. CONCLUSIONS: Moderate pain vigilance, as assessed in the present study, appears to be associated with appropriate detection of pain-related cues and not necessarily with the overinclusion of other negative cues. PMID:23717826
Recognition profile of emotions in natural and virtual faces.
Dyck, Miriam; Winbeck, Maren; Leiberg, Susanne; Chen, Yuhan; Gur, Ruben C; Gur, Rurben C; Mathiak, Klaus
2008-01-01
Computer-generated virtual faces become increasingly realistic including the simulation of emotional expressions. These faces can be used as well-controlled, realistic and dynamic stimuli in emotion research. However, the validity of virtual facial expressions in comparison to natural emotion displays still needs to be shown for the different emotions and different age groups. Thirty-two healthy volunteers between the age of 20 and 60 rated pictures of natural human faces and faces of virtual characters (avatars) with respect to the expressed emotions: happiness, sadness, anger, fear, disgust, and neutral. Results indicate that virtual emotions were recognized comparable to natural ones. Recognition differences in virtual and natural faces depended on specific emotions: whereas disgust was difficult to convey with the current avatar technology, virtual sadness and fear achieved better recognition results than natural faces. Furthermore, emotion recognition rates decreased for virtual but not natural faces in participants over the age of 40. This specific age effect suggests that media exposure has an influence on emotion recognition. Virtual and natural facial displays of emotion may be equally effective. Improved technology (e.g. better modelling of the naso-labial area) may lead to even better results as compared to trained actors. Due to the ease with which virtual human faces can be animated and manipulated, validated artificial emotional expressions will be of major relevance in future research and therapeutic applications.
Bernal-Mondragón, C; Arriaga-Avila, V; Martínez-Abundis, E; Barrera-Mera, B; Mercado-Gómez, O; Guevara-Guzmán, R
2017-02-01
We investigated the short- and long-term effects of extremely low-frequency electromagnetic fields (EMF) on social recognition behavior and expression of α- and β-estrogen receptors (ER). Rats were exposed to 60-Hz electromagnetic fields for 9 or 30 days and tested for social recognition behavior. Immunohistochemistry and western blot assays were performed to evaluate α- and β-ER expression in the olfactory bulb of intact, ovariectomized (OVX), and ovariectomized+estradiol (E2) replacement (OVX+E2). Ovariectomization showed impairment of social recognition after 9 days of EMF exposure and a complete recovery after E2 replacement and so did those after 30 days. Short EMF exposure increased expression of β-ER in intact, but not in the others. Longer exposure produced a decrease in intact but an increase in OVX and OVX+E2. Our findings suggest a significant role for β-estrogen receptors and a lack of effect for α-estrogen receptors on a social recognition task. EMF: extremely low frequency electromagnetic fields; ERs: estrogen receptors; OB: olfactory bulb; OVX: ovariectomized; OVX + E 2 : ovariectomized + estradiol replacement; IEI: interexposure interval; β-ER: beta estrogen receptor; E 2 : replacement of estradiol; GAPDH: glyceraldehyde-3-phosphate dehydrogenase; WB: Western blot; PBS: phosphate-buffer saline; PB: phosphate-buffer.
Recognition Profile of Emotions in Natural and Virtual Faces
Dyck, Miriam; Winbeck, Maren; Leiberg, Susanne; Chen, Yuhan; Gur, Rurben C.; Mathiak, Klaus
2008-01-01
Background Computer-generated virtual faces become increasingly realistic including the simulation of emotional expressions. These faces can be used as well-controlled, realistic and dynamic stimuli in emotion research. However, the validity of virtual facial expressions in comparison to natural emotion displays still needs to be shown for the different emotions and different age groups. Methodology/Principal Findings Thirty-two healthy volunteers between the age of 20 and 60 rated pictures of natural human faces and faces of virtual characters (avatars) with respect to the expressed emotions: happiness, sadness, anger, fear, disgust, and neutral. Results indicate that virtual emotions were recognized comparable to natural ones. Recognition differences in virtual and natural faces depended on specific emotions: whereas disgust was difficult to convey with the current avatar technology, virtual sadness and fear achieved better recognition results than natural faces. Furthermore, emotion recognition rates decreased for virtual but not natural faces in participants over the age of 40. This specific age effect suggests that media exposure has an influence on emotion recognition. Conclusions/Significance Virtual and natural facial displays of emotion may be equally effective. Improved technology (e.g. better modelling of the naso-labial area) may lead to even better results as compared to trained actors. Due to the ease with which virtual human faces can be animated and manipulated, validated artificial emotional expressions will be of major relevance in future research and therapeutic applications. PMID:18985152
Nagarajan, R; Hariharan, M; Satiyan, M
2012-08-01
Developing tools to assist physically disabled and immobilized people through facial expression is a challenging area of research and has attracted many researchers recently. In this paper, luminance stickers based facial expression recognition is proposed. Recognition of facial expression is carried out by employing Discrete Wavelet Transform (DWT) as a feature extraction method. Different wavelet families with their different orders (db1 to db20, Coif1 to Coif 5 and Sym2 to Sym8) are utilized to investigate their performance in recognizing facial expression and to evaluate their computational time. Standard deviation is computed for the coefficients of first level of wavelet decomposition for every order of wavelet family. This standard deviation is used to form a set of feature vectors for classification. In this study, conventional validation and cross validation are performed to evaluate the efficiency of the suggested feature vectors. Three different classifiers namely Artificial Neural Network (ANN), k-Nearest Neighborhood (kNN) and Linear Discriminant Analysis (LDA) are used to classify a set of eight facial expressions. The experimental results demonstrate that the proposed method gives very promising classification accuracies.
Facial Expression Influences Face Identity Recognition During the Attentional Blink
2014-01-01
Emotional stimuli (e.g., negative facial expressions) enjoy prioritized memory access when task relevant, consistent with their ability to capture attention. Whether emotional expression also impacts on memory access when task-irrelevant is important for arbitrating between feature-based and object-based attentional capture. Here, the authors address this question in 3 experiments using an attentional blink task with face photographs as first and second target (T1, T2). They demonstrate reduced neutral T2 identity recognition after angry or happy T1 expression, compared to neutral T1, and this supports attentional capture by a task-irrelevant feature. Crucially, after neutral T1, T2 identity recognition was enhanced and not suppressed when T2 was angry—suggesting that attentional capture by this task-irrelevant feature may be object-based and not feature-based. As an unexpected finding, both angry and happy facial expressions suppress memory access for competing objects, but only angry facial expression enjoyed privileged memory access. This could imply that these 2 processes are relatively independent from one another. PMID:25286076
Facial expression influences face identity recognition during the attentional blink.
Bach, Dominik R; Schmidt-Daffy, Martin; Dolan, Raymond J
2014-12-01
Emotional stimuli (e.g., negative facial expressions) enjoy prioritized memory access when task relevant, consistent with their ability to capture attention. Whether emotional expression also impacts on memory access when task-irrelevant is important for arbitrating between feature-based and object-based attentional capture. Here, the authors address this question in 3 experiments using an attentional blink task with face photographs as first and second target (T1, T2). They demonstrate reduced neutral T2 identity recognition after angry or happy T1 expression, compared to neutral T1, and this supports attentional capture by a task-irrelevant feature. Crucially, after neutral T1, T2 identity recognition was enhanced and not suppressed when T2 was angry-suggesting that attentional capture by this task-irrelevant feature may be object-based and not feature-based. As an unexpected finding, both angry and happy facial expressions suppress memory access for competing objects, but only angry facial expression enjoyed privileged memory access. This could imply that these 2 processes are relatively independent from one another.
Riemer, Valentin; Frommel, Julian; Layher, Georg; Neumann, Heiko; Schrader, Claudia
2017-01-01
The importance of emotions experienced by learners during their interaction with multimedia learning systems, such as serious games, underscores the need to identify sources of information that allow the recognition of learners’ emotional experience without interrupting the learning process. Bodily expression is gaining in attention as one of these sources of information. However, to date, the question of how bodily expression can convey different emotions has largely been addressed in research relying on acted emotion displays. Following a more contextualized approach, the present study aims to identify features of bodily expression (i.e., posture and activity of the upper body and the head) that relate to genuine emotional experience during interaction with a serious game. In a multimethod approach, 70 undergraduates played a serious game relating to financial education while their bodily expression was captured using an off-the-shelf depth-image sensor (Microsoft Kinect). In addition, self-reports of experienced enjoyment, boredom, and frustration were collected repeatedly during gameplay, to address the dynamic changes in emotions occurring in educational tasks. Results showed that, firstly, the intensities of all emotions indeed changed significantly over the course of the game. Secondly, by using generalized estimating equations, distinct features of bodily expression could be identified as significant indicators for each emotion under investigation. A participant keeping their head more turned to the right was positively related to frustration being experienced, whereas keeping their head more turned to the left was positively related to enjoyment. Furthermore, having their upper body positioned more closely to the gaming screen was also positively related to frustration. Finally, increased activity of a participant’s head emerged as a significant indicator of boredom being experienced. These results confirm the value of bodily expression as an indicator of emotional experience in multimedia learning systems. Furthermore, the findings may guide developers of emotion recognition procedures by focusing on the identified features of bodily expression. PMID:28798717
Children's understanding of facial expression of emotion: II. Drawing of emotion-faces.
Missaghi-Lakshman, M; Whissell, C
1991-06-01
67 children from Grades 2, 4, and 7 drew faces representing the emotional expressions of fear, anger, surprise, disgust, happiness, and sadness. The children themselves and 29 adults later decoded the drawings in an emotion-recognition task. Children were the more accurate decoders, and their accuracy and the accuracy of adults increased significantly for judgments of 7th-grade drawings. The emotions happy and sad were most accurately decoded. There were no significant differences associated with sex. In their drawings, children utilized a symbol system that seems to be based on a highlighting or exaggeration of features of the innately governed facial expression of emotion.
RIG-I in RNA virus recognition
Kell, Alison M.; Gale, Michael
2015-01-01
Antiviral immunity is initiated upon host recognition of viral products via non-self molecular patterns known as pathogen-associated molecular patterns (PAMPs). Such recognition initiates signaling cascades that induce intracellular innate immune defenses and an inflammatory response that facilitates development of the acquired immune response. The retinoic acid-inducible gene I (RIG-I) and the RIG-I-like receptor (RLR) protein family are key cytoplasmic pathogen recognition receptors that are implicated in the recognition of viruses across genera and virus families, including functioning as major sensors of RNA viruses, and promoting recognition of some DNA viruses. RIG-I, the charter member of the RLR family, is activated upon binding to PAMP RNA. Activated RIG-I signals by interacting with the adapter protein MAVS leading to a signaling cascade that activates the transcription factors IRF3 and NF-κB. These actions induce the expression of antiviral gene products and the production of type I and III interferons that lead to an antiviral state in the infected cell and surrounding tissue. RIG-I signaling is essential for the control of infection by many RNA viruses. Recently, RIG-I crosstalk with other pathogen recognition receptors and components of the inflammasome has been described. In this review, we discuss the current knowledge regarding the role of RIG-I in recognition of a variety of virus families and its role in programming the adaptive immune response through cross-talk with parallel arms of the innate immune system, including how RIG-I can be leveraged for antiviral therapy. PMID:25749629
Gamond, L; Cattaneo, Z
2016-12-01
Consistent evidence suggests that emotional facial expressions are better recognized when the expresser and the perceiver belong to the same social group (in-group advantage). In this study, we used transcranial magnetic stimulation (TMS) to investigate the possible causal involvement of the dorsomedial prefrontal cortex (dmPFC) and of the right temporo-parietal junction (TPJ), two main nodes of the mentalizing neural network, in mediating the in-group advantage in emotion recognition. Participants performed an emotion discrimination task in a minimal (blue/green) group paradigm. We found that interfering with activity in the dmPFC significantly interfered with the effect of minimal group-membership on emotion recognition, reducing participants' ability to discriminate emotions expressed by in-group members. In turn, rTPJ mainly affected emotion discrimination per se, irrespective of group membership. Overall, our results point to a causal role of the dmPFC in mediating the in-group advantage in emotion recognition, favoring intragroup communication. Copyright © 2016 Elsevier Ltd. All rights reserved.
End-to-End Multimodal Emotion Recognition Using Deep Neural Networks
NASA Astrophysics Data System (ADS)
Tzirakis, Panagiotis; Trigeorgis, George; Nicolaou, Mihalis A.; Schuller, Bjorn W.; Zafeiriou, Stefanos
2017-12-01
Automatic affect recognition is a challenging task due to the various modalities emotions can be expressed with. Applications can be found in many domains including multimedia retrieval and human computer interaction. In recent years, deep neural networks have been used with great success in determining emotional states. Inspired by this success, we propose an emotion recognition system using auditory and visual modalities. To capture the emotional content for various styles of speaking, robust features need to be extracted. To this purpose, we utilize a Convolutional Neural Network (CNN) to extract features from the speech, while for the visual modality a deep residual network (ResNet) of 50 layers. In addition to the importance of feature extraction, a machine learning algorithm needs also to be insensitive to outliers while being able to model the context. To tackle this problem, Long Short-Term Memory (LSTM) networks are utilized. The system is then trained in an end-to-end fashion where - by also taking advantage of the correlations of the each of the streams - we manage to significantly outperform the traditional approaches based on auditory and visual handcrafted features for the prediction of spontaneous and natural emotions on the RECOLA database of the AVEC 2016 research challenge on emotion recognition.
Dalkıran, Mihriban; Tasdemir, Akif; Salihoglu, Tamer; Emul, Murat; Duran, Alaattin; Ugur, Mufit; Yavuz, Ruhi
2017-09-01
People with schizophrenia have impairments in emotion recognition along with other social cognitive deficits. In the current study, we aimed to investigate the immediate benefits of ECT on facial emotion recognition ability. Thirty-two treatment resistant patients with schizophrenia who have been indicated for ECT enrolled in the study. Facial emotion stimuli were a set of 56 photographs that depicted seven basic emotions: sadness, anger, happiness, disgust, surprise, fear, and neutral faces. The average age of the participants was 33.4 ± 10.5 years. The rate of recognizing the disgusted facial expression increased significantly after ECT (p < 0.05) and no significant changes were found in the rest of the facial expressions (p > 0.05). After the ECT, the time period of responding to the fear and happy facial expressions were significantly shorter (p < 0.05). Facial emotion recognition ability is an important social cognitive skill for social harmony, proper relation and living independently. At least, the ECT sessions do not seem to affect facial emotion recognition ability negatively and seem to improve identifying disgusted facial emotion which is related with dopamine enriched regions in brain.
Nosratababadi, Reza; Bagheri, Vahid; Zare-Bidaki, Mohammad; Hakimi, Hamid; Zainodini, Nahid; Kazemi Arababadi, Mohammad
2017-04-01
Chlamydia species are obligate intracellular pathogens causing different infectious diseases particularly asymptomatic genital infections and are also responsible for a wide range of complications. Previous studies showed that there are different immune responses to Chlamydia species and their infections are limited to some cases. Moreover, Chlamydia species are able to alter immune responses through modulating the expression of some immune system related molecules including cytokines. Toll like receptors (TLRs) belonge to pathogen recognition receptors (PRRs) and play vital roles in recognition of microbes and stimulation of appropriate immune responses. Therefore, it appears that TLRs may be considered as important sensors for recognition of Chlamydia and promotion of immune responses against these bacterial infections. Accordingly, TLR4 detects several microbial PAMPs such as bacterial lipopolysacharide (LPS) and subsequently activates transcription from pro-inflammatory cytokines in both MYD88 and TRIF pathways dependent manner. The purpose of this review is to provide the recent data about the status and major roles played by TLR4 in Chlamydia species recognition and promotion of immune responses against these infections and also the relationship between TLR4 activities and pathogenesis of Chlamydia infections. Copyright © 2017 Elsevier Ltd. All rights reserved.
Spiegel, S; Chiu, A; James, A S; Jentsch, J D; Karlsgodt, K H
2015-11-01
Numerous studies have implicated DTNBP1, the gene encoding dystrobrevin-binding protein or dysbindin, as a candidate risk gene for schizophrenia, though this relationship remains somewhat controversial. Variation in dysbindin, and its location on chromosome 6p, has been associated with cognitive processes, including those relying on a complex system of glutamatergic and dopaminergic interactions. Dysbindin is one of the seven protein subunits that comprise the biogenesis of lysosome-related organelles complex 1 (BLOC-1). Dysbindin protein levels are lower in mice with null mutations in pallidin, another gene in the BLOC-1, and pallidin levels are lower in mice with null mutations in the dysbindin gene, suggesting that multiple subunit proteins must be present to form a functional oligomeric complex. Furthermore, pallidin and dysbindin have similar distribution patterns in a mouse and human brain. Here, we investigated whether the apparent correspondence of pallid and dysbindin at the level of gene expression is also found at the level of behavior. Hypothesizing a mutation leading to underexpression of either of these proteins should show similar phenotypic effects, we studied recognition memory in both strains using the novel object recognition task (NORT) and social novelty recognition task (SNRT). We found that mice with a null mutation in either gene are impaired on SNRT and NORT when compared with wild-type controls. These results support the conclusion that deficits consistent with recognition memory impairment, a cognitive function that is impaired in schizophrenia, result from either pallidin or dysbindin mutations, possibly through degradation of BLOC-1 expression and/or function. © 2015 John Wiley & Sons Ltd and International Behavioural and Neural Genetics Society.
Smitha, K G; Vinod, A P
2015-11-01
Children with autism spectrum disorder have difficulty in understanding the emotional and mental states from the facial expressions of the people they interact. The inability to understand other people's emotions will hinder their interpersonal communication. Though many facial emotion recognition algorithms have been proposed in the literature, they are mainly intended for processing by a personal computer, which limits their usability in on-the-move applications where portability is desired. The portability of the system will ensure ease of use and real-time emotion recognition and that will aid for immediate feedback while communicating with caretakers. Principal component analysis (PCA) has been identified as the least complex feature extraction algorithm to be implemented in hardware. In this paper, we present a detailed study of the implementation of serial and parallel implementation of PCA in order to identify the most feasible method for realization of a portable emotion detector for autistic children. The proposed emotion recognizer architectures are implemented on Virtex 7 XC7VX330T FFG1761-3 FPGA. We achieved 82.3% detection accuracy for a word length of 8 bits.
Martinelli, Eugenio; Mencattini, Arianna; Daprati, Elena; Di Natale, Corrado
2016-01-01
Humans can communicate their emotions by modulating facial expressions or the tone of their voice. Albeit numerous applications exist that enable machines to read facial emotions and recognize the content of verbal messages, methods for speech emotion recognition are still in their infancy. Yet, fast and reliable applications for emotion recognition are the obvious advancement of present 'intelligent personal assistants', and may have countless applications in diagnostics, rehabilitation and research. Taking inspiration from the dynamics of human group decision-making, we devised a novel speech emotion recognition system that applies, for the first time, a semi-supervised prediction model based on consensus. Three tests were carried out to compare this algorithm with traditional approaches. Labeling performances relative to a public database of spontaneous speeches are reported. The novel system appears to be fast, robust and less computationally demanding than traditional methods, allowing for easier implementation in portable voice-analyzers (as used in rehabilitation, research, industry, etc.) and for applications in the research domain (such as real-time pairing of stimuli to participants' emotional state, selective/differential data collection based on emotional content, etc.).
Violent media consumption and the recognition of dynamic facial expressions.
Kirsh, Steven J; Mounts, Jeffrey R W; Olczak, Paul V
2006-05-01
This study assessed the speed of recognition of facial emotional expressions (happy and angry) as a function of violent media consumption. Color photos of calm facial expressions morphed to either an angry or a happy facial expression. Participants were asked to make a speeded identification of the emotion (happiness or anger) during the morph. Results indicated that, independent of trait aggressiveness, participants high in violent media consumption responded slower to depictions of happiness and faster to depictions of anger than participants low in violent media consumption. Implications of these findings are discussed with respect to current models of aggressive behavior.
ERIC Educational Resources Information Center
Brosnan, Mark; Johnson, Hilary; Grawmeyer, Beate; Chapman, Emma; Benton, Laura
2015-01-01
There is equivocal evidence as to whether there is a deficit in recognising emotional expressions in Autism spectrum disorder (ASD). This study compared emotion recognition in ASD in three types of emotion expression media (still image, dynamic image, auditory) across human stimuli (e.g. photo of a human face) and animated stimuli (e.g. cartoon…
Engineering Translational Activators with CRISPR-Cas System.
Du, Pei; Miao, Chensi; Lou, Qiuli; Wang, Zefeng; Lou, Chunbo
2016-01-15
RNA parts often serve as critical components in genetic engineering. Here we report a design of translational activators which is composed of an RNA endoribonuclease (Csy4) and two exchangeable RNA modules. Csy4, a member of Cas endoribonuclease, cleaves at a specific recognition site; this cleavage releases a cis-repressive RNA module (crRNA) from the masked ribosome binding site (RBS), which subsequently allows the downstream translation initiation. Unlike small RNA as a translational activator, the endoribonuclease-based activator is able to efficiently unfold the perfect RBS-crRNA pairing. As an exchangeable module, the crRNA-RBS duplex was forwardly and reversely engineered to modulate the dynamic range of translational activity. We further showed that Csy4 and its recognition site, together as a module, can also be replaced by orthogonal endoribonuclease-recognition site homologues. These modularly structured, high-performance translational activators would endow the programming of gene expression in the translation level with higher feasibility.
Body Emotion Recognition Disproportionately Depends on Vertical Orientations during Childhood
ERIC Educational Resources Information Center
Balas, Benjamin; Auen, Amanda; Saville, Alyson; Schmidt, Jamie
2018-01-01
Children's ability to recognize emotional expressions from faces and bodies develops during childhood. However, the low-level features that support accurate body emotion recognition during development have not been well characterized. This is in marked contrast to facial emotion recognition, which is known to depend upon specific spatial frequency…
Towards Real-Time Speech Emotion Recognition for Affective E-Learning
ERIC Educational Resources Information Center
Bahreini, Kiavash; Nadolski, Rob; Westera, Wim
2016-01-01
This paper presents the voice emotion recognition part of the FILTWAM framework for real-time emotion recognition in affective e-learning settings. FILTWAM (Framework for Improving Learning Through Webcams And Microphones) intends to offer timely and appropriate online feedback based upon learner's vocal intonations and facial expressions in order…
Wang, Jing; Wang, Haoyuan; Yang, Le; Lv, Liping; Zhang, Zhe; Ren, Bin; Dong, Lichun; Li, Ning
2018-04-01
In this paper, a novel riboregulator Switch System of Gene Expression including an OFF-TO-ON switch and an ON-TO-OFF switch was designed to regulate the expression state of target genes between "ON" and "OFF" by switching the identifiability of ribosome recognition site (RBS) based on the thermodynamic stability of different RNA-RNA hybridizations between RBS and small noncoding RNAs. The proposed riboregulator switch system was employed for the fermentative production of succinic acid using an engineered strain of E. coli JW1021, during which the expression of mgtC gene was controlled at "ON" state and that of pepc and ecaA genes were controlled at the "OFF" state in the lag phase and switched to the "OFF" and "ON" state once the strain enters the logarithmic phase. The results showed that using the strain of JW1021, the yield and productivity of succinic acid can reach 0.91 g g -1 and 3.25 g L -1 h -1 , respectively, much higher than those using the strains without harboring the riboregulator switch system.
Calcaterra, Roberta; Di Girolamo, Michele; Mirisola, Concetta; Baggi, Luigi
2016-06-01
Gingival epithelial cells have a pivotal role in the recognition of microorganisms and damage-associated molecular pattern molecules and in the regulation of the immune response. The investigation of the behavior of Toll-like receptors (TLRs) and nucleotide oligomerization domain (NOD) like receptors (NLRs) around a healthy implant may help to address the first step of periimplantitis pathogenesis. To investigate by quantitative real-time polymerase chain reaction, the mRNA expressions of TLR2, TLR3, TLR4, TLR5, TLR6, TLR9, NOD1, NOD2, and NLRP3 from gingival epithelial cells of the sulcus around healthy implants and around healthy teeth. Two types of implant-abutment systems with tube-in-tube interface were tested. After 6 months of implant restoration, gingival epithelial cells were obtained from the gingival sulcus around the implants and around the adjacent teeth of 10 patients. Our results did not reach statistical significance among the mRNA expressions of TLR2, TLR3, TLR4, TLR5, TLR6, TLR9, NOD1, NOD2, and NLRP3 in epithelial cells around the implant versus around natural teeth. This study shows that the implant-abutment systems tested did not induce an immune response by the surrounding epithelial cells at 6 months since their positioning, as well as in the adjacent clincally healthy teeth.
Optimal Geometrical Set for Automated Marker Placement to Virtualized Real-Time Facial Emotions
Maruthapillai, Vasanthan; Murugappan, Murugappan
2016-01-01
In recent years, real-time face recognition has been a major topic of interest in developing intelligent human-machine interaction systems. Over the past several decades, researchers have proposed different algorithms for facial expression recognition, but there has been little focus on detection in real-time scenarios. The present work proposes a new algorithmic method of automated marker placement used to classify six facial expressions: happiness, sadness, anger, fear, disgust, and surprise. Emotional facial expressions were captured using a webcam, while the proposed algorithm placed a set of eight virtual markers on each subject’s face. Facial feature extraction methods, including marker distance (distance between each marker to the center of the face) and change in marker distance (change in distance between the original and new marker positions), were used to extract three statistical features (mean, variance, and root mean square) from the real-time video sequence. The initial position of each marker was subjected to the optical flow algorithm for marker tracking with each emotional facial expression. Finally, the extracted statistical features were mapped into corresponding emotional facial expressions using two simple non-linear classifiers, K-nearest neighbor and probabilistic neural network. The results indicate that the proposed automated marker placement algorithm effectively placed eight virtual markers on each subject’s face and gave a maximum mean emotion classification rate of 96.94% using the probabilistic neural network. PMID:26859884
Optimal Geometrical Set for Automated Marker Placement to Virtualized Real-Time Facial Emotions.
Maruthapillai, Vasanthan; Murugappan, Murugappan
2016-01-01
In recent years, real-time face recognition has been a major topic of interest in developing intelligent human-machine interaction systems. Over the past several decades, researchers have proposed different algorithms for facial expression recognition, but there has been little focus on detection in real-time scenarios. The present work proposes a new algorithmic method of automated marker placement used to classify six facial expressions: happiness, sadness, anger, fear, disgust, and surprise. Emotional facial expressions were captured using a webcam, while the proposed algorithm placed a set of eight virtual markers on each subject's face. Facial feature extraction methods, including marker distance (distance between each marker to the center of the face) and change in marker distance (change in distance between the original and new marker positions), were used to extract three statistical features (mean, variance, and root mean square) from the real-time video sequence. The initial position of each marker was subjected to the optical flow algorithm for marker tracking with each emotional facial expression. Finally, the extracted statistical features were mapped into corresponding emotional facial expressions using two simple non-linear classifiers, K-nearest neighbor and probabilistic neural network. The results indicate that the proposed automated marker placement algorithm effectively placed eight virtual markers on each subject's face and gave a maximum mean emotion classification rate of 96.94% using the probabilistic neural network.
Cao, Yan; Wu, Ruiyong; Tai, Fadao; Zhang, Xia; Yu, Peng; An, Xiaolei; Qiao, Xufeng; Hao, Ping
2014-01-01
Paternal care is necessary for the healthy development of social behavior in monogamous rodents and social recognition underpins social behavior in these animals. The effects of paternal care on the development of social recognition and underlying neuroendocrine mechanisms, especially the involvement of oxytocin and estrogen pathways, remain poorly understood. We investigated the effects of paternal deprivation (PD: father was removed from neonatal pups and mother alone raised the offspring) on social recognition in mandarin voles (Microtus mandarinus), a socially monogamous rodent. Paternal deprivation was found to inhibit the development of social recognition in female and male offspring according to a habituation-dishabituation paradigm. Paternal deprivation resulted in increased inactivity and reduced investigation during new encounters with other animals. Paternal deprivation reduced oxytocin receptor (OTR) and estrogen receptor α (ERα) mRNA expression in the medial amygdala and nucleus accumbens. Paternal deprivation reduced serum oxytocin (OT) concentration in females, but had no effect on males. Our results provide substantial evidence that paternal deprivation inhibits the development of social recognition in female and male mandarin voles and alters social behavior later in life. This is possibly the result of altered expression of central OTR and ERα and serum OT levels caused by paternal deprivation. Copyright © 2013 Elsevier Inc. All rights reserved.
Recognition, survival and persistence of Staphylococcus aureus in the model host Tenebrio molitor.
Dorling, Jack; Moraes, Caroline; Rolff, Jens
2015-02-01
The degree of specificity of any given immune response to a parasite is governed by the complexity and variation of interactions between host and pathogen derived molecules. Here, we assess the extent to which recognition and immuno-resistance of cell wall mutants of the pathogen Staphylococcus aureus may contribute to establishment and maintenance of persistent infection in the model insect host, Tenebrio molitor. The cell surface of S. aureus is decorated with various molecules, including glycopolymers such as wall teichoic acid (WTA). WTA is covalently bound to peptidoglycan (PGN) and its absence has been associated with increased recognition of PGN by host receptors (PGRPs). WTA is also further modified by other molecules such as D-alanine (D-alanylation). Both the level of WTA expression and its D-alanylation were found to be important in the mediation of the host-parasite interaction in this model system. Specifically, WTA itself was seen to influence immune recognition, while D-alanylation of WTA was found to increase immuno-resistance and was associated with prolonged persistence of S. aureus in T. molitor. These results implicate WTA and its D-alanylation as important factors in the establishment and maintenance of persistent infection, affecting different critical junctions in the immune response; through potential evasion of recognition by PGRPs and resistance to humoral immune effectors during prolonged exposure to the immune system. This highlights a mechanism by which specificity in this host-parasite interaction may arise. Copyright © 2014 Elsevier Ltd. All rights reserved.
An intrinsic vasopressin system in the olfactory bulb is involved in social recognition
Tobin, Vicky A.; Hashimoto, Hirofumi; Wacker, Douglas W.; Takayanagi, Yuki; Langnaese, Kristina; Caquineau, Celine; Noack, Julia; Landgraf, Rainer; Onaka, Tatsushi; Leng, Gareth; Meddle, Simone L.; Engelmann, Mario; Ludwig, Mike
2010-01-01
Many peptides, when released as chemical messengers within the brain, have powerful influences on complex behaviours. Most strikingly, vasopressin and oxytocin, once thought of as circulating hormones whose actions were confined to peripheral organs, are now known to be released in the brain where they play fundamentally important roles in social behaviours1. In humans, disruptions of these peptide systems have been linked to several neurobehavioural disorders, including Prader-Willi syndrome, affective disorders, and obsessive-compulsive disorder, and polymorphisms of the vasopressin V1a receptor have been linked to autism2,3. Here we report that the rat olfactory bulb contains a large population of interneurones which express vasopressin, that blocking the actions of vasopressin in the olfactory bulb impairs the social recognition abilities of rats, and that vasopressin agonists and antagonists can modulate the processing of information by olfactory bulb neurones. The findings indicate that social information is processed in part by a vasopressin system intrinsic to the olfactory system. PMID:20182426
Near infrared and visible face recognition based on decision fusion of LBP and DCT features
NASA Astrophysics Data System (ADS)
Xie, Zhihua; Zhang, Shuai; Liu, Guodong; Xiong, Jinquan
2018-03-01
Visible face recognition systems, being vulnerable to illumination, expression, and pose, can not achieve robust performance in unconstrained situations. Meanwhile, near infrared face images, being light- independent, can avoid or limit the drawbacks of face recognition in visible light, but its main challenges are low resolution and signal noise ratio (SNR). Therefore, near infrared and visible fusion face recognition has become an important direction in the field of unconstrained face recognition research. In order to extract the discriminative complementary features between near infrared and visible images, in this paper, we proposed a novel near infrared and visible face fusion recognition algorithm based on DCT and LBP features. Firstly, the effective features in near-infrared face image are extracted by the low frequency part of DCT coefficients and the partition histograms of LBP operator. Secondly, the LBP features of visible-light face image are extracted to compensate for the lacking detail features of the near-infrared face image. Then, the LBP features of visible-light face image, the DCT and LBP features of near-infrared face image are sent to each classifier for labeling. Finally, decision level fusion strategy is used to obtain the final recognition result. The visible and near infrared face recognition is tested on HITSZ Lab2 visible and near infrared face database. The experiment results show that the proposed method extracts the complementary features of near-infrared and visible face images and improves the robustness of unconstrained face recognition. Especially for the circumstance of small training samples, the recognition rate of proposed method can reach 96.13%, which has improved significantly than 92.75 % of the method based on statistical feature fusion.
The level of cognitive function and recognition of emotions in older adults
Singh-Manoux, Archana; Batty, G. David; Ebmeier, Klaus P.; Jokela, Markus; Harmer, Catherine J.; Kivimäki, Mika
2017-01-01
Background The association between cognitive decline and the ability to recognise emotions in interpersonal communication is not well understood. We aimed to investigate the association between cognitive function and the ability to recognise emotions in other people’s facial expressions across the full continuum of cognitive capacity. Methods Cross-sectional analysis of 4039 participants (3016 men, 1023 women aged 59 to 82 years) in the Whitehall II study. Cognitive function was assessed using a 30-item Mini-Mental State Examination (MMSE), further classified into 8 groups: 30, 29, 28, 27, 26, 25, 24, and <24 (possible dementia) MMSE points. The Facial Expression Recognition Task (FERT) was used to examine recognition of anger, fear, disgust, sadness, and happiness. Results The multivariable adjusted difference in the percentage of accurate recognition between the highest and lowest MMSE group was 14.9 (95%CI, 11.1–18.7) for anger, 15.5 (11.9–19.2) for fear, 18.5 (15.2–21.8) for disgust, 11.6 (7.3–16.0) for sadness, and 6.3 (3.1–9.4) for happiness. However, recognition of several emotions was reduced already after 1 to 2-point reduction in MMSE and with further points down in MMSE, the recognition worsened at an accelerated rate. Conclusions The ability to recognize emotion in facial expressions is affected at an early stage of cognitive impairment and might decline at an accelerated rate with the deterioration of cognitive function. Accurate recognition of happiness seems to be less affected by a severe decline in cognitive performance than recognition of negatively valued emotions. PMID:28977015
Inconsistent emotion recognition deficits across stimulus modalities in Huntington׳s disease.
Rees, Elin M; Farmer, Ruth; Cole, James H; Henley, Susie M D; Sprengelmeyer, Reiner; Frost, Chris; Scahill, Rachael I; Hobbs, Nicola Z; Tabrizi, Sarah J
2014-11-01
Recognition of negative emotions is impaired in Huntington׳s Disease (HD). It is unclear whether these emotion-specific problems are driven by dissociable cognitive deficits, emotion complexity, test cue difficulty, or visuoperceptual impairments. This study set out to further characterise emotion recognition in HD by comparing patterns of deficits across stimulus modalities; notably including for the first time in HD, the more ecologically and clinically relevant modality of film clips portraying dynamic facial expressions. Fifteen early HD and 17 control participants were tested on emotion recognition from static facial photographs, non-verbal vocal expressions and one second dynamic film clips, all depicting different emotions. Statistically significant evidence of impairment of anger, disgust and fear recognition was seen in HD participants compared with healthy controls across multiple stimulus modalities. The extent of the impairment, as measured by the difference in the number of errors made between HD participants and controls, differed according to the combination of emotion and modality (p=0.013, interaction test). The largest between-group difference was seen in the recognition of anger from film clips. Consistent with previous reports, anger, disgust and fear were the most poorly recognised emotions by the HD group. This impairment did not appear to be due to task demands or expression complexity as the pattern of between-group differences did not correspond to the pattern of errors made by either group; implicating emotion-specific cognitive processing pathology. There was however evidence that the extent of emotion recognition deficits significantly differed between stimulus modalities. The implications in terms of designing future tests of emotion recognition and care giving are discussed. Copyright © 2014 Elsevier Ltd. All rights reserved.
Modulation of α power and functional connectivity during facial affect recognition.
Popov, Tzvetan; Miller, Gregory A; Rockstroh, Brigitte; Weisz, Nathan
2013-04-03
Research has linked oscillatory activity in the α frequency range, particularly in sensorimotor cortex, to processing of social actions. Results further suggest involvement of sensorimotor α in the processing of facial expressions, including affect. The sensorimotor face area may be critical for perception of emotional face expression, but the role it plays is unclear. The present study sought to clarify how oscillatory brain activity contributes to or reflects processing of facial affect during changes in facial expression. Neuromagnetic oscillatory brain activity was monitored while 30 volunteers viewed videos of human faces that changed their expression from neutral to fearful, neutral, or happy expressions. Induced changes in α power during the different morphs, source analysis, and graph-theoretic metrics served to identify the role of α power modulation and cross-regional coupling by means of phase synchrony during facial affect recognition. Changes from neutral to emotional faces were associated with a 10-15 Hz power increase localized in bilateral sensorimotor areas, together with occipital power decrease, preceding reported emotional expression recognition. Graph-theoretic analysis revealed that, in the course of a trial, the balance between sensorimotor power increase and decrease was associated with decreased and increased transregional connectedness as measured by node degree. Results suggest that modulations in α power facilitate early registration, with sensorimotor cortex including the sensorimotor face area largely functionally decoupled and thereby protected from additional, disruptive input and that subsequent α power decrease together with increased connectedness of sensorimotor areas facilitates successful facial affect recognition.
Marsh, Abigail A
2016-06-01
Because the face is the central focus of human social interactions, emotional facial expressions provide a unique window into the emotional lives of others. They play a particularly important role in fostering empathy, which entails understanding and responding to others' emotions, especially distress-related emotions such as fear. This Review considers how fearful facial as well as vocal and postural expressions are interpreted, with an emphasis on the role of the amygdala. The amygdala may be best known for its role in the acquisition and expression of conditioned fear, but it also supports the perception and recognition of others' fear. Various explanations have been supplied for the amygdala's role in interpreting and responding to fearful expressions. They include theories that amygdala responses to fearful expressions 1) reflect heightened vigilance in response to uncertain danger, 2) promote heightened attention to the eye region of faces, 3) represent a response to an unconditioned aversive stimulus, or 4) reflect the generation of an empathic fear response. Among these, only empathic fear explains why amygdala lesions would impair fear recognition across modalities. Supporting the possibility of a link between fundamental empathic processes and amygdala responses to fear is evidence that impaired fear recognition in psychopathic individuals results from amygdala dysfunction, whereas enhanced fear recognition in altruistic individuals results from enhanced amygdala function. Empathic concern and caring behaviors may be fostered by sensitivity to signs of acute distress in others, which relies on intact functioning of the amygdala. © 2015 Wiley Periodicals, Inc.
Toll-Like Receptor Pathways in Autoimmune Diseases.
Chen, Ji-Qing; Szodoray, Peter; Zeher, Margit
2016-02-01
Autoimmune diseases are a family of chronic systemic inflammatory disorders, characterized by the dysregulation of the immune system which finally results in the break of tolerance to self-antigen. Several studies suggest that Toll-like receptors (TLRs) play an essential role in the pathogenesis of autoimmune diseases. TLRs belong to the family of pattern recognition receptors (PRRs) that recognize a wide range of pathogen-associated molecular patterns (PAMPs). TLRs are type I transmembrane proteins and located on various cellular membranes. Two main groups have been classified based on their location; the extracelluar group referred to the ones located on the plasma membrane while the intracellular group all located in endosomal compartments responsible for the recognition of nucleic acids. They are released by the host cells and trigger various intracellular pathways which results in the production of proinflammatory cytokines, chemokines, as well as the expression of co-stimulatory molecules to protect against invading microorganisms. In particular, TLR pathway-associated proteins, such as IRAK, TRAF, and SOCS, are often dysregulated in this group of diseases. TLR-associated gene expression profile analysis together with single nucleotide polymorphism (SNP) assessment could be important to explain the pathomechanism driving autoimmune diseases. In this review, we summarize recent findings on TLR pathway regulation in various autoimmune diseases, including Sjögren's syndrome (SS), systemic lupus erythematosus (SLE), multiple sclerosis (MS), rheumatoid arthritis (RA), systemic sclerosis (SSc), and psoriasis.
Behler, Friederike; Maus, Regina; Bohling, Jennifer; Knippenberg, Sarah; Kirchhof, Gabriele; Nagata, Masahiro; Jonigk, Danny; Izykowski, Nicole; Mägel, Lavinia; Welte, Tobias; Yamasaki, Sho
2014-01-01
The macrophage-inducible C-type lectin Mincle has recently been identified to be a pattern recognition receptor sensing mycobacterial infection via recognition of the mycobacterial cell wall component trehalose-6′,6-dimycolate (TDM). However, its role in systemic mycobacterial infections has not been examined so far. Mincle-knockout (KO) mice were infected intravenously with Mycobacterium bovis BCG to mimic the systemic spread of mycobacteria under defined experimental conditions. After intravenous infection with M. bovis BCG, Mincle-KO mice responded with significantly higher numbers of mycobacterial CFU in spleen and liver, while reduced granuloma formation was observed only in the spleen. At the same time, reduced Th1 cytokine production and decreased numbers of gamma interferon-producing T cells were observed in the spleens of Mincle-KO mice relative to the numbers in the spleens of wild-type (WT) mice. The effect of adoptive transfer of defined WT leukocyte subsets generated from bone marrow cells of zDC+/DTR mice (which bear the human diphtheria toxin receptor [DTR] under the control of the classical dendritic cell-specific zinc finger transcription factor zDC) to specifically deplete Mincle-expressing classical dendritic cells (cDCs) but not macrophages after diphtheria toxin application on the numbers of splenic and hepatic CFU and T cell subsets was then determined. Adoptive transfer experiments revealed that Mincle-expressing splenic cDCs rather than Mincle-expressing macrophages contributed to the reconstitution of attenuated splenic antimycobacterial immune responses in Mincle-KO mice after intravenous challenge with BCG. Collectively, we show that expression of Mincle, particularly by cDCs, contributes to the control of splenic M. bovis BCG infection in mice. PMID:25332121
Wells, Laura Jean; Gillespie, Steven Mark; Rotshtein, Pia
2016-01-01
The identification of emotional expressions is vital for social interaction, and can be affected by various factors, including the expressed emotion, the intensity of the expression, the sex of the face, and the gender of the observer. This study investigates how these factors affect the speed and accuracy of expression recognition, as well as dwell time on the two most significant areas of the face: the eyes and the mouth. Participants were asked to identify expressions from female and male faces displaying six expressions (anger, disgust, fear, happiness, sadness, and surprise), each with three levels of intensity (low, moderate, and normal). Overall, responses were fastest and most accurate for happy expressions, but slowest and least accurate for fearful expressions. More intense expressions were also classified most accurately. Reaction time showed a different pattern, with slowest response times recorded for expressions of moderate intensity. Overall, responses were slowest, but also most accurate, for female faces. Relative to male observers, women showed greater accuracy and speed when recognizing female expressions. Dwell time analyses revealed that attention to the eyes was about three times greater than on the mouth, with fearful eyes in particular attracting longer dwell times. The mouth region was attended to the most for fearful, angry, and disgusted expressions and least for surprise. These results extend upon previous findings to show important effects of expression, emotion intensity, and sex on expression recognition and gaze behaviour, and may have implications for understanding the ways in which emotion recognition abilities break down.
Rotshtein, Pia
2016-01-01
The identification of emotional expressions is vital for social interaction, and can be affected by various factors, including the expressed emotion, the intensity of the expression, the sex of the face, and the gender of the observer. This study investigates how these factors affect the speed and accuracy of expression recognition, as well as dwell time on the two most significant areas of the face: the eyes and the mouth. Participants were asked to identify expressions from female and male faces displaying six expressions (anger, disgust, fear, happiness, sadness, and surprise), each with three levels of intensity (low, moderate, and normal). Overall, responses were fastest and most accurate for happy expressions, but slowest and least accurate for fearful expressions. More intense expressions were also classified most accurately. Reaction time showed a different pattern, with slowest response times recorded for expressions of moderate intensity. Overall, responses were slowest, but also most accurate, for female faces. Relative to male observers, women showed greater accuracy and speed when recognizing female expressions. Dwell time analyses revealed that attention to the eyes was about three times greater than on the mouth, with fearful eyes in particular attracting longer dwell times. The mouth region was attended to the most for fearful, angry, and disgusted expressions and least for surprise. These results extend upon previous findings to show important effects of expression, emotion intensity, and sex on expression recognition and gaze behaviour, and may have implications for understanding the ways in which emotion recognition abilities break down. PMID:27942030
Associations between facial emotion recognition and young adolescents’ behaviors in bullying
Gini, Gianluca; Altoè, Gianmarco
2017-01-01
This study investigated whether different behaviors young adolescents can act during bullying episodes were associated with their ability to recognize morphed facial expressions of the six basic emotions, expressed at high and low intensity. The sample included 117 middle-school students (45.3% girls; mean age = 12.4 years) who filled in a peer nomination questionnaire and individually performed a computerized emotion recognition task. Bayesian generalized mixed-effects models showed a complex picture, in which type and intensity of emotions, students’ behavior and gender interacted in explaining recognition accuracy. Results were discussed with a particular focus on negative emotions and suggesting a “neutral” nature of emotion recognition ability, which does not necessarily lead to moral behavior but can also be used for pursuing immoral goals. PMID:29131871
[Face recognition in patients with schizophrenia].
Doi, Hirokazu; Shinohara, Kazuyuki
2012-07-01
It is well known that patients with schizophrenia show severe deficiencies in social communication skills. These deficiencies are believed to be partly derived from abnormalities in face recognition. However, the exact nature of these abnormalities exhibited by schizophrenic patients with respect to face recognition has yet to be clarified. In the present paper, we review the main findings on face recognition deficiencies in patients with schizophrenia, particularly focusing on abnormalities in the recognition of facial expression and gaze direction, which are the primary sources of information of others' mental states. The existing studies reveal that the abnormal recognition of facial expression and gaze direction in schizophrenic patients is attributable to impairments in both perceptual processing of visual stimuli, and cognitive-emotional responses to social information. Furthermore, schizophrenic patients show malfunctions in distributed neural regions, ranging from the fusiform gyrus recruited in the structural encoding of facial stimuli, to the amygdala which plays a primary role in the detection of the emotional significance of stimuli. These findings were obtained from research in patient groups with heterogeneous characteristics. Because previous studies have indicated that impairments in face recognition in schizophrenic patients might vary according to the types of symptoms, it is of primary importance to compare the nature of face recognition deficiencies and the impairments of underlying neural functions across sub-groups of patients.
Lin, Chia-Yao; Tien, Yi-Min; Huang, Jong-Tsun; Tsai, Chon-Haw; Hsu, Li-Chuan
2016-01-01
Because of dopaminergic neurodegeneration, patients with Parkinson's disease (PD) show impairment in the recognition of negative facial expressions. In the present study, we aimed to determine whether PD patients with more advanced motor problems would show a much greater deficit in recognition of emotional facial expressions than a control group and whether impairment of emotion recognition would extend to positive emotions. Twenty-nine PD patients and 29 age-matched healthy controls were recruited. Participants were asked to discriminate emotions in Experiment 1 and identify gender in Experiment 2. In Experiment 1, PD patients demonstrated a recognition deficit for negative (sadness and anger) and positive faces. Further analysis showed that only PD patients with high motor dysfunction performed poorly in recognition of happy faces. In Experiment 2, PD patients showed an intact ability for gender identification, and the results eliminated possible abilities in the functions measured in Experiment 2 as alternative explanations for the results of Experiment 1. We concluded that patients' ability to recognize emotions deteriorated as the disease progressed. Recognition of negative emotions was impaired first, and then the impairment extended to positive emotions.
Tien, Yi-Min; Huang, Jong-Tsun
2016-01-01
Because of dopaminergic neurodegeneration, patients with Parkinson's disease (PD) show impairment in the recognition of negative facial expressions. In the present study, we aimed to determine whether PD patients with more advanced motor problems would show a much greater deficit in recognition of emotional facial expressions than a control group and whether impairment of emotion recognition would extend to positive emotions. Twenty-nine PD patients and 29 age-matched healthy controls were recruited. Participants were asked to discriminate emotions in Experiment 1 and identify gender in Experiment 2. In Experiment 1, PD patients demonstrated a recognition deficit for negative (sadness and anger) and positive faces. Further analysis showed that only PD patients with high motor dysfunction performed poorly in recognition of happy faces. In Experiment 2, PD patients showed an intact ability for gender identification, and the results eliminated possible abilities in the functions measured in Experiment 2 as alternative explanations for the results of Experiment 1. We concluded that patients' ability to recognize emotions deteriorated as the disease progressed. Recognition of negative emotions was impaired first, and then the impairment extended to positive emotions. PMID:27555668
Age differences in right-wing authoritarianism and their relation to emotion recognition.
Ruffman, Ted; Wilson, Marc; Henry, Julie D; Dawson, Abigail; Chen, Yan; Kladnitski, Natalie; Myftari, Ella; Murray, Janice; Halberstadt, Jamin; Hunter, John A
2016-03-01
This study examined the correlates of right-wing authoritarianism (RWA) in older adults. Participants were given tasks measuring emotion recognition, executive functions and fluid IQ and questionnaires measuring RWA, perceived threat and social dominance orientation. Study 1 established higher age-related RWA across the age span in more than 2,600 New Zealanders. Studies 2 to 4 found that threat, education, social dominance and age all predicted unique variance in older adults' RWA, but the most consistent predictor was emotion recognition, predicting unique variance in older adults' RWA independent of all other variables. We argue that older adults' worse emotion recognition is associated with a more general change in social judgment. Expression of extreme attitudes (right- or left-wing) has the potential to antagonize others, but worse emotion recognition means that subtle signals will not be perceived, making the expression of extreme attitudes more likely. Our findings are consistent with other studies showing that worsening emotion recognition underlies age-related declines in verbosity, understanding of social gaffes, and ability to detect lies. Such results indicate that emotion recognition is a core social insight linked to many aspects of social cognition. (c) 2016 APA, all rights reserved).
Hyaluronan functionalizing QDs as turn-on fluorescent probe for targeted recognition CD44 receptor
NASA Astrophysics Data System (ADS)
Zhou, Shang; Huo, Danqun; Hou, Changjun; Yang, Mei; Fa, Huanbao
2017-09-01
The recognition of tumor markers in living cancer cells has attracted increasing interest. In the present study, the turn-on fluorescence probe was designed based on the fluorescence of thiolated chitosan-coated CdTe QDs (CdTe/TCS QDs) quenched by hyaluronan, which could provide the low background signal for sensitive cellular imaging. This system is expected to offer specific recognition of CD44 receptor over other substances owing to the specific affinity of hyaluronan and CD44 receptor ( 8-9 kcal/mol). The probe is stable in aqueous and has little toxicity to living cells; thus, it can be utilized for targeted cancer cell imaging. The living lung cancer cell imaging experiments further demonstrate its value in recognizing cell-surface CD44 receptor with turn-on mode. In addition, the probe can be used to recognize and differentiate the subtypes of lung cancer cells based on the difference of CD44 expression on the surface of lung cancer cells. And, the western blot test further confirmed that the expression level of the CD44 receptor in lung cancer cells is different. Therefore, this probe may be potentially applied in recognizing lung cancer cells with higher contrast and sensitivity and provide new tools for cancer prognosis and therapy. [Figure not available: see fulltext.
Yang, P-J; Zhan, M-Y; Ye, C; Yu, X-Q; Rao, X-J
2017-12-01
Peptidoglycan is the major bacterial component recognized by the insect immune system. Peptidoglycan recognition proteins (PGRPs) are a family of pattern-recognition receptors that recognize peptidoglycans and modulate innate immune responses. Some PGRPs retain N-acetylmuramoyl-L-alanine amidase (Enzyme Commission number: 3.5.1.28) activity to hydrolyse bacterial peptidoglycans. Others have lost the enzymatic activity and work only as immune receptors. They are all important modulators for innate immunity. Here, we report the cloning and functional analysis of PGRP-S4, a short-form PGRP from the domesticated silkworm, Bombyx mori. The PGRP-S4 gene encodes a protein of 199 amino acids with a signal peptide and a PGRP domain. PGRP-S4 was expressed in the fat body, haemocytes and midgut. Its expression level was significantly induced by bacterial challenges in the midgut. The recombinant PGRP-S4 bound bacteria and different peptidoglycans. In addition, it inhibited bacterial growth and hydrolysed an Escherichia coli peptidoglycan in the presence of Zn 2+ . Scanning electron microscopy showed that PGRP-S4 disrupted the bacterial cell surface. PGRP-S4 further increased prophenoloxidase activation caused by peptidoglycans. Taken together, our data suggest that B. mori PGRP-S4 has multiple functions in immunity. © 2017 The Royal Entomological Society.
Hot, Pascal; Klein-Koerkamp, Yanica; Borg, Céline; Richard-Mornas, Aurélie; Zsoldos, Isabella; Paignon Adeline, Adeline; Thomas Antérion, Catherine; Baciu, Monica
2013-06-01
A decline in the ability to identify fearful expression has been frequently reported in patients with Alzheimer's disease (AD). In patients with severe destruction of the bilateral amygdala, similar difficulties have been reduced by using an explicit visual exploration strategy focusing on gaze. The current study assessed the possibility of applying a similar strategy in AD patients to improve fear recognition. It also assessed the possibility of improving fear recognition when a visual exploration strategy induced AD patients to process the eyes region. Seventeen patients with mild AD and 34 healthy subjects (17 young adults and 17 older adults) performed a classical task of emotional identification of faces expressing happiness, anger, and fear in two conditions: The face appeared progressively from the eyes region to the periphery (eyes region condition) or it appeared as a whole (global condition). Specific impairment in identifying a fearful expression was shown in AD patients compared with older adult controls during the global condition. Fear expression recognition was significantly improved in AD patients during the eyes region condition, in which they performed similarly to older adult controls. Our results suggest that using a different strategy of face exploration, starting first with processing of the eyes region, may compensate for a fear recognition deficit in AD patients. Findings suggest that a part of this deficit could be related to visuo-perceptual impairments. Additionally, these findings suggest that the decline of fearful face recognition reported in both normal aging and in AD may result from impairment of non-amygdalar processing in both groups and impairment of amygdalar-dependent processing in AD. Copyright © 2013 Elsevier Inc. All rights reserved.
Kobayashi, Takehito; Yagi, Yusuke; Nakamura, Takahiro
2016-01-01
The pentatricopeptide repeat (PPR) motif is a sequence-specific RNA/DNA-binding module. Elucidation of the RNA/DNA recognition mechanism has enabled engineering of PPR motifs as new RNA/DNA manipulation tools in living cells, including for genome editing. However, the biochemical characteristics of PPR proteins remain unknown, mostly due to the instability and/or unfolding propensities of PPR proteins in heterologous expression systems such as bacteria and yeast. To overcome this issue, we constructed reporter systems using animal cultured cells. The cell-based system has highly attractive features for PPR engineering: robust eukaryotic gene expression; availability of various vectors, reagents, and antibodies; highly efficient DNA delivery ratio (>80 %); and rapid, high-throughput data production. In this chapter, we introduce an example of such reporter systems: a PPR-based sequence-specific translational activation system. The cell-based reporter system can be applied to characterize plant genes of interested and to PPR engineering.
Oliver, Lindsay D; Virani, Karim; Finger, Elizabeth C; Mitchell, Derek G V
2014-07-01
Frontotemporal dementia (FTD) is a debilitating neurodegenerative disorder characterized by severely impaired social and emotional behaviour, including emotion recognition deficits. Though fear recognition impairments seen in particular neurological and developmental disorders can be ameliorated by reallocating attention to critical facial features, the possibility that similar benefits can be conferred to patients with FTD has yet to be explored. In the current study, we examined the impact of presenting distinct regions of the face (whole face, eyes-only, and eyes-removed) on the ability to recognize expressions of anger, fear, disgust, and happiness in 24 patients with FTD and 24 healthy controls. A recognition deficit was demonstrated across emotions by patients with FTD relative to controls. Crucially, removal of diagnostic facial features resulted in an appropriate decline in performance for both groups; furthermore, patients with FTD demonstrated a lack of disproportionate improvement in emotion recognition accuracy as a result of isolating critical facial features relative to controls. Thus, unlike some neurological and developmental disorders featuring amygdala dysfunction, the emotion recognition deficit observed in FTD is not likely driven by selective inattention to critical facial features. Patients with FTD also mislabelled negative facial expressions as happy more often than controls, providing further evidence for abnormalities in the representation of positive affect in FTD. This work suggests that the emotional expression recognition deficit associated with FTD is unlikely to be rectified by adjusting selective attention to diagnostic features, as has proven useful in other select disorders. Copyright © 2014 Elsevier Ltd. All rights reserved.
Kawaharada, Yasuyuki; Nielsen, Mette W.; Kelly, Simon; James, Euan K.; Andersen, Kasper R.; Rasmussen, Sheena R.; Füchtbauer, Winnie; Madsen, Lene H.; Heckmann, Anne B.; Radutoiu, Simona; Stougaard, Jens
2017-01-01
In Lotus japonicus, a LysM receptor kinase, EPR3, distinguishes compatible and incompatible rhizobial exopolysaccharides at the epidermis. However, the role of this recognition system in bacterial colonization of the root interior is unknown. Here we show that EPR3 advances the intracellular infection mechanism that mediates infection thread invasion of the root cortex and nodule primordia. At the cellular level, Epr3 expression delineates progression of infection threads into nodule primordia and cortical infection thread formation is impaired in epr3 mutants. Genetic dissection of this developmental coordination showed that Epr3 is integrated into the symbiosis signal transduction pathways. Further analysis showed differential expression of Epr3 in the epidermis and cortical primordia and identified key transcription factors controlling this tissue specificity. These results suggest that exopolysaccharide recognition is reiterated during the progressing infection and that EPR3 perception of compatible exopolysaccharide promotes an intracellular cortical infection mechanism maintaining bacteria enclosed in plant membranes. PMID:28230048
Robust kernel representation with statistical local features for face recognition.
Yang, Meng; Zhang, Lei; Shiu, Simon Chi-Keung; Zhang, David
2013-06-01
Factors such as misalignment, pose variation, and occlusion make robust face recognition a difficult problem. It is known that statistical features such as local binary pattern are effective for local feature extraction, whereas the recently proposed sparse or collaborative representation-based classification has shown interesting results in robust face recognition. In this paper, we propose a novel robust kernel representation model with statistical local features (SLF) for robust face recognition. Initially, multipartition max pooling is used to enhance the invariance of SLF to image registration error. Then, a kernel-based representation model is proposed to fully exploit the discrimination information embedded in the SLF, and robust regression is adopted to effectively handle the occlusion in face images. Extensive experiments are conducted on benchmark face databases, including extended Yale B, AR (A. Martinez and R. Benavente), multiple pose, illumination, and expression (multi-PIE), facial recognition technology (FERET), face recognition grand challenge (FRGC), and labeled faces in the wild (LFW), which have different variations of lighting, expression, pose, and occlusions, demonstrating the promising performance of the proposed method.
Han, Xu; Kim, Jung-jae; Kwoh, Chee Keong
2016-01-01
Biomedical text mining may target various kinds of valuable information embedded in the literature, but a critical obstacle to the extension of the mining targets is the cost of manual construction of labeled data, which are required for state-of-the-art supervised learning systems. Active learning is to choose the most informative documents for the supervised learning in order to reduce the amount of required manual annotations. Previous works of active learning, however, focused on the tasks of entity recognition and protein-protein interactions, but not on event extraction tasks for multiple event types. They also did not consider the evidence of event participants, which might be a clue for the presence of events in unlabeled documents. Moreover, the confidence scores of events produced by event extraction systems are not reliable for ranking documents in terms of informativity for supervised learning. We here propose a novel committee-based active learning method that supports multi-event extraction tasks and employs a new statistical method for informativity estimation instead of using the confidence scores from event extraction systems. Our method is based on a committee of two systems as follows: We first employ an event extraction system to filter potential false negatives among unlabeled documents, from which the system does not extract any event. We then develop a statistical method to rank the potential false negatives of unlabeled documents 1) by using a language model that measures the probabilities of the expression of multiple events in documents and 2) by using a named entity recognition system that locates the named entities that can be event arguments (e.g. proteins). The proposed method further deals with unknown words in test data by using word similarity measures. We also apply our active learning method for the task of named entity recognition. We evaluate the proposed method against the BioNLP Shared Tasks datasets, and show that our method can achieve better performance than such previous methods as entropy and Gibbs error based methods and a conventional committee-based method. We also show that the incorporation of named entity recognition into the active learning for event extraction and the unknown word handling further improve the active learning method. In addition, the adaptation of the active learning method into named entity recognition tasks also improves the document selection for manual annotation of named entities.
Quinn, Laura L.; Zuo, Jianmin; Abbott, Rachel J. M.; Shannon-Lowe, Claire; Tierney, Rosemary J.; Hislop, Andrew D.; Rowe, Martin
2014-01-01
CD8+ T cell responses to Epstein-Barr virus (EBV) lytic cycle expressed antigens display a hierarchy of immunodominance, in which responses to epitopes of immediate-early (IE) and some early (E) antigens are more frequently observed than responses to epitopes of late (L) expressed antigens. It has been proposed that this hierarchy, which correlates with the phase-specific efficiency of antigen presentation, may be due to the influence of viral immune-evasion genes. At least three EBV-encoded genes, BNLF2a, BGLF5 and BILF1, have the potential to inhibit processing and presentation of CD8+ T cell epitopes. Here we examined the relative contribution of these genes to modulation of CD8+ T cell recognition of EBV lytic antigens expressed at different phases of the replication cycle in EBV-transformed B-cells (LCLs) which spontaneously reactivate lytic cycle. Selective shRNA-mediated knockdown of BNLF2a expression led to more efficient recognition of immediate-early (IE)- and early (E)-derived epitopes by CD8+ T cells, while knock down of BILF1 increased recognition of epitopes from E and late (L)-expressed antigens. Contrary to what might have been predicted from previous ectopic expression studies in EBV-negative model cell lines, the shRNA-mediated inhibition of BGLF5 expression in LCLs showed only modest, if any, increase in recognition of epitopes expressed in any phase of lytic cycle. These data indicate that whilst BNLF2a interferes with antigen presentation with diminishing efficiency as lytic cycle progresses (IE>E>>L), interference by BILF1 increases with progression through lytic cycle (IE
Oil-in-Water Emulsion Exhibits Bitterness-Suppressing Effects in a Sensory Threshold Study.
Torrico, Damir Dennis; Sae-Eaw, Amporn; Sriwattana, Sujinda; Boeneke, Charles; Prinyawiwatkul, Witoon
2015-06-01
Little is known about how emulsion characteristics affect saltiness/bitterness perception. Sensory detection and recognition thresholds of NaCl, caffeine, and KCl in aqueous solution compared with oil-in-water emulsion systems were evaluated. For emulsions, NaCl, KCl, or caffeine were dissolved in water + emulsifier and mixed with canola oil (20% by weight). Two emulsions were prepared: emulsion 1 (viscosity = 257 cP) and emulsion 2 (viscosity = 59 cP). The forced-choice ascending concentration series method of limits (ASTM E-679-04) was used to determine detection and/or recognition thresholds at 25 °C. Group best estimate threshold (GBET) geometric means were expressed as g/100 mL. Comparing NaCl with KCl, there were no significant differences in detection GBET values for all systems (0.0197 - 0.0354). For saltiness recognition thresholds, KCl GBET values were higher compared with NaCl GBET (0.0822 - 0.1070 compared with 0.0471 - 0.0501). For NaCl and KCl, emulsion 1 and/or emulsion 2 did not significantly affect the saltiness recognition threshold compared with that of the aqueous solution. However, the bitterness recognition thresholds of caffeine and KCl in solution were significantly lower than in the emulsions (0.0242 - 0.0586 compared with 0.0754 - 0.1025). Gender generally had a marginal effect on threshold values. This study showed that, compared with the aqueous solutions, emulsions did not significantly affect the saltiness recognition threshold of NaCl and KCl, but exhibited bitterness-suppressing effects on KCl and/or caffeine. © 2015 Institute of Food Technologists®
Age-congruency and contact effects in body expression recognition from point-light displays (PLD)
Hermens, Frouke; Willmott, Alexander P.
2016-01-01
Recognition of older people’s body expressions is a crucial social skill. We here investigate how age, not just of the observer, but also of the observed individual, affects this skill. Age may influence the ability to recognize other people’s body expressions by changes in one’s own ability to perform certain action over the life-span (i.e., an own-age bias may occur, with best recognition for one’s own age). Whole body point light displays of children, young adults and older adults (>70 years) expressing six different emotions were presented to observers of the same three age-groups. Across two variations of the paradigm, no evidence for the predicted own-age bias (a cross-over interaction between one’s own age and the observed person’s age) was found. Instead, experience effects were found with children better recognizing older actors’ expressions of ‘active emotions,’ such as anger and happiness with greater exposure in daily life. Together, the findings suggest that age-related changes in one own’s mobility only influences body expression categorization in young children who interact frequently with older adults. PMID:27994986
Antennal transcriptome analysis of the Asian longhorned beetle Anoplophora glabripennis
Hu, Ping; Wang, Jingzhen; Cui, Mingming; Tao, Jing; Luo, Youqing
2016-01-01
Olfactory proteins form the basis of insect olfactory recognition, which is crucial for host identification, mating, and oviposition. Using transcriptome analysis of Anoplophora glabripennis antenna, we identified 42 odorant-binding proteins (OBPs), 12 chemosensory proteins (CSPs), 14 pheromone-degrading enzymes (PDEs), 1 odorant-degrading enzymes (ODE), 37 odorant receptors (ORs), 11 gustatory receptors (GRs), 2 sensory neuron membrane proteins (SNMPs), and 4 ionotropic receptor (IR). All CSPs and PBPs were expressed in antennae, confirming the authenticity of the transcriptome data. CSP expression profiles showed that AglaCSP3, AglaCSP6, and AglaCSP12 were expressed preferentially in maxillary palps and AglaCSP7 and AglaCSP9 were strongly expressed in antennae. The vast majority of CSPs were highly expressed in multiple chemosensory tissues, suggesting their participation in olfactory recognition in almost all olfactory tissues. Intriguingly, the PBP AglaPBP2 was preferentially expressed in antenna, indicating that it is the main protein involved in efficient and sensitive pheromone recognition. Phylogenetic analysis of olfactory proteins indicated AglaGR1 may detect CO2. This study establishes a foundation for determining the chemoreception molecular mechanisms of A. glabripennis, which would provide a new perspective for controlling pest populations, especially those of borers. PMID:27222053
Structural analysis of online handwritten mathematical symbols based on support vector machines
NASA Astrophysics Data System (ADS)
Simistira, Foteini; Papavassiliou, Vassilis; Katsouros, Vassilis; Carayannis, George
2013-01-01
Mathematical expression recognition is still a very challenging task for the research community mainly because of the two-dimensional (2d) structure of mathematical expressions (MEs). In this paper, we present a novel approach for the structural analysis between two on-line handwritten mathematical symbols of a ME, based on spatial features of the symbols. We introduce six features to represent the spatial affinity of the symbols and compare two multi-class classification methods that employ support vector machines (SVMs): one based on the "one-against-one" technique and one based on the "one-against-all", in identifying the relation between a pair of symbols (i.e. subscript, numerator, etc). A dataset containing 1906 spatial relations derived from the Competition on Recognition of Online Handwritten Mathematical Expressions (CROHME) 2012 training dataset is constructed to evaluate the classifiers and compare them with the rule-based classifier of the ILSP-1 system participated in the contest. The experimental results give an overall mean error rate of 2.61% for the "one-against-one" SVM approach, 6.57% for the "one-against-all" SVM technique and 12.31% error rate for the ILSP-1 classifier.
Tanimizu, Toshiyuki; Kenney, Justin W; Okano, Emiko; Kadoma, Kazune; Frankland, Paul W; Kida, Satoshi
2017-04-12
Social recognition memory is an essential and basic component of social behavior that is used to discriminate familiar and novel animals/humans. Previous studies have shown the importance of several brain regions for social recognition memories; however, the mechanisms underlying the consolidation of social recognition memory at the molecular and anatomic levels remain unknown. Here, we show a brain network necessary for the generation of social recognition memory in mice. A mouse genetic study showed that cAMP-responsive element-binding protein (CREB)-mediated transcription is required for the formation of social recognition memory. Importantly, significant inductions of the CREB target immediate-early genes c-fos and Arc were observed in the hippocampus (CA1 and CA3 regions), medial prefrontal cortex (mPFC), anterior cingulate cortex (ACC), and amygdala (basolateral region) when social recognition memory was generated. Pharmacological experiments using a microinfusion of the protein synthesis inhibitor anisomycin showed that protein synthesis in these brain regions is required for the consolidation of social recognition memory. These findings suggested that social recognition memory is consolidated through the activation of CREB-mediated gene expression in the hippocampus/mPFC/ACC/amygdala. Network analyses suggested that these four brain regions show functional connectivity with other brain regions and, more importantly, that the hippocampus functions as a hub to integrate brain networks and generate social recognition memory, whereas the ACC and amygdala are important for coordinating brain activity when social interaction is initiated by connecting with other brain regions. We have found that a brain network composed of the hippocampus/mPFC/ACC/amygdala is required for the consolidation of social recognition memory. SIGNIFICANCE STATEMENT Here, we identify brain networks composed of multiple brain regions for the consolidation of social recognition memory. We found that social recognition memory is consolidated through CREB-meditated gene expression in the hippocampus, medial prefrontal cortex, anterior cingulate cortex (ACC), and amygdala. Importantly, network analyses based on c-fos expression suggest that functional connectivity of these four brain regions with other brain regions is increased with time spent in social investigation toward the generation of brain networks to consolidate social recognition memory. Furthermore, our findings suggest that hippocampus functions as a hub to integrate brain networks and generate social recognition memory, whereas ACC and amygdala are important for coordinating brain activity when social interaction is initiated by connecting with other brain regions. Copyright © 2017 the authors 0270-6474/17/374103-14$15.00/0.
A Neural Basis of Facial Action Recognition in Humans
Srinivasan, Ramprakash; Golomb, Julie D.
2016-01-01
By combining different facial muscle actions, called action units, humans can produce an extraordinarily large number of facial expressions. Computational models and studies in cognitive science and social psychology have long hypothesized that the brain needs to visually interpret these action units to understand other people's actions and intentions. Surprisingly, no studies have identified the neural basis of the visual recognition of these action units. Here, using functional magnetic resonance imaging and an innovative machine learning analysis approach, we identify a consistent and differential coding of action units in the brain. Crucially, in a brain region thought to be responsible for the processing of changeable aspects of the face, multivoxel pattern analysis could decode the presence of specific action units in an image. This coding was found to be consistent across people, facilitating the estimation of the perceived action units on participants not used to train the multivoxel decoder. Furthermore, this coding of action units was identified when participants attended to the emotion category of the facial expression, suggesting an interaction between the visual analysis of action units and emotion categorization as predicted by the computational models mentioned above. These results provide the first evidence for a representation of action units in the brain and suggest a mechanism for the analysis of large numbers of facial actions and a loss of this capacity in psychopathologies. SIGNIFICANCE STATEMENT Computational models and studies in cognitive and social psychology propound that visual recognition of facial expressions requires an intermediate step to identify visible facial changes caused by the movement of specific facial muscles. Because facial expressions are indeed created by moving one's facial muscles, it is logical to assume that our visual system solves this inverse problem. Here, using an innovative machine learning method and neuroimaging data, we identify for the first time a brain region responsible for the recognition of actions associated with specific facial muscles. Furthermore, this representation is preserved across subjects. Our machine learning analysis does not require mapping the data to a standard brain and may serve as an alternative to hyperalignment. PMID:27098688
Gillespie, Steven M.; Rotshtein, Pia; Satherley, Rose-Marie; Beech, Anthony R.; Mitchell, Ian J.
2015-01-01
Research with violent offenders has consistently shown impaired recognition of other’s facial expressions of emotion. However, the extent to which similar problems can be observed among sexual offenders remains unknown. Using a computerized task, we presented sexual and violent offenders, and non-offenders, with male and female expressions of anger, disgust, fear, happiness, sadness, and surprise, morphed with neutral expressions at varying levels of intensity (10, 55, and 90% expressive). Based on signal detection theory, we used hit rates and false alarms to calculate the sensitivity index d-prime (d′) and criterion (c) for each emotional expression. Overall, sexual offenders showed reduced sensitivity to emotional expressions across intensity, sex, and type of expression, compared with non-offenders, while both sexual and violent offenders showed particular reduced sensitivity to fearful expressions. We also observed specific effects for high (90%) intensity female faces, with sexual offenders showing reduced sensitivity to anger compared with non-offenders and violent offenders, and reduced sensitivity to disgust compared with non-offenders. Furthermore, both sexual and violent offenders showed impaired sensitivity to high intensity female fearful expressions compared with non-offenders. Violent offenders also showed a higher criterion for classifying moderate and high intensity male expressions as fearful, indicative of a more conservative response style, compared with angry, happy, or sad. These results suggest that both types of offender show problems in emotion recognition, and may have implications for understanding the inhibition of violent and sexually violent behaviors. PMID:26029137
Gillespie, Steven M; Rotshtein, Pia; Satherley, Rose-Marie; Beech, Anthony R; Mitchell, Ian J
2015-01-01
Research with violent offenders has consistently shown impaired recognition of other's facial expressions of emotion. However, the extent to which similar problems can be observed among sexual offenders remains unknown. Using a computerized task, we presented sexual and violent offenders, and non-offenders, with male and female expressions of anger, disgust, fear, happiness, sadness, and surprise, morphed with neutral expressions at varying levels of intensity (10, 55, and 90% expressive). Based on signal detection theory, we used hit rates and false alarms to calculate the sensitivity index d-prime (d') and criterion (c) for each emotional expression. Overall, sexual offenders showed reduced sensitivity to emotional expressions across intensity, sex, and type of expression, compared with non-offenders, while both sexual and violent offenders showed particular reduced sensitivity to fearful expressions. We also observed specific effects for high (90%) intensity female faces, with sexual offenders showing reduced sensitivity to anger compared with non-offenders and violent offenders, and reduced sensitivity to disgust compared with non-offenders. Furthermore, both sexual and violent offenders showed impaired sensitivity to high intensity female fearful expressions compared with non-offenders. Violent offenders also showed a higher criterion for classifying moderate and high intensity male expressions as fearful, indicative of a more conservative response style, compared with angry, happy, or sad. These results suggest that both types of offender show problems in emotion recognition, and may have implications for understanding the inhibition of violent and sexually violent behaviors.
Audio-visual affective expression recognition
NASA Astrophysics Data System (ADS)
Huang, Thomas S.; Zeng, Zhihong
2007-11-01
Automatic affective expression recognition has attracted more and more attention of researchers from different disciplines, which will significantly contribute to a new paradigm for human computer interaction (affect-sensitive interfaces, socially intelligent environments) and advance the research in the affect-related fields including psychology, psychiatry, and education. Multimodal information integration is a process that enables human to assess affective states robustly and flexibly. In order to understand the richness and subtleness of human emotion behavior, the computer should be able to integrate information from multiple sensors. We introduce in this paper our efforts toward machine understanding of audio-visual affective behavior, based on both deliberate and spontaneous displays. Some promising methods are presented to integrate information from both audio and visual modalities. Our experiments show the advantage of audio-visual fusion in affective expression recognition over audio-only or visual-only approaches.
Inuzuka, Saki; Kakizawa, Hitoshi; Nishimura, Kei-Ichiro; Naito, Takuto; Miyazaki, Katsushi; Furuta, Hiroyuki; Matsumura, Shigeyoshi; Ikawa, Yoshiya
2018-06-01
The riboswitch is a class of RNA-based gene regulatory machinery that is dependent on recognition of its target ligand by RNA tertiary structures. Ligand recognition is achieved by the aptamer domain, and ligand-dependent structural changes of the expression platform then usually mediate termination of transcription or translational initiation. Ligand-dependent structural changes of the aptamer domain and expression platform have been reported for several riboswitches with short (<40 nucleotides) expression platforms. In this study, we characterized structural changes of the Vc2 c-di-GMP riboswitch that represses translation of downstream open reading frames in a ligand-dependent manner. The Vc2 riboswitch has a long (97 nucleotides) expression platform, but its structure and function are largely unknown. Through mutational analysis and chemical probing, we identified its secondary structures that are possibly responsible for switch-OFF and switch-ON states of translational initiation. © 2018 Molecular Biology Society of Japan and John Wiley & Sons Australia, Ltd.
Stewart, Suzanne L K; Schepman, Astrid; Haigh, Matthew; McHugh, Rhian; Stewart, Andrew J
2018-03-14
The recognition of emotional facial expressions is often subject to contextual influence, particularly when the face and the context convey similar emotions. We investigated whether spontaneous, incidental affective theory of mind inferences made while reading vignettes describing social situations would produce context effects on the identification of same-valenced emotions (Experiment 1) as well as differently-valenced emotions (Experiment 2) conveyed by subsequently presented faces. Crucially, we found an effect of context on reaction times in both experiments while, in line with previous work, we found evidence for a context effect on accuracy only in Experiment 1. This demonstrates that affective theory of mind inferences made at the pragmatic level of a text can automatically, contextually influence the perceptual processing of emotional facial expressions in a separate task even when those emotions are of a distinctive valence. Thus, our novel findings suggest that language acts as a contextual influence to the recognition of emotional facial expressions for both same and different valences.
Rodrigo-Ruiz, D; Perez-Gonzalez, J C; Cejudo, J
2017-08-16
It has recently been warned that children with attention deficit hyperactivity disorder (ADHD) show a deficit in emotional competence and emotional intelligence, specifically in their ability to emotional recognition. A systematic review of the scientific literature in reference to the emotional recognition of facial expressions in children with ADHD is presented in order to establish or rule the existence of emotional deficits as primary dysfunction in this disorder and, where appropriate, the effect size of the differences against normal development or neurotypical children. The results reveal the recent interest in the issue and the lack of information. Although there is no complete agreement, most of the studies show that emotional recognition of facial expressions is affected in children with ADHD, showing them significantly less accurate than children from control groups in recognizing emotions communicated through facial expressions. A part of these studies make comparisons on the recognition of different discrete emotions; having observed that children with ADHD tend to a greater difficulty recognizing negative emotions, especially anger, fear, and disgust. These results have direct implications for the educational and clinical diagnosis of ADHD; and for the educational intervention for children with ADHD, emotional education might entail an advantageous aid.
Gender differences in facial emotion recognition in persons with chronic schizophrenia.
Weiss, Elisabeth M; Kohler, Christian G; Brensinger, Colleen M; Bilker, Warren B; Loughead, James; Delazer, Margarete; Nolan, Karen A
2007-03-01
The aim of the present study was to investigate possible sex differences in the recognition of facial expressions of emotion and to investigate the pattern of classification errors in schizophrenic males and females. Such an approach provides an opportunity to inspect the degree to which males and females differ in perceiving and interpreting the different emotions displayed to them and to analyze which emotions are most susceptible to recognition errors. Fifty six chronically hospitalized schizophrenic patients (38 men and 18 women) completed the Penn Emotion Recognition Test (ER40), a computerized emotion discrimination test presenting 40 color photographs of evoked happy, sad, anger, fear expressions and neutral expressions balanced for poser gender and ethnicity. We found a significant sex difference in the patterns of error rates in the Penn Emotion Recognition Test. Neutral faces were more commonly mistaken as angry in schizophrenic men, whereas schizophrenic women misinterpreted neutral faces more frequently as sad. Moreover, female faces were better recognized overall, but fear was better recognized in same gender photographs, whereas anger was better recognized in different gender photographs. The findings of the present study lend support to the notion that sex differences in aggressive behavior could be related to a cognitive style characterized by hostile attributions to neutral faces in schizophrenic men.
Does Facial Amimia Impact the Recognition of Facial Emotions? An EMG Study in Parkinson’s Disease
Argaud, Soizic; Delplanque, Sylvain; Houvenaghel, Jean-François; Auffret, Manon; Duprez, Joan; Vérin, Marc; Grandjean, Didier; Sauleau, Paul
2016-01-01
According to embodied simulation theory, understanding other people’s emotions is fostered by facial mimicry. However, studies assessing the effect of facial mimicry on the recognition of emotion are still controversial. In Parkinson’s disease (PD), one of the most distinctive clinical features is facial amimia, a reduction in facial expressiveness, but patients also show emotional disturbances. The present study used the pathological model of PD to examine the role of facial mimicry on emotion recognition by investigating EMG responses in PD patients during a facial emotion recognition task (anger, joy, neutral). Our results evidenced a significant decrease in facial mimicry for joy in PD, essentially linked to the absence of reaction of the zygomaticus major and the orbicularis oculi muscles in response to happy avatars, whereas facial mimicry for expressions of anger was relatively preserved. We also confirmed that PD patients were less accurate in recognizing positive and neutral facial expressions and highlighted a beneficial effect of facial mimicry on the recognition of emotion. We thus provide additional arguments for embodied simulation theory suggesting that facial mimicry is a potential lever for therapeutic actions in PD even if it seems not to be necessarily required in recognizing emotion as such. PMID:27467393
Astrocytes contribute to gamma oscillations and recognition memory.
Lee, Hosuk Sean; Ghetti, Andrea; Pinto-Duarte, António; Wang, Xin; Dziewczapolski, Gustavo; Galimi, Francesco; Huitron-Resendiz, Salvador; Piña-Crespo, Juan C; Roberts, Amanda J; Verma, Inder M; Sejnowski, Terrence J; Heinemann, Stephen F
2014-08-12
Glial cells are an integral part of functional communication in the brain. Here we show that astrocytes contribute to the fast dynamics of neural circuits that underlie normal cognitive behaviors. In particular, we found that the selective expression of tetanus neurotoxin (TeNT) in astrocytes significantly reduced the duration of carbachol-induced gamma oscillations in hippocampal slices. These data prompted us to develop a novel transgenic mouse model, specifically with inducible tetanus toxin expression in astrocytes. In this in vivo model, we found evidence of a marked decrease in electroencephalographic (EEG) power in the gamma frequency range in awake-behaving mice, whereas neuronal synaptic activity remained intact. The reduction in cortical gamma oscillations was accompanied by impaired behavioral performance in the novel object recognition test, whereas other forms of memory, including working memory and fear conditioning, remained unchanged. These results support a key role for gamma oscillations in recognition memory. Both EEG alterations and behavioral deficits in novel object recognition were reversed by suppression of tetanus toxin expression. These data reveal an unexpected role for astrocytes as essential contributors to information processing and cognitive behavior.
Ng, Wy Ching; Londrigan, Sarah L.; Nasr, Najla; Cunningham, Anthony L.; Turville, Stuart; Brooks, Andrew G.
2015-01-01
ABSTRACT It is well established that influenza A virus (IAV) attachment to and infection of epithelial cells is dependent on sialic acid (SIA) at the cell surface, although the specific receptors that mediate IAV entry have not been defined and multiple receptors may exist. Lec2 Chinese hamster ovary (CHO) cells are SIA deficient and resistant to IAV infection. Here we demonstrate that the expression of the C-type lectin receptor langerin in Lec2 cells (Lec2-Lg) rendered them permissive to IAV infection, as measured by replication of the viral genome, transcription of viral mRNA, and synthesis of viral proteins. Unlike SIA-dependent infection of parental CHO cells, IAV attachment and infection of Lec2-Lg cells was mediated via lectin-mediated recognition of mannose-rich glycans expressed by the viral hemagglutinin glycoprotein. Lec2 cells expressing endocytosis-defective langerin bound IAV efficiently but remained resistant to IAV infection, confirming that internalization via langerin was essential for infectious entry. Langerin-mediated infection of Lec2-Lg cells was pH and dynamin dependent, occurred via clathrin- and caveolin-mediated endocytic pathways, and utilized early (Rab5+) but not late (Rab7+) endosomes. This study is the first to demonstrate that langerin represents an authentic receptor that binds and internalizes IAV to facilitate infection. Moreover, it describes a unique experimental system to probe specific pathways and compartments involved in infectious entry following recognition of IAV by a single cell surface receptor. IMPORTANCE On the surface of host cells, sialic acid (SIA) functions as the major attachment factor for influenza A viruses (IAV). However, few studies have identified specific transmembrane receptors that bind and internalize IAV to facilitate infection. Here we identify human langerin as a transmembrane glycoprotein that can act as an attachment factor and a bone fide endocytic receptor for IAV infection. Expression of langerin by an SIA-deficient cell line resistant to IAV rendered cells permissive to infection. As langerin represented the sole receptor for IAV infection in this system, we have defined the pathways and compartments involved in infectious entry of IAV into cells following recognition by langerin. PMID:26468543
Evers, Kris; Kerkhof, Inneke; Steyaert, Jean; Noens, Ilse; Wagemans, Johan
2014-01-01
Emotion recognition problems are frequently reported in individuals with an autism spectrum disorder (ASD). However, this research area is characterized by inconsistent findings, with atypical emotion processing strategies possibly contributing to existing contradictions. In addition, an attenuated saliency of the eyes region is often demonstrated in ASD during face identity processing. We wanted to compare reliance on mouth versus eyes information in children with and without ASD, using hybrid facial expressions. A group of six-to-eight-year-old boys with ASD and an age- and intelligence-matched typically developing (TD) group without intellectual disability performed an emotion labelling task with hybrid facial expressions. Five static expressions were used: one neutral expression and four emotional expressions, namely, anger, fear, happiness, and sadness. Hybrid faces were created, consisting of an emotional face half (upper or lower face region) with the other face half showing a neutral expression. Results showed no emotion recognition problem in ASD. Moreover, we provided evidence for the existence of top- and bottom-emotions in children: correct identification of expressions mainly depends on information in the eyes (so-called top-emotions: happiness) or in the mouth region (so-called bottom-emotions: sadness, anger, and fear). No stronger reliance on mouth information was found in children with ASD.
Heterogeneity of long-history migration predicts emotion recognition accuracy.
Wood, Adrienne; Rychlowska, Magdalena; Niedenthal, Paula M
2016-06-01
Recent work (Rychlowska et al., 2015) demonstrated the power of a relatively new cultural dimension, historical heterogeneity, in predicting cultural differences in the endorsement of emotion expression norms. Historical heterogeneity describes the number of source countries that have contributed to a country's present-day population over the last 500 years. People in cultures originating from a large number of source countries may have historically benefited from greater and clearer emotional expressivity, because they lacked a common language and well-established social norms. We therefore hypothesized that in addition to endorsing more expressive display rules, individuals from heterogeneous cultures will also produce facial expressions that are easier to recognize by people from other cultures. By reanalyzing cross-cultural emotion recognition data from 92 papers and 82 cultures, we show that emotion expressions of people from heterogeneous cultures are more easily recognized by observers from other cultures than are the expressions produced in homogeneous cultures. Heterogeneity influences expression recognition rates alongside the individualism-collectivism of the perceivers' culture, as more individualistic cultures were more accurate in emotion judgments than collectivistic cultures. This work reveals the present-day behavioral consequences of long-term historical migration patterns and demonstrates the predictive power of historical heterogeneity. (PsycINFO Database Record (c) 2016 APA, all rights reserved).
Mutual information-based facial expression recognition
NASA Astrophysics Data System (ADS)
Hazar, Mliki; Hammami, Mohamed; Hanêne, Ben-Abdallah
2013-12-01
This paper introduces a novel low-computation discriminative regions representation for expression analysis task. The proposed approach relies on interesting studies in psychology which show that most of the descriptive and responsible regions for facial expression are located around some face parts. The contributions of this work lie in the proposition of new approach which supports automatic facial expression recognition based on automatic regions selection. The regions selection step aims to select the descriptive regions responsible or facial expression and was performed using Mutual Information (MI) technique. For facial feature extraction, we have applied Local Binary Patterns Pattern (LBP) on Gradient image to encode salient micro-patterns of facial expressions. Experimental studies have shown that using discriminative regions provide better results than using the whole face regions whilst reducing features vector dimension.
Impaired holistic coding of facial expression and facial identity in congenital prosopagnosia.
Palermo, Romina; Willis, Megan L; Rivolta, Davide; McKone, Elinor; Wilson, C Ellie; Calder, Andrew J
2011-04-01
We test 12 individuals with congenital prosopagnosia (CP), who replicate a common pattern of showing severe difficulty in recognising facial identity in conjunction with normal recognition of facial expressions (both basic and 'social'). Strength of holistic processing was examined using standard expression composite and identity composite tasks. Compared to age- and sex-matched controls, group analyses demonstrated that CPs showed weaker holistic processing, for both expression and identity information. Implications are (a) normal expression recognition in CP can derive from compensatory strategies (e.g., over-reliance on non-holistic cues to expression); (b) the split between processing of expression and identity information may take place after a common stage of holistic processing; and (c) contrary to a recent claim, holistic processing of identity is functionally involved in face identification ability. Copyright © 2011 Elsevier Ltd. All rights reserved.
Impaired holistic coding of facial expression and facial identity in congenital prosopagnosia
Palermo, Romina; Willis, Megan L.; Rivolta, Davide; McKone, Elinor; Wilson, C. Ellie; Calder, Andrew J.
2011-01-01
We test 12 individuals with congenital prosopagnosia (CP), who replicate a common pattern of showing severe difficulty in recognising facial identity in conjunction with normal recognition of facial expressions (both basic and ‘social’). Strength of holistic processing was examined using standard expression composite and identity composite tasks. Compared to age- and sex-matched controls, group analyses demonstrated that CPs showed weaker holistic processing, for both expression and identity information. Implications are (a) normal expression recognition in CP can derive from compensatory strategies (e.g., over-reliance on non-holistic cues to expression); (b) the split between processing of expression and identity information may take place after a common stage of holistic processing; and (c) contrary to a recent claim, holistic processing of identity is functionally involved in face identification ability. PMID:21333662
Katagiri, Fumiaki; Glazebrook, Jane
2003-01-01
A major task in computational analysis of mRNA expression profiles is definition of relationships among profiles on the basis of similarities among them. This is generally achieved by pattern recognition in the distribution of data points representing each profile in a high-dimensional space. Some drawbacks of commonly used pattern recognition algorithms stem from their use of a globally linear space and/or limited degrees of freedom. A pattern recognition method called Local Context Finder (LCF) is described here. LCF uses nonlinear dimensionality reduction for pattern recognition. Then it builds a network of profiles based on the nonlinear dimensionality reduction results. LCF was used to analyze mRNA expression profiles of the plant host Arabidopsis interacting with the bacterial pathogen Pseudomonas syringae. In one case, LCF revealed two dimensions essential to explain the effects of the NahG transgene and the ndr1 mutation on resistant and susceptible responses. In another case, plant mutants deficient in responses to pathogen infection were classified on the basis of LCF analysis of their profiles. The classification by LCF was consistent with the results of biological characterization of the mutants. Thus, LCF is a powerful method for extracting information from expression profile data. PMID:12960373
Decoding facial expressions based on face-selective and motion-sensitive areas.
Liang, Yin; Liu, Baolin; Xu, Junhai; Zhang, Gaoyan; Li, Xianglin; Wang, Peiyuan; Wang, Bin
2017-06-01
Humans can easily recognize others' facial expressions. Among the brain substrates that enable this ability, considerable attention has been paid to face-selective areas; in contrast, whether motion-sensitive areas, which clearly exhibit sensitivity to facial movements, are involved in facial expression recognition remained unclear. The present functional magnetic resonance imaging (fMRI) study used multi-voxel pattern analysis (MVPA) to explore facial expression decoding in both face-selective and motion-sensitive areas. In a block design experiment, participants viewed facial expressions of six basic emotions (anger, disgust, fear, joy, sadness, and surprise) in images, videos, and eyes-obscured videos. Due to the use of multiple stimulus types, the impacts of facial motion and eye-related information on facial expression decoding were also examined. It was found that motion-sensitive areas showed significant responses to emotional expressions and that dynamic expressions could be successfully decoded in both face-selective and motion-sensitive areas. Compared with static stimuli, dynamic expressions elicited consistently higher neural responses and decoding performance in all regions. A significant decrease in both activation and decoding accuracy due to the absence of eye-related information was also observed. Overall, the findings showed that emotional expressions are represented in motion-sensitive areas in addition to conventional face-selective areas, suggesting that motion-sensitive regions may also effectively contribute to facial expression recognition. The results also suggested that facial motion and eye-related information played important roles by carrying considerable expression information that could facilitate facial expression recognition. Hum Brain Mapp 38:3113-3125, 2017. © 2017 Wiley Periodicals, Inc. © 2017 Wiley Periodicals, Inc.
Identifying significant environmental features using feature recognition.
DOT National Transportation Integrated Search
2015-10-01
The Department of Environmental Analysis at the Kentucky Transportation Cabinet has expressed an interest in feature-recognition capability because it may help analysts identify environmentally sensitive features in the landscape, : including those r...
Kamboj, Sunjeev K; Joye, Alyssa; Bisby, James A; Das, Ravi K; Platt, Bradley; Curran, H Valerie
2013-05-01
Studies of affect recognition can inform our understanding of the interpersonal effects of alcohol and help develop a more complete neuropsychological profile of this drug. The objective of the study was to examine affect recognition in social drinkers using a novel dynamic affect-recognition task, sampling performance across a range of evolutionarily significant target emotions and neutral expressions. Participants received 0, 0.4 or 0.8 g/kg alcohol in a double-blind, independent groups design. Relatively naturalistic changes in facial expression-from neutral (mouth open) to increasing intensities of target emotions, as well as neutral (mouth closed)-were simulated using computer-generated dynamic morphs. Accuracy and reaction time were measured and a two-high-threshold model applied to hits and false-alarm data to determine sensitivity and response bias. While there was no effect on the principal emotion expressions (happiness, sadness, fear, anger and disgust), compared to those receiving 0.8 g/kg of alcohol and placebo, participants administered with 0.4 g/kg alcohol tended to show an enhanced response bias to neutral expressions. Exploration of this effect suggested an accompanying tendency to misattribute neutrality to sad expressions following the 0.4-g/kg dose. The 0.4-g/kg alcohol-but not 0.8 g/kg-produced a limited and specific modification in affect recognition evidenced by a neutral response bias and possibly an accompanying tendency to misclassify sad expressions as neutral. In light of previous findings on involuntary negative memory following the 0.4-g/kg dose, we suggest that moderate-but not high-doses of alcohol have a special relevance to emotional processing in social drinkers.
Albre, Jérôme; Liénard, Marjorie A.; Sirey, Tamara M.; Schmidt, Silvia; Tooman, Leah K.; Carraher, Colm; Greenwood, David R.; Löfstedt, Christer; Newcomb, Richard D.
2012-01-01
Chemical signals are prevalent in sexual communication systems. Mate recognition has been extensively studied within the Lepidoptera, where the production and recognition of species-specific sex pheromone signals are typically the defining character. While the specific blend of compounds that makes up the sex pheromones of many species has been characterized, the molecular mechanisms underpinning the evolution of pheromone-based mate recognition systems remain largely unknown. We have focused on two sets of sibling species within the leafroller moth genera Ctenopseustis and Planotortrix that have rapidly evolved the use of distinct sex pheromone blends. The compounds within these blends differ almost exclusively in the relative position of double bonds that are introduced by desaturase enzymes. Of the six desaturase orthologs isolated from all four species, functional analyses in yeast and gene expression in pheromone glands implicate three in pheromone biosynthesis, two Δ9-desaturases, and a Δ10-desaturase, while the remaining three desaturases include a Δ6-desaturase, a terminal desaturase, and a non-functional desaturase. Comparative quantitative real-time PCR reveals that the Δ10-desaturase is differentially expressed in the pheromone glands of the two sets of sibling species, consistent with differences in the pheromone blend in both species pairs. In the pheromone glands of species that utilize (Z)-8-tetradecenyl acetate as sex pheromone component (Ctenopseustis obliquana and Planotortrix octo), the expression levels of the Δ10-desaturase are significantly higher than in the pheromone glands of their respective sibling species (C. herana and P. excessana). Our results demonstrate that interspecific sex pheromone differences are associated with differential regulation of the same desaturase gene in two genera of moths. We suggest that differential gene regulation among members of a multigene family may be an important mechanism of molecular innovation in sex pheromone evolution and speciation. PMID:22291612
Zalcman, Gisela; Federman, Noel; de la Fuente, Verónica; Romano, Arturo
2015-03-01
Long-term memory formation requires gene expression after acquisition of new information. The first step in the regulation of gene expression is the participation of transcription factors (TFs) such as nuclear factor kappa B (NF-кB), which are present before the neuronal activity induced by training. It was proposed that the activation of these types of TFs allows a second step in gene regulation by induction of immediate-early genes (IEGs) whose protein products are, in turn, TFs. Between these IEGs, zif268 has been found to play a critical role in long-term memory formation and reprocessing after retrieval. Here we found in mice hippocampus that, on one hand, NF-кB was activated 45 min after training in a novel object recognition (NOR) task and that inhibiting NF-кB immediately after training by intrahippocampal administration of NF-кB Decoy DNA impaired NOR memory consolidation. On the other hand, Zif268 protein expression was induced 45 min after NOR training and the administration of DNA antisense to its mRNA post-training impaired recognition memory. Finally, we found that the inhibition of NF-кB by NF-кB Decoy DNA reduced significantly the training-induced Zif268 increment, indicating that NF-кB is involved in the regulation of Zif268 expression. Thus, the present results support the involvement of NF-кB activity-dependent Zif268 expression in the hippocampus during recognition memory consolidation. Copyright © 2015 Elsevier Inc. All rights reserved.
Chiranjeevi, Pojala; Gopalakrishnan, Viswanath; Moogi, Pratibha
2015-09-01
Facial expression recognition is one of the open problems in computer vision. Robust neutral face recognition in real time is a major challenge for various supervised learning-based facial expression recognition methods. This is due to the fact that supervised methods cannot accommodate all appearance variability across the faces with respect to race, pose, lighting, facial biases, and so on, in the limited amount of training data. Moreover, processing each and every frame to classify emotions is not required, as user stays neutral for majority of the time in usual applications like video chat or photo album/web browsing. Detecting neutral state at an early stage, thereby bypassing those frames from emotion classification would save the computational power. In this paper, we propose a light-weight neutral versus emotion classification engine, which acts as a pre-processer to the traditional supervised emotion classification approaches. It dynamically learns neutral appearance at key emotion (KE) points using a statistical texture model, constructed by a set of reference neutral frames for each user. The proposed method is made robust to various types of user head motions by accounting for affine distortions based on a statistical texture model. Robustness to dynamic shift of KE points is achieved by evaluating the similarities on a subset of neighborhood patches around each KE point using the prior information regarding the directionality of specific facial action units acting on the respective KE point. The proposed method, as a result, improves emotion recognition (ER) accuracy and simultaneously reduces computational complexity of the ER system, as validated on multiple databases.
Safe Gene Therapy for Type 1 Diabetes
2010-10-01
Baltimore, Maryland, USA. Skeletal muscle development, nutrient uptake, and nutrient utilization is largely coordinated by growth hor- mone (GH) and its...one of the many means the immune system utilizes to ensure self-recognition and maintenance of tolerance. In central lymphoid organ, the thymus, PTA...largely present in the marginal zone bordering the white pulp. Such geometrical location are consistent with a tolerogenic role of Aire-expressing cells
Phage-mediated Delivery of Targeted sRNA Constructs to Knock Down Gene Expression in E. coli.
Bernheim, Aude G; Libis, Vincent K; Lindner, Ariel B; Wintermute, Edwin H
2016-03-20
RNA-mediated knockdowns are widely used to control gene expression. This versatile family of techniques makes use of short RNA (sRNA) that can be synthesized with any sequence and designed to complement any gene targeted for silencing. Because sRNA constructs can be introduced to many cell types directly or using a variety of vectors, gene expression can be repressed in living cells without laborious genetic modification. The most common RNA knockdown technology, RNA interference (RNAi), makes use of the endogenous RNA-induced silencing complex (RISC) to mediate sequence recognition and cleavage of the target mRNA. Applications of this technique are therefore limited to RISC-expressing organisms, primarily eukaryotes. Recently, a new generation of RNA biotechnologists have developed alternative mechanisms for controlling gene expression through RNA, and so made possible RNA-mediated gene knockdowns in bacteria. Here we describe a method for silencing gene expression in E. coli that functionally resembles RNAi. In this system a synthetic phagemid is designed to express sRNA, which may designed to target any sequence. The expression construct is delivered to a population of E. coli cells with non-lytic M13 phage, after which it is able to stably replicate as a plasmid. Antisense recognition and silencing of the target mRNA is mediated by the Hfq protein, endogenous to E. coli. This protocol includes methods for designing the antisense sRNA, constructing the phagemid vector, packaging the phagemid into M13 bacteriophage, preparing a live cell population for infection, and performing the infection itself. The fluorescent protein mKate2 and the antibiotic resistance gene chloramphenicol acetyltransferase (CAT) are targeted to generate representative data and to quantify knockdown effectiveness.
Updating schematic emotional facial expressions in working memory: Response bias and sensitivity.
Tamm, Gerly; Kreegipuu, Kairi; Harro, Jaanus; Cowan, Nelson
2017-01-01
It is unclear if positive, negative, or neutral emotional expressions have an advantage in short-term recognition. Moreover, it is unclear from previous studies of working memory for emotional faces whether effects of emotions comprise response bias or sensitivity. The aim of this study was to compare how schematic emotional expressions (sad, angry, scheming, happy, and neutral) are discriminated and recognized in an updating task (2-back recognition) in a representative sample of birth cohort of young adults. Schematic facial expressions allow control of identity processing, which is separate from expression processing, and have been used extensively in attention research but not much, until now, in working memory research. We found that expressions with a U-curved mouth (i.e., upwardly curved), namely happy and scheming expressions, favoured a bias towards recognition (i.e., towards indicating that the probe and the stimulus in working memory are the same). Other effects of emotional expression were considerably smaller (1-2% of the variance explained)) compared to a large proportion of variance that was explained by the physical similarity of items being compared. We suggest that the nature of the stimuli plays a role in this. The present application of signal detection methodology with emotional, schematic faces in a working memory procedure requiring fast comparisons helps to resolve important contradictions that have emerged in the emotional perception literature. Copyright © 2016 Elsevier B.V. All rights reserved.
Research of Daily Conversation Transmitting System Based on Mouth Part Pattern Recognition
NASA Astrophysics Data System (ADS)
Watanabe, Mutsumi; Nishi, Natsuko
The authors are developing a vision-based intension transfer technique by recognizing user’s face expressions and movements, to help free and convenient communications with aged or disabled persons who find difficulties in talking, discriminating small character prints and operating keyboards by hands and fingers. In this paper we report a prototype system, where layered daily conversations are successively selected by recognizing the transition in shape of user’s mouth parts using camera image sequences settled in front of the user. Four mouth part patterns are used in the system. A method that automatically recognizes these patterns by analyzing the intensity histogram data around the mouth region is newly developed. The confirmation of a selection on the way is executed by detecting the open and shut movements of mouth through the temporal change in intensity histogram data. The method has been installed in a desktop PC by VC++ programs. Experimental results of mouth shape pattern recognition by twenty-five persons have shown the effectiveness of the method.
Does cortisol modulate emotion recognition and empathy?
Duesenberg, Moritz; Weber, Juliane; Schulze, Lars; Schaeuffele, Carmen; Roepke, Stefan; Hellmann-Regen, Julian; Otte, Christian; Wingenfeld, Katja
2016-04-01
Emotion recognition and empathy are important aspects in the interaction and understanding of other people's behaviors and feelings. The Human environment comprises of stressful situations that impact social interactions on a daily basis. Aim of the study was to examine the effects of the stress hormone cortisol on emotion recognition and empathy. In this placebo-controlled study, 40 healthy men and 40 healthy women (mean age 24.5 years) received either 10mg of hydrocortisone or placebo. We used the Multifaceted Empathy Test to measure emotional and cognitive empathy. Furthermore, we examined emotion recognition from facial expressions, which contained two emotions (anger and sadness) and two emotion intensities (40% and 80%). We did not find a main effect for treatment or sex on either empathy or emotion recognition but a sex × emotion interaction on emotion recognition. The main result was a four-way-interaction on emotion recognition including treatment, sex, emotion and task difficulty. At 40% task difficulty, women recognized angry faces better than men in the placebo condition. Furthermore, in the placebo condition, men recognized sadness better than anger. At 80% task difficulty, men and women performed equally well in recognizing sad faces but men performed worse compared to women with regard to angry faces. Apparently, our results did not support the hypothesis that increases in cortisol concentration alone influence empathy and emotion recognition in healthy young individuals. However, sex and task difficulty appear to be important variables in emotion recognition from facial expressions. Copyright © 2016 Elsevier Ltd. All rights reserved.
Recognition of DHN-melanin by a C-type lectin receptor is required for immunity to Aspergillus.
Stappers, Mark H T; Clark, Alexandra E; Aimanianda, Vishukumar; Bidula, Stefan; Reid, Delyth M; Asamaphan, Patawee; Hardison, Sarah E; Dambuza, Ivy M; Valsecchi, Isabel; Kerscher, Bernhard; Plato, Anthony; Wallace, Carol A; Yuecel, Raif; Hebecker, Betty; da Glória Teixeira Sousa, Maria; Cunha, Cristina; Liu, Yan; Feizi, Ten; Brakhage, Axel A; Kwon-Chung, Kyung J; Gow, Neil A R; Zanda, Matteo; Piras, Monica; Zanato, Chiara; Jaeger, Martin; Netea, Mihai G; van de Veerdonk, Frank L; Lacerda, João F; Campos, António; Carvalho, Agostinho; Willment, Janet A; Latgé, Jean-Paul; Brown, Gordon D
2018-03-15
Resistance to infection is critically dependent on the ability of pattern recognition receptors to recognize microbial invasion and induce protective immune responses. One such family of receptors are the C-type lectins, which are central to antifungal immunity. These receptors activate key effector mechanisms upon recognition of conserved fungal cell-wall carbohydrates. However, several other immunologically active fungal ligands have been described; these include melanin, for which the mechanism of recognition is hitherto undefined. Here we identify a C-type lectin receptor, melanin-sensing C-type lectin receptor (MelLec), that has an essential role in antifungal immunity through recognition of the naphthalene-diol unit of 1,8-dihydroxynaphthalene (DHN)-melanin. MelLec recognizes melanin in conidial spores of Aspergillus fumigatus as well as in other DHN-melanized fungi. MelLec is ubiquitously expressed by CD31 + endothelial cells in mice, and is also expressed by a sub-population of these cells that co-express epithelial cell adhesion molecule and are detected only in the lung and the liver. In mouse models, MelLec was required for protection against disseminated infection with A. fumigatus. In humans, MelLec is also expressed by myeloid cells, and we identified a single nucleotide polymorphism of this receptor that negatively affected myeloid inflammatory responses and significantly increased the susceptibility of stem-cell transplant recipients to disseminated Aspergillus infections. MelLec therefore recognizes an immunologically active component commonly found on fungi and has an essential role in protective antifungal immunity in both mice and humans.
Facial emotion recognition ability: psychiatry nurses versus nurses from other departments.
Gultekin, Gozde; Kincir, Zeliha; Kurt, Merve; Catal, Yasir; Acil, Asli; Aydin, Aybike; Özcan, Mualla; Delikkaya, Busra N; Kacar, Selma; Emul, Murat
2016-12-01
Facial emotion recognition is a basic element in non-verbal communication. Although some researchers have shown that recognizing facial expressions may be important in the interaction between doctors and patients, there are no studies concerning facial emotion recognition in nurses. Here, we aimed to investigate facial emotion recognition ability in nurses and compare the abilities between nurses from psychiatry and other departments. In this cross-sectional study, sixty seven nurses were divided into two groups according to their departments: psychiatry (n=31); and, other departments (n=36). A Facial Emotion Recognition Test, constructed from a set of photographs from Ekman and Friesen's book "Pictures of Facial Affect", was administered to all participants. In whole group, the highest mean accuracy rate of recognizing facial emotion was the happy (99.14%) while the lowest accurately recognized facial expression was fear (47.71%). There were no significant differences between two groups among mean accuracy rates in recognizing happy, sad, fear, angry, surprised facial emotion expressions (for all, p>0.05). The ability of recognizing disgusted and neutral facial emotions tended to be better in other nurses than psychiatry nurses (p=0.052 and p=0.053, respectively) Conclusion: This study was the first that revealed indifference in the ability of FER between psychiatry nurses and non-psychiatry nurses. In medical education curricula throughout the world, no specific training program is scheduled for recognizing emotional cues of patients. We considered that improving the ability of recognizing facial emotion expression in medical stuff might be beneficial in reducing inappropriate patient-medical stuff interaction.
Oszvald, Maria; Kang, Tae-Jin; Tomoskozi, Sandor; Tamas, Cecilia; Tamas, Laszlo; Kim, Tae-Geum; Yang, Moon-Sik
2007-03-01
Epitopes often require co-delivery with adjuvant and targeting proteins to enable recognition by the immune system, and this approach may also increase the efficacy of the antigen. In this study, we assess and describe the ability of transgenic rice plants to express a fusion protein consisting of the B-subunit of the Escherichia coli heat-labile enterotoxin (LTB) and a synthetic core-neutralizing epitope (COE) of porcine epidemic diarrhea virus (PEDV), inducing an enteric disease that is seen most predominantly in piglets. Both components of the fusion proteins were detected with Western blot analysis. The fusion protein was determined to assemble into pentamers, as was evidenced by its ability to bind to GM1 gangliosides, and evidenced an average level of expression in a transgenic rice endosperm. This indicates that the expression system of the plant is capable of generating a sizable amount of antigen, possibly allowing for the successful development of an edible vaccine.
Das, Krishna; Eisel, David; Lenkl, Clarissa; Goyal, Ashish; Diederichs, Sven; Dickes, Elke; Osen, Wolfram; Eichmüller, Stefan B
2017-01-01
In this study, the CRISPR/Cas9 technology was used to establish murine tumor cell lines, devoid of MHC I or MHC II surface expression, respectively. The melanoma cell line B16F10 and the murine breast cancer cell line EO-771, the latter stably expressing the tumor antigen NY-BR-1 (EO-NY), were transfected with an expression plasmid encoding a β2m-specific single guide (sg)RNA and Cas9. The resulting MHC I negative cells were sorted by flow cytometry to obtain single cell clones, and loss of susceptibility of peptide pulsed MHC I negative clones to peptide-specific CTL recognition was determined by IFNγ ELISpot assay. The β2m knockout (KO) clones did not give rise to tumors in syngeneic mice (C57BL/6N), unless NK cells were depleted, suggesting that outgrowth of the β2m KO cell lines was controlled by NK cells. Using sgRNAs targeting the β-chain encoding locus of the IAb molecule we also generated several B16F10 MHC II KO clones. Peptide loaded B16F10 MHC II KO cells were insusceptible to recognition by OT-II cells and tumor growth was unaltered compared to parental B16F10 cells. Thus, in our hands the CRISPR/Cas9 system has proven to be an efficient straight forward strategy for the generation of MHC knockout cell lines. Such cell lines could serve as parental cells for co-transfection of compatible HLA alleles together with human tumor antigens of interest, thereby facilitating the generation of HLA matched transplantable tumor models, e.g. in HLAtg mouse strains of the newer generation, lacking cell surface expression of endogenous H2 molecules. In addition, our tumor cell lines established might offer a useful tool to investigate tumor reactive T cell responses that function independently from MHC molecule surface expression by the tumor.
The integration of visual context information in facial emotion recognition in 5- to 15-year-olds.
Theurel, Anne; Witt, Arnaud; Malsert, Jennifer; Lejeune, Fleur; Fiorentini, Chiara; Barisnikov, Koviljka; Gentaz, Edouard
2016-10-01
The current study investigated the role of congruent visual context information in the recognition of facial emotional expression in 190 participants from 5 to 15years of age. Children performed a matching task that presented pictures with different facial emotional expressions (anger, disgust, happiness, fear, and sadness) in two conditions: with and without a visual context. The results showed that emotions presented with visual context information were recognized more accurately than those presented in the absence of visual context. The context effect remained steady with age but varied according to the emotion presented and the gender of participants. The findings demonstrated for the first time that children from the age of 5years are able to integrate facial expression and visual context information, and this integration improves facial emotion recognition. Copyright © 2016 Elsevier Inc. All rights reserved.
Test battery for measuring the perception and recognition of facial expressions of emotion
Wilhelm, Oliver; Hildebrandt, Andrea; Manske, Karsten; Schacht, Annekathrin; Sommer, Werner
2014-01-01
Despite the importance of perceiving and recognizing facial expressions in everyday life, there is no comprehensive test battery for the multivariate assessment of these abilities. As a first step toward such a compilation, we present 16 tasks that measure the perception and recognition of facial emotion expressions, and data illustrating each task's difficulty and reliability. The scoring of these tasks focuses on either the speed or accuracy of performance. A sample of 269 healthy young adults completed all tasks. In general, accuracy and reaction time measures for emotion-general scores showed acceptable and high estimates of internal consistency and factor reliability. Emotion-specific scores yielded lower reliabilities, yet high enough to encourage further studies with such measures. Analyses of task difficulty revealed that all tasks are suitable for measuring emotion perception and emotion recognition related abilities in normal populations. PMID:24860528
Enhancement of cell recognition in vitro by dual-ligand cancer targeting gold naoparticles
Li, Xi; Zhou, Hongyu; Yang, Lei; Du, Guoqing; Pai-Panandiker, Atmaram; Huang, Xuefei; Yan, Bing
2011-01-01
A dual-ligand gold nanoparticle (DLGNP) was designed and synthesized to explore the therapeutic benefits of multivalent interactions between gold nanoparticles (GNPs) and cancer cells. DLGNP was tested on human epidermal cancer cells (KB), which had high expression of folate receptor. The cellular uptake of DLGNP was increased by 3.9 and 12.7 folds compared with GNP-folate or GNP-glucose. The enhanced cell recognition was due to multivalent interactions between both ligands on GNPs and cancer cells as shown by the ligand competition experiments. Furthermore, the multivalent interactions increased contrast between cells with high and low expression of folate receptors. The enhanced cell recognition enabled DLGNP to kill KB cells under X-ray irradiation at a dose that was safe to folate receptor low-expression (such as normal) cells. Thus DLGP has the potential to be a cancer-specific nano-theranostic agent. PMID:21232787
Is synthetic biology mechanical biology?
Holm, Sune
2015-12-01
A widespread and influential characterization of synthetic biology emphasizes that synthetic biology is the application of engineering principles to living systems. Furthermore, there is a strong tendency to express the engineering approach to organisms in terms of what seems to be an ontological claim: organisms are machines. In the paper I investigate the ontological and heuristic significance of the machine analogy in synthetic biology. I argue that the use of the machine analogy and the aim of producing rationally designed organisms does not necessarily imply a commitment to mechanical biology. The ideal of applying engineering principles to biology is best understood as expressing recognition of the machine-unlikeness of natural organisms and the limits of human cognition. The paper suggests an interpretation of the identification of organisms with machines in synthetic biology according to which it expresses a strategy for representing, understanding, and constructing living systems that are more machine-like than natural organisms.
The Role of Progesterone in the Feto-Maternal Immunological Crosstalk.
Szekeres-Bartho, Julia
2018-06-27
This review aims to provide a brief historical overview of the feto-maternal immunological relationship, which profoundly influences the outcome of pregnancy. The initial question posed in the nineteen fifties by Medawar, was based on the assumption that the maternal immune system recognizes the fetus as an allograft. Indeed, based on the association between HLA-matching and spontaneous miscarriage, it became obvious that immunological recognition of pregnancy is required for a successful gestation. The restricted expression of polymorphic HLA antigens on the trophoblast, together with the presence of non-polymorphic MHC products excludes recognition by both T and NK cells of trophoblast presented antigens, however, T cells, which constitute the majority of decidual T cells are likely candidates. Indeed, a high number of activated, progesterone receptor-expressing T cells are present in peripheral blood of healthy pregnant women, and in the presence of progesterone, these cells secrete an immunomodulatory protein called Progesterone-induced Blocking Factor or PIBF. As early as in the peri-implantation period, the embryo communicates with the maternal immune system via PIBF containing extracellular vesicles. PIBF contributes to the dominance of Th2-type reactivity characterising normal pregnancy, by inducing increased production of Th2 cytokines. The high expression of this molecule in the decidua might be one of the reasons for the low cytotoxic activity of decidual NK cells. ©2018The Author(s). Published by S. Karger AG, Basel.
Martinelli, Eugenio; Mencattini, Arianna; Di Natale, Corrado
2016-01-01
Humans can communicate their emotions by modulating facial expressions or the tone of their voice. Albeit numerous applications exist that enable machines to read facial emotions and recognize the content of verbal messages, methods for speech emotion recognition are still in their infancy. Yet, fast and reliable applications for emotion recognition are the obvious advancement of present ‘intelligent personal assistants’, and may have countless applications in diagnostics, rehabilitation and research. Taking inspiration from the dynamics of human group decision-making, we devised a novel speech emotion recognition system that applies, for the first time, a semi-supervised prediction model based on consensus. Three tests were carried out to compare this algorithm with traditional approaches. Labeling performances relative to a public database of spontaneous speeches are reported. The novel system appears to be fast, robust and less computationally demanding than traditional methods, allowing for easier implementation in portable voice-analyzers (as used in rehabilitation, research, industry, etc.) and for applications in the research domain (such as real-time pairing of stimuli to participants’ emotional state, selective/differential data collection based on emotional content, etc.). PMID:27563724
Traffic Behavior Recognition Using the Pachinko Allocation Model
Huynh-The, Thien; Banos, Oresti; Le, Ba-Vui; Bui, Dinh-Mao; Yoon, Yongik; Lee, Sungyoung
2015-01-01
CCTV-based behavior recognition systems have gained considerable attention in recent years in the transportation surveillance domain for identifying unusual patterns, such as traffic jams, accidents, dangerous driving and other abnormal behaviors. In this paper, a novel approach for traffic behavior modeling is presented for video-based road surveillance. The proposed system combines the pachinko allocation model (PAM) and support vector machine (SVM) for a hierarchical representation and identification of traffic behavior. A background subtraction technique using Gaussian mixture models (GMMs) and an object tracking mechanism based on Kalman filters are utilized to firstly construct the object trajectories. Then, the sparse features comprising the locations and directions of the moving objects are modeled by PAM into traffic topics, namely activities and behaviors. As a key innovation, PAM captures not only the correlation among the activities, but also among the behaviors based on the arbitrary directed acyclic graph (DAG). The SVM classifier is then utilized on top to train and recognize the traffic activity and behavior. The proposed model shows more flexibility and greater expressive power than the commonly-used latent Dirichlet allocation (LDA) approach, leading to a higher recognition accuracy in the behavior classification. PMID:26151213
Romani, Maria; Vigliante, Miriam; Faedda, Noemi; Rossetti, Serena; Pezzuti, Lina; Guidetti, Vincenzo; Cardona, Francesco
2018-06-01
This review focuses on facial recognition abilities in children and adolescents with attention deficit hyperactivity disorder (ADHD). A systematic review, using PRISMA guidelines, was conducted to identify original articles published prior to May 2017 pertaining to memory, face recognition, affect recognition, facial expression recognition and recall of faces in children and adolescents with ADHD. The qualitative synthesis based on different studies shows a particular focus of the research on facial affect recognition without paying similar attention to the structural encoding of facial recognition. In this review, we further investigate facial recognition abilities in children and adolescents with ADHD, providing synthesis of the results observed in the literature, while detecting face recognition tasks used on face processing abilities in ADHD and identifying aspects not yet explored. Copyright © 2018 Elsevier Ltd. All rights reserved.
Characterization and recognition of mixed emotional expressions in thermal face image
NASA Astrophysics Data System (ADS)
Saha, Priya; Bhattacharjee, Debotosh; De, Barin K.; Nasipuri, Mita
2016-05-01
Facial expressions in infrared imaging have been introduced to solve the problem of illumination, which is an integral constituent of visual imagery. The paper investigates facial skin temperature distribution on mixed thermal facial expressions of our created face database where six are basic expressions and rest 12 are a mixture of those basic expressions. Temperature analysis has been performed on three facial regions of interest (ROIs); periorbital, supraorbital and mouth. Temperature variability of the ROIs in different expressions has been measured using statistical parameters. The temperature variation measurement in ROIs of a particular expression corresponds to a vector, which is later used in recognition of mixed facial expressions. Investigations show that facial features in mixed facial expressions can be characterized by positive emotion induced facial features and negative emotion induced facial features. Supraorbital is a useful facial region that can differentiate basic expressions from mixed expressions. Analysis and interpretation of mixed expressions have been conducted with the help of box and whisker plot. Facial region containing mixture of two expressions is generally less temperature inducing than corresponding facial region containing basic expressions.
Altering sensorimotor feedback disrupts visual discrimination of facial expressions.
Wood, Adrienne; Lupyan, Gary; Sherrin, Steven; Niedenthal, Paula
2016-08-01
Looking at another person's facial expression of emotion can trigger the same neural processes involved in producing the expression, and such responses play a functional role in emotion recognition. Disrupting individuals' facial action, for example, interferes with verbal emotion recognition tasks. We tested the hypothesis that facial responses also play a functional role in the perceptual processing of emotional expressions. We altered the facial action of participants with a gel facemask while they performed a task that involved distinguishing target expressions from highly similar distractors. Relative to control participants, participants in the facemask condition demonstrated inferior perceptual discrimination of facial expressions, but not of nonface stimuli. The findings suggest that somatosensory/motor processes involving the face contribute to the visual perceptual-and not just conceptual-processing of facial expressions. More broadly, our study contributes to growing evidence for the fundamentally interactive nature of the perceptual inputs from different sensory modalities.
ERIC Educational Resources Information Center
Wright, Barry; Clarke, Natalie; Jordan, Jo; Young, Andrew W.; Clarke, Paula; Miles, Jeremy; Nation, Kate; Clarke, Leesa; Williams, Christine
2008-01-01
We compared young people with high-functioning autism spectrum disorders (ASDs) with age, sex and IQ matched controls on emotion recognition of faces and pictorial context. Each participant completed two tests of emotion recognition. The first used Ekman series faces. The second used facial expressions in visual context. A control task involved…
Pathogen recognition in the innate immune response.
Kumar, Himanshu; Kawai, Taro; Akira, Shizuo
2009-04-28
Immunity against microbial pathogens primarily depends on the recognition of pathogen components by innate receptors expressed on immune and non-immune cells. Innate receptors are evolutionarily conserved germ-line-encoded proteins and include TLRs (Toll-like receptors), RLRs [RIG-I (retinoic acid-inducible gene-I)-like receptors] and NLRs (Nod-like receptors). These receptors recognize pathogens or pathogen-derived products in different cellular compartments, such as the plasma membrane, the endosomes or the cytoplasm, and induce the expression of cytokines, chemokines and co-stimulatory molecules to eliminate pathogens and instruct pathogen-specific adaptive immune responses. In the present review, we will discuss the recent progress in the study of pathogen recognition by TLRs, RLRs and NLRs and their signalling pathways.
Isaacowitz, Derek M.; Stanley, Jennifer Tehan
2011-01-01
Older adults perform worse on traditional tests of emotion recognition accuracy than do young adults. In this paper, we review descriptive research to date on age differences in emotion recognition from facial expressions, as well as the primary theoretical frameworks that have been offered to explain these patterns. We propose that this is an area of inquiry that would benefit from an ecological approach in which contextual elements are more explicitly considered and reflected in experimental methods. Use of dynamic displays and examination of specific cues to accuracy, for example, may reveal more nuanced age-related patterns and may suggest heretofore unexplored underlying mechanisms. PMID:22125354
Identities of P2 and P3 Residues of H-2Kb-Bound Peptides Determine Mouse Ly49C Recognition
Marquez, Elsa A.; Kane, Kevin P.
2015-01-01
Ly49 receptors can be peptide selective in their recognition of MHC-I-peptide complexes, affording them a level of discrimination beyond detecting the presence or absence of specific MHC-I allele products. Despite this ability, little is understood regarding the properties that enable some peptides, when bound to MHC-I molecules, to support Ly49 recognition, but not others. Using RMA-S target cells expressing MHC-I molecules loaded with individual peptides and effector cells expressing the ectodomain of the inhibitory Ly49C receptor, we found that two adjacent amino acid residues, P2 and P3, both buried in the peptide binding groove of H-2Kb, determine mouse Ly49C specificity. If both are aliphatic residues, this is supportive. Whereas, small amino acids at P2 and aromatic amino acids at the P3 auxiliary anchor residue are detrimental to Ly49C recognition. These results resemble those with a rat Ly49 where the identity of a peptide anchor residue determines recognition, suggesting that dependence on specific peptide residues buried in the MHC-I peptide-binding groove may be fundamental to Ly49 peptide selectivity and recognition. PMID:26147851
Ensemble training to improve recognition using 2D ear
NASA Astrophysics Data System (ADS)
Middendorff, Christopher; Bowyer, Kevin W.
2009-05-01
The ear has gained popularity as a biometric feature due to the robustness of the shape over time and across emotional expression. Popular methods of ear biometrics analyze the ear as a whole, leaving these methods vulnerable to error due to occlusion. Many researchers explore ear recognition using an ensemble, but none present a method for designing the individual parts that comprise the ensemble. In this work, we introduce a method of modifying the ensemble shapes to improve performance. We determine how different properties of an ensemble training system can affect overall performance. We show that ensembles built from small parts will outperform ensembles built with larger parts, and that incorporating a large number of parts improves the performance of the ensemble.
Introduction of statistical information in a syntactic analyzer for document image recognition
NASA Astrophysics Data System (ADS)
Maroneze, André O.; Coüasnon, Bertrand; Lemaitre, Aurélie
2011-01-01
This paper presents an improvement to document layout analysis systems, offering a possible solution to Sayre's paradox (which states that an element "must be recognized before it can be segmented; and it must be segmented before it can be recognized"). This improvement, based on stochastic parsing, allows integration of statistical information, obtained from recognizers, during syntactic layout analysis. We present how this fusion of numeric and symbolic information in a feedback loop can be applied to syntactic methods to improve document description expressiveness. To limit combinatorial explosion during exploration of solutions, we devised an operator that allows optional activation of the stochastic parsing mechanism. Our evaluation on 1250 handwritten business letters shows this method allows the improvement of global recognition scores.
Indersmitten, Tim; Gur, Ruben C
2003-05-01
Since the discovery of facial asymmetries in emotional expressions of humans and other primates, hypotheses have related the greater left-hemiface intensity to right-hemispheric dominance in emotion processing. However, the difficulty of creating true frontal views of facial expressions in two-dimensional photographs has confounded efforts to better understand the phenomenon. We have recently described a method for obtaining three-dimensional photographs of posed and evoked emotional expressions and used these stimuli to investigate both intensity of expression and accuracy of recognizing emotion in chimeric faces constructed from only left- or right-side composites. The participant population included 38 (19 male, 19 female) African-American, Caucasian, and Asian adults. They were presented with chimeric composites generated from faces of eight actors and eight actresses showing four emotions: happiness, sadness, anger, and fear, each in posed and evoked conditions. We replicated the finding that emotions are expressed more intensely in the left hemiface for all emotions and conditions, with the exception of evoked anger, which was expressed more intensely in the right hemiface. In contrast, the results indicated that emotional expressions are recognized more efficiently in the right hemiface, indicating that the right hemiface expresses emotions more accurately. The double dissociation between the laterality of expression intensity and that of recognition efficiency supports the notion that the two kinds of processes may have distinct neural substrates. Evoked anger is uniquely expressed more intensely and accurately on the side of the face that projects to the viewer's right hemisphere, dominant in emotion recognition.
Formal implementation of a performance evaluation model for the face recognition system.
Shin, Yong-Nyuo; Kim, Jason; Lee, Yong-Jun; Shin, Woochang; Choi, Jin-Young
2008-01-01
Due to usability features, practical applications, and its lack of intrusiveness, face recognition technology, based on information, derived from individuals' facial features, has been attracting considerable attention recently. Reported recognition rates of commercialized face recognition systems cannot be admitted as official recognition rates, as they are based on assumptions that are beneficial to the specific system and face database. Therefore, performance evaluation methods and tools are necessary to objectively measure the accuracy and performance of any face recognition system. In this paper, we propose and formalize a performance evaluation model for the biometric recognition system, implementing an evaluation tool for face recognition systems based on the proposed model. Furthermore, we performed evaluations objectively by providing guidelines for the design and implementation of a performance evaluation system, formalizing the performance test process.
Recognition Of Complex Three Dimensional Objects Using Three Dimensional Moment Invariants
NASA Astrophysics Data System (ADS)
Sadjadi, Firooz A.
1985-01-01
A technique for the recognition of complex three dimensional objects is presented. The complex 3-D objects are represented in terms of their 3-D moment invariants, algebraic expressions that remain invariant independent of the 3-D objects' orientations and locations in the field of view. The technique of 3-D moment invariants has been used successfully for simple 3-D object recognition in the past. In this work we have extended this method for the representation of more complex objects. Two complex objects are represented digitally; their 3-D moment invariants have been calculated, and then the invariancy of these 3-D invariant moment expressions is verified by changing the orientation and the location of the objects in the field of view. The results of this study have significant impact on 3-D robotic vision, 3-D target recognition, scene analysis and artificial intelligence.
Gender affects body language reading.
Sokolov, Arseny A; Krüger, Samuel; Enck, Paul; Krägeloh-Mann, Ingeborg; Pavlova, Marina A
2011-01-01
Body motion is a rich source of information for social cognition. However, gender effects in body language reading are largely unknown. Here we investigated whether, and, if so, how recognition of emotional expressions revealed by body motion is gender dependent. To this end, females and males were presented with point-light displays portraying knocking at a door performed with different emotional expressions. The findings show that gender affects accuracy rather than speed of body language reading. This effect, however, is modulated by emotional content of actions: males surpass in recognition accuracy of happy actions, whereas females tend to excel in recognition of hostile angry knocking. Advantage of women in recognition accuracy of neutral actions suggests that females are better tuned to the lack of emotional content in body actions. The study provides novel insights into understanding of gender effects in body language reading, and helps to shed light on gender vulnerability to neuropsychiatric and neurodevelopmental impairments in visual social cognition.
Zhang, Rui; Liu, Jialin; Huang, Yong; Wang, Miye; Shi, Qingke; Chen, Jun; Zeng, Zhi
2017-05-02
It has been shown that the entities in everyday clinical text are often expressed in a way that varies from how they are expressed in the nomenclature. Owing to lots of synonyms, abbreviations, medical jargons or even misspellings in the daily used physician notes in clinical information system (CIS), the terminology without enough synonyms may not be adequately suitable for the task of Chinese clinical term recognition. This paper demonstrates a validated system to retrieve the Chinese term of clinical finding (CTCF) from CIS and map them to the corresponding concepts of international clinical nomenclature, such as SNOMED CT. The system focuses on the SNOMED CT with Chinese synonyms enrichment (SCCSE). The literal similarity and the diagnosis-related similarity metrics were used for concept mapping. Two CTCF recognition methods, the rule- and terminology-based approach (RTBA) and the conditional random field machine learner (CRF), were adopted to identify the concepts in physician notes. The system was validated against the history of present illness annotated by clinical experts. The RTBA and CRF could be combined to predict new CTCFs besides SCCSE persistently. Around 59,000 CTCF candidates were accepted as valid and 39,000 of them occurred at least once in the history of present illness. 3,729 of them were accordant with the description in referenced Chinese clinical nomenclature, which could cross map to other international nomenclature such as SNOMED CT. With the hybrid similarity metrics, another 7,454 valid CTCFs (synonyms) were succeeded in concept mapping. For CTCF recognition in physician notes, a series of experiments were performed to find out the best CRF feature set, which gained an F-score of 0.887. The RTBA achieved a better F-score of 0.919 by the CTCF dictionary created in this research. This research demonstrated that it is feasible to help the SNOMED CT with Chinese synonyms enrichment based on physician notes in CIS. With continuous maintenance of SCCSE, the CTCFs could be precisely retrieved from free text, and the CTCFs arranged in semantic hierarchy of SNOMED CT could greatly improve the meaningful use of electronic health record in China. The methodology is also useful for clinical synonyms enrichment in other languages.
Impaired Perception of Emotional Expression in Amyotrophic Lateral Sclerosis.
Oh, Seong Il; Oh, Ki Wook; Kim, Hee Jin; Park, Jin Seok; Kim, Seung Hyun
2016-07-01
The increasing recognition that deficits in social emotions occur in amyotrophic lateral sclerosis (ALS) is helping to explain the spectrum of neuropsychological dysfunctions, thus supporting the view of ALS as a multisystem disorder involving neuropsychological deficits as well as motor deficits. The aim of this study was to characterize the emotion perception abilities of Korean patients with ALS based on the recognition of facial expressions. Twenty-four patients with ALS and 24 age- and sex-matched healthy controls completed neuropsychological tests and facial emotion recognition tasks [ChaeLee Korean Facial Expressions of Emotions (ChaeLee-E)]. The ChaeLee-E test includes facial expressions for seven emotions: happiness, sadness, anger, disgust, fear, surprise, and neutral. The ability to perceive facial emotions was significantly worse among ALS patients performed than among healthy controls [65.2±18.0% vs. 77.1±6.6% (mean±SD), p=0.009]. Eight of the 24 patients (33%) scored below the 5th percentile score of controls for recognizing facial emotions. Emotion perception deficits occur in Korean ALS patients, particularly regarding facial expressions of emotion. These findings expand the spectrum of cognitive and behavioral dysfunction associated with ALS into emotion processing dysfunction.
Franklin, Robert G; Adams, Reginald B; Steiner, Troy G; Zebrowitz, Leslie A
2018-05-14
Through 3 studies, we investigated whether angularity and roundness present in faces contributes to the perception of anger and joyful expressions, respectively. First, in Study 1 we found that angry expressions naturally contain more inward-pointing lines, whereas joyful expressions contain more outward-pointing lines. Then, using image-processing techniques in Studies 2 and 3, we filtered images to contain only inward-pointing or outward-pointing lines as a way to approximate angularity and roundness. We found that filtering images to be more angular increased how threatening and angry a neutral face was rated, increased how intense angry expressions were rated, and enhanced the recognition of anger. Conversely, filtering images to be rounder increased how warm and joyful a neutral face was rated, increased the intensity of joyful expressions, and enhanced recognition of joy. Together these findings show that angularity and roundness play a direct role in the recognition of angry and joyful expressions. Given evidence that angularity and roundness may play a biological role in indicating threat and safety in the environment, this suggests that angularity and roundness represent primitive facial cues used to signal threat-anger and warmth-joy pairings. (PsycINFO Database Record (c) 2018 APA, all rights reserved).
Vaccines against advanced melanoma.
Blanchard, Tatiana; Srivastava, Pramod K; Duan, Fei
2013-01-01
Research shows that cancers are recognized by the immune system but that the immune recognition of tumors does not uniformly result in tumor rejection or regression. Quantitating the success or failure of the immune system in tumor elimination is difficult because we do not really know the total numbers of encounters of the immune system with the tumors. Regardless of that important issue, recognition of the tumor by the immune system implicitly contains the idea of the tumor antigen, which is what is actually recognized. We review the molecular identity of all forms of tumor antigens (antigens with specific mutations, cancer-testis antigens, differentiation antigens, over-expressed antigens) and discuss the use of these multiple forms of antigens in experimental immunotherapy of mouse and human melanoma. These efforts have been uniformly unsuccessful; however, the approaches that have not worked or have somewhat worked have been the source of many new insights into melanoma immunology. From a critical review of the various approaches to vaccine therapy we conclude that individual cancer-specific mutations are truly the only sources of cancer-specific antigens, and therefore, the most attractive targets for immunotherapy. Published by Elsevier Inc.
Basic and complex emotion recognition in children with autism: cross-cultural findings.
Fridenson-Hayo, Shimrit; Berggren, Steve; Lassalle, Amandine; Tal, Shahar; Pigat, Delia; Bölte, Sven; Baron-Cohen, Simon; Golan, Ofer
2016-01-01
Children with autism spectrum conditions (ASC) have emotion recognition deficits when tested in different expression modalities (face, voice, body). However, these findings usually focus on basic emotions, using one or two expression modalities. In addition, cultural similarities and differences in emotion recognition patterns in children with ASC have not been explored before. The current study examined the similarities and differences in the recognition of basic and complex emotions by children with ASC and typically developing (TD) controls across three cultures: Israel, Britain, and Sweden. Fifty-five children with high-functioning ASC, aged 5-9, were compared to 58 TD children. On each site, groups were matched on age, sex, and IQ. Children were tested using four tasks, examining recognition of basic and complex emotions from voice recordings, videos of facial and bodily expressions, and emotional video scenarios including all modalities in context. Compared to their TD peers, children with ASC showed emotion recognition deficits in both basic and complex emotions on all three modalities and their integration in context. Complex emotions were harder to recognize, compared to basic emotions for the entire sample. Cross-cultural agreement was found for all major findings, with minor deviations on the face and body tasks. Our findings highlight the multimodal nature of ER deficits in ASC, which exist for basic as well as complex emotions and are relatively stable cross-culturally. Cross-cultural research has the potential to reveal both autism-specific universal deficits and the role that specific cultures play in the way empathy operates in different countries.
Appearance-based human gesture recognition using multimodal features for human computer interaction
NASA Astrophysics Data System (ADS)
Luo, Dan; Gao, Hua; Ekenel, Hazim Kemal; Ohya, Jun
2011-03-01
The use of gesture as a natural interface plays an utmost important role for achieving intelligent Human Computer Interaction (HCI). Human gestures include different components of visual actions such as motion of hands, facial expression, and torso, to convey meaning. So far, in the field of gesture recognition, most previous works have focused on the manual component of gestures. In this paper, we present an appearance-based multimodal gesture recognition framework, which combines the different groups of features such as facial expression features and hand motion features which are extracted from image frames captured by a single web camera. We refer 12 classes of human gestures with facial expression including neutral, negative and positive meanings from American Sign Languages (ASL). We combine the features in two levels by employing two fusion strategies. At the feature level, an early feature combination can be performed by concatenating and weighting different feature groups, and LDA is used to choose the most discriminative elements by projecting the feature on a discriminative expression space. The second strategy is applied on decision level. Weighted decisions from single modalities are fused in a later stage. A condensation-based algorithm is adopted for classification. We collected a data set with three to seven recording sessions and conducted experiments with the combination techniques. Experimental results showed that facial analysis improve hand gesture recognition, decision level fusion performs better than feature level fusion.
Applying Suffix Rules to Organization Name Recognition
NASA Astrophysics Data System (ADS)
Inui, Takashi; Murakami, Koji; Hashimoto, Taiichi; Utsumi, Kazuo; Ishikawa, Masamichi
This paper presents a method for boosting the performance of the organization name recognition, which is a part of named entity recognition (NER). Although gazetteers (lists of the NEs) have been known as one of the effective features for supervised machine learning approaches on the NER task, the previous methods which have applied the gazetteers to the NER were very simple. The gazetteers have been used just for searching the exact matches between input text and NEs included in them. The proposed method generates regular expression rules from gazetteers, and, with these rules, it can realize a high-coverage searches based on looser matches between input text and NEs. To generate these rules, we focus on the two well-known characteristics of NE expressions; 1) most of NE expressions can be divided into two parts, class-reference part and instance-reference part, 2) for most of NE expressions the class-reference parts are located at the suffix position of them. A pattern mining algorithm runs on the set of NEs in the gazetteers, and some frequent word sequences from which NEs are constructed are found. Then, we employ only word sequences which have the class-reference part at the suffix position as suffix rules. Experimental results showed that our proposed method improved the performance of the organization name recognition, and achieved the 84.58 F-value for evaluation data.
Customized Molecular Phenotyping by Quantitative Gene Expression and Pattern Recognition Analysis
Akilesh, Shreeram; Shaffer, Daniel J.; Roopenian, Derry
2003-01-01
Description of the molecular phenotypes of pathobiological processes in vivo is a pressing need in genomic biology. We have implemented a high-throughput real-time PCR strategy to establish quantitative expression profiles of a customized set of target genes. It enables rapid, reproducible data acquisition from limited quantities of RNA, permitting serial sampling of mouse blood during disease progression. We developed an easy to use statistical algorithm—Global Pattern Recognition—to readily identify genes whose expression has changed significantly from healthy baseline profiles. This approach provides unique molecular signatures for rheumatoid arthritis, systemic lupus erythematosus, and graft versus host disease, and can also be applied to defining the molecular phenotype of a variety of other normal and pathological processes. PMID:12840047
Network Security via Biometric Recognition of Patterns of Gene Expression
NASA Technical Reports Server (NTRS)
Shaw, Harry C.
2016-01-01
Molecular biology provides the ability to implement forms of information and network security completely outside the bounds of legacy security protocols and algorithms. This paper addresses an approach which instantiates the power of gene expression for security. Molecular biology provides a rich source of gene expression and regulation mechanisms, which can be adopted to use in the information and electronic communication domains. Conventional security protocols are becoming increasingly vulnerable due to more intensive, highly capable attacks on the underlying mathematics of cryptography. Security protocols are being undermined by social engineering and substandard implementations by IT organizations. Molecular biology can provide countermeasures to these weak points with the current security approaches. Future advances in instruments for analyzing assays will also enable this protocol to advance from one of cryptographic algorithms to an integrated system of cryptographic algorithms and real-time expression and assay of gene expression products.
Design method of ARM based embedded iris recognition system
NASA Astrophysics Data System (ADS)
Wang, Yuanbo; He, Yuqing; Hou, Yushi; Liu, Ting
2008-03-01
With the advantages of non-invasiveness, uniqueness, stability and low false recognition rate, iris recognition has been successfully applied in many fields. Up to now, most of the iris recognition systems are based on PC. However, a PC is not portable and it needs more power. In this paper, we proposed an embedded iris recognition system based on ARM. Considering the requirements of iris image acquisition and recognition algorithm, we analyzed the design method of the iris image acquisition module, designed the ARM processing module and its peripherals, studied the Linux platform and the recognition algorithm based on this platform, finally actualized the design method of ARM-based iris imaging and recognition system. Experimental results show that the ARM platform we used is fast enough to run the iris recognition algorithm, and the data stream can flow smoothly between the camera and the ARM chip based on the embedded Linux system. It's an effective method of using ARM to actualize portable embedded iris recognition system.
Gender Differences in the Recognition of Vocal Emotions
Lausen, Adi; Schacht, Annekathrin
2018-01-01
The conflicting findings from the few studies conducted with regard to gender differences in the recognition of vocal expressions of emotion have left the exact nature of these differences unclear. Several investigators have argued that a comprehensive understanding of gender differences in vocal emotion recognition can only be achieved by replicating these studies while accounting for influential factors such as stimulus type, gender-balanced samples, number of encoders, decoders, and emotional categories. This study aimed to account for these factors by investigating whether emotion recognition from vocal expressions differs as a function of both listeners' and speakers' gender. A total of N = 290 participants were randomly and equally allocated to two groups. One group listened to words and pseudo-words, while the other group listened to sentences and affect bursts. Participants were asked to categorize the stimuli with respect to the expressed emotions in a fixed-choice response format. Overall, females were more accurate than males when decoding vocal emotions, however, when testing for specific emotions these differences were small in magnitude. Speakers' gender had a significant impact on how listeners' judged emotions from the voice. The group listening to words and pseudo-words had higher identification rates for emotions spoken by male than by female actors, whereas in the group listening to sentences and affect bursts the identification rates were higher when emotions were uttered by female than male actors. The mixed pattern for emotion-specific effects, however, indicates that, in the vocal channel, the reliability of emotion judgments is not systematically influenced by speakers' gender and the related stereotypes of emotional expressivity. Together, these results extend previous findings by showing effects of listeners' and speakers' gender on the recognition of vocal emotions. They stress the importance of distinguishing these factors to explain recognition ability in the processing of emotional prosody. PMID:29922202
Effects of training set selection on pain recognition via facial expressions
NASA Astrophysics Data System (ADS)
Shier, Warren A.; Yanushkevich, Svetlana N.
2016-07-01
This paper presents an approach to pain expression classification based on Gabor energy filters with Support Vector Machines (SVMs), followed by analyzing the effects of training set variations on the systems classification rate. This approach is tested on the UNBC-McMaster Shoulder Pain Archive, which consists of spontaneous pain images, hand labelled using the Prkachin and Solomon Pain Intensity scale. In this paper, the subjects pain intensity level has been quantized into three disjoint groups: no pain, weak pain and strong pain. The results of experiments show that Gabor energy filters with SVMs provide comparable or better results to previous filter- based pain recognition methods, with precision rates of 74%, 30% and 78% for no pain, weak pain and strong pain, respectively. The study of effects of intra-class skew, or changing the number of images per subject, show that both completely removing and over-representing poor quality subjects in the training set has little effect on the overall accuracy of the system. This result suggests that poor quality subjects could be removed from the training set to save offline training time and that SVM is robust not only to outliers in training data, but also to significant amounts of poor quality data mixed into the training sets.
Identifying differences in biased affective information processing in major depression.
Gollan, Jackie K; Pane, Heather T; McCloskey, Michael S; Coccaro, Emil F
2008-05-30
This study investigates the extent to which participants with major depression differ from healthy comparison participants in the irregularities in affective information processing, characterized by deficits in facial expression recognition, intensity categorization, and reaction time to identifying emotionally salient and neutral information. Data on diagnoses, symptom severity, and affective information processing using a facial recognition task were collected from 66 participants, male and female between ages 18 and 54 years, grouped by major depressive disorder (N=37) or healthy non-psychiatric (N=29) status. Findings from MANCOVAs revealed that major depression was associated with a significantly longer reaction time to sad facial expressions compared with healthy status. Also, depressed participants demonstrated a negative bias towards interpreting neutral facial expressions as sad significantly more often than healthy participants. In turn, healthy participants interpreted neutral faces as happy significantly more often than depressed participants. No group differences were observed for facial expression recognition and intensity categorization. The observed effects suggest that depression has significant effects on the perception of the intensity of negative affective stimuli, delayed speed of processing sad affective information, and biases towards interpreting neutral faces as sad.
Efficient Spatio-Temporal Local Binary Patterns for Spontaneous Facial Micro-Expression Recognition
Wang, Yandan; See, John; Phan, Raphael C.-W.; Oh, Yee-Hui
2015-01-01
Micro-expression recognition is still in the preliminary stage, owing much to the numerous difficulties faced in the development of datasets. Since micro-expression is an important affective clue for clinical diagnosis and deceit analysis, much effort has gone into the creation of these datasets for research purposes. There are currently two publicly available spontaneous micro-expression datasets—SMIC and CASME II, both with baseline results released using the widely used dynamic texture descriptor LBP-TOP for feature extraction. Although LBP-TOP is popular and widely used, it is still not compact enough. In this paper, we draw further inspiration from the concept of LBP-TOP that considers three orthogonal planes by proposing two efficient approaches for feature extraction. The compact robust form described by the proposed LBP-Six Intersection Points (SIP) and a super-compact LBP-Three Mean Orthogonal Planes (MOP) not only preserves the essential patterns, but also reduces the redundancy that affects the discriminality of the encoded features. Through a comprehensive set of experiments, we demonstrate the strengths of our approaches in terms of recognition accuracy and efficiency. PMID:25993498
Multimodal emotional state recognition using sequence-dependent deep hierarchical features.
Barros, Pablo; Jirak, Doreen; Weber, Cornelius; Wermter, Stefan
2015-12-01
Emotional state recognition has become an important topic for human-robot interaction in the past years. By determining emotion expressions, robots can identify important variables of human behavior and use these to communicate in a more human-like fashion and thereby extend the interaction possibilities. Human emotions are multimodal and spontaneous, which makes them hard to be recognized by robots. Each modality has its own restrictions and constraints which, together with the non-structured behavior of spontaneous expressions, create several difficulties for the approaches present in the literature, which are based on several explicit feature extraction techniques and manual modality fusion. Our model uses a hierarchical feature representation to deal with spontaneous emotions, and learns how to integrate multiple modalities for non-verbal emotion recognition, making it suitable to be used in an HRI scenario. Our experiments show that a significant improvement of recognition accuracy is achieved when we use hierarchical features and multimodal information, and our model improves the accuracy of state-of-the-art approaches from 82.5% reported in the literature to 91.3% for a benchmark dataset on spontaneous emotion expressions. Copyright © 2015 The Authors. Published by Elsevier Ltd.. All rights reserved.
Castagna, Filomena; Montemagni, Cristiana; Maria Milani, Anna; Rocca, Giuseppe; Rocca, Paola; Casacchia, Massimo; Bogetto, Filippo
2013-02-28
This study aimed to evaluate the ability to decode emotion in the auditory and audiovisual modality in a group of patients with schizophrenia, and to explore the role of cognition and psychopathology in affecting these emotion recognition abilities. Ninety-four outpatients in a stable phase and 51 healthy subjects were recruited. Patients were assessed through a psychiatric evaluation and a wide neuropsychological battery. All subjects completed the comprehensive affect testing system (CATS), a group of computerized tests designed to evaluate emotion perception abilities. With respect to the controls, patients were not impaired in the CATS tasks involving discrimination of nonemotional prosody, naming of emotional stimuli expressed by voice and judging the emotional content of a sentence, whereas they showed a specific impairment in decoding emotion in a conflicting auditory condition and in the multichannel modality. Prosody impairment was affected by executive functions, attention and negative symptoms, while deficit in multisensory emotion recognition was affected by executive functions and negative symptoms. These emotion recognition deficits, rather than being associated purely with emotion perception disturbances in schizophrenia, are affected by core symptoms of the illness. Copyright © 2012 Elsevier Ireland Ltd. All rights reserved.
Construction of diagnosis system and gene regulatory networks based on microarray analysis.
Hong, Chun-Fu; Chen, Ying-Chen; Chen, Wei-Chun; Tu, Keng-Chang; Tsai, Meng-Hsiun; Chan, Yung-Kuan; Yu, Shyr Shen
2018-05-01
A microarray analysis generally contains expression data of thousands of genes, but most of them are irrelevant to the disease of interest, making analyzing the genes concerning specific diseases complicated. Therefore, filtering out a few essential genes as well as their regulatory networks is critical, and a disease can be easily diagnosed just depending on the expression profiles of a few critical genes. In this study, a target gene screening (TGS) system, which is a microarray-based information system that integrates F-statistics, pattern recognition matching, a two-layer K-means classifier, a Parameter Detection Genetic Algorithm (PDGA), a genetic-based gene selector (GBG selector) and the association rule, was developed to screen out a small subset of genes that can discriminate malignant stages of cancers. During the first stage, F-statistic, pattern recognition matching, and a two-layer K-means classifier were applied in the system to filter out the 20 critical genes most relevant to ovarian cancer from 9600 genes, and the PDGA was used to decide the fittest values of the parameters for these critical genes. Among the 20 critical genes, 15 are associated with cancer progression. In the second stage, we further employed a GBG selector and the association rule to screen out seven target gene sets, each with only four to six genes, and each of which can precisely identify the malignancy stage of ovarian cancer based on their expression profiles. We further deduced the gene regulatory networks of the 20 critical genes by applying the Pearson correlation coefficient to evaluate the correlationship between the expression of each gene at the same stages and at different stages. Correlationships between gene pairs were calculated, and then, three regulatory networks were deduced. Their correlationships were further confirmed by the Ingenuity pathway analysis. The prognostic significances of the genes identified via regulatory networks were examined using online tools, and most represented biomarker candidates. In summary, our proposed system provides a new strategy to identify critical genes or biomarkers, as well as their regulatory networks, from microarray data. Copyright © 2018. Published by Elsevier Inc.
Container-code recognition system based on computer vision and deep neural networks
NASA Astrophysics Data System (ADS)
Liu, Yi; Li, Tianjian; Jiang, Li; Liang, Xiaoyao
2018-04-01
Automatic container-code recognition system becomes a crucial requirement for ship transportation industry in recent years. In this paper, an automatic container-code recognition system based on computer vision and deep neural networks is proposed. The system consists of two modules, detection module and recognition module. The detection module applies both algorithms based on computer vision and neural networks, and generates a better detection result through combination to avoid the drawbacks of the two methods. The combined detection results are also collected for online training of the neural networks. The recognition module exploits both character segmentation and end-to-end recognition, and outputs the recognition result which passes the verification. When the recognition module generates false recognition, the result will be corrected and collected for online training of the end-to-end recognition sub-module. By combining several algorithms, the system is able to deal with more situations, and the online training mechanism can improve the performance of the neural networks at runtime. The proposed system is able to achieve 93% of overall recognition accuracy.
Recognizing Age-Separated Face Images: Humans and Machines
Yadav, Daksha; Singh, Richa; Vatsa, Mayank; Noore, Afzel
2014-01-01
Humans utilize facial appearance, gender, expression, aging pattern, and other ancillary information to recognize individuals. It is interesting to observe how humans perceive facial age. Analyzing these properties can help in understanding the phenomenon of facial aging and incorporating the findings can help in designing effective algorithms. Such a study has two components - facial age estimation and age-separated face recognition. Age estimation involves predicting the age of an individual given his/her facial image. On the other hand, age-separated face recognition consists of recognizing an individual given his/her age-separated images. In this research, we investigate which facial cues are utilized by humans for estimating the age of people belonging to various age groups along with analyzing the effect of one's gender, age, and ethnicity on age estimation skills. We also analyze how various facial regions such as binocular and mouth regions influence age estimation and recognition capabilities. Finally, we propose an age-invariant face recognition algorithm that incorporates the knowledge learned from these observations. Key observations of our research are: (1) the age group of newborns and toddlers is easiest to estimate, (2) gender and ethnicity do not affect the judgment of age group estimation, (3) face as a global feature, is essential to achieve good performance in age-separated face recognition, and (4) the proposed algorithm yields improved recognition performance compared to existing algorithms and also outperforms a commercial system in the young image as probe scenario. PMID:25474200
Recognizing age-separated face images: humans and machines.
Yadav, Daksha; Singh, Richa; Vatsa, Mayank; Noore, Afzel
2014-01-01
Humans utilize facial appearance, gender, expression, aging pattern, and other ancillary information to recognize individuals. It is interesting to observe how humans perceive facial age. Analyzing these properties can help in understanding the phenomenon of facial aging and incorporating the findings can help in designing effective algorithms. Such a study has two components--facial age estimation and age-separated face recognition. Age estimation involves predicting the age of an individual given his/her facial image. On the other hand, age-separated face recognition consists of recognizing an individual given his/her age-separated images. In this research, we investigate which facial cues are utilized by humans for estimating the age of people belonging to various age groups along with analyzing the effect of one's gender, age, and ethnicity on age estimation skills. We also analyze how various facial regions such as binocular and mouth regions influence age estimation and recognition capabilities. Finally, we propose an age-invariant face recognition algorithm that incorporates the knowledge learned from these observations. Key observations of our research are: (1) the age group of newborns and toddlers is easiest to estimate, (2) gender and ethnicity do not affect the judgment of age group estimation, (3) face as a global feature, is essential to achieve good performance in age-separated face recognition, and (4) the proposed algorithm yields improved recognition performance compared to existing algorithms and also outperforms a commercial system in the young image as probe scenario.
Experimental study on GMM-based speaker recognition
NASA Astrophysics Data System (ADS)
Ye, Wenxing; Wu, Dapeng; Nucci, Antonio
2010-04-01
Speaker recognition plays a very important role in the field of biometric security. In order to improve the recognition performance, many pattern recognition techniques have be explored in the literature. Among these techniques, the Gaussian Mixture Model (GMM) is proved to be an effective statistic model for speaker recognition and is used in most state-of-the-art speaker recognition systems. The GMM is used to represent the 'voice print' of a speaker through modeling the spectral characteristic of speech signals of the speaker. In this paper, we implement a speaker recognition system, which consists of preprocessing, Mel-Frequency Cepstrum Coefficients (MFCCs) based feature extraction, and GMM based classification. We test our system with TIDIGITS data set (325 speakers) and our own recordings of more than 200 speakers; our system achieves 100% correct recognition rate. Moreover, we also test our system under the scenario that training samples are from one language but test samples are from a different language; our system also achieves 100% correct recognition rate, which indicates that our system is language independent.
Oxytocin Reduces Face Processing Time but Leaves Recognition Accuracy and Eye-Gaze Unaffected.
Hubble, Kelly; Daughters, Katie; Manstead, Antony S R; Rees, Aled; Thapar, Anita; van Goozen, Stephanie H M
2017-01-01
Previous studies have found that oxytocin (OXT) can improve the recognition of emotional facial expressions; it has been proposed that this effect is mediated by an increase in attention to the eye-region of faces. Nevertheless, evidence in support of this claim is inconsistent, and few studies have directly tested the effect of oxytocin on emotion recognition via altered eye-gaze Methods: In a double-blind, within-subjects, randomized control experiment, 40 healthy male participants received 24 IU intranasal OXT and placebo in two identical experimental sessions separated by a 2-week interval. Visual attention to the eye-region was assessed on both occasions while participants completed a static facial emotion recognition task using medium intensity facial expressions. Although OXT had no effect on emotion recognition accuracy, recognition performance was improved because face processing was faster across emotions under the influence of OXT. This effect was marginally significant (p<.06). Consistent with a previous study using dynamic stimuli, OXT had no effect on eye-gaze patterns when viewing static emotional faces and this was not related to recognition accuracy or face processing time. These findings suggest that OXT-induced enhanced facial emotion recognition is not necessarily mediated by an increase in attention to the eye-region of faces, as previously assumed. We discuss several methodological issues which may explain discrepant findings and suggest the effect of OXT on visual attention may differ depending on task requirements. (JINS, 2017, 23, 23-33).
Ferrucci, Roberta; Giannicola, Gaia; Rosa, Manuela; Fumagalli, Manuela; Boggio, Paulo Sergio; Hallett, Mark; Zago, Stefano; Priori, Alberto
2012-01-01
Some evidence suggests that the cerebellum participates in the complex network processing emotional facial expression. To evaluate the role of the cerebellum in recognising facial expressions we delivered transcranial direct current stimulation (tDCS) over the cerebellum and prefrontal cortex. A facial emotion recognition task was administered to 21 healthy subjects before and after cerebellar tDCS; we also tested subjects with a visual attention task and a visual analogue scale (VAS) for mood. Anodal and cathodal cerebellar tDCS both significantly enhanced sensory processing in response to negative facial expressions (anodal tDCS, p=.0021; cathodal tDCS, p=.018), but left positive emotion and neutral facial expressions unchanged (p>.05). tDCS over the right prefrontal cortex left facial expressions of both negative and positive emotion unchanged. These findings suggest that the cerebellum is specifically involved in processing facial expressions of negative emotion.
Measuring facial expression of emotion.
Wolf, Karsten
2015-12-01
Research into emotions has increased in recent decades, especially on the subject of recognition of emotions. However, studies of the facial expressions of emotion were compromised by technical problems with visible video analysis and electromyography in experimental settings. These have only recently been overcome. There have been new developments in the field of automated computerized facial recognition; allowing real-time identification of facial expression in social environments. This review addresses three approaches to measuring facial expression of emotion and describes their specific contributions to understanding emotion in the healthy population and in persons with mental illness. Despite recent progress, studies on human emotions have been hindered by the lack of consensus on an emotion theory suited to examining the dynamic aspects of emotion and its expression. Studying expression of emotion in patients with mental health conditions for diagnostic and therapeutic purposes will profit from theoretical and methodological progress.
Unseen stimuli modulate conscious visual experience: evidence from inter-hemispheric summation.
de Gelder, B; Pourtois, G; van Raamsdonk, M; Vroomen, J; Weiskrantz, L
2001-02-12
Emotional facial expression can be discriminated despite extensive lesions of striate cortex. Here we report differential performance with recognition of facial stimuli in the intact visual field depending on simultaneous presentation of congruent or incongruent stimuli in the blind field. Three experiments were based on inter-hemispheric summation. Redundant stimulation in the blind field led to shorter latencies for stimulus detection in the intact field. Recognition of the expression of a half-face expression in the intact field was faster when the other half of the face presented to the blind field had a congruent expression. Finally, responses to the expression of whole faces to the intact field were delayed for incongruent facial expressions presented in the blind field. These results indicate that the neuro-anatomical pathways (extra-striate cortical and sub-cortical) sustaining inter-hemispheric summation can operate in the absence of striate cortex.
A Role for TLR4 in Clostridium difficile Infection and the Recognition of Surface Layer Proteins
Ryan, Anthony; Lynch, Mark; Smith, Sinead M.; Amu, Sylvie; Nel, Hendrik J.; McCoy, Claire E.; Dowling, Jennifer K.; Draper, Eve; O'Reilly, Vincent; McCarthy, Ciara; O'Brien, Julie; Ní Eidhin, Déirdre; O'Connell, Mary J.; Keogh, Brian; Morton, Charles O.; Rogers, Thomas R.; Fallon, Padraic G.; O'Neill, Luke A.
2011-01-01
Clostridium difficile is the etiological agent of antibiotic-associated diarrhoea (AAD) and pseudomembranous colitis in humans. The role of the surface layer proteins (SLPs) in this disease has not yet been fully explored. The aim of this study was to investigate a role for SLPs in the recognition of C. difficile and the subsequent activation of the immune system. Bone marrow derived dendritic cells (DCs) exposed to SLPs were assessed for production of inflammatory cytokines, expression of cell surface markers and their ability to generate T helper (Th) cell responses. DCs isolated from C3H/HeN and C3H/HeJ mice were used in order to examine whether SLPs are recognised by TLR4. The role of TLR4 in infection was examined in TLR4-deficient mice. SLPs induced maturation of DCs characterised by production of IL-12, TNFα and IL-10 and expression of MHC class II, CD40, CD80 and CD86. Furthermore, SLP-activated DCs generated Th cells producing IFNγ and IL-17. SLPs were unable to activate DCs isolated from TLR4-mutant C3H/HeJ mice and failed to induce a subsequent Th cell response. TLR4−/− and Myd88−/−, but not TRIF−/− mice were more susceptible than wild-type mice to C. difficile infection. Furthermore, SLPs activated NFκB, but not IRF3, downstream of TLR4. Our results indicate that SLPs isolated from C. difficile can activate innate and adaptive immunity and that these effects are mediated by TLR4, with TLR4 having a functional role in experimental C. difficile infection. This suggests an important role for SLPs in the recognition of C. difficile by the immune system. PMID:21738466
Tryptophan depletion decreases the recognition of fear in female volunteers.
Harmer, C J; Rogers, R D; Tunbridge, E; Cowen, P J; Goodwin, G M
2003-06-01
Serotonergic processes have been implicated in the modulation of fear conditioning in humans, postulated to occur at the level of the amygdala. The processing of other fear-relevant cues, such as facial expressions, has also been associated with amygdala function, but an effect of serotonin depletion on these processes has not been assessed. The present study investigated the effects of reducing serotonin function, using acute tryptophan depletion, on the recognition of basic facial expressions of emotions in healthy male and female volunteers. A double-blind between-groups design was used, with volunteers being randomly allocated to receive an amino acid drink specifically lacking tryptophan or a control mixture containing a balanced mixture of these amino acids. Participants were given a facial expression recognition task 5 h after drink administration. This task featured examples of six basic emotions (fear, anger, disgust, surprise, sadness and happiness) that had been morphed between each full emotion and neutral in 10% steps. As a control, volunteers were given a famous face classification task matched in terms of response selection and difficulty level. Tryptophan depletion significantly impaired the recognition of fearful facial expressions in female, but not male, volunteers. This was specific since recognition of other basic emotions was comparable in the two groups. There was also no effect of tryptophan depletion on the classification of famous faces or on subjective state ratings of mood or anxiety. These results confirm a role for serotonin in the processing of fear related cues, and in line with previous findings also suggest greater effects of tryptophan depletion in female volunteers. Although acute tryptophan depletion does not typically affect mood in healthy subjects, the present results suggest that subtle changes in the processing of emotional material may occur with this manipulation of serotonin function.
Culture but not gender modulates amygdala activation during explicit emotion recognition.
Derntl, Birgit; Habel, Ute; Robinson, Simon; Windischberger, Christian; Kryspin-Exner, Ilse; Gur, Ruben C; Moser, Ewald
2012-05-29
Mounting evidence indicates that humans have significant difficulties in understanding emotional expressions from individuals of different ethnic backgrounds, leading to reduced recognition accuracy and stronger amygdala activation. However, the impact of gender on the behavioral and neural reactions during the initial phase of cultural assimilation has not been addressed. Therefore, we investigated 24 Asians students (12 females) and 24 age-matched European students (12 females) during an explicit emotion recognition task, using Caucasian facial expressions only, on a high-field MRI scanner. Analysis of functional data revealed bilateral amygdala activation to emotional expressions in Asian and European subjects. However, in the Asian sample, a stronger response of the amygdala emerged and was paralleled by reduced recognition accuracy, particularly for angry male faces. Moreover, no significant gender difference emerged. We also observed a significant inverse correlation between duration of stay and amygdala activation. In this study we investigated the "alien-effect" as an initial problem during cultural assimilation and examined this effect on a behavioral and neural level. This study has revealed bilateral amygdala activation to emotional expressions in Asian and European females and males. In the Asian sample, a stronger response of the amygdala bilaterally was observed and this was paralleled by reduced performance, especially for anger and disgust depicted by male expressions. However, no gender difference occurred. Taken together, while gender exerts only a subtle effect, culture and duration of stay as well as gender of poser are shown to be relevant factors for emotion processing, influencing not only behavioral but also neural responses in female and male immigrants.
Culture but not gender modulates amygdala activation during explicit emotion recognition
2012-01-01
Background Mounting evidence indicates that humans have significant difficulties in understanding emotional expressions from individuals of different ethnic backgrounds, leading to reduced recognition accuracy and stronger amygdala activation. However, the impact of gender on the behavioral and neural reactions during the initial phase of cultural assimilation has not been addressed. Therefore, we investigated 24 Asians students (12 females) and 24 age-matched European students (12 females) during an explicit emotion recognition task, using Caucasian facial expressions only, on a high-field MRI scanner. Results Analysis of functional data revealed bilateral amygdala activation to emotional expressions in Asian and European subjects. However, in the Asian sample, a stronger response of the amygdala emerged and was paralleled by reduced recognition accuracy, particularly for angry male faces. Moreover, no significant gender difference emerged. We also observed a significant inverse correlation between duration of stay and amygdala activation. Conclusion In this study we investigated the “alien-effect” as an initial problem during cultural assimilation and examined this effect on a behavioral and neural level. This study has revealed bilateral amygdala activation to emotional expressions in Asian and European females and males. In the Asian sample, a stronger response of the amygdala bilaterally was observed and this was paralleled by reduced performance, especially for anger and disgust depicted by male expressions. However, no gender difference occurred. Taken together, while gender exerts only a subtle effect, culture and duration of stay as well as gender of poser are shown to be relevant factors for emotion processing, influencing not only behavioral but also neural responses in female and male immigrants. PMID:22642400
Uyangaa, Erdenebileg; Choi, Jin Young; Patil, Ajit Mahadev; Hossain, Ferdaus Mohd Altaf; Park, Sung OK; Kim, Bumseok; Kim, Koanhoi; Eo, Seong Kug
2018-01-01
The importance of TLR2 and TLR9 in the recognition of infection with herpes simplex virus (HSV) and HSV-caused diseases has been described, but some discrepancies remain concerning the benefits of these responses. Moreover, the impact of TLR2/9 on innate and adaptive immune responses within relevant mucosal tissues has not been elucidated using natural mucosal infection model of HSV. Here, we demonstrate that dual TLR2/9 recognition is essential to provide resistance against mucosal infection with HSV via an intravaginal route. Dual TLR2/9 ablation resulted in the highly enhanced mortality with exacerbated symptoms of encephalitis compared with TLR2 or TLR9 deficiency alone, coinciding with highly increased viral load in central nervous system tissues. TLR2 appeared to play a minor role in providing resistance against mucosal infection with HSV, since TLR2-ablated mice showed higher survival rate compared with TLR9-ablated mice. Also, the high mortality in dual TLR2/9-ablated mice was closely associated with the reduction in early monocyte and NK cell infiltration in the vaginal tract (VT), which was likely to correlate with low expression of cytokines and CCR2 ligands (CCL2 and CCL7). More interestingly, our data revealed that dual TLR2/9 recognition of HSV infection plays an important role in the functional maturation of TNF-α and iNOS-producing dendritic cells (Tip-DCs) from monocytes as well as NK cell activation in VT. TLR2/9-dependent maturation of Tip-DCs from monocytes appeared to specifically present cognate Ag, which effectively provided functional effector CD4+ and CD8+ T cells specific for HSV Ag in VT and its draining lymph nodes. TLR2/9 expressed in monocytes was likely to directly facilitate Tip-DC-like features after HSV infection. Also, dual TLR2/9 recognition of HSV infection directly activated NK cells without the aid of dendritic cells through activation of p38 MAPK pathway. Taken together, these results indicate that dual TLR2/9 recognition plays a critical role in providing resistance against mucosal infection with HSV, which may involve a direct regulation of Tip-DCs and NK cells in VT. Therefore, our data provide a more detailed understanding of TLR2/9 role in conferring antiviral immunity within relevant mucosal tissues after mucosal infection with HSV. PMID:29760708
Uyangaa, Erdenebileg; Choi, Jin Young; Patil, Ajit Mahadev; Hossain, Ferdaus Mohd Altaf; Park, Sung Ok; Kim, Bumseok; Kim, Koanhoi; Eo, Seong Kug
2018-01-01
The importance of TLR2 and TLR9 in the recognition of infection with herpes simplex virus (HSV) and HSV-caused diseases has been described, but some discrepancies remain concerning the benefits of these responses. Moreover, the impact of TLR2/9 on innate and adaptive immune responses within relevant mucosal tissues has not been elucidated using natural mucosal infection model of HSV. Here, we demonstrate that dual TLR2/9 recognition is essential to provide resistance against mucosal infection with HSV via an intravaginal route. Dual TLR2/9 ablation resulted in the highly enhanced mortality with exacerbated symptoms of encephalitis compared with TLR2 or TLR9 deficiency alone, coinciding with highly increased viral load in central nervous system tissues. TLR2 appeared to play a minor role in providing resistance against mucosal infection with HSV, since TLR2-ablated mice showed higher survival rate compared with TLR9-ablated mice. Also, the high mortality in dual TLR2/9-ablated mice was closely associated with the reduction in early monocyte and NK cell infiltration in the vaginal tract (VT), which was likely to correlate with low expression of cytokines and CCR2 ligands (CCL2 and CCL7). More interestingly, our data revealed that dual TLR2/9 recognition of HSV infection plays an important role in the functional maturation of TNF-α and iNOS-producing dendritic cells (Tip-DCs) from monocytes as well as NK cell activation in VT. TLR2/9-dependent maturation of Tip-DCs from monocytes appeared to specifically present cognate Ag, which effectively provided functional effector CD4 + and CD8 + T cells specific for HSV Ag in VT and its draining lymph nodes. TLR2/9 expressed in monocytes was likely to directly facilitate Tip-DC-like features after HSV infection. Also, dual TLR2/9 recognition of HSV infection directly activated NK cells without the aid of dendritic cells through activation of p38 MAPK pathway. Taken together, these results indicate that dual TLR2/9 recognition plays a critical role in providing resistance against mucosal infection with HSV, which may involve a direct regulation of Tip-DCs and NK cells in VT. Therefore, our data provide a more detailed understanding of TLR2/9 role in conferring antiviral immunity within relevant mucosal tissues after mucosal infection with HSV.
Lu, Lingling; Mamiya, Takayoshi; Lu, Ping; Toriumi, Kazuya; Mouri, Akihiro; Hiramatsu, Masayuki; Kim, Hyoung-Chun; Zou, Li-Bo; Nagai, Taku; Nabeshima, Toshitaka
2010-08-01
Several studies have shown the disruptive effects of non-competitive N-methyl-d-aspartate (NMDA) receptor antagonists on neurobehavioural development. Based on the neurodevelopment hypothesis of schizophrenia, there is growing interest in animal models treated with NMDA antagonists at developing stages to investigate the pathogenesis of psychological disturbances in humans. Previous studies have reported that perinatal treatment with phencyclidine (PCP) impairs the development of neuronal systems and induces schizophrenia-like behaviour. However, the adverse effects of prenatal exposure to PCP on behaviour and the function of NMDA receptors are not well understood. This study investigated the long-term effects of prenatal exposure to PCP in mice. The prenatal PCP-treated mice showed hypersensitivity to a low dose of PCP in locomotor activity and impairment of recognition memory in the novel object recognition test at age 7 wk. Meanwhile, the prenatal exposure reduced the phosphorylation of NR1, although it increased the expression of NR1 itself. Furthermore, these behavioural changes were attenuated by atypical antipsychotic treatment. Taken together, prenatal exposure to PCP produced long-lasting behavioural deficits, accompanied by the abnormal expression and dysfunction of NMDA receptors in postpubertal mice. It is worth investigating the influences of disrupted NMDA receptors during the prenatal period on behaviour in later life.
An Effective 3D Ear Acquisition System
Liu, Yahui; Lu, Guangming; Zhang, David
2015-01-01
The human ear is a new feature in biometrics that has several merits over the more common face, fingerprint and iris biometrics. It can be easily captured from a distance without a fully cooperative subject. Also, the ear has a relatively stable structure that does not change much with the age and facial expressions. In this paper, we present a novel method of 3D ear acquisition system by using triangulation imaging principle, and the experiment results show that this design is efficient and can be used for ear recognition. PMID:26061553
An Effective 3D Ear Acquisition System.
Liu, Yahui; Lu, Guangming; Zhang, David
2015-01-01
The human ear is a new feature in biometrics that has several merits over the more common face, fingerprint and iris biometrics. It can be easily captured from a distance without a fully cooperative subject. Also, the ear has a relatively stable structure that does not change much with the age and facial expressions. In this paper, we present a novel method of 3D ear acquisition system by using triangulation imaging principle, and the experiment results show that this design is efficient and can be used for ear recognition.
The Effects of Cognitive Reappraisal and Expressive Suppression on Memory of Emotional Pictures.
Wang, Yan Mei; Chen, Jie; Han, Ben Yue
2017-01-01
In the field of emotion research, the influence of emotion regulation strategies on memory with emotional materials has been widely discussed in recent years. However, existing studies have focused exclusively on regulating negative emotion but not positive emotion. Therefore, in the present study, we investigated the influence of emotion regulation strategies for positive emotion on memory. One hundred and twenty college students were selected as participants. Emotional pictures (positive, negative and neutral) were selected from Chinese Affective Picture System (CAPS) as experimental materials. We employed a mixed, 4 (emotion regulation strategies: cognitive up-regulation, cognitive down-regulation, expressive suppression, passive viewing) × 3 (emotional pictures: positive, neutral, negative) experimental design. We investigated the influences of different emotion regulation strategies on memory performance, using free recall and recognition tasks with pictures varying in emotional content. The results showed that recognition and free recall memory performance of the cognitive reappraisal groups (up-regulation and down-regulation) were both better than that of the passive viewing group for all emotional pictures. No significant differences were reported in the two kinds of memory scores between the expressive suppression and passive viewing groups. The results also showed that the memory performance with the emotional pictures differed according to the form of memory test. For the recognition test, participants performed better with positive images than with neutral images. Free recall scores with negative images were higher than those with neutral images. These results suggest that both cognitive reappraisal regulation strategies (up-regulation and down-regulation) promoted explicit memories of the emotional content of stimuli, and the form of memory test influenced performance with emotional pictures.
Wang, Mengqiang; Wang, Lingling; Huang, Mengmeng; Yi, Qilin; Guo, Ying; Gai, Yunchao; Wang, Hao; Zhang, Huan; Song, Linsheng
2016-08-01
Galectins are a family of β-galactoside binding lectins that function as pattern recognition receptors (PRRs) in innate immune system of both vertebrates and invertebrates. The cDNA of Chinese mitten crab Eriocheir sinensis galectin (designated as EsGal) was cloned via rapid amplification of cDNA ends (RACE) technique based on expressed sequence tags (ESTs) analysis. The full-length cDNA of EsGal was 999 bp. Its open reading frame encoded a polypeptide of 218 amino acids containing a GLECT/Gal-bind_lectin domain and a proline/glycine rich low complexity region. The deduced amino acid sequence and domain organization of EsGal were highly similar to those of crustacean galectins. The mRNA transcripts of EsGal were found to be constitutively expressed in a wide range of tissues and mainly in hepatopancreas, gill and haemocytes. The mRNA expression level of EsGal increased rapidly and significantly after crabs were stimulated by different microbes. The recombinant EsGal (rEsGal) could bind various pathogen-associated molecular patterns (PAMPs), including lipopolysaccharide (LPS), peptidoglycan (PGN) and glucan (GLU), and exhibited strong activity to agglutinate Escherichia coli, Vibrio anguillarum, Bacillus subtilis, Micrococcus luteus, Staphylococcus aureus and Pichia pastoris, and such agglutinating activity could be inhibited by both d-galactose and α-lactose. The in vitro encapsulation assay revealed that rEsGal could enhance the encapsulation of haemocytes towards agarose beads. These results collectively suggested that EsGal played crucial roles in the immune recognition and elimination of pathogens and contributed to the innate immune response against various microbes in crabs. Copyright © 2016 Elsevier Ltd. All rights reserved.
Matsubara, Victor H.; Ishikawa, Karin H.; Ando-Suguimoto, Ellen S.; Bueno-Silva, Bruno; Nakamae, Atlas E. M.; Mayer, Marcia P. A.
2017-01-01
Probiotics are live microorganisms that confer benefits to the host health. The infection rate of potentially pathogenic organisms such as Candida albicans, the most common agent associated with mucosal candidiasis, can be reduced by probiotics. However, the mechanisms by which the probiotics interfere with the immune system are largely unknown. We evaluated the effect of probiotic bacteria on C. albicans challenged human macrophages. Macrophages were pretreated with lactobacilli alone (Lactobacillus rhamnosus LR32, Lactobacillus casei L324m, or Lactobacillus acidophilus NCFM) or associated with Escherichia coli lipopolysaccharide (LPS), followed by the challenge with C. albicans or LPS in a co-culture assay. The expression of pattern-recognition receptors genes (CLE7A, TLR2, and TLR4) was determined by RT-qPCR, and dectin-1 reduced levels were confirmed by flow cytometry. The cytokine profile was determined by ELISA using the macrophage cell supernatant. Overall probiotic lactobacilli down-regulated the transcription of CLEC7A (p < 0.05), resulting in the decreased expression of dectin-1 on probiotic pretreated macrophages. The tested Lactobacillus species down-regulated TLR4, and increased TLR2 mRNA levels in macrophages challenged with C. albicans. The cytokines profile of macrophages challenged with C. albicans or LPS were altered by the probiotics, which generally led to increased levels of IL-10 and IL-1β, and reduction of IL-12 production by macrophages (p < 0.05). Our data suggest that probiotic lactobacilli impair the recognition of PAMPs by macrophages, and alter the production of pro/anti-inflammatory cytokines, thus modulating inflammation. PMID:29238325
Araújo, Welington Luiz; Santos, Daiene Souza; Dini-Andreote, Francisco; Salgueiro-Londoño, Jennifer Katherine; Camargo-Neves, Aline Aparecida; Andreote, Fernando Dini; Dourado, Manuella Nóbrega
2015-10-01
The genus Methylobacterium is composed of pink-pigmented methylotrophic bacterial species that are widespread in natural environments, such as soils, stream water and plants. When in association with plants, this genus colonizes the host plant epiphytically and/or endophytically. This association is known to promote plant growth, induce plant systemic resistance and inhibit plant infection by phytopathogens. In the present study, we focused on evaluating the colonization of soybean seedling-roots by Methylobacterium mesophilicum strain SR1.6/6. We focused on the identification of the key genes involved in the initial step of soybean colonization by methylotrophic bacteria, which includes the plant exudate recognition and adaptation by planktonic bacteria. Visualization by scanning electron microscopy revealed that M. mesophilicum SR1.6/6 colonizes soybean roots surface effectively at 48 h after inoculation, suggesting a mechanism for root recognition and adaptation before this period. The colonization proceeds by the development of a mature biofilm on roots at 96 h after inoculation. Transcriptomic analysis of the planktonic bacteria (with plant) revealed the expression of several genes involved in membrane transport, thus confirming an initial metabolic activation of bacterial responses when in the presence of plant root exudates. Moreover, antioxidant genes were mostly expressed during the interaction with the plant exudates. Further evaluation of stress- and methylotrophic-related genes expression by qPCR showed that glutathione peroxidase and glutathione synthetase genes were up-regulated during the Methylobacterium-soybean interaction. These findings support that glutathione (GSH) is potentially a key molecule involved in cellular detoxification during plant root colonization. In addition to methylotrophic metabolism, antioxidant genes, mainly glutathione-related genes, play a key role during soybean exudate recognition and adaptation, the first step in bacterial colonization.
Kosonogov, Vladimir; Titova, Alisa; Vorobyeva, Elena
2015-01-01
The current study addressed the hypothesis that empathy and the restriction of facial muscles of observers can influence recognition of emotional facial expressions. A sample of 74 participants recognized the subjective onset of emotional facial expressions (anger, disgust, fear, happiness, sadness, surprise, and neutral) in a series of morphed face photographs showing a gradual change (frame by frame) from one expression to another. The high-empathy (as measured by the Empathy Quotient) participants recognized emotional facial expressions at earlier photographs from the series than did low-empathy ones, but there was no difference in the exploration time. Restriction of facial muscles of observers (with plasters and a stick in mouth) did not influence the responses. We discuss these findings in the context of the embodied simulation theory and previous data on empathy.
Deng, Meihong; Loughran, Patricia; Gibson, Gregory; Sodhi, Chhinder; Watkins, Simon; Hackam, David
2013-01-01
The morbidity associated with bacterial sepsis is the result of host immune responses to pathogens, which are dependent on pathogen recognition by pattern recognition receptors, such as TLR4. TLR4 is expressed on a range of cell types, yet the mechanisms by which cell-specific functions of TLR4 lead to an integrated sepsis response are poorly understood. To address this, we generated mice in which TLR4 was specifically deleted from myeloid cells (LysMTLR4KO) or hepatocytes (HCTLR4KO) and then determined survival, bacterial counts, host inflammatory responses, and organ injury in a model of cecal ligation and puncture (CLP), with or without antibiotics. LysM-TLR4 was required for phagocytosis and efficient bacterial clearance in the absence of antibiotics. Survival, the magnitude of the systemic and local inflammatory responses, and liver damage were associated with bacterial levels. HCTLR4 was required for efficient LPS clearance from the circulation, and deletion of HCTLR4 was associated with enhanced macrophage phagocytosis, lower bacterial levels, and improved survival in CLP without antibiotics. Antibiotic administration during CLP revealed an important role for hepatocyte LPS clearance in limiting sepsis-induced inflammation and organ injury. Our work defines cell type–selective roles for TLR4 in coordinating complex immune responses to bacterial sepsis and suggests that future strategies for modulating microbial molecule recognition should account for varying roles of pattern recognition receptors in multiple cell populations. PMID:23562812
Bekinschtein, Pedro; Renner, Maria Constanza; Gonzalez, Maria Carolina; Weisstaub, Noelia
2013-10-02
Often, retrieval cues are not uniquely related to one specific memory, which could lead to memory interference. Controlling interference is particularly important during episodic memory retrieval or when remembering specific events in a spatiotemporal context. Despite a clear involvement of prefrontal cortex (PFC) in episodic memory in human studies, information regarding the mechanisms and neurotransmitter systems in PFC involved in memory is scarce. Although the serotoninergic system has been linked to PFC functionality and modulation, its role in memory processing is poorly understood. We hypothesized that the serotoninergic system in PFC, in particular the 5-HT2A receptor (5-HT2AR) could have a role in the control of memory retrieval. In this work we used different versions of the object recognition task in rats to study the role of the serotoninergic modulation in the medial PFC (mPFC) in memory retrieval. We found that blockade of 5-HT2AR in mPFC affects retrieval of an object in context memory in a spontaneous novelty preference task, while sparing single-item recognition memory. We also determined that 5-HT2ARs in mPFC are required for hippocampal-mPFC interaction during retrieval of this type of memory, suggesting that the mPFC controls the expression of memory traces stored in the hippocampus biasing retrieval to the most relevant one.
Busquets-Garcia, Arnau; Gomis-González, Maria; Salgado-Mendialdúa, Victòria; Galera-López, Lorena; Puighermanal, Emma; Martín-García, Elena; Maldonado, Rafael; Ozaita, Andrés
2018-04-01
Cannabis affects cognitive performance through the activation of the endocannabinoid system, and the molecular mechanisms involved in this process are poorly understood. Using the novel object-recognition memory test in mice, we found that the main psychoactive component of cannabis, delta9-tetrahydrocannabinol (THC), alters short-term object-recognition memory specifically involving protein kinase C (PKC)-dependent signaling. Indeed, the systemic or intra-hippocampal pre-treatment with the PKC inhibitors prevented the short-term, but not the long-term, memory impairment induced by THC. In contrast, systemic pre-treatment with mammalian target of rapamycin complex 1 inhibitors, known to block the amnesic-like effects of THC on long-term memory, did not modify such a short-term cognitive deficit. Immunoblot analysis revealed a transient increase in PKC signaling activity in the hippocampus after THC treatment. Thus, THC administration induced the phosphorylation of a specific Ser residue in the hydrophobic-motif at the C-terminal tail of several PKC isoforms. This significant immunoreactive band that paralleled cognitive performance did not match in size with the major PKC isoforms expressed in the hippocampus except for PKCθ. Moreover, THC transiently enhanced the phosphorylation of the postsynaptic calmodulin-binding protein neurogranin in a PKC dependent manner. These data demonstrate that THC alters short-term object-recognition memory through hippocampal PKC/neurogranin signaling.
Innate immunity and the sensing of infection, damage and danger in the female genital tract.
Sheldon, Iain Martin; Owens, Siân-Eleri; Turner, Matthew Lloyd
2017-02-01
Tissue homeostasis in the female genital tract is challenged by infection, damage, and even physiological events during reproductive cycles. We propose that the evolutionarily ancient system of innate immunity is sufficient to sense and respond to danger in the non-pregnant female genital tract. Innate immunity produces a rapidly inducible, non-specific response when cells sense danger. Here we provide a primer on innate immunity and discuss what is known about how danger signals are sensed in the endometrium and ovary, the impact of inflammatory responses on reproduction, and how endocrinology and innate immunity are integrated. Endometrial epithelial and stromal cells, and ovarian granulosa cells express pattern recognition receptors, similar to cells of the innate immune system. These pattern recognition receptors, such as the Toll-like receptors, bind pathogen-associated or damage-associated molecular patterns. Activation of pattern recognition receptors leads to inflammation, recruitment of immune cells from the peripheral circulation, and phagocytosis. Although the inflammatory response helps maintain or restore endometrial health, there may also be negative consequences for fertility, including perturbation of oocyte competence. The intensity of the inflammatory response reflects the balance between the level of danger and the systems that regulate innate immunity, including the endocrine environment. Understanding innate immunity is important because disease and inappropriate inflammatory responses in the endometrium or ovary cause infertility. Copyright © 2016 Elsevier Ireland Ltd. All rights reserved.
Codebook-based electrooculography data analysis towards cognitive activity recognition.
Lagodzinski, P; Shirahama, K; Grzegorzek, M
2018-04-01
With the advancement in mobile/wearable technology, people started to use a variety of sensing devices to track their daily activities as well as health and fitness conditions in order to improve the quality of life. This work addresses an idea of eye movement analysis, which due to the strong correlation with cognitive tasks can be successfully utilized in activity recognition. Eye movements are recorded using an electrooculographic (EOG) system built into the frames of glasses, which can be worn more unobtrusively and comfortably than other devices. Since the obtained information is low-level sensor data expressed as a sequence representing values in constant intervals (100 Hz), the cognitive activity recognition problem is formulated as sequence classification. However, it is unclear what kind of features are useful for accurate cognitive activity recognition. Thus, a machine learning algorithm like a codebook approach is applied, which instead of focusing on feature engineering is using a distribution of characteristic subsequences (codewords) to describe sequences of recorded EOG data, where the codewords are obtained by clustering a large number of subsequences. Further, statistical analysis of the codeword distribution results in discovering features which are characteristic to a certain activity class. Experimental results demonstrate good accuracy of the codebook-based cognitive activity recognition reflecting the effective usage of the codewords. Copyright © 2017 Elsevier Ltd. All rights reserved.
Face photo-sketch synthesis and recognition.
Wang, Xiaogang; Tang, Xiaoou
2009-11-01
In this paper, we propose a novel face photo-sketch synthesis and recognition method using a multiscale Markov Random Fields (MRF) model. Our system has three components: 1) given a face photo, synthesizing a sketch drawing; 2) given a face sketch drawing, synthesizing a photo; and 3) searching for face photos in the database based on a query sketch drawn by an artist. It has useful applications for both digital entertainment and law enforcement. We assume that faces to be studied are in a frontal pose, with normal lighting and neutral expression, and have no occlusions. To synthesize sketch/photo images, the face region is divided into overlapping patches for learning. The size of the patches decides the scale of local face structures to be learned. From a training set which contains photo-sketch pairs, the joint photo-sketch model is learned at multiple scales using a multiscale MRF model. By transforming a face photo to a sketch (or transforming a sketch to a photo), the difference between photos and sketches is significantly reduced, thus allowing effective matching between the two in face sketch recognition. After the photo-sketch transformation, in principle, most of the proposed face photo recognition approaches can be applied to face sketch recognition in a straightforward way. Extensive experiments are conducted on a face sketch database including 606 faces, which can be downloaded from our Web site (http://mmlab.ie.cuhk.edu.hk/facesketch.html).
Varga, Dániel; Herédi, Judit; Kánvási, Zita; Ruszka, Marian; Kis, Zsolt; Ono, Etsuro; Iwamori, Naoki; Iwamori, Tokuko; Takakuwa, Hiroki; Vécsei, László; Toldi, József; Gellért, Levente
2015-01-01
L-Kynurenine (L-KYN) is a central metabolite of tryptophan degradation through the kynurenine pathway (KP). The systemic administration of L-KYN sulfate (L-KYNs) leads to a rapid elevation of the neuroactive KP metabolite kynurenic acid (KYNA). An elevated level of KYNA may have multiple effects on the synaptic transmission, resulting in complex behavioral changes, such as hypoactivity or spatial working memory deficits. These results emerged from studies that focused on rats, after low-dose L-KYNs treatment. However, in several studies neuroprotection was achieved through the administration of high-dose L-KYNs. In the present study, our aim was to investigate whether the systemic administration of a high dose of L-KYNs (300 mg/bwkg; i.p.) would produce alterations in behavioral tasks (open field or object recognition) in C57Bl/6j mice. To evaluate the changes in neuronal activity after L-KYNs treatment, in a separate group of animals we estimated c-Fos expression levels in the corresponding subcortical brain areas. The L-KYNs treatment did not affect the general ambulatory activity of C57Bl/6j mice, whereas it altered their moving patterns, elevating the movement velocity and resting time. Additionally, it seemed to increase anxiety-like behavior, as peripheral zone preference of the open field arena emerged and the rearing activity was attenuated. The treatment also completely abolished the formation of object recognition memory and resulted in decreases in the number of c-Fos-immunopositive-cells in the dorsal part of the striatum and in the CA1 pyramidal cell layer of the hippocampus. We conclude that a single exposure to L-KYNs leads to behavioral disturbances, which might be related to the altered basal c-Fos protein expression in C57Bl/6j mice.
Yao, Shih-Ying; Bull, Rebecca; Khng, Kiat Hui; Rahim, Anisa
2018-01-01
Understanding a child's ability to decode emotion expressions is important to allow early interventions for potential difficulties in social and emotional functioning. This study applied the Rasch model to investigate the psychometric properties of the NEPSY-II Affect Recognition subtest, a U.S. normed measure for 3-16 year olds which assesses the ability to recognize facial expressions of emotion. Data were collected from 1222 children attending preschools in Singapore. We first performed the Rasch analysis with the raw item data, and examined the technical qualities and difficulty pattern of the studied items. We subsequently investigated the relation of the estimated affect recognition ability from the Rasch analysis to a teacher-reported measure of a child's behaviors, emotions, and relationships. Potential gender differences were also examined. The Rasch model fits our data well. Also, the NEPSY-II Affect Recognition subtest was found to have reasonable technical qualities, expected item difficulty pattern, and desired association with the external measure of children's behaviors, emotions, and relationships for both boys and girls. Overall, findings from this study suggest that the NEPSY-II Affect Recognition subtest is a promising measure of young children's affect recognition ability. Suggestions for future test improvement and research were discussed.
Impaired perception of facial emotion in developmental prosopagnosia.
Biotti, Federica; Cook, Richard
2016-08-01
Developmental prosopagnosia (DP) is a neurodevelopmental condition characterised by difficulties recognising faces. Despite severe difficulties recognising facial identity, expression recognition is typically thought to be intact in DP; case studies have described individuals who are able to correctly label photographic displays of facial emotion, and no group differences have been reported. This pattern of deficits suggests a locus of impairment relatively late in the face processing stream, after the divergence of expression and identity analysis pathways. To date, however, there has been little attempt to investigate emotion recognition systematically in a large sample of developmental prosopagnosics using sensitive tests. In the present study, we describe three complementary experiments that examine emotion recognition in a sample of 17 developmental prosopagnosics. In Experiment 1, we investigated observers' ability to make binary classifications of whole-face expression stimuli drawn from morph continua. In Experiment 2, observers judged facial emotion using only the eye-region (the rest of the face was occluded). Analyses of both experiments revealed diminished ability to classify facial expressions in our sample of developmental prosopagnosics, relative to typical observers. Imprecise expression categorisation was particularly evident in those individuals exhibiting apperceptive profiles, associated with problems encoding facial shape accurately. Having split the sample of prosopagnosics into apperceptive and non-apperceptive subgroups, only the apperceptive prosopagnosics were impaired relative to typical observers. In our third experiment, we examined the ability of observers' to classify the emotion present within segments of vocal affect. Despite difficulties judging facial emotion, the prosopagnosics exhibited excellent recognition of vocal affect. Contrary to the prevailing view, our results suggest that many prosopagnosics do experience difficulties classifying expressions, particularly those with apperceptive profiles. These individuals may have difficulties forming view-invariant structural descriptions at an early stage in the face processing stream, before identity and expression pathways diverge. Copyright © 2016 Elsevier Ltd. All rights reserved.
Hiscock, Simon J; McInnis, Stephanie M
2003-12-01
Many hermaphrodite flowering plants avoid self-fertilization through genetic systems of self-incompatibility (SI). SI allows a plant to recognize and to reject self or self-related pollen, thereby preserving its ovules for outcrossing. Genes situated at the S-locus encode the 'male' (pollen) and 'female' (pistil) recognition determinants of SI. In sporophytic SI (SSI) the male determinant is expressed in the diploid anther, therefore haploid pollen grains behave with a diploid S phenotype. In Brassica, the male and the female determinants of SSI have been identified as a peptide ligand and its cognate receptor, respectively, and recent studies have identified downstream signalling molecules involved in pollen rejection. It now needs to be established whether the Brassica mechanism is universal in species with SSI, or unique to the Brassicaceae.
Pathways for smiling, disgust and fear recognition in blindsight patients.
Gerbella, Marzio; Caruana, Fausto; Rizzolatti, Giacomo
2017-08-31
The aim of the present review is to discuss the localization of circuits that allow recognition of emotional facial expressions in blindsight patients. Because recognition of facial expressions is function of different centers, and their localization is not always clear, we decided to discuss here three emotional facial expression - smiling, disgust, and fear - whose anatomical localization in the pregenual sector of the anterior cingulate cortex (pACC), anterior insula (AI), and amygdala, respectively, is well established. We examined, then, the possible pathways that may convey affective visual information to these centers following lesions of V1. We concluded that the pathway leading to pACC, AI, and amygdala involves the deep layers of the superior colliculus, the medial pulvinar, and the superior temporal sulcus region. We suggest that this visual pathway provides an image of the observed affective faces, which, although deteriorated, is sufficient to determine some overt behavior, but not to provide conscious experience of the presented stimuli. Copyright © 2017 Elsevier Ltd. All rights reserved.
Emotion through locomotion: gender impact.
Krüger, Samuel; Sokolov, Alexander N; Enck, Paul; Krägeloh-Mann, Ingeborg; Pavlova, Marina A
2013-01-01
Body language reading is of significance for daily life social cognition and successful social interaction, and constitutes a core component of social competence. Yet it is unclear whether our ability for body language reading is gender specific. In the present work, female and male observers had to visually recognize emotions through point-light human locomotion performed by female and male actors with different emotional expressions. For subtle emotional expressions only, males surpass females in recognition accuracy and readiness to respond to happy walking portrayed by female actors, whereas females exhibit a tendency to be better in recognition of hostile angry locomotion expressed by male actors. In contrast to widespread beliefs about female superiority in social cognition, the findings suggest that gender effects in recognition of emotions from human locomotion are modulated by emotional content of actions and opposite actor gender. In a nutshell, the study makes a further step in elucidation of gender impact on body language reading and on neurodevelopmental and psychiatric deficits in visual social cognition.
Runsaeng, Phanthipha; Thepnarong, Supattra; Rattanaporn, Onnicha; Utarabhand, Prapaporn
2015-12-01
Crustaceans are deficient in an adaptive immune system and depend solely on their innate immunity. One kind of pattern recognition proteins which plays an important role in the shrimp immunity is lectin. A new C-type lectin called FmLC2 was cloned from the stomach of the banana shrimp Fenneropenaeus merguiensis by means of RT-PCR and 5' and 3' rapid amplification of cDNA ends (RACE). Its full-length cDNA contains 1098 bp with a single open reading frame of 738 bp, encoding a peptide of 245 amino acids. The deduced amino acid sequence of FmLC2 consists of a signal peptide of 17 amino acids with a molecular mass of 28,115 Da and an isoelectric point of 6.94. The primary structure of FmLC2 comprises a single carbohydrate recognition domain (CRD) with a QPD (Gln-Pro-Asp) motif and one Ca(2+) binding site. Like other C-type lectins, its CRD structure contains a double-loop characteristic being stabilized by two conserved disulfide linkages. The mRNA expression of FmLC2 was detected specifically in the stomach and gills, less was found in the hepatopancreas. Upon inoculation of shrimp with Vibrio harveyi or white spot syndrome virus (WSSV), the FmLC2 expression either in stomach or gills was higher than in the hepatopancreas. Besides, its expression in these tissues was up-regulated to reach the highest levels at 12 or 18 h for V. harveyi or WSSV stimulation, respectively. RNAi-based silencing of FmLC2 resulted in suppression of its expression, increases in mortality when the shrimp were challenged with V. harveyi or WSSV, and the median lethal time was reduced compared with controls. These results suggest that FmLC2 may serve as receptor molecules which recognize invading bacterial and viral pathogens and thus contribute a role in the shrimp immune response. Copyright © 2015 Elsevier Ltd. All rights reserved.
Real-time face and gesture analysis for human-robot interaction
NASA Astrophysics Data System (ADS)
Wallhoff, Frank; Rehrl, Tobias; Mayer, Christoph; Radig, Bernd
2010-05-01
Human communication relies on a large number of different communication mechanisms like spoken language, facial expressions, or gestures. Facial expressions and gestures are one of the main nonverbal communication mechanisms and pass large amounts of information between human dialog partners. Therefore, to allow for intuitive human-machine interaction, a real-time capable processing and recognition of facial expressions, hand and head gestures are of great importance. We present a system that is tackling these challenges. The input features for the dynamic head gestures and facial expressions are obtained from a sophisticated three-dimensional model, which is fitted to the user in a real-time capable manner. Applying this model different kinds of information are extracted from the image data and afterwards handed over to a real-time capable data-transferring framework, the so-called Real-Time DataBase (RTDB). In addition to the head and facial-related features, also low-level image features regarding the human hand - optical flow, Hu-moments are stored into the RTDB for the evaluation process of hand gestures. In general, the input of a single camera is sufficient for the parallel evaluation of the different gestures and facial expressions. The real-time capable recognition of the dynamic hand and head gestures are performed via different Hidden Markov Models, which have proven to be a quick and real-time capable classification method. On the other hand, for the facial expressions classical decision trees or more sophisticated support vector machines are used for the classification process. These obtained results of the classification processes are again handed over to the RTDB, where other processes (like a Dialog Management Unit) can easily access them without any blocking effects. In addition, an adjustable amount of history can be stored by the RTDB buffer unit.
Rotation-invariant neural pattern recognition system with application to coin recognition.
Fukumi, M; Omatu, S; Takeda, F; Kosaka, T
1992-01-01
In pattern recognition, it is often necessary to deal with problems to classify a transformed pattern. A neural pattern recognition system which is insensitive to rotation of input pattern by various degrees is proposed. The system consists of a fixed invariance network with many slabs and a trainable multilayered network. The system was used in a rotation-invariant coin recognition problem to distinguish between a 500 yen coin and a 500 won coin. The results show that the approach works well for variable rotation pattern recognition.
Feder-Mengus, C; Ghosh, S; Weber, W P; Wyler, S; Zajac, P; Terracciano, L; Oertli, D; Heberer, M; Martin, I; Spagnoli, G C; Reschner, A
2007-04-10
Cancer cells' growth in three-dimensional (3D) architectures promotes resistance to drugs, cytokines, or irradiation. We investigated effects of 3D culture as compared to monolayers (2D) on melanoma cells' recognition by tumour-associated antigen (TAA)-specific HLA-A(*)0201-restricted cytotoxic T-lymphocytes (CTL). Culture of HBL, D10 (both HLA-A(*)0201+, TAA+) and NA8 (HLA-A(*)0201+, TAA-) melanoma cells on polyHEMA-coated plates, resulted in generation of 3D multicellular tumour spheroids (MCTS). Interferon-gamma (IFN-gamma) production by HLA-A(*)0201-restricted Melan-A/MART-1(27-35) or gp 100(280-288)-specific CTL clones served as immunorecognition marker. Co-culture with melanoma MCTS, resulted in defective TAA recognition by CTL as compared to 2D as witnessed by decreased IFN-gamma production and decreased Fas Ligand, perforin and granzyme B gene expression. A multiplicity of mechanisms were potentially involved. First, MCTS per se limit CTL capacity of recognising HLA class I restricted antigens by reducing exposed cell surfaces. Second, expression of melanoma differentiation antigens is downregulated in MCTS. Third, expression of HLA class I molecules can be downregulated in melanoma MCTS, possibly due to decreased interferon-regulating factor-1 gene expression. Fourth, lactic acid production is increased in MCTS, as compared to 2D. These data suggest that melanoma cells growing in 3D, even in the absence of immune selection, feature characteristics capable of dramatically inhibiting TAA recognition by specific CTL.
Practical automatic Arabic license plate recognition system
NASA Astrophysics Data System (ADS)
Mohammad, Khader; Agaian, Sos; Saleh, Hani
2011-02-01
Since 1970's, the need of an automatic license plate recognition system, sometimes referred as Automatic License Plate Recognition system, has been increasing. A license plate recognition system is an automatic system that is able to recognize a license plate number, extracted from image sensors. In specific, Automatic License Plate Recognition systems are being used in conjunction with various transportation systems in application areas such as law enforcement (e.g. speed limit enforcement) and commercial usages such as parking enforcement and automatic toll payment private and public entrances, border control, theft and vandalism control. Vehicle license plate recognition has been intensively studied in many countries. Due to the different types of license plates being used, the requirement of an automatic license plate recognition system is different for each country. [License plate detection using cluster run length smoothing algorithm ].Generally, an automatic license plate localization and recognition system is made up of three modules; license plate localization, character segmentation and optical character recognition modules. This paper presents an Arabic license plate recognition system that is insensitive to character size, font, shape and orientation with extremely high accuracy rate. The proposed system is based on a combination of enhancement, license plate localization, morphological processing, and feature vector extraction using the Haar transform. The performance of the system is fast due to classification of alphabet and numerals based on the license plate organization. Experimental results for license plates of two different Arab countries show an average of 99 % successful license plate localization and recognition in a total of more than 20 different images captured from a complex outdoor environment. The results run times takes less time compared to conventional and many states of art methods.
The A2iA French handwriting recognition system at the Rimes-ICDAR2011 competition
NASA Astrophysics Data System (ADS)
Menasri, Farès; Louradour, Jérôme; Bianne-Bernard, Anne-Laure; Kermorvant, Christopher
2012-01-01
This paper describes the system for the recognition of French handwriting submitted by A2iA to the competition organized at ICDAR2011 using the Rimes database. This system is composed of several recognizers based on three different recognition technologies, combined using a novel combination method. A framework multi-word recognition based on weighted finite state transducers is presented, using an explicit word segmentation, a combination of isolated word recognizers and a language model. The system was tested both for isolated word recognition and for multi-word line recognition and submitted to the RIMES-ICDAR2011 competition. This system outperformed all previously proposed systems on these tasks.
Ng, Wy Ching; Londrigan, Sarah L; Nasr, Najla; Cunningham, Anthony L; Turville, Stuart; Brooks, Andrew G; Reading, Patrick C
2016-01-01
It is well established that influenza A virus (IAV) attachment to and infection of epithelial cells is dependent on sialic acid (SIA) at the cell surface, although the specific receptors that mediate IAV entry have not been defined and multiple receptors may exist. Lec2 Chinese hamster ovary (CHO) cells are SIA deficient and resistant to IAV infection. Here we demonstrate that the expression of the C-type lectin receptor langerin in Lec2 cells (Lec2-Lg) rendered them permissive to IAV infection, as measured by replication of the viral genome, transcription of viral mRNA, and synthesis of viral proteins. Unlike SIA-dependent infection of parental CHO cells, IAV attachment and infection of Lec2-Lg cells was mediated via lectin-mediated recognition of mannose-rich glycans expressed by the viral hemagglutinin glycoprotein. Lec2 cells expressing endocytosis-defective langerin bound IAV efficiently but remained resistant to IAV infection, confirming that internalization via langerin was essential for infectious entry. Langerin-mediated infection of Lec2-Lg cells was pH and dynamin dependent, occurred via clathrin- and caveolin-mediated endocytic pathways, and utilized early (Rab5(+)) but not late (Rab7(+)) endosomes. This study is the first to demonstrate that langerin represents an authentic receptor that binds and internalizes IAV to facilitate infection. Moreover, it describes a unique experimental system to probe specific pathways and compartments involved in infectious entry following recognition of IAV by a single cell surface receptor. On the surface of host cells, sialic acid (SIA) functions as the major attachment factor for influenza A viruses (IAV). However, few studies have identified specific transmembrane receptors that bind and internalize IAV to facilitate infection. Here we identify human langerin as a transmembrane glycoprotein that can act as an attachment factor and a bone fide endocytic receptor for IAV infection. Expression of langerin by an SIA-deficient cell line resistant to IAV rendered cells permissive to infection. As langerin represented the sole receptor for IAV infection in this system, we have defined the pathways and compartments involved in infectious entry of IAV into cells following recognition by langerin. Copyright © 2015, American Society for Microbiology. All Rights Reserved.
Zhao, Qi; Murakami, Yukihisa; Tohda, Michihisa; Obi, Ryosuke; Shimada, Yutaka; Matsumoto, Kinzo
2007-04-01
We previously demonstrated that the Kampo formula chotosan (CTS) ameliorated spatial cognitive impairment via central cholinergic systems in a chronic cerebral hypoperfusion (P2VO) mouse model. In this study, the object discrimination tasks were used to determine if the ameliorative effects of CTS on P2VO-induced cognitive deficits are a characteristic pharmacological profile of this formula, with the aim of clarifying the mechanisms by which CTS enhances central cholinergic function in P2VO mice. The cholinesterase inhibitor tacrine (THA) and Kampo formula saikokeishito (SKT) were used as controls. P2VO impaired object discrimination performance in the object recognition, location, and context tests. Daily administration of CTS (750 mg/kg, p.o.) and THA (2.5 mg/kg, i.p.) improved the object discrimination deficits, whereas SKT (750 mg/kg, p.o.) did not. In ex vivo assays, tacrine but not CTS or SKT inhibited cortical cholinesterase activity. P2VO reduced the mRNA expression of m(3) and m(5) muscarinic receptors and choline acetyltransferase but not that of other muscarinic receptor subtypes in the cerebral cortex. Daily administration of CTS and THA but not SKT reversed these expression changes. These results suggest that CTS and THA improve P2VO-induced cognitive impairment by normalizing the deficit of central cholinergic systems and that the beneficial effect on P2VO-induced cognitive deficits is a distinctive pharmacological characteristic of CTS.
Šilar, Radoslav; Holátko, Jiří; Rucká, Lenka; Rapoport, Andrey; Dostálová, Hana; Kadeřábková, Pavla; Nešvera, Jan; Pátek, Miroslav
2016-09-01
Promoter activities in Corynebacterium glutamicum strains with deletions of genes encoding sigma factors of RNA polymerase suggested that transcription from some promoters is controlled by two sigma factors. To prove that different sigma factors are involved in the recognition of selected Corynebacterium glutamicum promoters, in vitro transcription system was applied. It was found that a typical housekeeping promoter Pper interacts with the alternative sigma factor σ(B) in addition to the primary sigma factor σ(A). On the other way round, the σ(B)-dependent promoter of the pqo gene that is expressed mainly in the stationary growth phase was active also with σ(A). Some promoters of genes involved in stress responses (P1clgR, P2dnaK, and P2dnaJ2) were found to be recognized by two stress-responding sigma factors, σ(H) and σ(E). In vitro transcription system thus proved to be a useful direct technique for demonstrating the overlap of different sigma factors in recognition of individual promoters in C. glutamicum.
Random-Profiles-Based 3D Face Recognition System
Joongrock, Kim; Sunjin, Yu; Sangyoun, Lee
2014-01-01
In this paper, a noble nonintrusive three-dimensional (3D) face modeling system for random-profile-based 3D face recognition is presented. Although recent two-dimensional (2D) face recognition systems can achieve a reliable recognition rate under certain conditions, their performance is limited by internal and external changes, such as illumination and pose variation. To address these issues, 3D face recognition, which uses 3D face data, has recently received much attention. However, the performance of 3D face recognition highly depends on the precision of acquired 3D face data, while also requiring more computational power and storage capacity than 2D face recognition systems. In this paper, we present a developed nonintrusive 3D face modeling system composed of a stereo vision system and an invisible near-infrared line laser, which can be directly applied to profile-based 3D face recognition. We further propose a novel random-profile-based 3D face recognition method that is memory-efficient and pose-invariant. The experimental results demonstrate that the reconstructed 3D face data consists of more than 50 k 3D point clouds and a reliable recognition rate against pose variation. PMID:24691101
ERIC Educational Resources Information Center
Bloom, Elana; Heath, Nancy
2010-01-01
Children with nonverbal learning disabilities (NVLD) have been found to be worse at recognizing facial expressions than children with verbal learning disabilities (LD) and without LD. However, little research has been done with adolescents. In addition, expressing and understanding facial expressions is yet to be studied among adolescents with LD…
Image quality assessment for video stream recognition systems
NASA Astrophysics Data System (ADS)
Chernov, Timofey S.; Razumnuy, Nikita P.; Kozharinov, Alexander S.; Nikolaev, Dmitry P.; Arlazarov, Vladimir V.
2018-04-01
Recognition and machine vision systems have long been widely used in many disciplines to automate various processes of life and industry. Input images of optical recognition systems can be subjected to a large number of different distortions, especially in uncontrolled or natural shooting conditions, which leads to unpredictable results of recognition systems, making it impossible to assess their reliability. For this reason, it is necessary to perform quality control of the input data of recognition systems, which is facilitated by modern progress in the field of image quality evaluation. In this paper, we investigate the approach to designing optical recognition systems with built-in input image quality estimation modules and feedback, for which the necessary definitions are introduced and a model for describing such systems is constructed. The efficiency of this approach is illustrated by the example of solving the problem of selecting the best frames for recognition in a video stream for a system with limited resources. Experimental results are presented for the system for identity documents recognition, showing a significant increase in the accuracy and speed of the system under simulated conditions of automatic camera focusing, leading to blurring of frames.
Brain systems for assessing the affective value of faces
Said, Christopher P.; Haxby, James V.; Todorov, Alexander
2011-01-01
Cognitive neuroscience research on facial expression recognition and face evaluation has proliferated over the past 15 years. Nevertheless, large questions remain unanswered. In this overview, we discuss the current understanding in the field, and describe what is known and what remains unknown. In §2, we describe three types of behavioural evidence that the perception of traits in neutral faces is related to the perception of facial expressions, and may rely on the same mechanisms. In §3, we discuss cortical systems for the perception of facial expressions, and argue for a partial segregation of function in the superior temporal sulcus and the fusiform gyrus. In §4, we describe the current understanding of how the brain responds to emotionally neutral faces. To resolve some of the inconsistencies in the literature, we perform a large group analysis across three different studies, and argue that one parsimonious explanation of prior findings is that faces are coded in terms of their typicality. In §5, we discuss how these two lines of research—perception of emotional expressions and face evaluation—could be integrated into a common, cognitive neuroscience framework. PMID:21536552
Bioinformatic Analysis of the Human Recombinant Iduronate 2-Sulfate Sulfatase
Morales-Álvarez, Edwin D.; Rivera-Hoyos, Claudia M.; Landázuri, Patricia; Poutou-Piñales, Raúl A.; Pedroza-Rodríguez, Aura M.
2016-01-01
Mucopolysaccharidosis type II is a human recessive disease linked to the X chromosome caused by deficiency of lysosomal enzyme Iduronate 2-Sulfate Sulfatase (IDS), which leads to accumulation of glycosaminoglycans in tissues and organs. The human enzyme has been expressed in Escherichia coli and Pichia pastoris in attempt to develop more successful expression systems that allow the production of recombinant IDS for Enzyme Replacement Therapy (ERT). However, the preservation of native signal peptide in the sequence has caused conflicts in processing and recognition in the past, which led to problems in expression and enzyme activity. With the main object being the improvement of the expression system, we eliminate the native signal peptide of human recombinant IDS. The resulting sequence showed two modified codons, thus, our study aimed to analyze computationally the nucleotide sequence of the IDSnh without signal peptide in order to determine the 3D structure and other biochemical properties to compare them with the native human IDS (IDSnh). Results showed that there are no significant differences between both molecules in spite of the two-codon modifications detected in the recombinant DNA sequence. PMID:27335624
Kulkarni, Abhishek; Ertekin, Deniz; Lee, Chi-Hon; Hummel, Thomas
2016-03-17
The precise recognition of appropriate synaptic partner neurons is a critical step during neural circuit assembly. However, little is known about the developmental context in which recognition specificity is important to establish synaptic contacts. We show that in the Drosophila visual system, sequential segregation of photoreceptor afferents, reflecting their birth order, lead to differential positioning of their growth cones in the early target region. By combining loss- and gain-of-function analyses we demonstrate that relative differences in the expression of the transcription factor Sequoia regulate R cell growth cone segregation. This initial growth cone positioning is consolidated via cell-adhesion molecule Capricious in R8 axons. Further, we show that the initial growth cone positioning determines synaptic layer selection through proximity-based axon-target interactions. Taken together, we demonstrate that birth order dependent pre-patterning of afferent growth cones is an essential pre-requisite for the identification of synaptic partner neurons during visual map formation in Drosophila.
Yang, Jie; Wang, Xiaonan; Tang, Shunming; Shen, Zhongyuan; Wu, Jinmei
2015-01-01
Peptidoglycan recognition protein (PGRP) binds specifically to peptidoglycan and plays an important role as a pattern recognition receptor in the innate immunity of insects. The cDNA of a short-type PGRP, an open reading frame of 588 bp encoding a polypeptide of 196 amino acids, was cloned from Bombyx mori. A phylogenetic tree was constructed, and the results showed that BmPGRP-S2 was most similar to Drosophila melanogaster PGRP (DmPGRP-SA). The induced expression profile of BmPGRP-S2 in healthy Escherichia coli- and Bacillus subtilis-challenged B. mori was measured using semiquantitative reverse transcriptase polymerase chain reaction analysis. The expression of BmPGRP-S2 was upregulated at 24 h by E. coli and Ba. subtilis challenge. In addition, in the integument of B. mori, RNAi knockdown of BmPGRP-S2 caused an obvious reduction in the transcription expression of the transcription factor Relish and in antibacterial effector genes Attacin, Gloverin, and Moricin. The results indicated that BmPGRP-S2 participates in the signal transduction pathway of B. mori. PMID:25797797