Sample records for facial image analysis

  1. Infrared thermal facial image sequence registration analysis and verification

    NASA Astrophysics Data System (ADS)

    Chen, Chieh-Li; Jian, Bo-Lin

    2015-03-01

    To study the emotional responses of subjects to the International Affective Picture System (IAPS), infrared thermal facial image sequence is preprocessed for registration before further analysis such that the variance caused by minor and irregular subject movements is reduced. Without affecting the comfort level and inducing minimal harm, this study proposes an infrared thermal facial image sequence registration process that will reduce the deviations caused by the unconscious head shaking of the subjects. A fixed image for registration is produced through the localization of the centroid of the eye region as well as image translation and rotation processes. Thermal image sequencing will then be automatically registered using the two-stage genetic algorithm proposed. The deviation before and after image registration will be demonstrated by image quality indices. The results show that the infrared thermal image sequence registration process proposed in this study is effective in localizing facial images accurately, which will be beneficial to the correlation analysis of psychological information related to the facial area.

  2. Principal component analysis for surface reflection components and structure in facial images and synthesis of facial images for various ages

    NASA Astrophysics Data System (ADS)

    Hirose, Misa; Toyota, Saori; Ojima, Nobutoshi; Ogawa-Ochiai, Keiko; Tsumura, Norimichi

    2017-08-01

    In this paper, principal component analysis is applied to the distribution of pigmentation, surface reflectance, and landmarks in whole facial images to obtain feature values. The relationship between the obtained feature vectors and the age of the face is then estimated by multiple regression analysis so that facial images can be modulated for woman aged 10-70. In a previous study, we analyzed only the distribution of pigmentation, and the reproduced images appeared to be younger than the apparent age of the initial images. We believe that this happened because we did not modulate the facial structures and detailed surfaces, such as wrinkles. By considering landmarks and surface reflectance over the entire face, we were able to analyze the variation in the distributions of facial structures and fine asperity, and pigmentation. As a result, our method is able to appropriately modulate the appearance of a face so that it appears to be the correct age.

  3. Learning representative features for facial images based on a modified principal component analysis

    NASA Astrophysics Data System (ADS)

    Averkin, Anton; Potapov, Alexey

    2013-05-01

    The paper is devoted to facial image analysis and particularly deals with the problem of automatic evaluation of the attractiveness of human faces. We propose a new approach for automatic construction of feature space based on a modified principal component analysis. Input data sets for the algorithm are the learning data sets of facial images, which are rated by one person. The proposed approach allows one to extract features of the individual subjective face beauty perception and to predict attractiveness values for new facial images, which were not included into a learning data set. The Pearson correlation coefficient between values predicted by our method for new facial images and personal attractiveness estimation values equals to 0.89. This means that the new approach proposed is promising and can be used for predicting subjective face attractiveness values in real systems of the facial images analysis.

  4. Automated facial acne assessment from smartphone images

    NASA Astrophysics Data System (ADS)

    Amini, Mohammad; Vasefi, Fartash; Valdebran, Manuel; Huang, Kevin; Zhang, Haomiao; Kemp, William; MacKinnon, Nicholas

    2018-02-01

    A smartphone mobile medical application is presented, that provides analysis of the health of skin on the face using a smartphone image and cloud-based image processing techniques. The mobile application employs the use of the camera to capture a front face image of a subject, after which the captured image is spatially calibrated based on fiducial points such as position of the iris of the eye. A facial recognition algorithm is used to identify features of the human face image, to normalize the image, and to define facial regions of interest (ROI) for acne assessment. We identify acne lesions and classify them into two categories: those that are papules and those that are pustules. Automated facial acne assessment was validated by performing tests on images of 60 digital human models and 10 real human face images. The application was able to identify 92% of acne lesions within five facial ROIs. The classification accuracy for separating papules from pustules was 98%. Combined with in-app documentation of treatment, lifestyle factors, and automated facial acne assessment, the app can be used in both cosmetic and clinical dermatology. It allows users to quantitatively self-measure acne severity and treatment efficacy on an ongoing basis to help them manage their chronic facial acne.

  5. Dermatological Feasibility of Multimodal Facial Color Imaging Modality for Cross-Evaluation of Facial Actinic Keratosis

    PubMed Central

    Bae, Youngwoo; Son, Taeyoon; Nelson, J. Stuart; Kim, Jae-Hong; Choi, Eung Ho; Jung, Byungjo

    2010-01-01

    Background/Purpose Digital color image analysis is currently considered as a routine procedure in dermatology. In our previous study, a multimodal facial color imaging modality (MFCIM), which provides a conventional, parallel- and cross-polarization, and fluorescent color image, was introduced for objective evaluation of various facial skin lesions. This study introduces a commercial version of MFCIM, DermaVision-PRO, for routine clinical use in dermatology and demonstrates its dermatological feasibility for cross-evaluation of skin lesions. Methods/Results Sample images of subjects with actinic keratosis or non-melanoma skin cancers were obtained at four different imaging modes. Various image analysis methods were applied to cross-evaluate the skin lesion and, finally, extract valuable diagnostic information. DermaVision-PRO is potentially a useful tool as an objective macroscopic imaging modality for quick prescreening and cross-evaluation of facial skin lesions. Conclusion DermaVision-PRO may be utilized as a useful tool for cross-evaluation of widely distributed facial skin lesions and an efficient database management of patient information. PMID:20923462

  6. Automated diagnosis of fetal alcohol syndrome using 3D facial image analysis

    PubMed Central

    Fang, Shiaofen; McLaughlin, Jason; Fang, Jiandong; Huang, Jeffrey; Autti-Rämö, Ilona; Fagerlund, Åse; Jacobson, Sandra W.; Robinson, Luther K.; Hoyme, H. Eugene; Mattson, Sarah N.; Riley, Edward; Zhou, Feng; Ward, Richard; Moore, Elizabeth S.; Foroud, Tatiana

    2012-01-01

    Objectives Use three-dimensional (3D) facial laser scanned images from children with fetal alcohol syndrome (FAS) and controls to develop an automated diagnosis technique that can reliably and accurately identify individuals prenatally exposed to alcohol. Methods A detailed dysmorphology evaluation, history of prenatal alcohol exposure, and 3D facial laser scans were obtained from 149 individuals (86 FAS; 63 Control) recruited from two study sites (Cape Town, South Africa and Helsinki, Finland). Computer graphics, machine learning, and pattern recognition techniques were used to automatically identify a set of facial features that best discriminated individuals with FAS from controls in each sample. Results An automated feature detection and analysis technique was developed and applied to the two study populations. A unique set of facial regions and features were identified for each population that accurately discriminated FAS and control faces without any human intervention. Conclusion Our results demonstrate that computer algorithms can be used to automatically detect facial features that can discriminate FAS and control faces. PMID:18713153

  7. BMI and WHR Are Reflected in Female Facial Shape and Texture: A Geometric Morphometric Image Analysis.

    PubMed

    Mayer, Christine; Windhager, Sonja; Schaefer, Katrin; Mitteroecker, Philipp

    2017-01-01

    Facial markers of body composition are frequently studied in evolutionary psychology and are important in computational and forensic face recognition. We assessed the association of body mass index (BMI) and waist-to-hip ratio (WHR) with facial shape and texture (color pattern) in a sample of young Middle European women by a combination of geometric morphometrics and image analysis. Faces of women with high BMI had a wider and rounder facial outline relative to the size of the eyes and lips, and relatively lower eyebrows. Furthermore, women with high BMI had a brighter and more reddish skin color than women with lower BMI. The same facial features were associated with WHR, even though BMI and WHR were only moderately correlated. Yet BMI was better predictable than WHR from facial attributes. After leave-one-out cross-validation, we were able to predict 25% of variation in BMI and 10% of variation in WHR by facial shape. Facial texture predicted only about 3-10% of variation in BMI and WHR. This indicates that facial shape primarily reflects total fat proportion, rather than the distribution of fat within the body. The association of reddish facial texture in high-BMI women may be mediated by increased blood pressure and superficial blood flow as well as diet. Our study elucidates how geometric morphometric image analysis serves to quantify the effect of biological factors such as BMI and WHR to facial shape and color, which in turn contributes to social perception.

  8. Signatures of personality on dense 3D facial images.

    PubMed

    Hu, Sile; Xiong, Jieyi; Fu, Pengcheng; Qiao, Lu; Tan, Jingze; Jin, Li; Tang, Kun

    2017-03-06

    It has long been speculated that cues on the human face exist that allow observers to make reliable judgments of others' personality traits. However, direct evidence of association between facial shapes and personality is missing from the current literature. This study assessed the personality attributes of 834 Han Chinese volunteers (405 males and 429 females), utilising the five-factor personality model ('Big Five'), and collected their neutral 3D facial images. Dense anatomical correspondence was established across the 3D facial images in order to allow high-dimensional quantitative analyses of the facial phenotypes. In this paper, we developed a Partial Least Squares (PLS) -based method. We used composite partial least squares component (CPSLC) to test association between the self-tested personality scores and the dense 3D facial image data, then used principal component analysis (PCA) for further validation. Among the five personality factors, agreeableness and conscientiousness in males and extraversion in females were significantly associated with specific facial patterns. The personality-related facial patterns were extracted and their effects were extrapolated on simulated 3D facial models.

  9. Automated Facial Recognition of Computed Tomography-Derived Facial Images: Patient Privacy Implications.

    PubMed

    Parks, Connie L; Monson, Keith L

    2017-04-01

    The recognizability of facial images extracted from publically available medical scans raises patient privacy concerns. This study examined how accurately facial images extracted from computed tomography (CT) scans are objectively matched with corresponding photographs of the scanned individuals. The test subjects were 128 adult Americans ranging in age from 18 to 60 years, representing both sexes and three self-identified population (ancestral descent) groups (African, European, and Hispanic). Using facial recognition software, the 2D images of the extracted facial models were compared for matches against five differently sized photo galleries. Depending on the scanning protocol and gallery size, in 6-61 % of the cases, a correct life photo match for a CT-derived facial image was the top ranked image in the generated candidate lists, even when blind searching in excess of 100,000 images. In 31-91 % of the cases, a correct match was located within the top 50 images. Few significant differences (p > 0.05) in match rates were observed between the sexes or across the three age cohorts. Highly significant differences (p < 0.01) were, however, observed across the three ancestral cohorts and between the two CT scanning protocols. Results suggest that the probability of a match between a facial image extracted from a medical scan and a photograph of the individual is moderately high. The facial image data inherent in commonly employed medical imaging modalities may need to consider a potentially identifiable form of "comparable" facial imagery and protected as such under patient privacy legislation.

  10. Spoofing detection on facial images recognition using LBP and GLCM combination

    NASA Astrophysics Data System (ADS)

    Sthevanie, F.; Ramadhani, K. N.

    2018-03-01

    The challenge for the facial based security system is how to detect facial image falsification such as facial image spoofing. Spoofing occurs when someone try to pretend as a registered user to obtain illegal access and gain advantage from the protected system. This research implements facial image spoofing detection method by analyzing image texture. The proposed method for texture analysis combines the Local Binary Pattern (LBP) and Gray Level Co-occurrence Matrix (GLCM) method. The experimental results show that spoofing detection using LBP and GLCM combination achieves high detection rate compared to that of using only LBP feature or GLCM feature.

  11. BMI and WHR Are Reflected in Female Facial Shape and Texture: A Geometric Morphometric Image Analysis

    PubMed Central

    Mayer, Christine; Windhager, Sonja; Schaefer, Katrin; Mitteroecker, Philipp

    2017-01-01

    Facial markers of body composition are frequently studied in evolutionary psychology and are important in computational and forensic face recognition. We assessed the association of body mass index (BMI) and waist-to-hip ratio (WHR) with facial shape and texture (color pattern) in a sample of young Middle European women by a combination of geometric morphometrics and image analysis. Faces of women with high BMI had a wider and rounder facial outline relative to the size of the eyes and lips, and relatively lower eyebrows. Furthermore, women with high BMI had a brighter and more reddish skin color than women with lower BMI. The same facial features were associated with WHR, even though BMI and WHR were only moderately correlated. Yet BMI was better predictable than WHR from facial attributes. After leave-one-out cross-validation, we were able to predict 25% of variation in BMI and 10% of variation in WHR by facial shape. Facial texture predicted only about 3–10% of variation in BMI and WHR. This indicates that facial shape primarily reflects total fat proportion, rather than the distribution of fat within the body. The association of reddish facial texture in high-BMI women may be mediated by increased blood pressure and superficial blood flow as well as diet. Our study elucidates how geometric morphometric image analysis serves to quantify the effect of biological factors such as BMI and WHR to facial shape and color, which in turn contributes to social perception. PMID:28052103

  12. Hepatitis Diagnosis Using Facial Color Image

    NASA Astrophysics Data System (ADS)

    Liu, Mingjia; Guo, Zhenhua

    Facial color diagnosis is an important diagnostic method in traditional Chinese medicine (TCM). However, due to its qualitative, subjective and experi-ence-based nature, traditional facial color diagnosis has a very limited application in clinical medicine. To circumvent the subjective and qualitative problems of facial color diagnosis of Traditional Chinese Medicine, in this paper, we present a novel computer aided facial color diagnosis method (CAFCDM). The method has three parts: face Image Database, Image Preprocessing Module and Diagnosis Engine. Face Image Database is carried out on a group of 116 patients affected by 2 kinds of liver diseases and 29 healthy volunteers. The quantitative color feature is extracted from facial images by using popular digital image processing techni-ques. Then, KNN classifier is employed to model the relationship between the quantitative color feature and diseases. The results show that the method can properly identify three groups: healthy, severe hepatitis with jaundice and severe hepatitis without jaundice with accuracy higher than 73%.

  13. Imaging the Facial Nerve: A Contemporary Review

    PubMed Central

    Gupta, Sachin; Mends, Francine; Hagiwara, Mari; Fatterpekar, Girish; Roehm, Pamela C.

    2013-01-01

    Imaging plays a critical role in the evaluation of a number of facial nerve disorders. The facial nerve has a complex anatomical course; thus, a thorough understanding of the course of the facial nerve is essential to localize the sites of pathology. Facial nerve dysfunction can occur from a variety of causes, which can often be identified on imaging. Computed tomography and magnetic resonance imaging are helpful for identifying bony facial canal and soft tissue abnormalities, respectively. Ultrasound of the facial nerve has been used to predict functional outcomes in patients with Bell's palsy. More recently, diffusion tensor tractography has appeared as a new modality which allows three-dimensional display of facial nerve fibers. PMID:23766904

  14. Automatic Contour Extraction of Facial Organs for Frontal Facial Images with Various Facial Expressions

    NASA Astrophysics Data System (ADS)

    Kobayashi, Hiroshi; Suzuki, Seiji; Takahashi, Hisanori; Tange, Akira; Kikuchi, Kohki

    This study deals with a method to realize automatic contour extraction of facial features such as eyebrows, eyes and mouth for the time-wise frontal face with various facial expressions. Because Snakes which is one of the most famous methods used to extract contours, has several disadvantages, we propose a new method to overcome these issues. We define the elastic contour model in order to hold the contour shape and then determine the elastic energy acquired by the amount of modification of the elastic contour model. Also we utilize the image energy obtained by brightness differences of the control points on the elastic contour model. Applying the dynamic programming method, we determine the contour position where the total value of the elastic energy and the image energy becomes minimum. Employing 1/30s time-wise facial frontal images changing from neutral to one of six typical facial expressions obtained from 20 subjects, we have estimated our method and find it enables high accuracy automatic contour extraction of facial features.

  15. Recognizing Action Units for Facial Expression Analysis

    PubMed Central

    Tian, Ying-li; Kanade, Takeo; Cohn, Jeffrey F.

    2010-01-01

    Most automatic expression analysis systems attempt to recognize a small set of prototypic expressions, such as happiness, anger, surprise, and fear. Such prototypic expressions, however, occur rather infrequently. Human emotions and intentions are more often communicated by changes in one or a few discrete facial features. In this paper, we develop an Automatic Face Analysis (AFA) system to analyze facial expressions based on both permanent facial features (brows, eyes, mouth) and transient facial features (deepening of facial furrows) in a nearly frontal-view face image sequence. The AFA system recognizes fine-grained changes in facial expression into action units (AUs) of the Facial Action Coding System (FACS), instead of a few prototypic expressions. Multistate face and facial component models are proposed for tracking and modeling the various facial features, including lips, eyes, brows, cheeks, and furrows. During tracking, detailed parametric descriptions of the facial features are extracted. With these parameters as the inputs, a group of action units (neutral expression, six upper face AUs and 10 lower face AUs) are recognized whether they occur alone or in combinations. The system has achieved average recognition rates of 96.4 percent (95.4 percent if neutral expressions are excluded) for upper face AUs and 96.7 percent (95.6 percent with neutral expressions excluded) for lower face AUs. The generalizability of the system has been tested by using independent image databases collected and FACS-coded for ground-truth by different research teams. PMID:25210210

  16. Image ratio features for facial expression recognition application.

    PubMed

    Song, Mingli; Tao, Dacheng; Liu, Zicheng; Li, Xuelong; Zhou, Mengchu

    2010-06-01

    Video-based facial expression recognition is a challenging problem in computer vision and human-computer interaction. To target this problem, texture features have been extracted and widely used, because they can capture image intensity changes raised by skin deformation. However, existing texture features encounter problems with albedo and lighting variations. To solve both problems, we propose a new texture feature called image ratio features. Compared with previously proposed texture features, e.g., high gradient component features, image ratio features are more robust to albedo and lighting variations. In addition, to further improve facial expression recognition accuracy based on image ratio features, we combine image ratio features with facial animation parameters (FAPs), which describe the geometric motions of facial feature points. The performance evaluation is based on the Carnegie Mellon University Cohn-Kanade database, our own database, and the Japanese Female Facial Expression database. Experimental results show that the proposed image ratio feature is more robust to albedo and lighting variations, and the combination of image ratio features and FAPs outperforms each feature alone. In addition, we study asymmetric facial expressions based on our own facial expression database and demonstrate the superior performance of our combined expression recognition system.

  17. Multimodal digital color imaging system for facial skin lesion analysis

    NASA Astrophysics Data System (ADS)

    Bae, Youngwoo; Lee, Youn-Heum; Jung, Byungjo

    2008-02-01

    In dermatology, various digital imaging modalities have been used as an important tool to quantitatively evaluate the treatment effect of skin lesions. Cross-polarization color image was used to evaluate skin chromophores (melanin and hemoglobin) information and parallel-polarization image to evaluate skin texture information. In addition, UV-A induced fluorescent image has been widely used to evaluate various skin conditions such as sebum, keratosis, sun damages, and vitiligo. In order to maximize the evaluation efficacy of various skin lesions, it is necessary to integrate various imaging modalities into an imaging system. In this study, we propose a multimodal digital color imaging system, which provides four different digital color images of standard color image, parallel and cross-polarization color image, and UV-A induced fluorescent color image. Herein, we describe the imaging system and present the examples of image analysis. By analyzing the color information and morphological features of facial skin lesions, we are able to comparably and simultaneously evaluate various skin lesions. In conclusion, we are sure that the multimodal color imaging system can be utilized as an important assistant tool in dermatology.

  18. Tensor Rank Preserving Discriminant Analysis for Facial Recognition.

    PubMed

    Tao, Dapeng; Guo, Yanan; Li, Yaotang; Gao, Xinbo

    2017-10-12

    Facial recognition, one of the basic topics in computer vision and pattern recognition, has received substantial attention in recent years. However, for those traditional facial recognition algorithms, the facial images are reshaped to a long vector, thereby losing part of the original spatial constraints of each pixel. In this paper, a new tensor-based feature extraction algorithm termed tensor rank preserving discriminant analysis (TRPDA) for facial image recognition is proposed; the proposed method involves two stages: in the first stage, the low-dimensional tensor subspace of the original input tensor samples was obtained; in the second stage, discriminative locality alignment was utilized to obtain the ultimate vector feature representation for subsequent facial recognition. On the one hand, the proposed TRPDA algorithm fully utilizes the natural structure of the input samples, and it applies an optimization criterion that can directly handle the tensor spectral analysis problem, thereby decreasing the computation cost compared those traditional tensor-based feature selection algorithms. On the other hand, the proposed TRPDA algorithm extracts feature by finding a tensor subspace that preserves most of the rank order information of the intra-class input samples. Experiments on the three facial databases are performed here to determine the effectiveness of the proposed TRPDA algorithm.

  19. Discrimination of gender using facial image with expression change

    NASA Astrophysics Data System (ADS)

    Kuniyada, Jun; Fukuda, Takahiro; Terada, Kenji

    2005-12-01

    By carrying out marketing research, the managers of large-sized department stores or small convenience stores obtain the information such as ratio of men and women of visitors and an age group, and improve their management plan. However, these works are carried out in the manual operations, and it becomes a big burden to small stores. In this paper, the authors propose a method of men and women discrimination by extracting difference of the facial expression change from color facial images. Now, there are a lot of methods of the automatic recognition of the individual using a motion facial image or a still facial image in the field of image processing. However, it is very difficult to discriminate gender under the influence of the hairstyle and clothes, etc. Therefore, we propose the method which is not affected by personality such as size and position of facial parts by paying attention to a change of an expression. In this method, it is necessary to obtain two facial images with an expression and an expressionless. First, a region of facial surface and the regions of facial parts such as eyes, nose, and mouth are extracted in the facial image with color information of hue and saturation in HSV color system and emphasized edge information. Next, the features are extracted by calculating the rate of the change of each facial part generated by an expression change. In the last step, the values of those features are compared between the input data and the database, and the gender is discriminated. In this paper, it experimented for the laughing expression and smile expression, and good results were provided for discriminating gender.

  20. Investigation into the use of photoanthropometry in facial image comparison.

    PubMed

    Moreton, Reuben; Morley, Johanna

    2011-10-10

    Photoanthropometry is a metric based facial image comparison technique. Measurements of the face are taken from an image using predetermined facial landmarks. Measurements are then converted to proportionality indices (PIs) and compared to PIs from another facial image. Photoanthropometry has been presented as a facial image comparison technique in UK courts for over 15 years. It is generally accepted that extrinsic factors (e.g. orientation of the head, camera angle and distance from the camera) can cause discrepancies in anthropometric measurements of the face from photographs. However there has been limited empirical research into quantifying the influence of such variables. The aim of this study was to determine the reliability of photoanthropometric measurements between different images of the same individual taken with different angulations of the camera. The study examined the facial measurements of 25 individuals from high resolution photographs, taken at different horizontal and vertical camera angles in a controlled environment. Results show that the degree of variability in facial measurements of the same individual due to variations in camera angle can be as great as the variability of facial measurements between different individuals. Results suggest that photoanthropometric facial comparison, as it is currently practiced, is unsuitable for elimination purposes. Preliminary investigations into the effects of distance from camera and image resolution in poor quality images suggest that such images are not an accurate representation of an individuals face, however further work is required. Copyright © 2011 Elsevier Ireland Ltd. All rights reserved.

  1. Magnetic resonance imaging of facial nerve schwannoma.

    PubMed

    Thompson, Andrew L; Aviv, Richard I; Chen, Joseph M; Nedzelski, Julian M; Yuen, Heng-Wai; Fox, Allan J; Bharatha, Aditya; Bartlett, Eric S; Symons, Sean P

    2009-12-01

    This study characterizes the magnetic resonance (MR) appearances of facial nerve schwannoma (FNS). We hypothesize that the extent of FNS demonstrated on MR will be greater compared to prior computed tomography studies, that geniculate involvement will be most common, and that cerebellar pontine angle (CPA) and internal auditory canal (IAC) involvement will more frequently result in sensorineural hearing loss (SNHL). Retrospective study. Clinical, pathologic, and enhanced MR imaging records of 30 patients with FNS were analyzed. Morphologic characteristics and extent of segmental facial nerve involvement were documented. Median age at initial imaging was 51 years (range, 28-76 years). Pathologic confirmation was obtained in 14 patients (47%), and the diagnosis reached in the remainder by identification of a mass, thickening, and enhancement along the course of the facial nerve. All 30 lesions involved two or more contiguous segments of the facial nerve, with 28 (93%) involving three or more segments. The median segments involved per lesion was 4, mean of 3.83. Geniculate involvement was most common, in 29 patients (97%). CPA (P = .001) and IAC (P = .02) involvement was significantly related to SNHL. Seventeen patients (57%) presented with facial nerve dysfunction, manifesting in 12 patients as facial nerve weakness or paralysis, and/or in eight with involuntary movements of the facial musculature. This study highlights the morphologic heterogeneity and typical multisegment involvement of FNS. Enhanced MR is the imaging modality of choice for FNS. The neuroradiologist must accurately diagnose and characterize this lesion, and thus facilitate optimal preoperative planning and counseling.

  2. Person-independent facial expression analysis by fusing multiscale cell features

    NASA Astrophysics Data System (ADS)

    Zhou, Lubing; Wang, Han

    2013-03-01

    Automatic facial expression recognition is an interesting and challenging task. To achieve satisfactory accuracy, deriving a robust facial representation is especially important. A novel appearance-based feature, the multiscale cell local intensity increasing patterns (MC-LIIP), to represent facial images and conduct person-independent facial expression analysis is presented. The LIIP uses a decimal number to encode the texture or intensity distribution around each pixel via pixel-to-pixel intensity comparison. To boost noise resistance, MC-LIIP carries out comparison computation on the average values of scalable cells instead of individual pixels. The facial descriptor fuses region-based histograms of MC-LIIP features from various scales, so as to encode not only textural microstructures but also the macrostructures of facial images. Finally, a support vector machine classifier is applied for expression recognition. Experimental results on the CK+ and Karolinska directed emotional faces databases show the superiority of the proposed method.

  3. Autonomous facial recognition system inspired by human visual system based logarithmical image visualization technique

    NASA Astrophysics Data System (ADS)

    Wan, Qianwen; Panetta, Karen; Agaian, Sos

    2017-05-01

    Autonomous facial recognition system is widely used in real-life applications, such as homeland border security, law enforcement identification and authentication, and video-based surveillance analysis. Issues like low image quality, non-uniform illumination as well as variations in poses and facial expressions can impair the performance of recognition systems. To address the non-uniform illumination challenge, we present a novel robust autonomous facial recognition system inspired by the human visual system based, so called, logarithmical image visualization technique. In this paper, the proposed method, for the first time, utilizes the logarithmical image visualization technique coupled with the local binary pattern to perform discriminative feature extraction for facial recognition system. The Yale database, the Yale-B database and the ATT database are used for computer simulation accuracy and efficiency testing. The extensive computer simulation demonstrates the method's efficiency, accuracy, and robustness of illumination invariance for facial recognition.

  4. A novel method to measure conspicuous facial pores using computer analysis of digital-camera-captured images: the effect of glycolic acid chemical peeling.

    PubMed

    Kakudo, Natsuko; Kushida, Satoshi; Tanaka, Nobuko; Minakata, Tatsuya; Suzuki, Kenji; Kusumoto, Kenji

    2011-11-01

    Chemical peeling is becoming increasingly popular for skin rejuvenation in dermatological esthetic surgery. Conspicuous facial pores are one of the most frequently encountered skin problems in women of all ages. This study was performed to analyze the effectiveness of reducing conspicuous facial pores using glycolic acid chemical peeling (GACP) based on a novel computer analysis of digital-camera-captured images. GACP was performed a total of five times at 2-week intervals in 22 healthy women. Computerized image analysis of conspicuous, open, and darkened facial pores was performed using the Robo Skin Analyzer CS 50. The number of conspicuous facial pores decreased significantly in 19 (86%) of the 22 subjects, with a mean improvement rate of 34.6%. The number of open pores decreased significantly in 16 (72%) of the subjects, with a mean improvement rate of 11.0%. The number of darkened pores decreased significantly in 18 (81%) of the subjects, with a mean improvement rate of 34.3%. GACP significantly reduces the number of conspicuous facial pores. The Robo Skin Analyzer CS 50 is useful for the quantification and analysis of 'pore enlargement', a subtle finding in dermatological esthetic surgery. © 2011 John Wiley & Sons A/S.

  5. Facial identification in very low-resolution images simulating prosthetic vision.

    PubMed

    Chang, M H; Kim, H S; Shin, J H; Park, K S

    2012-08-01

    Familiar facial identification is important to blind or visually impaired patients and can be achieved using a retinal prosthesis. Nevertheless, there are limitations in delivering the facial images with a resolution sufficient to distinguish facial features, such as eyes and nose, through multichannel electrode arrays used in current visual prostheses. This study verifies the feasibility of familiar facial identification under low-resolution prosthetic vision and proposes an edge-enhancement method to deliver more visual information that is of higher quality. We first generated a contrast-enhanced image and an edge image by applying the Sobel edge detector and blocked each of them by averaging. Then, we subtracted the blocked edge image from the blocked contrast-enhanced image and produced a pixelized image imitating an array of phosphenes. Before subtraction, every gray value of the edge images was weighted as 50% (mode 2), 75% (mode 3) and 100% (mode 4). In mode 1, the facial image was blocked and pixelized with no further processing. The most successful identification was achieved with mode 3 at every resolution in terms of identification index, which covers both accuracy and correct response time. We also found that the subjects recognized a distinctive face especially more accurately and faster than the other given facial images even under low-resolution prosthetic vision. Every subject could identify familiar faces even in very low-resolution images. And the proposed edge-enhancement method seemed to contribute to intermediate-stage visual prostheses.

  6. Brain Responses to Dynamic Facial Expressions: A Normative Meta-Analysis.

    PubMed

    Zinchenko, Oksana; Yaple, Zachary A; Arsalidou, Marie

    2018-01-01

    Identifying facial expressions is crucial for social interactions. Functional neuroimaging studies show that a set of brain areas, such as the fusiform gyrus and amygdala, become active when viewing emotional facial expressions. The majority of functional magnetic resonance imaging (fMRI) studies investigating face perception typically employ static images of faces. However, studies that use dynamic facial expressions (e.g., videos) are accumulating and suggest that a dynamic presentation may be more sensitive and ecologically valid for investigating faces. By using quantitative fMRI meta-analysis the present study examined concordance of brain regions associated with viewing dynamic facial expressions. We analyzed data from 216 participants that participated in 14 studies, which reported coordinates for 28 experiments. Our analysis revealed bilateral fusiform and middle temporal gyri, left amygdala, left declive of the cerebellum and the right inferior frontal gyrus. These regions are discussed in terms of their relation to models of face processing.

  7. Reproducibility of the dynamics of facial expressions in unilateral facial palsy.

    PubMed

    Alagha, M A; Ju, X; Morley, S; Ayoub, A

    2018-02-01

    The aim of this study was to assess the reproducibility of non-verbal facial expressions in unilateral facial paralysis using dynamic four-dimensional (4D) imaging. The Di4D system was used to record five facial expressions of 20 adult patients. The system captured 60 three-dimensional (3D) images per second; each facial expression took 3-4seconds which was recorded in real time. Thus a set of 180 3D facial images was generated for each expression. The procedure was repeated after 30min to assess the reproducibility of the expressions. A mathematical facial mesh consisting of thousands of quasi-point 'vertices' was conformed to the face in order to determine the morphological characteristics in a comprehensive manner. The vertices were tracked throughout the sequence of the 180 images. Five key 3D facial frames from each sequence of images were analyzed. Comparisons were made between the first and second capture of each facial expression to assess the reproducibility of facial movements. Corresponding images were aligned using partial Procrustes analysis, and the root mean square distance between them was calculated and analyzed statistically (paired Student t-test, P<0.05). Facial expressions of lip purse, cheek puff, and raising of eyebrows were reproducible. Facial expressions of maximum smile and forceful eye closure were not reproducible. The limited coordination of various groups of facial muscles contributed to the lack of reproducibility of these facial expressions. 4D imaging is a useful clinical tool for the assessment of facial expressions. Copyright © 2017 International Association of Oral and Maxillofacial Surgeons. Published by Elsevier Ltd. All rights reserved.

  8. Automatic image assessment from facial attributes

    NASA Astrophysics Data System (ADS)

    Ptucha, Raymond; Kloosterman, David; Mittelstaedt, Brian; Loui, Alexander

    2013-03-01

    Personal consumer photography collections often contain photos captured by numerous devices stored both locally and via online services. The task of gathering, organizing, and assembling still and video assets in preparation for sharing with others can be quite challenging. Current commercial photobook applications are mostly manual-based requiring significant user interactions. To assist the consumer in organizing these assets, we propose an automatic method to assign a fitness score to each asset, whereby the top scoring assets are used for product creation. Our method uses cues extracted from analyzing pixel data, metadata embedded in the file, as well as ancillary tags or online comments. When a face occurs in an image, its features have a dominating influence on both aesthetic and compositional properties of the displayed image. As such, this paper will emphasize the contributions faces have on affecting the overall fitness score of an image. To understand consumer preference, we conducted a psychophysical study that spanned 27 judges, 5,598 faces, and 2,550 images. Preferences on a per-face and per-image basis were independently gathered to train our classifiers. We describe how to use machine learning techniques to merge differing facial attributes into a single classifier. Our novel methods of facial weighting, fusion of facial attributes, and dimensionality reduction produce stateof- the-art results suitable for commercial applications.

  9. Feature selection from a facial image for distinction of sasang constitution.

    PubMed

    Koo, Imhoi; Kim, Jong Yeol; Kim, Myoung Geun; Kim, Keun Ho

    2009-09-01

    Recently, oriental medicine has received attention for providing personalized medicine through consideration of the unique nature and constitution of individual patients. With the eventual goal of globalization, the current trend in oriental medicine research is the standardization by adopting western scientific methods, which could represent a scientific revolution. The purpose of this study is to establish methods for finding statistically significant features in a facial image with respect to distinguishing constitution and to show the meaning of those features. From facial photo images, facial elements are analyzed in terms of the distance, angle and the distance ratios, for which there are 1225, 61 250 and 749 700 features, respectively. Due to the very large number of facial features, it is quite difficult to determine truly meaningful features. We suggest a process for the efficient analysis of facial features including the removal of outliers, control for missing data to guarantee data confidence and calculation of statistical significance by applying ANOVA. We show the statistical properties of selected features according to different constitutions using the nine distances, 10 angles and 10 rates of distance features that are finally established. Additionally, the Sasang constitutional meaning of the selected features is shown here.

  10. Automated Video Based Facial Expression Analysis of Neuropsychiatric Disorders

    PubMed Central

    Wang, Peng; Barrett, Frederick; Martin, Elizabeth; Milanova, Marina; Gur, Raquel E.; Gur, Ruben C.; Kohler, Christian; Verma, Ragini

    2008-01-01

    Deficits in emotional expression are prominent in several neuropsychiatric disorders, including schizophrenia. Available clinical facial expression evaluations provide subjective and qualitative measurements, which are based on static 2D images that do not capture the temporal dynamics and subtleties of expression changes. Therefore, there is a need for automated, objective and quantitative measurements of facial expressions captured using videos. This paper presents a computational framework that creates probabilistic expression profiles for video data and can potentially help to automatically quantify emotional expression differences between patients with neuropsychiatric disorders and healthy controls. Our method automatically detects and tracks facial landmarks in videos, and then extracts geometric features to characterize facial expression changes. To analyze temporal facial expression changes, we employ probabilistic classifiers that analyze facial expressions in individual frames, and then propagate the probabilities throughout the video to capture the temporal characteristics of facial expressions. The applications of our method to healthy controls and case studies of patients with schizophrenia and Asperger’s syndrome demonstrate the capability of the video-based expression analysis method in capturing subtleties of facial expression. Such results can pave the way for a video based method for quantitative analysis of facial expressions in clinical research of disorders that cause affective deficits. PMID:18045693

  11. Dose and diagnostic image quality in digital tomosynthesis imaging of facial bones in pediatrics

    NASA Astrophysics Data System (ADS)

    King, J. M.; Hickling, S.; Elbakri, I. A.; Reed, M.; Wrogemann, J.

    2011-03-01

    The purpose of this study was to evaluate the use of digital tomosynthesis (DT) for pediatric facial bone imaging. We compared the eye lens dose and diagnostic image quality of DT facial bone exams relative to digital radiography (DR) and computed tomography (CT), and investigated whether we could modify our current DT imaging protocol to reduce patient dose while maintaining sufficient diagnostic image quality. We measured the dose to the eye lens for all three modalities using high-sensitivity thermoluminescent dosimeters (TLDs) and an anthropomorphic skull phantom. To assess the diagnostic image quality of DT compared to the corresponding DR and CT images, we performed an observer study where the visibility of anatomical structures in the DT phantom images were rated on a four-point scale. We then acquired DT images at lower doses and had radiologists indicate whether the visibility of each structure was adequate for diagnostic purposes. For typical facial bone exams, we measured eye lens doses of 0.1-0.4 mGy for DR, 0.3-3.7 mGy for DT, and 26 mGy for CT. In general, facial bone structures were visualized better with DT then DR, and the majority of structures were visualized well enough to avoid the need for CT. DT imaging provides high quality diagnostic images of the facial bones while delivering significantly lower doses to the lens of the eye compared to CT. In addition, we found that by adjusting the imaging parameters, the DT effective dose can be reduced by up to 50% while maintaining sufficient image quality.

  12. Evaluating visibility of age spot and freckle based on simulated spectral reflectance distribution and facial color image

    NASA Astrophysics Data System (ADS)

    Hirose, Misa; Toyota, Saori; Tsumura, Norimichi

    2018-02-01

    In this research, we evaluate the visibility of age spot and freckle with changing the blood volume based on simulated spectral reflectance distribution and the actual facial color images, and compare these results. First, we generate three types of spatial distribution of age spot and freckle in patch-like images based on the simulated spectral reflectance. The spectral reflectance is simulated using Monte Carlo simulation of light transport in multi-layered tissue. Next, we reconstruct the facial color image with changing the blood volume. We acquire the concentration distribution of melanin, hemoglobin and shading components by applying the independent component analysis on a facial color image. We reproduce images using the obtained melanin and shading concentration and the changed hemoglobin concentration. Finally, we evaluate the visibility of pigmentations using simulated spectral reflectance distribution and facial color images. In the result of simulated spectral reflectance distribution, we found that the visibility became lower as the blood volume increases. However, we can see that a specific blood volume reduces the visibility of the actual pigmentations from the result of the facial color images.

  13. Image analysis of skin color heterogeneity focusing on skin chromophores and the age-related changes in facial skin.

    PubMed

    Kikuchi, Kumiko; Masuda, Yuji; Yamashita, Toyonobu; Kawai, Eriko; Hirao, Tetsuji

    2015-05-01

    Heterogeneity with respect to skin color tone is one of the key factors in visual perception of facial attractiveness and age. However, there have been few studies on quantitative analyses of the color heterogeneity of facial skin. The purpose of this study was to develop image evaluation methods for skin color heterogeneity focusing on skin chromophores and then characterize ethnic differences and age-related changes. A facial imaging system equipped with an illumination unit and a high-resolution digital camera was used to develop image evaluation methods for skin color heterogeneity. First, melanin and/or hemoglobin images were obtained using pigment-specific image-processing techniques, which involved conversion from Commission Internationale de l'Eclairage XYZ color values to melanin and/or hemoglobin indexes as measures of their contents. Second, a spatial frequency analysis with threshold settings was applied to the individual images. Cheek skin images of 194 healthy Asian and Caucasian female subjects were acquired using the imaging system. Applying this methodology, the skin color heterogeneity of Asian and Caucasian faces was characterized. The proposed pigment-specific image-processing techniques allowed visual discrimination of skin redness from skin pigmentation. In the heterogeneity analyses of cheek skin color, age-related changes in melanin were clearly detected in Asian and Caucasian skin. Furthermore, it was found that the heterogeneity indexes of hemoglobin were significantly higher in Caucasian skin than in Asian skin. We have developed evaluation methods for skin color heterogeneity by image analyses based on the major chromophores, melanin and hemoglobin, with special reference to their size. This methodology focusing on skin color heterogeneity should be useful for better understanding of aging and ethnic differences. © 2014 John Wiley & Sons A/S. Published by John Wiley & Sons Ltd.

  14. Evaluation of facial attractiveness in black people according to the subjective facial analysis criteria.

    PubMed

    Melo, Andréa Reis de; Conti, Ana Cláudia de Castro Ferreira; Almeida-Pedrin, Renata Rodrigues; Didier, Victor; Valarelli, Danilo Pinelli; Capelozza Filho, Leopoldino

    2017-02-01

    The objective of this study was to evaluate the facial attractiveness in 30 black individuals, according to the Subjective Facial Analysis criteria. Frontal and profile view photographs of 30 black individuals were evaluated for facial attractiveness and classified as esthetically unpleasant, acceptable, or pleasant by 50 evaluators: the 30 individuals from the sample, 10 orthodontists, and 10 laymen. Besides assessing the facial attractiveness, the evaluators had to identify the structures responsible for the classification as unpleasant and pleasant. Intraexaminer agreement was assessed by using Spearman's correlation, correlation within each category using Kendall concordance coefficient, and correlation between the 3 categories using chi-square test and proportions. Most of the frontal (53. 5%) and profile view (54. 9%) photographs were classified as esthetically acceptable. The structures most identified as esthetically unpleasant were the mouth, lips, and face, in the frontal view; and nose and chin in the profile view. The structures most identified as esthetically pleasant were harmony, face, and mouth, in the frontal view; and harmony and nose in the profile view. The ratings by the examiners in the sample and laymen groups showed statistically significant correlation in both views. The orthodontists agreed with the laymen on the evaluation of the frontal view and disagreed on profile view, especially regarding whether the images were esthetically unpleasant or acceptable. Based on these results, the evaluation of facial attractiveness according to the Subjective Facial Analysis criteria proved to be applicable and to have a subjective influence; therefore, it is suggested that the patient's opinion regarding the facial esthetics should be considered in orthodontic treatmentplanning.

  15. Advances in computer imaging/applications in facial plastic surgery.

    PubMed

    Papel, I D; Jiannetto, D F

    1999-01-01

    Rapidly progressing computer technology, ever-increasing expectations of patients, and a confusing medicolegal environment requires a clarification of the role of computer imaging/applications. Advances in computer technology and its applications are reviewed. A brief historical discussion is included for perspective. Improvements in both hardware and software with the advent of digital imaging have allowed great increases in speed and accuracy in patient imaging. This facilitates doctor-patient communication and possibly realistic patient expectations. Patients seeking cosmetic surgery now often expect preoperative imaging. Although society in general has become more litigious, a literature search up to 1998 reveals no lawsuits directly involving computer imaging. It appears that conservative utilization of computer imaging by the facial plastic surgeon may actually reduce liability and promote communication. Recent advances have significantly enhanced the value of computer imaging in the practice of facial plastic surgery. These technological advances in computer imaging appear to contribute a useful technique for the practice of facial plastic surgery. Inclusion of computer imaging should be given serious consideration as an adjunct to clinical practice.

  16. Feature Selection from a Facial Image for Distinction of Sasang Constitution

    PubMed Central

    Koo, Imhoi; Kim, Jong Yeol; Kim, Myoung Geun

    2009-01-01

    Recently, oriental medicine has received attention for providing personalized medicine through consideration of the unique nature and constitution of individual patients. With the eventual goal of globalization, the current trend in oriental medicine research is the standardization by adopting western scientific methods, which could represent a scientific revolution. The purpose of this study is to establish methods for finding statistically significant features in a facial image with respect to distinguishing constitution and to show the meaning of those features. From facial photo images, facial elements are analyzed in terms of the distance, angle and the distance ratios, for which there are 1225, 61 250 and 749 700 features, respectively. Due to the very large number of facial features, it is quite difficult to determine truly meaningful features. We suggest a process for the efficient analysis of facial features including the removal of outliers, control for missing data to guarantee data confidence and calculation of statistical significance by applying ANOVA. We show the statistical properties of selected features according to different constitutions using the nine distances, 10 angles and 10 rates of distance features that are finally established. Additionally, the Sasang constitutional meaning of the selected features is shown here. PMID:19745013

  17. Facial fluid synthesis for assessment of acne vulgaris using luminescent visualization system through optical imaging and integration of fluorescent imaging system

    NASA Astrophysics Data System (ADS)

    Balbin, Jessie R.; Dela Cruz, Jennifer C.; Camba, Clarisse O.; Gozo, Angelo D.; Jimenez, Sheena Mariz B.; Tribiana, Aivje C.

    2017-06-01

    Acne vulgaris, commonly called as acne, is a skin problem that occurs when oil and dead skin cells clog up in a person's pores. This is because hormones change which makes the skin oilier. The problem is people really do not know the real assessment of sensitivity of their skin in terms of fluid development on their faces that tends to develop acne vulgaris, thus having more complications. This research aims to assess Acne Vulgaris using luminescent visualization system through optical imaging and integration of image processing algorithms. Specifically, this research aims to design a prototype for facial fluid analysis using luminescent visualization system through optical imaging and integration of fluorescent imaging system, and to classify different facial fluids present in each person. Throughout the process, some structures and layers of the face will be excluded, leaving only a mapped facial structure with acne regions. Facial fluid regions are distinguished from the acne region as they are characterized differently.

  18. Image Based Hair Segmentation Algorithm for the Application of Automatic Facial Caricature Synthesis

    PubMed Central

    Peng, Zhenyun; Zhang, Yaohui

    2014-01-01

    Hair is a salient feature in human face region and are one of the important cues for face analysis. Accurate detection and presentation of hair region is one of the key components for automatic synthesis of human facial caricature. In this paper, an automatic hair detection algorithm for the application of automatic synthesis of facial caricature based on a single image is proposed. Firstly, hair regions in training images are labeled manually and then the hair position prior distributions and hair color likelihood distribution function are estimated from these labels efficiently. Secondly, the energy function of the test image is constructed according to the estimated prior distributions of hair location and hair color likelihood. This energy function is further optimized according to graph cuts technique and initial hair region is obtained. Finally, K-means algorithm and image postprocessing techniques are applied to the initial hair region so that the final hair region can be segmented precisely. Experimental results show that the average processing time for each image is about 280 ms and the average hair region detection accuracy is above 90%. The proposed algorithm is applied to a facial caricature synthesis system. Experiments proved that with our proposed hair segmentation algorithm the facial caricatures are vivid and satisfying. PMID:24592182

  19. Effects of glycolic acid chemical peeling on facial pigment deposition: evaluation using novel computer analysis of digital-camera-captured images.

    PubMed

    Kakudo, Natsuko; Kushida, Satoshi; Suzuki, Kenji; Kusumoto, Kenji

    2013-12-01

    Chemical peeling is becoming increasingly popular for skin rejuvenation in dermatological cosmetic medicine. However, the improvements seen with chemical peeling are often very minor, and it is difficult to conduct a quantitative assessment of pre- and post-treatment appearance. We report the pre- and postpeeling effects for facial pigment deposition using a novel computer analysis method for digital-camera-captured images. Glycolic acid chemical peeling was performed a total of 5 times at 2-week intervals in 23 healthy women. We conducted a computer image analysis by utilizing Robo Skin Analyzer CS 50 and Clinical Suite 2.1 and then reviewed each parameter for the area of facial pigment deposition pre- and post-treatment. Parameters were pigmentation size and four pigmentation categories: little pigmentation and three levels of marked pigmentation (Lv1, 2, and 3) based on detection threshold. Each parameter was measured, and the total area of facial pigmentation was calculated. The total area of little pigmentation and marked pigmentation (Lv1) was significantly reduced. On the other hand, a significant difference was not observed for the total area of marked pigmentation Lv2 and Lv3. This suggests that glycolic acid chemical peeling has an effect on small facial pigment disposition or has an effect on light pigment deposition. As the Robo Skin Analyzer is useful for objectively quantifying and analyzing minor changes in facial skin, it is considered to be an effective tool for accumulating treatment evidence in the cosmetic and esthetic skin field. © 2013 Wiley Periodicals, Inc.

  20. Pose-variant facial expression recognition using an embedded image system

    NASA Astrophysics Data System (ADS)

    Song, Kai-Tai; Han, Meng-Ju; Chang, Shuo-Hung

    2008-12-01

    In recent years, one of the most attractive research areas in human-robot interaction is automated facial expression recognition. Through recognizing the facial expression, a pet robot can interact with human in a more natural manner. In this study, we focus on the facial pose-variant problem. A novel method is proposed in this paper to recognize pose-variant facial expressions. After locating the face position in an image frame, the active appearance model (AAM) is applied to track facial features. Fourteen feature points are extracted to represent the variation of facial expressions. The distance between feature points are defined as the feature values. These feature values are sent to a support vector machine (SVM) for facial expression determination. The pose-variant facial expression is classified into happiness, neutral, sadness, surprise or anger. Furthermore, in order to evaluate the performance for practical applications, this study also built a low resolution database (160x120 pixels) using a CMOS image sensor. Experimental results show that the recognition rate is 84% with the self-built database.

  1. Emotion Estimation Algorithm from Facial Image Analyses of e-Learning Users

    NASA Astrophysics Data System (ADS)

    Shigeta, Ayuko; Koike, Takeshi; Kurokawa, Tomoya; Nosu, Kiyoshi

    This paper proposes an emotion estimation algorithm from e-Learning user's facial image. The algorithm characteristics are as follows: The criteria used to relate an e-Learning use's emotion to a representative emotion were obtained from the time sequential analysis of user's facial expressions. By examining the emotions of the e-Learning users and the positional change of the facial expressions from the experiment results, the following procedures are introduce to improve the estimation reliability; (1) some effective features points are chosen by the emotion estimation (2) dividing subjects into two groups by the change rates of the face feature points (3) selection of the eigenvector of the variance-co-variance matrices (cumulative contribution rate>=95%) (4) emotion calculation using Mahalanobis distance.

  2. Macroscopic in vivo imaging of facial nerve regeneration in Thy1-GFP rats.

    PubMed

    Placheta, Eva; Wood, Matthew D; Lafontaine, Christine; Frey, Manfred; Gordon, Tessa; Borschel, Gregory H

    2015-01-01

    Facial nerve injury leads to severe functional and aesthetic deficits. The transgenic Thy1-GFP rat is a new model for facial nerve injury and reconstruction research that will help improve clinical outcomes through translational facial nerve injury research. To determine whether serial in vivo imaging of nerve regeneration in the transgenic rat model is possible, facial nerve regeneration was imaged under the main paradigms of facial nerve injury and reconstruction. Fifteen male Thy1-GFP rats, which express green fluorescent protein (GFP) in their neural structures, were divided into 3 groups in the laboratory: crush-injury, direct repair, and cross-face nerve grafting (30-mm graft length). The distal nerve stump or nerve graft was predegenerated for 2 weeks. The facial nerve of the transgenic rats was serially imaged at the time of operation and after 2, 4, and 8 weeks of regeneration. The imaging was performed under a GFP-MDS-96/BN excitation stand (BLS Ltd). Facial nerve injury. Optical fluorescence of regenerating facial nerve axons. Serial in vivo imaging of the regeneration of GFP-positive axons in the Thy1-GFP rat model is possible. All animals survived the short imaging procedures well, and nerve regeneration was followed over clinically relevant distances. The predegeneration of the distal nerve stump or the cross-face nerve graft was, however, necessary to image the regeneration front at early time points. Crush injury was not suitable to sufficiently predegenerate the nerve (and to allow for degradation of the GFP through Wallerian degeneration). After direct repair, axons regenerated over the coaptation site in between 2 and 4 weeks. The GFP-positive nerve fibers reached the distal end of the 30-mm-long cross-face nervegrafts after 4 to 8 weeks of regeneration. The time course of facial nerve regeneration was studied by serial in vivo imaging in the transgenic rat model. Nerve regeneration was followed over clinically relevant distances in a small

  3. Validation of image analysis techniques to measure skin aging features from facial photographs.

    PubMed

    Hamer, M A; Jacobs, L C; Lall, J S; Wollstein, A; Hollestein, L M; Rae, A R; Gossage, K W; Hofman, A; Liu, F; Kayser, M; Nijsten, T; Gunn, D A

    2015-11-01

    Accurate measurement of the extent skin has aged is crucial for skin aging research. Image analysis offers a quick and consistent approach for quantifying skin aging features from photographs, but is prone to technical bias and requires proper validation. Facial photographs of 75 male and 75 female North-European participants, randomly selected from the Rotterdam Study, were graded by two physicians using photonumeric scales for wrinkles (full face, forehead, crow's feet, nasolabial fold and upper lip), pigmented spots and telangiectasia. Image analysis measurements of the same features were optimized using photonumeric grades from 50 participants, then compared to photonumeric grading in the 100 remaining participants stratified by sex. The inter-rater reliability of the photonumeric grades was good to excellent (intraclass correlation coefficients 0.65-0.93). Correlations between the digital measures and the photonumeric grading were moderate to excellent for all the wrinkle comparisons (Spearman's rho ρ = 0.52-0.89) bar the upper lip wrinkles in the men (fair, ρ = 0.30). Correlations were moderate to good for pigmented spots and telangiectasia (ρ = 0.60-0.75). These comparisons demonstrate that all the image analysis measures, bar the upper lip measure in the men, are suitable for use in skin aging research and highlight areas of improvement for future refinements of the techniques. © 2015 John Wiley & Sons A/S. Published by John Wiley & Sons.

  4. Speech Signal and Facial Image Processing for Obstructive Sleep Apnea Assessment

    PubMed Central

    Espinoza-Cuadros, Fernando; Fernández-Pozo, Rubén; Toledano, Doroteo T.; Alcázar-Ramírez, José D.; López-Gonzalo, Eduardo; Hernández-Gómez, Luis A.

    2015-01-01

    Obstructive sleep apnea (OSA) is a common sleep disorder characterized by recurring breathing pauses during sleep caused by a blockage of the upper airway (UA). OSA is generally diagnosed through a costly procedure requiring an overnight stay of the patient at the hospital. This has led to proposing less costly procedures based on the analysis of patients' facial images and voice recordings to help in OSA detection and severity assessment. In this paper we investigate the use of both image and speech processing to estimate the apnea-hypopnea index, AHI (which describes the severity of the condition), over a population of 285 male Spanish subjects suspected to suffer from OSA and referred to a Sleep Disorders Unit. Photographs and voice recordings were collected in a supervised but not highly controlled way trying to test a scenario close to an OSA assessment application running on a mobile device (i.e., smartphones or tablets). Spectral information in speech utterances is modeled by a state-of-the-art low-dimensional acoustic representation, called i-vector. A set of local craniofacial features related to OSA are extracted from images after detecting facial landmarks using Active Appearance Models (AAMs). Support vector regression (SVR) is applied on facial features and i-vectors to estimate the AHI. PMID:26664493

  5. Speech Signal and Facial Image Processing for Obstructive Sleep Apnea Assessment.

    PubMed

    Espinoza-Cuadros, Fernando; Fernández-Pozo, Rubén; Toledano, Doroteo T; Alcázar-Ramírez, José D; López-Gonzalo, Eduardo; Hernández-Gómez, Luis A

    2015-01-01

    Obstructive sleep apnea (OSA) is a common sleep disorder characterized by recurring breathing pauses during sleep caused by a blockage of the upper airway (UA). OSA is generally diagnosed through a costly procedure requiring an overnight stay of the patient at the hospital. This has led to proposing less costly procedures based on the analysis of patients' facial images and voice recordings to help in OSA detection and severity assessment. In this paper we investigate the use of both image and speech processing to estimate the apnea-hypopnea index, AHI (which describes the severity of the condition), over a population of 285 male Spanish subjects suspected to suffer from OSA and referred to a Sleep Disorders Unit. Photographs and voice recordings were collected in a supervised but not highly controlled way trying to test a scenario close to an OSA assessment application running on a mobile device (i.e., smartphones or tablets). Spectral information in speech utterances is modeled by a state-of-the-art low-dimensional acoustic representation, called i-vector. A set of local craniofacial features related to OSA are extracted from images after detecting facial landmarks using Active Appearance Models (AAMs). Support vector regression (SVR) is applied on facial features and i-vectors to estimate the AHI.

  6. Interactive searching of facial image databases

    NASA Astrophysics Data System (ADS)

    Nicholls, Robert A.; Shepherd, John W.; Shepherd, Jean

    1995-09-01

    A set of psychological facial descriptors has been devised to enable computerized searching of criminal photograph albums. The descriptors have been used to encode image databased of up to twelve thousand images. Using a system called FACES, the databases are searched by translating a witness' verbal description into corresponding facial descriptors. Trials of FACES have shown that this coding scheme is more productive and efficient than searching traditional photograph albums. An alternative method of searching the encoded database using a genetic algorithm is currenly being tested. The genetic search method does not require the witness to verbalize a description of the target but merely to indicate a degree of similarity between the target and a limited selection of images from the database. The major drawback of FACES is that is requires a manual encoding of images. Research is being undertaken to automate the process, however, it will require an algorithm which can predict human descriptive values. Alternatives to human derived coding schemes exist using statistical classifications of images. Since databases encoded using statistical classifiers do not have an obvious direct mapping to human derived descriptors, a search method which does not require the entry of human descriptors is required. A genetic search algorithm is being tested for such a purpose.

  7. Facial Nerve Paralysis due to a Pleomorphic Adenoma with the Imaging Characteristics of a Facial Nerve Schwannoma

    PubMed Central

    Nader, Marc-Elie; Bell, Diana; Sturgis, Erich M.; Ginsberg, Lawrence E.; Gidley, Paul W.

    2014-01-01

    Background Facial nerve paralysis in a patient with a salivary gland mass usually denotes malignancy. However, facial paralysis can also be caused by benign salivary gland tumors. Methods We present a case of facial nerve paralysis due to a benign salivary gland tumor that had the imaging characteristics of an intraparotid facial nerve schwannoma. Results The patient presented to our clinic 4 years after the onset of facial nerve paralysis initially diagnosed as Bell palsy. Computed tomography demonstrated filling and erosion of the stylomastoid foramen with a mass on the facial nerve. Postoperative histopathology showed the presence of a pleomorphic adenoma. Facial paralysis was thought to be caused by extrinsic nerve compression. Conclusions This case illustrates the difficulty of accurate preoperative diagnosis of a parotid gland mass and reinforces the concept that facial nerve paralysis in the context of salivary gland tumors may not always indicate malignancy. PMID:25083397

  8. Facial Nerve Paralysis due to a Pleomorphic Adenoma with the Imaging Characteristics of a Facial Nerve Schwannoma.

    PubMed

    Nader, Marc-Elie; Bell, Diana; Sturgis, Erich M; Ginsberg, Lawrence E; Gidley, Paul W

    2014-08-01

    Background Facial nerve paralysis in a patient with a salivary gland mass usually denotes malignancy. However, facial paralysis can also be caused by benign salivary gland tumors. Methods We present a case of facial nerve paralysis due to a benign salivary gland tumor that had the imaging characteristics of an intraparotid facial nerve schwannoma. Results The patient presented to our clinic 4 years after the onset of facial nerve paralysis initially diagnosed as Bell palsy. Computed tomography demonstrated filling and erosion of the stylomastoid foramen with a mass on the facial nerve. Postoperative histopathology showed the presence of a pleomorphic adenoma. Facial paralysis was thought to be caused by extrinsic nerve compression. Conclusions This case illustrates the difficulty of accurate preoperative diagnosis of a parotid gland mass and reinforces the concept that facial nerve paralysis in the context of salivary gland tumors may not always indicate malignancy.

  9. Classifying Facial Actions

    PubMed Central

    Donato, Gianluca; Bartlett, Marian Stewart; Hager, Joseph C.; Ekman, Paul; Sejnowski, Terrence J.

    2010-01-01

    The Facial Action Coding System (FACS) [23] is an objective method for quantifying facial movement in terms of component actions. This system is widely used in behavioral investigations of emotion, cognitive processes, and social interaction. The coding is presently performed by highly trained human experts. This paper explores and compares techniques for automatically recognizing facial actions in sequences of images. These techniques include analysis of facial motion through estimation of optical flow; holistic spatial analysis, such as principal component analysis, independent component analysis, local feature analysis, and linear discriminant analysis; and methods based on the outputs of local filters, such as Gabor wavelet representations and local principal components. Performance of these systems is compared to naive and expert human subjects. Best performances were obtained using the Gabor wavelet representation and the independent component representation, both of which achieved 96 percent accuracy for classifying 12 facial actions of the upper and lower face. The results provide converging evidence for the importance of using local filters, high spatial frequencies, and statistical independence for classifying facial actions. PMID:21188284

  10. Novel Noninvasive Brain Disease Detection System Using a Facial Image Sensor

    PubMed Central

    Shu, Ting; Zhang, Bob; Tang, Yuan Yan

    2017-01-01

    Brain disease including any conditions or disabilities that affect the brain is fast becoming a leading cause of death. The traditional diagnostic methods of brain disease are time-consuming, inconvenient and non-patient friendly. As more and more individuals undergo examinations to determine if they suffer from any form of brain disease, developing noninvasive, efficient, and patient friendly detection systems will be beneficial. Therefore, in this paper, we propose a novel noninvasive brain disease detection system based on the analysis of facial colors. The system consists of four components. A facial image is first captured through a specialized sensor, where four facial key blocks are next located automatically from the various facial regions. Color features are extracted from each block to form a feature vector for classification via the Probabilistic Collaborative based Classifier. To thoroughly test the system and its performance, seven facial key block combinations were experimented. The best result was achieved using the second facial key block, where it showed that the Probabilistic Collaborative based Classifier is the most suitable. The overall performance of the proposed system achieves an accuracy −95%, a sensitivity −94.33%, a specificity −95.67%, and an average processing time (for one sample) of <1 min at brain disease detection. PMID:29292716

  11. Enhanced facial recognition for thermal imagery using polarimetric imaging.

    PubMed

    Gurton, Kristan P; Yuffa, Alex J; Videen, Gorden W

    2014-07-01

    We present a series of long-wave-infrared (LWIR) polarimetric-based thermal images of facial profiles in which polarization-state information of the image-forming radiance is retained and displayed. The resultant polarimetric images show enhanced facial features, additional texture, and details that are not present in corresponding conventional thermal imagery. It has been generally thought that conventional thermal imagery (MidIR or LWIR) could not produce the detailed spatial information required for reliable human identification due to the so-called "ghosting" effect often seen in thermal imagery of human subjects. By using polarimetric information, we are able to extract subtle surface features of the human face, thus improving subject identification. Polarimetric image sets considered include the conventional thermal intensity image, S0, the two Stokes images, S1 and S2, and a Stokes image product called the degree-of-linear-polarization image.

  12. Facial Asymmetry-Based Age Group Estimation: Role in Recognizing Age-Separated Face Images.

    PubMed

    Sajid, Muhammad; Taj, Imtiaz Ahmad; Bajwa, Usama Ijaz; Ratyal, Naeem Iqbal

    2018-04-23

    Face recognition aims to establish the identity of a person based on facial characteristics. On the other hand, age group estimation is the automatic calculation of an individual's age range based on facial features. Recognizing age-separated face images is still a challenging research problem due to complex aging processes involving different types of facial tissues, skin, fat, muscles, and bones. Certain holistic and local facial features are used to recognize age-separated face images. However, most of the existing methods recognize face images without incorporating the knowledge learned from age group estimation. In this paper, we propose an age-assisted face recognition approach to handle aging variations. Inspired by the observation that facial asymmetry is an age-dependent intrinsic facial feature, we first use asymmetric facial dimensions to estimate the age group of a given face image. Deeply learned asymmetric facial features are then extracted for face recognition using a deep convolutional neural network (dCNN). Finally, we integrate the knowledge learned from the age group estimation into the face recognition algorithm using the same dCNN. This integration results in a significant improvement in the overall performance compared to using the face recognition algorithm alone. The experimental results on two large facial aging datasets, the MORPH and FERET sets, show that the proposed age group estimation based on the face recognition approach yields superior performance compared to some existing state-of-the-art methods. © 2018 American Academy of Forensic Sciences.

  13. A Multivariate Analysis of Unilateral Cleft Lip and Palate Facial Skeletal Morphology.

    PubMed

    Starbuck, John M; Ghoneima, Ahmed; Kula, Katherine

    2015-07-01

    Unilateral cleft lip and palate (UCLP) occurs when the maxillary and nasal facial prominences fail to fuse correctly during development, resulting in a palatal cleft and clefted soft and hard tissues of the dentoalveolus. The UCLP deformity may compromise an individual's ability to eat, chew, and speak. In this retrospective cross-sectional study, cone beam computed tomography (CBCT) images of 7-17-year-old individuals born with UCLP (n = 24) and age- and sex-matched controls (n = 24) were assessed. Coordinate values of three-dimensional anatomical landmarks (n = 32) were recorded from each CBCT image. Data were evaluated using principal coordinates analysis (PCOORD) and Euclidean distance matrix analysis (EDMA). Approximately 40% of morphometric variation is captured by PCOORD axes 1-3, and the negative and positive ends of each axis are associated with specific patterns of morphological differences. Approximately 36% of facial skeletal measures significantly differ by confidence interval testing (α = 0.10) between samples. Although significant form differences occur across the facial skeleton, strong patterns of morphological differences were localized to the lateral and superioinferior aspects of the nasal aperture, particularly on the clefted side of the face. The UCLP deformity strongly influences facial skeletal morphology of the midface and oronasal facial regions, and to a lesser extent the upper and lower facial skeletons. The pattern of strong morphological differences in the oronasal region combined with differences across the facial complex suggests that craniofacial bones are integrated and covary, despite influences from the congenital cleft.

  14. Easy facial analysis using the facial golden mask.

    PubMed

    Kim, Yong-Ha

    2007-05-01

    For over 2000 years, many artists and scientists have tried to understand or quantify the form of the perfect, ideal, or most beautiful face both in art and in vivo (life). A mathematical relationship has been consistently and repeatedly reported to be present in beautiful things. This particular relationship is the golden ratio. It is a mathematical ratio of 1.618:1 that seems to appear recurrently in beautiful things in nature as well as in other things that are seen as beautiful. Dr. Marquardt made the facial golden mask that contains and includes all of the one-dimensional and two-dimensional geometric golden elements formed from the golden ratio. The purpose of this study is to evaluate the usefulness of the golden facial mask. In 40 cases, the authors applied the facial golden mask to preoperative and postoperative photographs and scored each photograph on a 1 to 5 scale from the perspective of their personal aesthetic views. The score was lower when the facial deformity was severe, whereas it was higher when the face was attractive. Compared with the average scores of facial mask applied photographs and nonapplied photographs using a nonparametric test, statistical significance was not reached (P > 0.05). This implies that the facial golden mask may be used as an analytical tool. The facial golden mask is easy to apply, inexpensive, and relatively objective. Therefore, the authors introduce it as a useful facial analysis.

  15. Quantified Facial Soft-tissue Strain in Animation Measured by Real-time Dynamic 3-Dimensional Imaging.

    PubMed

    Hsu, Vivian M; Wes, Ari M; Tahiri, Youssef; Cornman-Homonoff, Joshua; Percec, Ivona

    2014-09-01

    The aim of this study is to evaluate and quantify dynamic soft-tissue strain in the human face using real-time 3-dimensional imaging technology. Thirteen subjects (8 women, 5 men) between the ages of 18 and 70 were imaged using a dual-camera system and 3-dimensional optical analysis (ARAMIS, Trilion Quality Systems, Pa.). Each subject was imaged at rest and with the following facial expressions: (1) smile, (2) laughter, (3) surprise, (4) anger, (5) grimace, and (6) pursed lips. The facial strains defining stretch and compression were computed for each subject and compared. The areas of greatest strain were localized to the midface and lower face for all expressions. Subjects over the age of 40 had a statistically significant increase in stretch in the perioral region while lip pursing compared with subjects under the age of 40 (58.4% vs 33.8%, P = 0.015). When specific components of lip pursing were analyzed, there was a significantly greater degree of stretch in the nasolabial fold region in subjects over 40 compared with those under 40 (61.6% vs 32.9%, P = 0.007). Furthermore, we observed a greater degree of asymmetry of strain in the nasolabial fold region in the older age group (18.4% vs 5.4%, P = 0.03). This pilot study illustrates that the face can be objectively and quantitatively evaluated using dynamic major strain analysis. The technology of 3-dimensional optical imaging can be used to advance our understanding of facial soft-tissue dynamics and the effects of animation on facial strain over time.

  16. Quantitative Magnetic Resonance Imaging Volumetry of Facial Muscles in Healthy Patients with Facial Palsy

    PubMed Central

    Volk, Gerd F.; Karamyan, Inna; Klingner, Carsten M.; Reichenbach, Jürgen R.

    2014-01-01

    Background: Magnetic resonance imaging (MRI) has not yet been established systematically to detect structural muscular changes after facial nerve lesion. The purpose of this pilot study was to investigate quantitative assessment of MRI muscle volume data for facial muscles. Methods: Ten healthy subjects and 5 patients with facial palsy were recruited. Using manual or semiautomatic segmentation of 3T MRI, volume measurements were performed for the frontal, procerus, risorius, corrugator supercilii, orbicularis oculi, nasalis, zygomaticus major, zygomaticus minor, levator labii superioris, orbicularis oris, depressor anguli oris, depressor labii inferioris, and mentalis, as well as for the masseter and temporalis as masticatory muscles for control. Results: All muscles except the frontal (identification in 4/10 volunteers), procerus (4/10), risorius (6/10), and zygomaticus minor (8/10) were identified in all volunteers. Sex or age effects were not seen (all P > 0.05). There was no facial asymmetry with exception of the zygomaticus major (larger on the left side; P = 0.012). The exploratory examination of 5 patients revealed considerably smaller muscle volumes on the palsy side 2 months after facial injury. One patient with chronic palsy showed substantial muscle volume decrease, which also occurred in another patient with incomplete chronic palsy restricted to the involved facial area. Facial nerve reconstruction led to mixed results of decreased but also increased muscle volumes on the palsy side compared with the healthy side. Conclusions: First systematic quantitative MRI volume measures of 5 different clinical presentations of facial paralysis are provided. PMID:25289366

  17. [Magnetic resonance imaging in facial injuries and digital fusion CT/MRI].

    PubMed

    Kozakiewicz, Marcin; Olszycki, Marek; Arkuszewski, Piotr; Stefańczyk, Ludomir

    2006-01-01

    Magnetic resonance images [MRI] and their digital fusion with computed tomography [CT] data, observed in patients affected with facial injuries, are presented in this study. The MR imaging of 12 posttraumatic patients was performed in the same plains as their previous CT scans. Evaluation focused on quality of the facial soft tissues depicting, which was unsatisfactory in CT. Using the own "Dental Studio" programme the digital fusion of the both modalities was performed. Pathologic dislocations and injures of facial soft tissues are visualized better in MRI than in CT examination. Especially MRI properly reveals disturbances in intraorbital soft structures. MRI-based assessment is valuable in patients affected with facial soft tissues injuries, especially in case of orbita/sinuses hernia. Fusion CT/MRI scans allows to evaluate simultaneously bone structure and soft tissues of the same region.

  18. Quantified Facial Soft-tissue Strain in Animation Measured by Real-time Dynamic 3-Dimensional Imaging

    PubMed Central

    Hsu, Vivian M.; Wes, Ari M.; Tahiri, Youssef; Cornman-Homonoff, Joshua

    2014-01-01

    Background: The aim of this study is to evaluate and quantify dynamic soft-tissue strain in the human face using real-time 3-dimensional imaging technology. Methods: Thirteen subjects (8 women, 5 men) between the ages of 18 and 70 were imaged using a dual-camera system and 3-dimensional optical analysis (ARAMIS, Trilion Quality Systems, Pa.). Each subject was imaged at rest and with the following facial expressions: (1) smile, (2) laughter, (3) surprise, (4) anger, (5) grimace, and (6) pursed lips. The facial strains defining stretch and compression were computed for each subject and compared. Results: The areas of greatest strain were localized to the midface and lower face for all expressions. Subjects over the age of 40 had a statistically significant increase in stretch in the perioral region while lip pursing compared with subjects under the age of 40 (58.4% vs 33.8%, P = 0.015). When specific components of lip pursing were analyzed, there was a significantly greater degree of stretch in the nasolabial fold region in subjects over 40 compared with those under 40 (61.6% vs 32.9%, P = 0.007). Furthermore, we observed a greater degree of asymmetry of strain in the nasolabial fold region in the older age group (18.4% vs 5.4%, P = 0.03). Conclusions: This pilot study illustrates that the face can be objectively and quantitatively evaluated using dynamic major strain analysis. The technology of 3-dimensional optical imaging can be used to advance our understanding of facial soft-tissue dynamics and the effects of animation on facial strain over time. PMID:25426394

  19. A new method for automatic tracking of facial landmarks in 3D motion captured images (4D).

    PubMed

    Al-Anezi, T; Khambay, B; Peng, M J; O'Leary, E; Ju, X; Ayoub, A

    2013-01-01

    The aim of this study was to validate the automatic tracking of facial landmarks in 3D image sequences. 32 subjects (16 males and 16 females) aged 18-35 years were recruited. 23 anthropometric landmarks were marked on the face of each subject with non-permanent ink using a 0.5mm pen. The subjects were asked to perform three facial animations (maximal smile, lip purse and cheek puff) from rest position. Each animation was captured by the 3D imaging system. A single operator manually digitised the landmarks on the 3D facial models and their locations were compared with those of the automatically tracked ones. To investigate the accuracy of manual digitisation, the operator re-digitised the same set of 3D images of 10 subjects (5 male and 5 female) at 1 month interval. The discrepancies in x, y and z coordinates between the 3D position of the manual digitised landmarks and that of the automatic tracked facial landmarks were within 0.17mm. The mean distance between the manually digitised and the automatically tracked landmarks using the tracking software was within 0.55 mm. The automatic tracking of facial landmarks demonstrated satisfactory accuracy which would facilitate the analysis of the dynamic motion during facial animations. Copyright © 2012 International Association of Oral and Maxillofacial Surgeons. Published by Elsevier Ltd. All rights reserved.

  20. Two-dimensional auto-correlation analysis and Fourier-transform analysis of second-harmonic-generation image for quantitative analysis of collagen fiber in human facial skin

    NASA Astrophysics Data System (ADS)

    Ogura, Yuki; Tanaka, Yuji; Hase, Eiji; Yamashita, Toyonobu; Yasui, Takeshi

    2018-02-01

    We compare two-dimensional auto-correlation (2D-AC) analysis and two-dimensional Fourier transform (2D-FT) for evaluation of age-dependent structural change of facial dermal collagen fibers caused by intrinsic aging and extrinsic photo-aging. The age-dependent structural change of collagen fibers for female subjects' cheek skin in their 20s, 40s, and 60s were more noticeably reflected in 2D-AC analysis than in 2D-FT analysis. Furthermore, 2D-AC analysis indicated significantly higher correlation with the skin elasticity measured by Cutometer® than 2D-AC analysis. 2D-AC analysis of SHG image has a high potential for quantitative evaluation of not only age-dependent structural change of collagen fibers but also skin elasticity.

  1. Three-dimensional analysis of facial shape and symmetry in twins using laser surface scanning.

    PubMed

    Djordjevic, J; Jadallah, M; Zhurov, A I; Toma, A M; Richmond, S

    2013-08-01

    Three-dimensional analysis of facial shape and symmetry in twins. Faces of 37 twin pairs [19 monozygotic (MZ) and 18 dizygotic (DZ)] were laser scanned at the age of 15 during a follow-up of the Avon Longitudinal Study of Parents and Children (ALSPAC), South West of England. Facial shape was analysed using two methods: 1) Procrustes analysis of landmark configurations (63 x, y and z coordinates of 21 facial landmarks) and 2) three-dimensional comparisons of facial surfaces within each twin pair. Monozygotic and DZ twins were compared using ellipsoids representing 95% of the variation in landmark configurations and surface-based average faces. Facial symmetry was analysed by superimposing the original and mirror facial images. Both analyses showed greater similarity of facial shape in MZ twins, with lower third being the least similar. Procrustes analysis did not reveal any significant difference in facial landmark configurations of MZ and DZ twins. The average faces of MZ and DZ males were coincident in the forehead, supraorbital and infraorbital ridges, the bridge of the nose and lower lip. In MZ and DZ females, the eyes, supraorbital and infraorbital ridges, philtrum and lower part of the cheeks were coincident. Zygosity did not seem to influence the amount of facial symmetry. Lower facial third was the most asymmetrical. Three-dimensional analyses revealed differences in facial shapes of MZ and DZ twins. The relative contribution of genetic and environmental factors is different for the upper, middle and lower facial thirds. © 2012 John Wiley & Sons A/S. Published by John Wiley & Sons Ltd.

  2. Early Changes in Facial Profile Following Structured Filler Rhinoplasty: An Anthropometric Analysis Using a 3-Dimensional Imaging System.

    PubMed

    Rho, Nark Kyoung; Park, Je Young; Youn, Choon Shik; Lee, Soo-Keun; Kim, Hei Sung

    2017-02-01

    Quantitative measurements are important for objective evaluation of postprocedural outcomes. Three-dimensional (3D) imaging is known as an objective, accurate, and reliable system for quantifying the soft tissue dimensions of the face. To compare the preprocedural and acute postprocedural nasofrontal, nasofacial, nasolabial, and nasomental angles, early changes in the height and length of the nose, and nasal volume using a 3D surface imaging with a light-emitting diode. The 3D imaging analysis of 40 Korean women who underwent structured nonsurgical rhinoplasty was conducted. The 3D assessment was performed before, immediately after, 1 day, and 2 weeks after filler rhinoplasty with a Morpheus 3D scanner (Morpheus Co., Seoul, Korea). There were significant early changes in facial profile following nonsurgical rhinoplasty with a hyaluronic acid filler. An average increase of 6.03° in the nasofrontal angle, an increase of 3.79° in the nasolabial angle, increase of 0.88° in the nasomental angle, and a reduction of 0.83° in the nasofacial angle was observed at 2 weeks of follow-up. Increment in nasal volume and nose height was also found after 2 weeks. Side effects, such as hematoma, nodules, and skin necrosis, were not observed. The 3D surface imaging quantitatively demonstrated the early changes in facial profile after structured filler rhinoplasty. The study results describe significant acute spatial changes in nose shape following treatment.

  3. Recognition of children on age-different images: Facial morphology and age-stable features.

    PubMed

    Caplova, Zuzana; Compassi, Valentina; Giancola, Silvio; Gibelli, Daniele M; Obertová, Zuzana; Poppa, Pasquale; Sala, Remo; Sforza, Chiarella; Cattaneo, Cristina

    2017-07-01

    The situation of missing children is one of the most emotional social issues worldwide. The search for and identification of missing children is often hampered, among others, by the fact that the facial morphology of long-term missing children changes as they grow. Nowadays, the wide coverage by surveillance systems potentially provides image material for comparisons with images of missing children that may facilitate identification. The aim of study was to identify whether facial features are stable in time and can be utilized for facial recognition by comparing facial images of children at different ages as well as to test the possible use of moles in recognition. The study was divided into two phases (1) morphological classification of facial features using an Anthropological Atlas; (2) algorithm developed in MATLAB® R2014b for assessing the use of moles as age-stable features. The assessment of facial features by Anthropological Atlases showed high mismatch percentages among observers. On average, the mismatch percentages were lower for features describing shape than for those describing size. The nose tip cleft and the chin dimple showed the best agreement between observers regarding both categorization and stability over time. Using the position of moles as a reference point for recognition of the same person on age-different images seems to be a useful method in terms of objectivity and it can be concluded that moles represent age-stable facial features that may be considered for preliminary recognition. Copyright © 2017 The Chartered Society of Forensic Sciences. Published by Elsevier B.V. All rights reserved.

  4. Brain responses to facial attractiveness induced by facial proportions: evidence from an fMRI study

    PubMed Central

    Shen, Hui; Chau, Desmond K. P.; Su, Jianpo; Zeng, Ling-Li; Jiang, Weixiong; He, Jufang; Fan, Jintu; Hu, Dewen

    2016-01-01

    Brain responses to facial attractiveness induced by facial proportions are investigated by using functional magnetic resonance imaging (fMRI), in 41 young adults (22 males and 19 females). The subjects underwent fMRI while they were presented with computer-generated, yet realistic face images, which had varying facial proportions, but the same neutral facial expression, baldhead and skin tone, as stimuli. Statistical parametric mapping with parametric modulation was used to explore the brain regions with the response modulated by facial attractiveness ratings (ARs). The results showed significant linear effects of the ARs in the caudate nucleus and the orbitofrontal cortex for all of the subjects, and a non-linear response profile in the right amygdala for only the male subjects. Furthermore, canonical correlation analysis was used to learn the most relevant facial ratios that were best correlated with facial attractiveness. A regression model on the fMRI-derived facial ratio components demonstrated a strong linear relationship between the visually assessed mean ARs and the predictive ARs. Overall, this study provided, for the first time, direct neurophysiologic evidence of the effects of facial ratios on facial attractiveness and suggested that there are notable gender differences in perceiving facial attractiveness as induced by facial proportions. PMID:27779211

  5. Brain responses to facial attractiveness induced by facial proportions: evidence from an fMRI study.

    PubMed

    Shen, Hui; Chau, Desmond K P; Su, Jianpo; Zeng, Ling-Li; Jiang, Weixiong; He, Jufang; Fan, Jintu; Hu, Dewen

    2016-10-25

    Brain responses to facial attractiveness induced by facial proportions are investigated by using functional magnetic resonance imaging (fMRI), in 41 young adults (22 males and 19 females). The subjects underwent fMRI while they were presented with computer-generated, yet realistic face images, which had varying facial proportions, but the same neutral facial expression, baldhead and skin tone, as stimuli. Statistical parametric mapping with parametric modulation was used to explore the brain regions with the response modulated by facial attractiveness ratings (ARs). The results showed significant linear effects of the ARs in the caudate nucleus and the orbitofrontal cortex for all of the subjects, and a non-linear response profile in the right amygdala for only the male subjects. Furthermore, canonical correlation analysis was used to learn the most relevant facial ratios that were best correlated with facial attractiveness. A regression model on the fMRI-derived facial ratio components demonstrated a strong linear relationship between the visually assessed mean ARs and the predictive ARs. Overall, this study provided, for the first time, direct neurophysiologic evidence of the effects of facial ratios on facial attractiveness and suggested that there are notable gender differences in perceiving facial attractiveness as induced by facial proportions.

  6. Three-dimensional analysis of facial morphology.

    PubMed

    Liu, Yun; Kau, Chung How; Talbert, Leslie; Pan, Feng

    2014-09-01

    The objectives of this study were to evaluate sexual dimorphism for facial features within Chinese and African American populations and to compare the facial morphology by sex between these 2 populations. Three-dimensional facial images were acquired by using the portable 3dMDface System, which captured 189 subjects from 2 population groups of Chinese (n = 72) and African American (n = 117). Each population was categorized into male and female groups for evaluation. All subjects in the groups were aged between 18 and 30 years and had no apparent facial anomalies. A total of 23 anthropometric landmarks were identified on the three-dimensional faces of each subject. Twenty-one measurements in 4 regions, including 19 distances and 2 angles, were not only calculated but also compared within and between the Chinese and African American populations. The Student's t-test was used to analyze each data set obtained within each subgroup. Distinct facial differences were presented between the examined subgroups. When comparing the sex differences of facial morphology in the Chinese population, significant differences were noted in 71.43% of the parameters calculated, and the same proportion was found in the African American group. The facial morphologic differences between the Chinese and African American populations were evaluated by sex. The proportion of significant differences in the parameters calculated was 90.48% for females and 95.24% for males between the 2 populations. The African American population had a more convex profile and greater face width than those of the Chinese population. Sexual dimorphism for facial features was presented in both the Chinese and African American populations. In addition, there were significant differences in facial morphology between these 2 populations.

  7. Modelling the perceptual similarity of facial expressions from image statistics and neural responses.

    PubMed

    Sormaz, Mladen; Watson, David M; Smith, William A P; Young, Andrew W; Andrews, Timothy J

    2016-04-01

    The ability to perceive facial expressions of emotion is essential for effective social communication. We investigated how the perception of facial expression emerges from the image properties that convey this important social signal, and how neural responses in face-selective brain regions might track these properties. To do this, we measured the perceptual similarity between expressions of basic emotions, and investigated how this is reflected in image measures and in the neural response of different face-selective regions. We show that the perceptual similarity of different facial expressions (fear, anger, disgust, sadness, happiness) can be predicted by both surface and feature shape information in the image. Using block design fMRI, we found that the perceptual similarity of expressions could also be predicted from the patterns of neural response in the face-selective posterior superior temporal sulcus (STS), but not in the fusiform face area (FFA). These results show that the perception of facial expression is dependent on the shape and surface properties of the image and on the activity of specific face-selective regions. Copyright © 2016 Elsevier Inc. All rights reserved.

  8. Bilateral cleft lip and palate: A morphometric analysis of facial skeletal form using cone beam computed tomography.

    PubMed

    Starbuck, John M; Ghoneima, Ahmed; Kula, Katherine

    2015-07-01

    Bilateral cleft lip and palate (BCLP) is caused by a lack of merging of maxillary and nasal facial prominences during development and morphogenesis. BCLP is associated with congenital defects of the oronasal facial region that can impair ingestion, mastication, speech, and dentofacial development. Using cone beam computed tomography (CBCT) images, 7- to 18-year old individuals born with BCLP (n = 15) and age- and sex-matched controls (n = 15) were retrospectively assessed. Coordinate values of three-dimensional facial skeletal anatomical landmarks (n = 32) were measured from each CBCT image. Data were evaluated using principal coordinates analysis (PCOORD) and Euclidean Distance Matrix Analysis (EDMA). PCOORD axes 1-3 explain approximately 45% of the morphological variation between samples, and specific patterns of morphological differences were associated with each axis. Approximately, 30% of facial skeletal measures significantly differ by confidence interval testing (α = 0.10) between samples. While significant form differences occur across the facial skeleton, strong patterns of differences are localized to the lateral and superioinferior aspects of the nasal aperture. In conclusion, the BCLP deformity significantly alters facial skeletal morphology of the midface and oronasal regions of the face, but morphological differences were also found in the upper facial skeleton and to a lesser extent, the lower facial skeleton. This pattern of strong differences in the oronasal region of the facial skeleton combined with differences across the rest of the facial complex underscores the idea that bones of the craniofacial skeleton are integrated. © 2015 Wiley Periodicals, Inc.

  9. Comparative Accuracy of Facial Models Fabricated Using Traditional and 3D Imaging Techniques.

    PubMed

    Lincoln, Ketu P; Sun, Albert Y T; Prihoda, Thomas J; Sutton, Alan J

    2016-04-01

    The purpose of this investigation was to compare the accuracy of facial models fabricated using facial moulage impression methods to the three-dimensional printed (3DP) fabrication methods using soft tissue images obtained from cone beam computed tomography (CBCT) and 3D stereophotogrammetry (3D-SPG) scans. A reference phantom model was fabricated using a 3D-SPG image of a human control form with ten fiducial markers placed on common anthropometric landmarks. This image was converted into the investigation control phantom model (CPM) using 3DP methods. The CPM was attached to a camera tripod for ease of image capture. Three CBCT and three 3D-SPG images of the CPM were captured. The DICOM and STL files from the three 3dMD and three CBCT were imported to the 3DP, and six testing models were made. Reversible hydrocolloid and dental stone were used to make three facial moulages of the CPM, and the impressions/casts were poured in type IV gypsum dental stone. A coordinate measuring machine (CMM) was used to measure the distances between each of the ten fiducial markers. Each measurement was made using one point as a static reference to the other nine points. The same measuring procedures were accomplished on all specimens. All measurements were compared between specimens and the control. The data were analyzed using ANOVA and Tukey pairwise comparison of the raters, methods, and fiducial markers. The ANOVA multiple comparisons showed significant difference among the three methods (p < 0.05). Further, the interaction of methods versus fiducial markers also showed significant difference (p < 0.05). The CBCT and facial moulage method showed the greatest accuracy. 3DP models fabricated using 3D-SPG showed statistical difference in comparison to the models fabricated using the traditional method of facial moulage and 3DP models fabricated from CBCT imaging. 3DP models fabricated using 3D-SPG were less accurate than the CPM and models fabricated using facial moulage and CBCT

  10. Video analysis of the biomechanics of a bicycle accident resulting in significant facial fractures.

    PubMed

    Syed, Shameer H; Willing, Ryan; Jenkyn, Thomas R; Yazdani, Arjang

    2013-11-01

    This study aimed to use video analysis techniques to determine the velocity, impact force, angle of impact, and impulse to fracture involved in a video-recorded bicycle accident resulting in facial fractures. Computed tomographic images of the resulting facial injury are presented for correlation with data and calculations. To our knowledge, such an analysis of an actual recorded trauma has not been reported in the literature. A video recording of the accident was split into frames and analyzed using an image editing program. Measurements of velocity and angle of impact were obtained from this analysis, and the force of impact and impulse were calculated using the inverse dynamic method with connected rigid body segments. These results were then correlated with the actual fracture pattern found on computed tomographic imaging of the subject's face. There was an impact velocity of 6.25 m/s, impact angles of 14 and 6.3 degrees of neck extension and axial rotation, respectively, an impact force of 1910.4 N, and an impulse to fracture of 47.8 Ns. These physical parameters resulted in clinically significant bilateral mid-facial Le Fort II and III pattern fractures. These data confer further understanding of the biomechanics of bicycle-related accidents by correlating an actual clinical outcome with the kinematic and dynamic parameters involved in the accident itself and yielding a concrete evidence of the velocity, force, and impulse necessary to cause clinically significant facial trauma. These findings can aid in the design of protective equipment for bicycle riders to help avoid this type of injury.

  11. [Facial nerve neurinomas].

    PubMed

    Sokołowski, Jacek; Bartoszewicz, Robert; Morawski, Krzysztof; Jamróz, Barbara; Niemczyk, Kazimierz

    2013-01-01

    Evaluation of diagnostic, surgical technique, treatment results facial nerve neurinomas and its comparison with literature was the main purpose of this study. Seven cases of patients (2005-2011) with facial nerve schwannomas were included to retrospective analysis in the Department of Otolaryngology, Medical University of Warsaw. All patients were assessed with history of the disease, physical examination, hearing tests, computed tomography and/or magnetic resonance imaging, electronystagmography. Cases were observed in the direction of potential complications and recurrences. Neurinoma of the facial nerve occurred in the vertical segment (n=2), facial nerve geniculum (n=1) and the internal auditory canal (n=4). The symptoms observed in patients were analyzed: facial nerve paresis (n=3), hearing loss (n=2), dizziness (n=1). Magnetic resonance imaging and computed tomography allowed to confirm the presence of the tumor and to assess its staging. Schwannoma of the facial nerve has been surgically removed using the middle fossa approach (n=5) and by antromastoidectomy (n=2). Anatomical continuity of the facial nerve was achieved in 3 cases. In the twelve months after surgery, facial nerve paresis was rated at level II-III° HB. There was no recurrence of the tumor in radiological observation. Facial nerve neurinoma is a rare tumor. Currently surgical techniques allow in most cases, the radical removing of the lesion and reconstruction of the VII nerve function. The rate of recurrence is low. A tumor of the facial nerve should be considered in the differential diagnosis of nerve VII paresis. Copyright © 2013 Polish Otorhinolaryngology - Head and Neck Surgery Society. Published by Elsevier Urban & Partner Sp. z.o.o. All rights reserved.

  12. Enhancing facial features by using clear facial features

    NASA Astrophysics Data System (ADS)

    Rofoo, Fanar Fareed Hanna

    2017-09-01

    The similarity of features between individuals of same ethnicity motivated the idea of this project. The idea of this project is to extract features of clear facial image and impose them on blurred facial image of same ethnic origin as an approach to enhance a blurred facial image. A database of clear images containing 30 individuals equally divided to five different ethnicities which were Arab, African, Chines, European and Indian. Software was built to perform pre-processing on images in order to align the features of clear and blurred images. And the idea was to extract features of clear facial image or template built from clear facial images using wavelet transformation to impose them on blurred image by using reverse wavelet. The results of this approach did not come well as all the features did not align together as in most cases the eyes were aligned but the nose or mouth were not aligned. Then we decided in the next approach to deal with features separately but in the result in some cases a blocky effect was present on features due to not having close matching features. In general the available small database did not help to achieve the goal results, because of the number of available individuals. The color information and features similarity could be more investigated to achieve better results by having larger database as well as improving the process of enhancement by the availability of closer matches in each ethnicity.

  13. Evaluation of facial expression in acute pain in cats.

    PubMed

    Holden, E; Calvo, G; Collins, M; Bell, A; Reid, J; Scott, E M; Nolan, A M

    2014-12-01

    To describe the development of a facial expression tool differentiating pain-free cats from those in acute pain. Observers shown facial images from painful and pain-free cats were asked to identify if they were in pain or not. From facial images, anatomical landmarks were identified and distances between these were mapped. Selected distances underwent statistical analysis to identify features discriminating pain-free and painful cats. Additionally, thumbnail photographs were reviewed by two experts to identify discriminating facial features between the groups. Observers (n = 68) had difficulty in identifying pain-free from painful cats, with only 13% of observers being able to discriminate more than 80% of painful cats. Analysis of 78 facial landmarks and 80 distances identified six significant factors differentiating pain-free and painful faces including ear position and areas around the mouth/muzzle. Standardised mouth and ear distances when combined showed excellent discrimination properties, correctly differentiating pain-free and painful cats in 98% of cases. Expert review supported these findings and a cartoon-type picture scale was developed from thumbnail images. Initial investigation into facial features of painful and pain-free cats suggests potentially good discrimination properties of facial images. Further testing is required for development of a clinical tool. © 2014 British Small Animal Veterinary Association.

  14. Cortical representation of facial and tongue movements: a task functional magnetic resonance imaging study.

    PubMed

    Xiao, Fu-Long; Gao, Pei-Yi; Qian, Tian-Yi; Sui, Bin-Bin; Xue, Jing; Zhou, Jian; Lin, Yan

    2017-05-01

    Functional magnetic resonance imaging (fMRI) mapping can present the activated cortical area during movement, while little is known about precise location in facial and tongue movements. To investigate the representation of facial and tongue movements by task fMRI. Twenty right-handed healthy subjects were underwent block design task fMRI examination. Task movements included lip pursing, cheek bulging, grinning and vertical tongue excursion. Statistical parametric mapping (SPM8) was applied to analysis the data. One-sample t-test was used to calculate the common activation area between facial and tongue movements. Also, paired t-test was used to test for areas of over- or underactivation in tongue movement compared with each group of facial movements. The common areas within facial and tongue movements suggested the similar motor circuits of activation in both movements. Prior activation in tongue movement was situated laterally and inferiorly in sensorimotor area relative to facial movements. Prior activation of tongue movement was investigated in left superior parietal lobe relative to lip pursing. Also, prior activation in bilateral cuneus lobe in grinning compared with tongue movement was detected. © 2015 Scandinavian Society of Clinical Physiology and Nuclear Medicine. Published by John Wiley & Sons Ltd.

  15. Facial Expression Recognition using Multiclass Ensemble Least-Square Support Vector Machine

    NASA Astrophysics Data System (ADS)

    Lawi, Armin; Sya'Rani Machrizzandi, M.

    2018-03-01

    Facial expression is one of behavior characteristics of human-being. The use of biometrics technology system with facial expression characteristics makes it possible to recognize a person’s mood or emotion. The basic components of facial expression analysis system are face detection, face image extraction, facial classification and facial expressions recognition. This paper uses Principal Component Analysis (PCA) algorithm to extract facial features with expression parameters, i.e., happy, sad, neutral, angry, fear, and disgusted. Then Multiclass Ensemble Least-Squares Support Vector Machine (MELS-SVM) is used for the classification process of facial expression. The result of MELS-SVM model obtained from our 185 different expression images of 10 persons showed high accuracy level of 99.998% using RBF kernel.

  16. A Neuromonitoring Approach to Facial Nerve Preservation During Image-guided Robotic Cochlear Implantation.

    PubMed

    Ansó, Juan; Dür, Cilgia; Gavaghan, Kate; Rohrbach, Helene; Gerber, Nicolas; Williamson, Tom; Calvo, Enric M; Balmer, Thomas Wyss; Precht, Christina; Ferrario, Damien; Dettmer, Matthias S; Rösler, Kai M; Caversaccio, Marco D; Bell, Brett; Weber, Stefan

    2016-01-01

    A multielectrode probe in combination with an optimized stimulation protocol could provide sufficient sensitivity and specificity to act as an effective safety mechanism for preservation of the facial nerve in case of an unsafe drill distance during image-guided cochlear implantation. A minimally invasive cochlear implantation is enabled by image-guided and robotic-assisted drilling of an access tunnel to the middle ear cavity. The approach requires the drill to pass at distances below 1  mm from the facial nerve and thus safety mechanisms for protecting this critical structure are required. Neuromonitoring is currently used to determine facial nerve proximity in mastoidectomy but lacks sensitivity and specificity necessaries to effectively distinguish the close distance ranges experienced in the minimally invasive approach, possibly because of current shunting of uninsulated stimulating drilling tools in the drill tunnel and because of nonoptimized stimulation parameters. To this end, we propose an advanced neuromonitoring approach using varying levels of stimulation parameters together with an integrated bipolar and monopolar stimulating probe. An in vivo study (sheep model) was conducted in which measurements at specifically planned and navigated lateral distances from the facial nerve were performed to determine if specific sets of stimulation parameters in combination with the proposed neuromonitoring system could reliably detect an imminent collision with the facial nerve. For the accurate positioning of the neuromonitoring probe, a dedicated robotic system for image-guided cochlear implantation was used and drilling accuracy was corrected on postoperative microcomputed tomographic images. From 29 trajectories analyzed in five different subjects, a correlation between stimulus threshold and drill-to-facial nerve distance was found in trajectories colliding with the facial nerve (distance <0.1  mm). The shortest pulse duration that provided the highest

  17. Clinical significance of quantitative analysis of facial nerve enhancement on MRI in Bell's palsy.

    PubMed

    Song, Mee Hyun; Kim, Jinna; Jeon, Ju Hyun; Cho, Chang Il; Yoo, Eun Hye; Lee, Won-Sang; Lee, Ho-Ki

    2008-11-01

    Quantitative analysis of the facial nerve on the lesion side as well as the normal side, which allowed for more accurate measurement of facial nerve enhancement in patients with facial palsy, showed statistically significant correlation with the initial severity of facial nerve inflammation, although little prognostic significance was shown. This study investigated the clinical significance of quantitative measurement of facial nerve enhancement in patients with Bell's palsy by analyzing the enhancement pattern and correlating MRI findings with initial severity of facial palsy and clinical outcome. Facial nerve enhancement was measured quantitatively by using the region of interest on pre- and postcontrast T1-weighted images in 44 patients diagnosed with Bell's palsy. The signal intensity increase on the lesion side was first compared with that of the contralateral side and then correlated with the initial degree of facial palsy and prognosis. The lesion side showed significantly higher signal intensity increase compared with the normal side in all of the segments except for the mastoid segment. Signal intensity increase at the internal auditory canal and labyrinthine segments showed correlation with the initial degree of facial palsy but no significant difference was found between different prognostic groups.

  18. Use of Computer Imaging in Rhinoplasty: A Survey of the Practices of Facial Plastic Surgeons.

    PubMed

    Singh, Prabhjyot; Pearlman, Steven

    2017-08-01

    The objective of this study was to quantify the use of computer imaging by facial plastic surgeons. AAFPRS Facial plastic surgeons were surveyed about their use of computer imaging during rhinoplasty consultations. The survey collected information about surgeon demographics, practice settings, practice patterns, and rates of computer imaging (CI) for primary and revision rhinoplasty. For those surgeons who used CI, additional information was also collected, which included who performed the imaging and whether the patient was given the morphed images after the consultation. A total of 238 out of 1200 (19.8%) facial plastic surgeons responded to the survey. Out of those who responded, 195 surgeons (83%) were board certified by the American Board of Facial Plastic and Reconstructive Surgeons (ABFPRS). The majority of respondents (150 surgeons, 63%) used CI during rhinoplasty consultation. Of the surgeons who use CI, 92% performed the image morphing themselves. Approximately two-thirds of surgeons who use CI gave their patient a printout of the morphed images after the consultation. Computer imaging (CI) is a frequently utilized tool for facial plastic surgeons during cosmetic consultations with patients. Based on these results of this study, it can be suggested that the majority of facial plastic surgeons who use CI do so for both primary and revision rhinoplasty. As more sophisticated systems become available, it is possible that utilization of CI modalities will increase. This provides the surgeon with further tools to use at his or her disposal during discussion of aesthetic surgery. This journal requires that authors assign a level of evidence to each article. For a full description of these Evidence-Based Medicine ratings, please refer to the Table of Contents or the online Instructions to Authors www.springer.com/00266 .

  19. Correcting the planar perspective projection in geometric structures applied to forensic facial analysis.

    PubMed

    Baldasso, Rosane Pérez; Tinoco, Rachel Lima Ribeiro; Vieira, Cristina Saft Matos; Fernandes, Mário Marques; Oliveira, Rogério Nogueira

    2016-10-01

    The process of forensic facial analysis may be founded on several scientific techniques and imaging modalities, such as digital signal processing, photogrammetry and craniofacial anthropometry. However, one of the main limitations in this analysis is the comparison of images acquired with different angles of incidence. The present study aimed to explore a potential approach for the correction of the planar perspective projection (PPP) in geometric structures traced from the human face. A technique for the correction of the PPP was calibrated within photographs of two geometric structures obtained with angles of incidence distorted in 80°, 60° and 45°. The technique was performed using ImageJ ® 1.46r (National Institutes of Health, Bethesda, Maryland). The corrected images were compared with photographs of the same object obtained in 90° (reference). In a second step, the technique was validated in a digital human face created using MakeHuman ® 1.0.2 (Free Software Foundation, Massachusetts, EUA) and Blender ® 2.75 (Blender ® Foundation, Amsterdam, Nederland) software packages. The images registered with angular distortion presented a gradual decrease in height when compared to the reference. The digital technique for the correction of the PPP is a valuable tool for forensic applications using photographic imaging modalities, such as forensic facial analysis. Copyright © 2016 Elsevier Ireland Ltd. All rights reserved.

  20. Facial neuropathy with imaging enhancement of the facial nerve: a case report

    PubMed Central

    Mumtaz, Sehreen; Jensen, Matthew B

    2014-01-01

    A young women developed unilateral facial neuropathy 2 weeks after a motor vehicle collision involving fractures of the skull and mandible. MRI showed contrast enhancement of the facial nerve. We review the literature describing facial neuropathy after trauma and facial nerve enhancement patterns with different causes of facial neuropathy. PMID:25574155

  1. Retrospective case series of the imaging findings of facial nerve hemangioma.

    PubMed

    Yue, Yunlong; Jin, Yanfang; Yang, Bentao; Yuan, Hui; Li, Jiandong; Wang, Zhenchang

    2015-09-01

    The aim was to compare high-resolution computed tomography (HRCT) and thin-section magnetic resonance imaging (MRI) findings of facial nerve hemangioma. The HRCT and MRI characteristics of 17 facial nerve hemangiomas diagnosed between 2006 and 2013 were retrospectively analyzed. All patients included in the study suffered from a space-occupying lesion of soft tissues at the geniculate ganglion fossa. Affected nerve was compared for size and shape with the contralateral unaffected nerve. HRCT showed irregular expansion and broadening of the facial nerve canal, damage of the bone wall and destruction of adjacent bone, with "point"-like or "needle"-like calcifications in 14 cases. The average CT value was 320.9 ± 141.8 Hu. Fourteen patients had a widened labyrinthine segment; 6/17 had a tympanic segment widening; 2/17 had a greater superficial petrosal nerve canal involvement, and 2/17 had an affected internal auditory canal (IAC) segment. On MRI, all lesions were significantly enhanced due to high blood supply. Using 2D FSE T2WI, the lesion detection rate was 82.4 % (14/17). 3D fast imaging employing steady-state acquisition (3D FIESTA) revealed the lesions in all patients. HRCT showed that the average number of involved segments in the facial nerve canal was 2.41, while MRI revealed an average of 2.70 segments (P < 0.05). HRCT and MR findings of facial nerve hemangioma were typical, revealing irregular masses growing along the facial nerve canal, with calcifications and rich blood supply. Thin-section enhanced MRI was more accurate in lesion detection and assessment compared with HRCT.

  2. Modeling first impressions from highly variable facial images.

    PubMed

    Vernon, Richard J W; Sutherland, Clare A M; Young, Andrew W; Hartley, Tom

    2014-08-12

    First impressions of social traits, such as trustworthiness or dominance, are reliably perceived in faces, and despite their questionable validity they can have considerable real-world consequences. We sought to uncover the information driving such judgments, using an attribute-based approach. Attributes (physical facial features) were objectively measured from feature positions and colors in a database of highly variable "ambient" face photographs, and then used as input for a neural network to model factor dimensions (approachability, youthful-attractiveness, and dominance) thought to underlie social attributions. A linear model based on this approach was able to account for 58% of the variance in raters' impressions of previously unseen faces, and factor-attribute correlations could be used to rank attributes by their importance to each factor. Reversing this process, neural networks were then used to predict facial attributes and corresponding image properties from specific combinations of factor scores. In this way, the factors driving social trait impressions could be visualized as a series of computer-generated cartoon face-like images, depicting how attributes change along each dimension. This study shows that despite enormous variation in ambient images of faces, a substantial proportion of the variance in first impressions can be accounted for through linear changes in objectively defined features.

  3. Modeling first impressions from highly variable facial images

    PubMed Central

    Vernon, Richard J. W.; Sutherland, Clare A. M.; Young, Andrew W.; Hartley, Tom

    2014-01-01

    First impressions of social traits, such as trustworthiness or dominance, are reliably perceived in faces, and despite their questionable validity they can have considerable real-world consequences. We sought to uncover the information driving such judgments, using an attribute-based approach. Attributes (physical facial features) were objectively measured from feature positions and colors in a database of highly variable “ambient” face photographs, and then used as input for a neural network to model factor dimensions (approachability, youthful-attractiveness, and dominance) thought to underlie social attributions. A linear model based on this approach was able to account for 58% of the variance in raters’ impressions of previously unseen faces, and factor-attribute correlations could be used to rank attributes by their importance to each factor. Reversing this process, neural networks were then used to predict facial attributes and corresponding image properties from specific combinations of factor scores. In this way, the factors driving social trait impressions could be visualized as a series of computer-generated cartoon face-like images, depicting how attributes change along each dimension. This study shows that despite enormous variation in ambient images of faces, a substantial proportion of the variance in first impressions can be accounted for through linear changes in objectively defined features. PMID:25071197

  4. Facial recognition in education system

    NASA Astrophysics Data System (ADS)

    Krithika, L. B.; Venkatesh, K.; Rathore, S.; Kumar, M. Harish

    2017-11-01

    Human beings exploit emotions comprehensively for conveying messages and their resolution. Emotion detection and face recognition can provide an interface between the individuals and technologies. The most successful applications of recognition analysis are recognition of faces. Many different techniques have been used to recognize the facial expressions and emotion detection handle varying poses. In this paper, we approach an efficient method to recognize the facial expressions to track face points and distances. This can automatically identify observer face movements and face expression in image. This can capture different aspects of emotion and facial expressions.

  5. Low-level image properties in facial expressions.

    PubMed

    Menzel, Claudia; Redies, Christoph; Hayn-Leichsenring, Gregor U

    2018-06-04

    We studied low-level image properties of face photographs and analyzed whether they change with different emotional expressions displayed by an individual. Differences in image properties were measured in three databases that depicted a total of 167 individuals. Face images were used either in their original form, cut to a standard format or superimposed with a mask. Image properties analyzed were: brightness, redness, yellowness, contrast, spectral slope, overall power and relative power in low, medium and high spatial frequencies. Results showed that image properties differed significantly between expressions within each individual image set. Further, specific facial expressions corresponded to patterns of image properties that were consistent across all three databases. In order to experimentally validate our findings, we equalized the luminance histograms and spectral slopes of three images from a given individual who showed two expressions. Participants were significantly slower in matching the expression in an equalized compared to an original image triad. Thus, existing differences in these image properties (i.e., spectral slope, brightness or contrast) facilitate emotion detection in particular sets of face images. Copyright © 2018. Published by Elsevier B.V.

  6. Image-based Analysis of Emotional Facial Expressions in Full Face Transplants.

    PubMed

    Bedeloglu, Merve; Topcu, Çagdas; Akgul, Arzu; Döger, Ela Naz; Sever, Refik; Ozkan, Ozlenen; Ozkan, Omer; Uysal, Hilmi; Polat, Ovunc; Çolak, Omer Halil

    2018-01-20

    In this study, it is aimed to determine the degree of the development in emotional expression of full face transplant patients from photographs. Hence, a rehabilitation process can be planned according to the determination of degrees as a later work. As envisaged, in full face transplant cases, the determination of expressions can be confused or cannot be achieved as the healthy control group. In order to perform image-based analysis, a control group consist of 9 healthy males and 2 full-face transplant patients participated in the study. Appearance-based Gabor Wavelet Transform (GWT) and Local Binary Pattern (LBP) methods are adopted for recognizing neutral and 6 emotional expressions which consist of angry, scared, happy, hate, confused and sad. Feature extraction was carried out by using both methods and combination of these methods serially. In the performed expressions, the extracted features of the most distinct zones in the facial area where the eye and mouth region, have been used to classify the emotions. Also, the combination of these region features has been used to improve classifier performance. Control subjects and transplant patients' ability to perform emotional expressions have been determined with K-nearest neighbor (KNN) classifier with region-specific and method-specific decision stages. The results have been compared with healthy group. It has been observed that transplant patients don't reflect some emotional expressions. Also, there were confusions among expressions.

  7. The asymmetric facial skin perfusion distribution of Bell's palsy discovered by laser speckle imaging technology.

    PubMed

    Cui, Han; Chen, Yi; Zhong, Weizheng; Yu, Haibo; Li, Zhifeng; He, Yuhai; Yu, Wenlong; Jin, Lei

    2016-01-01

    Bell's palsy is a kind of peripheral neural disease that cause abrupt onset of unilateral facial weakness. In the pathologic study, it was evidenced that ischemia of facial nerve at the affected side of face existed in Bell's palsy patients. Since the direction of facial nerve blood flow is primarily proximal to distal, facial skin microcirculation would also be affected after the onset of Bell's palsy. Therefore, monitoring the full area of facial skin microcirculation would help to identify the condition of Bell's palsy patients. In this study, a non-invasive, real time and full field imaging technology - laser speckle imaging (LSI) technology was applied for measuring facial skin blood perfusion distribution of Bell's palsy patients. 85 participants with different stage of Bell's palsy were included. Results showed that Bell's palsy patients' facial skin perfusion of affected side was lower than that of the normal side at the region of eyelid, and that the asymmetric distribution of the facial skin perfusion between two sides of eyelid is positively related to the stage of the disease (P <  0.001). During the recovery, the perfusion of affected side of eyelid was increasing to nearly the same with the normal side. This study was a novel application of LSI in evaluating the facial skin perfusion of Bell's palsy patients, and we discovered that the facial skin blood perfusion could reflect the stage of Bell's palsy, which suggested that microcirculation should be investigated in patients with this neurological deficit. It was also suggested LSI as potential diagnostic tool for Bell's palsy.

  8. Analysis of Facial Expression by Taste Stimulation

    NASA Astrophysics Data System (ADS)

    Tobitani, Kensuke; Kato, Kunihito; Yamamoto, Kazuhiko

    In this study, we focused on the basic taste stimulation for the analysis of real facial expressions. We considered that the expressions caused by taste stimulation were unaffected by individuality or emotion, that is, such expressions were involuntary. We analyzed the movement of facial muscles by taste stimulation and compared real expressions with artificial expressions. From the result, we identified an obvious difference between real and artificial expressions. Thus, our method would be a new approach for facial expression recognition.

  9. What does magnetic resonance imaging add to the prenatal ultrasound diagnosis of facial clefts?

    PubMed

    Mailáth-Pokorny, M; Worda, C; Krampl-Bettelheim, E; Watzinger, F; Brugger, P C; Prayer, D

    2010-10-01

    Ultrasound is the modality of choice for prenatal detection of cleft lip and palate. Because its accuracy in detecting facial clefts, especially isolated clefts of the secondary palate, can be limited, magnetic resonance imaging (MRI) is used as an additional method for assessing the fetus. The aim of this study was to investigate the role of fetal MRI in the prenatal diagnosis of facial clefts. Thirty-four pregnant women with a mean gestational age of 26 (range, 19-34) weeks underwent in utero MRI, after ultrasound examination had identified either a facial cleft (n = 29) or another suspected malformation (micrognathia (n = 1), cardiac defect (n = 1), brain anomaly (n = 2) or diaphragmatic hernia (n = 1)). The facial cleft was classified postnatally and the diagnoses were compared with the previous ultrasound findings. There were 11 (32.4%) cases with cleft of the primary palate alone, 20 (58.8%) clefts of the primary and secondary palate and three (8.8%) isolated clefts of the secondary palate. In all cases the primary and secondary palate were visualized successfully with MRI. Ultrasound imaging could not detect five (14.7%) facial clefts and misclassified 15 (44.1%) facial clefts. The MRI classification correlated with the postnatal/postmortem diagnosis. In our hands MRI allows detailed prenatal evaluation of the primary and secondary palate. By demonstrating involvement of the palate, MRI provides better detection and classification of facial clefts than does ultrasound alone. Copyright © 2010 ISUOG. Published by John Wiley & Sons, Ltd.

  10. A neurophysiological study of facial numbness in multiple sclerosis: Integration with clinical data and imaging findings.

    PubMed

    Koutsis, Georgios; Kokotis, Panagiotis; Papagianni, Aikaterini E; Evangelopoulos, Maria-Eleftheria; Kilidireas, Constantinos; Karandreas, Nikolaos

    2016-09-01

    To integrate neurophysiological findings with clinical and imaging data in a consecutive series of multiple sclerosis (MS) patients developing facial numbness during the course of an MS attack. Nine consecutive patients with MS and recent-onset facial numbness were studied clinically, imaged with routine MRI, and assessed neurophysiologically with trigeminal somatosensory evoked potential (TSEP), blink reflex (BR), masseter reflex (MR), facial nerve conduction, facial muscle and masseter EMG studies. All patients had unilateral facial hypoesthesia on examination and lesions in the ipsilateral pontine tegmentum on MRI. All patients had abnormal TSEPs upon stimulation of the affected side, excepting one that was tested following remission of numbness. BR was the second most sensitive neurophysiological method with 6/9 examinations exhibiting an abnormal R1 component. The MR was abnormal in 3/6 patients, always on the affected side. Facial conduction and EMG studies were normal in all patients but one. Facial numbness was always related to abnormal TSEPs. A concomitant R1 abnormality on BR allowed localization of the responsible pontine lesion, which closely corresponded with MRI findings. We conclude that neurophysiological assessment of MS patients with facial numbness is a sensitive tool, which complements MRI, and can improve lesion localization. Copyright © 2016 Elsevier B.V. All rights reserved.

  11. Chondromyxoid fibroma of the mastoid facial nerve canal mimicking a facial nerve schwannoma.

    PubMed

    Thompson, Andrew L; Bharatha, Aditya; Aviv, Richard I; Nedzelski, Julian; Chen, Joseph; Bilbao, Juan M; Wong, John; Saad, Reda; Symons, Sean P

    2009-07-01

    Chondromyxoid fibroma of the skull base is a rare entity. Involvement of the temporal bone is particularly rare. We present an unusual case of progressive facial nerve paralysis with imaging and clinical findings most suggestive of a facial nerve schwannoma. The lesion was tubular in appearance, expanded the mastoid facial nerve canal, protruded out of the stylomastoid foramen, and enhanced homogeneously. The only unusual imaging feature was minor calcification within the tumor. Surgery revealed an irregular, cystic lesion. Pathology diagnosed a chondromyxoid fibroma involving the mastoid portion of the facial nerve canal, destroying the facial nerve.

  12. Association of Frontal and Lateral Facial Attractiveness.

    PubMed

    Gu, Jeffrey T; Avilla, David; Devcic, Zlatko; Karimi, Koohyar; Wong, Brian J F

    2018-01-01

    Despite the large number of studies focused on defining frontal or lateral facial attractiveness, no reports have examined whether a significant association between frontal and lateral facial attractiveness exists. To examine the association between frontal and lateral facial attractiveness and to identify anatomical features that may influence discordance between frontal and lateral facial beauty. Paired frontal and lateral facial synthetic images of 240 white women (age range, 18-25 years) were evaluated from September 30, 2004, to September 29, 2008, using an internet-based focus group (n = 600) on an attractiveness Likert scale of 1 to 10, with 1 being least attractive and 10 being most attractive. Data analysis was performed from December 6, 2016, to March 30, 2017. The association between frontal and lateral attractiveness scores was determined using linear regression. Outliers were defined as data outside the 95% individual prediction interval. To identify features that contribute to score discordance between frontal and lateral attractiveness scores, each of these image pairs were scrutinized by an evaluator panel for facial features that were present in the frontal or lateral projections and absent in the other respective facial projections. Attractiveness scores obtained from internet-based focus groups. For the 240 white women studied (mean [SD] age, 21.4 [2.2] years), attractiveness scores ranged from 3.4 to 9.5 for frontal images and 3.3 to 9.4 for lateral images. The mean (SD) frontal attractiveness score was 6.9 (1.4), whereas the mean (SD) lateral attractiveness score was 6.4 (1.3). Simple linear regression of frontal and lateral attractiveness scores resulted in a coefficient of determination of r2 = 0.749. Eight outlier pairs were identified and analyzed by panel evaluation. Panel evaluation revealed no clinically applicable association between frontal and lateral images among outliers; however, contributory facial features were suggested

  13. New method for analysis of facial growth in a pediatric reconstructed mandible.

    PubMed

    Kau, Chung How; Kamel, Sherif Galal; Wilson, Jim; Wong, Mark E

    2011-04-01

    The aim of this article was to present a new method of analysis for the assessment of facial growth and morphology after surgical resection of the mandible in a growing patient. This was a 2-year longitudinal study of facial growth in a child who had undergone segmental resection of the mandible with immediate reconstruction as a treatment for juvenile aggressive fibromatosis. Three-dimensional digital stereo-photogrammteric cameras were used for image acquisition at several follow-up intervals: immediate, 6 months, and 2 years postresection. After processing and superimposition, shell-to-shell deviation maps were used for the analysis of the facial growth pattern and its deviation from normal growth. The changes were seen as mean surface changes and color maps. An average constructed female face from a previous study was used as a reference for a normal growth pattern. The patient showed significant growth during this period. Positive changes took place around the nose, lateral brow area, and lower lip and chin, whereas negative changes were evident at the lower lips and cheeks area. An increase in the vertical dimension of the face at the chin region was also seen prominently. Three-dimensional digital stereo-photogrammetry can be used as an objective, noninvasive method for quantifying and monitoring facial growth and its abnormalities. Copyright © 2011 American Association of Orthodontists. Published by Mosby, Inc. All rights reserved.

  14. Combat-related facial burns: analysis of strategic pitfalls.

    PubMed

    Johnson, Benjamin W; Madson, Andrew Q; Bong-Thakur, Sarah; Tucker, David; Hale, Robert G; Chan, Rodney K

    2015-01-01

    Burns constitute approximately 10% of all combat-related injuries to the head and neck region. We postulated that the combat environment presents unique challenges not commonly encountered among civilian injuries. The purpose of the present study was to determine the features commonly seen among combat facial burns that will result in therapeutic challenges and might contribute to undesired outcomes. The present study was a retrospective study performed using a query of the Burn Registry at the US Army Institute of Surgical Research Burn Center for all active duty facial burn admissions from October 2001 to February 2011. The demographic data, total body surface area of the burn, facial region body surface area involvement, and dates of injury, first operation, and first facial operation were tabulated and compared. A subset analysis of severe facial burns, defined by a greater than 7% facial region body surface area, was performed with a thorough medical record review to determine the presence of associated injuries. Of all the military burn injuries, 67.1% (n = 558) involved the face. Of these, 81.3% (n = 454) were combat related. The combat facial burns had a mean total body surface area of 21.4% and a mean facial region body surface area of 3.2%. The interval from the date of the injury to the first operative encounter was 6.6 ± 0.8 days and was 19.8 ± 2.0 days to the first facial operation. A subset analysis of the severe facial burns revealed that the first facial operation and the definitive coverage operation was performed at 13.45 ± 2.6 days and 31.9 ± 4.1 days after the injury, respectively. The mortality rate for this subset of patients was 32% (n = 10), with a high rate of associated inhalational injuries (61%, n = 19), limb amputations (29%, n = 9), and facial allograft usage (48%, n = 15) and a mean facial autograft thickness of 10.5/1,000th in. Combat-related facial burns present multiple challenges, which can contribute to suboptimal long

  15. A facial expression image database and norm for Asian population: a preliminary report

    NASA Astrophysics Data System (ADS)

    Chen, Chien-Chung; Cho, Shu-ling; Horszowska, Katarzyna; Chen, Mei-Yen; Wu, Chia-Ching; Chen, Hsueh-Chih; Yeh, Yi-Yu; Cheng, Chao-Min

    2009-01-01

    We collected 6604 images of 30 models in eight types of facial expression: happiness, anger, sadness, disgust, fear, surprise, contempt and neutral. Among them, 406 most representative images from 12 models were rated by more than 200 human raters for perceived emotion category and intensity. Such large number of emotion categories, models and raters is sufficient for most serious expression recognition research both in psychology and in computer science. All the models and raters are of Asian background. Hence, this database can also be used when the culture background is a concern. In addition, 43 landmarks each of the 291 rated frontal view images were identified and recorded. This information should facilitate feature based research of facial expression. Overall, the diversity in images and richness in information should make our database and norm useful for a wide range of research.

  16. Changing perception: facial reanimation surgery improves attractiveness and decreases negative facial perception.

    PubMed

    Dey, Jacob K; Ishii, Masaru; Boahene, Kofi D O; Byrne, Patrick J; Ishii, Lisa E

    2014-01-01

    Determine the effect of facial reanimation surgery on observer-graded attractiveness and negative facial perception of patients with facial paralysis. Randomized controlled experiment. Ninety observers viewed images of paralyzed faces, smiling and in repose, before and after reanimation surgery, as well as normal comparison faces. Observers rated the attractiveness of each face and characterized the paralyzed faces by rating severity, disfigured/bothersome, and importance to repair. Iterated factor analysis indicated these highly correlated variables measure a common domain, so they were combined to create the disfigured, important to repair, bothersome, severity (DIBS) factor score. Mixed effects linear regression determined the effect of facial reanimation surgery on attractiveness and DIBS score. Facial paralysis induces an attractiveness penalty of 2.51 on a 10-point scale for faces in repose and 3.38 for smiling faces. Mixed effects linear regression showed that reanimation surgery improved attractiveness for faces both in repose and smiling by 0.84 (95% confidence interval [CI]: 0.67, 1.01) and 1.24 (95% CI: 1.07, 1.42) respectively. Planned hypothesis tests confirmed statistically significant differences in attractiveness ratings between postoperative and normal faces, indicating attractiveness was not completely normalized. Regression analysis also showed that reanimation surgery decreased DIBS by 0.807 (95% CI: 0.704, 0.911) for faces in repose and 0.989 (95% CI: 0.886, 1.093), an entire standard deviation, for smiling faces. Facial reanimation surgery increases attractiveness and decreases negative facial perception of patients with facial paralysis. These data emphasize the need to optimize reanimation surgery to restore not only function, but also symmetry and cosmesis to improve facial perception and patient quality of life. © 2013 The American Laryngological, Rhinological and Otological Society, Inc.

  17. Automatic three-dimensional quantitative analysis for evaluation of facial movement.

    PubMed

    Hontanilla, B; Aubá, C

    2008-01-01

    The aim of this study is to present a new 3D capture system of facial movements called FACIAL CLIMA. It is an automatic optical motion system that involves placing special reflecting dots on the subject's face and video recording with three infrared-light cameras the subject performing several face movements such as smile, mouth puckering, eye closure and forehead elevation. Images from the cameras are automatically processed with a software program that generates customised information such as 3D data on velocities and areas. The study has been performed in 20 healthy volunteers. The accuracy of the measurement process and the intrarater and interrater reliabilities have been evaluated. Comparison of a known distance and angle with those obtained by FACIAL CLIMA shows that this system is accurate to within 0.13 mm and 0.41 degrees . In conclusion, the accuracy of the FACIAL CLIMA system for evaluation of facial movements is demonstrated and also the high intrarater and interrater reliability. It has advantages with respect to other systems that have been developed for evaluation of facial movements, such as short calibration time, short measuring time, easiness to use and it provides not only distances but also velocities and areas. Thus the FACIAL CLIMA system could be considered as an adequate tool to assess the outcome of facial paralysis reanimation surgery. Thus, patients with facial paralysis could be compared between surgical centres such that effectiveness of facial reanimation operations could be evaluated.

  18. Traumatic facial nerve neuroma with facial palsy presenting in infancy.

    PubMed

    Clark, James H; Burger, Peter C; Boahene, Derek Kofi; Niparko, John K

    2010-07-01

    To describe the management of traumatic neuroma of the facial nerve in a child and literature review. Sixteen-month-old male subject. Radiological imaging and surgery. Facial nerve function. The patient presented at 16 months with a right facial palsy and was found to have a right facial nerve traumatic neuroma. A transmastoid, middle fossa resection of the right facial nerve lesion was undertaken with a successful facial nerve-to-hypoglossal nerve anastomosis. The facial palsy improved postoperatively. A traumatic neuroma should be considered in an infant who presents with facial palsy, even in the absence of an obvious history of trauma. The treatment of such lesion is complex in any age group but especially in young children. Symptoms, age, lesion size, growth rate, and facial nerve function determine the appropriate management.

  19. The effect of image quality and forensic expertise in facial image comparisons.

    PubMed

    Norell, Kristin; Läthén, Klas Brorsson; Bergström, Peter; Rice, Allyson; Natu, Vaidehi; O'Toole, Alice

    2015-03-01

    Images of perpetrators in surveillance video footage are often used as evidence in court. In this study, identification accuracy was compared for forensic experts and untrained persons in facial image comparisons as well as the impact of image quality. Participants viewed thirty image pairs and were asked to rate the level of support garnered from their observations for concluding whether or not the two images showed the same person. Forensic experts reached their conclusions with significantly fewer errors than did untrained participants. They were also better than novices at determining when two high-quality images depicted the same person. Notably, lower image quality led to more careful conclusions by experts, but not for untrained participants. In summary, the untrained participants had more false negatives and false positives than experts, which in the latter case could lead to a higher risk of an innocent person being convicted for an untrained witness. © 2014 American Academy of Forensic Sciences.

  20. Mutual information-based facial expression recognition

    NASA Astrophysics Data System (ADS)

    Hazar, Mliki; Hammami, Mohamed; Hanêne, Ben-Abdallah

    2013-12-01

    This paper introduces a novel low-computation discriminative regions representation for expression analysis task. The proposed approach relies on interesting studies in psychology which show that most of the descriptive and responsible regions for facial expression are located around some face parts. The contributions of this work lie in the proposition of new approach which supports automatic facial expression recognition based on automatic regions selection. The regions selection step aims to select the descriptive regions responsible or facial expression and was performed using Mutual Information (MI) technique. For facial feature extraction, we have applied Local Binary Patterns Pattern (LBP) on Gradient image to encode salient micro-patterns of facial expressions. Experimental studies have shown that using discriminative regions provide better results than using the whole face regions whilst reducing features vector dimension.

  1. Qualitative and Quantitative Analysis for Facial Complexion in Traditional Chinese Medicine

    PubMed Central

    Zhao, Changbo; Li, Guo-zheng; Li, Fufeng; Wang, Zhi; Liu, Chang

    2014-01-01

    Facial diagnosis is an important and very intuitive diagnostic method in Traditional Chinese Medicine (TCM). However, due to its qualitative and experience-based subjective property, traditional facial diagnosis has a certain limitation in clinical medicine. The computerized inspection method provides classification models to recognize facial complexion (including color and gloss). However, the previous works only study the classification problems of facial complexion, which is considered as qualitative analysis in our perspective. For quantitative analysis expectation, the severity or degree of facial complexion has not been reported yet. This paper aims to make both qualitative and quantitative analysis for facial complexion. We propose a novel feature representation of facial complexion from the whole face of patients. The features are established with four chromaticity bases splitting up by luminance distribution on CIELAB color space. Chromaticity bases are constructed from facial dominant color using two-level clustering; the optimal luminance distribution is simply implemented with experimental comparisons. The features are proved to be more distinctive than the previous facial complexion feature representation. Complexion recognition proceeds by training an SVM classifier with the optimal model parameters. In addition, further improved features are more developed by the weighted fusion of five local regions. Extensive experimental results show that the proposed features achieve highest facial color recognition performance with a total accuracy of 86.89%. And, furthermore, the proposed recognition framework could analyze both color and gloss degrees of facial complexion by learning a ranking function. PMID:24967342

  2. Effects of a small talking facial image on autonomic activity: the moderating influence of dispositional BIS and BAS sensitivities and emotions.

    PubMed

    Ravaja, Niklas

    2004-01-01

    We examined the moderating influence of dispositional behavioral inhibition system and behavioral activation system (BAS) sensitivities, Negative Affect, and Positive Affect on the relationship between a small moving vs. static facial image and autonomic responses when viewing/listening to news messages read by a newscaster among 36 young adults. Autonomic parameters measured were respiratory sinus arrhythmia (RSA), low-frequency (LF) component of heart rate variability (HRV), electrodermal activity, and pulse transit time (PTT). The results showed that dispositional BAS sensitivity, particularly BAS Fun Seeking, and Negative Affect interacted with facial image motion in predicting autonomic nervous system activity. A moving facial image was related to lower RSA and LF component of HRV and shorter PTTs as compared to a static facial image among high BAS individuals. Even a small talking facial image may contribute to sustained attentional engagement among high BAS individuals, given that the BAS directs attention toward the positive cue and a moving social stimulus may act as a positive incentive for high BAS individuals.

  3. Facial Phenotyping by Quantitative Photography Reflects Craniofacial Morphology Measured on Magnetic Resonance Imaging in Icelandic Sleep Apnea Patients

    PubMed Central

    Sutherland, Kate; Schwab, Richard J.; Maislin, Greg; Lee, Richard W.W.; Benedikstdsottir, Bryndis; Pack, Allan I.; Gislason, Thorarinn; Juliusson, Sigurdur; Cistulli, Peter A.

    2014-01-01

    Study Objectives: (1) To determine whether facial phenotype, measured by quantitative photography, relates to underlying craniofacial obstructive sleep apnea (OSA) risk factors, measured with magnetic resonance imaging (MRI); (2) To assess whether these associations are independent of body size and obesity. Design: Cross-sectional cohort. Setting: Landspitali, The National University Hospital, Iceland. Participants: One hundred forty patients (87.1% male) from the Icelandic Sleep Apnea Cohort who had both calibrated frontal and profile craniofacial photographs and upper airway MRI. Mean ± standard deviation age 56.1 ± 10.4 y, body mass index 33.5 ± 5.05 kg/m2, with on-average severe OSA (apnea-hypopnea index 45.4 ± 19.7 h-1). Interventions: N/A. Measurements and Results: Relationships between surface facial dimensions (photos) and facial bony dimensions and upper airway soft-tissue volumes (MRI) was assessed using canonical correlation analysis. Photo and MRI craniofacial datasets related in four significant canonical correlations, primarily driven by measurements of (1) maxillary-mandibular relationship (r = 0.8, P < 0.0001), (2) lower face height (r = 0.76, P < 0.0001), (3) mandibular length (r = 0.67, P < 0.0001), and (4) tongue volume (r = 0.52, P = 0.01). Correlations 1, 2, and 3 were unchanged when controlled for weight and neck and waist circumference. However, tongue volume was no longer significant, suggesting facial dimensions relate to tongue volume as a result of obesity. Conclusions: Significant associations were found between craniofacial variable sets from facial photography and MRI. This study confirms that facial photographic phenotype reflects underlying aspects of craniofacial skeletal abnormalities associated with OSA. Therefore, facial photographic phenotyping may be a useful tool to assess intermediate phenotypes for OSA, particularly in large-scale studies. Citation: Sutherland K, Schwab RJ, Maislin G, Lee RW, Benedikstdsottir B, Pack AI

  4. Police witness identification images: a geometric morphometric analysis.

    PubMed

    Hayes, Susan; Tullberg, Cameron

    2012-11-01

    Research into witness identification images typically occurs within the laboratory and involves subjective likeness and recognizability judgments. This study analyzed whether actual witness identification images systematically alter the facial shapes of the suspects described. The shape analysis tool, geometric morphometrics, was applied to 46 homologous facial landmarks displayed on 50 witness identification images and their corresponding arrest photographs, using principal component analysis and multivariate regressions. The results indicate that compared with arrest photographs, witness identification images systematically depict suspects with lowered and medially located eyebrows (p = <0.000001). This was found to occur independently of the Police Artist, and did not occur with composites produced under laboratory conditions. There are several possible explanations for this finding, including any, or all, of the following: The suspect was frowning at the time of the incident, the witness had negative feelings toward the suspect, this is an effect of unfamiliar face processing, the suspect displayed fear at the time of their arrest photograph. © 2012 American Academy of Forensic Sciences.

  5. Cost-effective system for facial imaging and three-dimensional reconstruction

    NASA Astrophysics Data System (ADS)

    Shokouhi, S. B.; Monro, D. M.; Sherlock, Barry G.

    1998-06-01

    Three dimensional (3-D) images have recently received wide attention in applications involving medical treatment. Most current 3-D imaging methods focus on the internal organs of the body. However, several medical image applications such as plastic surgery, body deformities, rehabilitation, dental surgery and orthodontics, make use of the surface contours of the body. Several techniques are currently available for producing 3-D images of the body surface and most of the systems which implement these techniques are expensive, requiring complex equipment with highly trained operators. The research involves the development of a simple, low cost and non-invasive contour capturing method for facial surfaces. This is achieved using the structured light technique, employing a standard commercial slide projector, CCD camera and a frame-grabber card linked to a PC. Structured light has already been used for many applications, but only to a limited extent in the clinical environment. All current implementations involve extensive manual intervention by highly skilled operators and this has proven to be a serious hindrance to clinical acceptance of 3-D imaging. A primary objective of this work is to minimize the amount of manual intervention required, so that the system can be used by clinicians who do not have specialist training in the use of this equipment. The eventual aim is to provide a software assisted surgical procedure, which by merging the facial data, allows the manipulation of soft tissue and gives the facility to predict and monitor post-surgical appearance. The research focuses on how the images are obtained using the structured light optic system and the subsequent image processing of data to give a realistic 3-D image.

  6. Computer-Aided Recognition of Facial Attributes for Fetal Alcohol Spectrum Disorders.

    PubMed

    Valentine, Matthew; Bihm, Dustin C J; Wolf, Lior; Hoyme, H Eugene; May, Philip A; Buckley, David; Kalberg, Wendy; Abdul-Rahman, Omar A

    2017-12-01

    To compare the detection of facial attributes by computer-based facial recognition software of 2-D images against standard, manual examination in fetal alcohol spectrum disorders (FASD). Participants were gathered from the Fetal Alcohol Syndrome Epidemiology Research database. Standard frontal and oblique photographs of children were obtained during a manual, in-person dysmorphology assessment. Images were submitted for facial analysis conducted by the facial dysmorphology novel analysis technology (an automated system), which assesses ratios of measurements between various facial landmarks to determine the presence of dysmorphic features. Manual blinded dysmorphology assessments were compared with those obtained via the computer-aided system. Areas under the curve values for individual receiver-operating characteristic curves revealed the computer-aided system (0.88 ± 0.02) to be comparable to the manual method (0.86 ± 0.03) in detecting patients with FASD. Interestingly, cases of alcohol-related neurodevelopmental disorder (ARND) were identified more efficiently by the computer-aided system (0.84 ± 0.07) in comparison to the manual method (0.74 ± 0.04). A facial gestalt analysis of patients with ARND also identified more generalized facial findings compared to the cardinal facial features seen in more severe forms of FASD. We found there was an increased diagnostic accuracy for ARND via our computer-aided method. As this category has been historically difficult to diagnose, we believe our experiment demonstrates that facial dysmorphology novel analysis technology can potentially improve ARND diagnosis by introducing a standardized metric for recognizing FASD-associated facial anomalies. Earlier recognition of these patients will lead to earlier intervention with improved patient outcomes. Copyright © 2017 by the American Academy of Pediatrics.

  7. Cerebral, facial, and orbital involvement in Erdheim-Chester disease: CT and MR imaging findings.

    PubMed

    Drier, Aurélie; Haroche, Julien; Savatovsky, Julien; Godenèche, Gaelle; Dormont, Didier; Chiras, Jacques; Amoura, Zahir; Bonneville, Fabrice

    2010-05-01

    To retrospectively review the brain magnetic resonance (MR) imaging and computed tomographic (CT) findings in patients with Erdheim-Chester disease (ECD). The ethics committee required neither institutional review board approval nor informed patient consent for retrospective analyses of the patients' medical records and imaging data. The patients' medical files were retrospectively reviewed in accordance with human subject research protocols. Three neuroradiologists in consensus analyzed the signal intensity, location, size, number, and gadolinium uptake of lesions detected on brain MR images obtained in 33 patients with biopsy-proved ECD. Thirty patients had intracranial, facial bone, and/or orbital involvement, and three had normal neurologic imaging findings. The hypothalamic-pituitary axis was involved in 16 (53%) of the 30 patients, with six (20%) cases of micronodular or nodular masses of the infundibular stalk. Meningeal lesions were observed in seven (23%) patients. Three (10%) patients had bilateral symmetric T2 high signal intensity in the dentate nucleus areas, and five (17%) had multiple intraaxial enhancing masses. Striking intracranial periarterial infiltration was observed in three (10%) patients. Another patient (3%) had a lesion in the lumen of the superior sagittal sinus. Nine (30%) patients had orbital involvement. Twenty-four (80%) patients had osteosclerosis of the facial and/or skull bones. At least two anatomic sites were involved in two-thirds (n = 20) of the patients. Osteosclerosis of the facial bones associated with orbital masses and either meningeal or infundibular stalk masses was seen in eight (27%) patients. Lesions of the brain, meninges, facial bones, and orbits are frequently observed and should be systematically sought on the brain MR and CT images obtained in patients with ECD, even if these patients are asymptomatic. Careful attention should be directed to the periarterial environment.

  8. Observer success rates for identification of 3D surface reconstructed facial images and implications for patient privacy and security

    NASA Astrophysics Data System (ADS)

    Chen, Joseph J.; Siddiqui, Khan M.; Fort, Leslie; Moffitt, Ryan; Juluru, Krishna; Kim, Woojin; Safdar, Nabile; Siegel, Eliot L.

    2007-03-01

    3D and multi-planar reconstruction of CT images have become indispensable in the routine practice of diagnostic imaging. These tools cannot only enhance our ability to diagnose diseases, but can also assist in therapeutic planning as well. The technology utilized to create these can also render surface reconstructions, which may have the undesired potential of providing sufficient detail to allow recognition of facial features and consequently patient identity, leading to violation of patient privacy rights as described in the HIPAA (Health Insurance Portability and Accountability Act) legislation. The purpose of this study is to evaluate whether 3D reconstructed images of a patient's facial features can indeed be used to reliably or confidently identify that specific patient. Surface reconstructed images of the study participants were created used as candidates for matching with digital photographs of participants. Data analysis was performed to determine the ability of observers to successfully match 3D surface reconstructed images of the face with facial photographs. The amount of time required to perform the match was recorded as well. We also plan to investigate the ability of digital masks or physical drapes to conceal patient identity. The recently expressed concerns over the inability to truly "anonymize" CT (and MRI) studies of the head/face/brain are yet to be tested in a prospective study. We believe that it is important to establish whether these reconstructed images are a "threat" to patient privacy/security and if so, whether minimal interventions from a clinical perspective can substantially reduce this possibility.

  9. An Assessment of How Facial Mimicry Can Change Facial Morphology: Implications for Identification.

    PubMed

    Gibelli, Daniele; De Angelis, Danilo; Poppa, Pasquale; Sforza, Chiarella; Cattaneo, Cristina

    2017-03-01

    The assessment of facial mimicry is important in forensic anthropology; in addition, the application of modern 3D image acquisition systems may help for the analysis of facial surfaces. This study aimed at exposing a novel method for comparing 3D profiles in different facial expressions. Ten male adults, aged between 30 and 40 years, underwent acquisitions by stereophotogrammetry (VECTRA-3D ® ) with different expressions (neutral, happy, sad, angry, surprised). The acquisition of each individual was then superimposed on the neutral one according to nine landmarks, and the root mean square (RMS) value between the two expressions was calculated. The highest difference in comparison with the neutral standard was shown by the happy expression (RMS 4.11 mm), followed by the surprised (RMS 2.74 mm), sad (RMS 1.3 mm), and angry ones (RMS 1.21 mm). This pilot study shows that the 3D-3D superimposition may provide reliable results concerning facial alteration due to mimicry. © 2016 American Academy of Forensic Sciences.

  10. A new approach for the analysis of facial growth and age estimation: Iris ratio

    PubMed Central

    Machado, Carlos Eduardo Palhares; Flores, Marta Regina Pinheiro; Lima, Laíse Nascimento Correia; Tinoco, Rachel Lima Ribeiro; Bezerra, Ana Cristina Barreto; Evison, Martin Paul; Guimarães, Marco Aurélio

    2017-01-01

    The study of facial growth is explored in many fields of science, including anatomy, genetics, and forensics. In the field of forensics, it acts as a valuable tool for combating child pornography. The present research proposes a new method, based on relative measurements and fixed references of the human face—specifically considering measurements of the diameter of the iris (iris ratio)—for the analysis of facial growth in association with age in children and sub-adults. The experimental sample consisted of digital photographs of 1000 Brazilian subjects, aged between 6 and 22 years, distributed equally by sex and divided into five specific age groups (6, 10, 14, 18, and 22 year olds ± one month). The software package SAFF-2D® (Forensic Facial Analysis System, Brazilian Federal Police, Brazil) was used for positioning 11 landmarks on the images. Ten measurements were calculated and used as fixed references to evaluate the growth of the other measurements for each age group, as well the accumulated growth (6–22 years old). The Intraclass Correlation Coefficient (ICC) was applied for the evaluation of intra-examiner and inter-examiner reliability within a specific set of images. Pearson’s Correlation Coefficient was used to assess the association between each measurement taken and the respective age groups. ANOVA and Post-hoc Tukey tests were used to search for statistical differences between the age groups. The outcomes indicated that facial structures grow with different timing in children and adolescents. Moreover, the growth allometry expressed in this study may be used to understand what structures have more or less proportional variation in function for the age ranges studied. The diameter of the iris was found to be the most stable measurement compared to the others and represented the best cephalometric measurement as a fixed reference for facial growth ratios (or indices). The method described shows promising potential for forensic applications

  11. A new approach for the analysis of facial growth and age estimation: Iris ratio.

    PubMed

    Machado, Carlos Eduardo Palhares; Flores, Marta Regina Pinheiro; Lima, Laíse Nascimento Correia; Tinoco, Rachel Lima Ribeiro; Franco, Ademir; Bezerra, Ana Cristina Barreto; Evison, Martin Paul; Guimarães, Marco Aurélio

    2017-01-01

    The study of facial growth is explored in many fields of science, including anatomy, genetics, and forensics. In the field of forensics, it acts as a valuable tool for combating child pornography. The present research proposes a new method, based on relative measurements and fixed references of the human face-specifically considering measurements of the diameter of the iris (iris ratio)-for the analysis of facial growth in association with age in children and sub-adults. The experimental sample consisted of digital photographs of 1000 Brazilian subjects, aged between 6 and 22 years, distributed equally by sex and divided into five specific age groups (6, 10, 14, 18, and 22 year olds ± one month). The software package SAFF-2D® (Forensic Facial Analysis System, Brazilian Federal Police, Brazil) was used for positioning 11 landmarks on the images. Ten measurements were calculated and used as fixed references to evaluate the growth of the other measurements for each age group, as well the accumulated growth (6-22 years old). The Intraclass Correlation Coefficient (ICC) was applied for the evaluation of intra-examiner and inter-examiner reliability within a specific set of images. Pearson's Correlation Coefficient was used to assess the association between each measurement taken and the respective age groups. ANOVA and Post-hoc Tukey tests were used to search for statistical differences between the age groups. The outcomes indicated that facial structures grow with different timing in children and adolescents. Moreover, the growth allometry expressed in this study may be used to understand what structures have more or less proportional variation in function for the age ranges studied. The diameter of the iris was found to be the most stable measurement compared to the others and represented the best cephalometric measurement as a fixed reference for facial growth ratios (or indices). The method described shows promising potential for forensic applications, especially

  12. Imitating expressions: emotion-specific neural substrates in facial mimicry.

    PubMed

    Lee, Tien-Wen; Josephs, Oliver; Dolan, Raymond J; Critchley, Hugo D

    2006-09-01

    Intentionally adopting a discrete emotional facial expression can modulate the subjective feelings corresponding to that emotion; however, the underlying neural mechanism is poorly understood. We therefore used functional brain imaging (functional magnetic resonance imaging) to examine brain activity during intentional mimicry of emotional and non-emotional facial expressions and relate regional responses to the magnitude of expression-induced facial movement. Eighteen healthy subjects were scanned while imitating video clips depicting three emotional (sad, angry, happy), and two 'ingestive' (chewing and licking) facial expressions. Simultaneously, facial movement was monitored from displacement of fiducial markers (highly reflective dots) on each subject's face. Imitating emotional expressions enhanced activity within right inferior prefrontal cortex. This pattern was absent during passive viewing conditions. Moreover, the magnitude of facial movement during emotion-imitation predicted responses within right insula and motor/premotor cortices. Enhanced activity in ventromedial prefrontal cortex and frontal pole was observed during imitation of anger, in ventromedial prefrontal and rostral anterior cingulate during imitation of sadness and in striatal, amygdala and occipitotemporal during imitation of happiness. Our findings suggest a central role for right inferior frontal gyrus in the intentional imitation of emotional expressions. Further, by entering metrics for facial muscular change into analysis of brain imaging data, we highlight shared and discrete neural substrates supporting affective, action and social consequences of somatomotor emotional expression.

  13. Effects of Objective 3-Dimensional Measures of Facial Shape and Symmetry on Perceptions of Facial Attractiveness.

    PubMed

    Hatch, Cory D; Wehby, George L; Nidey, Nichole L; Moreno Uribe, Lina M

    2017-09-01

    Meeting patient desires for enhanced facial esthetics requires that providers have standardized and objective methods to measure esthetics. The authors evaluated the effects of objective 3-dimensional (3D) facial shape and asymmetry measurements derived from 3D facial images on perceptions of facial attractiveness. The 3D facial images of 313 adults in Iowa were digitized with 32 landmarks, and objective 3D facial measurements capturing symmetric and asymmetric components of shape variation, centroid size, and fluctuating asymmetry were obtained from the 3D coordinate data using geo-morphometric analyses. Frontal and profile images of study participants were rated for facial attractiveness by 10 volunteers (5 women and 5 men) on a 5-point Likert scale and a visual analog scale. Multivariate regression was used to identify the effects of the objective 3D facial measurements on attractiveness ratings. Several objective 3D facial measurements had marked effects on attractiveness ratings. Shorter facial heights with protrusive chins, midface retrusion, faces with protrusive noses and thin lips, flat mandibular planes with deep labiomental folds, any cants of the lip commissures and floor of the nose, larger faces overall, and increased fluctuating asymmetry were rated as significantly (P < .001) less attractive. Perceptions of facial attractiveness can be explained by specific 3D measurements of facial shapes and fluctuating asymmetry, which have important implications for clinical practice and research. Copyright © 2017 American Association of Oral and Maxillofacial Surgeons. Published by Elsevier Inc. All rights reserved.

  14. A prospective analysis of physical examination findings in the diagnosis of facial fractures: Determining predictive value

    PubMed Central

    Timashpolsky, Alisa; Dagum, Alexander B; Sayeed, Syed M; Romeiser, Jamie L; Rosenfeld, Elisheva A; Conkling, Nicole

    2016-01-01

    BACKGROUND There are >150,000 patient visits per year to emergency rooms for facial trauma. The reliability of a computed tomography (CT) scan has made it the primary modality for diagnosing facial skeletal injury, with the physical examination playing more a cursory role. Knowing the predictive value of physical findings in facial skeletal injuries may enable more appropriate use of imaging and health care resources. OBJECTIVE A blinded prospective study was undertaken to assess the predictive value of physical examination findings in detecting maxillofacial fracture in trauma patients, and in determining whether a patient will require surgical intervention. METHODS Over a four-month period, the authors’ team examined patients admitted with facial trauma to the emergency department of their hospital. The evaluating physician completed a standardized physical examination evaluation form indicating the physical findings. Corresponding CT scans and surgical records were then reviewed, and the results recorded by a plastic surgeon who was blinded to the results of the physical examination. RESULTS A total of 57 patients met the inclusion criteria; there were 44 male and 13 female patients. The sensitivity, specificity, positive predictive value and negative predictive value of grouped physical examination findings were determined in major areas. In further analysis, specific examination findings with n≥9 (15%) were also reported. CONCLUSIONS The data demonstrated a high negative predictive value of at least 90% for orbital floor, zygomatic, mandibular and nasal bone fractures compared with CT scan. Furthermore, none of the patients who did not have a physical examination finding for a particular facial fracture required surgery for that fracture. Thus, the instrument performed well at ruling out fractures in these areas when there were none. Ultimately, these results may help reduce unnecessary radiation and costly imaging in patients with facial trauma without

  15. A prospective analysis of physical examination findings in the diagnosis of facial fractures: Determining predictive value.

    PubMed

    Timashpolsky, Alisa; Dagum, Alexander B; Sayeed, Syed M; Romeiser, Jamie L; Rosenfeld, Elisheva A; Conkling, Nicole

    2016-01-01

    There are >150,000 patient visits per year to emergency rooms for facial trauma. The reliability of a computed tomography (CT) scan has made it the primary modality for diagnosing facial skeletal injury, with the physical examination playing more a cursory role. Knowing the predictive value of physical findings in facial skeletal injuries may enable more appropriate use of imaging and health care resources. A blinded prospective study was undertaken to assess the predictive value of physical examination findings in detecting maxillofacial fracture in trauma patients, and in determining whether a patient will require surgical intervention. Over a four-month period, the authors' team examined patients admitted with facial trauma to the emergency department of their hospital. The evaluating physician completed a standardized physical examination evaluation form indicating the physical findings. Corresponding CT scans and surgical records were then reviewed, and the results recorded by a plastic surgeon who was blinded to the results of the physical examination. A total of 57 patients met the inclusion criteria; there were 44 male and 13 female patients. The sensitivity, specificity, positive predictive value and negative predictive value of grouped physical examination findings were determined in major areas. In further analysis, specific examination findings with n≥9 (15%) were also reported. The data demonstrated a high negative predictive value of at least 90% for orbital floor, zygomatic, mandibular and nasal bone fractures compared with CT scan. Furthermore, none of the patients who did not have a physical examination finding for a particular facial fracture required surgery for that fracture. Thus, the instrument performed well at ruling out fractures in these areas when there were none. Ultimately, these results may help reduce unnecessary radiation and costly imaging in patients with facial trauma without facial fractures.

  16. A Neural Basis of Facial Action Recognition in Humans

    PubMed Central

    Srinivasan, Ramprakash; Golomb, Julie D.

    2016-01-01

    By combining different facial muscle actions, called action units, humans can produce an extraordinarily large number of facial expressions. Computational models and studies in cognitive science and social psychology have long hypothesized that the brain needs to visually interpret these action units to understand other people's actions and intentions. Surprisingly, no studies have identified the neural basis of the visual recognition of these action units. Here, using functional magnetic resonance imaging and an innovative machine learning analysis approach, we identify a consistent and differential coding of action units in the brain. Crucially, in a brain region thought to be responsible for the processing of changeable aspects of the face, multivoxel pattern analysis could decode the presence of specific action units in an image. This coding was found to be consistent across people, facilitating the estimation of the perceived action units on participants not used to train the multivoxel decoder. Furthermore, this coding of action units was identified when participants attended to the emotion category of the facial expression, suggesting an interaction between the visual analysis of action units and emotion categorization as predicted by the computational models mentioned above. These results provide the first evidence for a representation of action units in the brain and suggest a mechanism for the analysis of large numbers of facial actions and a loss of this capacity in psychopathologies. SIGNIFICANCE STATEMENT Computational models and studies in cognitive and social psychology propound that visual recognition of facial expressions requires an intermediate step to identify visible facial changes caused by the movement of specific facial muscles. Because facial expressions are indeed created by moving one's facial muscles, it is logical to assume that our visual system solves this inverse problem. Here, using an innovative machine learning method and

  17. Retrospective single center study of the efficacy of large spot 532 nm laser for the treatment of facial capillary malformations in 44 patients with the use of three-dimensional image analysis.

    PubMed

    Kwiek, Bartłomiej; Rożalski, Michał; Kowalewski, Cezary; Ambroziak, Marcin

    2017-10-01

    We wanted to asses the efficacy of large spot 532 nm laser for the treatment of facial capillary malformations with the use of three-dimensional (3D) image analysis. Retrospective single center study on previously non-treated patients with facial capillary malformations (CM) was performed. A total of 44 consecutive Caucasian patients aged 5-66 were included. Patients had 3D photography performed before and after and had at least one single session of treatment with 532 nm neodymium-doped yttrium aluminum garnet (Nd:YAG) laser with contact cooling, fluencies ranging from 8 to 11.5 J/cm 2 , pulse duration ranging from 5 to 9 milliseconds and spot size ranging from 5 to 10 mm. Objective analysis of percentage improvement based on 3D digital assessment of combined color and area improvement (global clearance effect [GCE]) were performed. Median maximal improvement achieved during the treatment (GCE max ) was 70.4%. Mean number of laser procedures required to achieve this improvement was 7.1 (ranging from 2 to 14)). Improvement of minimum 25% (GCE 25) was achieved by all patients, of minimum 50% (GCE 50) by 77.3%, of minimum 75% (GCE 75) by 38.6%, and of minimum 90% (GCE 90) by 13.64. Large spot 532 nm laser is highly effective in the treatment of facial CM. 3D color and area image analysis provides an objective method to compare different methods of facial CM treatment in future studies. Lasers Surg. Med. 49:743-749, 2017. © 2017 Wiley Periodicals, Inc. © 2017 Wiley Periodicals, Inc.

  18. Case analysis of temporal bone lesions with facial paralysis as main manifestation and literature review.

    PubMed

    Chen, Wen-Jing; Ye, Jing-Ying; Li, Xin; Xu, Jia; Yi, Hai-Jin

    2017-08-23

    This study aims to discuss clinical characteristics, image manifestation and treatment methods of temporal bone lesions with facial paralysis as the main manifestation for deepening the understanding of such type of lesions and reducing erroneous and missed diagnosis. The clinical data of 16 patients with temporal bone lesions and facial paralysis as main manifestation, who were diagnosed and treated from 2009 to 2016, were retrospectively analyzed. Among these patients, six patients had congenital petrous bone cholesteatoma (PBC), nine patients had facial nerve schwannoma, and one patient had facial nerve hemangioma. All the patients had an experience of long-term erroneous diagnosis. The lesions were completely excised by surgery. PBC and primary facial nerve tumors were pathologically confirmed. Facial-hypoglossal nerve anastomosis was performed on two patients. HB grade VI was recovered to HB grade V in one patient. The anastomosis failed due to severe facial nerve fibrosis in one patient. Hence, HB remained at grade VI. Postoperative recovery was good for all patients. No lesion recurrence was observed after 1-6 years of follow-up. For the patients with progressive or complete facial paralysis, imaging examination should be perfected in a timely manner. Furthermore, PBC, primary facial nerve tumors and other temporal bone space-occupying lesions should be eliminated. Lesions should be timely detected and proper intervention should be conducted, in order to reduce operation difficulty and complications, and increase the opportunity of facial nerve function reconstruction.

  19. Facial morphologies of an adult Egyptian population and an adult Houstonian white population compared using 3D imaging.

    PubMed

    Seager, Dennis Craig; Kau, Chung How; English, Jeryl D; Tawfik, Wael; Bussa, Harry I; Ahmed, Abou El Yazeed M

    2009-09-01

    To compare the facial morphologies of an adult Egyptian population with those of a Houstonian white population. The three-dimensional (3D) images were acquired via a commercially available stereophotogrammetric camera capture system. The 3dMDface System photographed 186 subjects from two population groups (Egypt and Houston). All of the participants from both population groups were between 18 and 30 years of age and had no apparent facial anomalies. All facial images were overlaid and superimposed, and a complex mathematical algorithm was performed to generate a composite facial average (one male and one female) for each subgroup (EGY-M: Egyptian male subjects; EGY-F: Egyptian female subjects; HOU-M: Houstonian male subjects; and HOU-F: Houstonian female subjects). The computer-generated facial averages were superimposed based on a previously validated superimposition method, and the facial differences were evaluated and quantified. Distinct facial differences were evident between the subgroups evaluated, involving various regions of the face including the slant of the forehead, and the nasal, malar, and labial regions. Overall, the mean facial differences between the Egyptian and Houstonian female subjects were 1.33 +/- 0.93 mm, while the differences in Egyptian and Houstonian male subjects were 2.32 +/- 2.23 mm. The range of differences for the female population pairings and the male population pairings were 14.34 mm and 13.71 mm, respectively. The average adult Egyptian and white Houstonian face possess distinct differences. Different populations and ethnicities have different facial features and averages.

  20. Facial nerve paralysis secondary to occult malignant neoplasms.

    PubMed

    Boahene, Derek O; Olsen, Kerry D; Driscoll, Colin; Lewis, Jean E; McDonald, Thomas J

    2004-04-01

    This study reviewed patients with unilateral facial paralysis and normal clinical and imaging findings who underwent diagnostic facial nerve exploration. Study design and setting Fifteen patients with facial paralysis and normal findings were seen in the Mayo Clinic Department of Otorhinolaryngology. Eleven patients were misdiagnosed as having Bell palsy or idiopathic paralysis. Progressive facial paralysis with sequential involvement of adjacent facial nerve branches occurred in all 15 patients. Seven patients had a history of regional skin squamous cell carcinoma, 13 patients had surgical exploration to rule out a neoplastic process, and 2 patients had negative exploration. At last follow-up, 5 patients were alive. Patients with facial paralysis and normal clinical and imaging findings should be considered for facial nerve exploration when the patient has a history of pain or regional skin cancer, involvement of other cranial nerves, and prolonged facial paralysis. Occult malignancy of the facial nerve may cause unilateral facial paralysis in patients with normal clinical and imaging findings.

  1. Facial Structure Analysis Separates Autism Spectrum Disorders into Meaningful Clinical Subgroups

    ERIC Educational Resources Information Center

    Obafemi-Ajayi, Tayo; Miles, Judith H.; Takahashi, T. Nicole; Qi, Wenchuan; Aldridge, Kristina; Zhang, Minqi; Xin, Shi-Qing; He, Ying; Duan, Ye

    2015-01-01

    Varied cluster analysis were applied to facial surface measurements from 62 prepubertal boys with essential autism to determine whether facial morphology constitutes viable biomarker for delineation of discrete Autism Spectrum Disorders (ASD) subgroups. Earlier study indicated utility of facial morphology for autism subgrouping (Aldridge et al. in…

  2. Facial color processing in the face-selective regions: an fMRI study.

    PubMed

    Nakajima, Kae; Minami, Tetsuto; Tanabe, Hiroki C; Sadato, Norihiro; Nakauchi, Shigeki

    2014-09-01

    Facial color is important information for social communication as it provides important clues to recognize a person's emotion and health condition. Our previous EEG study suggested that N170 at the left occipito-temporal site is related to facial color processing (Nakajima et al., [2012]: Neuropsychologia 50:2499-2505). However, because of the low spatial resolution of EEG experiment, the brain region is involved in facial color processing remains controversial. In the present study, we examined the neural substrates of facial color processing using functional magnetic resonance imaging (fMRI). We measured brain activity from 25 subjects during the presentation of natural- and bluish-colored face and their scrambled images. The bilateral fusiform face (FFA) area and occipital face area (OFA) were localized by the contrast of natural-colored faces versus natural-colored scrambled images. Moreover, region of interest (ROI) analysis showed that the left FFA was sensitive to facial color, whereas the right FFA and the right and left OFA were insensitive to facial color. In combination with our previous EEG results, these data suggest that the left FFA may play an important role in facial color processing. Copyright © 2014 Wiley Periodicals, Inc.

  3. The assessment of facial variation in 4747 British school children.

    PubMed

    Toma, Arshed M; Zhurov, Alexei I; Playle, Rebecca; Marshall, David; Rosin, Paul L; Richmond, Stephen

    2012-12-01

    The aim of this study is to identify key components contributing to facial variation in a large population-based sample of 15.5-year-old children (2514 females and 2233 males). The subjects were recruited from the Avon Longitudinal Study of Parents and Children. Three-dimensional facial images were obtained for each subject using two high-resolution Konica Minolta laser scanners. Twenty-one reproducible facial landmarks were identified and their coordinates were recorded. The facial images were registered using Procrustes analysis. Principal component analysis was then employed to identify independent groups of correlated coordinates. For the total data set, 14 principal components (PCs) were identified which explained 82 per cent of the total variance, with the first three components accounting for 46 per cent of the variance. Similar results were obtained for males and females separately with only subtle gender differences in some PCs. Facial features may be treated as a multidimensional statistical continuum with respect to the PCs. The first three PCs characterize the face in terms of height, width, and prominence of the nose. The derived PCs may be useful to identify and classify faces according to a scale of normality.

  4. Facial Affect Recognition Using Regularized Discriminant Analysis-Based Algorithms

    NASA Astrophysics Data System (ADS)

    Lee, Chien-Cheng; Huang, Shin-Sheng; Shih, Cheng-Yuan

    2010-12-01

    This paper presents a novel and effective method for facial expression recognition including happiness, disgust, fear, anger, sadness, surprise, and neutral state. The proposed method utilizes a regularized discriminant analysis-based boosting algorithm (RDAB) with effective Gabor features to recognize the facial expressions. Entropy criterion is applied to select the effective Gabor feature which is a subset of informative and nonredundant Gabor features. The proposed RDAB algorithm uses RDA as a learner in the boosting algorithm. The RDA combines strengths of linear discriminant analysis (LDA) and quadratic discriminant analysis (QDA). It solves the small sample size and ill-posed problems suffered from QDA and LDA through a regularization technique. Additionally, this study uses the particle swarm optimization (PSO) algorithm to estimate optimal parameters in RDA. Experiment results demonstrate that our approach can accurately and robustly recognize facial expressions.

  5. Adaptation of facial synthesis to parameter analysis in MPEG-4 visual communication

    NASA Astrophysics Data System (ADS)

    Yu, Lu; Zhang, Jingyu; Liu, Yunhai

    2000-12-01

    In MPEG-4, Facial Definition Parameters (FDPs) and Facial Animation Parameters (FAPs) are defined to animate 1 a facial object. Most of the previous facial animation reconstruction systems were focused on synthesizing animation from manually or automatically generated FAPs but not the FAPs extracted from natural video scene. In this paper, an analysis-synthesis MPEG-4 visual communication system is established, in which facial animation is reconstructed from FAPs extracted from natural video scene.

  6. Preservation of Facial Nerve Function Repaired by Using Fibrin Glue-Coated Collagen Fleece for a Totally Transected Facial Nerve during Vestibular Schwannoma Surgery

    PubMed Central

    Choi, Kyung-Sik; Kim, Min-Su; Jang, Sung-Ho

    2014-01-01

    Recently, the increasing rates of facial nerve preservation after vestibular schwannoma (VS) surgery have been achieved. However, the management of a partially or completely damaged facial nerve remains an important issue. The authors report a patient who was had a good recovery after a facial nerve reconstruction using fibrin glue-coated collagen fleece for a totally transected facial nerve during VS surgery. And, we verifed the anatomical preservation and functional outcome of the facial nerve with postoperative diffusion tensor (DT) imaging facial nerve tractography, electroneurography (ENoG) and House-Brackmann (HB) grade. DT imaging tractography at the 3rd postoperative day revealed preservation of facial nerve. And facial nerve degeneration ratio was 94.1% at 7th postoperative day ENoG. At postoperative 3 months and 1 year follow-up examination with DT imaging facial nerve tractography and ENoG, good results for facial nerve function were observed. PMID:25024825

  7. Perceptual expertise in forensic facial image comparison

    PubMed Central

    White, David; Phillips, P. Jonathon; Hahn, Carina A.; Hill, Matthew; O'Toole, Alice J.

    2015-01-01

    Forensic facial identification examiners are required to match the identity of faces in images that vary substantially, owing to changes in viewing conditions and in a person's appearance. These identifications affect the course and outcome of criminal investigations and convictions. Despite calls for research on sources of human error in forensic examination, existing scientific knowledge of face matching accuracy is based, almost exclusively, on people without formal training. Here, we administered three challenging face matching tests to a group of forensic examiners with many years' experience of comparing face images for law enforcement and government agencies. Examiners outperformed untrained participants and computer algorithms, thereby providing the first evidence that these examiners are experts at this task. Notably, computationally fusing responses of multiple experts produced near-perfect performance. Results also revealed qualitative differences between expert and non-expert performance. First, examiners' superiority was greatest at longer exposure durations, suggestive of more entailed comparison in forensic examiners. Second, experts were less impaired by image inversion than non-expert students, contrasting with face memory studies that show larger face inversion effects in high performers. We conclude that expertise in matching identity across unfamiliar face images is supported by processes that differ qualitatively from those supporting memory for individual faces. PMID:26336174

  8. Perceptual expertise in forensic facial image comparison.

    PubMed

    White, David; Phillips, P Jonathon; Hahn, Carina A; Hill, Matthew; O'Toole, Alice J

    2015-09-07

    Forensic facial identification examiners are required to match the identity of faces in images that vary substantially, owing to changes in viewing conditions and in a person's appearance. These identifications affect the course and outcome of criminal investigations and convictions. Despite calls for research on sources of human error in forensic examination, existing scientific knowledge of face matching accuracy is based, almost exclusively, on people without formal training. Here, we administered three challenging face matching tests to a group of forensic examiners with many years' experience of comparing face images for law enforcement and government agencies. Examiners outperformed untrained participants and computer algorithms, thereby providing the first evidence that these examiners are experts at this task. Notably, computationally fusing responses of multiple experts produced near-perfect performance. Results also revealed qualitative differences between expert and non-expert performance. First, examiners' superiority was greatest at longer exposure durations, suggestive of more entailed comparison in forensic examiners. Second, experts were less impaired by image inversion than non-expert students, contrasting with face memory studies that show larger face inversion effects in high performers. We conclude that expertise in matching identity across unfamiliar face images is supported by processes that differ qualitatively from those supporting memory for individual faces. © 2015 The Author(s).

  9. Facial biometrics of peri-oral changes in Crohn's disease.

    PubMed

    Zou, L; Adegun, O K; Willis, A; Fortune, Farida

    2014-05-01

    Crohn's disease is a chronic relapsing and remitting inflammatory condition which affects any part of the gastrointestinal tract. In the oro-facial region, patients can present peri-oral swellings which results in severe facial disfigurement. To date, assessing the degree of facial changes and evaluation of treatment outcomes relies on clinical observation and semi-quantitative methods. In this paper, we describe the development of a robust and reproducible measurement strategy using 3-D facial biometrics to objectively quantify the extent and progression of oro-facial Crohn's disease. Using facial laser scanning, 32 serial images from 13 Crohn's patients attending the Oral Medicine clinic were acquired during relapse, remission, and post-treatment phases. Utilising theories of coordinate metrology, the facial images were subjected to registration, regions of interest identification, and reproducible repositioning prior to obtaining volume measurements. To quantify the changes in tissue volume, scan images from consecutive appointments were compared to the baseline (first scan image). Reproducibility test was performed to ascertain the degree of uncertainty in volume measurements. 3-D facial biometric imaging is a reliable method to identify and quantify peri-oral swelling in Crohn's patients. Comparison of facial scan images at different phases of the disease revealed precisely profile and volume changes. The volume measurements were highly reproducible as adjudged from the 1% standard deviation. 3-D facial biometrics measurements in Crohn's patients with oro-facial involvement offers a quick, robust, economical and objective approach for guided therapeutic intervention and routine assessment of treatment efficacy on the clinic.

  10. Validity and reliability of a structured-light 3D scanner and an ultrasound imaging system for measurements of facial skin thickness.

    PubMed

    Lee, Kang-Woo; Kim, Sang-Hwan; Gil, Young-Chun; Hu, Kyung-Seok; Kim, Hee-Jin

    2017-10-01

    Three-dimensional (3 D)-scanning-based morphological studies of the face are commonly included in various clinical procedures. This study evaluated validity and reliability of a 3 D scanning system by comparing the ultrasound (US) imaging system versus the direct measurement of facial skin. The facial skin thickness at 19 landmarks was measured using the three different methods in 10 embalmed adult Korean cadavers. Skin thickness was first measured using the ultrasound device, then 3 D scanning of the facial skin surface was performed. After the skin on the left half of face was gently dissected, deviating slightly right of the midline, to separate it from the subcutaneous layer, and the harvested facial skin's thickness was measured directly using neck calipers. The dissected specimen was then scanned again, then the scanned images of undissected and dissected faces were superimposed using Morpheus Plastic Solution (version 3.0) software. Finally, the facial skin thickness was calculated from the superimposed images. The ICC value for the correlations between the 3 D scanning system and direct measurement showed excellent reliability (0.849, 95% confidence interval = 0.799-0.887). Bland-Altman analysis showed a good level of agreement between the 3 D scanning system and direct measurement (bias = 0.49 ± 0.49 mm, mean±SD). These results demonstrate that the 3 D scanning system precisely reflects structural changes before and after skin dissection. Therefore, an in-depth morphological study using this 3 D scanning system could provide depth data about the main anatomical structures of face, thereby providing crucial anatomical knowledge for utilization in various clinical applications. Clin. Anat. 30:878-886, 2017. © 2017 Wiley Periodicals, Inc. © 2017 Wiley Periodicals, Inc.

  11. The Usefulness of MR Imaging of the Temporal Bone in the Evaluation of Patients with Facial and Audiovestibular Dysfunction

    PubMed Central

    Park, Sang Uk; Cho, Young Kuk; Lim, Myung Kwan; Kim, Won Hong; Suh, Chang Hae; Lee, Seung Chul

    2002-01-01

    Objective To evaluate the clinical utility of MR imaging of the temporal bone in patients with facial and audiovestibular dysfunction with particular emphasis on the importance of contrast enhancement. Materials and Methods We retrospectively reviewed the MR images of 179 patients [72 men, 107 women; average age, 44 (range, 1-77) years] who presented with peripheral facial palsy (n=15), audiometrically proven sensorineural hearing loss (n=104), vertigo (n=109), or tinnitus (n=92). Positive MR imaging findings possibly responsible for the patients clinical manifestations were categorized according to the anatomic sites and presumed etiologies of the lesions. We also assessed the utility of contrast-enhanced MR imaging by analyzing its contribution to the demonstration of lesions which would otherwise not have been apparent. All MR images were interpreted by two neuroradiologists, who reached their conclusions by consensus. Results MR images demonstrated positive findings, thought to account for the presenting symptoms, in 78 (44%) of 179 patients, including 15 (100%) of 15 with peripheral facial palsy, 43 (41%) of 104 with sensorineural hearing loss, 40 (37%) of 109 with vertigo, and 39 (42%) of 92 with tinnitus. Thirty (38%) of those 78 patients had lesions that could be confidently recognized only at contrast-enhanced MR imaging. Conclusion Even though its use led to positive findings in less than half of these patients, MR imaging of the temporal bone is a useful diagnostic procedure in the evaluation of those with facial and audiovestibular dysfunction. Because it was only at contrast-enhanced MR imaging that a significant number of patients showed positive imaging findings which explained their clinical manifestations, the use of contrast material is highly recommended. PMID:11919474

  12. Emotional facial activation induced by unconsciously perceived dynamic facial expressions.

    PubMed

    Kaiser, Jakob; Davey, Graham C L; Parkhouse, Thomas; Meeres, Jennifer; Scott, Ryan B

    2016-12-01

    Do facial expressions of emotion influence us when not consciously perceived? Methods to investigate this question have typically relied on brief presentation of static images. In contrast, real facial expressions are dynamic and unfold over several seconds. Recent studies demonstrate that gaze contingent crowding (GCC) can block awareness of dynamic expressions while still inducing behavioural priming effects. The current experiment tested for the first time whether dynamic facial expressions presented using this method can induce unconscious facial activation. Videos of dynamic happy and angry expressions were presented outside participants' conscious awareness while EMG measurements captured activation of the zygomaticus major (active when smiling) and the corrugator supercilii (active when frowning). Forced-choice classification of expressions confirmed they were not consciously perceived, while EMG revealed significant differential activation of facial muscles consistent with the expressions presented. This successful demonstration opens new avenues for research examining the unconscious emotional influences of facial expressions. Copyright © 2016 Elsevier B.V. All rights reserved.

  13. Estimation of human emotions using thermal facial information

    NASA Astrophysics Data System (ADS)

    Nguyen, Hung; Kotani, Kazunori; Chen, Fan; Le, Bac

    2014-01-01

    In recent years, research on human emotion estimation using thermal infrared (IR) imagery has appealed to many researchers due to its invariance to visible illumination changes. Although infrared imagery is superior to visible imagery in its invariance to illumination changes and appearance differences, it has difficulties in handling transparent glasses in the thermal infrared spectrum. As a result, when using infrared imagery for the analysis of human facial information, the regions of eyeglasses are dark and eyes' thermal information is not given. We propose a temperature space method to correct eyeglasses' effect using the thermal facial information in the neighboring facial regions, and then use Principal Component Analysis (PCA), Eigen-space Method based on class-features (EMC), and PCA-EMC method to classify human emotions from the corrected thermal images. We collected the Kotani Thermal Facial Emotion (KTFE) database and performed the experiments, which show the improved accuracy rate in estimating human emotions.

  14. Down syndrome detection from facial photographs using machine learning techniques

    NASA Astrophysics Data System (ADS)

    Zhao, Qian; Rosenbaum, Kenneth; Sze, Raymond; Zand, Dina; Summar, Marshall; Linguraru, Marius George

    2013-02-01

    Down syndrome is the most commonly occurring chromosomal condition; one in every 691 babies in United States is born with it. Patients with Down syndrome have an increased risk for heart defects, respiratory and hearing problems and the early detection of the syndrome is fundamental for managing the disease. Clinically, facial appearance is an important indicator in diagnosing Down syndrome and it paves the way for computer-aided diagnosis based on facial image analysis. In this study, we propose a novel method to detect Down syndrome using photography for computer-assisted image-based facial dysmorphology. Geometric features based on facial anatomical landmarks, local texture features based on the Contourlet transform and local binary pattern are investigated to represent facial characteristics. Then a support vector machine classifier is used to discriminate normal and abnormal cases; accuracy, precision and recall are used to evaluate the method. The comparison among the geometric, local texture and combined features was performed using the leave-one-out validation. Our method achieved 97.92% accuracy with high precision and recall for the combined features; the detection results were higher than using only geometric or texture features. The promising results indicate that our method has the potential for automated assessment for Down syndrome from simple, noninvasive imaging data.

  15. The perception of children's computer-imaged facial profiles by patients, mothers and clinicians.

    PubMed

    Miner, Robert M; Anderson, Nina K; Evans, Carla A; Giddon, Donald B

    2007-11-01

    To demonstrate the usefulness of a new imaging system for comparing the morphometric bases of children's self-perception of their facial profile with the perceptions of their mothers and treating clinicians. Rather than choosing among a series of static images, a computer imaging program was developed to elicit a range of acceptable responses or tolerance for change from which a midpoint of acceptability was derived. Using the method of Giddon et al, three profile features (upper and lower lips and mandible) from standardized images of 24 patients aged 8- 15 years were distorted and presented to patients, parents, and clinicians in random order as slowly moving images (four frames per second) from retrusive and protrusive extremes. Subjects clicked the mouse when the image became acceptable and released it when it was no longer acceptable. Subjects responded similarly to a neutral facial profile. Patients and their mothers overestimated the protrusiveness of the mandible of the actual pretreatment profile. Consistent with related studies, mothers had a smaller tolerance for change in the soft tissue profile than the children or clinicians. The magnitudes of the children's self-preference and preferred change in a neutral face were also significantly correlated. Both patients and mothers preferred a more protrusive profile than that of the actual or neutral face for the patient and neutral face. Imaging software can be used with children to compare their preferences with those of parents and clinicians to facilitate treatment planning and patient satisfaction.

  16. Comparison of facial morphologies between adult Chinese and Houstonian Caucasian populations using three-dimensional imaging.

    PubMed

    Wirthlin, J; Kau, C H; English, J D; Pan, F; Zhou, H

    2013-09-01

    The objective of this study was to compare the facial morphologies of an adult Chinese population to a Houstonian white population. Three-dimensional (3D) images were acquired via a commercially available stereophotogrammetric camera system, 3dMDface™. Using the system, 100 subjects from a Houstonian population and 71 subjects from a Chinese population were photographed. A complex mathematical algorithm was performed to generate a composite facial average (one for males and one for females) for each subgroup. The computer-generated facial averages were then superimposed based on a previously validated superimposition method. The facial averages were evaluated for differences. Distinct facial differences were evident between the subgroups evaluated. These areas included the nasal tip, the peri-orbital area, the malar process, the labial region, the forehead, and the chin. Overall, the mean facial difference between the Chinese and Houstonian female averages was 2.73±2.20mm, while the difference between the Chinese and Houstonian males was 2.83±2.20mm. The percent similarity for the female population pairings and male population pairings were 10.45% and 12.13%, respectively. The average adult Chinese and Houstonian faces possess distinct differences. Different populations and ethnicities have different facial features and averages that should be considered in the planning of treatment. Copyright © 2013 International Association of Oral and Maxillofacial Surgeons. Published by Elsevier Ltd. All rights reserved.

  17. Facial dynamics and emotional expressions in facial aging treatments.

    PubMed

    Michaud, Thierry; Gassia, Véronique; Belhaouari, Lakhdar

    2015-03-01

    Facial expressions convey emotions that form the foundation of interpersonal relationships, and many of these emotions promote and regulate our social linkages. Hence, the facial aging symptomatological analysis and the treatment plan must of necessity include knowledge of the facial dynamics and the emotional expressions of the face. This approach aims to more closely meet patients' expectations of natural-looking results, by correcting age-related negative expressions while observing the emotional language of the face. This article will successively describe patients' expectations, the role of facial expressions in relational dynamics, the relationship between facial structures and facial expressions, and the way facial aging mimics negative expressions. Eventually, therapeutic implications for facial aging treatment will be addressed. © 2015 Wiley Periodicals, Inc.

  18. Analysis of differences between Western and East-Asian faces based on facial region segmentation and PCA for facial expression recognition

    NASA Astrophysics Data System (ADS)

    Benitez-Garcia, Gibran; Nakamura, Tomoaki; Kaneko, Masahide

    2017-01-01

    Darwin was the first one to assert that facial expressions are innate and universal, which are recognized across all cultures. However, recent some cross-cultural studies have questioned this assumed universality. Therefore, this paper presents an analysis of the differences between Western and East-Asian faces of the six basic expressions (anger, disgust, fear, happiness, sadness and surprise) focused on three individual facial regions of eyes-eyebrows, nose and mouth. The analysis is conducted by applying PCA for two feature extraction methods: appearance-based by using the pixel intensities of facial parts, and geometric-based by handling 125 feature points from the face. Both methods are evaluated using 4 standard databases for both racial groups and the results are compared with a cross-cultural human study applied to 20 participants. Our analysis reveals that differences between Westerns and East-Asians exist mainly on the regions of eyes-eyebrows and mouth for expressions of fear and disgust respectively. This work presents important findings for a better design of automatic facial expression recognition systems based on the difference between two racial groups.

  19. ALE meta-analysis on facial judgments of trustworthiness and attractiveness.

    PubMed

    Bzdok, D; Langner, R; Caspers, S; Kurth, F; Habel, U; Zilles, K; Laird, A; Eickhoff, Simon B

    2011-01-01

    Faces convey a multitude of information in social interaction, among which are trustworthiness and attractiveness. Humans process and evaluate these two dimensions very quickly due to their great adaptive importance. Trustworthiness evaluation is crucial for modulating behavior toward strangers; attractiveness evaluation is a crucial factor for mate selection, possibly providing cues for reproductive success. As both dimensions rapidly guide social behavior, this study tests the hypothesis that both judgments may be subserved by overlapping brain networks. To this end, we conducted an activation likelihood estimation meta-analysis on 16 functional magnetic resonance imaging studies pertaining to facial judgments of trustworthiness and attractiveness. Throughout combined, individual, and conjunction analyses on those two facial judgments, we observed consistent maxima in the amygdala which corroborates our initial hypothesis. This finding supports the contemporary paradigm shift extending the amygdala's role from dominantly processing negative emotional stimuli to processing socially relevant ones. We speculate that the amygdala filters sensory information with evolutionarily conserved relevance. Our data suggest that such a role includes not only "fight-or-flight" decisions but also social behaviors with longer term pay-off schedules, e.g., trustworthiness and attractiveness evaluation. © Springer-Verlag 2010

  20. Laptop Computer - Based Facial Recognition System Assessment

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    R. A. Cain; G. B. Singleton

    2001-03-01

    The objective of this project was to assess the performance of the leading commercial-off-the-shelf (COTS) facial recognition software package when used as a laptop application. We performed the assessment to determine the system's usefulness for enrolling facial images in a database from remote locations and conducting real-time searches against a database of previously enrolled images. The assessment involved creating a database of 40 images and conducting 2 series of tests to determine the product's ability to recognize and match subject faces under varying conditions. This report describes the test results and includes a description of the factors affecting the results.more » After an extensive market survey, we selected Visionics' FaceIt{reg_sign} software package for evaluation and a review of the Facial Recognition Vendor Test 2000 (FRVT 2000). This test was co-sponsored by the US Department of Defense (DOD) Counterdrug Technology Development Program Office, the National Institute of Justice, and the Defense Advanced Research Projects Agency (DARPA). Administered in May-June 2000, the FRVT 2000 assessed the capabilities of facial recognition systems that were currently available for purchase on the US market. Our selection of this Visionics product does not indicate that it is the ''best'' facial recognition software package for all uses. It was the most appropriate package based on the specific applications and requirements for this specific application. In this assessment, the system configuration was evaluated for effectiveness in identifying individuals by searching for facial images captured from video displays against those stored in a facial image database. An additional criterion was that the system be capable of operating discretely. For this application, an operational facial recognition system would consist of one central computer hosting the master image database with multiple standalone systems configured with duplicates of the master

  1. An analysis of maxillary anterior teeth: facial and dental proportions.

    PubMed

    Hasanreisoglu, Ufuk; Berksun, Semih; Aras, Kerem; Arslan, Ilker

    2005-12-01

    The size and form of the maxillary anterior teeth are important in achieving pleasing dental and facial esthetics. However, little scientific data have been defined as criteria for evaluating these morphological features. This study analyzed the clinical crown dimensions of maxillary anterior teeth to determine whether consistent relationships exist between tooth width and several facial measurements in a subset of the Turkish population. Full-face and anterior tooth images of 100 Turkish dental students viewed from the front and engaged in maximum smiling were recorded with digital photography under standardized conditions. Gypsum casts of the maxillary arches of the subjects were also made. The dimensions of the anterior teeth, the occurrence of the golden ratio, the difference between the actual and perceived sizes, and the relationship between the anterior teeth and several facial measurements by gender were analyzed using the information obtained from both the computer images and the casts. One-sample, 2-sample, and paired t tests, and repeated-measures analysis of variance and Duncan multiple-range tests were performed to analyze the data (alpha=.05). The dimensions of the central incisors (P<.05) and canines (P<.01) varied by gender. The existence of the so-called "golden proportion" for the maxillary anterior teeth as a whole was not found. Significant differences emerged when the mean ratios between various perceived widths were compared with their ideal golden ratios (P<.01). Proportional relationships between the bizygomatic width and the width of the central incisor, and the intercanine distance and the interalar width in women were observed. The maxillary central incisor and canine dimensions of men were greater than those of women in the Turkish population studied, with the canines showing the greatest gender variation. Neither a golden proportion nor any other recurrent proportion for all anterior teeth was determined. Bizygomatic width and interalar

  2. Three-dimensional facial recognition using passive long-wavelength infrared polarimetric imaging.

    PubMed

    Yuffa, Alex J; Gurton, Kristan P; Videen, Gorden

    2014-12-20

    We use a polarimetric camera to record the Stokes parameters and the degree of linear polarization of long-wavelength infrared radiation emitted by human faces. These Stokes images are combined with Fresnel relations to extract the surface normal at each pixel. Integrating over these surface normals yields a three-dimensional facial image. One major difficulty of this technique is that the normal vectors determined from the polarizations are not unique. We overcome this problem by introducing an additional boundary condition on the subject. The major sources of error in producing inversions are noise in the images caused by scattering of the background signal and the ambiguity in determining the surface normals from the Fresnel coefficients.

  3. Three-dimensional Imaging Methods for Quantitative Analysis of Facial Soft Tissues and Skeletal Morphology in Patients with Orofacial Clefts: A Systematic Review

    PubMed Central

    Kuijpers, Mette A. R.; Chiu, Yu-Ting; Nada, Rania M.; Carels, Carine E. L.; Fudalej, Piotr S.

    2014-01-01

    Background Current guidelines for evaluating cleft palate treatments are mostly based on two-dimensional (2D) evaluation, but three-dimensional (3D) imaging methods to assess treatment outcome are steadily rising. Objective To identify 3D imaging methods for quantitative assessment of soft tissue and skeletal morphology in patients with cleft lip and palate. Data sources Literature was searched using PubMed (1948–2012), EMBASE (1980–2012), Scopus (2004–2012), Web of Science (1945–2012), and the Cochrane Library. The last search was performed September 30, 2012. Reference lists were hand searched for potentially eligible studies. There was no language restriction. Study selection We included publications using 3D imaging techniques to assess facial soft tissue or skeletal morphology in patients older than 5 years with a cleft lip with/or without cleft palate. We reviewed studies involving the facial region when at least 10 subjects in the sample size had at least one cleft type. Only primary publications were included. Data extraction Independent extraction of data and quality assessments were performed by two observers. Results Five hundred full text publications were retrieved, 144 met the inclusion criteria, with 63 high quality studies. There were differences in study designs, topics studied, patient characteristics, and success measurements; therefore, only a systematic review could be conducted. Main 3D-techniques that are used in cleft lip and palate patients are CT, CBCT, MRI, stereophotogrammetry, and laser surface scanning. These techniques are mainly used for soft tissue analysis, evaluation of bone grafting, and changes in the craniofacial skeleton. Digital dental casts are used to evaluate treatment and changes over time. Conclusion Available evidence implies that 3D imaging methods can be used for documentation of CLP patients. No data are available yet showing that 3D methods are more informative than conventional 2D methods. Further research

  4. Augmentation of linear facial anthropometrics through modern morphometrics: a facial convexity example.

    PubMed

    Wei, R; Claes, P; Walters, M; Wholley, C; Clement, J G

    2011-06-01

    The facial region has traditionally been quantified using linear anthropometrics. These are well established in dentistry, but require expertise to be used effectively. The aim of this study was to augment the utility of linear anthropometrics by applying them in conjunction with modern 3-D morphometrics. Facial images of 75 males and 94 females aged 18-25 years with self-reported Caucasian ancestry were used. An anthropometric mask was applied to establish corresponding quasi-landmarks on the images in the dataset. A statistical face-space, encoding shape covariation, was established. The facial median plane was extracted facilitating both manual and automated indication of commonly used midline landmarks. From both indications, facial convexity angles were calculated and compared. The angles were related to the face-space using a regression based pathway enabling the visualization of facial form associated with convexity variation. Good agreement between the manual and automated angles was found (Pearson correlation: 0.9478-0.9474, Dahlberg root mean squared error: 1.15°-1.24°). The population mean angle was 166.59°-166.29° (SD 5.09°-5.2°) for males-females. The angle-pathway provided valuable feedback. Linear facial anthropometrics can be extended when used in combination with a face-space derived from 3-D scans and the exploration of property pathways inferred in a statistically verifiable way. © 2011 Australian Dental Association.

  5. Hybrid Feature Extraction-based Approach for Facial Parts Representation and Recognition

    NASA Astrophysics Data System (ADS)

    Rouabhia, C.; Tebbikh, H.

    2008-06-01

    Face recognition is a specialized image processing which has attracted a considerable attention in computer vision. In this article, we develop a new facial recognition system from video sequences images dedicated to person identification whose face is partly occulted. This system is based on a hybrid image feature extraction technique called ACPDL2D (Rouabhia et al. 2007), it combines two-dimensional principal component analysis and two-dimensional linear discriminant analysis with neural network. We performed the feature extraction task on the eyes and the nose images separately then a Multi-Layers Perceptron classifier is used. Compared to the whole face, the results of simulation are in favor of the facial parts in terms of memory capacity and recognition (99.41% for the eyes part, 98.16% for the nose part and 97.25 % for the whole face).

  6. Sound-induced facial synkinesis following facial nerve paralysis.

    PubMed

    Ma, Ming-San; van der Hoeven, Johannes H; Nicolai, Jean-Philippe A; Meek, Marcel F

    2009-08-01

    Facial synkinesis (or synkinesia) (FS) occurs frequently after paresis or paralysis of the facial nerve and is in most cases due to aberrant regeneration of (branches of) the facial nerve. Patients suffer from inappropriate and involuntary synchronous facial muscle contractions. Here we describe two cases of sound-induced facial synkinesis (SFS) after facial nerve injury. As far as we know, this phenomenon has not been described in the English literature before. Patient A presented with right hemifacial palsy after lesion of the facial nerve due to skull base fracture. He reported involuntary muscle activity at the right corner of the mouth, specifically on hearing ringing keys. Patient B suffered from left hemifacial palsy following otitis media and developed involuntary muscle contraction in the facial musculature specifically on hearing clapping hands or a trumpet sound. Both patients were evaluated by means of video, audio and EMG analysis. Possible mechanisms in the pathophysiology of SFS are postulated and therapeutic options are discussed.

  7. A new quantitative evaluation method for age-related changes of individual pigmented spots in facial skin.

    PubMed

    Kikuchi, K; Masuda, Y; Yamashita, T; Sato, K; Katagiri, C; Hirao, T; Mizokami, Y; Yaguchi, H

    2016-08-01

    Facial skin pigmentation is one of the most prominent visible features of skin aging and often affects perception of health and beauty. To date, facial pigmentation has been evaluated using various image analysis methods developed for the cosmetic and esthetic fields. However, existing methods cannot provide precise information on pigmented spots, such as variations in size, color shade, and distribution pattern. The purpose of this study is the development of image evaluation methods to analyze individual pigmented spots and acquire detailed information on their age-related changes. To characterize the individual pigmented spots within a cheek image, we established a simple object-counting algorithm. First, we captured cheek images using an original imaging system equipped with an illumination unit and a high-resolution digital camera. The acquired images were converted into melanin concentration images using compensation formulae. Next, the melanin images were converted into binary images. The binary images were then subjected to noise reduction. Finally, we calculated parameters such as the melanin concentration, quantity, and size of individual pigmented spots using a connected-components labeling algorithm, which assigns a unique label to each separate group of connected pixels. The cheek image analysis was evaluated on 643 female Japanese subjects. We confirmed that the proposed method was sufficiently sensitive to measure the melanin concentration, and the numbers and sizes of individual pigmented spots through manual evaluation of the cheek images. The image analysis results for the 643 Japanese women indicated clear relationships between age and the changes in the pigmented spots. We developed a new quantitative evaluation method for individual pigmented spots in facial skin. This method facilitates the analysis of the characteristics of various pigmented facial spots and is directly applicable to the fields of dermatology, pharmacology, and esthetic

  8. Comparing Facial 3D Analysis With DNA Testing to Determine Zygosities of Twins.

    PubMed

    Vuollo, Ville; Sidlauskas, Mantas; Sidlauskas, Antanas; Harila, Virpi; Salomskiene, Loreta; Zhurov, Alexei; Holmström, Lasse; Pirttiniemi, Pertti; Heikkinen, Tuomo

    2015-06-01

    The aim of this study was to compare facial 3D analysis to DNA testing in twin zygosity determinations. Facial 3D images of 106 pairs of young adult Lithuanian twins were taken with a stereophotogrammetric device (3dMD, Atlanta, Georgia) and zygosity was determined according to similarity of facial form. Statistical pattern recognition methodology was used for classification. The results showed that in 75% to 90% of the cases, zygosity determinations were similar to DNA-based results. There were 81 different classification scenarios, including 3 groups, 3 features, 3 different scaling methods, and 3 threshold levels. It appeared that coincidence with 0.5 mm tolerance is the most suitable feature for classification. Also, leaving out scaling improves results in most cases. Scaling was expected to equalize the magnitude of differences and therefore lead to better recognition performance. Still, better classification features and a more effective scaling method or classification in different facial areas could further improve the results. In most of the cases, male pair zygosity recognition was at a higher level compared with females. Erroneously classified twin pairs appear to be obvious outliers in the sample. In particular, faces of young dizygotic (DZ) twins may be so similar that it is very hard to define a feature that would help classify the pair as DZ. Correspondingly, monozygotic (MZ) twins may have faces with quite different shapes. Such anomalous twin pairs are interesting exceptions, but they form a considerable portion in both zygosity groups.

  9. Measuring Facial Movement

    ERIC Educational Resources Information Center

    Ekman, Paul; Friesen, Wallace V.

    1976-01-01

    The Facial Action Code (FAC) was derived from an analysis of the anatomical basis of facial movement. The development of the method is explained, contrasting it to other methods of measuring facial behavior. An example of how facial behavior is measured is provided, and ideas about research applications are discussed. (Author)

  10. Automatically Log Off Upon Disappearance of Facial Image

    DTIC Science & Technology

    2005-03-01

    log off a PC when the user’s face disappears for an adjustable time interval. Among the fundamental technologies of biometrics, facial recognition is... facial recognition products. In this report, a brief overview of face detection technologies is provided. The particular neural network-based face...ensure that the user logging onto the system is the same person. Among the fundamental technologies of biometrics, facial recognition is the only

  11. Assessment of facial golden proportions among young Japanese women.

    PubMed

    Mizumoto, Yasushi; Deguchi, Toshio; Fong, Kelvin W C

    2009-08-01

    Facial proportions are of interest in orthodontics. The null hypothesis is that there is no difference in golden proportions of the soft-tissue facial balance between Japanese and white women. Facial proportions were assessed by examining photographs of 3 groups of Asian women: group 1, 30 young adult patients with a skeletal Class 1 occlusion; group 2, 30 models; and group 3, 14 popular actresses. Photographic prints or slides were digitized for image analysis. Group 1 subjects had standardized photos taken as part of their treatment. Photos of the subjects in groups 2 and 3 were collected from magazines and other sources and were of varying sizes; therefore, the output image size was not considered. The range of measurement errors was 0.17% to 1.16%. ANOVA was selected because the data set was normally distributed with homogeneous variances. The subjects in the 3 groups showed good total facial proportions. The proportions of the face-height components in group 1 were similar to the golden proportion, which indicated a longer, lower facial height and shorter nose. Group 2 differed from the golden proportion, with a short, lower facial height. Group 3 had golden proportions in all 7 measurements. The proportion of the face width deviated from the golden proportion, indicating a small mouth or wide-set eyes in groups 1 and 2. The null hypothesis was verified in the group 3 actresses in the facial height components. Some measurements in groups 1 and 2 showed different facial proportions that deviated from the golden proportion (ratio).

  12. Facial trauma.

    PubMed

    Peeters, N; Lemkens, P; Leach, R; Gemels B; Schepers, S; Lemmens, W

    Facial trauma. Patients with facial trauma must be assessed in a systematic way so as to avoid missing any injury. Severe and disfiguring facial injuries can be distracting. However, clinicians must first focus on the basics of trauma care, following the Advanced Trauma Life Support (ATLS) system of care. Maxillofacial trauma occurs in a significant number of severely injured patients. Life- and sight-threatening injuries must be excluded during the primary and secondary surveys. Special attention must be paid to sight-threatening injuries in stabilized patients through early referral to an appropriate specialist or the early initiation of emergency care treatment. The gold standard for the radiographic evaluation of facial injuries is computed tomography (CT) imaging. Nasal fractures are the most frequent isolated facial fractures. Isolated nasal fractures are principally diagnosed through history and clinical examination. Closed reduction is the most frequently performed treatment for isolated nasal fractures, with a fractured nasal septum as a predictor of failure. Ear, nose and throat surgeons, maxillofacial surgeons and ophthalmologists must all develop an adequate treatment plan for patients with complex maxillofacial trauma.

  13. Anthropometric Study of Three-Dimensional Facial Morphology in Malay Adults

    PubMed Central

    Majawit, Lynnora Patrick; Mohd Razi, Roziana

    2016-01-01

    Objectives To establish the three-dimensional (3D) facial soft tissue morphology of adult Malaysian subjects of the Malay ethnic group; and to determine the morphological differences between the genders, using a non-invasive stereo-photogrammetry 3D camera. Material and Methods One hundred and nine subjects participated in this research, 54 Malay men and 55 Malay women, aged 20–30 years old with healthy BMI and with no adverse skeletal deviation. Twenty-three facial landmarks were identified on 3D facial images captured using a VECTRA M5-360 Head System (Canfield Scientific Inc, USA). Two angular, 3 ratio and 17 linear measurements were identified using Canfield Mirror imaging software. Intra- and inter-examiner reliability tests were carried out using 10 randomly selected images, analyzed using the intra-class correlation coefficient (ICC). Multivariate analysis of variance (MANOVA) was carried out to investigate morphologic differences between genders. Results ICC scores were generally good for both intra-examiner (range 0.827–0.987) and inter-examiner reliability (range 0.700–0.983) tests. Generally, all facial measurements were larger in men than women, except the facial profile angle which was larger in women. Clinically significant gender dimorphisms existed in biocular width, nose height, nasal bridge length, face height and lower face height values (mean difference > 3mm). Clinical significance was set at 3mm. Conclusion Facial soft tissue morphological values can be gathered efficiently and measured effectively from images captured by a non-invasive stereo-photogrammetry 3D camera. Adult men in Malaysia when compared to women had a wider distance between the eyes, a longer and more prominent nose and a longer face. PMID:27706220

  14. Are Portable Stereophotogrammetric Devices Reliable in Facial Imaging? A Validation Study of VECTRA H1 Device.

    PubMed

    Gibelli, Daniele; Pucciarelli, Valentina; Cappella, Annalisa; Dolci, Claudia; Sforza, Chiarella

    2018-01-31

    Modern 3-dimensional (3D) image acquisition systems represent a crucial technologic development in facial anatomy because of their accuracy and precision. The recently introduced portable devices can improve facial databases by increasing the number of applications. In the present study, the VECTRA H1 portable stereophotogrammetric device was validated to verify its applicability to 3D facial analysis. Fifty volunteers underwent 4 facial scans using portable VECTRA H1 and static VECTRA M3 devices (2 for each instrument). Repeatability of linear, angular, surface area, and volume measurements was verified within the device and between devices using the Bland-Altman test and the calculation of absolute and relative technical errors of measurement (TEM and rTEM, respectively). In addition, the 2 scans obtained by the same device and the 2 scans obtained by different devices were registered and superimposed to calculate the root mean square (RMS; point-to-point) distance between the 2 surfaces. Most linear, angular, and surface area measurements had high repeatability in M3 versus M3, H1 versus H1, and M3 versus H1 comparisons (range, 82.2 to 98.7%; TEM range, 0.3 to 2.0 mm, 0.4° to 1.8°; rTEM range, 0.2 to 3.1%). In contrast, volumes and RMS distances showed evident differences in M3 versus M3 and H1 versus H1 comparisons and reached the maximum when scans from the 2 different devices were compared. The portable VECTRA H1 device proved reliable for assessing linear measurements, angles, and surface areas; conversely, the influence of involuntary facial movements on volumes and RMS distances was more important compared with the static device. Copyright © 2018 American Association of Oral and Maxillofacial Surgeons. Published by Elsevier Inc. All rights reserved.

  15. Impact of facial defect reconstruction on attractiveness and negative facial perception.

    PubMed

    Dey, Jacob K; Ishii, Masaru; Boahene, Kofi D O; Byrne, Patrick; Ishii, Lisa E

    2015-06-01

    Measure the impact of facial defect reconstruction on observer-graded attractiveness and negative facial perception. Prospective, randomized, controlled experiment. One hundred twenty casual observers viewed images of faces with defects of varying sizes and locations before and after reconstruction as well as normal comparison faces. Observers rated attractiveness, defect severity, and how disfiguring, bothersome, and important to repair they considered each face. Facial defects decreased attractiveness -2.26 (95% confidence interval [CI]: -2.45, -2.08) on a 10-point scale. Mixed effects linear regression showed this attractiveness penalty varied with defect size and location, with large and central defects generating the greatest penalty. Reconstructive surgery increased attractiveness 1.33 (95% CI: 1.18, 1.47), an improvement dependent upon size and location, restoring some defect categories to near normal ranges of attractiveness. Iterated principal factor analysis indicated the disfiguring, important to repair, bothersome, and severity variables were highly correlated and measured a common domain; thus, they were combined to create the disfigured, important to repair, bothersome, severity (DIBS) factor score, representing negative facial perception. The DIBS regression showed defect faces have a 1.5 standard deviation increase in negative perception (DIBS: 1.69, 95% CI: 1.61, 1.77) compared to normal faces, which decreased by a similar magnitude after surgery (DIBS: -1.44, 95% CI: -1.49, -1.38). These findings varied with defect size and location. Surgical reconstruction of facial defects increased attractiveness and decreased negative social facial perception, an impact that varied with defect size and location. These new social perception data add to the evidence base demonstrating the value of high-quality reconstructive surgery. NA. © 2015 The American Laryngological, Rhinological and Otological Society, Inc.

  16. Facial motion parameter estimation and error criteria in model-based image coding

    NASA Astrophysics Data System (ADS)

    Liu, Yunhai; Yu, Lu; Yao, Qingdong

    2000-04-01

    Model-based image coding has been given extensive attention due to its high subject image quality and low bit-rates. But the estimation of object motion parameter is still a difficult problem, and there is not a proper error criteria for the quality assessment that are consistent with visual properties. This paper presents an algorithm of the facial motion parameter estimation based on feature point correspondence and gives the motion parameter error criteria. The facial motion model comprises of three parts. The first part is the global 3-D rigid motion of the head, the second part is non-rigid translation motion in jaw area, and the third part consists of local non-rigid expression motion in eyes and mouth areas. The feature points are automatically selected by a function of edges, brightness and end-node outside the blocks of eyes and mouth. The numbers of feature point are adjusted adaptively. The jaw translation motion is tracked by the changes of the feature point position of jaw. The areas of non-rigid expression motion can be rebuilt by using block-pasting method. The estimation approach of motion parameter error based on the quality of reconstructed image is suggested, and area error function and the error function of contour transition-turn rate are used to be quality criteria. The criteria reflect the image geometric distortion caused by the error of estimated motion parameters properly.

  17. Local binary pattern variants-based adaptive texture features analysis for posed and nonposed facial expression recognition

    NASA Astrophysics Data System (ADS)

    Sultana, Maryam; Bhatti, Naeem; Javed, Sajid; Jung, Soon Ki

    2017-09-01

    Facial expression recognition (FER) is an important task for various computer vision applications. The task becomes challenging when it requires the detection and encoding of macro- and micropatterns of facial expressions. We present a two-stage texture feature extraction framework based on the local binary pattern (LBP) variants and evaluate its significance in recognizing posed and nonposed facial expressions. We focus on the parametric limitations of the LBP variants and investigate their effects for optimal FER. The size of the local neighborhood is an important parameter of the LBP technique for its extraction in images. To make the LBP adaptive, we exploit the granulometric information of the facial images to find the local neighborhood size for the extraction of center-symmetric LBP (CS-LBP) features. Our two-stage texture representations consist of an LBP variant and the adaptive CS-LBP features. Among the presented two-stage texture feature extractions, the binarized statistical image features and adaptive CS-LBP features were found showing high FER rates. Evaluation of the adaptive texture features shows competitive and higher performance than the nonadaptive features and other state-of-the-art approaches, respectively.

  18. Forming Facial Expressions Influences Assessment of Others' Dominance but Not Trustworthiness.

    PubMed

    Ueda, Yoshiyuki; Nagoya, Kie; Yoshikawa, Sakiko; Nomura, Michio

    2017-01-01

    Forming specific facial expressions influences emotions and perception. Bearing this in mind, studies should be reconsidered in which observers expressing neutral emotions inferred personal traits from the facial expressions of others. In the present study, participants were asked to make happy, neutral, and disgusted facial expressions: for "happy," they held a wooden chopstick in their molars to form a smile; for "neutral," they clasped the chopstick between their lips, making no expression; for "disgusted," they put the chopstick between their upper lip and nose and knit their brows in a scowl. However, they were not asked to intentionally change their emotional state. Observers judged happy expression images as more trustworthy, competent, warm, friendly, and distinctive than disgusted expression images, regardless of the observers' own facial expression. Observers judged disgusted expression images as more dominant than happy expression images. However, observers expressing disgust overestimated dominance in observed disgusted expression images and underestimated dominance in happy expression images. In contrast, observers with happy facial forms attenuated dominance for disgusted expression images. These results suggest that dominance inferred from facial expressions is unstable and influenced by not only the observed facial expression, but also the observers' own physiological states.

  19. Social Risk and Depression: Evidence from Manual and Automatic Facial Expression Analysis

    PubMed Central

    Girard, Jeffrey M.; Cohn, Jeffrey F.; Mahoor, Mohammad H.; Mavadati, Seyedmohammad; Rosenwald, Dean P.

    2014-01-01

    Investigated the relationship between change over time in severity of depression symptoms and facial expression. Depressed participants were followed over the course of treatment and video recorded during a series of clinical interviews. Facial expressions were analyzed from the video using both manual and automatic systems. Automatic and manual coding were highly consistent for FACS action units, and showed similar effects for change over time in depression severity. For both systems, when symptom severity was high, participants made more facial expressions associated with contempt, smiled less, and those smiles that occurred were more likely to be accompanied by facial actions associated with contempt. These results are consistent with the “social risk hypothesis” of depression. According to this hypothesis, when symptoms are severe, depressed participants withdraw from other people in order to protect themselves from anticipated rejection, scorn, and social exclusion. As their symptoms fade, participants send more signals indicating a willingness to affiliate. The finding that automatic facial expression analysis was both consistent with manual coding and produced the same pattern of depression effects suggests that automatic facial expression analysis may be ready for use in behavioral and clinical science. PMID:24598859

  20. Functional connectivity between amygdala and facial regions involved in recognition of facial threat

    PubMed Central

    Harada, Tokiko; Ruffman, Ted; Sadato, Norihiro; Iidaka, Tetsuya

    2013-01-01

    The recognition of threatening faces is important for making social judgments. For example, threatening facial features of defendants could affect the decisions of jurors during a trial. Previous neuroimaging studies using faces of members of the general public have identified a pivotal role of the amygdala in perceiving threat. This functional magnetic resonance imaging study used face photographs of male prisoners who had been convicted of first-degree murder (MUR) as threatening facial stimuli. We compared the subjective ratings of MUR faces with those of control (CON) faces and examined how they were related to brain activation, particularly, the modulation of the functional connectivity between the amygdala and other brain regions. The MUR faces were perceived to be more threatening than the CON faces. The bilateral amygdala was shown to respond to both MUR and CON faces, but subtraction analysis revealed no significant difference between the two. Functional connectivity analysis indicated that the extent of connectivity between the left amygdala and the face-related regions (i.e. the superior temporal sulcus, inferior temporal gyrus and fusiform gyrus) was correlated with the subjective threat rating for the faces. We have demonstrated that the functional connectivity is modulated by vigilance for threatening facial features. PMID:22156740

  1. Facial morphometry of Ecuadorian patients with growth hormone receptor deficiency/Laron syndrome.

    PubMed Central

    Schaefer, G B; Rosenbloom, A L; Guevara-Aguirre, J; Campbell, E A; Ullrich, F; Patil, K; Frias, J L

    1994-01-01

    Facial morphometry using computerised image analysis was performed on patients with growth hormone receptor deficiency (Laron syndrome) from an inbred population of southern Ecuador. Morphometrics were compared for 49 patients, 70 unaffected relatives, and 14 unrelated persons. Patients with growth hormone receptor deficiency showed significant decreases in measures of vertical facial growth as compared to unaffected relatives and unrelated persons with short stature from other causes. This report validates and quantifies the clinical impression of foreshortened facies in growth hormone receptor deficiency. Images PMID:7815422

  2. Facial recognition software success rates for the identification of 3D surface reconstructed facial images: implications for patient privacy and security.

    PubMed

    Mazura, Jan C; Juluru, Krishna; Chen, Joseph J; Morgan, Tara A; John, Majnu; Siegel, Eliot L

    2012-06-01

    Image de-identification has focused on the removal of textual protected health information (PHI). Surface reconstructions of the face have the potential to reveal a subject's identity even when textual PHI is absent. This study assessed the ability of a computer application to match research subjects' 3D facial reconstructions with conventional photographs of their face. In a prospective study, 29 subjects underwent CT scans of the head and had frontal digital photographs of their face taken. Facial reconstructions of each CT dataset were generated on a 3D workstation. In phase 1, photographs of the 29 subjects undergoing CT scans were added to a digital directory and tested for recognition using facial recognition software. In phases 2-4, additional photographs were added in groups of 50 to increase the pool of possible matches and the test for recognition was repeated. As an internal control, photographs of all subjects were tested for recognition against an identical photograph. Of 3D reconstructions, 27.5% were matched correctly to corresponding photographs (95% upper CL, 40.1%). All study subject photographs were matched correctly to identical photographs (95% lower CL, 88.6%). Of 3D reconstructions, 96.6% were recognized simply as a face by the software (95% lower CL, 83.5%). Facial recognition software has the potential to recognize features on 3D CT surface reconstructions and match these with photographs, with implications for PHI.

  3. Diagnosis and surgical outcomes of intraparotid facial nerve schwannoma showing normal facial nerve function.

    PubMed

    Lee, D W; Byeon, H K; Chung, H P; Choi, E C; Kim, S-H; Park, Y M

    2013-07-01

    The findings of intraparotid facial nerve schwannoma (FNS) using preoperative diagnostic tools, including ultrasonography (US)-guided fine needle aspiration biopsy, computed tomography (CT) scan, and magnetic resonance imaging (MRI), were analyzed to determine if there are any useful findings that might suggest the presence of a lesion. Treatment guidelines are suggested. The medical records of 15 patients who were diagnosed with an intraparotid FNS were retrospectively analyzed. US and CT scans provide clinicians with only limited information; gadolinium enhanced T1-weighted images from MRI provide more specific findings. Tumors could be removed successfully with surgical exploration, preserving facial nerve function at the same time. Gadolinium-enhanced T1-weighted MRI showed more characteristic findings for the diagnosis of intraparotid FNS. Intraparotid FNS without facial palsy can be diagnosed with MRI preoperatively, and surgical exploration is a suitable treatment modality which can remove the tumor and preserve facial nerve function. Crown Copyright © 2013. Published by Elsevier Ltd. All rights reserved.

  4. Facial Attractiveness Assessment using Illustrated Questionnairers

    PubMed Central

    MESAROS, ANCA; CORNEA, DANIELA; CIOARA, LIVIU; DUDEA, DIANA; MESAROS, MICHAELA; BADEA, MINDRA

    2015-01-01

    Introduction. An attractive facial appearance is considered nowadays to be a decisive factor in establishing successful interactions between humans. In relation to this topic, scientific literature states that some of the facial features have more impact then others, and important authors revealed that certain proportions between different anthropometrical landmarks are mandatory for an attractive facial appearance. Aim. Our study aims to assess if certain facial features count differently in people’s opinion while assessing facial attractiveness in correlation with factors such as age, gender, specific training and culture. Material and methods. A 5-item multiple choice illustrated questionnaire was presented to 236 dental students. The Photoshop CS3 software was used in order to obtain the sets of images for the illustrated questions. The original image was handpicked from the internet by a panel of young dentists from a series of 15 pictures of people considered to have attractive faces. For each of the questions, the images presented were simulating deviations from the ideally symmetric and proportionate face. The sets of images consisted in multiple variations of deviations mixed with the original photo. Junior and sophomore year students from our dental medical school, having different nationalities were required to participate in our questionnaire. Simple descriptive statistics were used to interpret the data. Results. Assessing the results obtained from the questionnaire it was observed that a majority of students considered as unattractive the overdevelopment of the lower third, while the initial image with perfect symmetry and proportion was considered as the most attractive by only 38.9% of the subjects. Likewise, regarding the symmetry 36.86% considered unattractive the canting of the inter-commissural line. The interviewed subjects considered that for a face to be attractive it needs to have harmonious proportions between the different facial

  5. Analysis of facial expressions in parkinson's disease through video-based automatic methods.

    PubMed

    Bandini, Andrea; Orlandi, Silvia; Escalante, Hugo Jair; Giovannelli, Fabio; Cincotta, Massimo; Reyes-Garcia, Carlos A; Vanni, Paola; Zaccara, Gaetano; Manfredi, Claudia

    2017-04-01

    The automatic analysis of facial expressions is an evolving field that finds several clinical applications. One of these applications is the study of facial bradykinesia in Parkinson's disease (PD), which is a major motor sign of this neurodegenerative illness. Facial bradykinesia consists in the reduction/loss of facial movements and emotional facial expressions called hypomimia. In this work we propose an automatic method for studying facial expressions in PD patients relying on video-based METHODS: 17 Parkinsonian patients and 17 healthy control subjects were asked to show basic facial expressions, upon request of the clinician and after the imitation of a visual cue on a screen. Through an existing face tracker, the Euclidean distance of the facial model from a neutral baseline was computed in order to quantify the changes in facial expressivity during the tasks. Moreover, an automatic facial expressions recognition algorithm was trained in order to study how PD expressions differed from the standard expressions. Results show that control subjects reported on average higher distances than PD patients along the tasks. This confirms that control subjects show larger movements during both posed and imitated facial expressions. Moreover, our results demonstrate that anger and disgust are the two most impaired expressions in PD patients. Contactless video-based systems can be important techniques for analyzing facial expressions also in rehabilitation, in particular speech therapy, where patients could get a definite advantage from a real-time feedback about the proper facial expressions/movements to perform. Copyright © 2017 Elsevier B.V. All rights reserved.

  6. Forensic facial comparison in South Africa: State of the science.

    PubMed

    Steyn, M; Pretorius, M; Briers, N; Bacci, N; Johnson, A; Houlton, T M R

    2018-06-01

    Forensic facial comparison (FFC) is a scientific technique used to link suspects to a crime scene based on the analysis of photos or video recordings from that scene. While basic guidelines on practice and training are provided by the Facial Identification Scientific Working Group, details of how these are applied across the world are scarce. FFC is frequently used in South Africa, with more than 700 comparisons conducted in the last two years alone. In this paper the standards of practice are outlined, with new proposed levels of agreement/conclusions. We outline three levels of training that were established, with training in facial anatomy, terminology, principles of image comparison, image science, facial recognition and computer skills being aimed at developing general competency. Training in generating court charts and understanding court case proceedings are being specifically developed for the South African context. Various shortcomings still exist, specifically with regard to knowledge of the reliability of the technique. These need to be addressed in future research. Copyright © 2018 Elsevier B.V. All rights reserved.

  7. Facial recognition from volume-rendered magnetic resonance imaging data.

    PubMed

    Prior, Fred W; Brunsden, Barry; Hildebolt, Charles; Nolan, Tracy S; Pringle, Michael; Vaishnavi, S Neil; Larson-Prior, Linda J

    2009-01-01

    Three-dimensional (3-D) reconstructions of computed tomography (CT) and magnetic resonance (MR) brain imaging studies are a routine component of both clinical practice and clinical and translational research. A side effect of such reconstructions is the creation of a potentially recognizable face. The Health Insurance Portability and Accountability Act of 1996 (HIPAA) Privacy Rule requires that individually identifiable health information may not be used for research unless identifiers that may be associated with the health information including "Full face photographic images and other comparable images ..." are removed (de-identification). Thus, a key question is: Are reconstructed facial images comparable to full-face photographs for the purpose of identification? To address this question, MR images were selected from existing research repositories and subjects were asked to pair an MR reconstruction with one of 40 photographs. The chance probability that an observer could match a photograph with its 3-D MR image was 1 in 40 (0.025), and we considered 4 successes out of 40 (4/40, 0.1) to indicate that a subject could identify persons' faces from their 3-D MR images. Forty percent of the subjects were able to successfully match photographs with MR images with success rates higher than the null hypothesis success rate. The Blyth-Still-Casella 95% confidence interval for the 40% success rate was 29%-52%, and the 40% success rate was significantly higher ( P < 0.001) than our null hypothesis success rate of 1 in 10 (0.10).

  8. Dynamic facial expression recognition based on geometric and texture features

    NASA Astrophysics Data System (ADS)

    Li, Ming; Wang, Zengfu

    2018-04-01

    Recently, dynamic facial expression recognition in videos has attracted growing attention. In this paper, we propose a novel dynamic facial expression recognition method by using geometric and texture features. In our system, the facial landmark movements and texture variations upon pairwise images are used to perform the dynamic facial expression recognition tasks. For one facial expression sequence, pairwise images are created between the first frame and each of its subsequent frames. Integration of both geometric and texture features further enhances the representation of the facial expressions. Finally, Support Vector Machine is used for facial expression recognition. Experiments conducted on the extended Cohn-Kanade database show that our proposed method can achieve a competitive performance with other methods.

  9. Facial expression system on video using widrow hoff

    NASA Astrophysics Data System (ADS)

    Jannah, M.; Zarlis, M.; Mawengkang, H.

    2018-03-01

    Facial expressions recognition is one of interesting research. This research contains human feeling to computer application Such as the interaction between human and computer, data compression, facial animation and facial detection from the video. The purpose of this research is to create facial expression system that captures image from the video camera. The system in this research uses Widrow-Hoff learning method in training and testing image with Adaptive Linear Neuron (ADALINE) approach. The system performance is evaluated by two parameters, detection rate and false positive rate. The system accuracy depends on good technique and face position that trained and tested.

  10. Judgment of Nasolabial Esthetics in Cleft Lip and Palate Is Not Influenced by Overall Facial Attractiveness.

    PubMed

    Kocher, Katharina; Kowalski, Piotr; Kolokitha, Olga-Elpis; Katsaros, Christos; Fudalej, Piotr S

    2016-05-01

    To determine whether judgment of nasolabial esthetics in cleft lip and palate (CLP) is influenced by overall facial attractiveness. Experimental study. University of Bern, Switzerland. Seventy-two fused images (36 of boys, 36 of girls) were constructed. Each image comprised (1) the nasolabial region of a treated child with complete unilateral CLP (UCLP) and (2) the external facial features, i.e., the face with masked nasolabial region, of a noncleft child. Photographs of the nasolabial region of six boys and six girls with UCLP representing a wide range of esthetic outcomes, i.e., from very good to very poor appearance, were randomly chosen from a sample of 60 consecutively treated patients in whom nasolabial esthetics had been rated in a previous study. Photographs of external facial features of six boys and six girls without UCLP with various esthetics were randomly selected from patients' files. Eight lay raters evaluated the fused images using a 100-mm visual analogue scale. Method reliability was assessed by reevaluation of fused images after >1 month. A regression model was used to analyze which elements of facial esthetics influenced the perception of nasolabial appearance. Method reliability was good. A regression analysis demonstrated that only the appearance of the nasolabial area affected the esthetic scores of fused images (coefficient = -11.44; P < .001; R(2) = 0.464). The appearance of the external facial features did not influence perceptions of fused images. Cropping facial images for assessment of nasolabial appearance in CLP seems unnecessary. Instead, esthetic evaluation can be performed on images of full faces.

  11. Facial Orientation and Facial Shape in Extant Great Apes: A Geometric Morphometric Analysis of Covariation

    PubMed Central

    Neaux, Dimitri; Guy, Franck; Gilissen, Emmanuel; Coudyzer, Walter; Vignaud, Patrick; Ducrocq, Stéphane

    2013-01-01

    The organization of the bony face is complex, its morphology being influenced in part by the rest of the cranium. Characterizing the facial morphological variation and craniofacial covariation patterns in extant hominids is fundamental to the understanding of their evolutionary history. Numerous studies on hominid facial shape have proposed hypotheses concerning the relationship between the anterior facial shape, facial block orientation and basicranial flexion. In this study we test these hypotheses in a sample of adult specimens belonging to three extant hominid genera (Homo, Pan and Gorilla). Intraspecific variation and covariation patterns are analyzed using geometric morphometric methods and multivariate statistics, such as partial least squared on three-dimensional landmarks coordinates. Our results indicate significant intraspecific covariation between facial shape, facial block orientation and basicranial flexion. Hominids share similar characteristics in the relationship between anterior facial shape and facial block orientation. Modern humans exhibit a specific pattern in the covariation between anterior facial shape and basicranial flexion. This peculiar feature underscores the role of modern humans' highly-flexed basicranium in the overall integration of the cranium. Furthermore, our results are consistent with the hypothesis of a relationship between the reduction of the value of the cranial base angle and a downward rotation of the facial block in modern humans, and to a lesser extent in chimpanzees. PMID:23441232

  12. Decoding facial expressions based on face-selective and motion-sensitive areas.

    PubMed

    Liang, Yin; Liu, Baolin; Xu, Junhai; Zhang, Gaoyan; Li, Xianglin; Wang, Peiyuan; Wang, Bin

    2017-06-01

    Humans can easily recognize others' facial expressions. Among the brain substrates that enable this ability, considerable attention has been paid to face-selective areas; in contrast, whether motion-sensitive areas, which clearly exhibit sensitivity to facial movements, are involved in facial expression recognition remained unclear. The present functional magnetic resonance imaging (fMRI) study used multi-voxel pattern analysis (MVPA) to explore facial expression decoding in both face-selective and motion-sensitive areas. In a block design experiment, participants viewed facial expressions of six basic emotions (anger, disgust, fear, joy, sadness, and surprise) in images, videos, and eyes-obscured videos. Due to the use of multiple stimulus types, the impacts of facial motion and eye-related information on facial expression decoding were also examined. It was found that motion-sensitive areas showed significant responses to emotional expressions and that dynamic expressions could be successfully decoded in both face-selective and motion-sensitive areas. Compared with static stimuli, dynamic expressions elicited consistently higher neural responses and decoding performance in all regions. A significant decrease in both activation and decoding accuracy due to the absence of eye-related information was also observed. Overall, the findings showed that emotional expressions are represented in motion-sensitive areas in addition to conventional face-selective areas, suggesting that motion-sensitive regions may also effectively contribute to facial expression recognition. The results also suggested that facial motion and eye-related information played important roles by carrying considerable expression information that could facilitate facial expression recognition. Hum Brain Mapp 38:3113-3125, 2017. © 2017 Wiley Periodicals, Inc. © 2017 Wiley Periodicals, Inc.

  13. Effects of cultural characteristics on building an emotion classifier through facial expression analysis

    NASA Astrophysics Data System (ADS)

    da Silva, Flávio Altinier Maximiano; Pedrini, Helio

    2015-03-01

    Facial expressions are an important demonstration of humanity's humors and emotions. Algorithms capable of recognizing facial expressions and associating them with emotions were developed and employed to compare the expressions that different cultural groups use to show their emotions. Static pictures of predominantly occidental and oriental subjects from public datasets were used to train machine learning algorithms, whereas local binary patterns, histogram of oriented gradients (HOGs), and Gabor filters were employed to describe the facial expressions for six different basic emotions. The most consistent combination, formed by the association of HOG filter and support vector machines, was then used to classify the other cultural group: there was a strong drop in accuracy, meaning that the subtle differences of facial expressions of each culture affected the classifier performance. Finally, a classifier was trained with images from both occidental and oriental subjects and its accuracy was higher on multicultural data, evidencing the need of a multicultural training set to build an efficient classifier.

  14. Virtual transplantation in designing a facial prosthesis for extensive maxillofacial defects that cross the facial midline using computer-assisted technology.

    PubMed

    Feng, Zhi-hong; Dong, Yan; Bai, Shi-zhu; Wu, Guo-feng; Bi, Yun-peng; Wang, Bo; Zhao, Yi-min

    2010-01-01

    The aim of this article was to demonstrate a novel approach to designing facial prostheses using the transplantation concept and computer-assisted technology for extensive, large, maxillofacial defects that cross the facial midline. The three-dimensional (3D) facial surface images of a patient and his relative were reconstructed using data obtained through optical scanning. Based on these images, the corresponding portion of the relative's face was transplanted to the patient's where the defect was located, which could not be rehabilitated using mirror projection, to design the virtual facial prosthesis without the eye. A 3D model of an artificial eye that mimicked the patient's remaining one was developed, transplanted, and fit onto the virtual prosthesis. A personalized retention structure for the artificial eye was designed on the virtual facial prosthesis. The wax prosthesis was manufactured through rapid prototyping, and the definitive silicone prosthesis was completed. The size, shape, and cosmetic appearance of the prosthesis were satisfactory and matched the defect area well. The patient's facial appearance was recovered perfectly with the prosthesis, as determined through clinical evaluation. The optical 3D imaging and computer-aided design/computer-assisted manufacturing system used in this study can design and fabricate facial prostheses more precisely than conventional manual sculpturing techniques. The discomfort generally associated with such conventional methods was decreased greatly. The virtual transplantation used to design the facial prosthesis for the maxillofacial defect, which crossed the facial midline, and the development of the retention structure for the eye were both feasible.

  15. Facial Fractures.

    PubMed

    Ghosh, Rajarshi; Gopalkrishnan, Kulandaswamy

    2018-06-01

    The aim of this study is to retrospectively analyze the incidence of facial fractures along with age, gender predilection, etiology, commonest site, associated dental injuries, and any complications of patients operated in Craniofacial Unit of SDM College of Dental Sciences and Hospital. This retrospective study was conducted at the Department of OMFS, SDM College of Dental Sciences, Dharwad from January 2003 to December 2013. Data were recorded for the cause of injury, age and gender distribution, frequency and type of injury, localization and frequency of soft tissue injuries, dentoalveolar trauma, facial bone fractures, complications, concomitant injuries, and different treatment protocols.All the data were analyzed using statistical analysis that is chi-squared test. A total of 1146 patients reported at our unit with facial fractures during these 10 years. Males accounted for a higher frequency of facial fractures (88.8%). Mandible was the commonest bone to be fractured among all the facial bones (71.2%). Maxillary central incisors were the most common teeth to be injured (33.8%) and avulsion was the most common type of injury (44.6%). Commonest postoperative complication was plate infection (11%) leading to plate removal. Other injuries associated with facial fractures were rib fractures, head injuries, upper and lower limb fractures, etc., among these rib fractures were seen most frequently (21.6%). This study was performed to compare the different etiologic factors leading to diverse facial fracture patterns. By statistical analysis of this record the authors come to know about the relationship of facial fractures with gender, age, associated comorbidities, etc.

  16. Scales of degree of facial paralysis: analysis of agreement.

    PubMed

    Fonseca, Kércia Melo de Oliveira; Mourão, Aline Mansueto; Motta, Andréa Rodrigues; Vicente, Laelia Cristina Caseiro

    2015-01-01

    It has become common to use scales to measure the degree of involvement of facial paralysis in phonoaudiological clinics. To analyze the inter- and intra-rater agreement of the scales of degree of facial paralysis and to elicit point of view of the appraisers regarding their use. Cross-sectional observational clinical study of the Chevalier and House & Brackmann scales performed by five speech therapists with clinical experience, who analyzed the facial expression of 30 adult subjects with impaired facial movements two times, with a one week interval between evaluations. The kappa analysis was employed. There was excellent inter-rater agreement for both scales (kappa>0.80), and on the Chevalier scale a substantial intra-rater agreement in the first assessment (kappa=0.792) and an excellent agreement in the second assessment (kappa=0.928). The House & Brackmann scale showed excellent agreement at both assessments (kappa=0.850 and 0.857). As for the appraisers' point of view, one appraiser thought prior training is necessary for the Chevalier scale and, four appraisers felt that training is important for the House & Brackmann scale. Both scales have good inter- and intra-rater agreement and most of the appraisers agree on the ease and relevance of the application of these scales. Copyright © 2014 Associação Brasileira de Otorrinolaringologia e Cirurgia Cérvico-Facial. Published by Elsevier Editora Ltda. All rights reserved.

  17. Evaluation of psychological stress in confined environments using salivary, skin, and facial image parameters.

    PubMed

    Egawa, Mariko; Haze, Shinichiro; Gozu, Yoko; Hosoi, Junichi; Onodera, Tomoko; Tojo, Yosuke; Katsuyama, Masako; Hara, Yusuke; Katagiri, Chika; Inoue, Natsuhiko; Furukawa, Satoshi; Suzuki, Go

    2018-05-29

    Detecting the influence of psychological stress is particularly important in prolonged space missions. In this study, we determined potential markers of psychological stress in a confined environment. We examined 23 Japanese subjects staying for 2 weeks in a confined facility at Tsukuba Space Center, measuring salivary, skin, and facial image parameters. Saliva was collected at four points in a single day to detect diurnal variation. Increases in salivary cortisol were detected after waking up on the 4th and 11th days, and at 15:30 on the 1st and in the second half of the stay. Transepidermal water loss (TEWL) and sebum content of the skin were higher compared with outside the facility on the 4th and 1st days respectively. Increased IL-1β in the stripped stratum corneum was observed on the 14th day, and 7 days after leaving. Differences in facial expression symmetry at the time of facial expression changes were observed on 11th and 14th days. Thus, we detected a transition of psychological stress using salivary cortisol profiles and skin physiological parameters. The results also suggested that IL-1β in the stripped stratum corneum and facial expression symmetry are possible novel markers for conveniently detecting psychological stress.

  18. Expression-dependent susceptibility to face distortions in processing of facial expressions of emotion.

    PubMed

    Guo, Kun; Soornack, Yoshi; Settle, Rebecca

    2018-03-05

    Our capability of recognizing facial expressions of emotion under different viewing conditions implies the existence of an invariant expression representation. As natural visual signals are often distorted and our perceptual strategy changes with external noise level, it is essential to understand how expression perception is susceptible to face distortion and whether the same facial cues are used to process high- and low-quality face images. We systematically manipulated face image resolution (experiment 1) and blur (experiment 2), and measured participants' expression categorization accuracy, perceived expression intensity and associated gaze patterns. Our analysis revealed a reasonable tolerance to face distortion in expression perception. Reducing image resolution up to 48 × 64 pixels or increasing image blur up to 15 cycles/image had little impact on expression assessment and associated gaze behaviour. Further distortion led to decreased expression categorization accuracy and intensity rating, increased reaction time and fixation duration, and stronger central fixation bias which was not driven by distortion-induced changes in local image saliency. Interestingly, the observed distortion effects were expression-dependent with less deterioration impact on happy and surprise expressions, suggesting this distortion-invariant facial expression perception might be achieved through the categorical model involving a non-linear configural combination of local facial features. Copyright © 2018 Elsevier Ltd. All rights reserved.

  19. Factors contributing to the adaptation aftereffects of facial expression.

    PubMed

    Butler, Andrea; Oruc, Ipek; Fox, Christopher J; Barton, Jason J S

    2008-01-29

    Previous studies have demonstrated the existence of adaptation aftereffects for facial expressions. Here we investigated which aspects of facial stimuli contribute to these aftereffects. In Experiment 1, we examined the role of local adaptation to image elements such as curvature, shape and orientation, independent of expression, by using hybrid faces constructed from either the same or opposing expressions. While hybrid faces made with consistent expressions generated aftereffects as large as those with normal faces, there were no aftereffects from hybrid faces made from different expressions, despite the fact that these contained the same local image elements. In Experiment 2, we examined the role of facial features independent of the normal face configuration by contrasting adaptation with whole faces to adaptation with scrambled faces. We found that scrambled faces also generated significant aftereffects, indicating that expressive features without a normal facial configuration could generate expression aftereffects. In Experiment 3, we examined the role of facial configuration by using schematic faces made from line elements that in isolation do not carry expression-related information (e.g. curved segments and straight lines) but that convey an expression when arranged in a normal facial configuration. We obtained a significant aftereffect for facial configurations but not scrambled configurations of these line elements. We conclude that facial expression aftereffects are not due to local adaptation to image elements but due to high-level adaptation of neural representations that involve both facial features and facial configuration.

  20. Misleading first impressions: different for different facial images of the same person.

    PubMed

    Todorov, Alexander; Porter, Jenny M

    2014-07-01

    Studies on first impressions from facial appearance have rapidly proliferated in the past decade. Almost all of these studies have relied on a single face image per target individual, and differences in impressions have been interpreted as originating in stable physiognomic differences between individuals. Here we show that images of the same individual can lead to different impressions, with within-individual image variance comparable to or exceeding between-individuals variance for a variety of social judgments (Experiment 1). We further show that preferences for images shift as a function of the context (e.g., selecting an image for online dating vs. a political campaign; Experiment 2), that preferences are predictably biased by the selection of the images (e.g., an image fitting a political campaign vs. a randomly selected image; Experiment 3), and that these biases are evident after extremely brief (40-ms) presentation of the images (Experiment 4). We discuss the implications of these findings for studies on the accuracy of first impressions. © The Author(s) 2014.

  1. Regional facial asymmetries and attractiveness of the face.

    PubMed

    Kaipainen, Anu E; Sieber, Kevin R; Nada, Rania M; Maal, Thomas J; Katsaros, Christos; Fudalej, Piotr S

    2016-12-01

    Facial attractiveness is an important factor in our social interactions. It is still not entirely clear which factors influence the attractiveness of a face and facial asymmetry appears to play a certain role. The aim of the present study was to assess the association between facial attractiveness and regional facial asymmetries evaluated on three-dimensional (3D) images. 3D facial images of 59 (23 male, 36 female) young adult patients (age 16-25 years) before orthodontic treatment were evaluated for asymmetry. The same 3D images were presented to 12 lay judges who rated the attractiveness of each subject on a 100mm visual analogue scale. Reliability of the method was assessed with Bland-Altman plots and Cronbach's alpha coefficient. All subjects showed a certain amount of asymmetry in all regions of the face; most asymmetry was found in the chin and cheek areas and less in the lip, nose and forehead areas. No statistically significant differences in regional facial asymmetries were found between male and female subjects (P > 0.05). Regression analyses demonstrated that the judgement of facial attractiveness was not influenced by absolute regional facial asymmetries when gender, facial width-to-height ratio and type of malocclusion were controlled (P > 0.05). A potential limitation of the study could be that other biologic and cultural factors influencing the perception of facial attractiveness were not controlled for. A small amount of asymmetry was present in all subjects assessed in this study, and asymmetry of this magnitude may not influence the assessment of facial attractiveness. © The Author 2015. Published by Oxford University Press on behalf of the European Orthodontic Society. All rights reserved. For permissions, please email: journals.permissions@oup.com.

  2. Facial soft biometric features for forensic face recognition.

    PubMed

    Tome, Pedro; Vera-Rodriguez, Ruben; Fierrez, Julian; Ortega-Garcia, Javier

    2015-12-01

    This paper proposes a functional feature-based approach useful for real forensic caseworks, based on the shape, orientation and size of facial traits, which can be considered as a soft biometric approach. The motivation of this work is to provide a set of facial features, which can be understood by non-experts such as judges and support the work of forensic examiners who, in practice, carry out a thorough manual comparison of face images paying special attention to the similarities and differences in shape and size of various facial traits. This new approach constitutes a tool that automatically converts a set of facial landmarks to a set of features (shape and size) corresponding to facial regions of forensic value. These features are furthermore evaluated in a population to generate statistics to support forensic examiners. The proposed features can also be used as additional information that can improve the performance of traditional face recognition systems. These features follow the forensic methodology and are obtained in a continuous and discrete manner from raw images. A statistical analysis is also carried out to study the stability, discrimination power and correlation of the proposed facial features on two realistic databases: MORPH and ATVS Forensic DB. Finally, the performance of both continuous and discrete features is analyzed using different similarity measures. Experimental results show high discrimination power and good recognition performance, especially for continuous features. A final fusion of the best systems configurations achieves rank 10 match results of 100% for ATVS database and 75% for MORPH database demonstrating the benefits of using this information in practice. Copyright © 2015 Elsevier Ireland Ltd. All rights reserved.

  3. Facial image of Biblical Jews from Israel.

    PubMed

    Kobyliansky, E; Balueva, T; Veselovskaya, E; Arensburg, B

    2008-06-01

    The present report deals with reconstructing the facial shapes of ancient inhabitants of Israel based on their cranial remains. The skulls of a male from the Hellenistic period and a female from the Roman period have been reconstructed. They were restored using the most recently developed programs in anthropological facial reconstruction, especially that of the Institute of Ethnology and Anthropology of the Russian Academy of Sciences (Balueva & Veselovskaya 2004). The basic craniometrical measurements of the two skulls were measured according to Martin & Saller (1957) and compared to the data from three ancient populations of Israel described by Arensburg et al. (1980): that of the Hellenistic period dating from 332 to 37 B.C., that of the Roman period, from 37 B.C. to 324 C.E., and that of the Byzantine period that continued until the Arab conquest in 640 C.E. Most of this osteological material was excavated in the Jordan River and the Dead Sea areas. A sample from the XVIIth century Jews from Prague (Matiegka 1926) was also used for osteometrical comparisons. The present study will characterize not only the osteological morphology of the material, but also the facial appearance of ancient inhabitants of Israel. From an anthropometric point of view, the two skulls studied here definitely belong to the same sample from the Hellenistic, Roman, and Byzantine populations of Israel as well as from Jews from Prague. Based on its facial reconstruction, the male skull may belong to the large Mediterranean group that inhabited this area from historic to modern times. The female skull also exhibits all the Mediterranean features but, in addition, probably some equatorial (African) mixture manifested by the shape of the reconstructed nose and the facial prognatism.

  4. Thermographic imaging of facial and ventilatory activity during vocalization, speech and expiration (Conference Presentation)

    NASA Astrophysics Data System (ADS)

    Izdebski, Krzysztof; Jarosz, Paweł; Usydus, Ireneusz

    2017-02-01

    Ventilation, speech and singing must use facial musculature to complete these motor tasks and these tasks are fueled by the air we inhale. This motor process requires increase in the blood flow as the muscles contract and relax, therefore skin surface temperature changes are expected. Hence, we used thermography to image these effects. The system used was the thermography camera model FLIR X6580sc with a chilled detector (FLIR Systems Advanced Thermal Solutions, 27700 SW Parkway Ave Wilsonville, OR 97070, USA). To assure improved imaging, the room temperature was air-conditioned to +18° C. All images were recoded at the speed of 30 f/s. Acquired data were analyzed with FLIR Research IR Max Version 4 software and software filters. In this preliminary study a male subject was imaged from frontal and lateral views simultaneously while he performed normal resting ventilation, speech and song. The lateral image was captured in a stainless steel mirror. Results showed different levels of heat flow in the facial musculature as a function of these three tasks. Also, we were able to capture the exalted air jet directionality. The breathing jet was discharged in horizontal direction, speaking voice jet was discharged downwards while singing jet went upward. We interpreted these jet directions as representing different gas content of air expired during these different tasks, with speech having less oxygen than singing. Further studies examining gas exchange during various forms of speech and song and emotional states are warranted.

  5. Composite Artistry Meets Facial Recognition Technology: Exploring the Use of Facial Recognition Technology to Identify Composite Images

    DTIC Science & Technology

    2011-09-01

    be submitted into a facial recognition program for comparison with millions of possible matches, offering abundant opportunities to identify the...to leverage the robust number of comparative opportunities associated with facial recognition programs. This research investigates the efficacy of...combining composite forensic artistry with facial recognition technology to create a viable investigative tool to identify suspects, as well as better

  6. Thermal imaging as a biometrics approach to facial signature authentication.

    PubMed

    Guzman, A M; Goryawala, M; Wang, Jin; Barreto, A; Andrian, J; Rishe, N; Adjouadi, M

    2013-01-01

    A new thermal imaging framework with unique feature extraction and similarity measurements for face recognition is presented. The research premise is to design specialized algorithms that would extract vasculature information, create a thermal facial signature and identify the individual. The proposed algorithm is fully integrated and consolidates the critical steps of feature extraction through the use of morphological operators, registration using the Linear Image Registration Tool and matching through unique similarity measures designed for this task. The novel approach at developing a thermal signature template using four images taken at various instants of time ensured that unforeseen changes in the vasculature over time did not affect the biometric matching process as the authentication process relied only on consistent thermal features. Thirteen subjects were used for testing the developed technique on an in-house thermal imaging system. The matching using the similarity measures showed an average accuracy of 88.46% for skeletonized signatures and 90.39% for anisotropically diffused signatures. The highly accurate results obtained in the matching process clearly demonstrate the ability of the thermal infrared system to extend in application to other thermal imaging based systems. Empirical results applying this approach to an existing database of thermal images proves this assertion.

  7. Analysis of facial motion patterns during speech using a matrix factorization algorithm

    PubMed Central

    Lucero, Jorge C.; Munhall, Kevin G.

    2008-01-01

    This paper presents an analysis of facial motion during speech to identify linearly independent kinematic regions. The data consists of three-dimensional displacement records of a set of markers located on a subject’s face while producing speech. A QR factorization with column pivoting algorithm selects a subset of markers with independent motion patterns. The subset is used as a basis to fit the motion of the other facial markers, which determines facial regions of influence of each of the linearly independent markers. Those regions constitute kinematic “eigenregions” whose combined motion produces the total motion of the face. Facial animations may be generated by driving the independent markers with collected displacement records. PMID:19062866

  8. Forensic facial reconstruction: Nasal projection in Brazilian adults.

    PubMed

    Tedeschi-Oliveira, Silvia Virginia; Beaini, Thiago Leite; Melani, Rodolfo Francisco Haltenhoff

    2016-09-01

    The nose has a marked cognitive influence on facial image; however, it loses its shape during cadaveric decomposition. The known methods of estimating nasal projection using Facial Reconstruction are lacking in practicality and reproducibility. We attempted to relate the points Rhinion, Pronasale and Prosthion by studying the angle formed by straight lines that connect them. Two examiners measured this angle with the help of analysis and image-processing software, Image J, directly from cephalometric radiographs. The sample consisted of 300 males, aged between 24 and 77 years, and 300 females, aged 24 to 69 years. The proposed angle ranged from 80° to 100° in both sexes and all ages. It was considered possible to use a 90° angle from projections of the Rhinion and Prosthion points in order to determine the Pronasale position, as well as to estimate the nasal projection of Brazilian adults. Copyright © 2016 Elsevier Ireland Ltd. All rights reserved.

  9. What's behind the mask? A look at blood flow changes with prolonged facial pressure and expression using laser Doppler imaging.

    PubMed

    Van-Buendia, Lan B; Allely, Rebekah R; Lassiter, Ronald; Weinand, Christian; Jordan, Marion H; Jeng, James C

    2010-01-01

    Clinically, the initial blanching in burn scar seen on transparent plastic face mask application seems to diminish with time and movement requiring mask alteration. To date, studies quantifying perfusion with prolonged mask use do not exist. This study used laser Doppler imaging (LDI) to assess perfusion through the transparent face mask and movement in subjects with and without burn over time. Five subjects fitted with transparent face masks were scanned with the LDI on four occasions. The four subjects without burn were scanned in the following manner: 1) no mask, 2) mask on while at rest, 3) mask on with alternating intervals of sustained facial expression and rest, and 4) after mask removal. Images were acquired every 3 minutes throughout the 85-minute study period. The subject with burn underwent a shortened scanning protocol to increase comfort. Each face was divided into five regions of interest for analysis. Compared with baseline, mask application decreased perfusion significantly in all subjects (P < .0001). Perfusion did not change during the rest period. There were no significant differences with changing facial expression in any of the regions of interest. On mask removal, all regions of the face demonstrated a hyperemic effect with the chin (P = .05) and each cheek (P < .0001) reaching statistical significance. Perfusion levels did not return to baseline in the chin and cheeks after 30 minutes of mask removal. Perfusions remain constantly low while wearing the face mask, despite changing facial expressions. Changing facial expressions with the mask on did not alter perfusion. Hyperemic response occurs on removal of the mask. This study exposed methodology and statistical issues worth considering when conducting future research with the face, pressure therapy, and with LDI technology.

  10. Real Time 3D Facial Movement Tracking Using a Monocular Camera.

    PubMed

    Dong, Yanchao; Wang, Yanming; Yue, Jiguang; Hu, Zhencheng

    2016-07-25

    The paper proposes a robust framework for 3D facial movement tracking in real time using a monocular camera. It is designed to estimate the 3D face pose and local facial animation such as eyelid movement and mouth movement. The framework firstly utilizes the Discriminative Shape Regression method to locate the facial feature points on the 2D image and fuses the 2D data with a 3D face model using Extended Kalman Filter to yield 3D facial movement information. An alternating optimizing strategy is adopted to fit to different persons automatically. Experiments show that the proposed framework could track the 3D facial movement across various poses and illumination conditions. Given the real face scale the framework could track the eyelid with an error of 1 mm and mouth with an error of 2 mm. The tracking result is reliable for expression analysis or mental state inference.

  11. Quantitative facial asymmetry: using three-dimensional photogrammetry to measure baseline facial surface symmetry.

    PubMed

    Taylor, Helena O; Morrison, Clinton S; Linden, Olivia; Phillips, Benjamin; Chang, Johnny; Byrne, Margaret E; Sullivan, Stephen R; Forrest, Christopher R

    2014-01-01

    Although symmetry is hailed as a fundamental goal of aesthetic and reconstructive surgery, our tools for measuring this outcome have been limited and subjective. With the advent of three-dimensional photogrammetry, surface geometry can be captured, manipulated, and measured quantitatively. Until now, few normative data existed with regard to facial surface symmetry. Here, we present a method for reproducibly calculating overall facial symmetry and present normative data on 100 subjects. We enrolled 100 volunteers who underwent three-dimensional photogrammetry of their faces in repose. We collected demographic data on age, sex, and race and subjectively scored facial symmetry. We calculated the root mean square deviation (RMSD) between the native and reflected faces, reflecting about a plane of maximum symmetry. We analyzed the interobserver reliability of the subjective assessment of facial asymmetry and the quantitative measurements and compared the subjective and objective values. We also classified areas of greatest asymmetry as localized to the upper, middle, or lower facial thirds. This cluster of normative data was compared with a group of patients with subtle but increasing amounts of facial asymmetry. We imaged 100 subjects by three-dimensional photogrammetry. There was a poor interobserver correlation between subjective assessments of asymmetry (r = 0.56). There was a high interobserver reliability for quantitative measurements of facial symmetry RMSD calculations (r = 0.91-0.95). The mean RMSD for this normative population was found to be 0.80 ± 0.24 mm. Areas of greatest asymmetry were distributed as follows: 10% upper facial third, 49% central facial third, and 41% lower facial third. Precise measurement permitted discrimination of subtle facial asymmetry within this normative group and distinguished norms from patients with subtle facial asymmetry, with placement of RMSDs along an asymmetry ruler. Facial surface symmetry, which is poorly assessed

  12. FaceWarehouse: a 3D facial expression database for visual computing.

    PubMed

    Cao, Chen; Weng, Yanlin; Zhou, Shun; Tong, Yiying; Zhou, Kun

    2014-03-01

    We present FaceWarehouse, a database of 3D facial expressions for visual computing applications. We use Kinect, an off-the-shelf RGBD camera, to capture 150 individuals aged 7-80 from various ethnic backgrounds. For each person, we captured the RGBD data of her different expressions, including the neutral expression and 19 other expressions such as mouth-opening, smile, kiss, etc. For every RGBD raw data record, a set of facial feature points on the color image such as eye corners, mouth contour, and the nose tip are automatically localized, and manually adjusted if better accuracy is required. We then deform a template facial mesh to fit the depth data as closely as possible while matching the feature points on the color image to their corresponding points on the mesh. Starting from these fitted face meshes, we construct a set of individual-specific expression blendshapes for each person. These meshes with consistent topology are assembled as a rank-3 tensor to build a bilinear face model with two attributes: identity and expression. Compared with previous 3D facial databases, for every person in our database, there is a much richer matching collection of expressions, enabling depiction of most human facial actions. We demonstrate the potential of FaceWarehouse for visual computing with four applications: facial image manipulation, face component transfer, real-time performance-based facial image animation, and facial animation retargeting from video to image.

  13. Do Dynamic Compared to Static Facial Expressions of Happiness and Anger Reveal Enhanced Facial Mimicry?

    PubMed Central

    Rymarczyk, Krystyna; Żurawski, Łukasz; Jankowiak-Siuda, Kamila; Szatkowska, Iwona

    2016-01-01

    Facial mimicry is the spontaneous response to others’ facial expressions by mirroring or matching the interaction partner. Recent evidence suggested that mimicry may not be only an automatic reaction but could be dependent on many factors, including social context, type of task in which the participant is engaged, or stimulus properties (dynamic vs static presentation). In the present study, we investigated the impact of dynamic facial expression and sex differences on facial mimicry and judgment of emotional intensity. Electromyography recordings were recorded from the corrugator supercilii, zygomaticus major, and orbicularis oculi muscles during passive observation of static and dynamic images of happiness and anger. The ratings of the emotional intensity of facial expressions were also analysed. As predicted, dynamic expressions were rated as more intense than static ones. Compared to static images, dynamic displays of happiness also evoked stronger activity in the zygomaticus major and orbicularis oculi, suggesting that subjects experienced positive emotion. No muscles showed mimicry activity in response to angry faces. Moreover, we found that women exhibited greater zygomaticus major muscle activity in response to dynamic happiness stimuli than static stimuli. Our data support the hypothesis that people mimic positive emotions and confirm the importance of dynamic stimuli in some emotional processing. PMID:27390867

  14. Facial Palsy Following Embolization of a Juvenile Nasopharyngeal Angiofibroma.

    PubMed

    Tawfik, Kareem O; Harmon, Jeffrey J; Walters, Zoe; Samy, Ravi; de Alarcon, Alessandro; Stevens, Shawn M; Abruzzo, Todd

    2018-05-01

    To describe a case of the rare complication of facial palsy following preoperative embolization of a juvenile nasopharyngeal angiofibroma (JNA). To illustrate the vascular supply to the facial nerve and as a result, highlight the etiology of the facial nerve palsy. The angiography and magnetic resonance (MR) imaging of a case of facial palsy following preoperative embolization of a JNA is reviewed. A 13-year-old male developed left-sided facial palsy following preoperative embolization of a left-sided JNA. Evaluation of MR imaging studies and retrospective review of the angiographic data suggested errant embolization of particles into the petrosquamosal branch of the middle meningeal artery (MMA), a branch of the internal maxillary artery (IMA), through collateral vasculature. The petrosquamosal branch of the MMA is the predominant blood supply to the facial nerve in the facial canal. The facial palsy resolved since complete infarction of the nerve was likely prevented by collateral blood supply from the stylomastoid artery. Facial palsy is a potential complication of embolization of the IMA, a branch of the external carotid artery (ECA). This is secondary to ischemia of the facial nerve due to embolization of its vascular supply. Clinicians should be aware of this potential complication and counsel patients accordingly prior to embolization for JNA.

  15. In search of Leonardo: computer-based facial image analysis of Renaissance artworks for identifying Leonardo as subject

    NASA Astrophysics Data System (ADS)

    Tyler, Christopher W.; Smith, William A. P.; Stork, David G.

    2012-03-01

    One of the enduring mysteries in the history of the Renaissance is the adult appearance of the archetypical "Renaissance Man," Leonardo da Vinci. His only acknowledged self-portrait is from an advanced age, and various candidate images of younger men are difficult to assess given the absence of documentary evidence. One clue about Leonardo's appearance comes from the remark of the contemporary historian, Vasari, that the sculpture of David by Leonardo's master, Andrea del Verrocchio, was based on the appearance of Leonardo when he was an apprentice. Taking a cue from this statement, we suggest that the more mature sculpture of St. Thomas, also by Verrocchio, might also have been a portrait of Leonardo. We tested the possibility Leonardo was the subject for Verrocchio's sculpture by a novel computational technique for the comparison of three-dimensional facial configurations. Based on quantitative measures of similarities, we also assess whether another pair of candidate two-dimensional images are plausibly attributable as being portraits of Leonardo as a young adult. Our results are consistent with the claim Leonardo is indeed the subject in these works, but we need comparisons with images in a larger corpora of candidate artworks before our results achieve statistical significance.

  16. Computerised analysis of facial emotion expression in eating disorders

    PubMed Central

    2017-01-01

    Background Problems with social-emotional processing are known to be an important contributor to the development and maintenance of eating disorders (EDs). Diminished facial communication of emotion has been frequently reported in individuals with anorexia nervosa (AN). Less is known about facial expressivity in bulimia nervosa (BN) and in people who have recovered from AN (RecAN). This study aimed to pilot the use of computerised facial expression analysis software to investigate emotion expression across the ED spectrum and recovery in a large sample of participants. Method 297 participants with AN, BN, RecAN, and healthy controls were recruited. Participants watched film clips designed to elicit happy or sad emotions, and facial expressions were then analysed using FaceReader. Results The finding mirrored those from previous work showing that healthy control and RecAN participants expressed significantly more positive emotions during the positive clip compared to the AN group. There were no differences in emotion expression during the sad film clip. Discussion These findings support the use of computerised methods to analyse emotion expression in EDs. The findings also demonstrate that reduced positive emotion expression is likely to be associated with the acute stage of AN illness, with individuals with BN showing an intermediate profile. PMID:28575109

  17. Computerised analysis of facial emotion expression in eating disorders.

    PubMed

    Leppanen, Jenni; Dapelo, Marcela Marin; Davies, Helen; Lang, Katie; Treasure, Janet; Tchanturia, Kate

    2017-01-01

    Problems with social-emotional processing are known to be an important contributor to the development and maintenance of eating disorders (EDs). Diminished facial communication of emotion has been frequently reported in individuals with anorexia nervosa (AN). Less is known about facial expressivity in bulimia nervosa (BN) and in people who have recovered from AN (RecAN). This study aimed to pilot the use of computerised facial expression analysis software to investigate emotion expression across the ED spectrum and recovery in a large sample of participants. 297 participants with AN, BN, RecAN, and healthy controls were recruited. Participants watched film clips designed to elicit happy or sad emotions, and facial expressions were then analysed using FaceReader. The finding mirrored those from previous work showing that healthy control and RecAN participants expressed significantly more positive emotions during the positive clip compared to the AN group. There were no differences in emotion expression during the sad film clip. These findings support the use of computerised methods to analyse emotion expression in EDs. The findings also demonstrate that reduced positive emotion expression is likely to be associated with the acute stage of AN illness, with individuals with BN showing an intermediate profile.

  18. Perception of health from facial cues

    PubMed Central

    Henderson, Audrey J.; Holzleitner, Iris J.; Talamas, Sean N.

    2016-01-01

    Impressions of health are integral to social interactions, yet poorly understood. A review of the literature reveals multiple facial characteristics that potentially act as cues to health judgements. The cues vary in their stability across time: structural shape cues including symmetry and sexual dimorphism alter slowly across the lifespan and have been found to have weak links to actual health, but show inconsistent effects on perceived health. Facial adiposity changes over a medium time course and is associated with both perceived and actual health. Skin colour alters over a short time and has strong effects on perceived health, yet links to health outcomes have barely been evaluated. Reviewing suggested an additional influence of demeanour as a perceptual cue to health. We, therefore, investigated the association of health judgements with multiple facial cues measured objectively from two-dimensional and three-dimensional facial images. We found evidence for independent contributions of face shape and skin colour cues to perceived health. Our empirical findings: (i) reinforce the role of skin yellowness; (ii) demonstrate the utility of global face shape measures of adiposity; and (iii) emphasize the role of affect in facial images with nominally neutral expression in impressions of health. PMID:27069057

  19. Looking Like a Leader–Facial Shape Predicts Perceived Height and Leadership Ability

    PubMed Central

    Re, Daniel E.; Hunter, David W.; Coetzee, Vinet; Tiddeman, Bernard P.; Xiao, Dengke; DeBruine, Lisa M.; Jones, Benedict C.; Perrett, David I.

    2013-01-01

    Judgments of leadership ability from face images predict the outcomes of actual political elections and are correlated with leadership success in the corporate world. The specific facial cues that people use to judge leadership remain unclear, however. Physical height is also associated with political and organizational success, raising the possibility that facial cues of height contribute to leadership perceptions. Consequently, we assessed whether cues to height exist in the face and, if so, whether they are associated with perception of leadership ability. We found that facial cues to perceived height had a strong relationship with perceived leadership ability. Furthermore, when allowed to manually manipulate faces, participants increased facial cues associated with perceived height in order to maximize leadership perception. A morphometric analysis of face shape revealed that structural facial masculinity was not responsible for the relationship between perceived height and perceived leadership ability. Given the prominence of facial appearance in making social judgments, facial cues to perceived height may have a significant influence on leadership selection. PMID:24324651

  20. Real Time 3D Facial Movement Tracking Using a Monocular Camera

    PubMed Central

    Dong, Yanchao; Wang, Yanming; Yue, Jiguang; Hu, Zhencheng

    2016-01-01

    The paper proposes a robust framework for 3D facial movement tracking in real time using a monocular camera. It is designed to estimate the 3D face pose and local facial animation such as eyelid movement and mouth movement. The framework firstly utilizes the Discriminative Shape Regression method to locate the facial feature points on the 2D image and fuses the 2D data with a 3D face model using Extended Kalman Filter to yield 3D facial movement information. An alternating optimizing strategy is adopted to fit to different persons automatically. Experiments show that the proposed framework could track the 3D facial movement across various poses and illumination conditions. Given the real face scale the framework could track the eyelid with an error of 1 mm and mouth with an error of 2 mm. The tracking result is reliable for expression analysis or mental state inference. PMID:27463714

  1. Automated facial recognition of manually generated clay facial approximations: Potential application in unidentified persons data repositories.

    PubMed

    Parks, Connie L; Monson, Keith L

    2018-01-01

    This research examined how accurately 2D images (i.e., photographs) of 3D clay facial approximations were matched to corresponding photographs of the approximated individuals using an objective automated facial recognition system. Irrespective of search filter (i.e., blind, sex, or ancestry) or rank class (R 1 , R 10 , R 25 , and R 50 ) employed, few operationally informative results were observed. In only a single instance of 48 potential match opportunities was a clay approximation matched to a corresponding life photograph within the top 50 images (R 50 ) of a candidate list, even with relatively small gallery sizes created from the application of search filters (e.g., sex or ancestry search restrictions). Increasing the candidate lists to include the top 100 images (R 100 ) resulted in only two additional instances of correct match. Although other untested variables (e.g., approximation method, 2D photographic process, and practitioner skill level) may have impacted the observed results, this study suggests that 2D images of manually generated clay approximations are not readily matched to life photos by automated facial recognition systems. Further investigation is necessary in order to identify the underlying cause(s), if any, of the poor recognition results observed in this study (e.g., potential inferior facial feature detection and extraction). Additional inquiry exploring prospective remedial measures (e.g., stronger feature differentiation) is also warranted, particularly given the prominent use of clay approximations in unidentified persons casework. Copyright © 2017. Published by Elsevier B.V.

  2. Empirical mode decomposition-based facial pose estimation inside video sequences

    NASA Astrophysics Data System (ADS)

    Qing, Chunmei; Jiang, Jianmin; Yang, Zhijing

    2010-03-01

    We describe a new pose-estimation algorithm via integration of the strength in both empirical mode decomposition (EMD) and mutual information. While mutual information is exploited to measure the similarity between facial images to estimate poses, EMD is exploited to decompose input facial images into a number of intrinsic mode function (IMF) components, which redistribute the effect of noise, expression changes, and illumination variations as such that, when the input facial image is described by the selected IMF components, all the negative effects can be minimized. Extensive experiments were carried out in comparisons to existing representative techniques, and the results show that the proposed algorithm achieves better pose-estimation performances with robustness to noise corruption, illumination variation, and facial expressions.

  3. Influence of anteroposterior mandibular positions on facial attractiveness in Japanese adults.

    PubMed

    Kuroda, Shingo; Sugahara, Takako; Takabatake, Souichirou; Taketa, Hiroaki; Ando, Ryoko; Takano-Yamamoto, Teruko

    2009-01-01

    Our aims in this study were to determine the anteroposterior facial relationship that is regarded as most attractive by Japanese laypersons in a questionnaire survey and to evaluate which analysis of the soft-tissue profile is most suitable for Japanese people. We showed 262 Japanese laypersons (121 male, 141 female) 9 morphed profile images with Point B and menton anteriorly or distally moved by software and asked them to number them sequentially according to their attractiveness. To examine which analysis best reflects facial attractiveness as judged by laypersons, we made 5 types of analyses of the facial profile with 11 variables in the 9 images. The normal face was judged favorably; however, an attractive profile might be different for each subject. The 3 highest ranking profiles (normal face and moderate mandibular retrusions) were often favorites, and 2 profiles (severe mandibular protrusions) were liked the least for most subjects. However, the other images showed a wide range of distribution. Mandibular retrusion was generally more favored than mandibular protrusion and bimaxillary protrusion (severe chin retrusion) had a high attractiveness ranking and was well accepted in the Japanese population. To evaluate the profiles of Japanese subjects, it is important to evaluate not only the esthetic line defined by the nose and chin, but also the balance of the upper and lower lips defined by the posterior reference line--ie, Burstone's Sn-Pog' line.

  4. Dependence of the appearance-based perception of criminality, suggestibility, and trustworthiness on the level of pixelation of facial images.

    PubMed

    Nurmoja, Merle; Eamets, Triin; Härma, Hanne-Loore; Bachmann, Talis

    2012-10-01

    While the dependence of face identification on the level of pixelation-transform of the images of faces has been well studied, similar research on face-based trait perception is underdeveloped. Because depiction formats used for hiding individual identity in visual media and evidential material recorded by surveillance cameras often consist of pixelized images, knowing the effects of pixelation on person perception has practical relevance. Here, the results of two experiments are presented showing the effect of facial image pixelation on the perception of criminality, trustworthiness, and suggestibility. It appears that individuals (N = 46, M age = 21.5 yr., SD = 3.1 for criminality ratings; N = 94, M age = 27.4 yr., SD = 10.1 for other ratings) have the ability to discriminate between facial cues ndicative of these perceived traits from the coarse level of image pixelation (10-12 pixels per face horizontally) and that the discriminability increases with a decrease in the coarseness of pixelation. Perceived criminality and trustworthiness appear to be better carried by the pixelized images than perceived suggestibility.

  5. Facial paralysis

    MedlinePlus

    ... a physical, speech, or occupational therapist. If facial paralysis from Bell palsy lasts for more than 6 to 12 months, plastic surgery may be recommended to help the eye close and improve the appearance of the face. Alternative Names Paralysis of the face Images Ptosis, drooping of the ...

  6. Wait, are you sad or angry? Large exposure time differences required for the categorization of facial expressions of emotion

    PubMed Central

    Du, Shichuan; Martinez, Aleix M.

    2013-01-01

    Abstract Facial expressions of emotion are essential components of human behavior, yet little is known about the hierarchical organization of their cognitive analysis. We study the minimum exposure time needed to successfully classify the six classical facial expressions of emotion (joy, surprise, sadness, anger, disgust, fear) plus neutral as seen at different image resolutions (240 × 160 to 15 × 10 pixels). Our results suggest a consistent hierarchical analysis of these facial expressions regardless of the resolution of the stimuli. Happiness and surprise can be recognized after very short exposure times (10–20 ms), even at low resolutions. Fear and anger are recognized the slowest (100–250 ms), even in high-resolution images, suggesting a later computation. Sadness and disgust are recognized in between (70–200 ms). The minimum exposure time required for successful classification of each facial expression correlates with the ability of a human subject to identify it correctly at low resolutions. These results suggest a fast, early computation of expressions represented mostly by low spatial frequencies or global configural cues and a later, slower process for those categories requiring a more fine-grained analysis of the image. We also demonstrate that those expressions that are mostly visible in higher-resolution images are not recognized as accurately. We summarize implications for current computational models. PMID:23509409

  7. External auditory canal cholesteatoma and keratosis obturans: the role of imaging in preventing facial nerve injury.

    PubMed

    McCoul, Edward D; Hanson, Matthew B

    2011-12-01

    We conducted a retrospective study to compare the clinical characteristics of external auditory canal cholesteatoma (EACC) with those of a similar entity, keratosis obturans (KO). We also sought to identify those aspects of each disease that may lead to complications. We identified 6 patients in each group. Imaging studies were reviewed for evidence of bony erosion and the proximity of disease to vital structures. All 6 patients in the EACC group had their diagnosis confirmed by computed tomography (CT), which demonstrated widening of the bony external auditory canal; 4 of these patients had critical erosion of bone adjacent to the facial nerve. Of the 6 patients with KO, only 2 had undergone CT, and neither exhibited any significant bony erosion or expansion; 1 of them developed osteomyelitis of the temporal bone and adjacent temporomandibular joint. Another patient manifested KO as part of a dermatophytid reaction. The essential component of treatment in all cases of EACC was microscopic debridement of the ear canal. We conclude that EACC may produce significant erosion of bone with exposure of vital structures, including the facial nerve. Because of the clinical similarity of EACC to KO, misdiagnosis is possible. Temporal bone imaging should be obtained prior to attempts at debridement of suspected EACC. Increased awareness of these uncommon conditions is warranted to prompt appropriate investigation and prevent iatrogenic complications such as facial nerve injury.

  8. Analysis of Facial Injuries Caused by Power Tools.

    PubMed

    Kim, Jiye; Choi, Jin-Hee; Hyun Kim, Oh; Won Kim, Sug

    2016-06-01

    The number of injuries caused by power tools is steadily increasing as more domestic woodwork is undertaken and more power tools are used recreationally. The injuries caused by the different power tools as a consequence of accidents are an issue, because they can lead to substantial costs for patients and the national insurance system. The increase in hand surgery as a consequence of the use of power tools and its economic impact, and the characteristics of the hand injuries caused by power saws have been described. In recent years, the authors have noticed that, in addition to hand injuries, facial injuries caused by power tools commonly present to the emergency room. This study aimed to review the data in relation to facial injuries caused by power saws that were gathered from patients who visited the trauma center at our hospital over the last 4 years, and to analyze the incidence and epidemiology of the facial injuries caused by power saws. The authors found that facial injuries caused by power tools have risen continually. Facial injuries caused by power tools are accidental, and they cause permanent facial disfigurements and functional disabilities. Accidents are almost inevitable in particular workplaces; however, most facial injuries could be avoided by providing sufficient operator training and by tool operators wearing suitable protective devices. The evaluation of the epidemiology and patterns of facial injuries caused by power tools in this study should provide the information required to reduce the number of accidental injuries.

  9. Three-dimensional gender differences in facial form of children in the North East of England.

    PubMed

    Bugaighis, Iman; Mattick, Clare R; Tiddeman, Bernard; Hobson, Ross

    2013-06-01

    The aim of the prospective cross-sectional morphometric study was to explore three dimensional (3D) facial shape and form (shape plus size) variation within and between 8- and 12-year-old Caucasian children; 39 males age-matched with 41 females. The 3D images were captured using a stereophotogrammeteric system, and facial form was recorded by digitizing 39 anthropometric landmarks for each scan. The x, y, z coordinates of each landmark were extracted and used to calculate linear and angular measurements. 3D landmark asymmetry was quantified using Generalized Procrustes Analysis (GPA) and an average face was constructed for each gender. The average faces were superimposed and differences were visualized and quantified. Shape variations were explored using GPA and PrincipalComponent Analysis. Analysis of covariance and Pearson correlation coefficients were used to explore gender differences and to determine any correlation between facial measurements and height or weight. Multivariate analysis was used to ascertain differences in facial measurements or 3D landmark asymmetry. There were no differences in height or weight between genders. There was a significant positive correlation between facial measurements and height and weight and statistically significant differences in linear facial width measurements between genders. These differences were related to the larger size of males rather than differences in shape. There were no age- or gender-linked significant differences in 3D landmark asymmetry. Shape analysis confirmed similarities between both males and females for facial shape and form in 8- to 12-year-old children. Any differences found were related to differences in facial size rather than shape.

  10. Population calcium imaging of spontaneous respiratory and novel motor activity in the facial nucleus and ventral brainstem in newborn mice

    PubMed Central

    Persson, Karin; Rekling, Jens C

    2011-01-01

    Abstract The brainstem contains rhythm and pattern forming circuits, which drive cranial and spinal motor pools to produce respiratory and other motor patterns. Here we used calcium imaging combined with nerve recordings in newborn mice to reveal spontaneous population activity in the ventral brainstem and in the facial nucleus. In Fluo-8 AM loaded brainstem–spinal cord preparations, respiratory activity on cervical nerves was synchronized with calcium signals at the ventrolateral brainstem surface. Individual ventrolateral neurons at the level of the parafacial respiratory group showed perfect or partial synchrony with respiratory nerve bursts. In brainstem–spinal cord preparations, cut at the level of the mid-facial nucleus, calcium signals were recorded in the dorsal, lateral and medial facial subnuclei during respiratory activity. Strong activity initiated in the dorsal subnucleus, followed by activity in lateral and medial subnuclei. Whole-cell recordings from facial motoneurons showed weak respiratory drives, and electrical field potential recordings confirmed respiratory drive to particularly the dorsal and lateral subnuclei. Putative facial premotoneurons showed respiratory-related calcium signals, and were predominantly located dorsomedial to the facial nucleus. A novel motor activity on facial, cervical and thoracic nerves was synchronized with calcium signals at the ventromedial brainstem extending from the level of the facial nucleus to the medulla–spinal cord border. Cervical dorsal root stimulation induced similar ventromedial activity. The medial facial subnucleus showed calcium signals synchronized with this novel motor activity on cervical nerves, and cervical dorsal root stimulation induced similar medial facial subnucleus activity. In conclusion, the dorsal and lateral facial subnuclei are strongly respiratory-modulated, and the brainstem contains a novel pattern forming circuit that drives the medial facial subnucleus and cervical motor

  11. Recognizing Facial Slivers.

    PubMed

    Gilad-Gutnick, Sharon; Harmatz, Elia Samuel; Tsourides, Kleovoulos; Yovel, Galit; Sinha, Pawan

    2018-07-01

    We report here an unexpectedly robust ability of healthy human individuals ( n = 40) to recognize extremely distorted needle-like facial images, challenging the well-entrenched notion that veridical spatial configuration is necessary for extracting facial identity. In face identification tasks of parametrically compressed internal and external features, we found that the sum of performances on each cue falls significantly short of performance on full faces, despite the equal visual information available from both measures (with full faces essentially being a superposition of internal and external features). We hypothesize that this large deficit stems from the use of positional information about how the internal features are positioned relative to the external features. To test this, we systematically changed the relations between internal and external features and found preferential encoding of vertical but not horizontal spatial relationships in facial representations ( n = 20). Finally, we employ magnetoencephalography imaging ( n = 20) to demonstrate a close mapping between the behavioral psychometric curve and the amplitude of the M250 face familiarity, but not M170 face-sensitive evoked response field component, providing evidence that the M250 can be modulated by faces that are perceptually identifiable, irrespective of extreme distortions to the face's veridical configuration. We theorize that the tolerance to compressive distortions has evolved from the need to recognize faces across varying viewpoints. Our findings help clarify the important, but poorly defined, concept of facial configuration and also enable an association between behavioral performance and previously reported neural correlates of face perception.

  12. High-resolution face verification using pore-scale facial features.

    PubMed

    Li, Dong; Zhou, Huiling; Lam, Kin-Man

    2015-08-01

    Face recognition methods, which usually represent face images using holistic or local facial features, rely heavily on alignment. Their performances also suffer a severe degradation under variations in expressions or poses, especially when there is one gallery per subject only. With the easy access to high-resolution (HR) face images nowadays, some HR face databases have recently been developed. However, few studies have tackled the use of HR information for face recognition or verification. In this paper, we propose a pose-invariant face-verification method, which is robust to alignment errors, using the HR information based on pore-scale facial features. A new keypoint descriptor, namely, pore-Principal Component Analysis (PCA)-Scale Invariant Feature Transform (PPCASIFT)-adapted from PCA-SIFT-is devised for the extraction of a compact set of distinctive pore-scale facial features. Having matched the pore-scale features of two-face regions, an effective robust-fitting scheme is proposed for the face-verification task. Experiments show that, with one frontal-view gallery only per subject, our proposed method outperforms a number of standard verification methods, and can achieve excellent accuracy even the faces are under large variations in expression and pose.

  13. Observers' response to facial disfigurement from head and neck cancer.

    PubMed

    Cho, Joowon; Fingeret, Michelle Cororve; Huang, Sheng-Cheng; Liu, Jun; Reece, Gregory P; Markey, Mia K

    2018-05-30

    Our long-term goal is to develop a normative feedback intervention to support head and neck cancer patients in forming realistic expectations about how other people in non-social group settings will respond to their appearance. This study aimed to evaluate the relationship between observer ratings of facial disfigurement and observer ratings of emotional response when viewing photographs of faces of head and neck cancer patients. Seventy-five (75) observers rated their emotional response to each of 144 facial photographs of head and neck cancer patients using the Self-Assessment-Manikin and rated severity of facial disfigurement on a 9-point scale. Body image investment of the observers was measured using the Appearance Schemas Inventory-Revised. A standardized multiple regression model was used to assess the relationship between observer ratings of facial disfigurement and observer ratings of emotional response, taking into consideration the age and sex of the patient depicted in the stimulus photograph, as well as the age, sex, and body image investment of the observer. Observers who had a strong emotional response to a patient's facial photograph tended to rate the patient's facial disfigurement as more severe (standardized regression coefficient β = 0.328, P < 0.001). Sex and age of the observer had more influence on the rating of facial disfigurement than did the patient's demographic characteristics. Observers more invested in their own body image tended to rate the facial disfigurement as more severe. This study lays the groundwork for a normative database of emotional response to facial disfigurement. Copyright © 2018 John Wiley & Sons, Ltd.

  14. Contributions of feature shapes and surface cues to the recognition of facial expressions.

    PubMed

    Sormaz, Mladen; Young, Andrew W; Andrews, Timothy J

    2016-10-01

    Theoretical accounts of face processing often emphasise feature shapes as the primary visual cue to the recognition of facial expressions. However, changes in facial expression also affect the surface properties of the face. In this study, we investigated whether this surface information can also be used in the recognition of facial expression. First, participants identified facial expressions (fear, anger, disgust, sadness, happiness) from images that were manipulated such that they varied mainly in shape or mainly in surface properties. We found that the categorization of facial expression is possible in either type of image, but that different expressions are relatively dependent on surface or shape properties. Next, we investigated the relative contributions of shape and surface information to the categorization of facial expressions. This employed a complementary method that involved combining the surface properties of one expression with the shape properties from a different expression. Our results showed that the categorization of facial expressions in these hybrid images was equally dependent on the surface and shape properties of the image. Together, these findings provide a direct demonstration that both feature shape and surface information make significant contributions to the recognition of facial expressions. Copyright © 2016 Elsevier Ltd. All rights reserved.

  15. Quantitative anatomical analysis of facial expression using a 3D motion capture system: Application to cosmetic surgery and facial recognition technology.

    PubMed

    Lee, Jae-Gi; Jung, Su-Jin; Lee, Hyung-Jin; Seo, Jung-Hyuk; Choi, You-Jin; Bae, Hyun-Sook; Park, Jong-Tae; Kim, Hee-Jin

    2015-09-01

    The topography of the facial muscles differs between males and females and among individuals of the same gender. To explain the unique expressions that people can make, it is important to define the shapes of the muscle, their associations with the skin, and their relative functions. Three-dimensional (3D) motion-capture analysis, often used to study facial expression, was used in this study to identify characteristic skin movements in males and females when they made six representative basic expressions. The movements of 44 reflective markers (RMs) positioned on anatomical landmarks were measured. Their mean displacement was large in males [ranging from 14.31 mm (fear) to 41.15 mm (anger)], and 3.35-4.76 mm smaller in females [ranging from 9.55 mm (fear) to 37.80 mm (anger)]. The percentages of RMs involved in the ten highest mean maximum displacement values in making at least one expression were 47.6% in males and 61.9% in females. The movements of the RMs were larger in males than females but were more limited. Expanding our understanding of facial expression requires morphological studies of facial muscles and studies of related complex functionality. Conducting these together with quantitative analyses, as in the present study, will yield data valuable for medicine, dentistry, and engineering, for example, for surgical operations on facial regions, software for predicting changes in facial features and expressions after corrective surgery, and the development of face-mimicking robots. © 2015 Wiley Periodicals, Inc.

  16. Gender differences in the motivational processing of facial beauty☆

    PubMed Central

    Levy, Boaz; Ariely, Dan; Mazar, Nina; Chi, Won; Lukas, Scott; Elman, Igor

    2013-01-01

    Gender may be involved in the motivational processing of facial beauty. This study applied a behavioral probe, known to activate brain motivational regions, to healthy heterosexual subjects. Matched samples of men and women were administered two tasks: (a) key pressing to change the viewing time of average or beautiful female or male facial images, and (b) rating the attractiveness of these images. Men expended more effort (via the key-press task) to extend the viewing time of the beautiful female faces. Women displayed similarly increased effort for beautiful male and female images, but the magnitude of this effort was substantially lower than that of men for beautiful females. Heterosexual facial attractiveness ratings were comparable in both groups. These findings demonstrate heterosexual specificity of facial motivational targets for men, but not for women. Moreover, heightened drive for the pursuit of heterosexual beauty in the face of regular valuational assessments, displayed by men, suggests a gender-specific incentive sensitization phenomenon. PMID:24282336

  17. Facial recognition using multisensor images based on localized kernel eigen spaces.

    PubMed

    Gundimada, Satyanadh; Asari, Vijayan K

    2009-06-01

    A feature selection technique along with an information fusion procedure for improving the recognition accuracy of a visual and thermal image-based facial recognition system is presented in this paper. A novel modular kernel eigenspaces approach is developed and implemented on the phase congruency feature maps extracted from the visual and thermal images individually. Smaller sub-regions from a predefined neighborhood within the phase congruency images of the training samples are merged to obtain a large set of features. These features are then projected into higher dimensional spaces using kernel methods. The proposed localized nonlinear feature selection procedure helps to overcome the bottlenecks of illumination variations, partial occlusions, expression variations and variations due to temperature changes that affect the visual and thermal face recognition techniques. AR and Equinox databases are used for experimentation and evaluation of the proposed technique. The proposed feature selection procedure has greatly improved the recognition accuracy for both the visual and thermal images when compared to conventional techniques. Also, a decision level fusion methodology is presented which along with the feature selection procedure has outperformed various other face recognition techniques in terms of recognition accuracy.

  18. Human Facial Shape and Size Heritability and Genetic Correlations.

    PubMed

    Cole, Joanne B; Manyama, Mange; Larson, Jacinda R; Liberton, Denise K; Ferrara, Tracey M; Riccardi, Sheri L; Li, Mao; Mio, Washington; Klein, Ophir D; Santorico, Stephanie A; Hallgrímsson, Benedikt; Spritz, Richard A

    2017-02-01

    The human face is an array of variable physical features that together make each of us unique and distinguishable. Striking familial facial similarities underscore a genetic component, but little is known of the genes that underlie facial shape differences. Numerous studies have estimated facial shape heritability using various methods. Here, we used advanced three-dimensional imaging technology and quantitative human genetics analysis to estimate narrow-sense heritability, heritability explained by common genetic variation, and pairwise genetic correlations of 38 measures of facial shape and size in normal African Bantu children from Tanzania. Specifically, we fit a linear mixed model of genetic relatedness between close and distant relatives to jointly estimate variance components that correspond to heritability explained by genome-wide common genetic variation and variance explained by uncaptured genetic variation, the sum representing total narrow-sense heritability. Our significant estimates for narrow-sense heritability of specific facial traits range from 28 to 67%, with horizontal measures being slightly more heritable than vertical or depth measures. Furthermore, for over half of facial traits, >90% of narrow-sense heritability can be explained by common genetic variation. We also find high absolute genetic correlation between most traits, indicating large overlap in underlying genetic loci. Not surprisingly, traits measured in the same physical orientation (i.e., both horizontal or both vertical) have high positive genetic correlations, whereas traits in opposite orientations have high negative correlations. The complex genetic architecture of facial shape informs our understanding of the intricate relationships among different facial features as well as overall facial development. Copyright © 2017 by the Genetics Society of America.

  19. Facial Emotions Recognition using Gabor Transform and Facial Animation Parameters with Neural Networks

    NASA Astrophysics Data System (ADS)

    Harit, Aditya; Joshi, J. C., Col; Gupta, K. K.

    2018-03-01

    The paper proposed an automatic facial emotion recognition algorithm which comprises of two main components: feature extraction and expression recognition. The algorithm uses a Gabor filter bank on fiducial points to find the facial expression features. The resulting magnitudes of Gabor transforms, along with 14 chosen FAPs (Facial Animation Parameters), compose the feature space. There are two stages: the training phase and the recognition phase. Firstly, for the present 6 different emotions, the system classifies all training expressions in 6 different classes (one for each emotion) in the training stage. In the recognition phase, it recognizes the emotion by applying the Gabor bank to a face image, then finds the fiducial points, and then feeds it to the trained neural architecture.

  20. A study of patient facial expressivity in relation to orthodontic/surgical treatment.

    PubMed

    Nafziger, Y J

    1994-09-01

    A dynamic analysis of the faces of patients seeking an aesthetic restoration of facial aberrations with orthognathic treatment requires (besides the routine static study, such as records, study models, photographs, and cephalometric tracings) the study of their facial expressions. To determine a classification method for the units of expressive facial behavior, the mobility of the face is studied with the aid of the facial action coding system (FACS) created by Ekman and Friesen. With video recordings of faces and photographic images taken from the video recordings, the authors have modified a technique of facial analysis structured on the visual observation of the anatomic basis of movement. The technique, itself, is based on the defining of individual facial expressions and then codifying such expressions through the use of minimal, anatomic action units. These action units actually combine to form facial expressions. With the help of FACS, the facial expressions of 18 patients before and after orthognathic surgery, and six control subjects without dentofacial deformation have been studied. I was able to register 6278 facial expressions and then further define 18,844 action units, from the 6278 facial expressions. A classification of the facial expressions made by subject groups and repeated in quantified time frames has allowed establishment of "rules" or "norms" relating to expression, thus further enabling the making of comparisons of facial expressiveness between patients and control subjects. This study indicates that the facial expressions of the patients were more similar to the facial expressions of the controls after orthognathic surgery. It was possible to distinguish changes in facial expressivity in patients after dentofacial surgery, the type and degree of change depended on the facial structure before surgery. Changes noted tended toward a functioning that is identical to that of subjects who do not suffer from dysmorphosis and toward greater lip

  1. Evaluation of facial attractiveness from end-of-treatment facial photographs.

    PubMed

    Shafiee, Roxanne; Korn, Edward L; Pearson, Helmer; Boyd, Robert L; Baumrind, Sheldon

    2008-04-01

    Orthodontists typically make judgments of facial attractiveness by examining groupings of profile, full-face, and smiling photographs considered together as a "triplet." The primary objective of this study was to determine the relative contributions of the 3 photographs-each considered separately-to the overall judgment a clinician forms by examining the combination of the 3. End-of-treatment triplet orthodontic photographs of 45 randomly selected orthodontic patients were duplicated. Copies of the profile, full-face, and smiling images were generated, and the images were separated and then pooled by image type for all subjects. Ten judges ranked the 45 photographs of each image type for facial attractiveness in groups of 9 to 12, from "most attractive" to "least attractive." Each judge also ranked the triplet groupings for the same 45 subjects. The mean attractiveness rankings for each type of photograph were then correlated with the mean rankings of each other and the triplets. The rankings of the 3 image types correlated highly with each other and the rankings of the triplets (P <.0001). The rankings of the smiling photographs were most predictive of the rankings of the triplets (r = 0.93); those of the profile photographs were the least predictive (r = 0.76). The difference between these correlations was highly statistically significant (P = .0003). It was also possible to test the extent to which the judges' rankings were influenced by sex, original Angle classification, and extraction status of each patient. No statistically significant preferences were found for sex or Angle classification, and only 1 marginally significant preference was found for extraction pattern. Clinician judges demonstrated a high level of agreement in ranking the facial attractiveness of profile, full-face, and smiling photographs of a group of orthodontically treated patients whose actual differences in physical dimensions were relatively small. The judges' rankings of the smiling

  2. Noncontact measurement of heart rate using facial video illuminated under natural light and signal weighted analysis.

    PubMed

    Yan, Yonggang; Ma, Xiang; Yao, Lifeng; Ouyang, Jianfei

    2015-01-01

    Non-contact and remote measurements of vital physical signals are important for reliable and comfortable physiological self-assessment. We presented a novel optical imaging-based method to measure the vital physical signals. Using a digital camera and ambient light, the cardiovascular pulse waves were extracted better from human color facial videos correctly. And the vital physiological parameters like heart rate were measured using a proposed signal-weighted analysis method. The measured HRs consistent with those measured simultaneously with reference technologies (r=0.94, p<0.001 for HR). The results show that the imaging-based method is suitable for measuring the physiological parameters, and provide a reliable and comfortable measurement mode. The study lays a physical foundation for measuring multi-physiological parameters of human noninvasively.

  3. The review and results of different methods for facial recognition

    NASA Astrophysics Data System (ADS)

    Le, Yifan

    2017-09-01

    In recent years, facial recognition draws much attention due to its wide potential applications. As a unique technology in Biometric Identification, facial recognition represents a significant improvement since it could be operated without cooperation of people under detection. Hence, facial recognition will be taken into defense system, medical detection, human behavior understanding, etc. Several theories and methods have been established to make progress in facial recognition: (1) A novel two-stage facial landmark localization method is proposed which has more accurate facial localization effect under specific database; (2) A statistical face frontalization method is proposed which outperforms state-of-the-art methods for face landmark localization; (3) It proposes a general facial landmark detection algorithm to handle images with severe occlusion and images with large head poses; (4) There are three methods proposed on Face Alignment including shape augmented regression method, pose-indexed based multi-view method and a learning based method via regressing local binary features. The aim of this paper is to analyze previous work of different aspects in facial recognition, focusing on concrete method and performance under various databases. In addition, some improvement measures and suggestions in potential applications will be put forward.

  4. Functional Alterations of Postcentral Gyrus Modulated by Angry Facial Expressions during Intraoral Tactile Stimuli in Patients with Burning Mouth Syndrome: A Functional Magnetic Resonance Imaging Study

    PubMed Central

    Yoshino, Atsuo; Okamoto, Yasumasa; Doi, Mitsuru; Okada, Go; Takamura, Masahiro; Ichikawa, Naho; Yamawaki, Shigeto

    2017-01-01

    Previous findings suggest that negative emotions could influence abnormal sensory perception in burning mouth syndrome (BMS). However, few studies have investigated the underlying neural mechanisms associated with BMS. We examined activation of brain regions in response to intraoral tactile stimuli when modulated by angry facial expressions. We performed functional magnetic resonance imaging on a group of 27 BMS patients and 21 age-matched healthy controls. Tactile stimuli were presented during different emotional contexts, which were induced via the continuous presentation of angry or neutral pictures of human faces. BMS patients exhibited higher tactile ratings and greater activation in the postcentral gyrus during the presentation of tactile stimuli involving angry faces relative to controls. Significant positive correlations between changes in brain activation elicited by angry facial images in the postcentral gyrus and changes in tactile rating scores by angry facial images were found for both groups. For BMS patients, there was a significant positive correlation between changes in tactile-related activation of the postcentral gyrus elicited by angry facial expressions and pain intensity in daily life. Findings suggest that neural responses in the postcentral gyrus are more strongly affected by angry facial expressions in BMS patients, which may reflect one possible mechanism underlying impaired somatosensory system function in this disorder. PMID:29163243

  5. Multivariate Pattern Classification of Facial Expressions Based on Large-Scale Functional Connectivity.

    PubMed

    Liang, Yin; Liu, Baolin; Li, Xianglin; Wang, Peiyuan

    2018-01-01

    It is an important question how human beings achieve efficient recognition of others' facial expressions in cognitive neuroscience, and it has been identified that specific cortical regions show preferential activation to facial expressions in previous studies. However, the potential contributions of the connectivity patterns in the processing of facial expressions remained unclear. The present functional magnetic resonance imaging (fMRI) study explored whether facial expressions could be decoded from the functional connectivity (FC) patterns using multivariate pattern analysis combined with machine learning algorithms (fcMVPA). We employed a block design experiment and collected neural activities while participants viewed facial expressions of six basic emotions (anger, disgust, fear, joy, sadness, and surprise). Both static and dynamic expression stimuli were included in our study. A behavioral experiment after scanning confirmed the validity of the facial stimuli presented during the fMRI experiment with classification accuracies and emotional intensities. We obtained whole-brain FC patterns for each facial expression and found that both static and dynamic facial expressions could be successfully decoded from the FC patterns. Moreover, we identified the expression-discriminative networks for the static and dynamic facial expressions, which span beyond the conventional face-selective areas. Overall, these results reveal that large-scale FC patterns may also contain rich expression information to accurately decode facial expressions, suggesting a novel mechanism, which includes general interactions between distributed brain regions, and that contributes to the human facial expression recognition.

  6. Multivariate Pattern Classification of Facial Expressions Based on Large-Scale Functional Connectivity

    PubMed Central

    Liang, Yin; Liu, Baolin; Li, Xianglin; Wang, Peiyuan

    2018-01-01

    It is an important question how human beings achieve efficient recognition of others’ facial expressions in cognitive neuroscience, and it has been identified that specific cortical regions show preferential activation to facial expressions in previous studies. However, the potential contributions of the connectivity patterns in the processing of facial expressions remained unclear. The present functional magnetic resonance imaging (fMRI) study explored whether facial expressions could be decoded from the functional connectivity (FC) patterns using multivariate pattern analysis combined with machine learning algorithms (fcMVPA). We employed a block design experiment and collected neural activities while participants viewed facial expressions of six basic emotions (anger, disgust, fear, joy, sadness, and surprise). Both static and dynamic expression stimuli were included in our study. A behavioral experiment after scanning confirmed the validity of the facial stimuli presented during the fMRI experiment with classification accuracies and emotional intensities. We obtained whole-brain FC patterns for each facial expression and found that both static and dynamic facial expressions could be successfully decoded from the FC patterns. Moreover, we identified the expression-discriminative networks for the static and dynamic facial expressions, which span beyond the conventional face-selective areas. Overall, these results reveal that large-scale FC patterns may also contain rich expression information to accurately decode facial expressions, suggesting a novel mechanism, which includes general interactions between distributed brain regions, and that contributes to the human facial expression recognition. PMID:29615882

  7. Facial Expression Presentation for Real-Time Internet Communication

    NASA Astrophysics Data System (ADS)

    Dugarry, Alexandre; Berrada, Aida; Fu, Shan

    2003-01-01

    Text, voice and video images are the most common forms of media content for instant communication on the Internet. Studies have shown that facial expressions convey much richer information than text and voice during a face-to-face conversation. The currently available real time means of communication (instant text messages, chat programs and videoconferencing), however, have major drawbacks in terms of exchanging facial expression. The first two means do not involve the image transmission, whilst video conferencing requires a large bandwidth that is not always available, and the transmitted image sequence is neither smooth nor without delay. The objective of the work presented here is to develop a technique that overcomes these limitations, by extracting the facial expression of speakers and to realise real-time communication. In order to get the facial expressions, the main characteristics of the image are emphasized. Interpolation is performed on edge points previously detected to create geometric shapes such as arcs, lines, etc. The regional dominant colours of the pictures are also extracted and the combined results are subsequently converted into Scalable Vector Graphics (SVG) format. The application based on the proposed technique aims at being used simultaneously with chat programs and being able to run on any platform.

  8. Enhanced facial texture illumination normalization for face recognition.

    PubMed

    Luo, Yong; Guan, Ye-Peng

    2015-08-01

    An uncontrolled lighting condition is one of the most critical challenges for practical face recognition applications. An enhanced facial texture illumination normalization method is put forward to resolve this challenge. An adaptive relighting algorithm is developed to improve the brightness uniformity of face images. Facial texture is extracted by using an illumination estimation difference algorithm. An anisotropic histogram-stretching algorithm is proposed to minimize the intraclass distance of facial skin and maximize the dynamic range of facial texture distribution. Compared with the existing methods, the proposed method can more effectively eliminate the redundant information of facial skin and illumination. Extensive experiments show that the proposed method has superior performance in normalizing illumination variation and enhancing facial texture features for illumination-insensitive face recognition.

  9. Three-dimensional assessment of facial asymmetry: A systematic review.

    PubMed

    Akhil, Gopi; Senthil Kumar, Kullampalayam Palanisamy; Raja, Subramani; Janardhanan, Kumaresan

    2015-08-01

    For patients with facial asymmetry, complete and precise diagnosis, and surgical treatments to correct the underlying cause of the asymmetry are significant. Conventional diagnostic radiographs (submento-vertex projections, posteroanterior radiography) have limitations in asymmetry diagnosis due to two-dimensional assessments of three-dimensional (3D) images. The advent of 3D images has greatly reduced the magnification and projection errors that are common in conventional radiographs making it as a precise diagnostic aid for assessment of facial asymmetry. Thus, this article attempts to review the newly introduced 3D tools in the diagnosis of more complex facial asymmetries.

  10. Ethnicity identification from face images

    NASA Astrophysics Data System (ADS)

    Lu, Xiaoguang; Jain, Anil K.

    2004-08-01

    Human facial images provide the demographic information, such as ethnicity and gender. Conversely, ethnicity and gender also play an important role in face-related applications. Image-based ethnicity identification problem is addressed in a machine learning framework. The Linear Discriminant Analysis (LDA) based scheme is presented for the two-class (Asian vs. non-Asian) ethnicity classification task. Multiscale analysis is applied to the input facial images. An ensemble framework, which integrates the LDA analysis for the input face images at different scales, is proposed to further improve the classification performance. The product rule is used as the combination strategy in the ensemble. Experimental results based on a face database containing 263 subjects (2,630 face images, with equal balance between the two classes) are promising, indicating that LDA and the proposed ensemble framework have sufficient discriminative power for the ethnicity classification problem. The normalized ethnicity classification scores can be helpful in the facial identity recognition. Useful as a "soft" biometric, face matching scores can be updated based on the output of ethnicity classification module. In other words, ethnicity classifier does not have to be perfect to be useful in practice.

  11. Effect of frontal facial type and sex on preferred chin projection.

    PubMed

    Choi, Jin-Young; Kim, Taeyun; Kim, Hyung-Mo; Lee, Sang-Hoon; Cho, Il-Sik; Baek, Seung-Hak

    2017-03-01

    To investigate the effects of frontal facial type (FFT) and sex on preferred chin projection (CP) in three-dimensional (3D) facial images. Six 3D facial images were acquired using a 3D facial scanner (euryprosopic [Eury-FFT], mesoprosopic [Meso-FFT], and leptoprosopic [Lepto-FFT] for each sex). After normal CP in each 3D facial image was set to 10° of the facial profile angle (glabella-subnasale-pogonion), CPs were morphed by gradations of 2° from normal (moderately protrusive [6°], slightly protrusive [8°], slightly retrusive [12°], and moderately retrusive [14°]). Seventy-five dental students (48 men and 27 women) were asked to rate the CPs (6°, 8°, 10°, 12°, and 14°) from the most to least preferred in each 3D image. Statistical analyses included the Kolmogorov-Smirnov test, Kruskal-Wallis test, and Bonferroni correction. No significant difference was observed in the distribution of preferred CP in the same FFT between male and female evaluators. In Meso-FFT, the normal CP was the most preferred without any sex difference. However, in Eury-FFT, the slightly protrusive CP was favored in male 3D images, but the normal CP was preferred in female 3D images. In Lepto-FFT, the normal CP was favored in male 3D images, whereas the slightly retrusive CP was favored in female 3D images. The mean preferred CP angle differed significantly according to FFT (Eury-FFT: male, 8.7°, female, 9.9°; Meso-FFT: male, 9.8°, female, 10.7°; Lepto-FFT: male, 10.8°, female, 11.4°; p < 0.001). Our findings might serve as guidelines for setting the preferred CP according to FFT and sex.

  12. Research on facial expression simulation based on depth image

    NASA Astrophysics Data System (ADS)

    Ding, Sha-sha; Duan, Jin; Zhao, Yi-wu; Xiao, Bo; Wang, Hao

    2017-11-01

    Nowadays, face expression simulation is widely used in film and television special effects, human-computer interaction and many other fields. Facial expression is captured by the device of Kinect camera .The method of AAM algorithm based on statistical information is employed to detect and track faces. The 2D regression algorithm is applied to align the feature points. Among them, facial feature points are detected automatically and 3D cartoon model feature points are signed artificially. The aligned feature points are mapped by keyframe techniques. In order to improve the animation effect, Non-feature points are interpolated based on empirical models. Under the constraint of Bézier curves we finish the mapping and interpolation. Thus the feature points on the cartoon face model can be driven if the facial expression varies. In this way the purpose of cartoon face expression simulation in real-time is came ture. The experiment result shows that the method proposed in this text can accurately simulate the facial expression. Finally, our method is compared with the previous method. Actual data prove that the implementation efficiency is greatly improved by our method.

  13. Dissociation between recognition and detection advantage for facial expressions: a meta-analysis.

    PubMed

    Nummenmaa, Lauri; Calvo, Manuel G

    2015-04-01

    Happy facial expressions are recognized faster and more accurately than other expressions in categorization tasks, whereas detection in visual search tasks is widely believed to be faster for angry than happy faces. We used meta-analytic techniques for resolving this categorization versus detection advantage discrepancy for positive versus negative facial expressions. Effect sizes were computed on the basis of the r statistic for a total of 34 recognition studies with 3,561 participants and 37 visual search studies with 2,455 participants, yielding a total of 41 effect sizes for recognition accuracy, 25 for recognition speed, and 125 for visual search speed. Random effects meta-analysis was conducted to estimate effect sizes at population level. For recognition tasks, an advantage in recognition accuracy and speed for happy expressions was found for all stimulus types. In contrast, for visual search tasks, moderator analysis revealed that a happy face detection advantage was restricted to photographic faces, whereas a clear angry face advantage was found for schematic and "smiley" faces. Robust detection advantage for nonhappy faces was observed even when stimulus emotionality was distorted by inversion or rearrangement of the facial features, suggesting that visual features primarily drive the search. We conclude that the recognition advantage for happy faces is a genuine phenomenon related to processing of facial expression category and affective valence. In contrast, detection advantages toward either happy (photographic stimuli) or nonhappy (schematic) faces is contingent on visual stimulus features rather than facial expression, and may not involve categorical or affective processing. (c) 2015 APA, all rights reserved).

  14. Non-invasive health status detection system using Gabor filters based on facial block texture features.

    PubMed

    Shu, Ting; Zhang, Bob

    2015-04-01

    Blood tests allow doctors to check for certain diseases and conditions. However, using a syringe to extract the blood can be deemed invasive, slightly painful, and its analysis time consuming. In this paper, we propose a new non-invasive system to detect the health status (Healthy or Diseased) of an individual based on facial block texture features extracted using the Gabor filter. Our system first uses a non-invasive capture device to collect facial images. Next, four facial blocks are located on these images to represent them. Afterwards, each facial block is convolved with a Gabor filter bank to calculate its texture value. Classification is finally performed using K-Nearest Neighbor and Support Vector Machines via a Library for Support Vector Machines (with four kernel functions). The system was tested on a dataset consisting of 100 Healthy and 100 Diseased (with 13 forms of illnesses) samples. Experimental results show that the proposed system can detect the health status with an accuracy of 93 %, a sensitivity of 94 %, a specificity of 92 %, using a combination of the Gabor filters and facial blocks.

  15. Effects of different rapid maxillary expansion appliances on facial soft tissues using three-dimensional imaging.

    PubMed

    Altındiş, Sedat; Toy, Ebubekir; Başçiftçi, Faruk Ayhan

    2016-07-01

    To determine three-dimensional (3D) effects of three different rapid maxillary expansion (RME) appliances on facial soft tissues. Forty-two children (18 boys, 24 girls) who required RME treatment were included in this study. Patients were randomly divided into three equal groups: banded RME, acrylic splint RME, and modified acrylic splint RME. For each patient, 3D images were obtained before treatment (T1) and at the end of the 3-month retention (T2) with the 3dMD system. When three RME appliances were compared in terms of the effects on the facial soft tissues, there were no significant differences among them. The mouth and nasal width showed a significant increase in all groups. Although the effect of the acrylic splint RME appliances on total face height was less than that of the banded RME, there was no significant difference between the appliances. The effect of the modified acrylic splint appliance on the upper lip was significant according to the volumetric measurements (P < .01). There were no significant differences among three RME appliances on the facial soft tissues. The modified acrylic splint RME produced a more protrusive effect on the upper lip.

  16. Factors Influencing Perception of Facial Attractiveness: Gender and Dental Education.

    PubMed

    Jung, Ga-Hee; Jung, Seunggon; Park, Hong-Ju; Oh, Hee-Kyun; Kook, Min-Suk

    2018-03-01

    This study was conducted to investigate the gender- and dental education-specific differences in perception of facial attractiveness for varying ratio of lower face contour. Two hundred eleven students (110 male respondents and 110 female respondents; aged between 20-38 years old) were requested to rate facial figures with alterations to the bigonial width and the vertical length of the lower face. We produced a standard figure which is based on the "golden ratio" and 4 additional series of figures with either horizontal or vertical alterations to the contour of lower face. The preference for each figure was evaluated using a Visual Analog Scale. The Kruskal Wallis test was used for differences in the preferences for each figure and the Mann-Whitney U test was used to evaluate gender-specific differences and differences by dental education. In general, the highest preference score was indicated for the standard figure, whereas facial figure with large bigonial width and chin length had the lowest score.Male respondents showed significantly higher preference score for facial contour that had a 0.1 proportional increase in the facial height-bigonial width ratio over that of the standard figure.For horizontal alterations to the facial profiles, there were no significant differences in the preferences by the level of dental education. For vertically altered images, the average Visual Analog Scale was significantly lower among the dentally-educated for facial image that had a proportional 0.22 and 0.42 increase in the ratio between the vertical length of the chin and the lip. Generally, the standard image based on the golden ratio was the most. Slender face was appealed more to males than to females, and facial image with an increased lower facial height were perceived to be much less attractive to the dentally-educated respondents, which suggests that the dental education might have some influence in sensitivity to vertical changes in lower face.

  17. 5-ALA induced fluorescent image analysis of actinic keratosis

    NASA Astrophysics Data System (ADS)

    Cho, Yong-Jin; Bae, Youngwoo; Choi, Eung-Ho; Jung, Byungjo

    2010-02-01

    In this study, we quantitatively analyzed 5-ALA induced fluorescent images of actinic keratosis using digital fluorescent color and hyperspectral imaging modalities. UV-A was utilized to induce fluorescent images and actinic keratosis (AK) lesions were demarcated from surrounding the normal region with different methods. Eight subjects with AK lesion were participated in this study. In the hyperspectral imaging modality, spectral analysis method was utilized for hyperspectral cube image and AK lesions were demarcated from the normal region. Before image acquisition, we designated biopsy position for histopathology of AK lesion and surrounding normal region. Erythema index (E.I.) values on both regions were calculated from the spectral cube data. Image analysis of subjects resulted in two different groups: the first group with the higher fluorescence signal and E.I. on AK lesion than the normal region; the second group with lower fluorescence signal and without big difference in E.I. between two regions. In fluorescent color image analysis of facial AK, E.I. images were calculated on both normal and AK lesions and compared with the results of hyperspectral imaging modality. The results might indicate that the different intensity of fluorescence and E.I. among the subjects with AK might be interpreted as different phases of morphological and metabolic changes of AK lesions.

  18. Quantitative Anthropometric Measures of Facial Appearance of Healthy Hispanic/Latino White Children: Establishing Reference Data for Care of Cleft Lip With or Without Cleft Palate

    NASA Astrophysics Data System (ADS)

    Lee, Juhun; Ku, Brian; Combs, Patrick D.; Da Silveira, Adriana. C.; Markey, Mia K.

    2017-06-01

    Cleft lip with or without cleft palate (CL ± P) is one of the most common congenital facial deformities worldwide. To minimize negative social consequences of CL ± P, reconstructive surgery is conducted to modify the face to a more normal appearance. Each race/ethnic group requires its own facial norm data, yet there are no existing facial norm data for Hispanic/Latino White children. The objective of this paper is to identify measures of facial appearance relevant for planning reconstructive surgery for CL ± P of Hispanic/Latino White children. Quantitative analysis was conducted on 3D facial images of 82 (41 girls, 41 boys) healthy Hispanic/Latino White children whose ages ranged from 7 to 12 years. Twenty-eight facial anthropometric features related to CL ± P (mainly in the nasal and mouth area) were measured from 3D facial images. In addition, facial aesthetic ratings were obtained from 16 non-clinical observers for the same 3D facial images using a 7-point Likert scale. Pearson correlation analysis was conducted to find features that were correlated with the panel ratings of observers. Boys with a longer face and nose, or thicker upper and lower lips are considered more attractive than others while girls with a less curved middle face contour are considered more attractive than others. Associated facial landmarks for these features are primary focus areas for reconstructive surgery for CL ± P. This study identified anthropometric measures of facial features of Hispanic/Latino White children that are pertinent to CL ± P and which correlate with the panel attractiveness ratings.

  19. Photometric facial analysis of the Igbo Nigerian adult male

    PubMed Central

    Ukoha, Ukoha Ukoha; Udemezue, Onochie Okwudili; Oranusi, Chidi Kingsley; Asomugha, Azuoma Lasbrey; Dimkpa, Uchechukwu; Nzeukwu, Lynda Chinenye

    2012-01-01

    Background: A carefully performed facial analysis can serve as a strong foundation for successful facial reconstructive and plastic surgeries, rhinoplasty or orthodontics. Aim: The purpose of this study is to determine the facial features and qualities of the Igbo Nigerian adult male using photometry. Materials and Methods: One hundred and twenty subjects aged between 18 and 28 years were studied at the Anambra State University, Uli, Nigeria. The frontal and right lateral view photographs of their faces were taken and traced out on tracing papers. On these, two vertical distances, nasion to subnasal and subnasale to menton, and four angles, nasofrontal (NF), nasofacial, nasomental (NM) and mentocervical, were measured. Results: The result showed that the Igbo Nigerian adult male had a middle face that was shorter than the lower one (41.76% vs.58.24%), a moderate glabella (NF=133.97°), a projected nose (NM=38.68°) and a less prominent chin (NM=125.87°). Conclusion: This study is very important in medical practice as it can be used to compare the pre- and post-operative results of plastic surgery and other related surgeries of the face. PMID:23661886

  20. Symmetrical and Asymmetrical Interactions between Facial Expressions and Gender Information in Face Perception.

    PubMed

    Liu, Chengwei; Liu, Ying; Iqbal, Zahida; Li, Wenhui; Lv, Bo; Jiang, Zhongqing

    2017-01-01

    To investigate the interaction between facial expressions and facial gender information during face perception, the present study matched the intensities of the two types of information in face images and then adopted the orthogonal condition of the Garner Paradigm to present the images to participants who were required to judge the gender and expression of the faces; the gender and expression presentations were varied orthogonally. Gender and expression processing displayed a mutual interaction. On the one hand, the judgment of angry expressions occurred faster when presented with male facial images; on the other hand, the classification of the female gender occurred faster when presented with a happy facial expression than when presented with an angry facial expression. According to the evoked-related potential results, the expression classification was influenced by gender during the face structural processing stage (as indexed by N170), which indicates the promotion or interference of facial gender with the coding of facial expression features. However, gender processing was affected by facial expressions in more stages, including the early (P1) and late (LPC) stages of perceptual processing, reflecting that emotional expression influences gender processing mainly by directing attention.

  1. Extraction and representation of common feature from uncertain facial expressions with cloud model.

    PubMed

    Wang, Shuliang; Chi, Hehua; Yuan, Hanning; Geng, Jing

    2017-12-01

    Human facial expressions are key ingredient to convert an individual's innate emotion in communication. However, the variation of facial expressions affects the reliable identification of human emotions. In this paper, we present a cloud model to extract facial features for representing human emotion. First, the uncertainties in facial expression are analyzed in the context of cloud model. The feature extraction and representation algorithm is established under cloud generators. With forward cloud generator, facial expression images can be re-generated as many as we like for visually representing the extracted three features, and each feature shows different roles. The effectiveness of the computing model is tested on Japanese Female Facial Expression database. Three common features are extracted from seven facial expression images. Finally, the paper is concluded and remarked.

  2. Acromegaly determination using discriminant analysis of the three-dimensional facial classification in Taiwanese.

    PubMed

    Wang, Ming-Hsu; Lin, Jen-Der; Chang, Chen-Nen; Chiou, Wen-Ko

    2017-08-01

    The aim of this study was to assess the size, angles and positional characteristics of facial anthropometry between "acromegalic" patients and control subjects. We also identify possible facial soft tissue measurements for generating discriminant functions toward acromegaly determination in males and females for acromegaly early self-awareness. This is a cross-sectional study. Subjects participating in this study included 70 patients diagnosed with acromegaly (35 females and 35 males) and 140 gender-matched control individuals. Three-dimensional facial images were collected via a camera system. Thirteen landmarks were selected. Eleven measurements from the three categories were selected and applied, including five frontal widths, three lateral depths and three lateral angular measurements. Descriptive analyses were conducted using means and standard deviations for each measurement. Univariate and multivariate discriminant function analyses were applied in order to calculate the accuracy of acromegaly detection. Patients with acromegaly exhibit soft-tissue facial enlargement and hypertrophy. Frontal widths as well as lateral depth and angle of facial changes were evident. The average accuracies of all functions for female patient detection ranged from 80.0-91.40%. The average accuracies of all functions for male patient detection were from 81.0-94.30%. The greatest anomaly observed was evidenced in the lateral angles, with greater enlargement of "nasofrontal" angles for females and greater "mentolabial" angles for males. Additionally, shapes of the lateral angles showed changes. The majority of the facial measurements proved dynamic for acromegaly patients; however, it is problematic to detect the disease with progressive body anthropometric changes. The discriminant functions of detection developed in this study could help patients, their families, medical practitioners and others to identify and track progressive facial change patterns before the possible patients

  3. Eigen-disfigurement model for simulating plausible facial disfigurement after reconstructive surgery.

    PubMed

    Lee, Juhun; Fingeret, Michelle C; Bovik, Alan C; Reece, Gregory P; Skoracki, Roman J; Hanasono, Matthew M; Markey, Mia K

    2015-03-27

    Patients with facial cancers can experience disfigurement as they may undergo considerable appearance changes from their illness and its treatment. Individuals with difficulties adjusting to facial cancer are concerned about how others perceive and evaluate their appearance. Therefore, it is important to understand how humans perceive disfigured faces. We describe a new strategy that allows simulation of surgically plausible facial disfigurement on a novel face for elucidating the human perception on facial disfigurement. Longitudinal 3D facial images of patients (N = 17) with facial disfigurement due to cancer treatment were replicated using a facial mannequin model, by applying Thin-Plate Spline (TPS) warping and linear interpolation on the facial mannequin model in polar coordinates. Principal Component Analysis (PCA) was used to capture longitudinal structural and textural variations found within each patient with facial disfigurement arising from the treatment. We treated such variations as disfigurement. Each disfigurement was smoothly stitched on a healthy face by seeking a Poisson solution to guided interpolation using the gradient of the learned disfigurement as the guidance field vector. The modeling technique was quantitatively evaluated. In addition, panel ratings of experienced medical professionals on the plausibility of simulation were used to evaluate the proposed disfigurement model. The algorithm reproduced the given face effectively using a facial mannequin model with less than 4.4 mm maximum error for the validation fiducial points that were not used for the processing. Panel ratings of experienced medical professionals on the plausibility of simulation showed that the disfigurement model (especially for peripheral disfigurement) yielded predictions comparable to the real disfigurements. The modeling technique of this study is able to capture facial disfigurements and its simulation represents plausible outcomes of reconstructive surgery

  4. Multi-layer sparse representation for weighted LBP-patches based facial expression recognition.

    PubMed

    Jia, Qi; Gao, Xinkai; Guo, He; Luo, Zhongxuan; Wang, Yi

    2015-03-19

    In this paper, a novel facial expression recognition method based on sparse representation is proposed. Most contemporary facial expression recognition systems suffer from limited ability to handle image nuisances such as low resolution and noise. Especially for low intensity expression, most of the existing training methods have quite low recognition rates. Motivated by sparse representation, the problem can be solved by finding sparse coefficients of the test image by the whole training set. Deriving an effective facial representation from original face images is a vital step for successful facial expression recognition. We evaluate facial representation based on weighted local binary patterns, and Fisher separation criterion is used to calculate the weighs of patches. A multi-layer sparse representation framework is proposed for multi-intensity facial expression recognition, especially for low-intensity expressions and noisy expressions in reality, which is a critical problem but seldom addressed in the existing works. To this end, several experiments based on low-resolution and multi-intensity expressions are carried out. Promising results on publicly available databases demonstrate the potential of the proposed approach.

  5. Assessing photoplethysmographic imaging performance beyond facial perfusion analysis

    NASA Astrophysics Data System (ADS)

    Amelard, Robert; Hughson, Richard L.; Greaves, Danielle K.; Clausi, David A.; Wong, Alexander

    2017-02-01

    Photoplethysmographic imaging (PPGI) systems are relatively new non-contact biophotonic diffuse reflectance systems able to assess arterial pulsations through transient changes in light-tissue interaction. Many PPGI studies have focused on extracting heart rate from the face or hand. Though PPGI systems can be used for widefield imaging of any anatomical area, whole-body investigations are lacking. Here, using a novel PPGI system, coded hemodynamic imaging (CHI), we explored and analyzed the pulsatility at major arterial locations across the whole body, including the neck (carotid artery), arm/wrist (brachial, radial and ulnar arteries), and leg/feet (popliteal and tibial arteries). CHI was positioned 1.5 m from the participant, and diffuse reactance from a broadband tungsten-halogen illumination was filtered using 850{1000 nm bandpass filter for deep tissue penetration. Images were acquired over a highly varying 24-participant sample (11/13 female/male, age 28.7+/-12.4 years, BMI 25.5+/-5.2 kg/m2), and a preliminary case study was performed. B-mode ultrasound images were acquired to validate observations through planar arterial characteristics.

  6. Is moral beauty different from facial beauty? Evidence from an fMRI study

    PubMed Central

    Wang, Tingting; Mo, Ce; Tan, Li Hai; Cant, Jonathan S.; Zhong, Luojin; Cupchik, Gerald

    2015-01-01

    Is moral beauty different from facial beauty? Two functional magnetic resonance imaging experiments were performed to answer this question. Experiment 1 investigated the network of moral aesthetic judgments and facial aesthetic judgments. Participants performed aesthetic judgments and gender judgments on both faces and scenes containing moral acts. The conjunction analysis of the contrasts ‘facial aesthetic judgment > facial gender judgment’ and ‘scene moral aesthetic judgment > scene gender judgment’ identified the common involvement of the orbitofrontal cortex (OFC), inferior temporal gyrus and medial superior frontal gyrus, suggesting that both types of aesthetic judgments are based on the orchestration of perceptual, emotional and cognitive components. Experiment 2 examined the network of facial beauty and moral beauty during implicit perception. Participants performed a non-aesthetic judgment task on both faces (beautiful vs common) and scenes (containing morally beautiful vs neutral information). We observed that facial beauty (beautiful faces > common faces) involved both the cortical reward region OFC and the subcortical reward region putamen, whereas moral beauty (moral beauty scenes > moral neutral scenes) only involved the OFC. Moreover, compared with facial beauty, moral beauty spanned a larger-scale cortical network, indicating more advanced and complex cerebral representations characterizing moral beauty. PMID:25298010

  7. Brief report: Representational momentum for dynamic facial expressions in pervasive developmental disorder.

    PubMed

    Uono, Shota; Sato, Wataru; Toichi, Motomi

    2010-03-01

    Individuals with pervasive developmental disorder (PDD) have difficulty with social communication via emotional facial expressions, but behavioral studies involving static images have reported inconsistent findings about emotion recognition. We investigated whether dynamic presentation of facial expression would enhance subjective perception of expressed emotion in 13 individuals with PDD and 13 typically developing controls. We presented dynamic and static emotional (fearful and happy) expressions. Participants were asked to match a changeable emotional face display with the last presented image. The results showed that both groups perceived the last image of dynamic facial expression to be more emotionally exaggerated than the static facial expression. This finding suggests that individuals with PDD have an intact perceptual mechanism for processing dynamic information in another individual's face.

  8. Regional Brain Responses Are Biased Toward Infant Facial Expressions Compared to Adult Facial Expressions in Nulliparous Women.

    PubMed

    Li, Bingbing; Cheng, Gang; Zhang, Dajun; Wei, Dongtao; Qiao, Lei; Wang, Xiangpeng; Che, Xianwei

    2016-01-01

    Recent neuroimaging studies suggest that neutral infant faces compared to neutral adult faces elicit greater activity in brain areas associated with face processing, attention, empathic response, reward, and movement. However, whether infant facial expressions evoke larger brain responses than adult facial expressions remains unclear. Here, we performed event-related functional magnetic resonance imaging in nulliparous women while they were presented with images of matched unfamiliar infant and adult facial expressions (happy, neutral, and uncomfortable/sad) in a pseudo-randomized order. We found that the bilateral fusiform and right lingual gyrus were overall more activated during the presentation of infant facial expressions compared to adult facial expressions. Uncomfortable infant faces compared to sad adult faces evoked greater activation in the bilateral fusiform gyrus, precentral gyrus, postcentral gyrus, posterior cingulate cortex-thalamus, and precuneus. Neutral infant faces activated larger brain responses in the left fusiform gyrus compared to neutral adult faces. Happy infant faces compared to happy adult faces elicited larger responses in areas of the brain associated with emotion and reward processing using a more liberal threshold of p < 0.005 uncorrected. Furthermore, the level of the test subjects' Interest-In-Infants was positively associated with the intensity of right fusiform gyrus response to infant faces and uncomfortable infant faces compared to sad adult faces. In addition, the Perspective Taking subscale score on the Interpersonal Reactivity Index-Chinese was significantly correlated with precuneus activity during uncomfortable infant faces compared to sad adult faces. Our findings suggest that regional brain areas may bias cognitive and emotional responses to infant facial expressions compared to adult facial expressions among nulliparous women, and this bias may be modulated by individual differences in Interest-In-Infants and

  9. Regional Brain Responses Are Biased Toward Infant Facial Expressions Compared to Adult Facial Expressions in Nulliparous Women

    PubMed Central

    Zhang, Dajun; Wei, Dongtao; Qiao, Lei; Wang, Xiangpeng; Che, Xianwei

    2016-01-01

    Recent neuroimaging studies suggest that neutral infant faces compared to neutral adult faces elicit greater activity in brain areas associated with face processing, attention, empathic response, reward, and movement. However, whether infant facial expressions evoke larger brain responses than adult facial expressions remains unclear. Here, we performed event-related functional magnetic resonance imaging in nulliparous women while they were presented with images of matched unfamiliar infant and adult facial expressions (happy, neutral, and uncomfortable/sad) in a pseudo-randomized order. We found that the bilateral fusiform and right lingual gyrus were overall more activated during the presentation of infant facial expressions compared to adult facial expressions. Uncomfortable infant faces compared to sad adult faces evoked greater activation in the bilateral fusiform gyrus, precentral gyrus, postcentral gyrus, posterior cingulate cortex-thalamus, and precuneus. Neutral infant faces activated larger brain responses in the left fusiform gyrus compared to neutral adult faces. Happy infant faces compared to happy adult faces elicited larger responses in areas of the brain associated with emotion and reward processing using a more liberal threshold of p < 0.005 uncorrected. Furthermore, the level of the test subjects’ Interest-In-Infants was positively associated with the intensity of right fusiform gyrus response to infant faces and uncomfortable infant faces compared to sad adult faces. In addition, the Perspective Taking subscale score on the Interpersonal Reactivity Index-Chinese was significantly correlated with precuneus activity during uncomfortable infant faces compared to sad adult faces. Our findings suggest that regional brain areas may bias cognitive and emotional responses to infant facial expressions compared to adult facial expressions among nulliparous women, and this bias may be modulated by individual differences in Interest-In-Infants and

  10. Features versus context: An approach for precise and detailed detection and delineation of faces and facial features.

    PubMed

    Ding, Liya; Martinez, Aleix M

    2010-11-01

    The appearance-based approach to face detection has seen great advances in the last several years. In this approach, we learn the image statistics describing the texture pattern (appearance) of the object class we want to detect, e.g., the face. However, this approach has had limited success in providing an accurate and detailed description of the internal facial features, i.e., eyes, brows, nose, and mouth. In general, this is due to the limited information carried by the learned statistical model. While the face template is relatively rich in texture, facial features (e.g., eyes, nose, and mouth) do not carry enough discriminative information to tell them apart from all possible background images. We resolve this problem by adding the context information of each facial feature in the design of the statistical model. In the proposed approach, the context information defines the image statistics most correlated with the surroundings of each facial component. This means that when we search for a face or facial feature, we look for those locations which most resemble the feature yet are most dissimilar to its context. This dissimilarity with the context features forces the detector to gravitate toward an accurate estimate of the position of the facial feature. Learning to discriminate between feature and context templates is difficult, however, because the context and the texture of the facial features vary widely under changing expression, pose, and illumination, and may even resemble one another. We address this problem with the use of subclass divisions. We derive two algorithms to automatically divide the training samples of each facial feature into a set of subclasses, each representing a distinct construction of the same facial component (e.g., closed versus open eyes) or its context (e.g., different hairstyles). The first algorithm is based on a discriminant analysis formulation. The second algorithm is an extension of the AdaBoost approach. We provide

  11. Professional assessment of facial profile attractiveness.

    PubMed

    Soh, Jen; Chew, Ming Tak; Wong, Hwee Bee

    2005-08-01

    The aim of this study was to compare the assessments of Chinese facial profile attractiveness by orthodontists and oral surgeons. The sample comprised 31 dental professionals (20 orthodontists, 11 oral surgeons) in an Asian community. Facial profile photographs and lateral cephalometric radiographs of 2 Chinese adults (1 man, 1 woman) with normal profiles, Class I incisor relationships, and Class I skeletal patterns were digitized. The digital images were modified by altering cephalometric skeletal and dental hard tissue Chinese normative values in increments of 2 standard deviations in the anteroposterior plane to obtain 7 facial profiles for each sex. The images were bimaxillary protrusion, protrusive mandible, retrusive mandible, normal profile (Class I incisor with Class I skeletal pattern), retrusive maxilla, protrusive maxilla, and bimaxillary retrusion. The Mann-Whitney U test was used to determine professional differences in assessment. Multiple regression analysis was performed with age, professional status, sex, and number of years in practice as independent variables. A strong correlation was found in the profile assessment between orthodontists and oral surgeons. Normal and bimaxillary retrusive Chinese male and female profiles were judged to be highly attractive by orthodontists and oral surgeons. Chinese male and female profiles with protrusive mandibles were judged the least attractive. There was a difference in professional opinion about the most attractive male profile (P < .05), with orthodontists preferring a flatter profile and oral surgeons preferring a fuller normal Chinese profile. Sex of dental professionals and number of years in clinical practice were found to affect profile rankings.

  12. Face-selective regions show invariance to linear, but not to non-linear, changes in facial images.

    PubMed

    Baseler, Heidi A; Young, Andrew W; Jenkins, Rob; Mike Burton, A; Andrews, Timothy J

    2016-12-01

    Familiar face recognition is remarkably invariant across huge image differences, yet little is understood concerning how image-invariant recognition is achieved. To investigate the neural correlates of invariance, we localized the core face-responsive regions and then compared the pattern of fMR-adaptation to different stimulus transformations in each region to behavioural data demonstrating the impact of the same transformations on familiar face recognition. In Experiment 1, we compared linear transformations of size and aspect ratio to a non-linear transformation affecting only part of the face. We found that adaptation to facial identity in face-selective regions showed invariance to linear changes, but there was no invariance to non-linear changes. In Experiment 2, we measured the sensitivity to non-linear changes that fell within the normal range of variation across face images. We found no adaptation to facial identity for any of the non-linear changes in the image, including to faces that varied in different levels of caricature. These results show a compelling difference in the sensitivity to linear compared to non-linear image changes in face-selective regions of the human brain that is only partially consistent with their effect on behavioural judgements of identity. We conclude that while regions such as the FFA may well be involved in the recognition of face identity, they are more likely to contribute to some form of normalisation that underpins subsequent recognition than to form the neural substrate of recognition per se. Copyright © 2016 Elsevier Ltd. All rights reserved.

  13. Facial discrimination in body dysmorphic, obsessive-compulsive and social anxiety disorders.

    PubMed

    Hübner, Claudia; Wiesendahl, Wiebke; Kleinstäuber, Maria; Stangier, Ulrich; Kathmann, Norbert; Buhlmann, Ulrike

    2016-02-28

    Body dysmorphic disorder (BDD) is characterized by preoccupation with perceived flaws in one's own appearance. Several risk factors such as aesthetic perceptual sensitivity have been proposed to explain BDD's unique symptomatology. Although research on facial discrimination is limited so far, the few existing studies have produced mixed results. Thus, the purpose of this study was to further examine facial discrimination in BDD. We administered a facial discrimination paradigm, which allows to assess the ability to identify slight to strong facial changes (e.g., hair loss, acne) when presented with an original (unmodified) facial image, relative to a changed (modified) facial image. The experiment was administered in individuals with BDD, social anxiety disorder, obsessive-compulsive disorder, and mentally healthy controls (32 per group, respectively). Overall, groups did not differ with respect to their ability to correctly identify facial aberrations when presented with other people's faces. Our findings do not support the hypothesis of enhanced general aesthetic perceptual sensitivity in individuals with (vs. without) BDD. Copyright © 2015 Elsevier Ireland Ltd. All rights reserved.

  14. A comparison study of different facial soft tissue analysis methods.

    PubMed

    Kook, Min-Suk; Jung, Seunggon; Park, Hong-Ju; Oh, Hee-Kyun; Ryu, Sun-Youl; Cho, Jin-Hyoung; Lee, Jae-Seo; Yoon, Suk-Ja; Kim, Min-Soo; Shin, Hyo-Keun

    2014-07-01

    The purpose of this study was to evaluate several different facial soft tissue measurement methods. After marking 15 landmarks in the facial area of 12 mannequin heads of different sizes and shapes, facial soft tissue measurements were performed by the following 5 methods: Direct anthropometry, Digitizer, 3D CT, 3D scanner, and DI3D system. With these measurement methods, 10 measurement values representing the facial width, height, and depth were determined twice with a one week interval by one examiner. These data were analyzed with the SPSS program. The position created based on multi-dimensional scaling showed that direct anthropometry, 3D CT, digitizer, 3D scanner demonstrated relatively similar values, while the DI3D system showed slightly different values. All 5 methods demonstrated good accuracy and had a high coefficient of reliability (>0.92) and a low technical error (<0.9 mm). The measured value of the distance between the right and left medial canthus obtained by using the DI3D system was statistically significantly different from that obtained by using the digital caliper, digitizer and laser scanner (p < 0.05), but the other measured values were not significantly different. On evaluating the reproducibility of measurement methods, two measurement values (Ls-Li, G-Pg) obtained by using direct anthropometry, one measurement value (N'-Prn) obtained by using the digitizer, and four measurement values (EnRt-EnLt, AlaRt-AlaLt, ChRt-ChLt, Sn-Pg) obtained by using the DI3D system, were statistically significantly different. However, the mean measurement error in every measurement method was low (<0.7 mm). All measurement values obtained by using the 3D CT and 3D scanner did not show any statistically significant difference. The results of this study show that all 3D facial soft tissue analysis methods demonstrate favorable accuracy and reproducibility, and hence they can be used in clinical practice and research studies. Copyright © 2013 European

  15. [Peripheral facial nerve lesion induced long-term dendritic retraction in pyramidal cortico-facial neurons].

    PubMed

    Urrego, Diana; Múnera, Alejandro; Troncoso, Julieta

    2011-01-01

    Little evidence is available concerning the morphological modifications of motor cortex neurons associated with peripheral nerve injuries, and the consequences of those injuries on post lesion functional recovery. Dendritic branching of cortico-facial neurons was characterized with respect to the effects of irreversible facial nerve injury. Twenty-four adult male rats were distributed into four groups: sham (no lesion surgery), and dendritic assessment at 1, 3 and 5 weeks post surgery. Eighteen lesion animals underwent surgical transection of the mandibular and buccal branches of the facial nerve. Dendritic branching was examined by contralateral primary motor cortex slices stained with the Golgi-Cox technique. Layer V pyramidal (cortico-facial) neurons from sham and injured animals were reconstructed and their dendritic branching was compared using Sholl analysis. Animals with facial nerve lesions displayed persistent vibrissal paralysis throughout the five week observation period. Compared with control animal neurons, cortico-facial pyramidal neurons of surgically injured animals displayed shrinkage of their dendritic branches at statistically significant levels. This shrinkage persisted for at least five weeks after facial nerve injury. Irreversible facial motoneuron axonal damage induced persistent dendritic arborization shrinkage in contralateral cortico-facial neurons. This morphological reorganization may be the physiological basis of functional sequelae observed in peripheral facial palsy patients.

  16. I Can Stomach That! Fearlessness About Death Predicts Attenuated Facial Electromyography Activity in Response to Death-Related Images.

    PubMed

    Velkoff, Elizabeth A; Forrest, Lauren N; Dodd, Dorian R; Smith, April R

    2016-06-01

    Objective measures of suicide risk can convey life-saving information to clinicians, but few such measures exist. This study examined an objective measure of fearlessness about death (FAD), testing whether FAD relates to self-reported and physiological aversion to death. Females (n = 87) reported FAD and disgust sensitivity, and facial electromyography was used to measure physiological facial responses consistent with disgust while viewing death-related images. FAD predicted attenuated expression of physiological death aversion, even when controlling for self-reported death-related disgust sensitivity. Diminished physiological aversion to death-related stimuli holds promise as an objective measure of FAD and suicide risk. © 2015 The American Association of Suicidology.

  17. Influencing Factors Analysis of Facial Nerve Function after the Microsurgical Resection of Acoustic Neuroma

    PubMed Central

    Hong, WenMing; Cheng, HongWei; Wang, XiaoJie; Feng, ChunGuo

    2017-01-01

    Objective To explore and analyze the influencing factors of facial nerve function retainment after microsurgery resection of acoustic neurinoma. Methods Retrospective analysis of our hospital 105 acoustic neuroma cases from October, 2006 to January 2012, in the group all patients were treated with suboccipital sigmoid sinus approach to acoustic neuroma microsurgery resection. We adopted researching individual patient data, outpatient review and telephone followed up and the House-Brackmann grading system to evaluate and analyze the facial nerve function. Results Among 105 patients in this study group, complete surgical resection rate was 80.9% (85/105), subtotal resection rate was 14.3% (15/105), and partial resection rate 4.8% (5/105). The rate of facial nerve retainment on neuroanatomy was 95.3% (100/105) and the mortality rate was 2.1% (2/105). Facial nerve function when the patient is discharged from the hospital, also known as immediate facial nerve function which was graded in House-Brackmann: excellent facial nerve function (House-Brackmann I–II level) cases accounted for 75.2% (79/105), facial nerve function III–IV level cases accounted for 22.9% (24/105), and V–VI cases accounted for 1.9% (2/105). Patients were followed up for more than one year, with excellent facial nerve function retention rate (H-B I–II level) was 74.4% (58/78). Conclusion Acoustic neuroma patients after surgery, the long-term (≥1 year) facial nerve function excellent retaining rate was closely related with surgical proficiency, post-operative immediate facial nerve function, diameter of tumor and whether to use electrophysiological monitoring techniques; while there was no significant correlation with the patient’s age, surgical approach, whether to stripping the internal auditory canal, whether there was cystic degeneration, tumor recurrence, whether to merge with obstructive hydrocephalus and the length of the duration of symptoms. PMID:28264236

  18. The use of three-dimensional imaging to evaluate the effect of conventional orthodontic approach in treating a subject with facial asymmetry

    PubMed Central

    Kheir, Nadia Abou; Kau, Chung How

    2016-01-01

    The growth of the craniofacial skeleton takes place from the 3rd week of intra-uterine life until 18 years of age. During this period, the craniofacial complex is affected by extrinsic and intrinsic factors which guide or alter the pattern of growth. Asymmetry can be encountered due to these multifactorial effects or as the normal divergence of the hemifacial counterpart occurs. At present, an orthodontist plays a major role not only in diagnosing dental asymmetry but also facial asymmetry. However, an orthodontist's role in treating or camouflaging the asymmetry can be limited due to the severity. The aim of this research is to report a technique for facial three-dimensional (3D) analysis used to measure the progress of nonsurgical orthodontic treatment approach for a subject with maxillary asymmetry combined with mandibular angular asymmetry. The facial analysis was composed of five parts: Upper face asymmetry analysis, maxillary analysis, maxillary cant analysis, mandibular cant analysis, and mandibular asymmetry analysis which were applied using 3D software InVivoDental 5.2.3 (Anatomage Company, San Jose, CA, USA). The five components of the facial analysis were applied in the initial cone-beam computed tomography (T1) for diagnosis. Maxillary analysis, maxillary cant analysis, and mandibular cant analysis were applied to measure the progress of the orthodontics treatment (T2). Twenty-two linear measurements bilaterally and sixteen angular criteria were used to analyze the facial structures using different anthropometric landmarks. Only angular mandibular asymmetry was reported. However, the subject had maxillary alveolar ridge cant of 9.96°and dental maxillary cant was 2.95° in T1. The mandibular alveolar ridge cant was 7.41° and the mandibular dental cant was 8.39°. Highest decrease in the cant was reported maxillary alveolar ridge around 2.35° and in the mandibular alveolar ridge around 3.96° in T2. Facial 3D analysis is considered a useful adjunct in

  19. Facial skin color measurement based on camera colorimetric characterization

    NASA Astrophysics Data System (ADS)

    Yang, Boquan; Zhou, Changhe; Wang, Shaoqing; Fan, Xin; Li, Chao

    2016-10-01

    The objective measurement of facial skin color and its variance is of great significance as much information can be obtained from it. In this paper, we developed a new skin color measurement procedure which includes following parts: first, a new skin tone color checker made of pantone Skin Tone Color Checker was designed for camera colorimetric characterization; second, the chromaticity of light source was estimated via a new scene illumination estimation method considering several previous algorithms; third, chromatic adaption was used to convert the input facial image into output facial image which appears taken under canonical light; finally the validity and accuracy of our method was verified by comparing the results obtained by our procedure with these by spectrophotometer.

  20. Facial Indicators of Positive Emotions in Rats

    PubMed Central

    Finlayson, Kathryn; Lampe, Jessica Frances; Hintze, Sara; Würbel, Hanno; Melotti, Luca

    2016-01-01

    Until recently, research in animal welfare science has mainly focused on negative experiences like pain and suffering, often neglecting the importance of assessing and promoting positive experiences. In rodents, specific facial expressions have been found to occur in situations thought to induce negatively valenced emotional states (e.g., pain, aggression and fear), but none have yet been identified for positive states. Thus, this study aimed to investigate if facial expressions indicative of positive emotional state are exhibited in rats. Adolescent male Lister Hooded rats (Rattus norvegicus, N = 15) were individually subjected to a Positive and a mildly aversive Contrast Treatment over two consecutive days in order to induce contrasting emotional states and to detect differences in facial expression. The Positive Treatment consisted of playful manual tickling administered by the experimenter, while the Contrast Treatment consisted of exposure to a novel test room with intermittent bursts of white noise. The number of positive ultrasonic vocalisations was greater in the Positive Treatment compared to the Contrast Treatment, indicating the experience of differentially valenced states in the two treatments. The main findings were that Ear Colour became significantly pinker and Ear Angle was wider (ears more relaxed) in the Positive Treatment compared to the Contrast Treatment. All other quantitative and qualitative measures of facial expression, which included Eyeball height to width Ratio, Eyebrow height to width Ratio, Eyebrow Angle, visibility of the Nictitating Membrane, and the established Rat Grimace Scale, did not show differences between treatments. This study contributes to the exploration of positive emotional states, and thus good welfare, in rats as it identified the first facial indicators of positive emotions following a positive heterospecific play treatment. Furthermore, it provides improvements to the photography technique and image analysis for the

  1. Accurate landmarking of three-dimensional facial data in the presence of facial expressions and occlusions using a three-dimensional statistical facial feature model.

    PubMed

    Zhao, Xi; Dellandréa, Emmanuel; Chen, Liming; Kakadiaris, Ioannis A

    2011-10-01

    Three-dimensional face landmarking aims at automatically localizing facial landmarks and has a wide range of applications (e.g., face recognition, face tracking, and facial expression analysis). Existing methods assume neutral facial expressions and unoccluded faces. In this paper, we propose a general learning-based framework for reliable landmark localization on 3-D facial data under challenging conditions (i.e., facial expressions and occlusions). Our approach relies on a statistical model, called 3-D statistical facial feature model, which learns both the global variations in configurational relationships between landmarks and the local variations of texture and geometry around each landmark. Based on this model, we further propose an occlusion classifier and a fitting algorithm. Results from experiments on three publicly available 3-D face databases (FRGC, BU-3-DFE, and Bosphorus) demonstrate the effectiveness of our approach, in terms of landmarking accuracy and robustness, in the presence of expressions and occlusions.

  2. Facial Scar Revision: Understanding Facial Scar Treatment

    MedlinePlus

    ... Contact Us Trust your face to a facial plastic surgeon Facial Scar Revision Understanding Facial Scar Treatment ... face like the eyes or lips. A facial plastic surgeon has many options for treating and improving ...

  3. Recognizing Facial Expressions Automatically from Video

    NASA Astrophysics Data System (ADS)

    Shan, Caifeng; Braspenning, Ralph

    Facial expressions, resulting from movements of the facial muscles, are the face changes in response to a person's internal emotional states, intentions, or social communications. There is a considerable history associated with the study on facial expressions. Darwin [22] was the first to describe in details the specific facial expressions associated with emotions in animals and humans, who argued that all mammals show emotions reliably in their faces. Since that, facial expression analysis has been a area of great research interest for behavioral scientists [27]. Psychological studies [48, 3] suggest that facial expressions, as the main mode for nonverbal communication, play a vital role in human face-to-face communication. For illustration, we show some examples of facial expressions in Fig. 1.

  4. Facial asymmetry and condylar hyperplasia: considerations for diagnosis in 27 consecutives patients

    PubMed Central

    Olate, Sergio; Almeida, Andrés; Alister, Juan Pablo; Navarro, Pablo; Netto, Henrique Duque; de Moraes, Márcio

    2013-01-01

    Facial asymmetry associated with condylar hyperplasia (CH) has been become the object of study in recent years. The aim of this study is to demonstrate the importance of analyzing the presence of CH in cases of facial asymmetry. Twenty-seven consecutive patients were studied without distinction of age or gender; all the patients consulted for treatment of facial and/or mandibular asymmetry and voluntarily agreed to participate in the study. All the patients underwent facial cone beam tomography and bilateral TMJ as well as a detailed history where they indicated the progression of the disease; in cases of active evolution determined by clinical analysis and imaging, a SPECT analysis was performed to define the isotope uptake. 29.6% of the subjects with scintigram exhibited active CH with a more than 10% difference in uptake between the two condyles; 18.5% presented differences in uptake between 5% and 10%. Active CH was related to the age and gender of the subjects, being more prevalent in women than in men. The aggression level of the uptake was also related to the subject’s age. 55% of the subjects presented with some type of orthodontic treatment with no diagnosis of TMJ pathology in the initial consultation. It can be concluded that CH is associated with facial asymmetries and must be studied integrally before assessing treatment options. PMID:24260600

  5. Finding Makhubu: A morphological forensic facial comparison.

    PubMed

    Houlton, T M R; Steyn, M

    2018-04-01

    June 16, 1976, marks the Soweto Youth Student Uprising in South Africa. A harrowing image capturing police brutality from that day comprises of 18-year-old Mbuyisa Makhubu carrying a dying 12-year-old Hector Peterson. This circulated international press and contributed to world pressure against the apartheid government. This elevated Makhubu's profile with the national security police and forced him to flee to Botswana, then Nigeria, before disappearing in 1978. In 1988, Victor Vinnetou illegally entered Canada and was later arrested on immigration charges in 2004. Evasive of his true identity, the Canadian Border Services Agency and Makhubu's family believe Vinnetou is Makhubu, linking them by a characteristic moon-shaped birthmark on his left chest. A performed DNA test however, was inconclusive. Following the continued 40-year mystery, Eye Witness News in 2016 requested further investigation. Using a limited series of portrait images, a forensic facial comparison (FFC) was conducted utilising South African Police Service (SAPS) protocols and Facial Identification Scientific Working Group (FISWG) guidelines. The images provided, presented a substantial time-lapse and generally low resolution, while being taken from irregular angles and distances, with different subject poses, orientations and environments. This enforced the use of a morphological analysis; a primary method of FFC that develops conclusions based on subjective observations. The results were fundamentally inconclusive, but multiple similarities and valid explanations for visible differences were identified. To enhance the investigation, visual evidence of the moon-shaped birthmark and further DNA analysis is required. Copyright © 2018 Elsevier B.V. All rights reserved.

  6. Molecular analysis of velo-cardio-facial syndrome patients with psychiatric disorders.

    PubMed Central

    Carlson, C; Papolos, D; Pandita, R K; Faedda, G L; Veit, S; Goldberg, R; Shprintzen, R; Kucherlapati, R; Morrow, B

    1997-01-01

    Velo-cardio-facial syndrome (VCFS) is characterized by conotruncal cardiac defects, cleft palate, learning disabilities, and characteristic facial appearance and is associated with hemizygous deletions within 22q11. A newly recognized clinical feature is the presence of psychiatric illness in children and adults with VCFS. To ascertain the relationship between psychiatric illness, VCFS, and chromosome 22 deletions, we evaluated 26 VCFS patients by clinical and molecular biological methods. The VCFS children and adolescents were found to share a set of psychiatric disorders, including bipolar spectrum disorders and attention-deficit disorder with hyperactivity. The adult patients, >18 years of age, were affected with bipolar spectrum disorders. Four of six adult patients had psychotic symptoms manifested as paranoid and grandiose delusions. Loss-of-heterozygosity analysis of all 26 patients revealed that all but 3 had a large 3-Mb common deletion. One patient had a nested distal deletion and two did not have a detectable deletion. Somatic cell hybrids were developed from the two patients who did not have a detectable deletion within 22q11 and were analyzed with a large number of sequence tagged sites. A deletion was not detected among the two patients at a resolution of 21 kb. There was no correlation between the phenotype and the presence of the deletion within 22q11. The remarkably high prevalence of bipolar spectrum disorders, in association with the congenital anomalies of VCFS and its occurrence among nondeleted VCFS patients, suggest a common genetic etiology. Images Figure 4 PMID:9106531

  7. Is moral beauty different from facial beauty? Evidence from an fMRI study.

    PubMed

    Wang, Tingting; Mo, Lei; Mo, Ce; Tan, Li Hai; Cant, Jonathan S; Zhong, Luojin; Cupchik, Gerald

    2015-06-01

    Is moral beauty different from facial beauty? Two functional magnetic resonance imaging experiments were performed to answer this question. Experiment 1 investigated the network of moral aesthetic judgments and facial aesthetic judgments. Participants performed aesthetic judgments and gender judgments on both faces and scenes containing moral acts. The conjunction analysis of the contrasts 'facial aesthetic judgment > facial gender judgment' and 'scene moral aesthetic judgment > scene gender judgment' identified the common involvement of the orbitofrontal cortex (OFC), inferior temporal gyrus and medial superior frontal gyrus, suggesting that both types of aesthetic judgments are based on the orchestration of perceptual, emotional and cognitive components. Experiment 2 examined the network of facial beauty and moral beauty during implicit perception. Participants performed a non-aesthetic judgment task on both faces (beautiful vs common) and scenes (containing morally beautiful vs neutral information). We observed that facial beauty (beautiful faces > common faces) involved both the cortical reward region OFC and the subcortical reward region putamen, whereas moral beauty (moral beauty scenes > moral neutral scenes) only involved the OFC. Moreover, compared with facial beauty, moral beauty spanned a larger-scale cortical network, indicating more advanced and complex cerebral representations characterizing moral beauty. © The Author (2014). Published by Oxford University Press. For Permissions, please email: journals.permissions@oup.com.

  8. Rhinoplasty and facial asymmetry: Analysis of subjective and anthropometric factors in the Caucasian nose

    PubMed Central

    Carvalho, Bettina; Ballin, Annelyse Christine; Becker, Renata Vecentin; Berger, Cezar Augusto Sarraff; Hurtado, Johann G. G. Melcherts; Mocellin, Marcos

    2012-01-01

    Summary Introduction: Anthropometric proportions and symmetry are considered determinants of beauty. These parameters have significant importance in facial plastic surgery, particularly in rhinoplasty. As the central organ of the face, the nose is especially important in determining facial symmetry, both through the perception of a crooked nose and through the determination of facial growth. The evaluation of the presence of facial asymmetry has great relevance preoperatively, both for surgical planning and counseling. Aim/Objective: To evaluate and document the presence of facial asymmetry in patients during rhinoplasty planning and to correlate the anthropometric measures with the perception of facial symmetry or asymmetry, assessing whether there is a higher prevalence of facial asymmetry in these patients compared to volunteers without nasal complaints. Methods: This prospective study was performed by comparing photographs of patients with rhinoplasty planning and volunteers (controls), n = 201, and by evaluating of anthropometric measurements taken from a line passing through the center of the face, until tragus, medial canthus, corner side wing margin, and oral commissure of each side, by statistical analysis (Z test and odds ratio). Results: None of the patients or volunteers had completely symmetric values. Subjectively, 59% of patients were perceived as asymmetric, against 54% of volunteers. Objectively, more than 89% of respondents had asymmetrical measures. Patients had greater RLMTr (MidLine Tragus Ratio) asymmetry than volunteers, which was statistically significant. Discussion/Conclusion: Facial asymmetries are very common in patients seeking rhinoplasty, and special attention should be paid to these aspects both for surgical planning and for counseling of patients. PMID:25991972

  9. Rhinoplasty and facial asymmetry: Analysis of subjective and anthropometric factors in the Caucasian nose.

    PubMed

    Carvalho, Bettina; Ballin, Annelyse Christine; Becker, Renata Vecentin; Berger, Cezar Augusto Sarraff; Hurtado, Johann G G Melcherts; Mocellin, Marcos

    2012-10-01

     Anthropometric proportions and symmetry are considered determinants of beauty. These parameters have significant importance in facial plastic surgery, particularly in rhinoplasty. As the central organ of the face, the nose is especially important in determining facial symmetry, both through the perception of a crooked nose and through the determination of facial growth. The evaluation of the presence of facial asymmetry has great relevance preoperatively, both for surgical planning and counseling.  To evaluate and document the presence of facial asymmetry in patients during rhinoplasty planning and to correlate the anthropometric measures with the perception of facial symmetry or asymmetry, assessing whether there is a higher prevalence of facial asymmetry in these patients compared to volunteers without nasal complaints.  This prospective study was performed by comparing photographs of patients with rhinoplasty planning and volunteers (controls), n = 201, and by evaluating of anthropometric measurements taken from a line passing through the center of the face, until tragus, medial canthus, corner side wing margin, and oral commissure of each side, by statistical analysis (Z test and odds ratio).  None of the patients or volunteers had completely symmetric values. Subjectively, 59% of patients were perceived as asymmetric, against 54% of volunteers. Objectively, more than 89% of respondents had asymmetrical measures. Patients had greater RLMTr (MidLine Tragus Ratio) asymmetry than volunteers, which was statistically significant.  Facial asymmetries are very common in patients seeking rhinoplasty, and special attention should be paid to these aspects both for surgical planning and for counseling of patients.

  10. Facial reanimation by muscle-nerve neurotization after facial nerve sacrifice. Case report.

    PubMed

    Taupin, A; Labbé, D; Babin, E; Fromager, G

    2016-12-01

    Recovering a certain degree of mimicry after sacrifice of the facial nerve is a clinically recognized finding. The authors report a case of hemifacial reanimation suggesting a phenomenon of neurotization from muscle-to-nerve. A woman benefited from a parotidectomy with sacrifice of the left facial nerve indicated for recurrent tumor in the gland. The distal branches of the facial nerve, isolated at the time of resection, were buried in the masseter muscle underneath. The patient recovered a voluntary hémifacial motricity. The electromyographic analysis of the motor activity of the zygomaticus major before and after block of the masseter nerve showed a dependence between mimic muscles and the masseter muscle. Several hypotheses have been advanced to explain the spontaneous reanimation of facial paralysis. The clinical case makes it possible to argue in favor of muscle-to-nerve neurotization from masseter muscle to distal branches of the facial nerve. It illustrates the quality of motricity that can be obtained thanks to this procedure. The authors describe a simple implantation technique of distal branches of the facial nerve in the masseter muscle during a radical parotidectomy with facial nerve sacrifice and recovery of resting tone but also a quality voluntary mimicry. Copyright © 2015 Elsevier Masson SAS. All rights reserved.

  11. Functional Analysis of SPECC1L in Craniofacial Development and Oblique Facial Cleft Pathogenesis

    PubMed Central

    Gfrerer, Lisa; Shubinets, Valeriy; Hoyos, Tatiana; Kong, Yawei; Nguyen, Christina; Pietschmann, Peter; Morton, Cynthia C.; Maas, Richard L.; Liao, Eric C.

    2015-01-01

    Background Oblique facial clefts, also known as Tessier clefts, are severe orofacial clefts, the genetic basis of which is poorly understood. Human genetics studies revealed that disruption in SPECC1L resulted in oblique facial clefts, demonstrating that oblique facial cleft malformation has a genetic basis. An important step toward innovation in treatment of oblique facial clefts would be improved understanding of its genetic pathogenesis. The authors exploit the zebrafish model to elucidate the function of SPECC1L by studying its homolog, specc1lb. Methods Gene and protein expression analysis was carried out by reverse-transcriptase polymerase chain reaction and immunohistochemistry staining. Morpholino knockdown, mRNA rescue, lineage tracing and terminal deoxynucleotidyl transferase-mediated dUTP nick end-labeling assays were performed for functional analysis. Results Expression of specc1lb was detected in epithelia juxtaposed to chondrocytes. Knockdown of specc1lb resulted in bilateral clefts between median and lateral elements of the ethmoid plate, structures analogous to the frontonasal process and the paired maxillary processes. Lineage tracing analysis revealed that cranial neural crest cells contributing to the frontonasal prominence failed to integrate with the maxillary prominence populations. Cells contributing to lower jaw structures were able to migrate to their destined pharyngeal segment but failed to converge to form mandibular elements. Conclusions These results demonstrate that specc1lb is required for integration of frontonasal and maxillary elements and convergence of mandibular prominences. The authors confirm the role of SPECC1L in orofacial cleft pathogenesis in the first animal model of Tessier cleft, providing morphogenetic insight into the mechanisms of normal craniofacial development and oblique facial cleft pathogenesis. PMID:25357034

  12. Intra-temporal facial nerve centerline segmentation for navigated temporal bone surgery

    NASA Astrophysics Data System (ADS)

    Voormolen, Eduard H. J.; van Stralen, Marijn; Woerdeman, Peter A.; Pluim, Josien P. W.; Noordmans, Herke J.; Regli, Luca; Berkelbach van der Sprenkel, Jan W.; Viergever, Max A.

    2011-03-01

    Approaches through the temporal bone require surgeons to drill away bone to expose a target skull base lesion while evading vital structures contained within it, such as the sigmoid sinus, jugular bulb, and facial nerve. We hypothesize that an augmented neuronavigation system that continuously calculates the distance to these structures and warns if the surgeon drills too close, will aid in making safe surgical approaches. Contemporary image guidance systems are lacking an automated method to segment the inhomogeneous and complexly curved facial nerve. Therefore, we developed a segmentation method to delineate the intra-temporal facial nerve centerline from clinically available temporal bone CT images semi-automatically. Our method requires the user to provide the start- and end-point of the facial nerve in a patient's CT scan, after which it iteratively matches an active appearance model based on the shape and texture of forty facial nerves. Its performance was evaluated on 20 patients by comparison to our gold standard: manually segmented facial nerve centerlines. Our segmentation method delineates facial nerve centerlines with a maximum error along its whole trajectory of 0.40+/-0.20 mm (mean+/-standard deviation). These results demonstrate that our model-based segmentation method can robustly segment facial nerve centerlines. Next, we can investigate whether integration of this automated facial nerve delineation with a distance calculating neuronavigation interface results in a system that can adequately warn surgeons during temporal bone drilling, and effectively diminishes risks of iatrogenic facial nerve palsy.

  13. Importance of the brow in facial expressiveness during human communication.

    PubMed

    Neely, John Gail; Lisker, Paul; Drapekin, Jesse

    2014-03-01

    The objective of this study was to evaluate laterality and upper/lower face dominance of expressiveness during prescribed speech using a unique validated image subtraction system capable of sensitive and reliable measurement of facial surface deformation. Observations and experiments of central control of facial expressions during speech and social utterances in humans and animals suggest that the right mouth moves more than the left during nonemotional speech. However, proficient lip readers seem to attend to the whole face to interpret meaning from expressed facial cues, also implicating a horizontal (upper face-lower face) axis. Prospective experimental design. Experimental maneuver: recited speech. image-subtraction strength-duration curve amplitude. Thirty normal human adults were evaluated during memorized nonemotional recitation of 2 short sentences. Facial movements were assessed using a video-image subtractions system capable of simultaneously measuring upper and lower specific areas of each hemiface. The results demonstrate both axes influence facial expressiveness in human communication; however, the horizontal axis (upper versus lower face) would appear dominant, especially during what would appear to be spontaneous breakthrough unplanned expressiveness. These data are congruent with the concept that the left cerebral hemisphere has control over nonemotionally stimulated speech; however, the multisynaptic brainstem extrapyramidal pathways may override hemiface laterality and preferentially take control of the upper face. Additionally, these data demonstrate the importance of the often-ignored brow in facial expressiveness. Experimental study. EBM levels not applicable.

  14. Reduced white matter integrity and facial emotion perception in never-medicated patients with first-episode schizophrenia: A diffusion tensor imaging study.

    PubMed

    Zhao, Xiaoxin; Sui, Yuxiu; Yao, Jingjing; Lv, Yiding; Zhang, Xinyue; Jin, Zhuma; Chen, Lijun; Zhang, Xiangrong

    2017-07-03

    Facial emotion perception is impaired in schizophrenia. Although the pathology of schizophrenia is thought to involve abnormality in white matter (WM), few studies have examined the correlation between facial emotion perception and WM abnormalities in never-medicated patients with first-episode schizophrenia. The present study tested associations between facial emotion perception and WM integrity in order to investigate the neural basis of impaired facial emotion perception in schizophrenia. Sixty-three schizophrenic patients and thirty control subjects underwent facial emotion categorization (FEC). The FEC data was inserted into a logistic function model with subsequent analysis by independent-samples T test and the shift point and slope as outcome measurements. Severity of symptoms was measured using a five-factor model of the Positive and Negative Syndrome Scale (PANSS). Voxelwise group comparison of WM fractional anisotropy (FA) was operated using tract-based spatial statistics (TBSS). The correlation between impaired facial emotion perception and FA reduction was examined in patients using simple regression analysis within brain areas that showed a significant FA reduction in patients compared with controls. The same correlation analysis was also performed for control subjects in the whole brain. The patients with schizophrenia reported a higher shift point and a steeper slope than control subjects in FEC. The patients showed a significant FA reduction in left deep WM in the parietal, temporal and occipital lobes, a small portion of the corpus callosum (CC), and the corona radiata. In voxelwise correlation analysis, we found that facial emotion perception significantly correlated with reduced FA in various WM regions, including left forceps major (FM), inferior longitudinal fasciculus (ILF), inferior fronto-occipital fasciculus (IFOF), Left splenium of CC, and left ILF. The correlation analyses in healthy controls revealed no significant correlation of FA with

  15. Eyeglasses Lens Contour Extraction from Facial Images Using an Efficient Shape Description

    PubMed Central

    Borza, Diana; Darabant, Adrian Sergiu; Danescu, Radu

    2013-01-01

    This paper presents a system that automatically extracts the position of the eyeglasses and the accurate shape and size of the frame lenses in facial images. The novelty brought by this paper consists in three key contributions. The first one is an original model for representing the shape of the eyeglasses lens, using Fourier descriptors. The second one is a method for generating the search space starting from a finite, relatively small number of representative lens shapes based on Fourier morphing. Finally, we propose an accurate lens contour extraction algorithm using a multi-stage Monte Carlo sampling technique. Multiple experiments demonstrate the effectiveness of our approach. PMID:24152926

  16. Facial Nerve Schwannoma: A Case Report, Radiological Features and Literature Review.

    PubMed

    Pilloni, Giulia; Mico, Barbara Massa; Altieri, Roberto; Zenga, Francesco; Ducati, Alessandro; Garbossa, Diego; Tartara, Fulvio

    2017-12-22

    Facial nerve schwannoma localized in the middle fossa is a rare lesion. We report a case of a facial nerve schwannoma in a 30-year-old male presenting with facial nerve palsy. Magnetic resonance imaging (MRI) showed a 3 cm diameter tumor of the right middle fossa. The tumor was removed using a sub-temporal approach. Intraoperative monitoring allowed for identification of the facial nerve, so it was not damaged during the surgical excision. Neurological clinical examination at discharge demonstrated moderate facial nerve improvement (Grade III House-Brackmann).

  17. Facial First Impressions Across Culture: Data-Driven Modeling of Chinese and British Perceivers' Unconstrained Facial Impressions.

    PubMed

    Sutherland, Clare A M; Liu, Xizi; Zhang, Lingshan; Chu, Yingtung; Oldmeadow, Julian A; Young, Andrew W

    2018-04-01

    People form first impressions from facial appearance rapidly, and these impressions can have considerable social and economic consequences. Three dimensions can explain Western perceivers' impressions of Caucasian faces: approachability, youthful-attractiveness, and dominance. Impressions along these dimensions are theorized to be based on adaptive cues to threat detection or sexual selection, making it likely that they are universal. We tested whether the same dimensions of facial impressions emerge across culture by building data-driven models of first impressions of Asian and Caucasian faces derived from Chinese and British perceivers' unconstrained judgments. We then cross-validated the dimensions with computer-generated average images. We found strong evidence for common approachability and youthful-attractiveness dimensions across perceiver and face race, with some evidence of a third dimension akin to capability. The models explained ~75% of the variance in facial impressions. In general, the findings demonstrate substantial cross-cultural agreement in facial impressions, especially on the most salient dimensions.

  18. Small vestibular schwannomas presenting with facial nerve palsy.

    PubMed

    Espahbodi, Mana; Carlson, Matthew L; Fang, Te-Yung; Thompson, Reid C; Haynes, David S

    2014-06-01

    To describe the surgical management and convalescence of two patients presenting with severe facial nerve weakness associated with small intracanalicular vestibular schwannomas (VS). Retrospective review. Two adult female patients presenting with audiovestibular symptoms and subacute facial nerve paralysis (House-Brackmann Grade IV and V). In both cases, post-contrast T1-weighted magnetic resonance imaging revealed an enhancing lesion within the internal auditory canal without lateral extension beyond the fundus. Translabyrinthine exploration demonstrated vestibular nerve origin of tumor, extrinsic to the facial nerve, and frozen section pathology confirmed schwannoma. Gross total tumor resection with VIIth cranial nerve preservation and decompression of the labyrinthine segment of the facial nerve was performed. Both patients recovered full motor function between 6 and 8 months after surgery. Although rare, small VS may cause severe facial neuropathy, mimicking the presentation of facial nerve schwannomas and other less common pathologies. In the absence of labyrinthine extension on MRI, surgical exploration is the only reliable means of establishing a diagnosis. In the case of confirmed VS, early gross total resection with facial nerve preservation and labyrinthine segment decompression may afford full motor recovery-an outcome that cannot be achieved with facial nerve grafting.

  19. A 3D analysis of Caucasian and African American facial morphologies in a US population.

    PubMed

    Talbert, Leslie; Kau, Chung How; Christou, Terpsithea; Vlachos, Christos; Souccar, Nada

    2014-03-01

    This study aimed to compare facial morphologies of an adult African-American population to an adult Caucasian-American population using three-dimensional (3D) surface imaging. The images were captured using a stereophotogrammetric system (3dMDface(TM) system). Subjects were aged 19-30 years, with normal body mass index and no gross craniofacial anomalies. Images were aligned and combined using RF6 Plus Pack 2 software to produce a male and female facial average for each population. The averages were superimposed and the differences were assessed. The most distinct differences were in the forehead, alar base and perioricular regions. The average difference between African-American and Caucasian-American females was 1·18±0·98 mm. The African-American females had a broader face, wider alar base and more protrusive lips. The Caucasian-American females had a more prominent chin, malar region and lower forehead. The average difference between African-American and Caucasian-American males was 1·11±1·04 mm. The African-American males had a more prominent upper forehead and periocular region, wider alar base and more protrusive lips. No notable difference occurred between chin points of the two male populations. Average faces were created from 3D photographs, and the facial morphological differences between populations and genders were compared. African-American males had a more prominent upper forehead and periocular region, wider alar base and more protrusive lips. Caucasian-American males showed a more prominent nasal tip and malar area. African-American females had broader face, wider alar base and more protrusive lips. Caucasian-American females showed a more prominent chin point, malar region and lower forehead.

  20. [Effects of a Facial Muscle Exercise Program including Facial Massage for Patients with Facial Palsy].

    PubMed

    Choi, Hyoung Ju; Shin, Sung Hee

    2016-08-01

    The purpose of this study was to examine the effects of a facial muscle exercise program including facial massage on the facial muscle function, subjective symptoms related to paralysis and depression in patients with facial palsy. This study was a quasi-experimental research with a non-equivalent control group non-synchronized design. Participants were 70 patients with facial palsy (experimental group 35, control group 35). For the experimental group, the facial muscular exercise program including facial massage was performed 20 minutes a day, 3 times a week for two weeks. Data were analyzed using descriptive statistics, χ²-test, Fisher's exact test and independent sample t-test with the SPSS 18.0 program. Facial muscular function of the experimental group improved significantly compared to the control group. There was no significant difference in symptoms related to paralysis between the experimental group and control group. The level of depression in the experimental group was significantly lower than the control group. Results suggest that a facial muscle exercise program including facial massage is an effective nursing intervention to improve facial muscle function and decrease depression in patients with facial palsy.

  1. Middle ear osteoma causing progressive facial nerve weakness: a case report.

    PubMed

    Curtis, Kate; Bance, Manohar; Carter, Michael; Hong, Paul

    2014-09-18

    Facial nerve weakness is most commonly due to Bell's palsy or cerebrovascular accidents. Rarely, middle ear tumor presents with facial nerve dysfunction. We report a very unusual case of middle ear osteoma in a 49-year-old Caucasian woman causing progressive facial nerve deficit. A subtle middle ear lesion was observed on otoscopy and computed tomographic images demonstrated an osseous middle ear tumor. Complete surgical excision resulted in the partial recovery of facial nerve function. Facial nerve dysfunction is rarely caused by middle ear tumors. The weakness is typically due to a compressive effect on the middle ear portion of the facial nerve. Early recognition is crucial since removal of these lesions may lead to the recuperation of facial nerve function.

  2. Sequential Change of Wound Calculated by Image Analysis Using a Color Patch Method during a Secondary Intention Healing.

    PubMed

    Yang, Sejung; Park, Junhee; Lee, Hanuel; Kim, Soohyun; Lee, Byung-Uk; Chung, Kee-Yang; Oh, Byungho

    2016-01-01

    Photographs of skin wounds have the most important information during the secondary intention healing (SIH). However, there is no standard method for handling those images and analyzing them efficiently and conveniently. To investigate the sequential changes of SIH depending on the body sites using a color patch method. We performed retrospective reviews of 30 patients (11 facial and 19 non-facial areas) who underwent SIH for the restoration of skin defects and captured sequential photographs with a color patch which is specially designed for automatically calculating defect and scar sizes. Using a novel image analysis method with a color patch, skin defects were calculated more accurately (range of error rate: -3.39% ~ + 3.05%). All patients had smaller scar size than the original defect size after SIH treatment (rates of decrease: 18.8% ~ 86.1%), and facial area showed significantly higher decrease rate compared with the non-facial area such as scalp and extremities (67.05 ± 12.48 vs. 53.29 ± 18.11, P < 0.05). From the result of estimating the date corresponding to the half of the final decrement, all of the facial area showed improvements within two weeks (8.45 ± 3.91), and non-facial area needed 14.33 ± 9.78 days. From the results of sequential changes of skin defects, SIH can be recommended as an alternative treatment method for restoration with more careful dressing for initial two weeks.

  3. Influence of facial convexity on facial attractiveness in Japanese.

    PubMed

    Ioi, H; Nakata, S; Nakasima, A; Counts, Al

    2007-11-01

    The purpose of this study was to assess and determine the range of the top three most-favored facial profiles for each sex from a series of varying facial convexity, and to evaluate the clinically acceptable facial profiles for Japanese adults. Questionnaire-based study. Silhouettes of average male and female profiles were constructed from the profiles of 30 Japanese males and females with normal occlusions. Chin positions were protruded or retruded by 2 degrees , 4 degrees , 6 degrees , 8 degrees and 10 degrees , respectively, from the average profile. Forty-one orthodontists and 50 dental students were asked to select the three most-favored profiles for each sex, and they were also asked to indicate whether they would seek surgical orthodontic treatment if that image represented their own profile. For males, both the orthodontists and dental students chose the average profile as the most-favored profile. For females, both the orthodontists and dental students chose a slightly more retruded chin position as the most-favored profile. Japanese raters tended to choose class II profiles as more acceptable profiles than class III profiles for both males and females. These findings suggest that Japanese patients with class III profiles tend to seek surgical orthodontic treatment more often.

  4. Comparison of 3D Scanning Versus 2D Photography for the Identification of Facial Soft-Tissue Landmarks.

    PubMed

    Zogheib, T; Jacobs, R; Bornstein, M M; Agbaje, J O; Anumendem, D; Klazen, Y; Politis, C

    2018-01-01

    Three dimensional facial scanning is an innovation that provides opportunity for digital data acquisition, smile analysis and communication of treatment plan and outcome with patients. To assess the applicability of 3D facial scanning as compared to 2D clinical photography. Sample consisted of thirty Caucasians aged between 25 and 50 years old, without any dentofacial deformities. Fifteen soft-tissue facial landmarks were identified twice by 3 observers on 2D and 3D images of the 30 subjects. Five linear proportions and nine angular measurements were established in the orbital, nasal and oral regions. These data were compared to anthropometric norms of young Caucasians. Furthermore, a questionnaire was completed by 14 other observers, according to their personal judgment of the 2D and 3D images. Quantitatively, proportions linking the three facial regions in 3D were closer to the clinical standard (for 2D 3.3% and for 3D 1.8% error rate). Qualitatively, in 67% of the cases, observers were as confident about 3D as they were about 2D. Intra-observer Correlation Coefficient (ICC) revealed a better agreement between observers in 3D for the questions related to facial form, lip step and chin posture. The laser facial scanning could be a useful and reliable tool to analyze the circumoral region for orthodontic and orthognathic treatments as well as for plastic surgery planning and outcome.

  5. 3D FaceCam: a fast and accurate 3D facial imaging device for biometrics applications

    NASA Astrophysics Data System (ADS)

    Geng, Jason; Zhuang, Ping; May, Patrick; Yi, Steven; Tunnell, David

    2004-08-01

    Human faces are fundamentally three-dimensional (3D) objects, and each face has its unique 3D geometric profile. The 3D geometric features of a human face can be used, together with its 2D texture, for rapid and accurate face recognition purposes. Due to the lack of low-cost and robust 3D sensors and effective 3D facial recognition (FR) algorithms, almost all existing FR systems use 2D face images. Genex has developed 3D solutions that overcome the inherent problems in 2D while also addressing limitations in other 3D alternatives. One important aspect of our solution is a unique 3D camera (the 3D FaceCam) that combines multiple imaging sensors within a single compact device to provide instantaneous, ear-to-ear coverage of a human face. This 3D camera uses three high-resolution CCD sensors and a color encoded pattern projection system. The RGB color information from each pixel is used to compute the range data and generate an accurate 3D surface map. The imaging system uses no moving parts and combines multiple 3D views to provide detailed and complete 3D coverage of the entire face. Images are captured within a fraction of a second and full-frame 3D data is produced within a few seconds. This described method provides much better data coverage and accuracy in feature areas with sharp features or details (such as the nose and eyes). Using this 3D data, we have been able to demonstrate that a 3D approach can significantly improve the performance of facial recognition. We have conducted tests in which we have varied the lighting conditions and angle of image acquisition in the "field." These tests have shown that the matching results are significantly improved when enrolling a 3D image rather than a single 2D image. With its 3D solutions, Genex is working toward unlocking the promise of powerful 3D FR and transferring FR from a lab technology into a real-world biometric solution.

  6. Accuracy of computer-assisted navigation: significant augmentation by facial recognition software.

    PubMed

    Glicksman, Jordan T; Reger, Christine; Parasher, Arjun K; Kennedy, David W

    2017-09-01

    Over the past 20 years, image guidance navigation has been used with increasing frequency as an adjunct during sinus and skull base surgery. These devices commonly utilize surface registration, where varying pressure of the registration probe and loss of contact with the face during the skin tracing process can lead to registration inaccuracies, and the number of registration points incorporated is necessarily limited. The aim of this study was to evaluate the use of novel facial recognition software for image guidance registration. Consecutive adults undergoing endoscopic sinus surgery (ESS) were prospectively studied. Patients underwent image guidance registration via both conventional surface registration and facial recognition software. The accuracy of both registration processes were measured at the head of the middle turbinate (MTH), middle turbinate axilla (MTA), anterior wall of sphenoid sinus (SS), and nasal tip (NT). Forty-five patients were included in this investigation. Facial recognition was accurate to within a mean of 0.47 mm at the MTH, 0.33 mm at the MTA, 0.39 mm at the SS, and 0.36 mm at the NT. Facial recognition was more accurate than surface registration at the MTH by an average of 0.43 mm (p = 0.002), at the MTA by an average of 0.44 mm (p < 0.001), and at the SS by an average of 0.40 mm (p < 0.001). The integration of facial recognition software did not adversely affect registration time. In this prospective study, automated facial recognition software significantly improved the accuracy of image guidance registration when compared to conventional surface registration. © 2017 ARS-AAOA, LLC.

  7. Toward DNA-based facial composites: preliminary results and validation.

    PubMed

    Claes, Peter; Hill, Harold; Shriver, Mark D

    2014-11-01

    The potential of constructing useful DNA-based facial composites is forensically of great interest. Given the significant identity information coded in the human face these predictions could help investigations out of an impasse. Although, there is substantial evidence that much of the total variation in facial features is genetically mediated, the discovery of which genes and gene variants underlie normal facial variation has been hampered primarily by the multipartite nature of facial variation. Traditionally, such physical complexity is simplified by simple scalar measurements defined a priori, such as nose or mouth width or alternatively using dimensionality reduction techniques such as principal component analysis where each principal coordinate is then treated as a scalar trait. However, as shown in previous and related work, a more impartial and systematic approach to modeling facial morphology is available and can facilitate both the gene discovery steps, as we recently showed, and DNA-based facial composite construction, as we show here. We first use genomic ancestry and sex to create a base-face, which is simply an average sex and ancestry matched face. Subsequently, the effects of 24 individual SNPs that have been shown to have significant effects on facial variation are overlaid on the base-face forming the predicted-face in a process akin to a photomontage or image blending. We next evaluate the accuracy of predicted faces using cross-validation. Physical accuracy of the facial predictions either locally in particular parts of the face or in terms of overall similarity is mainly determined by sex and genomic ancestry. The SNP-effects maintain the physical accuracy while significantly increasing the distinctiveness of the facial predictions, which would be expected to reduce false positives in perceptual identification tasks. To the best of our knowledge this is the first effort at generating facial composites from DNA and the results are preliminary

  8. Hereditary family signature of facial expression

    PubMed Central

    Peleg, Gili; Katzir, Gadi; Peleg, Ofer; Kamara, Michal; Brodsky, Leonid; Hel-Or, Hagit; Keren, Daniel; Nevo, Eviatar

    2006-01-01

    Although facial expressions of emotion are universal, individual differences create a facial expression “signature” for each person; but, is there a unique family facial expression signature? Only a few family studies on the heredity of facial expressions have been performed, none of which compared the gestalt of movements in various emotional states; they compared only a few movements in one or two emotional states. No studies, to our knowledge, have compared movements of congenitally blind subjects with their relatives to our knowledge. Using two types of analyses, we show a correlation between movements of congenitally blind subjects with those of their relatives in think-concentrate, sadness, anger, disgust, joy, and surprise and provide evidence for a unique family facial expression signature. In the analysis “in-out family test,” a particular movement was compared each time across subjects. Results show that the frequency of occurrence of a movement of a congenitally blind subject in his family is significantly higher than that outside of his family in think-concentrate, sadness, and anger. In the analysis “the classification test,” in which congenitally blind subjects were classified to their families according to the gestalt of movements, results show 80% correct classification over the entire interview and 75% in anger. Analysis of the movements' frequencies in anger revealed a correlation between the movements' frequencies of congenitally blind individuals and those of their relatives. This study anticipates discovering genes that influence facial expressions, understanding their evolutionary significance, and elucidating repair mechanisms for syndromes lacking facial expression, such as autism. PMID:17043232

  9. Impaired social brain network for processing dynamic facial expressions in autism spectrum disorders.

    PubMed

    Sato, Wataru; Toichi, Motomi; Uono, Shota; Kochiyama, Takanori

    2012-08-13

    Impairment of social interaction via facial expressions represents a core clinical feature of autism spectrum disorders (ASD). However, the neural correlates of this dysfunction remain unidentified. Because this dysfunction is manifested in real-life situations, we hypothesized that the observation of dynamic, compared with static, facial expressions would reveal abnormal brain functioning in individuals with ASD.We presented dynamic and static facial expressions of fear and happiness to individuals with high-functioning ASD and to age- and sex-matched typically developing controls and recorded their brain activities using functional magnetic resonance imaging (fMRI). Regional analysis revealed reduced activation of several brain regions in the ASD group compared with controls in response to dynamic versus static facial expressions, including the middle temporal gyrus (MTG), fusiform gyrus, amygdala, medial prefrontal cortex, and inferior frontal gyrus (IFG). Dynamic causal modeling analyses revealed that bi-directional effective connectivity involving the primary visual cortex-MTG-IFG circuit was enhanced in response to dynamic as compared with static facial expressions in the control group. Group comparisons revealed that all these modulatory effects were weaker in the ASD group than in the control group. These results suggest that weak activity and connectivity of the social brain network underlie the impairment in social interaction involving dynamic facial expressions in individuals with ASD.

  10. A unified probabilistic framework for spontaneous facial action modeling and understanding.

    PubMed

    Tong, Yan; Chen, Jixu; Ji, Qiang

    2010-02-01

    Facial expression is a natural and powerful means of human communication. Recognizing spontaneous facial actions, however, is very challenging due to subtle facial deformation, frequent head movements, and ambiguous and uncertain facial motion measurements. Because of these challenges, current research in facial expression recognition is limited to posed expressions and often in frontal view. A spontaneous facial expression is characterized by rigid head movements and nonrigid facial muscular movements. More importantly, it is the coherent and consistent spatiotemporal interactions among rigid and nonrigid facial motions that produce a meaningful facial expression. Recognizing this fact, we introduce a unified probabilistic facial action model based on the Dynamic Bayesian network (DBN) to simultaneously and coherently represent rigid and nonrigid facial motions, their spatiotemporal dependencies, and their image measurements. Advanced machine learning methods are introduced to learn the model based on both training data and subjective prior knowledge. Given the model and the measurements of facial motions, facial action recognition is accomplished through probabilistic inference by systematically integrating visual measurements with the facial action model. Experiments show that compared to the state-of-the-art techniques, the proposed system yields significant improvements in recognizing both rigid and nonrigid facial motions, especially for spontaneous facial expressions.

  11. Facial expression recognition based on weber local descriptor and sparse representation

    NASA Astrophysics Data System (ADS)

    Ouyang, Yan

    2018-03-01

    Automatic facial expression recognition has been one of the research hotspots in the area of computer vision for nearly ten years. During the decade, many state-of-the-art methods have been proposed which perform very high accurate rate based on the face images without any interference. Nowadays, many researchers begin to challenge the task of classifying the facial expression images with corruptions and occlusions and the Sparse Representation based Classification framework has been wildly used because it can robust to the corruptions and occlusions. Therefore, this paper proposed a novel facial expression recognition method based on Weber local descriptor (WLD) and Sparse representation. The method includes three parts: firstly the face images are divided into many local patches, and then the WLD histograms of each patch are extracted, finally all the WLD histograms features are composed into a vector and combined with SRC to classify the facial expressions. The experiment results on the Cohn-Kanade database show that the proposed method is robust to occlusions and corruptions.

  12. Association Among Facial Paralysis, Depression, and Quality of Life in Facial Plastic Surgery Patients

    PubMed Central

    Nellis, Jason C.; Ishii, Masaru; Byrne, Patrick J.; Boahene, Kofi D. O.; Dey, Jacob K.; Ishii, Lisa E.

    2017-01-01

    IMPORTANCE Though anecdotally linked, few studies have investigated the impact of facial paralysis on depression and quality of life (QOL). OBJECTIVE To measure the association between depression, QOL, and facial paralysis in patients seeking treatment at a facial plastic surgery clinic. DESIGN, SETTING, PARTICIPANTS Data were prospectively collected for patients with all-cause facial paralysis and control patients initially presenting to a facial plastic surgery clinic from 2013 to 2015. The control group included a heterogeneous patient population presenting to facial plastic surgery clinic for evaluation. Patients who had prior facial reanimation surgery or missing demographic and psychometric data were excluded from analysis. MAIN OUTCOMES AND MEASURES Demographics, facial paralysis etiology, facial paralysis severity (graded on the House-Brackmann scale), Beck depression inventory, and QOL scores in both groups were examined. Potential confounders, including self-reported attractiveness and mood, were collected and analyzed. Self-reported scores were measured using a 0 to 100 visual analog scale. RESULTS There was a total of 263 patients (mean age, 48.8 years; 66.9% were female) were analyzed. There were 175 control patients and 88 patients with facial paralysis. Sex distributions were not significantly different between the facial paralysis and control groups. Patients with facial paralysis had significantly higher depression, lower self-reported attractiveness, lower mood, and lower QOL scores. Overall, 37 patients with facial paralysis (42.1%) screened positive for depression, with the greatest likelihood in patients with House-Brackmann grade 3 or greater (odds ratio, 10.8; 95% CI, 5.13–22.75) compared with 13 control patients (8.1%) (P < .001). In multivariate regression, facial paralysis and female sex were significantly associated with higher depression scores (constant, 2.08 [95% CI, 0.77–3.39]; facial paralysis effect, 5.98 [95% CI, 4.38–7

  13. Visible skin colouration predicts perception of male facial age, health and attractiveness.

    PubMed

    Fink, B; Bunse, L; Matts, P J; D'Emiliano, D

    2012-08-01

    Although there is evidence that perception of facial age, health and attractiveness is informed by shape characteristics as well as by visible skin condition, studies on the latter have focused almost exclusively on female skin. Recent research, however, suggests that a decrease in skin colour homogeneity leads to older, less healthy and less attractive ratings of facial skin in both women and men. Here, we elaborate on the significance of the homogeneity of visible skin colouration in men by testing the hypothesis that perception of age, health and attractiveness of (non-contextual) digitally isolated fields of cheek skin only can predict that of whole facial images. Facial digital images of 160 British men (all Caucasian) aged between 10 and 70 were blind-rated for age, health and attractiveness by a total of 147 men and 154 women (mean age = 22.95, SD = 4.26), and these ratings were related to those of corresponding images of cheek skin reported by Fink et al. (J. Eur. Acad. Dermatol. Venereol. in press). Linear regression analysis showed that age, health and attractiveness perception of men's faces could be predicted by the ratings of cheek skin only, such that older men were viewed as older, less healthy and less attractive. This result underlines once again the potent signalling role of skin in its own right, independent of shape or other factors and suggests strongly that visible skin condition, and skin colour homogeneity in particular, plays a significant role in the perception of men's faces. © 2012 The Authors. ICS © 2012 Society of Cosmetic Scientists and the Société Française de Cosmétologie.

  14. Angular photogrammetric analysis of the soft-tissue facial profile of Indian adults.

    PubMed

    Pandian, K Saravana; Krishnan, Sindhuja; Kumar, S Aravind

    2018-01-01

    Soft-tissue analysis has become an important component of orthodontic diagnosis and treatment planning. Photographic evaluation of an orthodontic patient is a very close representation of the appearance of the person. The previously established norms for soft-tissue analysis will vary for different ethnic groups. Thus, there is a need to develop soft-tissue facial profile norms pertaining to Indian ethnic groups. The aim of this study is to establish the angular photogrammetric standards of soft-tissue facial profile for Indian males and females and also to compare sexual dimorphism present between them. The lateral profile photographs of 300 random participants (150 males and 150 females) between ages 18 and 25 years were taken and analyzed using FACAD tracing software. Inclusion criteria were angles Class I molar occlusion with acceptable crowding and proclination, normal growth and development with well-aligned dental arches, and full complements of permanent teeth irrespective of third molar status. This study was conducted in Indian population, and samples were taken from various cities across India. Descriptive statistical analysis was carried out, and sexual dimorphism was evaluated by Student's t-test between males and females. The results of the present study showed statistically significant (P < 0.05) gender difference in 5 parameters out of 12 parameters in Indian population. In the present study, soft-tissue facial measurements were established by means of photogrammetric analysis to facilitate orthodontists to carry out more quantitative evaluation and make disciplined decisions. The mean values obtained can be used for comparison with records of participants with the same characteristics by following this photogrammetric technique.

  15. Proposal of Self-Learning and Recognition System of Facial Expression

    NASA Astrophysics Data System (ADS)

    Ogawa, Yukihiro; Kato, Kunihito; Yamamoto, Kazuhiko

    We describe realization of more complicated function by using the information acquired from some equipped unripe functions. The self-learning and recognition system of the human facial expression, which achieved under the natural relation between human and robot, are proposed. The robot with this system can understand human facial expressions and behave according to their facial expressions after the completion of learning process. The system modelled after the process that a baby learns his/her parents’ facial expressions. Equipping the robot with a camera the system can get face images and equipping the CdS sensors on the robot’s head the robot can get the information of human action. Using the information of these sensors, the robot can get feature of each facial expression. After self-learning is completed, when a person changed his facial expression in front of the robot, the robot operates actions under the relevant facial expression.

  16. History and current concepts in the analysis of facial attractiveness.

    PubMed

    Bashour, Mounir

    2006-09-01

    Facial attractiveness research has yielded many discoveries in the past 30 years, and facial cosmetic, plastic, and reconstructive surgeons should have a thorough understanding of these findings. Many of the recent studies were conducted by social, developmental, cognitive, and evolutionary psychologists, and although the findings have been published in the psychology literature, they have not been presented in a comprehensive manner appropriate to surgeons. The author reviews the findings of facial attractiveness research from antiquity to the present day and highlights and analyzes important concepts necessary for a thorough understanding of facial attractiveness. Four important cues emerge as being the most important determinants of attractiveness: averageness (prototypicality), sexual dimorphism, youthfulness, and symmetry. A surgeon planning facial cosmetic, plastic, or reconstructive surgery can potentially gain both profound insight and better quality surgical results by appreciating these findings.

  17. Coding and quantification of a facial expression for pain in lambs.

    PubMed

    Guesgen, M J; Beausoleil, N J; Leach, M; Minot, E O; Stewart, M; Stafford, K J

    2016-11-01

    Facial expressions are routinely used to assess pain in humans, particularly those who are non-verbal. Recently, there has been an interest in developing coding systems for facial grimacing in non-human animals, such as rodents, rabbits, horses and sheep. The aims of this preliminary study were to: 1. Qualitatively identify facial feature changes in lambs experiencing pain as a result of tail-docking and compile these changes to create a Lamb Grimace Scale (LGS); 2. Determine whether human observers can use the LGS to differentiate tail-docked lambs from control lambs and differentiate lambs before and after docking; 3. Determine whether changes in facial action units of the LGS can be objectively quantified in lambs before and after docking; 4. Evaluate effects of restraint of lambs on observers' perceptions of pain using the LGS and on quantitative measures of facial action units. By comparing images of lambs before (no pain) and after (pain) tail-docking, the LGS was devised in consultation with scientists experienced in assessing facial expression in other species. The LGS consists of five facial action units: Orbital Tightening, Mouth Features, Nose Features, Cheek Flattening and Ear Posture. The aims of the study were addressed in two experiments. In Experiment I, still images of the faces of restrained lambs were taken from video footage before and after tail-docking (n=4) or sham tail-docking (n=3). These images were scored by a group of five naïve human observers using the LGS. Because lambs were restrained for the duration of the experiment, Ear Posture was not scored. The scores for the images were averaged to provide one value per feature per period and then scores for the four LGS action units were averaged to give one LGS score per lamb per period. In Experiment II, still images of the faces nine lambs were taken before and after tail-docking. Stills were taken when lambs were restrained and unrestrained in each period. A different group of five

  18. Face-selective regions differ in their ability to classify facial expressions

    PubMed Central

    Zhang, Hui; Japee, Shruti; Nolan, Rachel; Chu, Carlton; Liu, Ning; Ungerleider, Leslie G

    2016-01-01

    Recognition of facial expressions is crucial for effective social interactions. Yet, the extent to which the various face-selective regions in the human brain classify different facial expressions remains unclear. We used functional magnetic resonance imaging (fMRI) and support vector machine pattern classification analysis to determine how well face-selective brain regions are able to decode different categories of facial expression. Subjects participated in a slow event-related fMRI experiment in which they were shown 32 face pictures, portraying four different expressions: neutral, fearful, angry, and happy and belonging to eight different identities. Our results showed that only the amygdala and the posterior superior temporal sulcus (STS) were able to accurately discriminate between these expressions, albeit in different ways: The amygdala discriminated fearful faces from non-fearful faces, whereas STS discriminated neutral from emotional (fearful, angry and happy) faces. In contrast to these findings on the classification of emotional expression, only the fusiform face area (FFA) and anterior inferior temporal cortex (aIT) could discriminate among the various facial identities. Further, the amygdala and STS were better than FFA and aIT at classifying expression, while FFA and aIT were better than the amygdala and STS at classifying identity. Taken together, our findings indicate that the decoding of facial emotion and facial identity occurs in different neural substrates: the amygdala and STS for the former and FFA and aIT for the latter. PMID:26826513

  19. Face-selective regions differ in their ability to classify facial expressions.

    PubMed

    Zhang, Hui; Japee, Shruti; Nolan, Rachel; Chu, Carlton; Liu, Ning; Ungerleider, Leslie G

    2016-04-15

    Recognition of facial expressions is crucial for effective social interactions. Yet, the extent to which the various face-selective regions in the human brain classify different facial expressions remains unclear. We used functional magnetic resonance imaging (fMRI) and support vector machine pattern classification analysis to determine how well face-selective brain regions are able to decode different categories of facial expression. Subjects participated in a slow event-related fMRI experiment in which they were shown 32 face pictures, portraying four different expressions: neutral, fearful, angry, and happy and belonging to eight different identities. Our results showed that only the amygdala and the posterior superior temporal sulcus (STS) were able to accurately discriminate between these expressions, albeit in different ways: the amygdala discriminated fearful faces from non-fearful faces, whereas STS discriminated neutral from emotional (fearful, angry and happy) faces. In contrast to these findings on the classification of emotional expression, only the fusiform face area (FFA) and anterior inferior temporal cortex (aIT) could discriminate among the various facial identities. Further, the amygdala and STS were better than FFA and aIT at classifying expression, while FFA and aIT were better than the amygdala and STS at classifying identity. Taken together, our findings indicate that the decoding of facial emotion and facial identity occurs in different neural substrates: the amygdala and STS for the former and FFA and aIT for the latter. Published by Elsevier Inc.

  20. Facial asymmetry quantitative evaluation in oculoauriculovertebral spectrum.

    PubMed

    Manara, Renzo; Schifano, Giovanni; Brotto, Davide; Mardari, Rodica; Ghiselli, Sara; Gerunda, Antonio; Ghirotto, Cristina; Fusetti, Stefano; Piacentile, Katherine; Scienza, Renato; Ermani, Mario; Martini, Alessandro

    2016-03-01

    Facial asymmetries in oculoauriculovertebral spectrum (OAVS) patients might require surgical corrections that are mostly based on qualitative approach and surgeon's experience. The present study aimed to develop a quantitative 3D CT imaging-based procedure suitable for maxillo-facial surgery planning in OAVS patients. Thirteen OAVS patients (mean age 3.5 ± 4.0 years; range 0.2-14.2, 6 females) and 13 controls (mean age 7.1 ± 5.3 years; range 0.6-15.7, 5 females) who underwent head CT examination were retrospectively enrolled. Eight bilateral anatomical facial landmarks were defined on 3D CT images (porion, orbitale, most anterior point of frontozygomatic suture, most superior point of temporozygomatic suture, most posterior-lateral point of the maxilla, gonion, condylion, mental foramen) and distance from orthogonal planes (in millimeters) was used to evaluate the asymmetry on each axis and to calculate a global asymmetry index of each anatomical landmark. Mean asymmetry values and relative confidence intervals were obtained from the control group. OAVS patients showed 2.5 ± 1.8 landmarks above the confidence interval while considering the global asymmetry values; 12 patients (92%) showed at least one pathologically asymmetric landmark. Considering each axis, the mean number of pathologically asymmetric landmarks increased to 5.5 ± 2.6 (p = 0.002) and all patients presented at least one significant landmark asymmetry. Modern CT-based 3D reconstructions allow accurate assessment of facial bone asymmetries in patients affected by OAVS. The evaluation as a global score and in different orthogonal axes provides precise quantitative data suitable for maxillo-facial surgical planning. CT-based 3D reconstruction might allow a quantitative approach for planning and following-up maxillo-facial surgery in OAVS patients.

  1. Facial nerve hemangioma: a rare case involving the vertical segment.

    PubMed

    Ahmadi, Neda; Newkirk, Kenneth; Kim, H Jeffrey

    2013-02-01

    This case report and literature review reports on a rare case of facial nerve hemangioma (FNH) involving the vertical facial nerve (FN) segment, and discusses the clinical presentation, imaging, pathogenesis, and management of these rare lesions. A 53-year-old male presented with a 10-year history of right hemifacial twitching and progressive facial paresis (House-Brackmann grading score V/VI). The computed tomography and magnetic resonance imaging studies confirmed an expansile lesion along the vertical FN segment. Excision and histopathologic examination demonstrated FNH. FNHs involving the vertical FN segment are extremely rare. Despite being rare lesions, we believe that familiarity with the presentation and management of FNHs are imperative. Laryngoscope, 2012. Copyright © 2012 The American Laryngological, Rhinological, and Otological Society, Inc.

  2. [Facial palsy].

    PubMed

    Cavoy, R

    2013-09-01

    Facial palsy is a daily challenge for the clinicians. Determining whether facial nerve palsy is peripheral or central is a key step in the diagnosis. Central nervous lesions can give facial palsy which may be easily differentiated from peripheral palsy. The next question is the peripheral facial paralysis idiopathic or symptomatic. A good knowledge of anatomy of facial nerve is helpful. A structure approach is given to identify additional features that distinguish symptomatic facial palsy from idiopathic one. The main cause of peripheral facial palsies is idiopathic one, or Bell's palsy, which remains a diagnosis of exclusion. The most common cause of symptomatic peripheral facial palsy is Ramsay-Hunt syndrome. Early identification of symptomatic facial palsy is important because of often worst outcome and different management. The prognosis of Bell's palsy is on the whole favorable and is improved with a prompt tapering course of prednisone. In Ramsay-Hunt syndrome, an antiviral therapy is added along with prednisone. We also discussed of current treatment recommendations. We will review short and long term complications of peripheral facial palsy.

  3. A large-scale analysis of sex differences in facial expressions

    PubMed Central

    Kodra, Evan; el Kaliouby, Rana; LaFrance, Marianne

    2017-01-01

    There exists a stereotype that women are more expressive than men; however, research has almost exclusively focused on a single facial behavior, smiling. A large-scale study examines whether women are consistently more expressive than men or whether the effects are dependent on the emotion expressed. Studies of gender differences in expressivity have been somewhat restricted to data collected in lab settings or which required labor-intensive manual coding. In the present study, we analyze gender differences in facial behaviors as over 2,000 viewers watch a set of video advertisements in their home environments. The facial responses were recorded using participants’ own webcams. Using a new automated facial coding technology we coded facial activity. We find that women are not universally more expressive across all facial actions. Nor are they more expressive in all positive valence actions and less expressive in all negative valence actions. It appears that generally women express actions more frequently than men, and in particular express more positive valence actions. However, expressiveness is not greater in women for all negative valence actions and is dependent on the discrete emotional state. PMID:28422963

  4. Direction of Amygdala-Neocortex Interaction During Dynamic Facial Expression Processing.

    PubMed

    Sato, Wataru; Kochiyama, Takanori; Uono, Shota; Yoshikawa, Sakiko; Toichi, Motomi

    2017-03-01

    Dynamic facial expressions of emotion strongly elicit multifaceted emotional, perceptual, cognitive, and motor responses. Neuroimaging studies revealed that some subcortical (e.g., amygdala) and neocortical (e.g., superior temporal sulcus and inferior frontal gyrus) brain regions and their functional interaction were involved in processing dynamic facial expressions. However, the direction of the functional interaction between the amygdala and the neocortex remains unknown. To investigate this issue, we re-analyzed functional magnetic resonance imaging (fMRI) data from 2 studies and magnetoencephalography (MEG) data from 1 study. First, a psychophysiological interaction analysis of the fMRI data confirmed the functional interaction between the amygdala and neocortical regions. Then, dynamic causal modeling analysis was used to compare models with forward, backward, or bidirectional effective connectivity between the amygdala and neocortical networks in the fMRI and MEG data. The results consistently supported the model of effective connectivity from the amygdala to the neocortex. Further increasing time-window analysis of the MEG demonstrated that this model was valid after 200 ms from the stimulus onset. These data suggest that emotional processing in the amygdala rapidly modulates some neocortical processing, such as perception, recognition, and motor mimicry, when observing dynamic facial expressions of emotion. © The Author 2016. Published by Oxford University Press. All rights reserved. For Permissions, please e-mail: journals.permissions@oup.com.

  5. Differential amygdala response during facial recognition in patients with schizophrenia: an fMRI study.

    PubMed

    Kosaka, H; Omori, M; Murata, T; Iidaka, T; Yamada, H; Okada, T; Takahashi, T; Sadato, N; Itoh, H; Yonekura, Y; Wada, Y

    2002-09-01

    Human lesion or neuroimaging studies suggest that amygdala is involved in facial emotion recognition. Although impairments in recognition of facial and/or emotional expression have been reported in schizophrenia, there are few neuroimaging studies that have examined differential brain activation during facial recognition between patients with schizophrenia and normal controls. To investigate amygdala responses during facial recognition in schizophrenia, we conducted a functional magnetic resonance imaging (fMRI) study with 12 right-handed medicated patients with schizophrenia and 12 age- and sex-matched healthy controls. The experiment task was a type of emotional intensity judgment task. During the task period, subjects were asked to view happy (or angry/disgusting/sad) and neutral faces simultaneously presented every 3 s and to judge which face was more emotional (positive or negative face discrimination). Imaging data were investigated in voxel-by-voxel basis for single-group analysis and for between-group analysis according to the random effect model using Statistical Parametric Mapping (SPM). No significant difference in task accuracy was found between the schizophrenic and control groups. Positive face discrimination activated the bilateral amygdalae of both controls and schizophrenics, with more prominent activation of the right amygdala shown in the schizophrenic group. Negative face discrimination activated the bilateral amygdalae in the schizophrenic group whereas the right amygdala alone in the control group, although no significant group difference was found. Exaggerated amygdala activation during emotional intensity judgment found in the schizophrenic patients may reflect impaired gating of sensory input containing emotion. Copyright 2002 Elsevier Science B.V.

  6. The telltale face: possible mechanisms behind defector and cooperator recognition revealed by emotional facial expression metrics.

    PubMed

    Kovács-Bálint, Zsófia; Bereczkei, Tamás; Hernádi, István

    2013-11-01

    In this study, we investigated the role of facial cues in cooperator and defector recognition. First, a face image database was constructed from pairs of full face portraits of target subjects taken at the moment of decision-making in a prisoner's dilemma game (PDG) and in a preceding neutral task. Image pairs with no deficiencies (n = 67) were standardized for orientation and luminance. Then, confidence in defector and cooperator recognition was tested with image rating in a different group of lay judges (n = 62). Results indicate that (1) defectors were better recognized (58% vs. 47%), (2) they looked different from cooperators (p < .01), (3) males but not females evaluated the images with a relative bias towards the cooperator category (p < .01), and (4) females were more confident in detecting defectors (p < .05). According to facial microexpression analysis, defection was strongly linked with depressed lower lips and less opened eyes. Significant correlation was found between the intensity of micromimics and the rating of images in the cooperator-defector dimension. In summary, facial expressions can be considered as reliable indicators of momentary social dispositions in the PDG. Females may exhibit an evolutionary-based overestimation bias to detecting social visual cues of the defector face. © 2012 The British Psychological Society.

  7. Social perception of morbidity in facial nerve paralysis.

    PubMed

    Li, Matthew Ka Ki; Niles, Navin; Gore, Sinclair; Ebrahimi, Ardalan; McGuinness, John; Clark, Jonathan Robert

    2016-08-01

    There are many patient-based and clinician-based scales measuring the severity of facial nerve paralysis and the impact on quality of life, however, the social perception of facial palsy has received little attention. The purpose of this pilot study was to measure the consequences of facial paralysis on selected domains of social perception and compare the social impact of paralysis of the different components. Four patients with typical facial palsies (global, marginal mandibular, zygomatic/buccal, and frontal) and 1 control were photographed. These images were each shown to 100 participants who subsequently rated variables of normality, perceived distress, trustworthiness, intelligence, interaction, symmetry, and disability. Statistical analysis was performed to compare the results among each palsy. Paralyzed faces were considered less normal compared to the control on a scale of 0 to 10 (mean, 8.6; 95% confidence interval [CI] = 8.30-8.86) with global paralysis (mean, 3.4; 95% CI = 3.08-3.80) rated as the most disfiguring, followed by the zygomatic/buccal (mean, 6.0; 95% CI = 5.68-6.37), marginal (mean, 6.5; 95% CI = 6.08-6.86), and then temporal palsies (mean, 6.9; 95% CI = 6.57-7.21). Similar trends were seen when analyzing these palsies for perceived distress, intelligence, and trustworthiness, using a random effects regression model. Our sample suggests that society views paralyzed faces as less normal, less trustworthy, and more distressed. Different components of facial paralysis are worse than others and surgical correction may need to be prioritized in an evidence-based manner with social morbidity in mind. © 2016 Wiley Periodicals, Inc. Head Neck 38:1158-1163, 2016. © 2016 Wiley Periodicals, Inc.

  8. Multiracial Facial Golden Ratio and Evaluation of Facial Appearance.

    PubMed

    Alam, Mohammad Khursheed; Mohd Noor, Nor Farid; Basri, Rehana; Yew, Tan Fo; Wen, Tay Hui

    2015-01-01

    This study aimed to investigate the association of facial proportion and its relation to the golden ratio with the evaluation of facial appearance among Malaysian population. This was a cross-sectional study with 286 randomly selected from Universiti Sains Malaysia (USM) Health Campus students (150 females and 136 males; 100 Malaysian Chinese, 100 Malaysian Malay and 86 Malaysian Indian), with the mean age of 21.54 ± 1.56 (Age range, 18-25). Facial indices obtained from direct facial measurements were used for the classification of facial shape into short, ideal and long. A validated structured questionnaire was used to assess subjects' evaluation of their own facial appearance. The mean facial indices of Malaysian Indian (MI), Malaysian Chinese (MC) and Malaysian Malay (MM) were 1.59 ± 0.19, 1.57 ± 0.25 and 1.54 ± 0.23 respectively. Only MC showed significant sexual dimorphism in facial index (P = 0.047; P<0.05) but no significant difference was found between races. Out of the 286 subjects, 49 (17.1%) were of ideal facial shape, 156 (54.5%) short and 81 (28.3%) long. The facial evaluation questionnaire showed that MC had the lowest satisfaction with mean score of 2.18 ± 0.97 for overall impression and 2.15 ± 1.04 for facial parts, compared to MM and MI, with mean score of 1.80 ± 0.97 and 1.64 ± 0.74 respectively for overall impression; 1.75 ± 0.95 and 1.70 ± 0.83 respectively for facial parts. 1) Only 17.1% of Malaysian facial proportion conformed to the golden ratio, with majority of the population having short face (54.5%); 2) Facial index did not depend significantly on races; 3) Significant sexual dimorphism was shown among Malaysian Chinese; 4) All three races are generally satisfied with their own facial appearance; 5) No significant association was found between golden ratio and facial evaluation score among Malaysian population.

  9. Multiracial Facial Golden Ratio and Evaluation of Facial Appearance

    PubMed Central

    2015-01-01

    This study aimed to investigate the association of facial proportion and its relation to the golden ratio with the evaluation of facial appearance among Malaysian population. This was a cross-sectional study with 286 randomly selected from Universiti Sains Malaysia (USM) Health Campus students (150 females and 136 males; 100 Malaysian Chinese, 100 Malaysian Malay and 86 Malaysian Indian), with the mean age of 21.54 ± 1.56 (Age range, 18–25). Facial indices obtained from direct facial measurements were used for the classification of facial shape into short, ideal and long. A validated structured questionnaire was used to assess subjects’ evaluation of their own facial appearance. The mean facial indices of Malaysian Indian (MI), Malaysian Chinese (MC) and Malaysian Malay (MM) were 1.59 ± 0.19, 1.57 ± 0.25 and 1.54 ± 0.23 respectively. Only MC showed significant sexual dimorphism in facial index (P = 0.047; P<0.05) but no significant difference was found between races. Out of the 286 subjects, 49 (17.1%) were of ideal facial shape, 156 (54.5%) short and 81 (28.3%) long. The facial evaluation questionnaire showed that MC had the lowest satisfaction with mean score of 2.18 ± 0.97 for overall impression and 2.15 ± 1.04 for facial parts, compared to MM and MI, with mean score of 1.80 ± 0.97 and 1.64 ± 0.74 respectively for overall impression; 1.75 ± 0.95 and 1.70 ± 0.83 respectively for facial parts. In conclusion: 1) Only 17.1% of Malaysian facial proportion conformed to the golden ratio, with majority of the population having short face (54.5%); 2) Facial index did not depend significantly on races; 3) Significant sexual dimorphism was shown among Malaysian Chinese; 4) All three races are generally satisfied with their own facial appearance; 5) No significant association was found between golden ratio and facial evaluation score among Malaysian population. PMID:26562655

  10. Survey on RGB, 3D, Thermal, and Multimodal Approaches for Facial Expression Recognition: History, Trends, and Affect-Related Applications.

    PubMed

    Corneanu, Ciprian Adrian; Simon, Marc Oliu; Cohn, Jeffrey F; Guerrero, Sergio Escalera

    2016-08-01

    Facial expressions are an important way through which humans interact socially. Building a system capable of automatically recognizing facial expressions from images and video has been an intense field of study in recent years. Interpreting such expressions remains challenging and much research is needed about the way they relate to human affect. This paper presents a general overview of automatic RGB, 3D, thermal and multimodal facial expression analysis. We define a new taxonomy for the field, encompassing all steps from face detection to facial expression recognition, and describe and classify the state of the art methods accordingly. We also present the important datasets and the bench-marking of most influential methods. We conclude with a general discussion about trends, important questions and future lines of research.

  11. Responses in the right posterior superior temporal sulcus show a feature-based response to facial expression.

    PubMed

    Flack, Tessa R; Andrews, Timothy J; Hymers, Mark; Al-Mosaiwi, Mohammed; Marsden, Samuel P; Strachan, James W A; Trakulpipat, Chayanit; Wang, Liang; Wu, Tian; Young, Andrew W

    2015-08-01

    The face-selective region of the right posterior superior temporal sulcus (pSTS) plays an important role in analysing facial expressions. However, it is less clear how facial expressions are represented in this region. In this study, we used the face composite effect to explore whether the pSTS contains a holistic or feature-based representation of facial expression. Aligned and misaligned composite images were created from the top and bottom halves of faces posing different expressions. In Experiment 1, participants performed a behavioural matching task in which they judged whether the top half of two images was the same or different. The ability to discriminate the top half of the face was affected by changes in the bottom half of the face when the images were aligned, but not when they were misaligned. This shows a holistic behavioural response to expression. In Experiment 2, we used fMR-adaptation to ask whether the pSTS has a corresponding holistic neural representation of expression. Aligned or misaligned images were presented in blocks that involved repeating the same image or in which the top or bottom half of the images changed. Increased neural responses were found in the right pSTS regardless of whether the change occurred in the top or bottom of the image, showing that changes in expression were detected across all parts of the face. However, in contrast to the behavioural data, the pattern did not differ between aligned and misaligned stimuli. This suggests that the pSTS does not encode facial expressions holistically. In contrast to the pSTS, a holistic pattern of response to facial expression was found in the right inferior frontal gyrus (IFG). Together, these results suggest that pSTS reflects an early stage in the processing of facial expression in which facial features are represented independently. Copyright © 2015 Elsevier Ltd. All rights reserved.

  12. Effect analysis of intradermal hyaluronic acid injection to treat enlarged facial pores.

    PubMed

    Qian, Wei; Zhang, Yan-Kun; Hou, Ying; Lyu, Wei; Cao, Qian; Li, Yan-Qi; Fan, Ju-Feng

    2017-08-08

    To investigate the clinical application and efficacy of intradermal injection of low molecular weight hyaluronic acid (LMW-HA) for treating enlarged facial pores. From January 2015 to May 2016, 42 subjects who sought aesthetic treatment underwent intradermal injection of LMW-HA to improve enlarged facial pores. For each treatment, 2.5 mL (25 mg) of LMW-HA was injected into the skin of the full face. The treatment was repeated 2-5 times with an interval of 1 to 1.5 months between consecutive treatments. The postoperative follow-up period was 1 to 6 months. Statistical analysis was used to compare the degree of enlargement of facial pores before and after injection. The clinical efficacy and adverse effects were recorded. The enlarged facial pores before and after treatment were categorized and subjected to the Wilcoxon matched-pairs signed-rank test. The difference was statistically significant (P<.01). The improvement rate was 40.03±18.41%. No infection, nodules, or pigmentation was reported at the injection sites in the subjects who sought aesthetic treatment. The overall satisfaction rate was 92.8%. Intradermal injection of LMW-HA can significantly improve skin texture, reduce pore size, and enhance skin radiance. The injection technique was simple, safe, and effective and could easily be extended to clinical practice. © 2017 Wiley Periodicals, Inc.

  13. Combat-Related Facial Burns: Analysis of Strategic Pitfalls

    DTIC Science & Technology

    2015-01-01

    10), with a high rate of associated inhalational injuries (61%, n = 19), limb amputations (29%, n = 9), and facial allograft usage (48%, n = 15...Combat injuries result in greater burn depths, a greater incidence of inhalation injuries, a greater incidence of concomitant nonburn injuries, and a...associated inhalational injury, of which, 7 were mild, 6 were moderate, and 6 were severe. Of the 31 patients, 6 (19%) had concomitant facial fractures

  14. Facial Anthropometric Evaluation of Unilateral Cleft Lip and Palate Patients: Infancy Through Adolescence.

    PubMed

    Dehghani, Mahboobe; Jahanbin, Arezoo; Omidkhoda, Maryam; Entezari, Mostafa; Shadkam, Elaheh

    2018-03-01

    Craniofacial anthropometric studies measure the differences in humans' craniofacial dimensions. The aim of this study was to determine facial anthropometric dimensions of newborn to 12-year-old girls with nonsyndromic unilateral cleft lip and palate (UCLP). In this cross-sectional analytical study, data was collected from 65 infant to 12-year old girls with UCLP. Digital frontal and profile facial photographs were transferred to a computer and desired anthropometric landmarks were traced on each image. Fifteen anthropometric parameters were measured which were the angles of facial, nasofacial, nasomental, Z, nasolabial, inclination of nasal base and labial fissure, nasal deviation, mentocervical, facial convexity and also ratios of nasal prominence relative to nasal height, middle to lower facial third, upper lip to lower lip height, columellar length relative to upper lip, and incisal show relative to incisal width. Pearson coefficient and linear regression were used for statistical analysis. Upper lip to lower lip height ratio and angles of nasofacial, nasolabial, and facial convexity decreased with the age of the patients. In contrast, nasomental angle and the ratios of columellar length to upper lip length, middle facial height to lower facial height, and incisal show relative to incisal width increased. Other parameters studied did not appear to have any significant correlation with age. In the girls with UCLP, various craniofacial dimensions have different growth rates with some parts growing slower than others. Some of the parameters studied were significantly correlated with age, thus growth-related curves and equations were obtained and presented.

  15. Cone beam tomographic study of facial structures characteristics at rest and wide smile, and their correlation with the facial types.

    PubMed

    Martins, Luciana Flaquer; Vigorito, Julio Wilson

    2013-01-01

    To determine the characteristics of facial soft tissues at rest and wide smile, and their possible relation to the facial type. We analyzed a sample of forty-eight young female adults, aged between 19.10 and 40 years old, with a mean age of 30.9 years, who had balanced profile and passive lip seal. Cone beam computed tomographies were performed at rest and wide smile postures on the entire sample which was divided into three groups according to individual facial types. Soft tissue features analysis of the lips, nose, zygoma and chin were done in sagittal, axial and frontal axis tomographic views. No differences were observed in any of the facial type variables for the static analysis of facial structures at both rest and wide smile postures. Dynamic analysis showed that brachifacial types are more sensitive to movement, presenting greater sagittal lip contraction. However, the lip movement produced by this type of face results in a narrow smile, with smaller tooth exposure area when compared with other facial types. Findings pointed out that the position of the upper lip should be ahead of the lower lip, and the latter, ahead of the pogonion. It was also found that the facial type does not impact the positioning of these structures. Additionally, the use of cone beam computed tomography may be a valuable method to study craniofacial features.

  16. A small-world network model of facial emotion recognition.

    PubMed

    Takehara, Takuma; Ochiai, Fumio; Suzuki, Naoto

    2016-01-01

    Various models have been proposed to increase understanding of the cognitive basis of facial emotions. Despite those efforts, interactions between facial emotions have received minimal attention. If collective behaviours relating to each facial emotion in the comprehensive cognitive system could be assumed, specific facial emotion relationship patterns might emerge. In this study, we demonstrate that the frameworks of complex networks can effectively capture those patterns. We generate 81 facial emotion images (6 prototypes and 75 morphs) and then ask participants to rate degrees of similarity in 3240 facial emotion pairs in a paired comparison task. A facial emotion network constructed on the basis of similarity clearly forms a small-world network, which features an extremely short average network distance and close connectivity. Further, even if two facial emotions have opposing valences, they are connected within only two steps. In addition, we show that intermediary morphs are crucial for maintaining full network integration, whereas prototypes are not at all important. These results suggest the existence of collective behaviours in the cognitive systems of facial emotions and also describe why people can efficiently recognize facial emotions in terms of information transmission and propagation. For comparison, we construct three simulated networks--one based on the categorical model, one based on the dimensional model, and one random network. The results reveal that small-world connectivity in facial emotion networks is apparently different from those networks, suggesting that a small-world network is the most suitable model for capturing the cognitive basis of facial emotions.

  17. Promising Technique for Facial Nerve Reconstruction in Extended Parotidectomy

    PubMed Central

    Villarreal, Ithzel Maria; Rodríguez-Valiente, Antonio; Castelló, Jose Ramon; Górriz, Carmen; Montero, Oscar Alvarez; García-Berrocal, Jose Ramon

    2015-01-01

    Introduction: Malignant tumors of the parotid gland account scarcely for 5% of all head and neck tumors. Most of these neoplasms have a high tendency for recurrence, local infiltration, perineural extension, and metastasis. Although uncommon, these malignant tumors require complex surgical treatment sometimes involving a total parotidectomy including a complete facial nerve resection. Severe functional and aesthetic facial defects are the result of a complete sacrifice or injury to isolated branches becoming an uncomfortable distress for patients and a major challenge for reconstructive surgeons. Case Report: A case of a 54-year-old, systemically healthy male patient with a 4 month complaint of pain and swelling on the right side of the face is presented. The patient reported a rapid increase in the size of the lesion over the past 2 months. Imaging tests and histopathological analysis reported an adenoid cystic carcinoma. A complete parotidectomy was carried out with an intraoperative notice of facial nerve infiltration requiring a second intervention for nerve and defect reconstruction. A free ALT flap with vascularized nerve grafts was the surgical choice. A 6 month follow-up showed partial facial movement recovery and the facial defect mended. Conclusion: It is of critical importance to restore function to patients with facial nerve injury. Vascularized nerve grafts, in many clinical and experimental studies, have shown to result in better nerve regeneration than conventional non-vascularized nerve grafts. Nevertheless, there are factors that may affect the degree, speed and regeneration rate regarding the free fasciocutaneous flap. In complex head and neck defects following a total parotidectomy, the extended free fasciocutaneous ALT (anterior-lateral thigh) flap with a vascularized nerve graft is ideally suited for the reconstruction of the injured site. Donor–site morbidity is low and additional surgical time is minimal compared with the time of a single

  18. Promising Technique for Facial Nerve Reconstruction in Extended Parotidectomy.

    PubMed

    Villarreal, Ithzel Maria; Rodríguez-Valiente, Antonio; Castelló, Jose Ramon; Górriz, Carmen; Montero, Oscar Alvarez; García-Berrocal, Jose Ramon

    2015-11-01

    Malignant tumors of the parotid gland account scarcely for 5% of all head and neck tumors. Most of these neoplasms have a high tendency for recurrence, local infiltration, perineural extension, and metastasis. Although uncommon, these malignant tumors require complex surgical treatment sometimes involving a total parotidectomy including a complete facial nerve resection. Severe functional and aesthetic facial defects are the result of a complete sacrifice or injury to isolated branches becoming an uncomfortable distress for patients and a major challenge for reconstructive surgeons. A case of a 54-year-old, systemically healthy male patient with a 4 month complaint of pain and swelling on the right side of the face is presented. The patient reported a rapid increase in the size of the lesion over the past 2 months. Imaging tests and histopathological analysis reported an adenoid cystic carcinoma. A complete parotidectomy was carried out with an intraoperative notice of facial nerve infiltration requiring a second intervention for nerve and defect reconstruction. A free ALT flap with vascularized nerve grafts was the surgical choice. A 6 month follow-up showed partial facial movement recovery and the facial defect mended. It is of critical importance to restore function to patients with facial nerve injury. Vascularized nerve grafts, in many clinical and experimental studies, have shown to result in better nerve regeneration than conventional non-vascularized nerve grafts. Nevertheless, there are factors that may affect the degree, speed and regeneration rate regarding the free fasciocutaneous flap. In complex head and neck defects following a total parotidectomy, the extended free fasciocutaneous ALT (anterior-lateral thigh) flap with a vascularized nerve graft is ideally suited for the reconstruction of the injured site. Donor-site morbidity is low and additional surgical time is minimal compared with the time of a single ALT flap transfer.

  19. Improvement of the facial evenness of leave-on skincare products by a modified application method in Chinese women.

    PubMed

    Zou, Y; Wang, X; Fan, G

    2015-04-01

    To understand the habits of Chinese women applying leave-on skincare products (LOSCP) and to improve female facial evenness of anti-ageing cosmetics through modifying facial skincare smear ways. A questionnaire on the method of applying LOSCP was distributed to 60 women with habit of using LOSCP. Their facial images before and after applying LOSCP were taken, and their positioning and grey value were used to analyse the effects of different applying methods on the uniformity of facial LOSCP. LOSCP including anti-ageing cosmetics have been widely used among Chinese women for a long time. However, some women do not concern how to properly apply LOSCP. In our survey, the main focal points of the face are forehead, malar region, cheek, mouth corners and chin when they looking into the mirror, and mouth corners and inner canthus are often overlooked when applying cosmetic products. The image analysis found that after applying the LOSCP, the greyscale of the forehead, glabella, malar region, upper lip region and jaw changed significantly whereas that of canthus, mouth corners and lateral cheek region was not significantly different. Applying an improved smear method (11-point method)could significantly increase the grey values of various facial areas. The way of Chinese women applying LOSCP may result in facial unevenness of skin products. By improving facial skincare smear method, one can make the products even in all facial areas, thereby ensuring the efficacy of anti-ageing cosmetics. Thus, further improvement and education regarding skincare is required. © 2014 Society of Cosmetic Scientists and the Société Française de Cosmétologie.

  20. Impaired social brain network for processing dynamic facial expressions in autism spectrum disorders

    PubMed Central

    2012-01-01

    Background Impairment of social interaction via facial expressions represents a core clinical feature of autism spectrum disorders (ASD). However, the neural correlates of this dysfunction remain unidentified. Because this dysfunction is manifested in real-life situations, we hypothesized that the observation of dynamic, compared with static, facial expressions would reveal abnormal brain functioning in individuals with ASD. We presented dynamic and static facial expressions of fear and happiness to individuals with high-functioning ASD and to age- and sex-matched typically developing controls and recorded their brain activities using functional magnetic resonance imaging (fMRI). Result Regional analysis revealed reduced activation of several brain regions in the ASD group compared with controls in response to dynamic versus static facial expressions, including the middle temporal gyrus (MTG), fusiform gyrus, amygdala, medial prefrontal cortex, and inferior frontal gyrus (IFG). Dynamic causal modeling analyses revealed that bi-directional effective connectivity involving the primary visual cortex–MTG–IFG circuit was enhanced in response to dynamic as compared with static facial expressions in the control group. Group comparisons revealed that all these modulatory effects were weaker in the ASD group than in the control group. Conclusions These results suggest that weak activity and connectivity of the social brain network underlie the impairment in social interaction involving dynamic facial expressions in individuals with ASD. PMID:22889284

  1. Cultural similarities and differences in perceiving and recognizing facial expressions of basic emotions.

    PubMed

    Yan, Xiaoqian; Andrews, Timothy J; Young, Andrew W

    2016-03-01

    The ability to recognize facial expressions of basic emotions is often considered a universal human ability. However, recent studies have suggested that this commonality has been overestimated and that people from different cultures use different facial signals to represent expressions (Jack, Blais, Scheepers, Schyns, & Caldara, 2009; Jack, Caldara, & Schyns, 2012). We investigated this possibility by examining similarities and differences in the perception and categorization of facial expressions between Chinese and white British participants using whole-face and partial-face images. Our results showed no cultural difference in the patterns of perceptual similarity of expressions from whole-face images. When categorizing the same expressions, however, both British and Chinese participants were slightly more accurate with whole-face images of their own ethnic group. To further investigate potential strategy differences, we repeated the perceptual similarity and categorization tasks with presentation of only the upper or lower half of each face. Again, the perceptual similarity of facial expressions was similar between Chinese and British participants for both the upper and lower face regions. However, participants were slightly better at categorizing facial expressions of their own ethnic group for the lower face regions, indicating that the way in which culture shapes the categorization of facial expressions is largely driven by differences in information decoding from this part of the face. (c) 2016 APA, all rights reserved).

  2. Deficits in facial affect recognition among antisocial populations: a meta-analysis.

    PubMed

    Marsh, Abigail A; Blair, R J R

    2008-01-01

    Individuals with disorders marked by antisocial behavior frequently show deficits in recognizing displays of facial affect. Antisociality may be associated with specific deficits in identifying fearful expressions, which would implicate dysfunction in neural structures that subserve fearful expression processing. A meta-analysis of 20 studies was conducted to assess: (a) if antisocial populations show any consistent deficits in recognizing six emotional expressions; (b) beyond any generalized impairment, whether specific fear recognition deficits are apparent; and (c) if deficits in fear recognition are a function of task difficulty. Results show a robust link between antisocial behavior and specific deficits in recognizing fearful expressions. This impairment cannot be attributed solely to task difficulty. These results suggest dysfunction among antisocial individuals in specified neural substrates, namely the amygdala, involved in processing fearful facial affect.

  3. The effects of a daily facial lotion containing vitamins B3 and E and provitamin B5 on the facial skin of Indian women: a randomized, double-blind trial.

    PubMed

    Jerajani, Hemangi R; Mizoguchi, Haruko; Li, James; Whittenbarger, Debora J; Marmor, Michael J

    2010-01-01

    The B vitamins niacinamide and panthenol have been shown to reduce many signs of skin aging, including hyperpigmentation and redness. To measure the facial skin effects in Indian women of the daily use of a lotion containing niacinamide, panthenol, and tocopherol acetate using quantitative image analysis. Adult women 30-60 years of age with epidermal hyperpigmentation were recruited in Mumbai and randomly assigned to apply a test or control lotion to the face daily for 10 weeks. Effects on skin tone were measured using an image capturing system and associated software. Skin texture was assessed by expert graders. Barrier function was evaluated by transepithelial water loss measurements. Subjects and evaluators were blinded to the product assignment. Of 246 women randomized to treatment, 207 (84%) completed the study. Women who used the test lotion experienced significantly reduced appearance of hyperpigmentation, improved skin tone evenness, appearance of lightening of skin, and positive effects on skin texture. Improvements versus control were seen as early as 6 weeks. The test lotion was well tolerated. The most common adverse event was a transient, mild burning sensation. Daily use of a facial lotion containing niacinamide, panthenol, and tocopheryl acetate improved skin tone and texture and was well tolerated in Indian women with facial signs of aging.

  4. Delayed presentation of traumatic facial nerve (CN VII) paralysis.

    PubMed

    Napoli, Anthony M; Panagos, Peter

    2005-11-01

    Facial nerve paralysis (Cranial Nerve VII, CN VII) can be a disfiguring disorder with profound impact upon the patient. The etiology of facial nerve paralysis may be congenital, iatrogenic, or result from neoplasm, infection, trauma, or toxic exposure. In the emergency department, the most common cause of unilateral facial paralysis is Bell's palsy, also known as idiopathic facial paralysis (IFP). We report a case of delayed presentation of unilateral facial nerve paralysis 3 days after sustaining a traumatic head injury. Re-evaluation and imaging of this patient revealed a full facial paralysis and temporal bone fracture extending into the facial canal. Because cranial nerve injuries occur in approximately 5-10% of head-injured patients, a good history and physical examination is important to differentiate IFP from another etiology. Newer generation high-resolution computed tomography (CT) scans are commonly demonstrating these fractures. An understanding of this complication, appropriate patient follow-up, and early involvement of the Otolaryngologist is important in management of these patients. The mechanism as well as the timing of facial nerve paralysis will determine the proper evaluation, consultation, and management for the patient. Patients with total or immediate paralysis as well as those with poorly prognostic audiogram results are good candidates for surgical repair.

  5. Brief Report: Representational Momentum for Dynamic Facial Expressions in Pervasive Developmental Disorder

    ERIC Educational Resources Information Center

    Uono, Shota; Sato, Wataru; Toichi, Motomi

    2010-01-01

    Individuals with pervasive developmental disorder (PDD) have difficulty with social communication via emotional facial expressions, but behavioral studies involving static images have reported inconsistent findings about emotion recognition. We investigated whether dynamic presentation of facial expression would enhance subjective perception of…

  6. Characterization and recognition of mixed emotional expressions in thermal face image

    NASA Astrophysics Data System (ADS)

    Saha, Priya; Bhattacharjee, Debotosh; De, Barin K.; Nasipuri, Mita

    2016-05-01

    Facial expressions in infrared imaging have been introduced to solve the problem of illumination, which is an integral constituent of visual imagery. The paper investigates facial skin temperature distribution on mixed thermal facial expressions of our created face database where six are basic expressions and rest 12 are a mixture of those basic expressions. Temperature analysis has been performed on three facial regions of interest (ROIs); periorbital, supraorbital and mouth. Temperature variability of the ROIs in different expressions has been measured using statistical parameters. The temperature variation measurement in ROIs of a particular expression corresponds to a vector, which is later used in recognition of mixed facial expressions. Investigations show that facial features in mixed facial expressions can be characterized by positive emotion induced facial features and negative emotion induced facial features. Supraorbital is a useful facial region that can differentiate basic expressions from mixed expressions. Analysis and interpretation of mixed expressions have been conducted with the help of box and whisker plot. Facial region containing mixture of two expressions is generally less temperature inducing than corresponding facial region containing basic expressions.

  7. Analysis of Facial Asymmetry in Deformational Plagiocephaly Using Three-Dimensional Computed Tomographic Review

    PubMed Central

    Moon, Il Yung; Oh, Kap Sung

    2014-01-01

    Background Infants with deformational plagiocephaly (DP) usually present with cranial vault deformities as well as facial asymmetry. The purpose of this study was to use three-dimensional anthropometric data to evaluate the influence of cranial deformities on facial asymmetry. Methods We analyzed three-dimensional computed tomography data for infants with DP (n=48) and without DP (n=30, control). Using 16 landmarks and 3 reference planes, 22 distance parameters and 2 angular parameters were compared. This cephalometric assessment focused on asymmetry of the orbits, nose, ears, maxilla, and mandible. We then assessed the correlation between 23 of the measurements and cranial vault asymmetry (CVA) for statistical significance using relative differences and correlation analysis. Results With the exception of few orbital asymmetry variables, most measurements indicated that the facial asymmetry was greater in infants with DP. Mandibular and nasal asymmetry was correlated highly with severity of CVA. Shortening of the ipsilateral mandibular body was particularly significant. There was no significant deformity in the maxilla or ear. Conclusion This study demonstrated that the cranial vault deformity in DP is associated with facial asymmetry. Compared with the control group, the infants with DP were found to have prominent asymmetry of the nose and mandible. PMID:28913202

  8. Treatment of Previously Treated Facial Capillary Malformations: Results of Single-Center Retrospective Objective 3-Dimensional Analysis of the Efficacy of Large Spot 532 nm Lasers.

    PubMed

    Kwiek, Bartłomiej; Ambroziak, Marcin; Osipowicz, Katarzyna; Kowalewski, Cezary; Rożalski, Michał

    2018-06-01

    Current treatment of facial capillary malformations (CM) has limited efficacy. To assess the efficacy of large spot 532 nm lasers for the treatment of previously treated facial CM with the use of 3-dimensional (3D) image analysis. Forty-three white patients aged 6 to 59 were included in this study. Patients had 3D photography performed before and after treatment with a 532 nm Nd:YAG laser with large spot and contact cooling. Objective analysis of percentage improvement based on 3D digital assessment of combined color and area improvement (global clearance effect [GCE]) were performed. The median maximal improvement achieved during the treatment (GCE) was 59.1%. The mean number of laser procedures required to achieve this improvement was 6.2 (range 1-16). Improvement of minimum 25% (GCE25) was achieved by 88.4% of patients, a minimum of 50% (GCE50) by 61.1%, a minimum of 75% (GCE75) by 25.6%, and a minimum of 90% (GCE90) by 4.6%. Patients previously treated with pulsed dye lasers had a significantly less response than those treated with other modalities (GCE 37.3% vs 61.8%, respectively). A large spot 532 nm laser is effective in previously treated patients with facial CM.

  9. Comparative analysis between scientific and the media communication following facial transplantation.

    PubMed

    Belanger, M; Harris, P G; Nikolis, A; Danino, A M

    2009-03-01

    Our aim was to analyze the communications about three outstanding medical reports. Was there any difference in the reports of the three allografts? Was there a correlation between the media and the scientific world? The Internet sites of three major newspapers were used for the media database. Those results were compared with PubMed between 2005 and 2007 using these key words: "facial graft," "facial allograft," "composite tissue allograft," and names of surgeons of the graft. We did a comparative analysis using a word processor and a quality analysis software. We analyzed 51 articles from the media and six from the PubMed database. In PubMed, 100% of the articles were on the first graft and respected the privacy of the patient compared to 67% of the media who unveiled the identity. The communication following a medical premiere depends on the team, which performes the act. We observed a major difference between the three cases. Ethical considerations are different for the media and for scientists. The communication management of a medical premiere takes preparation and evaluation.

  10. Spontaneous Facial Mimicry in Response to Dynamic Facial Expressions

    ERIC Educational Resources Information Center

    Sato, Wataru; Yoshikawa, Sakiko

    2007-01-01

    Based on previous neuroscientific evidence indicating activation of the mirror neuron system in response to dynamic facial actions, we hypothesized that facial mimicry would occur while subjects viewed dynamic facial expressions. To test this hypothesis, dynamic/static facial expressions of anger/happiness were presented using computer-morphing…

  11. Common and distinct neural correlates of facial emotion processing in social anxiety disorder and Williams syndrome: A systematic review and voxel-based meta-analysis of functional resonance imaging studies.

    PubMed

    Binelli, C; Subirà, S; Batalla, A; Muñiz, A; Sugranyés, G; Crippa, J A; Farré, M; Pérez-Jurado, L; Martín-Santos, R

    2014-11-01

    Social Anxiety Disorder (SAD) and Williams-Beuren Syndrome (WS) are two conditions which seem to be at opposite ends in the continuum of social fear but show compromised abilities in some overlapping areas, including some social interactions, gaze contact and processing of facial emotional cues. The increase in the number of neuroimaging studies has greatly expanded our knowledge of the neural bases of facial emotion processing in both conditions. However, to date, SAD and WS have not been compared. We conducted a systematic review of functional magnetic resonance imaging (fMRI) studies comparing SAD and WS cases to healthy control participants (HC) using facial emotion processing paradigms. Two researchers conducted comprehensive PubMed/Medline searches to identify all fMRI studies of facial emotion processing in SAD and WS. The following search key-words were used: "emotion processing"; "facial emotion"; "social anxiety"; "social phobia"; "Williams syndrome"; "neuroimaging"; "functional magnetic resonance"; "fMRI" and their combinations, as well as terms specifying individual facial emotions. We extracted spatial coordinates from each study and conducted two separate voxel-wise activation likelihood estimation meta-analyses, one for SAD and one for WS. Twenty-two studies met the inclusion criteria: 17 studies of SAD and five of WS. We found evidence for both common and distinct patterns of neural activation. Limbic engagement was common to SAD and WS during facial emotion processing, although we observed opposite patterns of activation for each disorder. Compared to HC, SAD cases showed hyperactivation of the amygdala, the parahippocampal gyrus and the globus pallidus. Compared to controls, participants with WS showed hypoactivation of these regions. Differential activation in a number of regions specific to either condition was also identified: SAD cases exhibited greater activation of the insula, putamen, the superior temporal gyrus, medial frontal regions and

  12. Electrical stimulation treatment for facial palsy after revision pleomorphic adenoma surgery

    PubMed Central

    Goldie, Simon; Sandeman, Jack; Cole, Richard; Dennis, Simon; Swain, Ian

    2016-01-01

    Surgery for pleomorphic adenoma recurrence presents a significant risk of facial nerve damage that can result in facial weakness effecting patients’ ability to communicate, mental health and self-image. We report two case studies that had marked facial weakness after resection of recurrent pleomorphic adenoma and their progress with electrical stimulation. Subjects received electrical stimulation twice daily for 24 weeks during which photographs of expressions, facial measurements and Sunnybrook scores were recorded. Both subjects recovered good facial function demonstrating Sunnybrook scores of 54 and 64 that improved to 88 and 96, respectively. Neither subjects demonstrated adverse effects of treatment. We conclude that electrical stimulation is a safe treatment and may improve facial palsy in patients after resection of recurrent pleomorphic adenoma. Larger studies would be difficult to pursue due to the low incidence of cases. PMID:27106613

  13. Dermatoscopic features of cutaneous non-facial non-acral lentiginous growth pattern melanomas

    PubMed Central

    Keir, Jeff

    2014-01-01

    Background: The dermatoscopic features of facial lentigo maligna (LM), facial lentigo maligna melanoma (LMM) and acral lentiginous melanoma (ALM) have been well described. This is the first description of the dermatoscopic appearance of a clinical series of cutaneous non-facial non-acral lentiginous growth pattern melanomas. Objective: To describe the dermatoscopic features of a series of cutaneous non-facial non-acral lentiginous growth pattern melanomas in an Australian skin cancer practice. Method: Single observer retrospective analysis of dermatoscopic images of a one-year series of cutaneous non-facial, non-acral melanomas reported as having a lentiginous growth pattern detected in an open access primary care skin cancer clinic in Australia. Lesions were scored for presence of classical criteria for facial LM; modified pattern analysis (“Chaos and Clues”) criteria; and the presence of two novel criteria: a lentigo-like pigment pattern lacking a lentigo-like border, and large polygons. Results: 20 melanomas occurring in 14 female and 6 male patients were included. Average patient age was 64 years (range: 44–83). Lesion distribution was: trunk 35%; upper limb 40%; and lower limb 25%. The incidences of criteria identified were: asymmetry of color or pattern (100%); lentigo-like pigment pattern lacking a lentigo-like border (90%); asymmetrically pigmented follicular openings (APFO’s) (70%); grey blue structures (70%); large polygons (45%); eccentric structureless area (15%); bright white lines (5%). 20% of the lesions had only the novel criteria and/or APFO’s. Limitations: Single observer, single center retrospective study. Conclusions: Cutaneous non-facial non-acral melanomas with a lentiginous growth pattern may have none or very few traditional criteria for the diagnosis of melanoma. Criteria that are logically expected in lesions with a lentiginous growth pattern (lentigo-like pigment pattern lacking a lentigo-like border, APFO’s) and the novel

  14. Dermatoscopic features of cutaneous non-facial non-acral lentiginous growth pattern melanomas.

    PubMed

    Keir, Jeff

    2014-01-01

    The dermatoscopic features of facial lentigo maligna (LM), facial lentigo maligna melanoma (LMM) and acral lentiginous melanoma (ALM) have been well described. This is the first description of the dermatoscopic appearance of a clinical series of cutaneous non-facial non-acral lentiginous growth pattern melanomas. To describe the dermatoscopic features of a series of cutaneous non-facial non-acral lentiginous growth pattern melanomas in an Australian skin cancer practice. Single observer retrospective analysis of dermatoscopic images of a one-year series of cutaneous non-facial, non-acral melanomas reported as having a lentiginous growth pattern detected in an open access primary care skin cancer clinic in Australia. Lesions were scored for presence of classical criteria for facial LM; modified pattern analysis ("Chaos and Clues") criteria; and the presence of two novel criteria: a lentigo-like pigment pattern lacking a lentigo-like border, and large polygons. 20 melanomas occurring in 14 female and 6 male patients were included. Average patient age was 64 years (range: 44-83). Lesion distribution was: trunk 35%; upper limb 40%; and lower limb 25%. The incidences of criteria identified were: asymmetry of color or pattern (100%); lentigo-like pigment pattern lacking a lentigo-like border (90%); asymmetrically pigmented follicular openings (APFO's) (70%); grey blue structures (70%); large polygons (45%); eccentric structureless area (15%); bright white lines (5%). 20% of the lesions had only the novel criteria and/or APFO's. Single observer, single center retrospective study. Cutaneous non-facial non-acral melanomas with a lentiginous growth pattern may have none or very few traditional criteria for the diagnosis of melanoma. Criteria that are logically expected in lesions with a lentiginous growth pattern (lentigo-like pigment pattern lacking a lentigo-like border, APFO's) and the novel criterion of large polygons may be useful in increasing sensitivity and

  15. Comparison of different methods for gender estimation from face image of various poses

    NASA Astrophysics Data System (ADS)

    Ishii, Yohei; Hongo, Hitoshi; Niwa, Yoshinori; Yamamoto, Kazuhiko

    2003-04-01

    Recently, gender estimation from face images has been studied for frontal facial images. However, it is difficult to obtain such facial images constantly in the case of application systems for security, surveillance and marketing research. In order to build such systems, a method is required to estimate gender from the image of various facial poses. In this paper, three different classifiers are compared in appearance-based gender estimation, which use four directional features (FDF). The classifiers are linear discriminant analysis (LDA), Support Vector Machines (SVMs) and Sparse Network of Winnows (SNoW). Face images used for experiments were obtained from 35 viewpoints. The direction of viewpoints varied +/-45 degrees horizontally, +/-30 degrees vertically at 15 degree intervals respectively. Although LDA showed the best performance for frontal facial images, SVM with Gaussian kernel was found the best performance (86.0%) for the facial images of 35 viewpoints. It is considered that SVM with Gaussian kernel is robust to changes in viewpoint when estimating gender from these results. Furthermore, the estimation rate was quite close to the average estimation rate at 35 viewpoints respectively. It is supposed that the methods are reasonable to estimate gender within the range of experimented viewpoints by learning face images from multiple directions by one class.

  16. Correlation between facial morphology and gene polymorphisms in the Uygur youth population.

    PubMed

    He, Huiyu; Mi, Xue; Zhang, Jiayu; Zhang, Qin; Yao, Yuan; Zhang, Xu; Xiao, Feng; Zhao, Chunping; Zheng, Shutao

    2017-04-25

    Human facial morphology varies considerably among individuals and can be influenced by gene polymorphisms. We explored the effects of single nucleotide polymorphisms (SNPs) on facial features in the Uygur youth population of the Kashi area in Xinjiang, China. Saliva samples were collected from 578 volunteers, and 10 SNPs previously associated with variations in facial physiognomy were genotyped. In parallel, 3D images of the subjects' faces were obtained using grating facial scanning technology. After delimitation of 15 salient landmarks, the correlation between SNPs and the distances between facial landmark pairs was assessed. Analysis of variance revealed that ENPP1 rs7754561 polymorphism was significantly associated with RAla-RLipCn and RLipCn-Sbn linear distances (p = 0.044 and p = 0.012, respectively) as well as RLipCn-Stm curve distance (p = 0.042). The GHR rs6180 polymorphism correlated with RLipCn-Stm linear distance (p = 0.04), while the GHR rs6184 polymorphism correlated with RLipCn-ULipP curve distance (p = 0.047). The FGFR1 rs4647905 polymorphism was associated with LLipCn-Nsn linear distance (p = 0.042). These results reveal that ENPP1 and FGFR1 influence lower anterior face height, the distance from the upper lip to the nasal floor, and lip shape. FGFR1 also influences the lower anterior face height, while GHR is associated with the length and width of the lip.

  17. MR relaxometry for the facial ageing assessment: the preliminary study of the age dependency in the MR relaxometry parameters within the facial soft tissue.

    PubMed

    Watanabe, M; Buch, K; Fujita, A; Christiansen, C L; Jara, H; Sakai, O

    2015-01-01

    To investigate the location-specific tissue properties and age-related changes of the facial fat and facial muscles using quantitative MRI (qMRI) analysis of longitudinal magnetization (T1) and transverse magnetization (T2) values. 38 subjects (20 males and 18 females, 0.5-87 years old) were imaged with a mixed turbo-spin echo sequence at 1.5 T. T1 and T2 measurements were obtained within regions of interest in six facial fat regions including the buccal fat and subcutaneous cheek fat, four eyelid fat regions (lateral upper, medial upper, lateral lower and medial lower) and five facial muscles including the orbicularis oculi, orbicularis oris, buccinator, zygomaticus major and masseter muscles bilaterally. Within the zygomaticus major muscle, age-associated T1 decreases in females and T1 increases in males were observed in later life with an increase in T2 values with age. The orbicularis oculi muscles showed lower T1 and higher T2 values compared to the masseter, orbicularis oris and buccinator muscles, which demonstrated small age-related changes. The dramatic age-related changes were also observed in the eyelid fat regions, particularly within the lower eyelid fat; negative correlations with age in T1 values (p<0.0001 for age) and prominent positive correlation in T2 values in male subjects (p<0.0001 for male×age). Age-related changes were not observed in T2 values within the subcutaneous cheek fat. This study demonstrates proof of concept using T1 and T2 values to assess age-related changes of the facial soft tissues, demonstrating tissue-specific qMRI measurements and non-uniform ageing patterns within different regions of facial soft tissues.

  18. Association of impaired facial affect recognition with basic facial and visual processing deficits in schizophrenia.

    PubMed

    Norton, Daniel; McBain, Ryan; Holt, Daphne J; Ongur, Dost; Chen, Yue

    2009-06-15

    Impaired emotion recognition has been reported in schizophrenia, yet the nature of this impairment is not completely understood. Recognition of facial emotion depends on processing affective and nonaffective facial signals, as well as basic visual attributes. We examined whether and how poor facial emotion recognition in schizophrenia is related to basic visual processing and nonaffective face recognition. Schizophrenia patients (n = 32) and healthy control subjects (n = 29) performed emotion discrimination, identity discrimination, and visual contrast detection tasks, where the emotionality, distinctiveness of identity, or visual contrast was systematically manipulated. Subjects determined which of two presentations in a trial contained the target: the emotional face for emotion discrimination, a specific individual for identity discrimination, and a sinusoidal grating for contrast detection. Patients had significantly higher thresholds (worse performance) than control subjects for discriminating both fearful and happy faces. Furthermore, patients' poor performance in fear discrimination was predicted by performance in visual detection and face identity discrimination. Schizophrenia patients require greater emotional signal strength to discriminate fearful or happy face images from neutral ones. Deficient emotion recognition in schizophrenia does not appear to be determined solely by affective processing but is also linked to the processing of basic visual and facial information.

  19. Fast 3D NIR systems for facial measurement and lip-reading

    NASA Astrophysics Data System (ADS)

    Brahm, Anika; Ramm, Roland; Heist, Stefan; Rulff, Christian; Kühmstedt, Peter; Notni, Gunther

    2017-05-01

    Structured-light projection is a well-established optical method for the non-destructive contactless three-dimensional (3D) measurement of object surfaces. In particular, there is a great demand for accurate and fast 3D scans of human faces or facial regions of interest in medicine, safety, face modeling, games, virtual life, or entertainment. New developments of facial expression detection and machine lip-reading can be used for communication tasks, future machine control, or human-machine interactions. In such cases, 3D information may offer more detailed information than 2D images which can help to increase the power of current facial analysis algorithms. In this contribution, we present new 3D sensor technologies based on three different methods of near-infrared projection technologies in combination with a stereo vision setup of two cameras. We explain the optical principles of an NIR GOBO projector, an array projector and a modified multi-aperture projection method and compare their performance parameters to each other. Further, we show some experimental measurement results of applications where we realized fast, accurate, and irritation-free measurements of human faces.

  20. The mysterious noh mask: contribution of multiple facial parts to the recognition of emotional expressions.

    PubMed

    Miyata, Hiromitsu; Nishimura, Ritsuko; Okanoya, Kazuo; Kawai, Nobuyuki

    2012-01-01

    A Noh mask worn by expert actors when performing on a Japanese traditional Noh drama is suggested to convey countless different facial expressions according to different angles of head/body orientation. The present study addressed the question of how different facial parts of a Noh mask, including the eyebrows, the eyes, and the mouth, may contribute to different emotional expressions. Both experimental situations of active creation and passive recognition of emotional facial expressions were introduced. In Experiment 1, participants either created happy or sad facial expressions, or imitated a face that looked up or down, by actively changing each facial part of a Noh mask image presented on a computer screen. For an upward tilted mask, the eyebrows and the mouth shared common features with sad expressions, whereas the eyes with happy expressions. This contingency tended to be reversed for a downward tilted mask. Experiment 2 further examined which facial parts of a Noh mask are crucial in determining emotional expressions. Participants were exposed to the synthesized Noh mask images with different facial parts expressing different emotions. Results clearly revealed that participants primarily used the shape of the mouth in judging emotions. The facial images having the mouth of an upward/downward tilted Noh mask strongly tended to be evaluated as sad/happy, respectively. The results suggest that Noh masks express chimeric emotional patterns, with different facial parts conveying different emotions This appears consistent with the principles of Noh which highly appreciate subtle and composite emotional expressions, as well as with the mysterious facial expressions observed in Western art. It was further demonstrated that the mouth serves as a diagnostic feature in characterizing the emotional expressions. This indicates the superiority of biologically-driven factors over the traditionally formulated performing styles when evaluating the emotions of the Noh masks.

  1. The Mysterious Noh Mask: Contribution of Multiple Facial Parts to the Recognition of Emotional Expressions

    PubMed Central

    Miyata, Hiromitsu; Nishimura, Ritsuko; Okanoya, Kazuo; Kawai, Nobuyuki

    2012-01-01

    Background A Noh mask worn by expert actors when performing on a Japanese traditional Noh drama is suggested to convey countless different facial expressions according to different angles of head/body orientation. The present study addressed the question of how different facial parts of a Noh mask, including the eyebrows, the eyes, and the mouth, may contribute to different emotional expressions. Both experimental situations of active creation and passive recognition of emotional facial expressions were introduced. Methodology/Principal Findings In Experiment 1, participants either created happy or sad facial expressions, or imitated a face that looked up or down, by actively changing each facial part of a Noh mask image presented on a computer screen. For an upward tilted mask, the eyebrows and the mouth shared common features with sad expressions, whereas the eyes with happy expressions. This contingency tended to be reversed for a downward tilted mask. Experiment 2 further examined which facial parts of a Noh mask are crucial in determining emotional expressions. Participants were exposed to the synthesized Noh mask images with different facial parts expressing different emotions. Results clearly revealed that participants primarily used the shape of the mouth in judging emotions. The facial images having the mouth of an upward/downward tilted Noh mask strongly tended to be evaluated as sad/happy, respectively. Conclusions/Significance The results suggest that Noh masks express chimeric emotional patterns, with different facial parts conveying different emotions This appears consistent with the principles of Noh which highly appreciate subtle and composite emotional expressions, as well as with the mysterious facial expressions observed in Western art. It was further demonstrated that the mouth serves as a diagnostic feature in characterizing the emotional expressions. This indicates the superiority of biologically-driven factors over the traditionally

  2. Mapping the impairment in decoding static facial expressions of emotion in prosopagnosia.

    PubMed

    Fiset, Daniel; Blais, Caroline; Royer, Jessica; Richoz, Anne-Raphaëlle; Dugas, Gabrielle; Caldara, Roberto

    2017-08-01

    Acquired prosopagnosia is characterized by a deficit in face recognition due to diverse brain lesions, but interestingly most prosopagnosic patients suffering from posterior lesions use the mouth instead of the eyes for face identification. Whether this bias is present for the recognition of facial expressions of emotion has not yet been addressed. We tested PS, a pure case of acquired prosopagnosia with bilateral occipitotemporal lesions anatomically sparing the regions dedicated for facial expression recognition. PS used mostly the mouth to recognize facial expressions even when the eye area was the most diagnostic. Moreover, PS directed most of her fixations towards the mouth. Her impairment was still largely present when she was instructed to look at the eyes, or when she was forced to look at them. Control participants showed a performance comparable to PS when only the lower part of the face was available. These observations suggest that the deficits observed in PS with static images are not solely attentional, but are rooted at the level of facial information use. This study corroborates neuroimaging findings suggesting that the Occipital Face Area might play a critical role in extracting facial features that are integrated for both face identification and facial expression recognition in static images. © The Author (2017). Published by Oxford University Press.

  3. Holistic face processing can inhibit recognition of forensic facial composites.

    PubMed

    McIntyre, Alex H; Hancock, Peter J B; Frowd, Charlie D; Langton, Stephen R H

    2016-04-01

    Facial composite systems help eyewitnesses to show the appearance of criminals. However, likenesses created by unfamiliar witnesses will not be completely accurate, and people familiar with the target can find them difficult to identify. Faces are processed holistically; we explore whether this impairs identification of inaccurate composite images and whether recognition can be improved. In Experiment 1 (n = 64) an imaging technique was used to make composites of celebrity faces more accurate and identification was contrasted with the original composite images. Corrected composites were better recognized, confirming that errors in production of the likenesses impair identification. The influence of holistic face processing was explored by misaligning the top and bottom parts of the composites (cf. Young, Hellawell, & Hay, 1987). Misalignment impaired recognition of corrected composites but identification of the original, inaccurate composites significantly improved. This effect was replicated with facial composites of noncelebrities in Experiment 2 (n = 57). We conclude that, like real faces, facial composites are processed holistically: recognition is impaired because unlike real faces, composites contain inaccuracies and holistic face processing makes it difficult to perceive identifiable features. This effect was consistent across composites of celebrities and composites of people who are personally familiar. Our findings suggest that identification of forensic facial composites can be enhanced by presenting composites in a misaligned format. (c) 2016 APA, all rights reserved).

  4. Estimation of 2D to 3D dimensions and proportionality indices for facial examination.

    PubMed

    Martos, Rubén; Valsecchi, Andrea; Ibáñez, Oscar; Alemán, Inmaculada

    2018-06-01

    Photo-anthropometry is a metric-based facial image comparison technique where measurements of the face are taken from an image using predetermined facial landmarks. In particular, dimensions and proportionality indices (DPIs) are compared to DPIs from another facial image. Different studies concluded that photo-anthropometric facial comparison, as it is currently practiced, is unsuitable for elimination purposes. The major limitation is the need for images acquired under very restrictive, controlled conditions. To overcome this latter issue, we propose a novel methodology to estimate 3D DPIs from 2D ones. It uses computer graphic techniques to simulate thousands of facial photographs under known camera conditions and regression to derive the mathematical relationship between 2D and 3D DPIs automatically. Additionally, we present a methodology that makes use of the estimated 3D DPIs for reducing the number of potential matches of a given unknown facial photograph within a set of known candidates. The error in the estimation of the 3D DPIs can be as large as 35%, but both I and III quartiles are consistently inside the ±5% range. The methodology for filtering cases has demonstrated to be useful in the task of narrowing down the list of possible candidates for a given photograph. It is able to remove on average (validated using cross-validation technique) 57% and 24% of the negative cases, depending on the amounts of DPIs available. Limitations of the work developed together with open research lines are included within the Discussion section. Copyright © 2018 Elsevier B.V. All rights reserved.

  5. The role of great auricular-facial nerve neurorrhaphy in facial nerve damage.

    PubMed

    Sun, Yan; Liu, Limei; Han, Yuechen; Xu, Lei; Zhang, Daogong; Wang, Haibo

    2015-01-01

    Facial nerve is easy to be damaged, and there are many reconstructive methods for facial nerve reconstructive, such as facial nerve end to end anastomosis, the great auricular nerve graft, the sural nerve graft, or hypoglossal-facial nerve anastomosis. However, there is still little study about great auricular-facial nerve neurorrhaphy. The aim of the present study was to identify the role of great auricular-facial nerve neurorrhaphy and the mechanism. Rat models of facial nerve cut (FC), facial nerve end to end anastomosis (FF), facial-great auricular neurorrhaphy (FG), and control (Ctrl) were established. Apex nasi amesiality observation, electrophysiology and immunofluorescence assays were employed to investigate the function and mechanism. In apex nasi amesiality observation, it was found apex nasi amesiality of FG group was partly recovered. Additionally, electrophysiology and immunofluorescence assays revealed that facial-great auricular neurorrhaphy could transfer nerve impulse and express AChR which was better than facial nerve cut and worse than facial nerve end to end anastomosis. The present study indicated that great auricular-facial nerve neurorrhaphy is a substantial solution for facial lesion repair, as it is efficiently preventing facial muscles atrophy by generating neurotransmitter like ACh.

  6. Extracranial Facial Nerve Schwannoma Treated by Hypo-fractionated CyberKnife Radiosurgery.

    PubMed

    Sasaki, Ayaka; Miyazaki, Shinichiro; Hori, Tomokatsu

    2016-09-21

    Facial nerve schwannoma is a rare intracranial tumor. Treatment for this benign tumor has been controversial. Here, we report a case of extracranial facial nerve schwannoma treated successfully by hypo-fractionated CyberKnife (Accuray, Sunnyvale, CA) radiosurgery and discuss the efficacy of this treatment. A 34-year-old female noticed a swelling in her right mastoid process. The lesion enlarged over a seven-month period, and she experienced facial spasm on the right side. She was diagnosed with a facial schwannoma via a magnetic resonance imaging (MRI) scan of the head and neck and was told to wait until the facial nerve palsy subsides. She was referred to our hospital for radiation therapy. We planned a fractionated CyberKnife radiosurgery for three consecutive days. After CyberKnife radiosurgery, the mass in the right parotid gradually decreased in size, and the facial nerve palsy disappeared. At her eight-month follow-up, her facial spasm had completely disappeared. There has been no recurrence and the facial nerve function has been normal. We successfully demonstrated the efficacy of CyberKnife radiosurgery as an alternative treatment that also preserves neurofunction for facial nerve schwannomas.

  7. Utilization of Facial Image Analysis Technology for Blink Detection: A Validation Study.

    PubMed

    Kitazawa, Momoko; Yoshimura, Michitaka; Liang, Kuo-Ching; Wada, Satoshi; Mimura, Masaru; Tsubota, Kazuo; Kishimoto, Taishiro

    2018-06-25

    The assessment of anterior eye diseases and the understanding of psychological functions of blinking can benefit greatly from a validated blinking detection technology. In this work, we proposed an algorithm based on facial recognition built on current video processing technologies to automatically filter and analyze blinking movements. We compared electrooculography (EOG), the gold standard of blinking measurement, with manual video tape recording counting (mVTRc) and our proposed automated video tape recording analysis (aVTRa) in both static and dynamic conditions to validate our aVTRa method. We measured blinking in both static condition, where the subject was sitting still with chin fixed on the table, and dynamic condition, where the subject's face was not fixed and natural communication was taking place between the subject and interviewer. We defined concordance of blinks between measurement methods as having less than 50 ms difference between eyes opening and closing. The subjects consisted of seven healthy Japanese volunteers (3 male, four female) without significant eye disease with average age of 31.4±7.2. The concordance of EOG vs. aVTRa, EOG vs. mVTRc, and aVTRa vs. mVTRc (average±SD) were found to be 92.2±10.8%, 85.0±16.5%, and 99.6±1.0% in static conditions and 32.6±31.0%, 28.0±24.2%, and 98.5±2.7% in dynamic conditions, respectively. In static conditions, we have found a high blink concordance rate between the proposed aVTRa versus EOG, and confirmed the validity of aVTRa in both static and dynamic conditions.

  8. Facial approximation-from facial reconstruction synonym to face prediction paradigm.

    PubMed

    Stephan, Carl N

    2015-05-01

    Facial approximation was first proposed as a synonym for facial reconstruction in 1987 due to dissatisfaction with the connotations the latter label held. Since its debut, facial approximation's identity has morphed as anomalies in face prediction have accumulated. Now underpinned by differences in what problems are thought to count as legitimate, facial approximation can no longer be considered a synonym for, or subclass of, facial reconstruction. Instead, two competing paradigms of face prediction have emerged, namely: facial approximation and facial reconstruction. This paper shines a Kuhnian lens across the discipline of face prediction to comprehensively review these developments and outlines the distinguishing features between the two paradigms. © 2015 American Academy of Forensic Sciences.

  9. Biometric identification based on novel frequency domain facial asymmetry measures

    NASA Astrophysics Data System (ADS)

    Mitra, Sinjini; Savvides, Marios; Vijaya Kumar, B. V. K.

    2005-03-01

    In the modern world, the ever-growing need to ensure a system's security has spurred the growth of the newly emerging technology of biometric identification. The present paper introduces a novel set of facial biometrics based on quantified facial asymmetry measures in the frequency domain. In particular, we show that these biometrics work well for face images showing expression variations and have the potential to do so in presence of illumination variations as well. A comparison of the recognition rates with those obtained from spatial domain asymmetry measures based on raw intensity values suggests that the frequency domain representation is more robust to intra-personal distortions and is a novel approach for performing biometric identification. In addition, some feature analysis based on statistical methods comparing the asymmetry measures across different individuals and across different expressions is presented.

  10. The effect of width of facial canal in patients with idiopathic peripheral facial paralysis on the development of paralysis.

    PubMed

    Eksi, Guldem; Akbay, Ercan; Bayarogullari, Hanifi; Cevik, Cengiz; Yengil, Erhan; Ozler, Gul Soylu

    2015-09-01

    The aim of this prospective study is to investigate whether the possible stenosis due to anatomic variations of labyrinthine segment (LS), tympanic segment (TS) and mastoid segment (MS) of the facial canal in the temporal bone is a predisposing factor in the development of paralysis. 22 patients with idiopathic peripheral facial paralysis (IPFP) were included in the study. Multi-slice computed tomography (MSCT) with 64 detectors was used for temporal bone imaging of the patients. Reconstruction images in axial, coronal and sagittal planes were created in workstation computers from the captured images. The diameters and lengths of LS, TS and MS of the facial canal were measured. The mean values of LD, ND and SL of LS were 1.31 ± 0.39, 0.91 ± 0.27, 4.17 ± 0.48 in patient group and 1.26 ± 0.29, 0.95 ± 0.21, 4.60 ± 1.36 in control group, respectively. The mean values of LD, ND and SL of TS were 1.11 ± 0.22, 0.90 ± 0.14, 12.63 ± 1.47 in patient group and 1.17 ± 0.23, 0.85 ± 0.24, 12.10 ± 1.79 in control group, respectively. The mean values of LD, ND and SL of MS were 1.80 ± 0.30, 1.44 ± 0.29 vs. 14.3 ± 1.90 in patient group 1.74 ± 0.38, 1.40 ± 0.29, 14.15 ± 2.16 in control group, respectively. The measurements of the parameters of all three segments in patient group and control group were similar. Similar results between patient and control group were obtained in this study investigating the effect of stenosis in facial canal in the development of IPFP.

  11. Human Facial Expressions as Adaptations:Evolutionary Questions in Facial Expression Research

    PubMed Central

    SCHMIDT, KAREN L.; COHN, JEFFREY F.

    2007-01-01

    The importance of the face in social interaction and social intelligence is widely recognized in anthropology. Yet the adaptive functions of human facial expression remain largely unknown. An evolutionary model of human facial expression as behavioral adaptation can be constructed, given the current knowledge of the phenotypic variation, ecological contexts, and fitness consequences of facial behavior. Studies of facial expression are available, but results are not typically framed in an evolutionary perspective. This review identifies the relevant physical phenomena of facial expression and integrates the study of this behavior with the anthropological study of communication and sociality in general. Anthropological issues with relevance to the evolutionary study of facial expression include: facial expressions as coordinated, stereotyped behavioral phenotypes, the unique contexts and functions of different facial expressions, the relationship of facial expression to speech, the value of facial expressions as signals, and the relationship of facial expression to social intelligence in humans and in nonhuman primates. Human smiling is used as an example of adaptation, and testable hypotheses concerning the human smile, as well as other expressions, are proposed. PMID:11786989

  12. Forming impressions: effects of facial expression and gender stereotypes.

    PubMed

    Hack, Tay

    2014-04-01

    The present study of 138 participants explored how facial expressions and gender stereotypes influence impressions. It was predicted that images of smiling women would be evaluated more favorably on traits reflecting warmth, and that images of non-smiling men would be evaluated more favorably on traits reflecting competence. As predicted, smiling female faces were rated as more warm; however, contrary to prediction, perceived competence of male faces was not affected by facial expression. Participants' female stereotype endorsement was a significant predictor for evaluations of female faces; those who ascribed more strongly to traditional female stereotypes reported the most positive impressions of female faces displaying a smiling expression. However, a similar effect was not found for images of men; endorsement of traditional male stereotypes did not predict participants' impressions of male faces.

  13. Botulinum toxin treatment for facial palsy: A systematic review.

    PubMed

    Cooper, Lilli; Lui, Michael; Nduka, Charles

    2017-06-01

    Facial palsy may be complicated by ipsilateral synkinesis or contralateral hyperkinesis. Botulinum toxin is increasingly used in the management of facial palsy; however, the optimum dose, treatment interval, adjunct therapy and performance as compared with alternative treatments have not been well established. This study aimed to systematically review the evidence for the use of botulinum toxin in facial palsy. The Cochrane central register of controlled trials (CENTRAL), MEDLINE(R) (1946 to September 2015) and Embase Classic + Embase (1947 to September 2015) were searched for randomised studies using botulinum toxin in facial palsy. Forty-seven studies were identified, and three included. Their physical and patient-reported outcomes are described, and observations and cautions are discussed. Facial asymmetry has a strong correlation to subjective domains such as impairment in social interaction and perception of self-image and appearance. Botulinum toxin injections represent a minimally invasive technique that is helpful in restoring facial symmetry at rest and during movement in chronic, and potentially acute, facial palsy. Botulinum toxin in combination with physical therapy may be particularly helpful. Currently, there is a paucity of data; areas for further research are suggested. A strong body of evidence may allow botulinum toxin treatment to be nationally standardised and recommended in the management of facial palsy. Copyright © 2017 British Association of Plastic, Reconstructive and Aesthetic Surgeons. Published by Elsevier Ltd. All rights reserved.

  14. 3D Face Model Dataset: Automatic Detection of Facial Expressions and Emotions for Educational Environments

    ERIC Educational Resources Information Center

    Chickerur, Satyadhyan; Joshi, Kartik

    2015-01-01

    Emotion detection using facial images is a technique that researchers have been using for the last two decades to try to analyze a person's emotional state given his/her image. Detection of various kinds of emotion using facial expressions of students in educational environment is useful in providing insight into the effectiveness of tutoring…

  15. Electrical stimulation treatment for facial palsy after revision pleomorphic adenoma surgery.

    PubMed

    Goldie, Simon; Sandeman, Jack; Cole, Richard; Dennis, Simon; Swain, Ian

    2016-04-22

    Surgery for pleomorphic adenoma recurrence presents a significant risk of facial nerve damage that can result in facial weakness effecting patients' ability to communicate, mental health and self-image. We report two case studies that had marked facial weakness after resection of recurrent pleomorphic adenoma and their progress with electrical stimulation. Subjects received electrical stimulation twice daily for 24 weeks during which photographs of expressions, facial measurements and Sunnybrook scores were recorded. Both subjects recovered good facial function demonstrating Sunnybrook scores of 54 and 64 that improved to 88 and 96, respectively. Neither subjects demonstrated adverse effects of treatment. We conclude that electrical stimulation is a safe treatment and may improve facial palsy in patients after resection of recurrent pleomorphic adenoma. Larger studies would be difficult to pursue due to the low incidence of cases. Published by Oxford University Press and JSCR Publishing Ltd. All rights reserved. © The Author 2016.

  16. Effective connectivity of facial expression network by using Granger causality analysis

    NASA Astrophysics Data System (ADS)

    Zhang, Hui; Li, Xiaoting

    2013-10-01

    Functional magnetic resonance imaging (fMRI) is an advanced non-invasive data acquisition technique to investigate the neural activity in human brain. In addition to localize the functional brain regions that is activated by specific cognitive task, fMRI can also be utilized to measure the task-related functional interactions among the active regions of interest (ROI) in the brain. Among the variety of analysis tools proposed for modeling the connectivity of brain regions, Granger causality analysis (GCA) measure the directions of information interactions by looking for the lagged effect among the brain regions. In this study, we use fMRI and Granger Causality analysis to investigate the effective connectivity of brain network induced by viewing several kinds of expressional faces. We focus on four kinds of facial expression stimuli: fearful, angry, happy and neutral faces. Five face selective regions of interest are localized and the effective connectivity within these regions is measured for the expressional faces. Our result based on 8 subjects showed that there is significant effective connectivity from STS to amygdala, from amygdala to OFA, aFFA and pFFA, from STS to aFFA and from pFFA to aFFA. This result suggested that there is an information flow from the STS to the amygdala when perusing expressional faces. This emotional expressional information flow that is conveyed by STS and amygdala, flow back to the face selective regions in occipital-temporal lobes, which constructed a emotional face processing network.

  17. Realistic facial animation generation based on facial expression mapping

    NASA Astrophysics Data System (ADS)

    Yu, Hui; Garrod, Oliver; Jack, Rachael; Schyns, Philippe

    2014-01-01

    Facial expressions reflect internal emotional states of a character or in response to social communications. Though much effort has been taken to generate realistic facial expressions, it still remains a challenging topic due to human being's sensitivity to subtle facial movements. In this paper, we present a method for facial animation generation, which reflects true facial muscle movements with high fidelity. An intermediate model space is introduced to transfer captured static AU peak frames based on FACS to the conformed target face. And then dynamic parameters derived using a psychophysics method is integrated to generate facial animation, which is assumed to represent natural correlation of multiple AUs. Finally, the animation sequence in the intermediate model space is mapped to the target face to produce final animation.

  18. The role of great auricular-facial nerve neurorrhaphy in facial nerve damage

    PubMed Central

    Sun, Yan; Liu, Limei; Han, Yuechen; Xu, Lei; Zhang, Daogong; Wang, Haibo

    2015-01-01

    Background: Facial nerve is easy to be damaged, and there are many reconstructive methods for facial nerve reconstructive, such as facial nerve end to end anastomosis, the great auricular nerve graft, the sural nerve graft, or hypoglossal-facial nerve anastomosis. However, there is still little study about great auricular-facial nerve neurorrhaphy. The aim of the present study was to identify the role of great auricular-facial nerve neurorrhaphy and the mechanism. Methods: Rat models of facial nerve cut (FC), facial nerve end to end anastomosis (FF), facial-great auricular neurorrhaphy (FG), and control (Ctrl) were established. Apex nasi amesiality observation, electrophysiology and immunofluorescence assays were employed to investigate the function and mechanism. Results: In apex nasi amesiality observation, it was found apex nasi amesiality of FG group was partly recovered. Additionally, electrophysiology and immunofluorescence assays revealed that facial-great auricular neurorrhaphy could transfer nerve impulse and express AChR which was better than facial nerve cut and worse than facial nerve end to end anastomosis. Conclusions: The present study indicated that great auricular-facial nerve neurorrhaphy is a substantial solution for facial lesion repair, as it is efficiently preventing facial muscles atrophy by generating neurotransmitter like ACh. PMID:26550216

  19. A Genome-Wide Association Study Identifies Five Loci Influencing Facial Morphology in Europeans

    PubMed Central

    Liu, Fan; van der Lijn, Fedde; Schurmann, Claudia; Zhu, Gu; Chakravarty, M. Mallar; Hysi, Pirro G.; Wollstein, Andreas; Lao, Oscar; de Bruijne, Marleen; Ikram, M. Arfan; van der Lugt, Aad; Rivadeneira, Fernando; Uitterlinden, André G.; Hofman, Albert; Niessen, Wiro J.; Homuth, Georg; de Zubicaray, Greig; McMahon, Katie L.; Thompson, Paul M.; Daboul, Amro; Puls, Ralf; Hegenscheid, Katrin; Bevan, Liisa; Pausova, Zdenka; Medland, Sarah E.; Montgomery, Grant W.; Wright, Margaret J.; Wicking, Carol; Boehringer, Stefan; Spector, Timothy D.; Paus, Tomáš; Martin, Nicholas G.; Biffar, Reiner; Kayser, Manfred

    2012-01-01

    Inter-individual variation in facial shape is one of the most noticeable phenotypes in humans, and it is clearly under genetic regulation; however, almost nothing is known about the genetic basis of normal human facial morphology. We therefore conducted a genome-wide association study for facial shape phenotypes in multiple discovery and replication cohorts, considering almost ten thousand individuals of European descent from several countries. Phenotyping of facial shape features was based on landmark data obtained from three-dimensional head magnetic resonance images (MRIs) and two-dimensional portrait images. We identified five independent genetic loci associated with different facial phenotypes, suggesting the involvement of five candidate genes—PRDM16, PAX3, TP63, C5orf50, and COL17A1—in the determination of the human face. Three of them have been implicated previously in vertebrate craniofacial development and disease, and the remaining two genes potentially represent novel players in the molecular networks governing facial development. Our finding at PAX3 influencing the position of the nasion replicates a recent GWAS of facial features. In addition to the reported GWA findings, we established links between common DNA variants previously associated with NSCL/P at 2p21, 8q24, 13q31, and 17q22 and normal facial-shape variations based on a candidate gene approach. Overall our study implies that DNA variants in genes essential for craniofacial development contribute with relatively small effect size to the spectrum of normal variation in human facial morphology. This observation has important consequences for future studies aiming to identify more genes involved in the human facial morphology, as well as for potential applications of DNA prediction of facial shape such as in future forensic applications. PMID:23028347

  20. [Presurgical orthodontics for facial asymmetry].

    PubMed

    Labarrère, H

    2003-03-01

    As with the treatment of all facial deformities, orthodontic pre-surgical preparation for facial asymmetry should aim at correcting severe occlusal discrepancies not solely on the basis of a narrow occlusal analysis but also in a way that will not disturb the proposed surgical protocol. In addition, facial asymmetries require specific adjustments, difficult to derive and to apply because of their inherent atypical morphological orientation of both alveolar and basal bony support. Three treated cases illustrate different solutions to problems posed by pathological torque: this torque must be considered with respect to proposed surgical changes, within the framework of their limitations and their possible contra-indications.

  1. Multilevel analysis of facial expressions of emotion and script: self-report (arousal and valence) and psychophysiological correlates.

    PubMed

    Balconi, Michela; Vanutelli, Maria Elide; Finocchiaro, Roberta

    2014-09-26

    The paper explored emotion comprehension in children with regard to facial expression of emotion. The effect of valence and arousal evaluation, of context and of psychophysiological measures was monitored. Indeed subjective evaluation of valence (positive vs. negative) and arousal (high vs. low), and contextual (facial expression vs. facial expression and script) variables were supposed to modulate the psychophysiological responses. Self-report measures (in terms of correct recognition, arousal and valence attribution) and psychophysiological correlates (facial electromyography, EMG, skin conductance response, SCR, and heart rate, HR) were observed when children (N = 26; mean age = 8.75 y; range 6-11 y) looked at six facial expressions of emotions (happiness, anger, fear, sadness, surprise, and disgust) and six emotional scripts (contextualized facial expressions). The competencies about the recognition, the evaluation on valence and arousal was tested in concomitance with psychophysiological variations. Specifically, we tested for the congruence of these multiple measures. Log-linear analysis and repeated measure ANOVAs showed different representations across the subjects, as a function of emotion. Specifically, children' recognition and attribution were well developed for some emotions (such as anger, fear, surprise and happiness), whereas some other emotions (mainly disgust and sadness) were less clearly represented. SCR, HR and EMG measures were modulated by the evaluation based on valence and arousal, with increased psychophysiological values mainly in response to anger, fear and happiness. As shown by multiple regression analysis, a significant consonance was found between self-report measures and psychophysiological behavior, mainly for emotions rated as more arousing and negative in valence. The multilevel measures were discussed at light of dimensional attribution model.

  2. Multilevel analysis of facial expressions of emotion and script: self-report (arousal and valence) and psychophysiological correlates

    PubMed Central

    2014-01-01

    Background The paper explored emotion comprehension in children with regard to facial expression of emotion. The effect of valence and arousal evaluation, of context and of psychophysiological measures was monitored. Indeed subjective evaluation of valence (positive vs. negative) and arousal (high vs. low), and contextual (facial expression vs. facial expression and script) variables were supposed to modulate the psychophysiological responses. Methods Self-report measures (in terms of correct recognition, arousal and valence attribution) and psychophysiological correlates (facial electromyography, EMG, skin conductance response, SCR, and heart rate, HR) were observed when children (N = 26; mean age = 8.75 y; range 6-11 y) looked at six facial expressions of emotions (happiness, anger, fear, sadness, surprise, and disgust) and six emotional scripts (contextualized facial expressions). The competencies about the recognition, the evaluation on valence and arousal was tested in concomitance with psychophysiological variations. Specifically, we tested for the congruence of these multiple measures. Results Log-linear analysis and repeated measure ANOVAs showed different representations across the subjects, as a function of emotion. Specifically, children’ recognition and attribution were well developed for some emotions (such as anger, fear, surprise and happiness), whereas some other emotions (mainly disgust and sadness) were less clearly represented. SCR, HR and EMG measures were modulated by the evaluation based on valence and arousal, with increased psychophysiological values mainly in response to anger, fear and happiness. Conclusions As shown by multiple regression analysis, a significant consonance was found between self-report measures and psychophysiological behavior, mainly for emotions rated as more arousing and negative in valence. The multilevel measures were discussed at light of dimensional attribution model. PMID:25261242

  3. CT detection of facial canal dehiscence and semicircular canal fistula: Comparison with surgical findings

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Fuse, Takeo; Tada, Yuichiro; Aoyagi, Masaru

    1996-03-01

    The purpose of this study was to determine the accuracy of high resolution CT (HRCT) in the detection of facial canal dehiscence and semicircular canal fistula, the preoperative evaluation of both of which is clinically very important for ear surgery. We retrospectively reviewed the HRCT findings in 61 patients who underwent mastoidectomy at Yamagata University between 1989 and 1993. The HRCT images were obtained in the axial and semicoronal planes using 1 mm slice thickness and 1 mm intersection gap. In 46 (75%) of the 61 patients, the HRCT image-based assessment of the facial canal dehiscence coincided with the surgicalmore » findings. The data for the facial canal revealed sensitivity of 66% and specificity of 84%. For semicircular canal fistula. in 59 (97%) of the 61 patients, the HRCT image-based assessment and the surgical findings coincided. The image-based assessment in the remaining two patients, who both had massive cholesteatoma, was false-positive. HRCT is useful in the diagnosis of facial canal dehiscence and labyrinthine fistula, but its limitations should also be recognized. 12 refs., 3 figs., 6 tabs.« less

  4. Dynamic texture recognition using local binary patterns with an application to facial expressions.

    PubMed

    Zhao, Guoying; Pietikäinen, Matti

    2007-06-01

    Dynamic texture (DT) is an extension of texture to the temporal domain. Description and recognition of DTs have attracted growing attention. In this paper, a novel approach for recognizing DTs is proposed and its simplifications and extensions to facial image analysis are also considered. First, the textures are modeled with volume local binary patterns (VLBP), which are an extension of the LBP operator widely used in ordinary texture analysis, combining motion and appearance. To make the approach computationally simple and easy to extend, only the co-occurrences of the local binary patterns on three orthogonal planes (LBP-TOP) are then considered. A block-based method is also proposed to deal with specific dynamic events such as facial expressions in which local information and its spatial locations should also be taken into account. In experiments with two DT databases, DynTex and Massachusetts Institute of Technology (MIT), both the VLBP and LBP-TOP clearly outperformed the earlier approaches. The proposed block-based method was evaluated with the Cohn-Kanade facial expression database with excellent results. The advantages of our approach include local processing, robustness to monotonic gray-scale changes, and simple computation.

  5. Facial soft tissue thickness in skeletal type I Japanese children.

    PubMed

    Utsuno, Hajime; Kageyama, Toru; Deguchi, Toshio; Umemura, Yasunobu; Yoshino, Mineo; Nakamura, Hiroshi; Miyazawa, Hiroo; Inoue, Katsuhiro

    2007-10-25

    Facial reconstruction techniques used in forensic anthropology require knowledge of the facial soft tissue thickness of each race if facial features are to be reconstructed correctly. If this is inaccurate, so also will be the reconstructed face. Knowledge of differences by age and sex are also required. Therefore, when unknown human skeletal remains are found, the forensic anthropologist investigates for race, sex, and age, and for other variables of relevance. Cephalometric X-ray images of living persons can help to provide this information. They give an approximately 10% enlargement from true size and can demonstrate the relationship between soft and hard tissue. In the present study, facial soft tissue thickness in Japanese children was measured at 12 anthropological points using X-ray cephalometry in order to establish a database for facial soft tissue thickness. This study of both boys and girls, aged from 6 to 18 years, follows a previous study of Japanese female children only, and focuses on facial soft tissue thickness in only one skeletal type. Sex differences in thickness of tissue were found from 12 years of age upwards. The study provides more detailed and accurate measurements than past reports of facial soft tissue thickness, and reveals the uniqueness of the Japanese child's facial profile.

  6. Preoperative Identification of Facial Nerve in Vestibular Schwannomas Surgery Using Diffusion Tensor Tractography

    PubMed Central

    Choi, Kyung-Sik; Kim, Min-Su; Kwon, Hyeok-Gyu; Jang, Sung-Ho

    2014-01-01

    Objective Facial nerve palsy is a common complication of treatment for vestibular schwannoma (VS), so preserving facial nerve function is important. The preoperative visualization of the course of facial nerve in relation to VS could help prevent injury to the nerve during the surgery. In this study, we evaluate the accuracy of diffusion tensor tractography (DTT) for preoperative identification of facial nerve. Methods We prospectively collected data from 11 patients with VS, who underwent preoperative DTT for facial nerve. Imaging results were correlated with intraoperative findings. Postoperative DTT was performed at postoperative 3 month. Facial nerve function was clinically evaluated according to the House-Brackmann (HB) facial nerve grading system. Results Facial nerve courses on preoperative tractography were entirely correlated with intraoperative findings in all patients. Facial nerve was located on the anterior of the tumor surface in 5 cases, on anteroinferior in 3 cases, on anterosuperior in 2 cases, and on posteroinferior in 1 case. In postoperative facial nerve tractography, preservation of facial nerve was confirmed in all patients. No patient had severe facial paralysis at postoperative one year. Conclusion This study shows that DTT for preoperative identification of facial nerve in VS surgery could be a very accurate and useful radiological method and could help to improve facial nerve preservation. PMID:25289119

  7. Effect of facial neuromuscular re-education on facial symmetry in patients with Bell's palsy: a randomized controlled trial.

    PubMed

    Manikandan, N

    2007-04-01

    To determine the effect of facial neuromuscular re-education over conventional therapeutic measures in improving facial symmetry in patients with Bell's palsy. Randomized controlled trial. Neurorehabilitation unit. Fifty-nine patients diagnosed with Bell's palsy were included in the study after they met the inclusion criteria. Patients were randomly divided into two groups: control (n = 30) and experimental (n = 29). Control group patients received conventional therapeutic measures while the facial neuromuscular re-education group patients received techniques that were tailored to each patient in three sessions per day for six days per week for a period of two weeks. All the patients were evaluated using a Facial Grading Scale before treatment and after three months. The Facial Grading Scale scores showed significant improvement in both control (mean 32 (range 9.7-54) to 54.5 (42.2-71.7)) and the experimental (33 (18-43.5) to 66 (54-76.7)) group. Facial Grading Scale change scores showed that experimental group (27.5 (20-43.77)) improved significantly more than the control group (16.5 (12.2-24.7)). Analysis of Facial Grading Scale subcomponents did not show statistical significance, except in the movement score (12 (8-16) to 24 (12-18)). Individualized facial neuromuscular re-education is more effective in improving facial symmetry in patients with Bell's palsy than conventional therapeutic measures.

  8. Parameterized Facial Expression Synthesis Based on MPEG-4

    NASA Astrophysics Data System (ADS)

    Raouzaiou, Amaryllis; Tsapatsoulis, Nicolas; Karpouzis, Kostas; Kollias, Stefanos

    2002-12-01

    In the framework of MPEG-4, one can include applications where virtual agents, utilizing both textual and multisensory data, including facial expressions and nonverbal speech help systems become accustomed to the actual feelings of the user. Applications of this technology are expected in educational environments, virtual collaborative workplaces, communities, and interactive entertainment. Facial animation has gained much interest within the MPEG-4 framework; with implementation details being an open research area (Tekalp, 1999). In this paper, we describe a method for enriching human computer interaction, focusing on analysis and synthesis of primary and intermediate facial expressions (Ekman and Friesen (1978)). To achieve this goal, we utilize facial animation parameters (FAPs) to model primary expressions and describe a rule-based technique for handling intermediate ones. A relation between FAPs and the activation parameter proposed in classical psychological studies is established, leading to parameterized facial expression analysis and synthesis notions, compatible with the MPEG-4 standard.

  9. Extra-facial melasma: clinical, histopathological, and immunohistochemical case-control study.

    PubMed

    Ritter, C G; Fiss, D V C; Borges da Costa, J A T; de Carvalho, R R; Bauermann, G; Cestari, T F

    2013-09-01

    Extra-facial melasma is a prevalent dermatosis in some populations with special characteristics in relation to its clinical aspects and probable etiopathogenic factors. Few studies have attempted to address this alteration of pigmentation, which has become a challenge in clinical Dermatology. To assess the clinical histopathological and immunohistochemical characteristics of extra-facial melasma, comparing affected, and unaffected sites. Case-control study with 45 patients in each group (melasma and disease-free volunteers), assessing their clinical characteristics. In 36 patients, biopsies were performed on the lesion and the normal perilesional skin. Specimens were stained with HE and Fontana-Masson, and melanocytes analysed by immunohistochemistry. Objective measurements were accomplished by a specifically designed image analysis software. The melasma group had a mean age ± SD of 56.67 ± 8 years, the majority of them were women (86.7%) and 82.1% of the female cases had reached menopause. There were no significant differences between groups in terms of presence of comorbidities, use of medications or hormone therapies. For extra-facial melasma patients, family history of this dermatose and of previous facial melasma was significantly higher than in the control group (P < 0.05). The HE staining showed increased rectification and basal hyperpigmentation, solar elastosis, and collagen degeneration in the pigmented area (P < 0.05). There was a significant increase in melanin density in melasma biopsies, but the immunohistochemical tests did not detect a difference between the groups in terms of number of melanocytes. Extra-facial melasma appears to be related to menopause, family history, and personal history of facial melasma, in the studied population. Histopathology revealed a pattern similar to what has been described for facial melasma, with signs of solar degeneration, and a similar number of melanocytes, when comparing patients, and controls, suggesting that

  10. [Using infrared thermal asymmetry analysis for objective assessment of the lesion of facial nerve function].

    PubMed

    Liu, Xu-long; Hong, Wen-xue; Song, Jia-lin; Wu, Zhen-ying

    2012-03-01

    The skin temperature distribution of a healthy human body exhibits a contralateral symmetry. Some lesions of facial nerve function are associated with an alteration of the thermal distribution of the human body. Since the dissipation of heat through the skin occurs for the most part in the form of infrared radiation, infrared thermography is the method of choice to capture the alteration of the infrared thermal distribution. This paper presents a new method of analysis of the thermal asymmetry named effective thermal area ratio, which is a product of two variables. The first variable is mean temperature difference between the specific facial region and its contralateral region. The second variable is a ratio, which is equal to the area of the abnormal region divided by the total area. Using this new method, we performed a controlled trial to assess the facial nerve function of the healthy subjects and the patients with Bell's palsy respectively. The results show: that the mean specificity and sensitivity of this method are 0.90 and 0.87 respectively, improved by 7% and 26% compared with conventional methods. Spearman correlation coefficient between effective thermal area ratio and the degree of facial nerve function is an average of 0.664. Hence, concerning the diagnosis and assessment of facial nerve function, infrared thermography is a powerful tool; while the effective ther mal area ratio is an efficient clinical indicator.

  11. Measuring facial expression of emotion.

    PubMed

    Wolf, Karsten

    2015-12-01

    Research into emotions has increased in recent decades, especially on the subject of recognition of emotions. However, studies of the facial expressions of emotion were compromised by technical problems with visible video analysis and electromyography in experimental settings. These have only recently been overcome. There have been new developments in the field of automated computerized facial recognition; allowing real-time identification of facial expression in social environments. This review addresses three approaches to measuring facial expression of emotion and describes their specific contributions to understanding emotion in the healthy population and in persons with mental illness. Despite recent progress, studies on human emotions have been hindered by the lack of consensus on an emotion theory suited to examining the dynamic aspects of emotion and its expression. Studying expression of emotion in patients with mental health conditions for diagnostic and therapeutic purposes will profit from theoretical and methodological progress.

  12. Can We Distinguish Emotions from Faces? Investigation of Implicit and Explicit Processes of Peak Facial Expressions.

    PubMed

    Xiao, Ruiqi; Li, Xianchun; Li, Lin; Wang, Yanmei

    2016-01-01

    Most previous studies on facial expression recognition have focused on the moderate emotions; to date, few studies have been conducted to investigate the explicit and implicit processes of peak emotions. In the current study, we used transiently peak intense expression images of athletes at the winning or losing point in competition as materials, and investigated the diagnosability of peak facial expressions at both implicit and explicit levels. In Experiment 1, participants were instructed to evaluate isolated faces, isolated bodies, and the face-body compounds, and eye-tracking movement was recorded. The results revealed that the isolated body and face-body congruent images were better recognized than isolated face and face-body incongruent images, indicating that the emotional information conveyed by facial cues was ambiguous, and the body cues influenced facial emotion recognition. Furthermore, eye movement records showed that the participants displayed distinct gaze patterns for the congruent and incongruent compounds. In Experiment 2A, the subliminal affective priming task was used, with faces as primes and bodies as targets, to investigate the unconscious emotion perception of peak facial expressions. The results showed that winning face prime facilitated reaction to winning body target, whereas losing face prime inhibited reaction to winning body target, suggesting that peak facial expressions could be perceived at the implicit level. In general, the results indicate that peak facial expressions cannot be consciously recognized but can be perceived at the unconscious level. In Experiment 2B, revised subliminal affective priming task and a strict awareness test were used to examine the validity of unconscious perception of peak facial expressions found in Experiment 2A. Results of Experiment 2B showed that reaction time to both winning body targets and losing body targets was influenced by the invisibly peak facial expression primes, which indicated the

  13. Can We Distinguish Emotions from Faces? Investigation of Implicit and Explicit Processes of Peak Facial Expressions

    PubMed Central

    Xiao, Ruiqi; Li, Xianchun; Li, Lin; Wang, Yanmei

    2016-01-01

    Most previous studies on facial expression recognition have focused on the moderate emotions; to date, few studies have been conducted to investigate the explicit and implicit processes of peak emotions. In the current study, we used transiently peak intense expression images of athletes at the winning or losing point in competition as materials, and investigated the diagnosability of peak facial expressions at both implicit and explicit levels. In Experiment 1, participants were instructed to evaluate isolated faces, isolated bodies, and the face-body compounds, and eye-tracking movement was recorded. The results revealed that the isolated body and face-body congruent images were better recognized than isolated face and face-body incongruent images, indicating that the emotional information conveyed by facial cues was ambiguous, and the body cues influenced facial emotion recognition. Furthermore, eye movement records showed that the participants displayed distinct gaze patterns for the congruent and incongruent compounds. In Experiment 2A, the subliminal affective priming task was used, with faces as primes and bodies as targets, to investigate the unconscious emotion perception of peak facial expressions. The results showed that winning face prime facilitated reaction to winning body target, whereas losing face prime inhibited reaction to winning body target, suggesting that peak facial expressions could be perceived at the implicit level. In general, the results indicate that peak facial expressions cannot be consciously recognized but can be perceived at the unconscious level. In Experiment 2B, revised subliminal affective priming task and a strict awareness test were used to examine the validity of unconscious perception of peak facial expressions found in Experiment 2A. Results of Experiment 2B showed that reaction time to both winning body targets and losing body targets was influenced by the invisibly peak facial expression primes, which indicated the

  14. [An individual facial shield for a sportsman with an orofacial injury].

    PubMed

    de Baat, C; Peters, R; van Iperen-Keiman, C M; de Vleeschouwer, M

    2005-05-01

    Facial shields are used when practising contact sports, high speed sports, sports using hard balls, sticks or bats, sports using protective shields or covers, and sports using hard boardings around the sports ground. Examples of facial shields are commercially available, per branch of sport standardised helmets. Fabricating individual protective shields is primarily restricted to mouth guards. In individual cases a more extensive facial shield is demanded, for instance in case of a surgically stabilised facial bone fracture. In order to be able to fabricate an extensive individual facial shield, an accurate to the nearest model of the anterior part of the head is required. An accurate model can be provided by making an impression of the face, which is poured in dental stone. Another method is producing a stereolithographic model using computertomography or magnetic resonance imaging. On the accurate model the facial shield can be designed and fabricated from a strictly safe material, such as polyvinylchloride or polycarbonate.

  15. Geometric facial comparisons in speed-check photographs.

    PubMed

    Buck, Ursula; Naether, Silvio; Kreutz, Kerstin; Thali, Michael

    2011-11-01

    In many cases, it is not possible to call the motorists to account for their considerable excess in speeding, because they deny being the driver on the speed-check photograph. An anthropological comparison of facial features using a photo-to-photo comparison can be very difficult depending on the quality of the photographs. One difficulty of that analysis method is that the comparison photographs of the presumed driver are taken with a different camera or camera lens and from a different angle than for the speed-check photo. To take a comparison photograph with exactly the same camera setup is almost impossible. Therefore, only an imprecise comparison of the individual facial features is possible. The geometry and position of each facial feature, for example the distances between the eyes or the positions of the ears, etc., cannot be taken into consideration. We applied a new method using 3D laser scanning, optical surface digitalization, and photogrammetric calculation of the speed-check photo, which enables a geometric comparison. Thus, the influence of the focal length and the distortion of the objective lens are eliminated and the precise position and the viewing direction of the speed-check camera are calculated. Even in cases of low-quality images or when the face of the driver is partly hidden, good results are delivered using this method. This new method, Geometric Comparison, is evaluated and validated in a prepared study which is described in this article.

  16. Assessing facial attractiveness: individual decisions and evolutionary constraints

    PubMed Central

    Kocsor, Ferenc; Feldmann, Adam; Bereczkei, Tamas; Kállai, János

    2013-01-01

    Background Several studies showed that facial attractiveness, as a highly salient social cue, influences behavioral responses. It has also been found that attractive faces evoke distinctive neural activation compared to unattractive or neutral faces. Objectives Our aim was to design a face recognition task where individual preferences for facial cues are controlled for, and to create conditions that are more similar to natural circumstances in terms of decision making. Design In an event-related functional magnetic resonance imaging (fMRI) experiment, subjects were shown attractive and unattractive faces, categorized on the basis of their own individual ratings. Results Statistical analysis of all subjects showed elevated brain activation for attractive opposite-sex faces in contrast to less attractive ones in regions that previously have been reported to show enhanced activation with increasing attractiveness level (e.g. the medial and superior occipital gyri, fusiform gyrus, precentral gyrus, and anterior cingular cortex). Besides these, females showed additional brain activation in areas thought to be involved in basic emotions and desires (insula), detection of facial emotions (superior temporal gyrus), and memory retrieval (hippocampus). Conclusions From these data, we speculate that because of the risks involving mate choice faced by women during evolutionary times, selection might have preferred the development of an elaborated neural system in females to assess the attractiveness and social value of male faces. PMID:24693356

  17. Facial trauma as physical violence markers against elderly Brazilians: A comparative analysis between genders.

    PubMed

    de Sousa, Rayanne Izabel Maciel; de Macedo Bernardino, Ítalo; Castro, Ricardo Dias; Cavalcanti, Alessandro Leite; Bento, Patricia Meira; d'Ávila, Sérgio

    2016-01-01

    The aim of this study was to characterize the profile of elderly Brazilians with injuries resulting from physical violence and identify victimization differences. A descriptive and exploratory study was conducted involving the analysis of medico-legal and social records of 259 elderly victims of physical violence treated at an Institute of Forensic Medicine and Dentistry over four years (from January 2008 to December 2011). The forensic service database was evaluated by researchers properly trained and calibrated to perform this function between January and March 2013. Socio-demographic variables of victims, aggression characteristics, aggressor's profile and types of lesions were evaluated. Descriptive and multivariate statistics using Multiple Correspondence Analysis (MCA) were performed. The prevalence of facial trauma was 42.9%. Based on the MCA results, two groups with different victimization profiles were identified: married men aged 70-79 years, victims of community violence at night, suffering facial injuries; and single, widowed or separated women aged 60-69 years, victims of domestic violence during the day, suffering trauma in other areas of the body. The results suggest that there is a high prevalence of facial injuries among elderly Brazilians victims of physical violence and there are important differences related to victimization characteristics according to gender. Copyright © 2016 Elsevier Ireland Ltd. All rights reserved.

  18. The Relative Importance of Sexual Dimorphism, Fluctuating Asymmetry, and Color Cues to Health during Evaluation of Potential Partners' Facial Photographs : A Conjoint Analysis Study.

    PubMed

    Mogilski, Justin K; Welling, Lisa L M

    2017-03-01

    Sexual dimorphism, symmetry, and coloration in human faces putatively signal information relevant to mate selection and reproduction. Although the independent contributions of these characteristics to judgments of attractiveness are well established, relatively few studies have examined whether individuals prioritize certain features over others. Here, participants (N = 542, 315 female) ranked six sets of facial photographs (3 male, 3 female) by their preference for starting long- and short-term romantic relationships with each person depicted. Composite-based digital transformations were applied such that each image set contained 11 different versions of the same identity. Each photograph in each image set had a unique combination of three traits: sexual dimorphism, symmetry, and color cues to health. Using conjoint analysis to evaluate participants' ranking decisions, we found that participants prioritized cues to sexual dimorphism over symmetry and color cues to health. Sexual dimorphism was also found to be relatively more important for the evaluation of male faces than for female faces, whereas symmetry and color cues to health were relatively more important for the evaluation of female faces than for male faces. Symmetry and color cues to health were more important for long-term versus short-term evaluations for female faces, but not male faces. Analyses of utility estimates reveal that our data are consistent with research showing that preferences for facial masculinity and femininity in male and female faces vary according to relationship context. These findings are interpreted in the context of previous work examining the influence of these facial attributes on romantic partner perception.

  19. The Prevalence of Cosmetic Facial Plastic Procedures among Facial Plastic Surgeons.

    PubMed

    Moayer, Roxana; Sand, Jordan P; Han, Albert; Nabili, Vishad; Keller, Gregory S

    2018-04-01

    This is the first study to report on the prevalence of cosmetic facial plastic surgery use among facial plastic surgeons. The aim of this study is to determine the frequency with which facial plastic surgeons have cosmetic procedures themselves. A secondary aim is to determine whether trends in usage of cosmetic facial procedures among facial plastic surgeons are similar to that of nonsurgeons. The study design was an anonymous, five-question, Internet survey distributed via email set in a single academic institution. Board-certified members of the American Academy of Facial Plastic and Reconstructive Surgery (AAFPRS) were included in this study. Self-reported history of cosmetic facial plastic surgery or minimally invasive procedures were recorded. The survey also queried participants for demographic data. A total of 216 members of the AAFPRS responded to the questionnaire. Ninety percent of respondents were male ( n  = 192) and 10.3% were female ( n  = 22). Thirty-three percent of respondents were aged 31 to 40 years ( n  = 70), 25% were aged 41 to 50 years ( n  = 53), 21.4% were aged 51 to 60 years ( n  = 46), and 20.5% were older than 60 years ( n  = 44). Thirty-six percent of respondents had a surgical cosmetic facial procedure and 75% has at least one minimally invasive cosmetic facial procedure. Facial plastic surgeons are frequent users of cosmetic facial plastic surgery. This finding may be due to access, knowledge base, values, or attitudes. By better understanding surgeon attitudes toward facial plastic surgery, we can improve communication with patients and delivery of care. This study is a first step in understanding use of facial plastic procedures among facial plastic surgeons. Thieme Medical Publishers 333 Seventh Avenue, New York, NY 10001, USA.

  20. Efficient quantitative assessment of facial paralysis using iris segmentation and active contour-based key points detection with hybrid classifier.

    PubMed

    Barbosa, Jocelyn; Lee, Kyubum; Lee, Sunwon; Lodhi, Bilal; Cho, Jae-Gu; Seo, Woo-Keun; Kang, Jaewoo

    2016-03-12

    Facial palsy or paralysis (FP) is a symptom that loses voluntary muscles movement in one side of the human face, which could be very devastating in the part of the patients. Traditional methods are solely dependent to clinician's judgment and therefore time consuming and subjective in nature. Hence, a quantitative assessment system becomes apparently invaluable for physicians to begin the rehabilitation process; and to produce a reliable and robust method is challenging and still underway. We introduce a novel approach for a quantitative assessment of facial paralysis that tackles classification problem for FP type and degree of severity. Specifically, a novel method of quantitative assessment is presented: an algorithm that extracts the human iris and detects facial landmarks; and a hybrid approach combining the rule-based and machine learning algorithm to analyze and prognosticate facial paralysis using the captured images. A method combining the optimized Daugman's algorithm and Localized Active Contour (LAC) model is proposed to efficiently extract the iris and facial landmark or key points. To improve the performance of LAC, appropriate parameters of initial evolving curve for facial features' segmentation are automatically selected. The symmetry score is measured by the ratio between features extracted from the two sides of the face. Hybrid classifiers (i.e. rule-based with regularized logistic regression) were employed for discriminating healthy and unhealthy subjects, FP type classification, and for facial paralysis grading based on House-Brackmann (H-B) scale. Quantitative analysis was performed to evaluate the performance of the proposed approach. Experiments show that the proposed method demonstrates its efficiency. Facial movement feature extraction on facial images based on iris segmentation and LAC-based key point detection along with a hybrid classifier provides a more efficient way of addressing classification problem on facial palsy type and degree

  1. Role of facial attractiveness in patients with slight-to-borderline treatment need according to the Aesthetic Component of the Index of Orthodontic Treatment Need as judged by eye tracking.

    PubMed

    Johnson, Elizabeth K; Fields, Henry W; Beck, F Michael; Firestone, Allen R; Rosenstiel, Stephen F

    2017-02-01

    Previous eye-tracking research has demonstrated that laypersons view the range of dental attractiveness levels differently depending on facial attractiveness levels. How the borderline levels of dental attractiveness are viewed has not been evaluated in the context of facial attractiveness and compared with those with near-ideal esthetics or those in definite need of orthodontic treatment according to the Aesthetic Component of the Index of Orthodontic Treatment Need scale. Our objective was to determine the level of viewers' visual attention in its treatment need categories levels 3 to 7 for persons considered "attractive," "average," or "unattractive." Facial images of persons at 3 facial attractiveness levels were combined with 5 levels of dental attractiveness (dentitions representing Aesthetic Component of the Index of Orthodontic Treatment Need levels 3-7) using imaging software to form 15 composite images. Each image was viewed twice by 66 lay participants using eye tracking. Both the fixation density (number of fixations per facial area) and the fixation duration (length of time for each facial area) were quantified for each image viewed. Repeated-measures analysis of variance was used to determine how fixation density and duration varied among the 6 facial interest areas (chin, ear, eye, mouth, nose, and other). Viewers demonstrated excellent to good reliability among the 6 interest areas (intraviewer reliability, 0.70-0.96; interviewer reliability, 0.56-0.93). Between Aesthetic Component of the Index of Orthodontic Treatment Need levels 3 and 7, viewers of all facial attractiveness levels showed an increase in attention to the mouth. However, only with the attractive models were significant differences in fixation density and duration found between borderline levels with female viewers. Female viewers paid attention to different areas of the face than did male viewers. The importance of dental attractiveness is amplified in facially attractive female

  2. People with chronic facial pain perform worse than controls at a facial emotion recognition task, but it is not all about the emotion.

    PubMed

    von Piekartz, H; Wallwork, S B; Mohr, G; Butler, D S; Moseley, G L

    2015-04-01

    Alexithymia, or a lack of emotional awareness, is prevalent in some chronic pain conditions and has been linked to poor recognition of others' emotions. Recognising others' emotions from their facial expression involves both emotional and motor processing, but the possible contribution of motor disruption has not been considered. It is possible that poor performance on emotional recognition tasks could reflect problems with emotional processing, motor processing or both. We hypothesised that people with chronic facial pain would be less accurate in recognising others' emotions from facial expressions, would be less accurate in a motor imagery task involving the face, and that performance on both tasks would be positively related. A convenience sample of 19 people (15 females) with chronic facial pain and 19 gender-matched controls participated. They undertook two tasks; in the first task, they identified the facial emotion presented in a photograph. In the second, they identified whether the person in the image had a facial feature pointed towards their left or right side, a well-recognised paradigm to induce implicit motor imagery. People with chronic facial pain performed worse than controls at both tasks (Facially Expressed Emotion Labelling (FEEL) task P < 0·001; left/right judgment task P < 0·001). Participants who were more accurate at one task were also more accurate at the other, regardless of group (P < 0·001, r(2)  = 0·523). Participants with chronic facial pain were worse than controls at both the FEEL emotion recognition task and the left/right facial expression task and performance covaried within participants. We propose that disrupted motor processing may underpin or at least contribute to the difficulty that facial pain patients have in emotion recognition and that further research that tests this proposal is warranted. © 2014 John Wiley & Sons Ltd.

  3. Whole-face procedures for recovering facial images from memory.

    PubMed

    Frowd, Charlie D; Skelton, Faye; Hepton, Gemma; Holden, Laura; Minahil, Simra; Pitchford, Melanie; McIntyre, Alex; Brown, Charity; Hancock, Peter J B

    2013-06-01

    Research has indicated that traditional methods for accessing facial memories usually yield unidentifiable images. Recent research, however, has made important improvements in this area to the witness interview, method used for constructing the face and recognition of finished composites. Here, we investigated whether three of these improvements would produce even-more recognisable images when used in conjunction with each other. The techniques are holistic in nature: they involve processes which operate on an entire face. Forty participants first inspected an unfamiliar target face. Nominally 24h later, they were interviewed using a standard type of cognitive interview (CI) to recall the appearance of the target, or an enhanced 'holistic' interview where the CI was followed by procedures for focussing on the target's character. Participants then constructed a composite using EvoFIT, a recognition-type system that requires repeatedly selecting items from face arrays, with 'breeding', to 'evolve' a composite. They either saw faces in these arrays with blurred external features, or an enhanced method where these faces were presented with masked external features. Then, further participants attempted to name the composites, first by looking at the face front-on, the normal method, and then for a second time by looking at the face side-on, which research demonstrates facilitates recognition. All techniques improved correct naming on their own, but together promoted highly-recognisable composites with mean naming at 74% correct. The implication is that these techniques, if used together by practitioners, should substantially increase the detection of suspects using this forensic method of person identification. Copyright © 2013 Forensic Science Society. Published by Elsevier Ireland Ltd. All rights reserved.

  4. [Surgical treatment in otogenic facial nerve palsy].

    PubMed

    Feng, Guo-Dong; Gao, Zhi-Qiang; Zhai, Meng-Yao; Lü, Wei; Qi, Fang; Jiang, Hong; Zha, Yang; Shen, Peng

    2008-06-01

    To study the character of facial nerve palsy due to four different auris diseases including chronic otitis media, Hunt syndrome, tumor and physical or chemical factors, and to discuss the principles of the surgical management of otogenic facial nerve palsy. The clinical characters of 24 patients with otogenic facial nerve palsy because of the four different auris diseases were retrospectively analyzed, all the cases were performed surgical management from October 1991 to March 2007. Facial nerve function was evaluated with House-Brackmann (HB) grading system. The 24 patients including 10 males and 14 females were analysis, of whom 12 cases due to cholesteatoma, 3 cases due to chronic otitis media, 3 cases due to Hunt syndrome, 2 cases resulted from acute otitis media, 2 cases due to physical or chemical factors and 2 cases due to tumor. All cases were treated with operations included facial nerve decompression, lesion resection with facial nerve decompression and lesion resection without facial nerve decompression, 1 patient's facial nerve was resected because of the tumor. According to HB grade system, I degree recovery was attained in 4 cases, while II degree in 10 cases, III degree in 6 cases, IV degree in 2 cases, V degree in 2 cases and VI degree in 1 case. Removing the lesions completely was the basic factor to the surgery of otogenic facial palsy, moreover, it was important to have facial nerve decompression soon after lesion removal.

  5. Processing of Facial Emotion in Bipolar Depression and Euthymia.

    PubMed

    Robinson, Lucy J; Gray, John M; Burt, Mike; Ferrier, I Nicol; Gallagher, Peter

    2015-10-01

    Previous studies of facial emotion processing in bipolar disorder (BD) have reported conflicting findings. In independently conducted studies, we investigate facial emotion labeling in euthymic and depressed BD patients using tasks with static and dynamically morphed images of different emotions displayed at different intensities. Study 1 included 38 euthymic BD patients and 28 controls. Participants completed two tasks: labeling of static images of basic facial emotions (anger, disgust, fear, happy, sad) shown at different expression intensities; the Eyes Test (Baron-Cohen, Wheelwright, Hill, Raste, & Plumb, 2001), which involves recognition of complex emotions using only the eye region of the face. Study 2 included 53 depressed BD patients and 47 controls. Participants completed two tasks: labeling of "dynamic" facial expressions of the same five basic emotions; the Emotional Hexagon test (Young, Perret, Calder, Sprengelmeyer, & Ekman, 2002). There were no significant group differences on any measures of emotion perception/labeling, compared to controls. A significant group by intensity interaction was observed in both emotion labeling tasks (euthymia and depression), although this effect did not survive the addition of measures of executive function/psychomotor speed as covariates. Only 2.6-15.8% of euthymic patients and 7.8-13.7% of depressed patients scored below the 10th percentile of the controls for total emotion recognition accuracy. There was no evidence of specific deficits in facial emotion labeling in euthymic or depressed BD patients. Methodological variations-including mood state, sample size, and the cognitive demands of the tasks-may contribute significantly to the variability in findings between studies.

  6. Outcome-dependent coactivation of lip and tongue primary somatosensory representation following hypoglossal-facial transfer after peripheral facial palsy.

    PubMed

    Rottler, Philipp; Schroeder, Henry W S; Lotze, Martin

    2014-02-01

    A hypoglossal-facial transfer is a common surgical strategy for reanimating the face after persistent total hemifacial palsy. We were interested in how motor recovery is associated with cortical reorganization of lip and tongue representation in the primary sensorimotor cortex after the transfer. Therefore, we used functional magnetic resonance imaging (fMRI) in 13 patients who underwent a hypoglossal-facial transfer after unilateral peripheral facial palsy. To identify primary motor and somatosensory tongue and lip representation sites, we measured repetitive tongue and lip movements during fMRI. Electromyography (EMG) of the perioral muscles during tongue and lip movements and standardized evaluation of lip elevation served as outcome parameters. We found an association of cortical representation sites in the pre- and postcentral gyrus (decreased distance of lip and tongue representation) with symmetry of recovered lip movements (lip elevation) and coactivation of the lip during voluntary tongue movements (EMG-activity of the lip during tongue movements). Overall, our study shows that hypoglossal-facial transfer resulted in an outcome-dependent cortical reorganization with activation of the cortical tongue area for restituded movement of the lip. Copyright © 2012 Wiley Periodicals, Inc.

  7. Exaggerated perception of facial expressions is increased in individuals with schizotypal traits

    PubMed Central

    Uono, Shota; Sato, Wataru; Toichi, Motomi

    2015-01-01

    Emotional facial expressions are indispensable communicative tools, and social interactions involving facial expressions are impaired in some psychiatric disorders. Recent studies revealed that the perception of dynamic facial expressions was exaggerated in normal participants, and this exaggerated perception is weakened in autism spectrum disorder (ASD). Based on the notion that ASD and schizophrenia spectrum disorder are at two extremes of the continuum with respect to social impairment, we hypothesized that schizophrenic characteristics would strengthen the exaggerated perception of dynamic facial expressions. To test this hypothesis, we investigated the relationship between the perception of facial expressions and schizotypal traits in a normal population. We presented dynamic and static facial expressions, and asked participants to change an emotional face display to match the perceived final image. The presence of schizotypal traits was positively correlated with the degree of exaggeration for dynamic, as well as static, facial expressions. Among its subscales, the paranoia trait was positively correlated with the exaggerated perception of facial expressions. These results suggest that schizotypal traits, specifically the tendency to over-attribute mental states to others, exaggerate the perception of emotional facial expressions. PMID:26135081

  8. Exaggerated perception of facial expressions is increased in individuals with schizotypal traits.

    PubMed

    Uono, Shota; Sato, Wataru; Toichi, Motomi

    2015-07-02

    Emotional facial expressions are indispensable communicative tools, and social interactions involving facial expressions are impaired in some psychiatric disorders. Recent studies revealed that the perception of dynamic facial expressions was exaggerated in normal participants, and this exaggerated perception is weakened in autism spectrum disorder (ASD). Based on the notion that ASD and schizophrenia spectrum disorder are at two extremes of the continuum with respect to social impairment, we hypothesized that schizophrenic characteristics would strengthen the exaggerated perception of dynamic facial expressions. To test this hypothesis, we investigated the relationship between the perception of facial expressions and schizotypal traits in a normal population. We presented dynamic and static facial expressions, and asked participants to change an emotional face display to match the perceived final image. The presence of schizotypal traits was positively correlated with the degree of exaggeration for dynamic, as well as static, facial expressions. Among its subscales, the paranoia trait was positively correlated with the exaggerated perception of facial expressions. These results suggest that schizotypal traits, specifically the tendency to over-attribute mental states to others, exaggerate the perception of emotional facial expressions.

  9. Dynamic facial expressions evoke distinct activation in the face perception network: a connectivity analysis study.

    PubMed

    Foley, Elaine; Rippon, Gina; Thai, Ngoc Jade; Longe, Olivia; Senior, Carl

    2012-02-01

    Very little is known about the neural structures involved in the perception of realistic dynamic facial expressions. In the present study, a unique set of naturalistic dynamic facial emotional expressions was created. Through fMRI and connectivity analysis, a dynamic face perception network was identified, which is demonstrated to extend Haxby et al.'s [Haxby, J. V., Hoffman, E. A., & Gobbini, M. I. The distributed human neural system for face perception. Trends in Cognitive Science, 4, 223-233, 2000] distributed neural system for face perception. This network includes early visual regions, such as the inferior occipital gyrus, which is identified as insensitive to motion or affect but sensitive to the visual stimulus, the STS, identified as specifically sensitive to motion, and the amygdala, recruited to process affect. Measures of effective connectivity between these regions revealed that dynamic facial stimuli were associated with specific increases in connectivity between early visual regions, such as the inferior occipital gyrus and the STS, along with coupling between the STS and the amygdala, as well as the inferior frontal gyrus. These findings support the presence of a distributed network of cortical regions that mediate the perception of different dynamic facial expressions.

  10. Cerebro-facio-thoracic dysplasia (Pascual-Castroviejo syndrome): Identification of a novel mutation, use of facial recognition analysis, and review of the literature.

    PubMed

    Tender, Jennifer A F; Ferreira, Carlos R

    2018-04-13

    Cerebro-facio-thoracic dysplasia (CFTD) is a rare, autosomal recessive disorder characterized by facial dysmorphism, cognitive impairment and distinct skeletal anomalies and has been linked to the TMCO1 defect syndrome. To describe two siblings with features consistent with CFTD with a novel homozygous p.Arg114* pathogenic variant in the TMCO1 gene. We conducted a literature review and summarized the clinical features and laboratory results of two siblings with a novel pathogenic variant in the TMCO1 gene. Facial recognition analysis was utilized to assess the specificity of facial traits. The novel homozygous p.Arg114* pathogenic variant in the TMCO1 gene is responsible for the clinical features of CFTD in two siblings. Facial recognition analysis allows unambiguous distinction of this syndrome against controls.

  11. MPEG-4-based 2D facial animation for mobile devices

    NASA Astrophysics Data System (ADS)

    Riegel, Thomas B.

    2005-03-01

    The enormous spread of mobile computing devices (e.g. PDA, cellular phone, palmtop, etc.) emphasizes scalable applications, since users like to run their favorite programs on the terminal they operate at that moment. Therefore appliances are of interest, which can be adapted to the hardware realities without loosing a lot of their functionalities. A good example for this is "Facial Animation," which offers an interesting way to achieve such "scalability." By employing MPEG-4, which provides an own profile for facial animation, a solution for low power terminals including mobile phones is demonstrated. From the generic 3D MPEG-4 face a specific 2D head model is derived, which consists primarily of a portrait image superposed by a suited warping mesh and adapted 2D animation rules. Thus the animation process of MPEG-4 need not be changed and standard compliant facial animation parameters can be used to displace the vertices of the mesh and warp the underlying image accordingly.

  12. Fourier analysis of human soft tissue facial shape: sex differences in normal adults.

    PubMed Central

    Ferrario, V F; Sforza, C; Schmitz, J H; Miani, A; Taroni, G

    1995-01-01

    Sexual dimorphism in human facial form involves both size and shape variations of the soft tissue structures. These variations are conventionally appreciated using linear and angular measurements, as well as ratios, taken from photographs or radiographs. Unfortunately this metric approach provides adequate quantitative information about size only, eluding the problems of shape definition. Mathematical methods such as the Fourier series allow a correct quantitative analysis of shape and of its changes. A method for the reconstruction of outlines starting from selected landmarks and for their Fourier analysis has been developed, and applied to analyse sex differences in shape of the soft tissue facial contour in a group of healthy young adults. When standardised for size, no sex differences were found between both cosine and sine coefficients of the Fourier series expansion. This shape similarity was largely overwhelmed by the very evident size differences and it could be measured only using the proper mathematical methods. PMID:8586558

  13. Cognitive penetrability and emotion recognition in human facial expressions

    PubMed Central

    Marchi, Francesco

    2015-01-01

    Do our background beliefs, desires, and mental images influence our perceptual experience of the emotions of others? In this paper, we will address the possibility of cognitive penetration (CP) of perceptual experience in the domain of social cognition. In particular, we focus on emotion recognition based on the visual experience of facial expressions. After introducing the current debate on CP, we review examples of perceptual adaptation for facial expressions of emotion. This evidence supports the idea that facial expressions are perceptually processed as wholes. That is, the perceptual system integrates lower-level facial features, such as eyebrow orientation, mouth angle etc., into facial compounds. We then present additional experimental evidence showing that in some cases, emotion recognition on the basis of facial expression is sensitive to and modified by the background knowledge of the subject. We argue that such sensitivity is best explained as a difference in the visual experience of the facial expression, not just as a modification of the judgment based on this experience. The difference in experience is characterized as the result of the interference of background knowledge with the perceptual integration process for faces. Thus, according to the best explanation, we have to accept CP in some cases of emotion recognition. Finally, we discuss a recently proposed mechanism for CP in the face-based recognition of emotion. PMID:26150796

  14. Adding Injury to Insult: A National Analysis of Combat Sport-Related Facial Injury.

    PubMed

    Hojjat, Houmehr; Svider, Peter F; Lin, Ho-Sheng; Folbe, Adam J; Shkoukani, Mahdi A; Eloy, Jean Anderson; Zuliani, Giancarlo

    2016-08-01

    To estimate the incidence of patients presenting to emergency departments (EDs) for facial trauma sustained from participation in combat sports and evaluate injury patterns and patient demographics. The National Electronic Injury Surveillance System (NEISS) was evaluated for facial injuries from wrestling, boxing, and martial arts leading to ED visits from 2008 to 2013. Relevant entries were examined for injury mechanism, location, type, as well as other patient characteristics. There were 1143 entries extrapolating to an estimated 42 395 ED visits from 2008 to 2013. Injury rates for boxing, marital arts, and wrestling were, respectively, 44, 56, and 120 injuries per 100 000 participants. Males comprised the majority (93.7%). A plurality of injuries involved lacerations (46.0%), followed by fractures (26.2%) and contusions/abrasions (19.3%). The proportion of fractures was highest among boxers (36.9%). Overall, the most common mechanisms of injury were punching, kicking, and head butting. The significant number of ED visits resulted from combat sports facial trauma, reinforcing the importance of familiarity with injury patterns among practitioners managing facial trauma. As most injuries involve individuals younger than 19 despite guidelines suggesting children and adolescents avoid combat sports, these findings may be used for patient education and encouragement of the use of personal protective equipment. Furthermore, injury patterns reported in this analysis may serve as an adjunct for enhancing clinical history taking and physical examination. © The Author(s) 2016.

  15. Changes in Lower Facial Height and Facial Esthetics with Incremental Increases in Occlusal Vertical Dimension in Dentate Subjects.

    PubMed

    Orenstein, Noah P; Bidra, Avinash S; Agar, John R; Taylor, Thomas D; Uribe, Flavio; Litt, Mark D; Little, Mark D

    2015-01-01

    To determine if there are objective changes in lower facial height and subjective changes in facial esthetics with incremental increases in occlusal vertical dimension in dentate subjects. Twenty subjects of four different races and both sexes with a Class I dental occlusion had custom diagnostic occlusal prostheses (mandibular overlays) fabricated on casts mounted on a semi-adjustable articulator. The overlays were fabricated at 2-mm, 3-mm, 4-mm, and 5-mm openings of the anterior guide pin of a semi-adjustable articulator. Direct facial measurements were made between pronasale and menton on each subject while wearing the four different overlays. Thereafter, two digital photographs (frontal and profile) were taken for each subject at maximum intercuspation (baseline) and wearing each of the four mandibular overlays. The photographs of eight subjects were standardized and displayed in a random order to 60 judges comprising 30 laypeople, 15 general dentists, and 15 prosthodontists. Using a visual analog scale, each judge was asked to rate the facial esthetics twice for each of the 80 images. For objective changes, although an anterior guide pin-lower facial height relationship of 1:0.63 mm was observed, the findings were not correlated (P>.20). For subjective changes, the visual analog scale ratings of judges were uncorrelated with increases in anterior guide pin opening up to 5 mm, irrespective of the judge's background status or the sexes of the judges or the subjects (P>.80). Incremental increases in anterior guide pin opening up to 5 mm did not correlate to similar increases in lower facial height. Additionally, it made no difference in a judge's evaluation of facial esthetics irrespective of the judge's background status (layperson, general dentist, or prosthodontist) or sex.

  16. Human (Homo sapiens) facial attractiveness in relation to skin texture and color.

    PubMed

    Fink, B; Grammer, K; Thornhill, R

    2001-03-01

    The notion that surface texture may provide important information about the geometry of visible surfaces has attracted considerable attention for a long time. The present study shows that skin texture plays a significant role in the judgment of female facial beauty. Following research in clinical dermatology, the authors developed a computer program that implemented an algorithm based on co-occurrence matrices for the analysis of facial skin texture. Homogeneity and contrast features as well as color parameters were extracted out of stimulus faces. Attractiveness ratings of the images made by male participants relate positively to parameters of skin homogeneity. The authors propose that skin texture is a cue to fertility and health. In contrast to some previous studies, the authors found that dark skin, not light skin, was rated as most attractive.

  17. Emotion unfolded by motion: a role for parietal lobe in decoding dynamic facial expressions.

    PubMed

    Sarkheil, Pegah; Goebel, Rainer; Schneider, Frank; Mathiak, Klaus

    2013-12-01

    Facial expressions convey important emotional and social information and are frequently applied in investigations of human affective processing. Dynamic faces may provide higher ecological validity to examine perceptual and cognitive processing of facial expressions. Higher order processing of emotional faces was addressed by varying the task and virtual face models systematically. Blood oxygenation level-dependent activation was assessed using functional magnetic resonance imaging in 20 healthy volunteers while viewing and evaluating either emotion or gender intensity of dynamic face stimuli. A general linear model analysis revealed that high valence activated a network of motion-responsive areas, indicating that visual motion areas support perceptual coding for the motion-based intensity of facial expressions. The comparison of emotion with gender discrimination task revealed increased activation of inferior parietal lobule, which highlights the involvement of parietal areas in processing of high level features of faces. Dynamic emotional stimuli may help to emphasize functions of the hypothesized 'extended' over the 'core' system for face processing.

  18. A method of assessing facial profile attractiveness and its application in comparing the aesthetic preferences of two samples of South Africans.

    PubMed

    Morar, Ajay; Stein, Errol

    2011-06-01

    Numerous studies have evaluated the perception of facial attractiveness. However, many of the instruments previously used have limitations. This study introduces an improved tool and describes its application in the assessment of the preferred facial profile in two sample groups. Cross-sectional study. Two sites were involved: a rural healthcare facility (Winterveldt, Northwest Province) and the campus of the University of the Witwatersrand (Johannesburg, Gauteng Province). Adult females and males selected from amongst first, attendees at the healthcare facility, and second, staff of the University of the Witwatersrand. Eight androgynous lateral facial profile images were created using a morphing software programme representing six transitions between two anchoring extremes in terms of lip retrusion/protrusion vs protrusion/retrusion. These images were presented to, and rated by, two mixed male/female groups of rural and of urban habitat using a pre-piloted form. Statistical analysis of the responses obtained established the preferred facial profile by gender in each group. The perception of facial attractiveness varied marginally between rural and urban black South Africans. There was no statistically significant difference between females and males in the rural group (P=0·2353) and those in the urban sample (P=0·1318) with respect to their choice of ideal facial profile. Females and males in both the rural and urban groups found extreme profile convexity unappealing. By contrast, a larger proportion of rural females, rural males and urban females demonstrated a preference for extreme profile concavity. The research tool described is a useful instrument in the assessment of facial profile attractiveness.

  19. Autologous Fat Used for Facial Filling Can Lead to Massive Cerebral Infarction Through Middle Cerebral Artery or Facial Intracranial Branches.

    PubMed

    Wang, Xian; Wu, Min; Zhou, Xing; Liu, Hengdeng; Zhang, Yongchao; Wang, Haiping

    2018-05-31

    Autologous fat injection is a procedure aimed at eliminating grave defects in the skin surface by subcutaneous injection of the patient's fatty tissue. Fat embolism is a rare but severe complication of this procedure, especially cerebral infarction. It is first reported by Thaunat in 2004. were presented to the hospital with sudden unconsciousness and left limb weakness in 24 hours after facial fat injection. Brain computed tomography and magnetic resonance imaging were performed immediately after admission. Frontal temporoparietal decompressive craniectomy plus multiple treatments scheduled for patients. Pictures and videos were taken during follow-up. Figures are edited with Adobe Photograph CS6. Patients were diagnosed with extensive cerebral infarction of the right hemisphere through the middle cerebral artery or facial-intracranial branches. Routine cosmetic procedures of facial fat injections could cause devastating and even fatal complications to patients. The small volume of fat grafts can be inserted through the internal carotid artery or go through the communicating branches between the facial artery and the intracranial artery into the brain.

  20. Fusiform gyrus volume reduction and facial recognition in chronic schizophrenia.

    PubMed

    Onitsuka, Toshiaki; Shenton, Martha E; Kasai, Kiyoto; Nestor, Paul G; Toner, Sarah K; Kikinis, Ron; Jolesz, Ferenc A; McCarley, Robert W

    2003-04-01

    The fusiform gyrus (FG), or occipitotemporal gyrus, is thought to subserve the processing and encoding of faces. Of note, several studies have reported that patients with schizophrenia show deficits in facial processing. It is thus hypothesized that the FG might be one brain region underlying abnormal facial recognition in schizophrenia. The objectives of this study were to determine whether there are abnormalities in gray matter volumes for the anterior and the posterior FG in patients with chronic schizophrenia and to investigate relationships between FG subregions and immediate and delayed memory for faces. Patients were recruited from the Boston VA Healthcare System, Brockton Division, and control subjects were recruited through newspaper advertisement. Study participants included 21 male patients diagnosed as having chronic schizophrenia and 28 male controls. Participants underwent high-spatial-resolution magnetic resonance imaging, and facial recognition memory was evaluated. Main outcome measures included anterior and posterior FG gray matter volumes based on high-spatial-resolution magnetic resonance imaging, a detailed and reliable manual delineation using 3-dimensional information, and correlation coefficients between FG subregions and raw scores on immediate and delayed facial memory derived from the Wechsler Memory Scale III. Patients with chronic schizophrenia had overall smaller FG gray matter volumes (10%) than normal controls. Additionally, patients with schizophrenia performed more poorly than normal controls in both immediate and delayed facial memory tests. Moreover, the degree of poor performance on delayed memory for faces was significantly correlated with the degree of bilateral anterior FG reduction in patients with schizophrenia. These results suggest that neuroanatomic FG abnormalities underlie at least some of the deficits associated with facial recognition in schizophrenia.

  1. Design and fabrication of facial prostheses for cancer patient applying computer aided method and manufacturing (CADCAM)

    NASA Astrophysics Data System (ADS)

    Din, Tengku Noor Daimah Tengku; Jamayet, Nafij; Rajion, Zainul Ahmad; Luddin, Norhayati; Abdullah, Johari Yap; Abdullah, Abdul Manaf; Yahya, Suzana

    2016-12-01

    Facial defects are either congenital or caused by trauma or cancer where most of them affect the person appearance. The emotional pressure and low self-esteem are problems commonly related to patient with facial defect. To overcome this problem, silicone prosthesis was designed to cover the defect part. This study describes the techniques in designing and fabrication for facial prosthesis applying computer aided method and manufacturing (CADCAM). The steps of fabricating the facial prosthesis were based on a patient case. The patient was diagnosed for Gorlin Gotz syndrome and came to Hospital Universiti Sains Malaysia (HUSM) for prosthesis. The 3D image of the patient was reconstructed from CT data using MIMICS software. Based on the 3D image, the intercanthal and zygomatic measurements of the patient were compared with available data in the database to find the suitable nose shape. The normal nose shape for the patient was retrieved from the nasal digital library. Mirror imaging technique was used to mirror the facial part. The final design of facial prosthesis including eye, nose and cheek was superimposed to see the result virtually. After the final design was confirmed, the mould design was created. The mould of nasal prosthesis was printed using Objet 3D printer. Silicone casting was done using the 3D print mould. The final prosthesis produced from the computer aided method was acceptable to be used for facial rehabilitation to provide better quality of life.

  2. Reaction of facial soft tissues to treatment with a Herbst appliance.

    PubMed

    Meyer-Marcotty, P; Kochel, J; Richter, U; Richter, F; Stellzig-Eisenhauer, Angelika

    2012-04-01

    The objective of this prospective longitudinal study was to investigate the reaction of facial soft tissues to treatment with a Herbst appliance. We aimed to quantify three-dimensionally (3D) the isolated effect of the Herbst appliance and volume changes in the lip profile. The 3D data of the facial soft tissues of 34 patients with skeletal Class II (17 female and 17 male, mean age 13.5 ± 1.8 years) were prepared in a standardized manner immediately before (T1) and after (T2) treatment with a Herbst appliance. Anthropometric evaluation was carried out in sagittal and vertical dimensions. To quantify volume changes, pretherapeutic and posttherapeutic images were superimposed three-dimensionally and the difference volumes calculated. Following testing for normal distribution, a statistical analysis was carried out using the paired t test. We observed ventral development of the soft tissues of the lower jaw with flattening of the profile curvature and anterior displacement of the sublabial region in a total of 27 patients. Anterior facial height was lengthened and the facial depth at the lower jaw increased. The largest percentage changes were noted in the lip profile, with a reduction in the red margin of the upper lip and an increase in lower lip height. We also observed a reduction of the sublabial fold in conjunction with a simultaneous increase in volume. The influence of the Herbst appliance on the facial soft tissues is expected to result in a positive treatment outcome, particularly in patients with a convex profile, a retrusive lower lip, and a marked sublabial fold. We observed a broad clinical spectrum of individual reactions in the facial soft tissues. It is, thus, not possible to detect a linear relationship between the Herbst treatment and soft tissue changes, making soft tissue changes difficult to predict.

  3. Effect of a Facial Muscle Exercise Device on Facial Rejuvenation

    PubMed Central

    Hwang, Ui-jae; Kwon, Oh-yun; Jung, Sung-hoon; Ahn, Sun-hee; Gwak, Gyeong-tae

    2018-01-01

    Abstract Background The efficacy of facial muscle exercises (FMEs) for facial rejuvenation is controversial. In the majority of previous studies, nonquantitative assessment tools were used to assess the benefits of FMEs. Objectives This study examined the effectiveness of FMEs using a Pao (MTG, Nagoya, Japan) device to quantify facial rejuvenation. Methods Fifty females were asked to perform FMEs using a Pao device for 30 seconds twice a day for 8 weeks. Facial muscle thickness and cross-sectional area were measured sonographically. Facial surface distance, surface area, and volumes were determined using a laser scanning system before and after FME. Facial muscle thickness, cross-sectional area, midfacial surface distances, jawline surface distance, and lower facial surface area and volume were compared bilaterally before and after FME using a paired Student t test. Results The cross-sectional areas of the zygomaticus major and digastric muscles increased significantly (right: P < 0.001, left: P = 0.015), while the midfacial surface distances in the middle (right: P = 0.005, left: P = 0.047) and lower (right: P = 0.028, left: P = 0.019) planes as well as the jawline surface distances (right: P = 0.004, left: P = 0.003) decreased significantly after FME using the Pao device. The lower facial surface areas (right: P = 0.005, left: P = 0.006) and volumes (right: P = 0.001, left: P = 0.002) were also significantly reduced after FME using the Pao device. Conclusions FME using the Pao device can increase facial muscle thickness and cross-sectional area, thus contributing to facial rejuvenation. Level of Evidence: 4 PMID:29365050

  4. Objectifying Facial Expressivity Assessment of Parkinson's Patients: Preliminary Study

    PubMed Central

    Patsis, Georgios; Jiang, Dongmei; Sahli, Hichem; Kerckhofs, Eric; Vandekerckhove, Marie

    2014-01-01

    Patients with Parkinson's disease (PD) can exhibit a reduction of spontaneous facial expression, designated as “facial masking,” a symptom in which facial muscles become rigid. To improve clinical assessment of facial expressivity of PD, this work attempts to quantify the dynamic facial expressivity (facial activity) of PD by automatically recognizing facial action units (AUs) and estimating their intensity. Spontaneous facial expressivity was assessed by comparing 7 PD patients with 8 control participants. To voluntarily produce spontaneous facial expressions that resemble those typically triggered by emotions, six emotions (amusement, sadness, anger, disgust, surprise, and fear) were elicited using movie clips. During the movie clips, physiological signals (facial electromyography (EMG) and electrocardiogram (ECG)) and frontal face video of the participants were recorded. The participants were asked to report on their emotional states throughout the experiment. We first examined the effectiveness of the emotion manipulation by evaluating the participant's self-reports. Disgust-induced emotions were significantly higher than the other emotions. Thus we focused on the analysis of the recorded data during watching disgust movie clips. The proposed facial expressivity assessment approach captured differences in facial expressivity between PD patients and controls. Also differences between PD patients with different progression of Parkinson's disease have been observed. PMID:25478003

  5. Objectifying facial expressivity assessment of Parkinson's patients: preliminary study.

    PubMed

    Wu, Peng; Gonzalez, Isabel; Patsis, Georgios; Jiang, Dongmei; Sahli, Hichem; Kerckhofs, Eric; Vandekerckhove, Marie

    2014-01-01

    Patients with Parkinson's disease (PD) can exhibit a reduction of spontaneous facial expression, designated as "facial masking," a symptom in which facial muscles become rigid. To improve clinical assessment of facial expressivity of PD, this work attempts to quantify the dynamic facial expressivity (facial activity) of PD by automatically recognizing facial action units (AUs) and estimating their intensity. Spontaneous facial expressivity was assessed by comparing 7 PD patients with 8 control participants. To voluntarily produce spontaneous facial expressions that resemble those typically triggered by emotions, six emotions (amusement, sadness, anger, disgust, surprise, and fear) were elicited using movie clips. During the movie clips, physiological signals (facial electromyography (EMG) and electrocardiogram (ECG)) and frontal face video of the participants were recorded. The participants were asked to report on their emotional states throughout the experiment. We first examined the effectiveness of the emotion manipulation by evaluating the participant's self-reports. Disgust-induced emotions were significantly higher than the other emotions. Thus we focused on the analysis of the recorded data during watching disgust movie clips. The proposed facial expressivity assessment approach captured differences in facial expressivity between PD patients and controls. Also differences between PD patients with different progression of Parkinson's disease have been observed.

  6. Reduction of facial redness with resveratrol added to topical product containing green tea polyphenols and caffeine.

    PubMed

    Ferzli, Georgina; Patel, Mital; Phrsai, Natasha; Brody, Neil

    2013-07-01

    Many topical formulations include antioxidants to improve the antioxidant capability of the skin. This study evaluated the ability of a unique combination of antioxidants including resveratrol, green tea polyphenols, and caffeine to reduce facial redness. Subjects (n=16) presenting with facial redness applied the resveratrol-enriched product twice daily to the entire face. Reduction in redness was evaluated by trained staff members and dermatology house staff officers. Evaluators compared clinical photographs and spectrally enhanced images taken before treatment and at 2-week intervals for up to 12 weeks. 16 of 16 clinical images showed improvement and 13 of 16 spectrally enhanced images were improved. Reduction in facial redness continued to evolve over the duration of the study period but was generally detectable by 6 weeks of treatment. Adverse effects were not observed in any subject. The skin product combination of resveratrol, green tea polyphenols, and caffeine safely reduces facial redness in most patients by 6 weeks of continuous treatment and may provide further improvement with additional treatment.

  7. Classifying and Standardizing Panfacial Trauma With a New Bony Facial Trauma Score.

    PubMed

    Casale, Garrett G A; Fishero, Brian A; Park, Stephen S; Sochor, Mark; Heltzel, Sara B; Christophel, J Jared

    2017-01-01

    The practice of facial trauma surgery would benefit from a useful quantitative scale that measures the extent of injury. To develop a facial trauma scale that incorporates only reducible fractures and is able to be reliably communicated to health care professionals. A cadaveric tissue study was conducted from October 1 to 3, 2014. Ten cadaveric heads were subjected to various degrees of facial trauma by dropping a fixed mass onto each head. The heads were then imaged with fine-cut computed tomography. A Bony Facial Trauma Scale (BFTS) for grading facial trauma was developed based only on clinically relevant (reducible) fractures. The traumatized cadaveric heads were then scored using this scale as well as 3 existing scoring systems. Regression analysis was used to determine correlation between degree of incursion of the fixed mass on the cadaveric heads and trauma severity as rated by the scoring systems. Statistical analysis was performed to determine correlation of the scores obtained using the BFTS with those of the 3 existing scoring systems. Scores obtained using the BFTS were not correlated with dentition (95% CI, -0.087 to 1.053; P = .08; measured as absolute number of teeth) or age of the cadaveric donor (95% CI, -0.068 to 0.944; P = .08). Facial trauma scores. Among all 10 cadaveric specimens (9 male donors and 1 female donor; age range, 41-87 years; mean age, 57.2 years), the facial trauma scores obtained using the BFTS correlated with depth of penetration of the mass into the face (odds ratio, 4.071; 95% CI, 1.676-6.448) P = .007) when controlling for presence of dentition and age. The BFTS scores also correlated with scores obtained using 3 existing facial trauma models (Facial Fracture Severity Scale, rs = 0.920; Craniofacial Disruption Score, rs = 0.945; and ZS Score, rs = 0.902; P < .001 for all 3 models). In addition, the BFTS was found to have excellent interrater reliability (0.908; P = .001), which was similar to

  8. Facial emotion processing in pediatric social anxiety disorder: Relevance of situational context.

    PubMed

    Schwab, Daniela; Schienle, Anne

    2017-08-01

    Social anxiety disorder (SAD) typically begins in childhood. Previous research has demonstrated that adult patients respond with elevated late positivity (LP) to negative facial expressions. In the present study on pediatric SAD, we investigated responses to negative facial expressions and the role of social context information. Fifteen children with SAD and 15 non-anxious controls were first presented with images of negative facial expressions with masked backgrounds. Following this, the complete images which included context information, were shown. The negative expressions were either a result of an emotion-relevant (e.g., social exclusion) or emotion-irrelevant elicitor (e.g., weight lifting). Relative to controls, the clinical group showed elevated parietal LP during face processing with and without context information. Both groups differed in their frontal LP depending on the type of context. In SAD patients, frontal LP was lower in emotion-relevant than emotion-irrelevant contexts. We conclude that SAD patients direct more automatic attention towards negative facial expressions (parietal effect) and are less capable in integrating affective context information (frontal effect). Copyright © 2017 Elsevier Ltd. All rights reserved.

  9. Facial emotion identification in early-onset and first-episode psychosis: a systematic review with meta-analysis.

    PubMed

    Barkl, Sophie J; Lah, Suncica; Harris, Anthony W F; Williams, Leanne M

    2014-10-01

    Patients with chronic schizophrenia are characterized by deficits in identifying facial expressions of emotion, and these deficits relate to impaired social and occupational function. It is not yet known if these deficits are trait-like and present at the onset of psychosis, preceding a subsequent diagnosis of schizophrenia. Our objective was to systematically review and analyze the extant literature to assess if there is a consistent profile of emotion identification problems in early-onset and first-episode psychosis. We conducted a systematic review and meta-analysis of 12 peer-reviewed studies of facial emotion identification in early-onset and first-episode psychosis, published between 1980 and March 2013. We examined the average mean difference between patients and controls on measures of facial emotion identification. Findings suggest that patients with early-onset and first-episode psychosis have impairment in identifying facial expressions of biologically salient emotion. Across the 12 studies, the onset of psychosis was distinguished by a generalized effect of significantly poorer accuracy for identifying facial expressions of emotion than healthy controls, and this difference had a substantial effect size (d=-0.88, N=378, 95% CI=-1.42 to -0.32). Within this general effect some emotions were also harder for patients to identify than others, with the magnitude of impairment found to be (i) large for disgust, fear and surprise, and (ii) medium for sadness, and happiness. No between groups mean differences were found for anger or neutral facial expressions. Deficits in facial emotion identification are evident at first onset of a psychotic episode. The findings suggest that, over and above a generalized deficit in identifying facial emotion, patients may find some emotions harder to identifying than others. This reflects findings with chronic schizophrenia populations and suggests that emotion identification impairment represents a trait susceptibility

  10. Positive facial expressions during retrieval of self-defining memories.

    PubMed

    Gandolphe, Marie Charlotte; Nandrino, Jean Louis; Delelis, Gérald; Ducro, Claire; Lavallee, Audrey; Saloppe, Xavier; Moustafa, Ahmed A; El Haj, Mohamad

    2017-11-14

    In this study, we investigated, for the first time, facial expressions during the retrieval of Self-defining memories (i.e., those vivid and emotionally intense memories of enduring concerns or unresolved conflicts). Participants self-rated the emotional valence of their Self-defining memories and autobiographical retrieval was analyzed with a facial analysis software. This software (Facereader) synthesizes the facial expression information (i.e., cheek, lips, muscles, eyebrow muscles) to describe and categorize facial expressions (i.e., neutral, happy, sad, surprised, angry, scared, and disgusted facial expressions). We found that participants showed more emotional than neutral facial expressions during the retrieval of Self-defining memories. We also found that participants showed more positive than negative facial expressions during the retrieval of Self-defining memories. Interestingly, participants attributed positive valence to the retrieved memories. These findings are the first to demonstrate the consistency between facial expressions and the emotional subjective experience of Self-defining memories. These findings provide valuable physiological information about the emotional experience of the past.

  11. Vertical facial height and its correlation with facial width and depth: Three dimensional cone beam computed tomography evaluation based on dry skulls.

    PubMed

    Wang, Ming Feng; Otsuka, Takero; Akimoto, Susumu; Sato, Sadao

    2013-01-01

    The aim of the present study was to evaluate how vertical facial height correlates with mandibular plane angle, facial width and depth from a three dimensional (3D) viewing angle. In this study 3D cephalometric landmarks were identified and measurements from 43 randomly selected cone beam computed tomography (CBCT) images of dry skulls from the Weisbach collection of Vienna Natural History Museum were analyzed. Pearson correlation coefficients of facial height measurements and mandibular plane angle and the correlation coefficients of height-width and height-depth were calculated, respectively. The mandibular plane angle (MP-SN) significantly correlated with ramus height (Co-Go) and posterior facial height (PFH) but not with anterior lower face height (ALFH) or anterior total face height (ATFH). The ALFH and ATFH showed significant correlation with anterior cranial base length (S-N), whereas PFH showed significant correlation with the mandible (S-B) and maxilla (S-A) anteroposterior position. High or low mandibular plane angle might not necessarily be accompanied by long or short anterior face height, respectively. The PFH rather than AFH is assumed to play a key role in the vertical facial type whereas AFH seems to undergo relatively intrinsic growth.

  12. Enhancement pattern of the normal facial nerve at 3.0 T temporal MRI.

    PubMed

    Hong, H S; Yi, B-H; Cha, J-G; Park, S-J; Kim, D H; Lee, H K; Lee, J-D

    2010-02-01

    The purpose of this study was to evaluate the enhancement pattern of the normal facial nerve at 3.0 T temporal MRI. We reviewed the medical records of 20 patients and evaluated 40 clinically normal facial nerves demonstrated by 3.0 T temporal MRI. The grade of enhancement of the facial nerve was visually scaled from 0 to 3. The patients comprised 11 men and 9 women, and the mean age was 39.7 years. The reasons for the MRI were sudden hearing loss (11 patients), Méniàre's disease (6) and tinnitus (7). Temporal MR scans were obtained by fluid-attenuated inversion-recovery (FLAIR) and diffusion-weighted imaging of the brain; three-dimensional (3D) fast imaging employing steady-state acquisition (FIESTA) images of the temporal bone with a 0.77 mm thickness, and pre-contrast and contrast-enhanced 3D spoiled gradient record acquisition in the steady state (SPGR) of the temporal bone with a 1 mm thickness, were obtained with 3.0 T MR scanning. 40 nerves (100%) were visibly enhanced along at least one segment of the facial nerve. The enhanced segments included the geniculate ganglion (77.5%), tympanic segment (37.5%) and mastoid segment (100%). Even the facial nerve in the internal auditory canal (15%) and labyrinthine segments (5%) showed mild enhancement. The use of high-resolution, high signal-to-noise ratio (with 3 T MRI), thin-section contrast-enhanced 3D SPGR sequences showed enhancement of the normal facial nerve along the whole course of the nerve; however, only mild enhancement was observed in areas associated with acute neuritis, namely the canalicular and labyrinthine segment.

  13. Cerebro-facio-thoracic dysplasia (Pascual-Castroviejo syndrome): Identification of a novel mutation, use of facial recognition analysis, and review of the literature

    PubMed Central

    Tender, Jennifer A.F.; Ferreira, Carlos R.

    2018-01-01

    BACKGROUND: Cerebro-facio-thoracic dysplasia (CFTD) is a rare, autosomal recessive disorder characterized by facial dysmorphism, cognitive impairment and distinct skeletal anomalies and has been linked to the TMCO1 defect syndrome. OBJECTIVE: To describe two siblings with features consistent with CFTD with a novel homozygous p.Arg114* pathogenic variant in the TMCO1 gene. METHODS: We conducted a literature review and summarized the clinical features and laboratory results of two siblings with a novel pathogenic variant in the TMCO1 gene. Facial recognition analysis was utilized to assess the specificity of facial traits. CONCLUSION: The novel homozygous p.Arg114* pathogenic variant in the TMCO1 gene is responsible for the clinical features of CFTD in two siblings. Facial recognition analysis allows unambiguous distinction of this syndrome against controls. PMID:29682451

  14. More emotional facial expressions during episodic than during semantic autobiographical retrieval.

    PubMed

    El Haj, Mohamad; Antoine, Pascal; Nandrino, Jean Louis

    2016-04-01

    There is a substantial body of research on the relationship between emotion and autobiographical memory. Using facial analysis software, our study addressed this relationship by investigating basic emotional facial expressions that may be detected during autobiographical recall. Participants were asked to retrieve 3 autobiographical memories, each of which was triggered by one of the following cue words: happy, sad, and city. The autobiographical recall was analyzed by a software for facial analysis that detects and classifies basic emotional expressions. Analyses showed that emotional cues triggered the corresponding basic facial expressions (i.e., happy facial expression for memories cued by happy). Furthermore, we dissociated episodic and semantic retrieval, observing more emotional facial expressions during episodic than during semantic retrieval, regardless of the emotional valence of cues. Our study provides insight into facial expressions that are associated with emotional autobiographical memory. It also highlights an ecological tool to reveal physiological changes that are associated with emotion and memory.

  15. Does Andrews facial analysis predict esthetic sagittal maxillary position?

    PubMed

    Resnick, Cory M; Daniels, Kimberly M; Vlahos, Maryann

    2018-04-01

    Cephalometric analyses have limited utility in planning maxillary sagittal position for orthognathic surgery. In Six Elements of Orofacial Harmony, Andrews quantified maxillary position relative to forehead projection and angulation and proposed an ideal relationship. The purpose of this study was to investigate the ability of this technique to predict esthetic sagittal maxillary position. Survey study including a male and female with straight facial profiles, normal maxillary incisor angulations, and Angle's Class I. Maxillary position was modified on lateral photographs to create 5 images for each participant with incisor-goal anterior limit line (GALL) distances of -4, -2, 0, +2, and +4 mm. A series of health care professionals and laypeople were asked to rate each photo in order of attractiveness. A total of 100 complete responses were received. Incisor-GALL distances of +4 mm (41%) and +2 mm (40%) were most commonly considered "most esthetic" for the female volunteer (P < .001). For the male volunteer, there were 2 peak "most esthetic" responses: incisor-GALL distances of 0 mm (37%) and -4 mm (32%) (P < .001). Respondents considered maxillary incisor position 2 to 4 mm anterior to GALL most attractive in a woman and 0 to 4 mm posterior to GALL most esthetic in a man. Using these modified target distances, this analysis may be useful for orthognathic surgery planning. Copyright © 2018 Elsevier Inc. All rights reserved.

  16. Effects of Orientation on Recognition of Facial Affect

    NASA Technical Reports Server (NTRS)

    Cohen, M. M.; Mealey, J. B.; Hargens, Alan R. (Technical Monitor)

    1997-01-01

    The ability to discriminate facial features is often degraded when the orientation of the face and/or the observer is altered. Previous studies have shown that gross distortions of facial features can go unrecognized when the image of the face is inverted, as exemplified by the 'Margaret Thatcher' effect. This study examines how quickly erect and supine observers can distinguish between smiling and frowning faces that are presented at various orientations. The effects of orientation are of particular interest in space, where astronauts frequently view one another in orientations other than the upright. Sixteen observers viewed individual facial images of six people on a computer screen; on a given trial, the image was either smiling or frowning. Each image was viewed when it was erect and when it was rotated (rolled) by 45 degrees, 90 degrees, 135 degrees, 180 degrees, 225 degrees and 270 degrees about the line of sight. The observers were required to respond as rapidly and accurately as possible to identify if the face presented was smiling or frowning. Measures of reaction time were obtained when the observers were both upright and supine. Analyses of variance revealed that mean reaction time, which increased with stimulus rotation (F=18.54, df 7/15, p (is less than) 0.001), was 22% longer when the faces were inverted than when they were erect, but that the orientation of the observer had no significant effect on reaction time (F=1.07, df 1/15, p (is greater than) .30). These data strongly suggest that the orientation of the image of a face on the observer's retina, but not its orientation with respect to gravity, is important in identifying the expression on the face.

  17. Cognitive factors associated with facial pain.

    PubMed

    Schwartz, S M; Gramling, S E

    1997-07-01

    Most well-accepted etiological models of facial pain (e.g., temporomandibular disorders and headache) implicate emotional distress as an important factor in the development and maintenance of pain. Data exists to support the notion that some facial pain sufferers are more emotionally distressed than no pain controls. However, many of these dependent measures of emotional distress are either lengthy assessment batteries, lack clear cut psychotherapeutic treatment implications, or focus exclusively on pain related sequela. As cognitive-behavioral interventions become more integrated into the treatment of chronic pain conditions, including various facial pain conditions, it becomes more imperative that the tools used to assess psychological functioning provide the clinician with specific cognitive/behavioral targets for change. The purpose of this study was to assess the degree to which symptomatic treatment seeking facial pain sufferers (N = 25), symptomatic non-treatment seeking facial pain sufferers (N = 48), and healthy pain-free controls (N = 70) differed on the Rational Beliefs Inventory (RBI). The RBI is a reliable, valid questionnaire assessing rational beliefs that are operationalized within a Rational Emotive Therapy (RET) framework. RET is a cognitive-behavioral treatment paradigm that focuses on how an individual's maladaptive cognitive errors or distortions exacerbate emotional distress. Group differences were assessed using a oneway Analysis of Covariance (ANCOVA) with the total RBI score serving as the dependent measure, and a Multivariate Analysis of Covariance (MANCOVA) using individual RBI belief subscales as dependent measures. These results indicated that groups differed significantly on the total score and several of the individual belief subscales. These findings indicated that facial pain sufferers generally hold maladaptive beliefs that may be of clinical significance for cognitive/behavioral treatment approaches.

  18. Facial anthropometric measurements in Iranian male workers using Digimizer version 4.1.1.0 image analysis software: a pilot study.

    PubMed

    Salvarzi, Elham; Choobineh, Alireza; Jahangiri, Mehdi; Keshavarzi, Sareh

    2018-02-26

    Craniometry is a subset of anthropometry, which measures the anatomical sizes of the head and face (craniofacial indicators). These dimensions are used in designing devices applied in the facial area, including respirators. This study was conducted to measure craniofacial dimensions of Iranian male workers required for face protective equipment design. In this study, facial anthropometric dimensions of 50 randomly selected Iranian male workers were measured by photographic method and Digimizer version 4.1.1.0. Ten facial dimensions were extracted from photographs and measured by Digimizer version 4.1.1.0. Mean, standard deviation and 5th, 50th and 95th percentiles for each dimension were determined and the relevant data bank was established. The anthropometric data bank for the 10 dimensions required for respirator design was provided for the target group with photo-anthropometric methods. The results showed that Iranian face dimensions were different from those of other nations and ethnicities. In this pilot study, anthropometric dimensions required for half-mask respirator design for Iranian male workers were measured by Digimizer version 4.1.1.0. The obtained anthropometric tables could be useful for the design of personal face protective equipment.

  19. The association between PTSD and facial affect recognition.

    PubMed

    Williams, Christian L; Milanak, Melissa E; Judah, Matt R; Berenbaum, Howard

    2018-05-05

    The major aims of this study were to examine how, if at all, having higher levels of PTSD would be associated with performance on a facial affect recognition task in which facial expressions of emotion are superimposed on emotionally valenced, non-face images. College students with trauma histories (N = 90) completed a facial affect recognition task as well as measures of exposure to traumatic events, and PTSD symptoms. When the face and context matched, participants with higher levels of PTSD were significantly more accurate. When the face and context were mismatched, participants with lower levels of PTSD were more accurate than were those with higher levels of PTSD. These findings suggest that PTSD is associated with how people process affective information. Furthermore, these results suggest that the enhanced attention of people with higher levels of PTSD to affective information can be either beneficial or detrimental to their ability to accurately identify facial expressions of emotion. Limitations, future directions and clinical implications are discussed. Copyright © 2018 Elsevier B.V. All rights reserved.

  20. Facial Age Synthesis Using Sparse Partial Least Squares (The Case of Ben Needham).

    PubMed

    Bukar, Ali M; Ugail, Hassan

    2017-09-01

    Automatic facial age progression (AFAP) has been an active area of research in recent years. This is due to its numerous applications which include searching for missing. This study presents a new method of AFAP. Here, we use an active appearance model (AAM) to extract facial features from available images. An aging function is then modelled using sparse partial least squares regression (sPLS). Thereafter, the aging function is used to render new faces at different ages. To test the accuracy of our algorithm, extensive evaluation is conducted using a database of 500 face images with known ages. Furthermore, the algorithm is used to progress Ben Needham's facial image that was taken when he was 21 months old to the ages of 6, 14, and 22 years. The algorithm presented in this study could potentially be used to enhance the search for missing people worldwide. © 2017 American Academy of Forensic Sciences.

  1. The Chinese Facial Emotion Recognition Database (CFERD): a computer-generated 3-D paradigm to measure the recognition of facial emotional expressions at different intensities.

    PubMed

    Huang, Charles Lung-Cheng; Hsiao, Sigmund; Hwu, Hai-Gwo; Howng, Shen-Long

    2012-12-30

    The Chinese Facial Emotion Recognition Database (CFERD), a computer-generated three-dimensional (3D) paradigm, was developed to measure the recognition of facial emotional expressions at different intensities. The stimuli consisted of 3D colour photographic images of six basic facial emotional expressions (happiness, sadness, disgust, fear, anger and surprise) and neutral faces of the Chinese. The purpose of the present study is to describe the development and validation of CFERD with nonclinical healthy participants (N=100; 50 men; age ranging between 18 and 50 years), and to generate normative data set. The results showed that the sensitivity index d' [d'=Z(hit rate)-Z(false alarm rate), where function Z(p), p∈[0,1

  2. Restoration of the Donor Face After Facial Allotransplantation

    PubMed Central

    Grant, Gerald T.; Liacouras, Peter; Santiago, Gabriel F.; Garcia, Juan R.; Al Rakan, Mohammed; Murphy, Ryan; Armand, Mehran; Gordon, Chad R.

    2014-01-01

    Introduction Current protocols for facial transplantation include the mandatory fabrication of an alloplastic “mask” to restore the congruency of the donor site in the setting of “open casket” burial. However, there is currently a paucity of literature describing the current state-of-the-art and available options. Methods During this study, we identified that most of donor masks are fabricated using conventional methods of impression, molds, silicone, and/or acrylic application by an experienced anaplastologist or maxillofacial prosthetics technician. However, with the recent introduction of several enhanced computer-assisted technologies, our facial transplant team hypothesized that there were areas for improvement with respect to cost and preparation time. Results The use of digital imaging for virtual surgical manipulation, computer-assisted planning, and prefabricated surgical cutting guides—in the setting of facial transplantation—provided us a novel opportunity for digital design and fabrication of a donor mask. The results shown here demonstrate an acceptable appearance for “open-casket” burial while maintaining donor identity after facial organ recovery. Conclusions Several newer techniques for fabrication of facial transplant donor masks exist currently and are described within the article. These encompass digital impression, digital design, and additive manufacturing technology. PMID:24835867

  3. [Idiopathic facial paralysis in children].

    PubMed

    Achour, I; Chakroun, A; Ayedi, S; Ben Rhaiem, Z; Mnejja, M; Charfeddine, I; Hammami, B; Ghorbel, A

    2015-05-01

    Idiopathic facial palsy is the most common cause of facial nerve palsy in children. Controversy exists regarding treatment options. The objectives of this study were to review the epidemiological and clinical characteristics as well as the outcome of idiopathic facial palsy in children to suggest appropriate treatment. A retrospective study was conducted on children with a diagnosis of idiopathic facial palsy from 2007 to 2012. A total of 37 cases (13 males, 24 females) with a mean age of 13.9 years were included in this analysis. The mean duration between onset of Bell's palsy and consultation was 3 days. Of these patients, 78.3% had moderately severe (grade IV) or severe paralysis (grade V on the House and Brackmann grading). Twenty-seven patients were treated in an outpatient context, three patients were hospitalized, and seven patients were treated as outpatients and subsequently hospitalized. All patients received corticosteroids. Eight of them also received antiviral treatment. The complete recovery rate was 94.6% (35/37). The duration of complete recovery was 7.4 weeks. Children with idiopathic facial palsy have a very good prognosis. The complete recovery rate exceeds 90%. However, controversy exists regarding treatment options. High-quality studies have been conducted on adult populations. Medical treatment based on corticosteroids alone or combined with antiviral treatment is certainly effective in improving facial function outcomes in adults. In children, the recommendation for prescription of steroids and antiviral drugs based on adult treatment appears to be justified. Randomized controlled trials in the pediatric population are recommended to define a strategy for management of idiopathic facial paralysis. Copyright © 2015 Elsevier Masson SAS. All rights reserved.

  4. Improved Facial Nerve Identification During Parotidectomy With Fluorescently Labeled Peptide

    PubMed Central

    Hussain, Timon; Nguyen, Linda T.; Whitney, Michael; Hasselmann, Jonathan; Nguyen, Quyen T.

    2016-01-01

    Objectives/Hypothesis Additional intraoperative guidance could reduce the risk of iatrogenic injury during parotid gland cancer surgery. We evaluated the intraoperative use of fluorescently labeled nerve binding peptide NP41 to aid facial nerve identification and preservation during parotidectomy in an orthotopic model of murine parotid gland cancer. We also quantified the accuracy of intraoperative nerve detection for surface and buried nerves in the head and neck with NP41 versus white light (WL) alone. Study Design Twenty-eight mice underwent parotid gland cancer surgeries with additional fluorescence (FL) guidance versus WL reflectance (WLR) alone. Eight mice were used for additional nerve-imaging experiments. Methods Twenty-eight parotid tumor-bearing mice underwent parotidectomy. Eight mice underwent imaging of both sides of the face after skin removal. Postoperative assessment of facial nerve function measured by automated whisker tracking were compared between FL guidance (n = 13) versus WL alone (n = 15). In eight mice, nerve to surrounding tissue contrast was measured under FL versus WLR for all nerve branches detectable in the field of view. Results Postoperative facial nerve function after parotid gland cancer surgery tended to be better with additional FL guidance. Fluorescent labeling significantly improved nerve to surrounding tissue contrast for both large and smaller buried nerve branches compared to WLR visualization and improved detection sensitivity and specificity. Conclusions NP41 FL imaging significantly aids the intraoperative identification of nerve braches otherwise nearly invisible to the naked eye. Its application in a murine model of parotid gland cancer surgery tended to improve functional preservation of the facial nerve. PMID:27171862

  5. Overview of pediatric peripheral facial nerve paralysis: analysis of 40 patients.

    PubMed

    Özkale, Yasemin; Erol, İlknur; Saygı, Semra; Yılmaz, İsmail

    2015-02-01

    Peripheral facial nerve paralysis in children might be an alarming sign of serious disease such as malignancy, systemic disease, congenital anomalies, trauma, infection, middle ear surgery, and hypertension. The cases of 40 consecutive children and adolescents who were diagnosed with peripheral facial nerve paralysis at Baskent University Adana Hospital Pediatrics and Pediatric Neurology Unit between January 2010 and January 2013 were retrospectively evaluated. We determined that the most common cause was Bell palsy, followed by infection, tumor lesion, and suspected chemotherapy toxicity. We noted that younger patients had generally poorer outcome than older patients regardless of disease etiology. Peripheral facial nerve paralysis has been reported in many countries in America and Europe; however, knowledge about its clinical features, microbiology, neuroimaging, and treatment in Turkey is incomplete. The present study demonstrated that Bell palsy and infection were the most common etiologies of peripheral facial nerve paralysis. © The Author(s) 2014.

  6. Perceived functional impact of abnormal facial appearance.

    PubMed

    Rankin, Marlene; Borah, Gregory L

    2003-06-01

    Functional facial deformities are usually described as those that impair respiration, eating, hearing, or speech. Yet facial scars and cutaneous deformities have a significant negative effect on social functionality that has been poorly documented in the scientific literature. Insurance companies are declining payments for reconstructive surgical procedures for facial deformities caused by congenital disabilities and after cancer or trauma operations that do not affect mechanical facial activity. The purpose of this study was to establish a large, sample-based evaluation of the perceived social functioning, interpersonal characteristics, and employability indices for a range of facial appearances (normal and abnormal). Adult volunteer evaluators (n = 210) provided their subjective perceptions based on facial physical appearance, and an analysis of the consequences of facial deformity on parameters of preferential treatment was performed. A two-group comparative research design rated the differences among 10 examples of digitally altered facial photographs of actual patients among various age and ethnic groups with "normal" and "abnormal" congenital deformities or posttrauma scars. Photographs of adult patients with observable congenital and posttraumatic deformities (abnormal) were digitally retouched to eliminate the stigmatic defects (normal). The normal and abnormal photographs of identical patients were evaluated by the large sample study group on nine parameters of social functioning, such as honesty, employability, attractiveness, and effectiveness, using a visual analogue rating scale. Patients with abnormal facial characteristics were rated as significantly less honest (p = 0.007), less employable (p = 0.001), less trustworthy (p = 0.01), less optimistic (p = 0.001), less effective (p = 0.02), less capable (p = 0.002), less intelligent (p = 0.03), less popular (p = 0.001), and less attractive (p = 0.001) than were the same patients with normal facial

  7. Facial measurement differences between patients with schizophrenia and non-psychiatric controls.

    PubMed

    Compton, Michael T; Brudno, Jennifer; Kryda, Aimee D; Bollini, Annie M; Walker, Elaine F

    2007-07-01

    Several previous reports suggest that facial measurements in patients with schizophrenia differ from those of non-psychiatric controls. Because the face and brain develop in concert from the same ectodermal tissue, the study of quantitative craniofacial abnormalities may give clues to genetic and/or environmental factors predisposing to schizophrenia. Using a predominantly African American sample, the present research question was two-fold: (1) Do patients differ from controls in terms of a number of specific facial measurements?, and (2) Does cluster analysis based on these facial measurements reveal distinct facial morphologies that significantly discriminate patients from controls? Facial dimensions were measured in 73 patients with schizophrenia and related psychotic disorders (42 males and 31 females) and 69 non-psychiatric controls (35 males and 34 females) using a 25-cm head and neck caliper. Due to differences in facial dimensions by gender, separate independent samples Student's t-tests and logistic regression analyses were employed to discern differences in facial measures between the patient and control groups in women and men. Findings were further explored using cluster analysis. Given an association between age and some facial dimensions, the effect of age was controlled. In unadjusted bivariate tests, female patients differed from female controls on several facial dimensions, though male patients did not differ significantly from male controls for any facial measure. Controlling for age using logistic regression, female patients had a greater mid-facial depth (tragus-subnasale) compared to female controls; male patients had lesser upper facial (trichion-glabella) and lower facial (subnasale-gnathion) heights compared to male controls. Among females, cluster analysis revealed two facial morphologies that significantly discriminated patients from controls, though this finding was not evident when employing further cluster analyses using secondary

  8. [Neurological disease and facial recognition].

    PubMed

    Kawamura, Mitsuru; Sugimoto, Azusa; Kobayakawa, Mutsutaka; Tsuruya, Natsuko

    2012-07-01

    To discuss the neurological basis of facial recognition, we present our case reports of impaired recognition and a review of previous literature. First, we present a case of infarction and discuss prosopagnosia, which has had a large impact on face recognition research. From a study of patient symptoms, we assume that prosopagnosia may be caused by unilateral right occipitotemporal lesion and right cerebral dominance of facial recognition. Further, circumscribed lesion and degenerative disease may also cause progressive prosopagnosia. Apperceptive prosopagnosia is observed in patients with posterior cortical atrophy (PCA), pathologically considered as Alzheimer's disease, and associative prosopagnosia in frontotemporal lobar degeneration (FTLD). Second, we discuss face recognition as part of communication. Patients with Parkinson disease show social cognitive impairments, such as difficulty in facial expression recognition and deficits in theory of mind as detected by the reading the mind in the eyes test. Pathological and functional imaging studies indicate that social cognitive impairment in Parkinson disease is possibly related to damages in the amygdalae and surrounding limbic system. The social cognitive deficits can be observed in the early stages of Parkinson disease, and even in the prodromal stage, for example, patients with rapid eye movement (REM) sleep behavior disorder (RBD) show impairment in facial expression recognition. Further, patients with myotonic dystrophy type 1 (DM 1), which is a multisystem disease that mainly affects the muscles, show social cognitive impairment similar to that of Parkinson disease. Our previous study showed that facial expression recognition impairment of DM 1 patients is associated with lesion in the amygdalae and insulae. Our study results indicate that behaviors and personality traits in DM 1 patients, which are revealed by social cognitive impairment, are attributable to dysfunction of the limbic system.

  9. Heritabilities of Facial Measurements and Their Latent Factors in Korean Families

    PubMed Central

    Kim, Hyun-Jin; Im, Sun-Wha; Jargal, Ganchimeg; Lee, Siwoo; Yi, Jae-Hyuk; Park, Jeong-Yeon; Sung, Joohon; Cho, Sung-Il; Kim, Jong-Yeol; Kim, Jong-Il; Seo, Jeong-Sun

    2013-01-01

    Genetic studies on facial morphology targeting healthy populations are fundamental in understanding the specific genetic influences involved; yet, most studies to date, if not all, have been focused on congenital diseases accompanied by facial anomalies. To study the specific genetic cues determining facial morphology, we estimated familial correlations and heritabilities of 14 facial measurements and 3 latent factors inferred from a factor analysis in a subset of the Korean population. The study included a total of 229 individuals from 38 families. We evaluated a total of 14 facial measurements using 2D digital photographs. We performed factor analysis to infer common latent variables. The heritabilities of 13 facial measurements were statistically significant (p < 0.05) and ranged from 0.25 to 0.61. Of these, the heritability of intercanthal width in the orbital region was found to be the highest (h2 = 0.61, SE = 0.14). Three factors (lower face portion, orbital region, and vertical length) were obtained through factor analysis, where the heritability values ranged from 0.45 to 0.55. The heritability values for each factor were higher than the mean heritability value of individual original measurements. We have confirmed the genetic influence on facial anthropometric traits and suggest a potential way to categorize and analyze the facial portions into different groups. PMID:23843774

  10. Facial expression recognition under partial occlusion based on fusion of global and local features

    NASA Astrophysics Data System (ADS)

    Wang, Xiaohua; Xia, Chen; Hu, Min; Ren, Fuji

    2018-04-01

    Facial expression recognition under partial occlusion is a challenging research. This paper proposes a novel framework for facial expression recognition under occlusion by fusing the global and local features. In global aspect, first, information entropy are employed to locate the occluded region. Second, principal Component Analysis (PCA) method is adopted to reconstruct the occlusion region of image. After that, a replace strategy is applied to reconstruct image by replacing the occluded region with the corresponding region of the best matched image in training set, Pyramid Weber Local Descriptor (PWLD) feature is then extracted. At last, the outputs of SVM are fitted to the probabilities of the target class by using sigmoid function. For the local aspect, an overlapping block-based method is adopted to extract WLD features, and each block is weighted adaptively by information entropy, Chi-square distance and similar block summation methods are then applied to obtain the probabilities which emotion belongs to. Finally, fusion at the decision level is employed for the data fusion of the global and local features based on Dempster-Shafer theory of evidence. Experimental results on the Cohn-Kanade and JAFFE databases demonstrate the effectiveness and fault tolerance of this method.

  11. Three-Dimensional Accuracy of Facial Scan for Facial Deformities in Clinics: A New Evaluation Method for Facial Scanner Accuracy.

    PubMed

    Zhao, Yi-Jiao; Xiong, Yu-Xue; Wang, Yong

    2017-01-01

    In this study, the practical accuracy (PA) of optical facial scanners for facial deformity patients in oral clinic was evaluated. Ten patients with a variety of facial deformities from oral clinical were included in the study. For each patient, a three-dimensional (3D) face model was acquired, via a high-accuracy industrial "line-laser" scanner (Faro), as the reference model and two test models were obtained, via a "stereophotography" (3dMD) and a "structured light" facial scanner (FaceScan) separately. Registration based on the iterative closest point (ICP) algorithm was executed to overlap the test models to reference models, and "3D error" as a new measurement indicator calculated by reverse engineering software (Geomagic Studio) was used to evaluate the 3D global and partial (upper, middle, and lower parts of face) PA of each facial scanner. The respective 3D accuracy of stereophotography and structured light facial scanners obtained for facial deformities was 0.58±0.11 mm and 0.57±0.07 mm. The 3D accuracy of different facial partitions was inconsistent; the middle face had the best performance. Although the PA of two facial scanners was lower than their nominal accuracy (NA), they all met the requirement for oral clinic use.

  12. An Automatic Diagnosis Method of Facial Acne Vulgaris Based on Convolutional Neural Network.

    PubMed

    Shen, Xiaolei; Zhang, Jiachi; Yan, Chenjun; Zhou, Hong

    2018-04-11

    In this paper, we present a new automatic diagnosis method for facial acne vulgaris which is based on convolutional neural networks (CNNs). To overcome the shortcomings of previous methods which were the inability to classify enough types of acne vulgaris. The core of our method is to extract features of images based on CNNs and achieve classification by classifier. A binary-classifier of skin-and-non-skin is used to detect skin area and a seven-classifier is used to achieve the classification task of facial acne vulgaris and healthy skin. In the experiments, we compare the effectiveness of our CNN and the VGG16 neural network which is pre-trained on the ImageNet data set. We use a ROC curve to evaluate the performance of binary-classifier and use a normalized confusion matrix to evaluate the performance of seven-classifier. The results of our experiments show that the pre-trained VGG16 neural network is effective in extracting features from facial acne vulgaris images. And the features are very useful for the follow-up classifiers. Finally, we try applying the classifiers both based on the pre-trained VGG16 neural network to assist doctors in facial acne vulgaris diagnosis.

  13. Body size and allometric variation in facial shape in children.

    PubMed

    Larson, Jacinda R; Manyama, Mange F; Cole, Joanne B; Gonzalez, Paula N; Percival, Christopher J; Liberton, Denise K; Ferrara, Tracey M; Riccardi, Sheri L; Kimwaga, Emmanuel A; Mathayo, Joshua; Spitzmacher, Jared A; Rolian, Campbell; Jamniczky, Heather A; Weinberg, Seth M; Roseman, Charles C; Klein, Ophir; Lukowiak, Ken; Spritz, Richard A; Hallgrimsson, Benedikt

    2018-02-01

    Morphological integration, or the tendency for covariation, is commonly seen in complex traits such as the human face. The effects of growth on shape, or allometry, represent a ubiquitous but poorly understood axis of integration. We address the question of to what extent age and measures of size converge on a single pattern of allometry for human facial shape. Our study is based on two large cross-sectional cohorts of children, one from Tanzania and the other from the United States (N = 7,173). We employ 3D facial imaging and geometric morphometrics to relate facial shape to age and anthropometric measures. The two populations differ significantly in facial shape, but the magnitude of this difference is small relative to the variation within each group. Allometric variation for facial shape is similar in both populations, representing a small but significant proportion of total variation in facial shape. Different measures of size are associated with overlapping but statistically distinct aspects of shape variation. Only half of the size-related variation in facial shape can be explained by the first principal component of four size measures and age while the remainder associates distinctly with individual measures. Allometric variation in the human face is complex and should not be regarded as a singular effect. This finding has important implications for how size is treated in studies of human facial shape and for the developmental basis for allometric variation more generally. © 2017 Wiley Periodicals, Inc.

  14. Analysis of 809 Facial Bone Fractures in a Pediatric and Adolescent Population

    PubMed Central

    Kim, Sang Hun; Lee, Soo Hyang

    2012-01-01

    Background Facial fractures are infrequent in children and adolescents and have different clinical features from those in adults. The low incidence in children and adolescents reflects the flexibility and underdevelopment of their facial skeletons, as well as their more protected environments. Only a few reports have reviewed such patients in Korea. The authors performed a retrospective study to analyze the characteristics of facial fractures in the Korean pediatric population. Methods We conducted a retrospective review on a series of 741 patients, aged <18 years, with facial fractures who had been treated at our hospital between 2006 and 2010. The following parameters were evaluated: age, sex, cause, location and type of fractures, associated injuries, treatment and complications. Results A total of 741 consecutive patients met the inclusion criteria. The ratio of boys to girls was 5.7:1. Facial fractures most commonly occurred in patients between 13 and 15 years of age (36.3%). The most common causes of injury was violence. The nasal fracture was the most common type of fracture (69%) and the blowout fracture was the second most common (20%). Associated injuries occurred in 156 patients (21%). Conclusions The incidence of pediatric facial fractures caused by violence is high in Korea. Our results show that as age increases, etiological factors and fracture patterns gradually shift towards those found in adults. This study provides an overview of facial fractures in these age groups that helps illustrate the trends and characteristics of the fractures and may be helpful in further evaluation and management. PMID:23233885

  15. Under pressure: progressively enlarging facial mass following high-pressure paint injection injury.

    PubMed

    Mushtaq, Jameel; Walker, Abigail; Hunter, Ben

    2016-01-19

    High-pressure paint injection injuries are relatively rare industrial accidents and almost exclusively occur on the non-dominant hand. A rarely documented complication of these injuries is the formation of a foreign body granuloma. We report a case of a 33-year-old man presenting with extensive facial scarring and progressive right paranasal swelling 7 years after a high-pressure paint injury. After imaging investigations, an excision of the mass and revision of scarring was performed. Access to the mass was gained indirectly through existing scarring over the nose to ensure an aesthetic result. Histological analysis revealed a florid granulomatous foreign body reaction to retained paint. To the best of our knowledge, this is the first reported case of a facial high-pressure paint injury with consequent formation of a foreign body granuloma. 2016 BMJ Publishing Group Ltd.

  16. Capturing Physiology of Emotion along Facial Muscles: A Method of Distinguishing Feigned from Involuntary Expressions

    NASA Astrophysics Data System (ADS)

    Khan, Masood Mehmood; Ward, Robert D.; Ingleby, Michael

    The ability to distinguish feigned from involuntary expressions of emotions could help in the investigation and treatment of neuropsychiatric and affective disorders and in the detection of malingering. This work investigates differences in emotion-specific patterns of thermal variations along the major facial muscles. Using experimental data extracted from 156 images, we attempted to classify patterns of emotion-specific thermal variations into neutral, and voluntary and involuntary expressions of positive and negative emotive states. Initial results suggest (i) each facial muscle exhibits a unique thermal response to various emotive states; (ii) the pattern of thermal variances along the facial muscles may assist in classifying voluntary and involuntary facial expressions; and (iii) facial skin temperature measurements along the major facial muscles may be used in automated emotion assessment.

  17. 3D imaging acquisition, modeling, and prototyping for facial defects reconstruction

    NASA Astrophysics Data System (ADS)

    Sansoni, Giovanna; Trebeschi, Marco; Cavagnini, Gianluca; Gastaldi, Giorgio

    2009-01-01

    A novel approach that combines optical three-dimensional imaging, reverse engineering (RE) and rapid prototyping (RP) for mold production in the prosthetic reconstruction of facial prostheses is presented. A commercial laser-stripe digitizer is used to perform the multiview acquisition of the patient's face; the point clouds are aligned and merged in order to obtain a polygonal model, which is then edited to sculpture the virtual prothesis. Two physical models of both the deformed face and the 'repaired' face are obtained: they differ only in the defect zone. Depending on the material used for the actual prosthesis, the two prototypes can be used either to directly cast the final prosthesis or to fabricate the positive wax pattern. Two case studies are presented, referring to prostetic reconstructions of an eye and of a nose. The results demonstrate the advantages over conventional techniques as well as the improvements with respect to known automated manufacturing techniques in the mold construction. The proposed method results into decreased patient's disconfort, reduced dependence on the anaplasthologist skill, increased repeatability and efficiency of the whole process.

  18. Diminished neural and cognitive responses to facial expressions of disgust in patients with psoriasis: a functional magnetic resonance imaging study.

    PubMed

    Kleyn, C Elise; McKie, Shane; Ross, Andrew R; Montaldi, Daniela; Gregory, Lloyd J; Elliott, Rebecca; Isaacs, Clare L; Anderson, Ian M; Richards, Helen L; Deakin, J F William; Fortune, Donal G; Griffiths, Christopher E M

    2009-11-01

    Psoriasis produces significant psychosocial disability; however, little is understood about the neurocognitive mechanisms that mediate the adverse consequences of the social stigma associated with visible skin lesions, such as disgusted facial expressions of others. Both the feeling of disgust and the observation of disgust in others are known to activate the insula cortex. We investigated whether the social impact of psoriasis is associated with altered cognitive processing of disgust using (i) a covert recognition of faces task conducted using functional magnetic resonance imaging (fMRI) and (ii) the facial expression recognition task (FERT), a decision-making task, conducted outside the scanner to assess the ability to recognize overtly different intensities of disgust. Thirteen right-handed male patients with psoriasis and 13 age-matched male controls were included. In the fMRI study, psoriasis patients had significantly (P<0.005) smaller signal responses to disgusted faces in the bilateral insular cortex compared with healthy controls. These data were corroborated by FERT, in that patients were less able than controls to identify all intensities of disgust tested. We hypothesize that patients with psoriasis, in this case male patients, develop a coping mechanism to protect them from stressful emotional responses by blocking the processing of disgusted facial expressions.

  19. Development of facial aging simulation system combined with three-dimensional shape prediction from facial photographs

    NASA Astrophysics Data System (ADS)

    Nagata, Takeshi; Matsuzaki, Kazutoshi; Taniguchi, Kei; Ogawa, Yoshinori; Imaizumi, Kazuhiko

    2017-03-01

    3D Facial aging changes in more than 10 years of identical persons are being measured at National Research Institute of Police Science. We performed machine learning using such measured data as teacher data and have developed the system which convert input 2D face image into 3D face model and simulate aging. Here, we report about processing and accuracy of our system.

  20. Facial attractiveness, symmetry, and physical fitness in young women.

    PubMed

    Hönekopp, Johannes; Bartholomé, Tobias; Jansen, Gregor

    2004-06-01

    This study explores the evolutionary-based hypothesis that facial attractiveness (a guiding force in mate selection) is a cue for physical fitness (presumably an important contributor to mate value in ancestral times). Since fluctuating asymmetry, a measure of developmental stability, is known to be a valid cue for fitness in several biological domains, we scrutinized facial asymmetry as a potential mediator between attractiveness and fitness. In our sample of young women, facial beauty indeed indicated physical fitness. The relationships that pertained to asymmetry were in the expected direction. However, a closer analysis revealed that facial asymmetry did not mediate the relationship between fitness and attractiveness. Unexpected problems regarding the measurement of facial asymmetry are discussed.

  1. Effect of Restraining Devices on Facial Fractures in Motor Vehicle Collisions.

    PubMed

    Hwang, Kun; Kim, Joo Ho

    2015-09-01

    The aim of this systematic review is to summarize and critically evaluate the evidence for or against the effectiveness of restraining devices on facial fractures in motor vehicle collisions (MVCs).In a PubMed search, the search terms "facial bone fracture and seat belt," "facial bone fracture and air bag," and "facial bone fracture and restraining" were used. The authors abstracted the odds ratio (OR) and 95% confidence intervals (CIs) from each study. Weighted mean differences and 95% CIs were also calculated. The statistical analysis was performed with Review Manager (The Nordic Cochrane Centre).The authors found 30 potentially relevant articles, of which 6 articles met our inclusion criteria. Five studies were subgrouped, and a meta-analysis of these data suggested beneficial effects of seat belts on decreasing facial fractures in MVCs (n = 15,768,960, OR, 0.46, 95% CI = 0.35-0.60). Three studies were subgrouped, and a meta-analysis of these data suggested that there were beneficial effects of seat belts and air bags on decreasing facial fractures in MVCs (n = 15,768,021, OR, 0.59, 95% CI = 0.47-0.74). Four studies were subgrouped, and a meta-analysis of these data suggested there were no significant effects of an air bag on decreasing facial fracture in MVCs (n = 15,932,259, OR, 1.00, 95% CI = 0.72-1.39).A seat belt alone (OR, 0.46) or a seat belt and an air bag (OR, 0.59) were effective to decrease facial fractures in MVCs. However, air bags alone had no significant effect (OR, 1.00). In using air bags, seat belt should be applied together to prevent facial fractures in motor vehicle injuries.

  2. Genetic Factors That Increase Male Facial Masculinity Decrease Facial Attractiveness of Female Relatives

    PubMed Central

    Lee, Anthony J.; Mitchem, Dorian G.; Wright, Margaret J.; Martin, Nicholas G.; Keller, Matthew C.; Zietsch, Brendan P.

    2014-01-01

    For women, choosing a facially masculine man as a mate is thought to confer genetic benefits to offspring. Crucial assumptions of this hypothesis have not been adequately tested. It has been assumed that variation in facial masculinity is due to genetic variation and that genetic factors that increase male facial masculinity do not increase facial masculinity in female relatives. We objectively quantified the facial masculinity in photos of identical (n = 411) and nonidentical (n = 782) twins and their siblings (n = 106). Using biometrical modeling, we found that much of the variation in male and female facial masculinity is genetic. However, we also found that masculinity of male faces is unrelated to their attractiveness and that facially masculine men tend to have facially masculine, less-attractive sisters. These findings challenge the idea that facially masculine men provide net genetic benefits to offspring and call into question this popular theoretical framework. PMID:24379153

  3. Genetic factors that increase male facial masculinity decrease facial attractiveness of female relatives.

    PubMed

    Lee, Anthony J; Mitchem, Dorian G; Wright, Margaret J; Martin, Nicholas G; Keller, Matthew C; Zietsch, Brendan P

    2014-02-01

    For women, choosing a facially masculine man as a mate is thought to confer genetic benefits to offspring. Crucial assumptions of this hypothesis have not been adequately tested. It has been assumed that variation in facial masculinity is due to genetic variation and that genetic factors that increase male facial masculinity do not increase facial masculinity in female relatives. We objectively quantified the facial masculinity in photos of identical (n = 411) and nonidentical (n = 782) twins and their siblings (n = 106). Using biometrical modeling, we found that much of the variation in male and female facial masculinity is genetic. However, we also found that masculinity of male faces is unrelated to their attractiveness and that facially masculine men tend to have facially masculine, less-attractive sisters. These findings challenge the idea that facially masculine men provide net genetic benefits to offspring and call into question this popular theoretical framework.

  4. Computerized measurement of facial expression of emotions in schizophrenia.

    PubMed

    Alvino, Christopher; Kohler, Christian; Barrett, Frederick; Gur, Raquel E; Gur, Ruben C; Verma, Ragini

    2007-07-30

    Deficits in the ability to express emotions characterize several neuropsychiatric disorders and are a hallmark of schizophrenia, and there is need for a method of quantifying expression, which is currently done by clinical ratings. This paper presents the development and validation of a computational framework for quantifying emotional expression differences between patients with schizophrenia and healthy controls. Each face is modeled as a combination of elastic regions, and expression changes are modeled as a deformation between a neutral face and an expressive face. Functions of these deformations, known as the regional volumetric difference (RVD) functions, form distinctive quantitative profiles of expressions. Employing pattern classification techniques, we have designed expression classifiers for the four universal emotions of happiness, sadness, anger and fear by training on RVD functions of expression changes. The classifiers were cross-validated and then applied to facial expression images of patients with schizophrenia and healthy controls. The classification score for each image reflects the extent to which the expressed emotion matches the intended emotion. Group-wise statistical analysis revealed this score to be significantly different between healthy controls and patients, especially in the case of anger. This score correlated with clinical severity of flat affect. These results encourage the use of such deformation based expression quantification measures for research in clinical applications that require the automated measurement of facial affect.

  5. Quantitative analysis of facial paralysis using local binary patterns in biomedical videos.

    PubMed

    He, Shu; Soraghan, John J; O'Reilly, Brian F; Xing, Dongshan

    2009-07-01

    Facial paralysis is the loss of voluntary muscle movement of one side of the face. A quantitative, objective, and reliable assessment system would be an invaluable tool for clinicians treating patients with this condition. This paper presents a novel framework for objective measurement of facial paralysis. The motion information in the horizontal and vertical directions and the appearance features on the apex frames are extracted based on the local binary patterns (LBPs) on the temporal-spatial domain in each facial region. These features are temporally and spatially enhanced by the application of novel block processing schemes. A multiresolution extension of uniform LBP is proposed to efficiently combine the micropatterns and large-scale patterns into a feature vector. The symmetry of facial movements is measured by the resistor-average distance (RAD) between LBP features extracted from the two sides of the face. Support vector machine is applied to provide quantitative evaluation of facial paralysis based on the House-Brackmann (H-B) scale. The proposed method is validated by experiments with 197 subject videos, which demonstrates its accuracy and efficiency.

  6. Contralateral botulinum toxin injection to improve facial asymmetry after acute facial paralysis.

    PubMed

    Kim, Jin

    2013-02-01

    The application of botulinum toxin to the healthy side of the face in patients with long-standing facial paralysis has been shown to be a minimally invasive technique that improves facial symmetry at rest and during facial motion, but our experience using botulinum toxin therapy for facial sequelae prompted the idea that botulinum toxin might be useful in acute cases of facial paralysis, leading to improve facial asymmetry. In cases in which medical or surgical treatment options are limited because of existing medical problems or advanced age, most patients with acute facial palsy are advised to await spontaneous recovery or are informed that no effective intervention exists. The purpose of this study was to evaluate the effect of botulinum toxin treatment for facial asymmetry in 18 patients after acute facial palsy who could not be optimally treated by medical or surgical management because of severe medical or other problems. From 2009 to 2011, nine patients with Bell's palsy, 5 with herpes zoster oticus and 4 with traumatic facial palsy (10 men and 8 women; age range, 22-82 yr; mean, 50.8 yr) participated in this study. Botulinum toxin A (Botox; Allergan Incorporated, Irvine, CA, USA) was injected using a tuberculin syringe with a 27-gauge needle. The amount injected per site varied from 2.5 to 3 U, and the total dose used per patient was 32 to 68 U (mean, 47.5 +/- 8.4 U). After administration of a single dose of botulinum toxin A on the nonparalyzed side of 18 patients with acute facial paralysis, marked relief of facial asymmetry was observed in 8 patients within 1 month of injection. Decreased facial asymmetry and strengthened facial function on the paralyzed side led to an increased HB and SB grade within 6 months after injection. Use of botulinum toxin after acute facial palsy cases is of great value. Such therapy decreases the relative hyperkinesis contralateral to the paralysis, leading to greater symmetric function. Especially in patients with medical

  7. Analysis on unevenness of skin color using the melanin and hemoglobin components separated by independent component analysis of skin color image

    NASA Astrophysics Data System (ADS)

    Ojima, Nobutoshi; Fujiwara, Izumi; Inoue, Yayoi; Tsumura, Norimichi; Nakaguchi, Toshiya; Iwata, Kayoko

    2011-03-01

    Uneven distribution of skin color is one of the biggest concerns about facial skin appearance. Recently several techniques to analyze skin color have been introduced by separating skin color information into chromophore components, such as melanin and hemoglobin. However, there are not many reports on quantitative analysis of unevenness of skin color by considering type of chromophore, clusters of different sizes and concentration of the each chromophore. We propose a new image analysis and simulation method based on chromophore analysis and spatial frequency analysis. This method is mainly composed of three techniques: independent component analysis (ICA) to extract hemoglobin and melanin chromophores from a single skin color image, an image pyramid technique which decomposes each chromophore into multi-resolution images, which can be used for identifying different sizes of clusters or spatial frequencies, and analysis of the histogram obtained from each multi-resolution image to extract unevenness parameters. As the application of the method, we also introduce an image processing technique to change unevenness of melanin component. As the result, the method showed high capabilities to analyze unevenness of each skin chromophore: 1) Vague unevenness on skin could be discriminated from noticeable pigmentation such as freckles or acne. 2) By analyzing the unevenness parameters obtained from each multi-resolution image for Japanese ladies, agerelated changes were observed in the parameters of middle spatial frequency. 3) An image processing system modulating the parameters was proposed to change unevenness of skin images along the axis of the obtained age-related change in real time.

  8. The Perception of Dynamic and Static Facial Expressions of Happiness and Disgust Investigated by ERPs and fMRI Constrained Source Analysis

    PubMed Central

    Trautmann-Lengsfeld, Sina Alexa; Domínguez-Borràs, Judith; Escera, Carles; Herrmann, Manfred; Fehr, Thorsten

    2013-01-01

    A recent functional magnetic resonance imaging (fMRI) study by our group demonstrated that dynamic emotional faces are more accurately recognized and evoked more widespread patterns of hemodynamic brain responses than static emotional faces. Based on this experimental design, the present study aimed at investigating the spatio-temporal processing of static and dynamic emotional facial expressions in 19 healthy women by means of multi-channel electroencephalography (EEG), event-related potentials (ERP) and fMRI-constrained regional source analyses. ERP analysis showed an increased amplitude of the LPP (late posterior positivity) over centro-parietal regions for static facial expressions of disgust compared to neutral faces. In addition, the LPP was more widespread and temporally prolonged for dynamic compared to static faces of disgust and happiness. fMRI constrained source analysis on static emotional face stimuli indicated the spatio-temporal modulation of predominantly posterior regional brain activation related to the visual processing stream for both emotional valences when compared to the neutral condition in the fusiform gyrus. The spatio-temporal processing of dynamic stimuli yielded enhanced source activity for emotional compared to neutral conditions in temporal (e.g., fusiform gyrus), and frontal regions (e.g., ventromedial prefrontal cortex, medial and inferior frontal cortex) in early and again in later time windows. The present data support the view that dynamic facial displays trigger more information reflected in complex neural networks, in particular because of their changing features potentially triggering sustained activation related to a continuing evaluation of those faces. A combined fMRI and EEG approach thus provides an advanced insight to the spatio-temporal characteristics of emotional face processing, by also revealing additional neural generators, not identifiable by the only use of an fMRI approach. PMID:23818974

  9. The 100 Most Cited Articles in Facial Trauma: A Bibliometric Analysis.

    PubMed

    Tahim, Arpan; Patel, Kush; Bridle, Christopher; Holmes, Simon

    2016-11-01

    The number of citations an article receives has been used as a marker of its influence within a surgical specialty. Currently, there is limited citation analysis in oral and maxillofacial trauma surgery. The purpose of this study was to determine the 100 most cited articles in facial trauma surgery and their characteristics. Articles were identified from the Science Citation Index of the Institute for Scientific Information using the Thomson Reuters Web of Science search engine. All articles until 2015 were included. Then, the 100 most cited articles were assessed for title, author, journal, country of origin, and number of citations. A citation index (number of citations received per year) also was calculated. The 100 most cited articles in facial trauma received 9,933 citations (range, 66 to 297). They were published from 1942 through 2008, with 1990 through 1999 being the commonest decade. Articles were cited on average 4.6 times per year. Articles were published in 28 different journals, with impact factors ranging from 0.94 to 35.3. Most articles were observational research studies. These findings reflect the attention that articles have received during the past half century in oral and maxillofacial trauma research, shedding light on often-read articles in this field. In addition to current bibliometric indices, it could provide a useful evidence base for facial surgeons, represent key educational material for aspiring trainees, and be used to help guide future research efforts. Copyright © 2016 American Association of Oral and Maxillofacial Surgeons. Published by Elsevier Inc. All rights reserved.

  10. Facial dysmorphism in Leigh syndrome with SURF-1 mutation and COX deficiency.

    PubMed

    Yüksel, Adnan; Seven, Mehmet; Cetincelik, Umran; Yeşil, Gözde; Köksal, Vedat

    2006-06-01

    Leigh syndrome is an inherited, progressive neurodegenerative disorder of infancy and childhood. Mutations in the nuclear SURF-1 gene are specifically associated with cytochrome C oxidase-deficient Leigh syndrome. This report describes two patients with similar facial features. One of them was a 2(1/2)-year-old male, and the other was a 3-year-old male with a mutation in SURF-1 gene and facial dysmorphism including frontal bossing, brachycephaly, hypertrichosis, lateral displacement of inner canthi, esotropia, maxillary hypoplasia, hypertrophic gums, irregularly placed teeth, upturned nostril, low-set big ears, and retrognathi. The first patient's magnetic resonance imaging at 15 months of age indicated mild symmetric T2 prolongation involving the subthalamic nuclei. His second magnetic resonance imaging at 2 years old revealed a symmetric T2 prolongation involving the subthalamic nuclei, substantia nigra, and medulla lesions. In the second child, at the age of 2 the first magnetic resonance imaging documented heavy brainstem and subthalamic nuclei involvement. A second magnetic resonance imaging, performed when he was 3 years old, revealed diffuse involvement of the substantia nigra and hyperintense lesions of the central tegmental tract in addition to previous lesions. Facial dysmorphism and magnetic resonance imaging findings, observed in these cases, can be specific findings in Leigh syndrome patients with cytochrome C oxidase deficiency. SURF-1 gene mutations must be particularly reviewed in such patients.

  11. An optimized ERP brain-computer interface based on facial expression changes.

    PubMed

    Jin, Jing; Daly, Ian; Zhang, Yu; Wang, Xingyu; Cichocki, Andrzej

    2014-06-01

    Interferences from spatially adjacent non-target stimuli are known to evoke event-related potentials (ERPs) during non-target flashes and, therefore, lead to false positives. This phenomenon was commonly seen in visual attention-based brain-computer interfaces (BCIs) using conspicuous stimuli and is known to adversely affect the performance of BCI systems. Although users try to focus on the target stimulus, they cannot help but be affected by conspicuous changes of the stimuli (such as flashes or presenting images) which were adjacent to the target stimulus. Furthermore, subjects have reported that conspicuous stimuli made them tired and annoyed. In view of this, the aim of this study was to reduce adjacent interference, annoyance and fatigue using a new stimulus presentation pattern based upon facial expression changes. Our goal was not to design a new pattern which could evoke larger ERPs than the face pattern, but to design a new pattern which could reduce adjacent interference, annoyance and fatigue, and evoke ERPs as good as those observed during the face pattern. Positive facial expressions could be changed to negative facial expressions by minor changes to the original facial image. Although the changes are minor, the contrast is big enough to evoke strong ERPs. In this paper, a facial expression change pattern between positive and negative facial expressions was used to attempt to minimize interference effects. This was compared against two different conditions, a shuffled pattern containing the same shapes and colours as the facial expression change pattern, but without the semantic content associated with a change in expression, and a face versus no face pattern. Comparisons were made in terms of classification accuracy and information transfer rate as well as user supplied subjective measures. The results showed that interferences from adjacent stimuli, annoyance and the fatigue experienced by the subjects could be reduced significantly (p < 0.05) by

  12. An optimized ERP brain-computer interface based on facial expression changes

    NASA Astrophysics Data System (ADS)

    Jin, Jing; Daly, Ian; Zhang, Yu; Wang, Xingyu; Cichocki, Andrzej

    2014-06-01

    Objective. Interferences from spatially adjacent non-target stimuli are known to evoke event-related potentials (ERPs) during non-target flashes and, therefore, lead to false positives. This phenomenon was commonly seen in visual attention-based brain-computer interfaces (BCIs) using conspicuous stimuli and is known to adversely affect the performance of BCI systems. Although users try to focus on the target stimulus, they cannot help but be affected by conspicuous changes of the stimuli (such as flashes or presenting images) which were adjacent to the target stimulus. Furthermore, subjects have reported that conspicuous stimuli made them tired and annoyed. In view of this, the aim of this study was to reduce adjacent interference, annoyance and fatigue using a new stimulus presentation pattern based upon facial expression changes. Our goal was not to design a new pattern which could evoke larger ERPs than the face pattern, but to design a new pattern which could reduce adjacent interference, annoyance and fatigue, and evoke ERPs as good as those observed during the face pattern. Approach. Positive facial expressions could be changed to negative facial expressions by minor changes to the original facial image. Although the changes are minor, the contrast is big enough to evoke strong ERPs. In this paper, a facial expression change pattern between positive and negative facial expressions was used to attempt to minimize interference effects. This was compared against two different conditions, a shuffled pattern containing the same shapes and colours as the facial expression change pattern, but without the semantic content associated with a change in expression, and a face versus no face pattern. Comparisons were made in terms of classification accuracy and information transfer rate as well as user supplied subjective measures. Main results. The results showed that interferences from adjacent stimuli, annoyance and the fatigue experienced by the subjects could be

  13. A 3-dimensional anthropometric evaluation of facial morphology among Chinese and Greek population.

    PubMed

    Liu, Yun; Kau, Chung How; Pan, Feng; Zhou, Hong; Zhang, Qiang; Zacharopoulos, Georgios Vasileiou

    2013-07-01

    The use of 3-dimensional (3D) facial imaging has taken greater importance as orthodontists use the soft tissue paradigm in the evaluation of skeletal disproportion. Studies have shown that faces defer in populations. To date, no anthropometric evaluations have been made of Chinese and Greek faces. The aim of this study was to compare facial morphologies of Greeks and Chinese using 3D facial anthropometric landmarks. Three-dimensional facial images were acquired via a commercially available stereophotogrammetric camera capture system. The 3dMD face system captured 245 subjects from 2 population groups (Chinese [n = 72] and Greek [n = 173]), and each population was categorized into male and female groups for evaluation. All subjects in the group were between 18 and 30 years old and had no apparent facial anomalies. Twenty-five anthropometric landmarks were identified on the 3D faces of each subject. Soft tissue nasion was set as the "zeroed" reference landmark. Twenty landmark distances were constructed and evaluated within 3 dimensions of space. Six angles, 4 proportions, and 1 construct were also calculated. Student t test was used to analyze each data set obtained within each subgroup. Distinct facial differences were noted between the subgroups evaluated. When comparing differences of sexes in 2 populations (eg, male Greeks and male Chinese), significant differences were noted in more than 80% of the landmark distances calculated. One hundred percent of the angular were significant, and the Chinese were broader in width to height facial proportions. In evaluating the lips to the esthetic line, the Chinese population had more protrusive lips. There are differences in the facial morphologies of subjects obtained from a Chinese population versus that of a Greek population.

  14. Perception of Age, Attractiveness, and Tiredness After Isolated and Combined Facial Subunit Aging.

    PubMed

    Forte, Antonio Jorge; Andrew, Tom W; Colasante, Cesar; Persing, John A

    2015-12-01

    Patients often seek help to redress aging that affects various regions of the face (subunits). The purpose of this study was to determine how aging of different facial subunits impacts perception of age, attractiveness, and tiredness. Frontal and lateral view facial photographs of a middle-aged woman were modified using imaging software to independently age different facial features. Sixty-six subjects were administered with a questionnaire, and presented with a baseline unmodified picture and others containing different individual or grouped aging of facial subunits. Test subjects were asked to estimate the age of the subject in the image and quantify (0-10 scale) how "tired" and "attractive" they appeared. Facial subunits were organized following rank assignment regarding impact on perception of age, attractiveness, and tiredness. The correlation coefficient between age and attractiveness had a strong inverse relationship of approximately -0.95 in both lateral and frontal views. From most to least impact in age, the rank assignment for frontal view facial subunits was full facial aging, middle third, lower third, upper third, vertical lip rhytides, horizontal forehead rhytides, jowls, upper eyelid ptosis, loss of malar volume, lower lid fat herniation, deepening glabellar furrows, and deepening nasolabial folds. From most to least impact in age, the rank assignment for lateral view facial subunits was severe neck ptosis, jowls, moderate neck ptosis, vertical lip rhytides, crow's feet, lower lid fat herniation, loss of malar volume, and elongated earlobe. This study provides a preliminary template for further research to determine which anatomical subunit will have the most substantial effect on an aged appearance, as well as on the perception of tiredness and attractiveness. This journal requires that authors assign a level of evidence to each article. For a full description of these Evidence-Based Medicine ratings, please refer to the Table of Contents or the

  15. Sad Facial Expressions Increase Choice Blindness

    PubMed Central

    Wang, Yajie; Zhao, Song; Zhang, Zhijie; Feng, Wenfeng

    2018-01-01

    Previous studies have discovered a fascinating phenomenon known as choice blindness—individuals fail to detect mismatches between the face they choose and the face replaced by the experimenter. Although previous studies have reported a couple of factors that can modulate the magnitude of choice blindness, the potential effect of facial expression on choice blindness has not yet been explored. Using faces with sad and neutral expressions (Experiment 1) and faces with happy and neutral expressions (Experiment 2) in the classic choice blindness paradigm, the present study investigated the effects of facial expressions on choice blindness. The results showed that the detection rate was significantly lower on sad faces than neutral faces, whereas no significant difference was observed between happy faces and neutral faces. The exploratory analysis of verbal reports found that participants who reported less facial features for sad (as compared to neutral) expressions also tended to show a lower detection rate of sad (as compared to neutral) faces. These findings indicated that sad facial expressions increased choice blindness, which might have resulted from inhibition of further processing of the detailed facial features by the less attractive sad expressions (as compared to neutral expressions). PMID:29358926

  16. Sad Facial Expressions Increase Choice Blindness.

    PubMed

    Wang, Yajie; Zhao, Song; Zhang, Zhijie; Feng, Wenfeng

    2017-01-01

    Previous studies have discovered a fascinating phenomenon known as choice blindness-individuals fail to detect mismatches between the face they choose and the face replaced by the experimenter. Although previous studies have reported a couple of factors that can modulate the magnitude of choice blindness, the potential effect of facial expression on choice blindness has not yet been explored. Using faces with sad and neutral expressions (Experiment 1) and faces with happy and neutral expressions (Experiment 2) in the classic choice blindness paradigm, the present study investigated the effects of facial expressions on choice blindness. The results showed that the detection rate was significantly lower on sad faces than neutral faces, whereas no significant difference was observed between happy faces and neutral faces. The exploratory analysis of verbal reports found that participants who reported less facial features for sad (as compared to neutral) expressions also tended to show a lower detection rate of sad (as compared to neutral) faces. These findings indicated that sad facial expressions increased choice blindness, which might have resulted from inhibition of further processing of the detailed facial features by the less attractive sad expressions (as compared to neutral expressions).

  17. Analysis of the efficacy of marketing tools in facial plastic surgery.

    PubMed

    Zavod, Matthew B; Adamson, Peter A

    2008-06-01

    To compare referral sources to a facial plastic surgery practice and to develop models correlating the referral source with the decision for surgery. Retrospective descriptive study. Well-established, metropolitan, private facial plastic surgery practice with training fellowship affiliated with an academic centre. One-thousand eighty-nine new consecutive patients presenting between January 2001 and December 2005 recorded intake data including age, gender, and chief complaint. Final data input was their decision for or against surgery. Main outcome measures included differences in referral sources based on data collected and how those sources related to decision for surgery. A 50% conversion rate was found. Women and older patients were more likely to be referred from magazines, television, and newspapers and for facial rejuvenation. Men and younger patients were more likely to be referred from the website and for rhinoplasty. For facial rejuvenation, both the number of patients interested in and the probability that they agreed to the procedure increased with age. For rhinoplasty, the converse was true. The most likely patients to schedule surgery were those who were referred from other patients, friends, or family members in our practice. The data confirm that word-of-mouth referrals are the most important source for predicting which patients will elect to proceed with surgery in this established facial cosmetic surgery practice.

  18. Differences in holistic processing do not explain cultural differences in the recognition of facial expression.

    PubMed

    Yan, Xiaoqian; Young, Andrew W; Andrews, Timothy J

    2017-12-01

    The aim of this study was to investigate the causes of the own-race advantage in facial expression perception. In Experiment 1, we investigated Western Caucasian and Chinese participants' perception and categorization of facial expressions of six basic emotions that included two pairs of confusable expressions (fear and surprise; anger and disgust). People were slightly better at identifying facial expressions posed by own-race members (mainly in anger and disgust). In Experiment 2, we asked whether the own-race advantage was due to differences in the holistic processing of facial expressions. Participants viewed composite faces in which the upper part of one expression was combined with the lower part of a different expression. The upper and lower parts of the composite faces were either aligned or misaligned. Both Chinese and Caucasian participants were better at identifying the facial expressions from the misaligned images, showing interference on recognizing the parts of the expressions created by holistic perception of the aligned composite images. However, this interference from holistic processing was equivalent across expressions of own-race and other-race faces in both groups of participants. Whilst the own-race advantage in recognizing facial expressions does seem to reflect the confusability of certain emotions, it cannot be explained by differences in holistic processing.

  19. Outcome of facial physiotherapy in patients with prolonged idiopathic facial palsy.

    PubMed

    Watson, G J; Glover, S; Allen, S; Irving, R M

    2015-04-01

    This study investigated whether patients who remain symptomatic more than a year following idiopathic facial paralysis gain benefit from tailored facial physiotherapy. A two-year retrospective review was conducted of all symptomatic patients. Data collected included: age, gender, duration of symptoms, Sunnybrook facial grading system scores pre-treatment and at last visit, and duration of treatment. The study comprised 22 patients (with a mean age of 50.5 years (range, 22-75 years)) who had been symptomatic for more than a year following idiopathic facial paralysis. The mean duration of symptoms was 45 months (range, 12-240 months). The mean duration of follow up was 10.4 months (range, 2-36 months). Prior to treatment, the mean Sunnybrook facial grading system score was 59 (standard deviation = 3.5); this had increased to 83 (standard deviation = 2.7) at the last visit, with an average improvement in score of 23 (standard deviation = 2.9). This increase was significant (p < 0.001). Tailored facial therapy can improve facial grading scores in patients who remain symptomatic for prolonged periods.

  20. Clinical outcome of continuous facial nerve monitoring during primary parotidectomy.

    PubMed

    Terrell, J E; Kileny, P R; Yian, C; Esclamado, R M; Bradford, C R; Pillsbury, M S; Wolf, G T

    1997-10-01

    To assess whether continuous facial nerve monitoring during parotidectomy is associated with a lower incidence of facial nerve paresis or paralysis compared with parotidectomy without monitoring and to assess the cost of such monitoring. A retrospective analysis of outcomes for patients who underwent parotidectomy with or without continuous facial nerve monitoring. University medical center. Fifty-six patients undergoing parotidectomy in whom continuous electromyographic monitoring was used and 61 patients in whom it was not used. (1) The incidence of early and persistent facial nerve paresis or paralysis and (2) the cost associated with facial nerve monitoring. Early, unintentional facial weakness was significantly lower in the group monitored by electromyograpy (43.6%) than in the unmonitored group (62.3%) (P=.04). In the subgroup of patients without comorbid conditions or surgeries, early weakness in the monitored group (33.3%) remained statistically lower than the rate of early weakness in the unmonitored group (57.5%) (P=.03). There was no statistical difference in the final facial nerve function or incidence of permanent nerve injury between the groups or subgroups. After multivariate analysis, nonmonitored status (odds ratio [OR], 3.22), advancing age (OR, 1.47 per 10 years), and longer operative times (OR, 1.3 per hour) were the only significant independent predictive variables significantly associated with early postoperative facial weakness. The incremental cost of facial nerve monitoring was $379. The results suggest that continuous electromyographic monitoring of facial muscle during primary parotidectomy reduces the incidence of short-term postoperative facial paresis. Advantages and disadvantages of this technique need to be considered together with the additional costs in deciding whether routine use of continuous monitoring is a useful, cost-effective adjunct to parotid surgery.

  1. iFER: facial expression recognition using automatically selected geometric eye and eyebrow features

    NASA Astrophysics Data System (ADS)

    Oztel, Ismail; Yolcu, Gozde; Oz, Cemil; Kazan, Serap; Bunyak, Filiz

    2018-03-01

    Facial expressions have an important role in interpersonal communications and estimation of emotional states or intentions. Automatic recognition of facial expressions has led to many practical applications and became one of the important topics in computer vision. We present a facial expression recognition system that relies on geometry-based features extracted from eye and eyebrow regions of the face. The proposed system detects keypoints on frontal face images and forms a feature set using geometric relationships among groups of detected keypoints. Obtained feature set is refined and reduced using the sequential forward selection (SFS) algorithm and fed to a support vector machine classifier to recognize five facial expression classes. The proposed system, iFER (eye-eyebrow only facial expression recognition), is robust to lower face occlusions that may be caused by beards, mustaches, scarves, etc. and lower face motion during speech production. Preliminary experiments on benchmark datasets produced promising results outperforming previous facial expression recognition studies using partial face features, and comparable results to studies using whole face information, only slightly lower by ˜ 2.5 % compared to the best whole face facial recognition system while using only ˜ 1 / 3 of the facial region.

  2. Tuberous Sclerosis Complex in 29 Children: Clinical and Genetic Analysis and Facial Angiofibroma Responses to Topical Sirolimus.

    PubMed

    Wang, Senfen; Liu, Yuanxiang; Wei, Jinghai; Zhang, Jian; Wang, Zhaoyang; Xu, Zigang

    2017-09-01

    Tuberous sclerosis complex (TSC) is a genetic disorder and facial angiofibromas are disfiguring facial lesions. The aim of this study was to analyze the clinical and genetic features of TSC and to assess the treatment of facial angiofibromas using topical sirolimus in Chinese children. Information was collected on 29 patients with TSC. Genetic analyses were performed in 12 children and their parents. Children were treated with 0.1% sirolimus ointment for 36 weeks. Clinical efficacy and plasma sirolimus concentrations were evaluated at baseline and 12, 24, and 36 weeks. Twenty-seven (93%) of the 29 patients had hypomelanotic macules and 15 (52%) had shagreen patch; 11 of the 12 (92%) who underwent genetic analysis had gene mutations in the TSC1 or TSC2 gene. Twenty-four children completed 36 weeks of treatment with topical sirolimus; facial angiofibromas were clinically undetectable in four (17%). The mean decrease in the Facial Angiofibroma Severity Index (FASI) score at 36 weeks was 47.6 ± 30.4%. There was no significant difference in the FASI score between weeks 24 and 36 (F = 1.00, p = 0.33). There was no detectable systemic absorption of sirolimus. Hypomelanotic macules are often the first sign of TSC. Genetic testing has a high detection rate in patients with a clinical diagnosis of TSC. Topical sirolimus appears to be both effective and well-tolerated as a treatment of facial angiofibromas in children with TSC. The response typically plateaus after 12 to 24 weeks of treatment. © 2017 Wiley Periodicals, Inc.

  3. Improved facial nerve identification during parotidectomy with fluorescently labeled peptide.

    PubMed

    Hussain, Timon; Nguyen, Linda T; Whitney, Michael; Hasselmann, Jonathan; Nguyen, Quyen T

    2016-12-01

    Additional intraoperative guidance could reduce the risk of iatrogenic injury during parotid gland cancer surgery. We evaluated the intraoperative use of fluorescently labeled nerve binding peptide NP41 to aid facial nerve identification and preservation during parotidectomy in an orthotopic model of murine parotid gland cancer. We also quantified the accuracy of intraoperative nerve detection for surface and buried nerves in the head and neck with NP41 versus white light (WL) alone. Twenty-eight mice underwent parotid gland cancer surgeries with additional fluorescence (FL) guidance versus WL reflectance (WLR) alone. Eight mice were used for additional nerve-imaging experiments. Twenty-eight parotid tumor-bearing mice underwent parotidectomy. Eight mice underwent imaging of both sides of the face after skin removal. Postoperative assessment of facial nerve function measured by automated whisker tracking were compared between FL guidance (n = 13) versus WL alone (n=15). In eight mice, nerve to surrounding tissue contrast was measured under FL versus WLR for all nerve branches detectable in the field of view. Postoperative facial nerve function after parotid gland cancer surgery tended to be better with additional FL guidance. Fluorescent labeling significantly improved nerve to surrounding tissue contrast for both large and smaller buried nerve branches compared to WLR visualization and improved detection sensitivity and specificity. NP41 FL imaging significantly aids the intraoperative identification of nerve braches otherwise nearly invisible to the naked eye. Its application in a murine model of parotid gland cancer surgery tended to improve functional preservation of the facial nerve. NA Laryngoscope, 126:2711-2717, 2016. © 2016 The American Laryngological, Rhinological and Otological Society, Inc.

  4. Brain anomalies in velo-cardio-facial syndrome

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Mitnick, R.J.; Bello, J.A.; Shprintzen, R.J.

    Magnetic resonance imaging of the brain in 11 consecutively referred patients with velo-cardio-facial syndrome (VCF) showed anomalies in nine cases including small vermis, cysts adjacent to the frontal horns, and small posterior fossa. Focal signal hyperintensities in the white matter on long TR images were also noted. The nine patients showed a variety of behavioral abnormalities including mild development delay, learning disabilities, and characteristic personality traits typical of this common multiple anomaly syndrome which has been related to a microdeletion at 22q11. Analysis of the behavorial findings showed no specific pattern related to the brain anomalies, and the patients withmore » VCF who did not have detectable brain lesions also had behavioral abnormalities consistent with VCF. The significance of the lesions is not yet known, but the high prevalence of anomalies in this sample suggests that structural brain abnormalities are probably common in VCF. 25 refs.« less

  5. Assessment of the facial features and chin development of fetuses with use of serial three-dimensional sonography and the mandibular size monogram in a Chinese population.

    PubMed

    Tsai, Meng-Yin; Lan, Kuo-Chung; Ou, Chia-Yo; Chen, Jen-Huang; Chang, Shiuh-Young; Hsu, Te-Yao

    2004-02-01

    Our purpose was to evaluate whether the application of serial three-dimensional (3D) sonography and the mandibular size monogram can allow observation of dynamic changes in facial features, as well as chin development in utero. The mandibular size monogram has been established through a cross-sectional study involving 183 fetal images. The serial changes of facial features and chin development are assessed in a cohort study involving 40 patients. The monogram reveals that the Biparietal distance (BPD)/Mandibular body length (MBL) ratio is gradually decreased with the advance of gestational age. The cohort study conducted with serial 3D sonography shows the same tendency. Both the images and the results of paired-samples t test (P<.001) statistical analysis suggest that the fetuses develop wider chins and broader facial features in later weeks. The serial 3D sonography and mandibular size monogram display disproportionate growth of the fetal head and chin that leads to changes in facial features in late gestation. This fact must be considered when we evaluate fetuses at risk for development of micrognathia.

  6. Preliminary analysis of facial hair follicle distribution for forensic identification using OCT

    NASA Astrophysics Data System (ADS)

    Zhang, Ning; Wang, Chengming; Li, Zhigang; Xie, Lanchi; Guo, Jingjing; Xu, Lei; Yan, Yuwen; Li, Zhihui; Huang, Wei; Xu, Xiaojing

    2018-02-01

    In most deaths caused by explosions, the fragment of explosive remained on the victim's body can provide valuable clues to forensic investigation. However, the examination of the skin and appendages at the scene of explosion, which may reveal clue to the identity of an individual, has not been extensively studied. Compared with visual appearance of the epidermis surface that affected by various wounds, skin adnexa embedded in the dermis has a more stable morphology as an inner biometric. Hair follicles are formed when a fetus is 5 months old and distributed fairly evenly throughout the body, with the exception of hairless palms and soles. Therefore, we focus on the distribution of hair follicles in order to infer information of age, gender, and race/ethnicity. Optical coherence tomography (OCT) is a novel forensic imaging method, which achieves non-destructive, high-resolution and most importantly cross-sectional imaging. In this study, we design and develop a custom-built spectral-domain three-dimensional (3D) OCT system with a portable handheld probe to detect and reconstruct the hair follicles in the facial skin. We test our system on the forehead and preauricular skin of 2 adult volunteers and demonstrate the high quality visualization of hair follicles beneath the epidermis. The diameter, orientation, density and shape of hair follicles can be extracted from the 3D volume data. The preliminary analysis suggests that these parameters vary from different part of body and have individual difference. Eventually, we believe 3D OCT is promising tool for the examination of hair follicles for forensic purpose.

  7. Morphologic evaluation and classification of facial asymmetry using 3-dimensional computed tomography.

    PubMed

    Baek, Chaehwan; Paeng, Jun-Young; Lee, Janice S; Hong, Jongrak

    2012-05-01

    A systematic classification is needed for the diagnosis and surgical treatment of facial asymmetry. The purposes of this study were to analyze the skeletal structures of patients with facial asymmetry and to objectively classify these patients into groups according to these structural characteristics. Patients with facial asymmetry and recent computed tomographic images from 2005 through 2009 were included in this study, which was approved by the institutional review board. Linear measurements, angles, and reference planes on 3-dimensional computed tomograms were obtained, including maxillary (upper midline deviation, maxilla canting, and arch form discrepancy) and mandibular (menton deviation, gonion to midsagittal plane, ramus height, and frontal ramus inclination) measurements. All measurements were analyzed using paired t tests with Bonferroni correction followed by K-means cluster analysis using SPSS 13.0 to determine an objective classification of facial asymmetry in the enrolled patients. Kruskal-Wallis test was performed to verify differences among clustered groups. P < .05 was considered statistically significant. Forty-three patients (18 male, 25 female) were included in the study. They were classified into 4 groups based on cluster analysis. Their mean age was 24.3 ± 4.4 years. Group 1 included subjects (44% of patients) with asymmetry caused by a shift or lateralization of the mandibular body. Group 2 included subjects (39%) with a significant difference between the left and right ramus height with menton deviation to the short side. Group 3 included subjects (12%) with atypical asymmetry, including deviation of the menton to the short side, prominence of the angle/gonion on the larger side, and reverse maxillary canting. Group 4 included subjects (5%) with severe maxillary canting, ramus height differences, and menton deviation to the short side. In this study, patients with asymmetry were classified into 4 statistically distinct groups according to

  8. Use of 3-dimensional surface acquisition to study facial morphology in 5 populations.

    PubMed

    Kau, Chung How; Richmond, Stephen; Zhurov, Alexei; Ovsenik, Maja; Tawfik, Wael; Borbely, Peter; English, Jeryl D

    2010-04-01

    The aim of this study was to assess the use of 3-dimensional facial averages for determining morphologic differences from various population groups. We recruited 473 subjects from 5 populations. Three-dimensional images of the subjects were obtained in a reproducible and controlled environment with a commercially available stereo-photogrammetric camera capture system. Minolta VI-900 (Konica Minolta, Tokyo, Japan) and 3dMDface (3dMD LLC, Atlanta, Ga) systems were used. Each image was obtained as a facial mesh and orientated along a triangulated axis. All faces were overlaid, one on top of the other, and a complex mathematical algorithm was performed until average composite faces of 1 man and 1 woman were achieved for each subgroup. These average facial composites were superimposed based on a previously validated superimposition method, and the facial differences were quantified. Distinct facial differences were observed among the groups. The linear differences between surface shells ranged from 0.37 to 1.00 mm for the male groups. The linear differences ranged from 0.28 and 0.87 mm for the women. The color histograms showed that the similarities in facial shells between the subgroups by sex ranged from 26.70% to 70.39% for men and 36.09% to 79.83% for women. The average linear distance from the signed color histograms for the male subgroups ranged from -6.30 to 4.44 mm. The female subgroups ranged from -6.32 to 4.25 mm. Average faces can be efficiently and effectively created from a sample of 3-dimensional faces. Average faces can be used to compare differences in facial morphologies for various populations and sexes. Facial morphologic differences were greatest when totally different ethnic variations were compared. Facial morphologic similarities were present in comparable groups, but there were large variations in concentrated areas of the face. Copyright 2010 American Association of Orthodontists. Published by Mosby, Inc. All rights reserved.

  9. [Analysis of volcanic-ash-based insoluble ingredients of facial cleansers].

    PubMed

    Ikarashi, Yoshiaki; Uchino, Tadashi; Nishimura, Tetsuji

    2011-01-01

    The substance termed "Shirasu balloons", produced by the heat treatment of volcanic silicates, is in the form of hollow glass microspheres. Recently, this substance has gained popularity as an ingredient of facial cleansers currently available in the market, because it lends a refreshing and smooth feeling after use. However, reports of eye injury after use of a facial cleanser containing a substance made from volcanic ashes are on the rise. We presumed that the shape and size of these volcanic-ash-based ingredients would be the cause of such injuries. Therefore, in this study, we first developed a method for extracting water-insoluble ingredients such as "Shirasu balloons" from the facial cleansers, and then, we examined their shapes and sizes. The insoluble ingredients extracted from the cleansers were mainly those derived from volcanic silicates. A part of the ingredients remained in the form of glass microspheres, but for the most part, the ingredients were present in various forms, such as fragments of broken glass. Some of the fragments were larger than 75 microm in length. Foreign objects having a certain hardness, shape, and size (e.g., size greater than 75 microm) can possibly cause eye injury. We further examined insoluble ingredients of facial scrubs, such as artificial mineral complexes, mud, charcoal, and polymers, except for volcanic-silicate-based ingredients. The amounts of insoluble ingredients extracted from these scrubs were small and did not have a sharp edge. Some scrubs had ingredients with particles larger than 75 microm in size, but their specific gravities were small and their hardness values were much lower than those of glass microspheres of ingredients such as "Shirasu balloons". Because the fragments of glass microspheres can possibly cause eye injury, the facial cleansers containing large insoluble ingredients derived from volcanic ashes should be avoided to use around eyes.

  10. Nerve crush but not displacement-induced stretch of the intra-arachnoidal facial nerve promotes facial palsy after cerebellopontine angle surgery.

    PubMed

    Bendella, Habib; Brackmann, Derald E; Goldbrunner, Roland; Angelov, Doychin N

    2016-10-01

    Little is known about the reasons for occurrence of facial nerve palsy after removal of cerebellopontine angle tumors. Since the intra-arachnoidal portion of the facial nerve is considered to be so vulnerable that even the slightest tension or pinch may result in ruptured axons, we tested whether a graded stretch or controlled crush would affect the postoperative motor performance of the facial (vibrissal) muscle in rats. Thirty Wistar rats, divided into five groups (one with intact controls and four with facial nerve lesions), were used. Under inhalation anesthesia, the occipital squama was opened, the cerebellum gently retracted to the left, and the intra-arachnoidal segment of the right facial nerve exposed. A mechanical displacement of the brainstem with 1 or 3 mm toward the midline or an electromagnet-controlled crush of the facial nerve with a tweezers at a closure velocity of 50 and 100 mm/s was applied. On the next day, whisking motor performance was determined by video-based motion analysis. Even the larger (with 3 mm) mechanical displacement of the brainstem had no harmful effect: The amplitude of the vibrissal whisks was in the normal range of 50°-60°. On the other hand, even the light nerve crush (50 mm/s) injured the facial nerve and resulted in paralyzed vibrissal muscles (amplitude of 10°-15°). We conclude that, contrary to the generally acknowledged assumptions, it is the nerve crush but not the displacement-induced stretching of the intra-arachnoidal facial trunk that promotes facial palsy after cerebellopontine angle surgery in rats.

  11. Randomized clinical trial of facial acupuncture with or without body acupuncture for treatment of melasma.

    PubMed

    Rerksuppaphol, Lakkana; Charoenpong, Theekapun; Rerksuppaphol, Sanguansak

    2016-02-01

    To evaluate the efficacy of acupuncture treatments in treating facial melasma, contrasting treatments involving facial acupuncture with facial/body acupuncture. Women suffering with melasma were randomly assigned into: 1) facial acupuncture (n = 20); or 2) facial/body acupuncture (n = 21). Each group was given 2 sessions per week for 8 weeks. Melasma area and darkness of its pigmentation were assessed using digital images. 95.2% and 90% of participants in facial/body and facial acupuncture, respectively, had decreased melasma areas, with a mean reduction area being 2.6 cm(2) (95%CI 1.6-3.6 cm(2)) and 2.4 cm(2) (95%CI 1.6-3.3 cm(2)), respectively. 66.7% (facial/body acupuncture) and 80.0% (facial acupuncture) of participants had lighter melasma pigmentation compared to their baselines (p-value = 0.482). Facial acupuncture, with or without body acupuncture, was shown to be effective in decreasing the size of melasma areas. This study is registered with the Thai Clinical Trial Registry (TCTR20140903004). Copyright © 2015 Elsevier Ltd. All rights reserved.

  12. Facial recognition using simulated prosthetic pixelized vision.

    PubMed

    Thompson, Robert W; Barnett, G David; Humayun, Mark S; Dagnelie, Gislin

    2003-11-01

    To evaluate a model of simulated pixelized prosthetic vision using noncontiguous circular phosphenes, to test the effects of phosphene and grid parameters on facial recognition. A video headset was used to view a reference set of four faces, followed by a partially averted image of one of those faces viewed through a square pixelizing grid that contained 10x10 to 32x32 dots separated by gaps. The grid size, dot size, gap width, dot dropout rate, and gray-scale resolution were varied separately about a standard test condition, for a total of 16 conditions. All tests were first performed at 99% contrast and then repeated at 12.5% contrast. Discrimination speed and performance were influenced by all stimulus parameters. The subjects achieved highly significant facial recognition accuracy for all high-contrast tests except for grids with 70% random dot dropout and two gray levels. In low-contrast tests, significant facial recognition accuracy was achieved for all but the most adverse grid parameters: total grid area less than 17% of the target image, 70% dropout, four or fewer gray levels, and a gap of 40.5 arcmin. For difficult test conditions, a pronounced learning effect was noticed during high-contrast trials, and a more subtle practice effect on timing was evident during subsequent low-contrast trials. These findings suggest that reliable face recognition with crude pixelized grids can be learned and may be possible, even with a crude visual prosthesis.

  13. Initial assessment of facial nerve paralysis based on motion analysis using an optical flow method.

    PubMed

    Samsudin, Wan Syahirah W; Sundaraj, Kenneth; Ahmad, Amirozi; Salleh, Hasriah

    2016-01-01

    An initial assessment method that can classify as well as categorize the severity of paralysis into one of six levels according to the House-Brackmann (HB) system based on facial landmarks motion using an Optical Flow (OF) algorithm is proposed. The desired landmarks were obtained from the video recordings of 5 normal and 3 Bell's Palsy subjects and tracked using the Kanade-Lucas-Tomasi (KLT) method. A new scoring system based on the motion analysis using area measurement is proposed. This scoring system uses the individual scores from the facial exercises and grades the paralysis based on the HB system. The proposed method has obtained promising results and may play a pivotal role towards improved rehabilitation programs for patients.

  14. Local intensity area descriptor for facial recognition in ideal and noise conditions

    NASA Astrophysics Data System (ADS)

    Tran, Chi-Kien; Tseng, Chin-Dar; Chao, Pei-Ju; Ting, Hui-Min; Chang, Liyun; Huang, Yu-Jie; Lee, Tsair-Fwu

    2017-03-01

    We propose a local texture descriptor, local intensity area descriptor (LIAD), which is applied for human facial recognition in ideal and noisy conditions. Each facial image is divided into small regions from which LIAD histograms are extracted and concatenated into a single feature vector to represent the facial image. The recognition is performed using a nearest neighbor classifier with histogram intersection and chi-square statistics as dissimilarity measures. Experiments were conducted with LIAD using the ORL database of faces (Olivetti Research Laboratory, Cambridge), the Face94 face database, the Georgia Tech face database, and the FERET database. The results demonstrated the improvement in accuracy of our proposed descriptor compared to conventional descriptors [local binary pattern (LBP), uniform LBP, local ternary pattern, histogram of oriented gradients, and local directional pattern]. Moreover, the proposed descriptor was less sensitive to noise and had low histogram dimensionality. Thus, it is expected to be a powerful texture descriptor that can be used for various computer vision problems.

  15. How Beauty Determines Gaze! Facial Attractiveness and Gaze Duration in Images of Real World Scenes

    PubMed Central

    Mitrovic, Aleksandra; Goller, Jürgen

    2016-01-01

    We showed that the looking time spent on faces is a valid covariate of beauty by testing the relation between facial attractiveness and gaze behavior. We presented natural scenes which always pictured two people, encompassing a wide range of facial attractiveness. Employing measurements of eye movements in a free viewing paradigm, we found a linear relation between facial attractiveness and gaze behavior: The more attractive the face, the longer and the more often it was looked at. In line with evolutionary approaches, the positive relation was particularly pronounced when participants viewed other sex faces. PMID:27698984

  16. Slowing down presentation of facial movements and vocal sounds enhances facial expression recognition and induces facial-vocal imitation in children with autism.

    PubMed

    Tardif, Carole; Lainé, France; Rodriguez, Mélissa; Gepner, Bruno

    2007-09-01

    This study examined the effects of slowing down presentation of facial expressions and their corresponding vocal sounds on facial expression recognition and facial and/or vocal imitation in children with autism. Twelve autistic children and twenty-four normal control children were presented with emotional and non-emotional facial expressions on CD-Rom, under audio or silent conditions, and under dynamic visual conditions (slowly, very slowly, at normal speed) plus a static control. Overall, children with autism showed lower performance in expression recognition and more induced facial-vocal imitation than controls. In the autistic group, facial expression recognition and induced facial-vocal imitation were significantly enhanced in slow conditions. Findings may give new perspectives for understanding and intervention for verbal and emotional perceptive and communicative impairments in autistic populations.

  17. An extensive analysis of various texture feature extractors to detect Diabetes Mellitus using facial specific regions.

    PubMed

    Shu, Ting; Zhang, Bob; Yan Tang, Yuan

    2017-04-01

    Researchers have recently discovered that Diabetes Mellitus can be detected through non-invasive computerized method. However, the focus has been on facial block color features. In this paper, we extensively study the effects of texture features extracted from facial specific regions at detecting Diabetes Mellitus using eight texture extractors. The eight methods are from four texture feature families: (1) statistical texture feature family: Image Gray-scale Histogram, Gray-level Co-occurance Matrix, and Local Binary Pattern, (2) structural texture feature family: Voronoi Tessellation, (3) signal processing based texture feature family: Gaussian, Steerable, and Gabor filters, and (4) model based texture feature family: Markov Random Field. In order to determine the most appropriate extractor with optimal parameter(s), various parameter(s) of each extractor are experimented. For each extractor, the same dataset (284 Diabetes Mellitus and 231 Healthy samples), classifiers (k-Nearest Neighbors and Support Vector Machines), and validation method (10-fold cross validation) are used. According to the experiments, the first and third families achieved a better outcome at detecting Diabetes Mellitus than the other two. The best texture feature extractor for Diabetes Mellitus detection is the Image Gray-scale Histogram with bin number=256, obtaining an accuracy of 99.02%, a sensitivity of 99.64%, and a specificity of 98.26% by using SVM. Copyright © 2017 Elsevier Ltd. All rights reserved.

  18. Tissue-Engineered Autologous Grafts for Facial Bone Reconstruction

    PubMed Central

    Bhumiratana, Sarindr; Bernhard, Jonathan C.; Alfi, David M.; Yeager, Keith; Eton, Ryan E.; Bova, Jonathan; Shah, Forum; Gimble, Jeffrey M.; Lopez, Mandi J.; Eisig, Sidney B.; Vunjak-Novakovic, Gordana

    2016-01-01

    Facial deformities require precise reconstruction of the appearance and function of the original tissue. The current standard of care—the use of bone harvested from another region in the body—has major limitations, including pain and comorbidities associated with surgery. We have engineered one of the most geometrically complex facial bones by using autologous stromal/stem cells, without bone morphogenic proteins, using native bovine bone matrix and a perfusion bioreactor for the growth and transport of living grafts. The ramus-condyle unit (RCU), the most eminent load-bearing bone in the skull, was reconstructed using an image-guided personalized approach in skeletally mature Yucatan minipigs (human-scale preclinical model). We used clinically approved decellularized bovine trabecular bone as a scaffolding material, and crafted it into an anatomically correct shape using image-guided micromilling, to fit the defect. Autologous adipose-derived stromal/stem cells were seeded into the scaffold and cultured in perfusion for 3 weeks in a specialized bioreactor to form immature bone tissue. Six months after implantation, the engineered grafts maintained their anatomical structure, integrated with native tissues, and generated greater volume of new bone and greater vascular infiltration than either non-seeded anatomical scaffolds or untreated defects. This translational study demonstrates feasibility of facial bone reconstruction using autologous, anatomically shaped, living grafts formed in vitro, and presents a platform for personalized bone tissue engineering. PMID:27306665

  19. Facial fractures in children.

    PubMed

    Boyette, Jennings R

    2014-10-01

    Facial trauma in children differs from adults. The growing facial skeleton presents several challenges to the reconstructive surgeon. A thorough understanding of the patterns of facial growth and development is needed to form an individualized treatment strategy. A proper diagnosis must be made and treatment options weighed against the risk of causing further harm to facial development. This article focuses on the management of facial fractures in children. Discussed are common fracture patterns based on the development of the facial structure, initial management, diagnostic strategies, new concepts and old controversies regarding radiologic examinations, conservative versus operative intervention, risks of growth impairment, and resorbable fixation. Copyright © 2014 Elsevier Inc. All rights reserved.

  20. Slowing down Presentation of Facial Movements and Vocal Sounds Enhances Facial Expression Recognition and Induces Facial-Vocal Imitation in Children with Autism

    ERIC Educational Resources Information Center

    Tardif, Carole; Laine, France; Rodriguez, Melissa; Gepner, Bruno

    2007-01-01

    This study examined the effects of slowing down presentation of facial expressions and their corresponding vocal sounds on facial expression recognition and facial and/or vocal imitation in children with autism. Twelve autistic children and twenty-four normal control children were presented with emotional and non-emotional facial expressions on…

  1. Facial and bodily correlates of family background.

    PubMed

    Boothroyd, Lynda G; Perrett, David I

    2006-09-22

    It has been suggested that absence of the father during early childhood has long-reaching effects on reproductive strategy and development of offspring. This paper reports two studies designed to investigate the physical characteristics of daughters associated with father absence. Study 1 used a facial averaging method to produce composite images of faces of women whose parents separated during their childhood (who were 'father absent'), women whose parents remained together, but had poor quality relationships and women whose parents were together and had good quality relationships. Images were then rated by male and female judges. Father absence and poor parental relationships were associated with apparent facial masculinity and reduced attractiveness in daughters. Poor parental relationships were also associated with reduced apparent health. Study 2 compared family background with body measurements and found that father absence or a poor quality relationship between parents were associated with body masculinity (high waist-to-hip ratio) and increased weight-for-height and adiposity. These results highlight the possibility of physical masculinization being associated with purported father absence 'effects'.

  2. Assessment of Facial Golden Proportions among North Maharashtri-an Population

    PubMed Central

    Sunilkumar, L N; Jadhav, Kalpana S; Nazirkar, Girish; Singh, Shailendra; Nagmode, Pradnya S; Ali, Fareedi Mukram

    2013-01-01

    Background: Divine Proportion in Orthodontics and Prosthodontics has always been intriguing. This was applied to the North Maharashtrian population to evaluate the relationship between facial esthetics and the golden proportions. Materials & Methods: Facial proportions were assessed by examining photographs of sum total of 300 subjects of North Maharashtrian population. Young adults with a skeletal and dental Class 1 occlusion, competent lips, and balanced facial proportion were selected. Photographic prints were taken and manually parameters were plotted and analysis was done. Results: The measurements of anterior facial height showed proportionality with the total facial height. The values showed shorter lower anterior facial height and deviation of facial width parameters from the divine proportion indicating small mouth, nose, and narrow-set eyes with respect to the inter-temporal width. Conclusion: There is soft-tissue facial balance of North Maharashtrian population in comparison with the golden proportion. However, certain parameters show some deviation from the divine proportion.. How to cite this article: Sunilkumar L N, Jadhav K S, Nazirkar G, Singh S, Nagmode P S, Ali F M. Assessment of Facial Golden Proportions among North Maharashtrian Population. J Int Oral Health 2013; 5(3):48-54. PMID:24155602

  3. Alagille syndrome in a Vietnamese cohort: mutation analysis and assessment of facial features.

    PubMed

    Lin, Henry C; Le Hoang, Phuc; Hutchinson, Anne; Chao, Grace; Gerfen, Jennifer; Loomes, Kathleen M; Krantz, Ian; Kamath, Binita M; Spinner, Nancy B

    2012-05-01

    Alagille syndrome (ALGS, OMIM #118450) is an autosomal dominant disorder that affects multiple organ systems including the liver, heart, eyes, vertebrae, and face. ALGS is caused by mutations in one of two genes in the Notch Signaling Pathway, Jagged1 (JAG1) or NOTCH2. In this study, analysis of 21 Vietnamese ALGS individuals led to the identification of 19 different mutations (18 JAG1 and 1 NOTCH2), 17 of which are novel, including the third reported NOTCH2 mutation in Alagille Syndrome. The spectrum of JAG1 mutations in the Vietnamese patients is similar to that previously reported, including nine frameshift, three missense, two splice site, one nonsense, two whole gene, and one partial gene deletion. The missense mutations are all likely to be disease causing, as two are loss of cysteines (C22R and C78G) and the third creates a cryptic splice site in exon 9 (G386R). No correlation between genotype and phenotype was observed. Assessment of clinical phenotype revealed that skeletal manifestations occur with a higher frequency than in previously reported Alagille cohorts. Facial features were difficult to assess and a Vietnamese pediatric gastroenterologist was only able to identify the facial phenotype in 61% of the cohort. To assess the agreement among North American dysmorphologists at detecting the presence of ALGS facial features in the Vietnamese patients, 37 clinical dysmorphologists evaluated a photographic panel of 20 Vietnamese children with and without ALGS. The dysmorphologists were unable to identify the individuals with ALGS in the majority of cases, suggesting that evaluation of facial features should not be used in the diagnosis of ALGS in this population. This is the first report of mutations and phenotypic spectrum of ALGS in a Vietnamese population. Copyright © 2012 Wiley Periodicals, Inc.

  4. The role of the posed smile in overall facial esthetics.

    PubMed

    Havens, David C; McNamara, James A; Sigler, Lauren M; Baccetti, Tiziano

    2010-03-01

    To evaluate the role of the posed smile in overall facial esthetics, as determined by laypersons and orthodontists. Twenty orthodontists and 20 lay evaluators were asked to perform six Q-sorts on different photographs of 48 white female subjects. The six Q-sorts consisted of three different photographs for each of two time points (pre- and posttreatment), as follows: (1) smile-only, (2) face without the smile, and (3) face with the smile. The evaluators determined a split-line for attractive and unattractive images at the end of each Q-sort. The proportions of attractive patients were compared across Q-sorts using a Wilcoxon signed-rank test for paired data. The evaluators also ranked nine facial/dental characteristics at the completion of the six Q-sorts. Evaluators found the pretreatment face without the smile to be significantly more attractive than the face with the smile or the smile-only photographs. Dissimilar results were seen posttreatment; there was not a significant difference between the three posttreatment photographs. The two panels agreed on the proportion of "attractive" subjects but differed on the attractiveness level of each individual subject. The presence of a malocclusion has a negative impact on facial attractiveness. Orthodontic correction of a malocclusion affects overall facial esthetics positively. Laypeople and orthodontists agree on what is attractive. Overall facial harmony is the most important characteristic used in deciding facial attractiveness.

  5. A three-dimensional look for facial differences between males and females in a British-Caucasian sample aged 151/2 years old.

    PubMed

    Toma, A M; Zhurov, A; Playle, R; Richmond, S

    2008-08-01

    Optical surface scanning accurately records the three-dimension (3D) shape of the face non-invasively. Many software programs have been developed to process and analyze the 3D data, enabling the clinicians to create average templates for groups of subjects to provide a comparison of facial shape. Differences in facial morphology of males and females were identified using a laser scan imaging technology. This study was undertaken on 380 British-Caucasian children aged 15 and a half year old, recruited from the Avon Longitudinal Study of Parents and Children (ALSPAC). 3D facial images were obtained for each individual using two high resolution Konica/Minolta laser scanners. The scan quality was assessed and any unsuitable scans were excluded from the study. Average facial templates were created for males and females, and a registration technique was used to superimpose the facial shells of males and females so that facial differences can be quantified. Thirty unsuitable scans were excluded from the study. The final sample consisted of 350 subjects (166 females, 184 males). Females tend to have more prominent eyes and cheeks in relation to males with a maximum difference of 2.4 mm. Males tend to have more prominent noses and mouths with a maximum difference of 2.7 mm. About 31% of the facial shells match exactly (no difference), mainly in the forehead and chin regions of the face. Differences in facial morphology can be accurately quantified and visualized using 3D imaging technology. This method of facial assessment can be recommended and applied for future research studies to assess facial soft tissue changes because of growth or healthcare intervention.

  6. Why the long face? The importance of vertical image structure for biological "barcodes" underlying face recognition.

    PubMed

    Spence, Morgan L; Storrs, Katherine R; Arnold, Derek H

    2014-07-29

    Humans are experts at face recognition. The mechanisms underlying this complex capacity are not fully understood. Recently, it has been proposed that face recognition is supported by a coarse-scale analysis of visual information contained in horizontal bands of contrast distributed along the vertical image axis-a biological facial "barcode" (Dakin & Watt, 2009). A critical prediction of the facial barcode hypothesis is that the distribution of image contrast along the vertical axis will be more important for face recognition than image distributions along the horizontal axis. Using a novel paradigm involving dynamic image distortions, a series of experiments are presented examining famous face recognition impairments from selectively disrupting image distributions along the vertical or horizontal image axes. Results show that disrupting the image distribution along the vertical image axis is more disruptive for recognition than matched distortions along the horizontal axis. Consistent with the facial barcode hypothesis, these results suggest that human face recognition relies disproportionately on appropriately scaled distributions of image contrast along the vertical image axis. © 2014 ARVO.

  7. Chronic, burning facial pain following cosmetic facial surgery.

    PubMed

    Eisenberg, E; Yaari, A; Har-Shai, Y

    1996-01-01

    Chronic, burning facial pain as a result of cosmetic facial surgery has rarely been reported. During the year of 1994, two female patients presented themselves at our Pain Relief Clinic with chronic facial pain that developed following aesthetic facial surgery. One patient underwent bilateral transpalpebral surgery for removal of intraorbital fat for the correction of the exophthalmus, and the other had classical face and anterior hairline forehead lifts. Pain in both patients was similar in that it was bilateral, symmetric, burning in quality, and aggravated by external stimuli, mainly light touch. It was resistant to multiple analgesic medications, and was associated with significant depression and disability. Diagnostic local (lidocaine) and systemic (lidocaine and phentolamine) nerve blocks failed to provide relief. Psychological evaluation revealed that the two patients had clear psychosocial factors that seemed to have further compounded their pain complaints. Tricyclic antidepressants (and biofeedback training in one patient) were modestly effective and produced only partial pain relief.

  8. Facial attractiveness.

    PubMed

    Little, Anthony C

    2014-11-01

    Facial attractiveness has important social consequences. Despite a widespread belief that beauty cannot be defined, in fact, there is considerable agreement across individuals and cultures on what is found attractive. By considering that attraction and mate choice are critical components of evolutionary selection, we can better understand the importance of beauty. There are many traits that are linked to facial attractiveness in humans and each may in some way impart benefits to individuals who act on their preferences. If a trait is reliably associated with some benefit to the perceiver, then we would expect individuals in a population to find that trait attractive. Such an approach has highlighted face traits such as age, health, symmetry, and averageness, which are proposed to be associated with benefits and so associated with facial attractiveness. This view may postulate that some traits will be universally attractive; however, this does not preclude variation. Indeed, it would be surprising if there existed a template of a perfect face that was not affected by experience, environment, context, or the specific needs of an individual. Research on facial attractiveness has documented how various face traits are associated with attractiveness and various factors that impact on an individual's judgments of facial attractiveness. Overall, facial attractiveness is complex, both in the number of traits that determine attraction and in the large number of factors that can alter attraction to particular faces. A fuller understanding of facial beauty will come with an understanding of how these various factors interact with each other. WIREs Cogn Sci 2014, 5:621-634. doi: 10.1002/wcs.1316 CONFLICT OF INTEREST: The author has declared no conflicts of interest for this article. For further resources related to this article, please visit the WIREs website. © 2014 John Wiley & Sons, Ltd.

  9. Three-dimensional evaluation of the relationship between jaw divergence and facial soft tissue dimensions.

    PubMed

    Rongo, Roberto; Antoun, Joseph Saswat; Lim, Yi Xin; Dias, George; Valletta, Rosa; Farella, Mauro

    2014-09-01

    To evaluate the relationship between mandibular divergence and vertical and transverse dimensions of the face. A sample was recruited from the orthodontic clinic of the University of Otago, New Zealand. The recruited participants (N  =  60) were assigned to three different groups based on the mandibular plane angle (hyperdivergent, n  =  20; normodivergent, n  =  20; and hypodivergent, n  =  20). The sample consisted of 31 females and 29 males, with a mean age of 21.1 years (SD ± 5.0). Facial scans were recorded for each participant using a three-dimensional (3D) white-light scanner and then merged to form a single 3D image of the face. Vertical and transverse measurements of the face were assessed from the 3D facial image. The hyperdivergent sample had a significantly larger total and lower anterior facial height than the other two groups (P < .05), although no difference was found for the middle facial height (P > .05). Similarly, there were no significant differences in the transverse measurements of the three study groups (P > .05). Both gender and body mass index (BMI) had a greater influence on the transverse dimension. Hyperdivergent facial types are associated with a long face but not necessarily a narrow face. Variations in facial soft tissue vertical and transversal dimensions are more likely to be due to gender. Body mass index has a role in mandibular width (GoGo) assessment.

  10. Rigid Facial Motion Influences Featural, But Not Holistic, Face Processing

    PubMed Central

    Xiao, Naiqi; Quinn, Paul C.; Ge, Liezhong; Lee, Kang

    2012-01-01

    We report three experiments in which we investigated the effect of rigid facial motion on face processing. Specifically, we used the face composite effect to examine whether rigid facial motion influences primarily featural or holistic processing of faces. In Experiments 1, 2, and 3, participants were first familiarized with dynamic displays in which a target face turned from one side to another; then at test, participants judged whether the top half of a composite face (the top half of the target face aligned or misaligned with the bottom half of a foil face) belonged to the target face. We compared performance in the dynamic condition to various static control conditions in Experiments 1, 2, and 3, which differed from each other in terms of the display order of the multiple static images or the inter stimulus interval (ISI) between the images. We found that the size of the face composite effect in the dynamic condition was significantly smaller than that in the static conditions. In other words, the dynamic face display influenced participants to process the target faces in a part-based manner and consequently their recognition of the upper portion of the composite face at test became less interfered with by the aligned lower part of the foil face. The findings from the present experiments provide the strongest evidence to date to suggest that the rigid facial motion mainly influences facial featural, but not holistic, processing. PMID:22342561

  11. [Establishment of the database of the 3D facial models for the plastic surgery based on network].

    PubMed

    Liu, Zhe; Zhang, Hai-Lin; Zhang, Zheng-Guo; Qiao, Qun

    2008-07-01

    To collect the three-dimensional (3D) facial data of 30 facial deformity patients by the 3D scanner and establish a professional database based on Internet. It can be helpful for the clinical intervention. The primitive point data of face topography were collected by the 3D scanner. Then the 3D point cloud was edited by reverse engineering software to reconstruct the 3D model of the face. The database system was divided into three parts, including basic information, disease information and surgery information. The programming language of the web system is Java. The linkages between every table of the database are credibility. The query operation and the data mining are convenient. The users can visit the database via the Internet and use the image analysis system to observe the 3D facial models interactively. In this paper we presented a database and a web system adapt to the plastic surgery of human face. It can be used both in clinic and in basic research.

  12. Facial anatomy.

    PubMed

    Marur, Tania; Tuna, Yakup; Demirci, Selman

    2014-01-01

    Dermatologic problems of the face affect both function and aesthetics, which are based on complex anatomical features. Treating dermatologic problems while preserving the aesthetics and functions of the face requires knowledge of normal anatomy. When performing successfully invasive procedures of the face, it is essential to understand its underlying topographic anatomy. This chapter presents the anatomy of the facial musculature and neurovascular structures in a systematic way with some clinically important aspects. We describe the attachments of the mimetic and masticatory muscles and emphasize their functions and nerve supply. We highlight clinically relevant facial topographic anatomy by explaining the course and location of the sensory and motor nerves of the face and facial vasculature with their relations. Additionally, this chapter reviews the recent nomenclature of the branching pattern of the facial artery. © 2013 Elsevier Inc. All rights reserved.

  13. Children's Facial Trustworthiness Judgments: Agreement and Relationship with Facial Attractiveness.

    PubMed

    Ma, Fengling; Xu, Fen; Luo, Xianming

    2016-01-01

    This study examined developmental changes in children's abilities to make trustworthiness judgments based on faces and the relationship between a child's perception of trustworthiness and facial attractiveness. One hundred and one 8-, 10-, and 12-year-olds, along with 37 undergraduates, were asked to judge the trustworthiness of 200 faces. Next, they issued facial attractiveness judgments. The results indicated that children made consistent trustworthiness and attractiveness judgments based on facial appearance, but with-adult and within-age agreement levels of facial judgments increased with age. Additionally, the agreement levels of judgments made by girls were higher than those by boys. Furthermore, the relationship between trustworthiness and attractiveness judgments increased with age, and the relationship between two judgments made by girls was closer than those by boys. These findings suggest that face-based trait judgment ability develops throughout childhood and that, like adults, children may use facial attractiveness as a heuristic cue that signals a stranger's trustworthiness.

  14. Children's Facial Trustworthiness Judgments: Agreement and Relationship with Facial Attractiveness

    PubMed Central

    Ma, Fengling; Xu, Fen; Luo, Xianming

    2016-01-01

    This study examined developmental changes in children's abilities to make trustworthiness judgments based on faces and the relationship between a child's perception of trustworthiness and facial attractiveness. One hundred and one 8-, 10-, and 12-year-olds, along with 37 undergraduates, were asked to judge the trustworthiness of 200 faces. Next, they issued facial attractiveness judgments. The results indicated that children made consistent trustworthiness and attractiveness judgments based on facial appearance, but with-adult and within-age agreement levels of facial judgments increased with age. Additionally, the agreement levels of judgments made by girls were higher than those by boys. Furthermore, the relationship between trustworthiness and attractiveness judgments increased with age, and the relationship between two judgments made by girls was closer than those by boys. These findings suggest that face-based trait judgment ability develops throughout childhood and that, like adults, children may use facial attractiveness as a heuristic cue that signals a stranger's trustworthiness. PMID:27148111

  15. Objective assessment of the contribution of dental esthetics and facial attractiveness in men via eye tracking.

    PubMed

    Baker, Robin S; Fields, Henry W; Beck, F Michael; Firestone, Allen R; Rosenstiel, Stephen F

    2018-04-01

    Recently, greater emphasis has been placed on smile esthetics in dentistry. Eye tracking has been used to objectively evaluate attention to the dentition (mouth) in female models with different levels of dental esthetics quantified by the aesthetic component of the Index of Orthodontic Treatment Need (IOTN). This has not been accomplished in men. Our objective was to determine the visual attention to the mouth in men with different levels of dental esthetics (IOTN levels) and background facial attractiveness, for both male and female raters, using eye tracking. Facial images of men rated as unattractive, average, and attractive were digitally manipulated and paired with validated oral images, IOTN levels 1 (no treatment need), 7 (borderline treatment need), and 10 (definite treatment need). Sixty-four raters meeting the inclusion criteria were included in the data analysis. Each rater was calibrated in the eye tracker and randomly viewed the composite images for 3 seconds, twice for reliability. Reliability was good or excellent (intraclass correlation coefficients, 0.6-0.9). Significant interactions were observed with factorial repeated-measures analysis of variance and the Tukey-Kramer method for density and duration of fixations in the interactions of model facial attractiveness by area of the face (P <0.0001, P <0.0001, respectively), dental esthetics (IOTN) by area of the face (P <0.0001, P <0.0001, respectively), and rater sex by area of the face (P = 0.0166, P = 0.0290, respectively). For area by facial attractiveness, the hierarchy of visual attention in unattractive and attractive models was eye, mouth, and nose, but for men of average attractiveness, it was mouth, eye, and nose. For dental esthetics by area, at IOTN 7, the mouth had significantly more visual attention than it did at IOTN 1 and significantly more than the nose. At IOTN 10, the mouth received significantly more attention than at IOTN 7 and surpassed the nose and eye. These

  16. Morphological Integration of Soft-Tissue Facial Morphology in Down Syndrome and Siblings

    PubMed Central

    Starbuck, John; Reeves, Roger H.; Richtsmeier, Joan

    2011-01-01

    Down syndrome (DS), resulting from trisomy of chromosome 21, is the most common live-born human aneuploidy. The phenotypic expression of trisomy 21 produces variable, though characteristic, facial morphology. Although certain facial features have been documented quantitatively and qualitatively as characteristic of DS (e.g., epicanthic folds, macroglossia, and hypertelorism), all of these traits occur in other craniofacial conditions with an underlying genetic cause. We hypothesize that the typical DS face is integrated differently than the face of non-DS siblings, and that the pattern of morphological integration unique to individuals with DS will yield information about underlying developmental associations between facial regions. We statistically compared morphological integration patterns of immature DS faces (N = 53) with those of non-DS siblings (N = 54), aged 6–12 years using 31 distances estimated from 3D coordinate data representing 17 anthropometric landmarks recorded on 3D digital photographic images. Facial features are affected differentially in DS, as evidenced by statistically significant differences in integration both within and between facial regions. Our results suggest a differential affect of trisomy on facial prominences during craniofacial development. PMID:21996933

  17. Morphological integration of soft-tissue facial morphology in Down Syndrome and siblings.

    PubMed

    Starbuck, John; Reeves, Roger H; Richtsmeier, Joan

    2011-12-01

    Down syndrome (DS), resulting from trisomy of chromosome 21, is the most common live-born human aneuploidy. The phenotypic expression of trisomy 21 produces variable, though characteristic, facial morphology. Although certain facial features have been documented quantitatively and qualitatively as characteristic of DS (e.g., epicanthic folds, macroglossia, and hypertelorism), all of these traits occur in other craniofacial conditions with an underlying genetic cause. We hypothesize that the typical DS face is integrated differently than the face of non-DS siblings, and that the pattern of morphological integration unique to individuals with DS will yield information about underlying developmental associations between facial regions. We statistically compared morphological integration patterns of immature DS faces (N = 53) with those of non-DS siblings (N = 54), aged 6-12 years using 31 distances estimated from 3D coordinate data representing 17 anthropometric landmarks recorded on 3D digital photographic images. Facial features are affected differentially in DS, as evidenced by statistically significant differences in integration both within and between facial regions. Our results suggest a differential affect of trisomy on facial prominences during craniofacial development. 2011 Wiley Periodicals, Inc.

  18. Facial Structure Predicts Sexual Orientation in Both Men and Women.

    PubMed

    Skorska, Malvina N; Geniole, Shawn N; Vrysen, Brandon M; McCormick, Cheryl M; Bogaert, Anthony F

    2015-07-01

    Biological models have typically framed sexual orientation in terms of effects of variation in fetal androgen signaling on sexual differentiation, although other biological models exist. Despite marked sex differences in facial structure, the relationship between sexual orientation and facial structure is understudied. A total of 52 lesbian women, 134 heterosexual women, 77 gay men, and 127 heterosexual men were recruited at a Canadian campus and various Canadian Pride and sexuality events. We found that facial structure differed depending on sexual orientation; substantial variation in sexual orientation was predicted using facial metrics computed by a facial modelling program from photographs of White faces. At the univariate level, lesbian and heterosexual women differed in 17 facial features (out of 63) and four were unique multivariate predictors in logistic regression. Gay and heterosexual men differed in 11 facial features at the univariate level, of which three were unique multivariate predictors. Some, but not all, of the facial metrics differed between the sexes. Lesbian women had noses that were more turned up (also more turned up in heterosexual men), mouths that were more puckered, smaller foreheads, and marginally more masculine face shapes (also in heterosexual men) than heterosexual women. Gay men had more convex cheeks, shorter noses (also in heterosexual women), and foreheads that were more tilted back relative to heterosexual men. Principal components analysis and discriminant functions analysis generally corroborated these results. The mechanisms underlying variation in craniofacial structure--both related and unrelated to sexual differentiation--may thus be important in understanding the development of sexual orientation.

  19. A comparison of facial expression properties in five hylobatid species.

    PubMed

    Scheider, Linda; Liebal, Katja; Oña, Leonardo; Burrows, Anne; Waller, Bridget

    2014-07-01

    Little is known about facial communication of lesser apes (family Hylobatidae) and how their facial expressions (and use of) relate to social organization. We investigated facial expressions (defined as combinations of facial movements) in social interactions of mated pairs in five different hylobatid species belonging to three different genera using a recently developed objective coding system, the Facial Action Coding System for hylobatid species (GibbonFACS). We described three important properties of their facial expressions and compared them between genera. First, we compared the rate of facial expressions, which was defined as the number of facial expressions per units of time. Second, we compared their repertoire size, defined as the number of different types of facial expressions used, independent of their frequency. Third, we compared the diversity of expression, defined as the repertoire weighted by the rate of use for each type of facial expression. We observed a higher rate and diversity of facial expression, but no larger repertoire, in Symphalangus (siamangs) compared to Hylobates and Nomascus species. In line with previous research, these results suggest siamangs differ from other hylobatids in certain aspects of their social behavior. To investigate whether differences in facial expressions are linked to hylobatid socio-ecology, we used a Phylogenetic General Least Square (PGLS) regression analysis to correlate those properties with two social factors: group-size and level of monogamy. No relationship between the properties of facial expressions and these socio-ecological factors was found. One explanation could be that facial expressions in hylobatid species are subject to phylogenetic inertia and do not differ sufficiently between species to reveal correlations with factors such as group size and monogamy level. © 2014 Wiley Periodicals, Inc.

  20. Visual adaptation provides objective electrophysiological evidence of facial identity discrimination.

    PubMed

    Retter, Talia L; Rossion, Bruno

    2016-07-01

    Discrimination of facial identities is a fundamental function of the human brain that is challenging to examine with macroscopic measurements of neural activity, such as those obtained with functional magnetic resonance imaging (fMRI) and electroencephalography (EEG). Although visual adaptation or repetition suppression (RS) stimulation paradigms have been successfully implemented to this end with such recording techniques, objective evidence of an identity-specific discrimination response due to adaptation at the level of the visual representation is lacking. Here, we addressed this issue with fast periodic visual stimulation (FPVS) and EEG recording combined with a symmetry/asymmetry adaptation paradigm. Adaptation to one facial identity is induced through repeated presentation of that identity at a rate of 6 images per second (6 Hz) over 10 sec. Subsequently, this identity is presented in alternation with another facial identity (i.e., its anti-face, both faces being equidistant from an average face), producing an identity repetition rate of 3 Hz over a 20 sec testing sequence. A clear EEG response at 3 Hz is observed over the right occipito-temporal (ROT) cortex, indexing discrimination between the two facial identities in the absence of an explicit behavioral discrimination measure. This face identity discrimination occurs immediately after adaptation and disappears rapidly within 20 sec. Importantly, this 3 Hz response is not observed in a control condition without the single-identity 10 sec adaptation period. These results indicate that visual adaptation to a given facial identity produces an objective (i.e., at a pre-defined stimulation frequency) electrophysiological index of visual discrimination between that identity and another, and provides a unique behavior-free quantification of the effect of visual adaptation. Copyright © 2016 Elsevier Ltd. All rights reserved.