Directly Reconstructing Principal Components of Heterogeneous Particles from Cryo-EM Images
Tagare, Hemant D.; Kucukelbir, Alp; Sigworth, Fred J.; Wang, Hongwei; Rao, Murali
2015-01-01
Structural heterogeneity of particles can be investigated by their three-dimensional principal components. This paper addresses the question of whether, and with what algorithm, the three-dimensional principal components can be directly recovered from cryo-EM images. The first part of the paper extends the Fourier slice theorem to covariance functions showing that the three-dimensional covariance, and hence the principal components, of a heterogeneous particle can indeed be recovered from two-dimensional cryo-EM images. The second part of the paper proposes a practical algorithm for reconstructing the principal components directly from cryo-EM images without the intermediate step of calculating covariances. This algorithm is based on maximizing the (posterior) likelihood using the Expectation-Maximization algorithm. The last part of the paper applies this algorithm to simulated data and to two real cryo-EM data sets: a data set of the 70S ribosome with and without Elongation Factor-G (EF-G), and a data set of the inluenza virus RNA dependent RNA Polymerase (RdRP). The first principal component of the 70S ribosome data set reveals the expected conformational changes of the ribosome as the EF-G binds and unbinds. The first principal component of the RdRP data set reveals a conformational change in the two dimers of the RdRP. PMID:26049077
NASA Astrophysics Data System (ADS)
Ji, Yi; Sun, Shanlin; Xie, Hong-Bo
2017-06-01
Discrete wavelet transform (WT) followed by principal component analysis (PCA) has been a powerful approach for the analysis of biomedical signals. Wavelet coefficients at various scales and channels were usually transformed into a one-dimensional array, causing issues such as the curse of dimensionality dilemma and small sample size problem. In addition, lack of time-shift invariance of WT coefficients can be modeled as noise and degrades the classifier performance. In this study, we present a stationary wavelet-based two-directional two-dimensional principal component analysis (SW2D2PCA) method for the efficient and effective extraction of essential feature information from signals. Time-invariant multi-scale matrices are constructed in the first step. The two-directional two-dimensional principal component analysis then operates on the multi-scale matrices to reduce the dimension, rather than vectors in conventional PCA. Results are presented from an experiment to classify eight hand motions using 4-channel electromyographic (EMG) signals recorded in healthy subjects and amputees, which illustrates the efficiency and effectiveness of the proposed method for biomedical signal analysis.
Directly reconstructing principal components of heterogeneous particles from cryo-EM images.
Tagare, Hemant D; Kucukelbir, Alp; Sigworth, Fred J; Wang, Hongwei; Rao, Murali
2015-08-01
Structural heterogeneity of particles can be investigated by their three-dimensional principal components. This paper addresses the question of whether, and with what algorithm, the three-dimensional principal components can be directly recovered from cryo-EM images. The first part of the paper extends the Fourier slice theorem to covariance functions showing that the three-dimensional covariance, and hence the principal components, of a heterogeneous particle can indeed be recovered from two-dimensional cryo-EM images. The second part of the paper proposes a practical algorithm for reconstructing the principal components directly from cryo-EM images without the intermediate step of calculating covariances. This algorithm is based on maximizing the posterior likelihood using the Expectation-Maximization algorithm. The last part of the paper applies this algorithm to simulated data and to two real cryo-EM data sets: a data set of the 70S ribosome with and without Elongation Factor-G (EF-G), and a data set of the influenza virus RNA dependent RNA Polymerase (RdRP). The first principal component of the 70S ribosome data set reveals the expected conformational changes of the ribosome as the EF-G binds and unbinds. The first principal component of the RdRP data set reveals a conformational change in the two dimers of the RdRP. Copyright © 2015 Elsevier Inc. All rights reserved.
The Complexity of Human Walking: A Knee Osteoarthritis Study
Kotti, Margarita; Duffell, Lynsey D.; Faisal, Aldo A.; McGregor, Alison H.
2014-01-01
This study proposes a framework for deconstructing complex walking patterns to create a simple principal component space before checking whether the projection to this space is suitable for identifying changes from the normality. We focus on knee osteoarthritis, the most common knee joint disease and the second leading cause of disability. Knee osteoarthritis affects over 250 million people worldwide. The motivation for projecting the highly dimensional movements to a lower dimensional and simpler space is our belief that motor behaviour can be understood by identifying a simplicity via projection to a low principal component space, which may reflect upon the underlying mechanism. To study this, we recruited 180 subjects, 47 of which reported that they had knee osteoarthritis. They were asked to walk several times along a walkway equipped with two force plates that capture their ground reaction forces along 3 axes, namely vertical, anterior-posterior, and medio-lateral, at 1000 Hz. Data when the subject does not clearly strike the force plate were excluded, leaving 1–3 gait cycles per subject. To examine the complexity of human walking, we applied dimensionality reduction via Probabilistic Principal Component Analysis. The first principal component explains 34% of the variance in the data, whereas over 80% of the variance is explained by 8 principal components or more. This proves the complexity of the underlying structure of the ground reaction forces. To examine if our musculoskeletal system generates movements that are distinguishable between normal and pathological subjects in a low dimensional principal component space, we applied a Bayes classifier. For the tested cross-validated, subject-independent experimental protocol, the classification accuracy equals 82.62%. Also, a novel complexity measure is proposed, which can be used as an objective index to facilitate clinical decision making. This measure proves that knee osteoarthritis subjects exhibit more variability in the two-dimensional principal component space. PMID:25232949
Complexity of free energy landscapes of peptides revealed by nonlinear principal component analysis.
Nguyen, Phuong H
2006-12-01
Employing the recently developed hierarchical nonlinear principal component analysis (NLPCA) method of Saegusa et al. (Neurocomputing 2004;61:57-70 and IEICE Trans Inf Syst 2005;E88-D:2242-2248), the complexities of the free energy landscapes of several peptides, including triglycine, hexaalanine, and the C-terminal beta-hairpin of protein G, were studied. First, the performance of this NLPCA method was compared with the standard linear principal component analysis (PCA). In particular, we compared two methods according to (1) the ability of the dimensionality reduction and (2) the efficient representation of peptide conformations in low-dimensional spaces spanned by the first few principal components. The study revealed that NLPCA reduces the dimensionality of the considered systems much better, than did PCA. For example, in order to get the similar error, which is due to representation of the original data of beta-hairpin in low dimensional space, one needs 4 and 21 principal components of NLPCA and PCA, respectively. Second, by representing the free energy landscapes of the considered systems as a function of the first two principal components obtained from PCA, we obtained the relatively well-structured free energy landscapes. In contrast, the free energy landscapes of NLPCA are much more complicated, exhibiting many states which are hidden in the PCA maps, especially in the unfolded regions. Furthermore, the study also showed that many states in the PCA maps are mixed up by several peptide conformations, while those of the NLPCA maps are more pure. This finding suggests that the NLPCA should be used to capture the essential features of the systems. (c) 2006 Wiley-Liss, Inc.
An efficient classification method based on principal component and sparse representation.
Zhai, Lin; Fu, Shujun; Zhang, Caiming; Liu, Yunxian; Wang, Lu; Liu, Guohua; Yang, Mingqiang
2016-01-01
As an important application in optical imaging, palmprint recognition is interfered by many unfavorable factors. An effective fusion of blockwise bi-directional two-dimensional principal component analysis and grouping sparse classification is presented. The dimension reduction and normalizing are implemented by the blockwise bi-directional two-dimensional principal component analysis for palmprint images to extract feature matrixes, which are assembled into an overcomplete dictionary in sparse classification. A subspace orthogonal matching pursuit algorithm is designed to solve the grouping sparse representation. Finally, the classification result is gained by comparing the residual between testing and reconstructed images. Experiments are carried out on a palmprint database, and the results show that this method has better robustness against position and illumination changes of palmprint images, and can get higher rate of palmprint recognition.
Dynamic of consumer groups and response of commodity markets by principal component analysis
NASA Astrophysics Data System (ADS)
Nobi, Ashadun; Alam, Shafiqul; Lee, Jae Woo
2017-09-01
This study investigates financial states and group dynamics by applying principal component analysis to the cross-correlation coefficients of the daily returns of commodity futures. The eigenvalues of the cross-correlation matrix in the 6-month timeframe displays similar values during 2010-2011, but decline following 2012. A sharp drop in eigenvalue implies the significant change of the market state. Three commodity sectors, energy, metals and agriculture, are projected into two dimensional spaces consisting of two principal components (PC). We observe that they form three distinct clusters in relation to various sectors. However, commodities with distinct features have intermingled with one another and scattered during severe crises, such as the European sovereign debt crises. We observe the notable change of the position of two dimensional spaces of groups during financial crises. By considering the first principal component (PC1) within the 6-month moving timeframe, we observe that commodities of the same group change states in a similar pattern, and the change of states of one group can be used as a warning for other group.
Information extraction from multivariate images
NASA Technical Reports Server (NTRS)
Park, S. K.; Kegley, K. A.; Schiess, J. R.
1986-01-01
An overview of several multivariate image processing techniques is presented, with emphasis on techniques based upon the principal component transformation (PCT). Multiimages in various formats have a multivariate pixel value, associated with each pixel location, which has been scaled and quantized into a gray level vector, and the bivariate of the extent to which two images are correlated. The PCT of a multiimage decorrelates the multiimage to reduce its dimensionality and reveal its intercomponent dependencies if some off-diagonal elements are not small, and for the purposes of display the principal component images must be postprocessed into multiimage format. The principal component analysis of a multiimage is a statistical analysis based upon the PCT whose primary application is to determine the intrinsic component dimensionality of the multiimage. Computational considerations are also discussed.
Guo, Zhiqiang; Wang, Huaiqing; Yang, Jie; Miller, David J
2015-01-01
In this paper, we propose and implement a hybrid model combining two-directional two-dimensional principal component analysis ((2D)2PCA) and a Radial Basis Function Neural Network (RBFNN) to forecast stock market behavior. First, 36 stock market technical variables are selected as the input features, and a sliding window is used to obtain the input data of the model. Next, (2D)2PCA is utilized to reduce the dimension of the data and extract its intrinsic features. Finally, an RBFNN accepts the data processed by (2D)2PCA to forecast the next day's stock price or movement. The proposed model is used on the Shanghai stock market index, and the experiments show that the model achieves a good level of fitness. The proposed model is then compared with one that uses the traditional dimension reduction method principal component analysis (PCA) and independent component analysis (ICA). The empirical results show that the proposed model outperforms the PCA-based model, as well as alternative models based on ICA and on the multilayer perceptron.
Guo, Zhiqiang; Wang, Huaiqing; Yang, Jie; Miller, David J.
2015-01-01
In this paper, we propose and implement a hybrid model combining two-directional two-dimensional principal component analysis ((2D)2PCA) and a Radial Basis Function Neural Network (RBFNN) to forecast stock market behavior. First, 36 stock market technical variables are selected as the input features, and a sliding window is used to obtain the input data of the model. Next, (2D)2PCA is utilized to reduce the dimension of the data and extract its intrinsic features. Finally, an RBFNN accepts the data processed by (2D)2PCA to forecast the next day's stock price or movement. The proposed model is used on the Shanghai stock market index, and the experiments show that the model achieves a good level of fitness. The proposed model is then compared with one that uses the traditional dimension reduction method principal component analysis (PCA) and independent component analysis (ICA). The empirical results show that the proposed model outperforms the PCA-based model, as well as alternative models based on ICA and on the multilayer perceptron. PMID:25849483
Principal component analysis and the locus of the Fréchet mean in the space of phylogenetic trees.
Nye, Tom M W; Tang, Xiaoxian; Weyenberg, Grady; Yoshida, Ruriko
2017-12-01
Evolutionary relationships are represented by phylogenetic trees, and a phylogenetic analysis of gene sequences typically produces a collection of these trees, one for each gene in the analysis. Analysis of samples of trees is difficult due to the multi-dimensionality of the space of possible trees. In Euclidean spaces, principal component analysis is a popular method of reducing high-dimensional data to a low-dimensional representation that preserves much of the sample's structure. However, the space of all phylogenetic trees on a fixed set of species does not form a Euclidean vector space, and methods adapted to tree space are needed. Previous work introduced the notion of a principal geodesic in this space, analogous to the first principal component. Here we propose a geometric object for tree space similar to the [Formula: see text]th principal component in Euclidean space: the locus of the weighted Fréchet mean of [Formula: see text] vertex trees when the weights vary over the [Formula: see text]-simplex. We establish some basic properties of these objects, in particular showing that they have dimension [Formula: see text], and propose algorithms for projection onto these surfaces and for finding the principal locus associated with a sample of trees. Simulation studies demonstrate that these algorithms perform well, and analyses of two datasets, containing Apicomplexa and African coelacanth genomes respectively, reveal important structure from the second principal components.
Principal component analysis on a torus: Theory and application to protein dynamics.
Sittel, Florian; Filk, Thomas; Stock, Gerhard
2017-12-28
A dimensionality reduction method for high-dimensional circular data is developed, which is based on a principal component analysis (PCA) of data points on a torus. Adopting a geometrical view of PCA, various distance measures on a torus are introduced and the associated problem of projecting data onto the principal subspaces is discussed. The main idea is that the (periodicity-induced) projection error can be minimized by transforming the data such that the maximal gap of the sampling is shifted to the periodic boundary. In a second step, the covariance matrix and its eigendecomposition can be computed in a standard manner. Adopting molecular dynamics simulations of two well-established biomolecular systems (Aib 9 and villin headpiece), the potential of the method to analyze the dynamics of backbone dihedral angles is demonstrated. The new approach allows for a robust and well-defined construction of metastable states and provides low-dimensional reaction coordinates that accurately describe the free energy landscape. Moreover, it offers a direct interpretation of covariances and principal components in terms of the angular variables. Apart from its application to PCA, the method of maximal gap shifting is general and can be applied to any other dimensionality reduction method for circular data.
Principal component analysis on a torus: Theory and application to protein dynamics
NASA Astrophysics Data System (ADS)
Sittel, Florian; Filk, Thomas; Stock, Gerhard
2017-12-01
A dimensionality reduction method for high-dimensional circular data is developed, which is based on a principal component analysis (PCA) of data points on a torus. Adopting a geometrical view of PCA, various distance measures on a torus are introduced and the associated problem of projecting data onto the principal subspaces is discussed. The main idea is that the (periodicity-induced) projection error can be minimized by transforming the data such that the maximal gap of the sampling is shifted to the periodic boundary. In a second step, the covariance matrix and its eigendecomposition can be computed in a standard manner. Adopting molecular dynamics simulations of two well-established biomolecular systems (Aib9 and villin headpiece), the potential of the method to analyze the dynamics of backbone dihedral angles is demonstrated. The new approach allows for a robust and well-defined construction of metastable states and provides low-dimensional reaction coordinates that accurately describe the free energy landscape. Moreover, it offers a direct interpretation of covariances and principal components in terms of the angular variables. Apart from its application to PCA, the method of maximal gap shifting is general and can be applied to any other dimensionality reduction method for circular data.
Fast, Exact Bootstrap Principal Component Analysis for p > 1 million
Fisher, Aaron; Caffo, Brian; Schwartz, Brian; Zipunnikov, Vadim
2015-01-01
Many have suggested a bootstrap procedure for estimating the sampling variability of principal component analysis (PCA) results. However, when the number of measurements per subject (p) is much larger than the number of subjects (n), calculating and storing the leading principal components from each bootstrap sample can be computationally infeasible. To address this, we outline methods for fast, exact calculation of bootstrap principal components, eigenvalues, and scores. Our methods leverage the fact that all bootstrap samples occupy the same n-dimensional subspace as the original sample. As a result, all bootstrap principal components are limited to the same n-dimensional subspace and can be efficiently represented by their low dimensional coordinates in that subspace. Several uncertainty metrics can be computed solely based on the bootstrap distribution of these low dimensional coordinates, without calculating or storing the p-dimensional bootstrap components. Fast bootstrap PCA is applied to a dataset of sleep electroencephalogram recordings (p = 900, n = 392), and to a dataset of brain magnetic resonance images (MRIs) (p ≈ 3 million, n = 352). For the MRI dataset, our method allows for standard errors for the first 3 principal components based on 1000 bootstrap samples to be calculated on a standard laptop in 47 minutes, as opposed to approximately 4 days with standard methods. PMID:27616801
How Adequate are One- and Two-Dimensional Free Energy Landscapes for Protein Folding Dynamics?
NASA Astrophysics Data System (ADS)
Maisuradze, Gia G.; Liwo, Adam; Scheraga, Harold A.
2009-06-01
The molecular dynamics trajectories of protein folding or unfolding, generated with the coarse-grained united-residue force field for the B domain of staphylococcal protein A, were analyzed by principal component analysis (PCA). The folding or unfolding process was examined by using free-energy landscapes (FELs) in PC space. By introducing a novel multidimensional FEL, it was shown that the low-dimensional FELs are not always sufficient for the description of folding or unfolding processes. Similarities between the topographies of FELs along low- and high-indexed principal components were observed.
Wei, Fang; Hu, Na; Lv, Xin; Dong, Xu-Yan; Chen, Hong
2015-07-24
In this investigation, off-line comprehensive two-dimensional liquid chromatography-atmospheric pressure chemical ionization mass spectrometry using a single column has been applied for the identification and quantification of triacylglycerols in edible oils. A novel mixed-mode phenyl-hexyl chromatographic column was employed in this off-line two-dimensional separation system. The phenyl-hexyl column combined the features of traditional C18 and silver-ion columns, which could provide hydrophobic interactions with triacylglycerols under acetonitrile conditions and can offer π-π interactions with triacylglycerols under methanol conditions. When compared with traditional off-line comprehensive two-dimensional liquid chromatography employing two different chromatographic columns (C18 and silver-ion column) and using elution solvents comprised of two phases (reversed-phase/normal-phase) for triacylglycerols separation, the novel off-line comprehensive two-dimensional liquid chromatography using a single column can be achieved by simply altering the mobile phase between acetonitrile and methanol, which exhibited a much higher selectivity for the separation of triacylglycerols with great efficiency and rapid speed. In addition, an approach based on the use of response factor with atmospheric pressure chemical ionization mass spectrometry has been developed for triacylglycerols quantification. Due to the differences between saturated and unsaturated acyl chains, the use of response factors significantly improves the quantitation of triacylglycerols. This two-dimensional liquid chromatography-mass spectrometry system was successfully applied for the profiling of triacylglycerols in soybean oils, peanut oils and lord oils. A total of 68 triacylglycerols including 40 triacylglycerols in soybean oils, 50 triacylglycerols in peanut oils and 44 triacylglycerols in lord oils have been identified and quantified. The liquid chromatography-mass spectrometry data were analyzed using principal component analysis. The results of the principal component analysis enabled a clear identification of different plant oils. By using this two-dimensional liquid chromatography-mass spectrometry system coupled with principal component analysis, adulterated soybean oils with 5% added lord oil and peanut oils with 5% added soybean oil can be clearly identified. Copyright © 2015 Elsevier B.V. All rights reserved.
ERIC Educational Resources Information Center
Chou, Yeh-Tai; Wang, Wen-Chung
2010-01-01
Dimensionality is an important assumption in item response theory (IRT). Principal component analysis on standardized residuals has been used to check dimensionality, especially under the family of Rasch models. It has been suggested that an eigenvalue greater than 1.5 for the first eigenvalue signifies a violation of unidimensionality when there…
NASA Astrophysics Data System (ADS)
Hristian, L.; Ostafe, M. M.; Manea, L. R.; Apostol, L. L.
2017-06-01
The work pursued the distribution of combed wool fabrics destined to manufacturing of external articles of clothing in terms of the values of durability and physiological comfort indices, using the mathematical model of Principal Component Analysis (PCA). Principal Components Analysis (PCA) applied in this study is a descriptive method of the multivariate analysis/multi-dimensional data, and aims to reduce, under control, the number of variables (columns) of the matrix data as much as possible to two or three. Therefore, based on the information about each group/assortment of fabrics, it is desired that, instead of nine inter-correlated variables, to have only two or three new variables called components. The PCA target is to extract the smallest number of components which recover the most of the total information contained in the initial data.
Morin, R.H.
1997-01-01
Returns from drilling in unconsolidated cobble and sand aquifers commonly do not identify lithologic changes that may be meaningful for Hydrogeologic investigations. Vertical resolution of saturated, Quaternary, coarse braided-slream deposits is significantly improved by interpreting natural gamma (G), epithermal neutron (N), and electromagnetically induced resistivity (IR) logs obtained from wells at the Capital Station site in Boise, Idaho. Interpretation of these geophysical logs is simplified because these sediments are derived largely from high-gamma-producing source rocks (granitics of the Boise River drainage), contain few clays, and have undergone little diagenesis. Analysis of G, N, and IR data from these deposits with principal components analysis provides an objective means to determine if units can be recognized within the braided-stream deposits. In particular, performing principal components analysis on G, N, and IR data from eight wells at Capital Station (1) allows the variable system dimensionality to be reduced from three to two by selecting the two eigenvectors with the greatest variance as axes for principal component scatterplots, (2) generates principal components with interpretable physical meanings, (3) distinguishes sand from cobble-dominated units, and (4) provides a means to distinguish between cobble-dominated units.
Lü, Gui-Cai; Zhao, Wei-Hong; Wang, Jiang-Tao
2011-01-01
The identification techniques for 10 species of red tide algae often found in the coastal areas of China were developed by combining the three-dimensional fluorescence spectra of fluorescence dissolved organic matter (FDOM) from the cultured red tide algae with principal component analysis. Based on the results of principal component analysis, the first principal component loading spectrum of three-dimensional fluorescence spectrum was chosen as the identification characteristic spectrum for red tide algae, and the phytoplankton fluorescence characteristic spectrum band was established. Then the 10 algae species were tested using Bayesian discriminant analysis with a correct identification rate of more than 92% for Pyrrophyta on the level of species, and that of more than 75% for Bacillariophyta on the level of genus in which the correct identification rates were more than 90% for the phaeodactylum and chaetoceros. The results showed that the identification techniques for 10 species of red tide algae based on the three-dimensional fluorescence spectra of FDOM from the cultured red tide algae and principal component analysis could work well.
A Multi-Dimensional Functional Principal Components Analysis of EEG Data
Hasenstab, Kyle; Scheffler, Aaron; Telesca, Donatello; Sugar, Catherine A.; Jeste, Shafali; DiStefano, Charlotte; Şentürk, Damla
2017-01-01
Summary The electroencephalography (EEG) data created in event-related potential (ERP) experiments have a complex high-dimensional structure. Each stimulus presentation, or trial, generates an ERP waveform which is an instance of functional data. The experiments are made up of sequences of multiple trials, resulting in longitudinal functional data and moreover, responses are recorded at multiple electrodes on the scalp, adding an electrode dimension. Traditional EEG analyses involve multiple simplifications of this structure to increase the signal-to-noise ratio, effectively collapsing the functional and longitudinal components by identifying key features of the ERPs and averaging them across trials. Motivated by an implicit learning paradigm used in autism research in which the functional, longitudinal and electrode components all have critical interpretations, we propose a multidimensional functional principal components analysis (MD-FPCA) technique which does not collapse any of the dimensions of the ERP data. The proposed decomposition is based on separation of the total variation into subject and subunit level variation which are further decomposed in a two-stage functional principal components analysis. The proposed methodology is shown to be useful for modeling longitudinal trends in the ERP functions, leading to novel insights into the learning patterns of children with Autism Spectrum Disorder (ASD) and their typically developing peers as well as comparisons between the two groups. Finite sample properties of MD-FPCA are further studied via extensive simulations. PMID:28072468
A multi-dimensional functional principal components analysis of EEG data.
Hasenstab, Kyle; Scheffler, Aaron; Telesca, Donatello; Sugar, Catherine A; Jeste, Shafali; DiStefano, Charlotte; Şentürk, Damla
2017-09-01
The electroencephalography (EEG) data created in event-related potential (ERP) experiments have a complex high-dimensional structure. Each stimulus presentation, or trial, generates an ERP waveform which is an instance of functional data. The experiments are made up of sequences of multiple trials, resulting in longitudinal functional data and moreover, responses are recorded at multiple electrodes on the scalp, adding an electrode dimension. Traditional EEG analyses involve multiple simplifications of this structure to increase the signal-to-noise ratio, effectively collapsing the functional and longitudinal components by identifying key features of the ERPs and averaging them across trials. Motivated by an implicit learning paradigm used in autism research in which the functional, longitudinal, and electrode components all have critical interpretations, we propose a multidimensional functional principal components analysis (MD-FPCA) technique which does not collapse any of the dimensions of the ERP data. The proposed decomposition is based on separation of the total variation into subject and subunit level variation which are further decomposed in a two-stage functional principal components analysis. The proposed methodology is shown to be useful for modeling longitudinal trends in the ERP functions, leading to novel insights into the learning patterns of children with Autism Spectrum Disorder (ASD) and their typically developing peers as well as comparisons between the two groups. Finite sample properties of MD-FPCA are further studied via extensive simulations. © 2017, The International Biometric Society.
[A study of Boletus bicolor from different areas using Fourier transform infrared spectrometry].
Zhou, Zai-Jin; Liu, Gang; Ren, Xian-Pei
2010-04-01
It is hard to differentiate the same species of wild growing mushrooms from different areas by macromorphological features. In this paper, Fourier transform infrared (FTIR) spectroscopy combined with principal component analysis was used to identify 58 samples of boletus bicolor from five different areas. Based on the fingerprint infrared spectrum of boletus bicolor samples, principal component analysis was conducted on 58 boletus bicolor spectra in the range of 1 350-750 cm(-1) using the statistical software SPSS 13.0. According to the result, the accumulated contributing ratio of the first three principal components accounts for 88.87%. They included almost all the information of samples. The two-dimensional projection plot using first and second principal component is a satisfactory clustering effect for the classification and discrimination of boletus bicolor. All boletus bicolor samples were divided into five groups with a classification accuracy of 98.3%. The study demonstrated that wild growing boletus bicolor at species level from different areas can be identified by FTIR spectra combined with principal components analysis.
Tan, Xinran; Zhu, Fan; Wang, Chao; Yu, Yang; Shi, Jian; Qi, Xue; Yuan, Feng; Tan, Jiubin
2017-11-19
This study presents a two-dimensional micro-/nanoradian angle generator (2D-MNAG) that achieves high angular displacement resolution and repeatability using a piezo-driven flexure hinge for two-dimensional deflections and three capacitive sensors for output angle monitoring and feedback control. The principal error of the capacitive sensor for precision microangle measurement is analyzed and compensated for; so as to achieve a high angle output resolution of 10 nrad (0.002 arcsec) and positioning repeatability of 120 nrad (0.024 arcsec) over a large angular range of ±4363 μrad (±900 arcsec) for the 2D-MNAG. The impact of each error component, together with the synthetic error of the 2D-MNAG after principal error compensation are determined using Monte Carlo simulation for further improvement of the 2D-MNAG.
Tan, Xinran; Zhu, Fan; Wang, Chao; Yu, Yang; Shi, Jian; Qi, Xue; Yuan, Feng; Tan, Jiubin
2017-01-01
This study presents a two-dimensional micro-/nanoradian angle generator (2D-MNAG) that achieves high angular displacement resolution and repeatability using a piezo-driven flexure hinge for two-dimensional deflections and three capacitive sensors for output angle monitoring and feedback control. The principal error of the capacitive sensor for precision microangle measurement is analyzed and compensated for; so as to achieve a high angle output resolution of 10 nrad (0.002 arcsec) and positioning repeatability of 120 nrad (0.024 arcsec) over a large angular range of ±4363 μrad (±900 arcsec) for the 2D-MNAG. The impact of each error component, together with the synthetic error of the 2D-MNAG after principal error compensation are determined using Monte Carlo simulation for further improvement of the 2D-MNAG. PMID:29156595
Catanuto, Giuseppe; Taher, Wafa; Rocco, Nicola; Catalano, Francesca; Allegra, Dario; Milotta, Filippo Luigi Maria; Stanco, Filippo; Gallo, Giovanni; Nava, Maurizio Bruno
2018-03-20
Breast shape is defined utilizing mainly qualitative assessment (full, flat, ptotic) or estimates, such as volume or distances between reference points, that cannot describe it reliably. We will quantitatively describe breast shape with two parameters derived from a statistical methodology denominated principal component analysis (PCA). We created a heterogeneous dataset of breast shapes acquired with a commercial infrared 3-dimensional scanner on which PCA was performed. We plotted on a Cartesian plane the two highest values of PCA for each breast (principal components 1 and 2). Testing of the methodology on a preoperative and postoperative surgical case and test-retest was performed by two operators. The first two principal components derived from PCA are able to characterize the shape of the breast included in the dataset. The test-retest demonstrated that different operators are able to obtain very similar values of PCA. The system is also able to identify major changes in the preoperative and postoperative stages of a two-stage reconstruction. Even minor changes were correctly detected by the system. This methodology can reliably describe the shape of a breast. An expert operator and a newly trained operator can reach similar results in a test/re-testing validation. Once developed and after further validation, this methodology could be employed as a good tool for outcome evaluation, auditing, and benchmarking.
Similarities between principal components of protein dynamics and random diffusion
NASA Astrophysics Data System (ADS)
Hess, Berk
2000-12-01
Principal component analysis, also called essential dynamics, is a powerful tool for finding global, correlated motions in atomic simulations of macromolecules. It has become an established technique for analyzing molecular dynamics simulations of proteins. The first few principal components of simulations of large proteins often resemble cosines. We derive the principal components for high-dimensional random diffusion, which are almost perfect cosines. This resemblance between protein simulations and noise implies that for many proteins the time scales of current simulations are too short to obtain convergence of collective motions.
Li, Ziyi; Safo, Sandra E; Long, Qi
2017-07-11
Sparse principal component analysis (PCA) is a popular tool for dimensionality reduction, pattern recognition, and visualization of high dimensional data. It has been recognized that complex biological mechanisms occur through concerted relationships of multiple genes working in networks that are often represented by graphs. Recent work has shown that incorporating such biological information improves feature selection and prediction performance in regression analysis, but there has been limited work on extending this approach to PCA. In this article, we propose two new sparse PCA methods called Fused and Grouped sparse PCA that enable incorporation of prior biological information in variable selection. Our simulation studies suggest that, compared to existing sparse PCA methods, the proposed methods achieve higher sensitivity and specificity when the graph structure is correctly specified, and are fairly robust to misspecified graph structures. Application to a glioblastoma gene expression dataset identified pathways that are suggested in the literature to be related with glioblastoma. The proposed sparse PCA methods Fused and Grouped sparse PCA can effectively incorporate prior biological information in variable selection, leading to improved feature selection and more interpretable principal component loadings and potentially providing insights on molecular underpinnings of complex diseases.
Asymptotics of empirical eigenstructure for high dimensional spiked covariance.
Wang, Weichen; Fan, Jianqing
2017-06-01
We derive the asymptotic distributions of the spiked eigenvalues and eigenvectors under a generalized and unified asymptotic regime, which takes into account the magnitude of spiked eigenvalues, sample size, and dimensionality. This regime allows high dimensionality and diverging eigenvalues and provides new insights into the roles that the leading eigenvalues, sample size, and dimensionality play in principal component analysis. Our results are a natural extension of those in Paul (2007) to a more general setting and solve the rates of convergence problems in Shen et al. (2013). They also reveal the biases of estimating leading eigenvalues and eigenvectors by using principal component analysis, and lead to a new covariance estimator for the approximate factor model, called shrinkage principal orthogonal complement thresholding (S-POET), that corrects the biases. Our results are successfully applied to outstanding problems in estimation of risks of large portfolios and false discovery proportions for dependent test statistics and are illustrated by simulation studies.
Asymptotics of empirical eigenstructure for high dimensional spiked covariance
Wang, Weichen
2017-01-01
We derive the asymptotic distributions of the spiked eigenvalues and eigenvectors under a generalized and unified asymptotic regime, which takes into account the magnitude of spiked eigenvalues, sample size, and dimensionality. This regime allows high dimensionality and diverging eigenvalues and provides new insights into the roles that the leading eigenvalues, sample size, and dimensionality play in principal component analysis. Our results are a natural extension of those in Paul (2007) to a more general setting and solve the rates of convergence problems in Shen et al. (2013). They also reveal the biases of estimating leading eigenvalues and eigenvectors by using principal component analysis, and lead to a new covariance estimator for the approximate factor model, called shrinkage principal orthogonal complement thresholding (S-POET), that corrects the biases. Our results are successfully applied to outstanding problems in estimation of risks of large portfolios and false discovery proportions for dependent test statistics and are illustrated by simulation studies. PMID:28835726
PCA based clustering for brain tumor segmentation of T1w MRI images.
Kaya, Irem Ersöz; Pehlivanlı, Ayça Çakmak; Sekizkardeş, Emine Gezmez; Ibrikci, Turgay
2017-03-01
Medical images are huge collections of information that are difficult to store and process consuming extensive computing time. Therefore, the reduction techniques are commonly used as a data pre-processing step to make the image data less complex so that a high-dimensional data can be identified by an appropriate low-dimensional representation. PCA is one of the most popular multivariate methods for data reduction. This paper is focused on T1-weighted MRI images clustering for brain tumor segmentation with dimension reduction by different common Principle Component Analysis (PCA) algorithms. Our primary aim is to present a comparison between different variations of PCA algorithms on MRIs for two cluster methods. Five most common PCA algorithms; namely the conventional PCA, Probabilistic Principal Component Analysis (PPCA), Expectation Maximization Based Principal Component Analysis (EM-PCA), Generalize Hebbian Algorithm (GHA), and Adaptive Principal Component Extraction (APEX) were applied to reduce dimensionality in advance of two clustering algorithms, K-Means and Fuzzy C-Means. In the study, the T1-weighted MRI images of the human brain with brain tumor were used for clustering. In addition to the original size of 512 lines and 512 pixels per line, three more different sizes, 256 × 256, 128 × 128 and 64 × 64, were included in the study to examine their effect on the methods. The obtained results were compared in terms of both the reconstruction errors and the Euclidean distance errors among the clustered images containing the same number of principle components. According to the findings, the PPCA obtained the best results among all others. Furthermore, the EM-PCA and the PPCA assisted K-Means algorithm to accomplish the best clustering performance in the majority as well as achieving significant results with both clustering algorithms for all size of T1w MRI images. Copyright © 2016 Elsevier Ireland Ltd. All rights reserved.
NASA Astrophysics Data System (ADS)
Fang, Leyuan; Wang, Chong; Li, Shutao; Yan, Jun; Chen, Xiangdong; Rabbani, Hossein
2017-11-01
We present an automatic method, termed as the principal component analysis network with composite kernel (PCANet-CK), for the classification of three-dimensional (3-D) retinal optical coherence tomography (OCT) images. Specifically, the proposed PCANet-CK method first utilizes the PCANet to automatically learn features from each B-scan of the 3-D retinal OCT images. Then, multiple kernels are separately applied to a set of very important features of the B-scans and these kernels are fused together, which can jointly exploit the correlations among features of the 3-D OCT images. Finally, the fused (composite) kernel is incorporated into an extreme learning machine for the OCT image classification. We tested our proposed algorithm on two real 3-D spectral domain OCT (SD-OCT) datasets (of normal subjects and subjects with the macular edema and age-related macular degeneration), which demonstrated its effectiveness.
Classification of fMRI resting-state maps using machine learning techniques: A comparative study
NASA Astrophysics Data System (ADS)
Gallos, Ioannis; Siettos, Constantinos
2017-11-01
We compare the efficiency of Principal Component Analysis (PCA) and nonlinear learning manifold algorithms (ISOMAP and Diffusion maps) for classifying brain maps between groups of schizophrenia patients and healthy from fMRI scans during a resting-state experiment. After a standard pre-processing pipeline, we applied spatial Independent component analysis (ICA) to reduce (a) noise and (b) spatial-temporal dimensionality of fMRI maps. On the cross-correlation matrix of the ICA components, we applied PCA, ISOMAP and Diffusion Maps to find an embedded low-dimensional space. Finally, support-vector-machines (SVM) and k-NN algorithms were used to evaluate the performance of the algorithms in classifying between the two groups.
NASA Astrophysics Data System (ADS)
Chung, Hyunkoo; Lu, Guolan; Tian, Zhiqiang; Wang, Dongsheng; Chen, Zhuo Georgia; Fei, Baowei
2016-03-01
Hyperspectral imaging (HSI) is an emerging imaging modality for medical applications. HSI acquires two dimensional images at various wavelengths. The combination of both spectral and spatial information provides quantitative information for cancer detection and diagnosis. This paper proposes using superpixels, principal component analysis (PCA), and support vector machine (SVM) to distinguish regions of tumor from healthy tissue. The classification method uses 2 principal components decomposed from hyperspectral images and obtains an average sensitivity of 93% and an average specificity of 85% for 11 mice. The hyperspectral imaging technology and classification method can have various applications in cancer research and management.
Nguyen, Phuong H
2007-05-15
Principal component analysis is a powerful method for projecting multidimensional conformational space of peptides or proteins onto lower dimensional subspaces in which the main conformations are present, making it easier to reveal the structures of molecules from e.g. molecular dynamics simulation trajectories. However, the identification of all conformational states is still difficult if the subspaces consist of more than two dimensions. This is mainly due to the fact that the principal components are not independent with each other, and states in the subspaces cannot be visualized. In this work, we propose a simple and fast scheme that allows one to obtain all conformational states in the subspaces. The basic idea is that instead of directly identifying the states in the subspace spanned by principal components, we first transform this subspace into another subspace formed by components that are independent of one other. These independent components are obtained from the principal components by employing the independent component analysis method. Because of independence between components, all states in this new subspace are defined as all possible combinations of the states obtained from each single independent component. This makes the conformational analysis much simpler. We test the performance of the method by analyzing the conformations of the glycine tripeptide and the alanine hexapeptide. The analyses show that our method is simple and quickly reveal all conformational states in the subspaces. The folding pathways between the identified states of the alanine hexapeptide are analyzed and discussed in some detail. 2007 Wiley-Liss, Inc.
Use of multivariate statistics to identify unreliable data obtained using CASA.
Martínez, Luis Becerril; Crispín, Rubén Huerta; Mendoza, Maximino Méndez; Gallegos, Oswaldo Hernández; Martínez, Andrés Aragón
2013-06-01
In order to identify unreliable data in a dataset of motility parameters obtained from a pilot study acquired by a veterinarian with experience in boar semen handling, but without experience in the operation of a computer assisted sperm analysis (CASA) system, a multivariate graphical and statistical analysis was performed. Sixteen boar semen samples were aliquoted then incubated with varying concentrations of progesterone from 0 to 3.33 µg/ml and analyzed in a CASA system. After standardization of the data, Chernoff faces were pictured for each measurement, and a principal component analysis (PCA) was used to reduce the dimensionality and pre-process the data before hierarchical clustering. The first twelve individual measurements showed abnormal features when Chernoff faces were drawn. PCA revealed that principal components 1 and 2 explained 63.08% of the variance in the dataset. Values of principal components for each individual measurement of semen samples were mapped to identify differences among treatment or among boars. Twelve individual measurements presented low values of principal component 1. Confidence ellipses on the map of principal components showed no statistically significant effects for treatment or boar. Hierarchical clustering realized on two first principal components produced three clusters. Cluster 1 contained evaluations of the two first samples in each treatment, each one of a different boar. With the exception of one individual measurement, all other measurements in cluster 1 were the same as observed in abnormal Chernoff faces. Unreliable data in cluster 1 are probably related to the operator inexperience with a CASA system. These findings could be used to objectively evaluate the skill level of an operator of a CASA system. This may be particularly useful in the quality control of semen analysis using CASA systems.
Two-dimensional PCA-based human gait identification
NASA Astrophysics Data System (ADS)
Chen, Jinyan; Wu, Rongteng
2012-11-01
It is very necessary to recognize person through visual surveillance automatically for public security reason. Human gait based identification focus on recognizing human by his walking video automatically using computer vision and image processing approaches. As a potential biometric measure, human gait identification has attracted more and more researchers. Current human gait identification methods can be divided into two categories: model-based methods and motion-based methods. In this paper a two-Dimensional Principal Component Analysis and temporal-space analysis based human gait identification method is proposed. Using background estimation and image subtraction we can get a binary images sequence from the surveillance video. By comparing the difference of two adjacent images in the gait images sequence, we can get a difference binary images sequence. Every binary difference image indicates the body moving mode during a person walking. We use the following steps to extract the temporal-space features from the difference binary images sequence: Projecting one difference image to Y axis or X axis we can get two vectors. Project every difference image in the difference binary images sequence to Y axis or X axis difference binary images sequence we can get two matrixes. These two matrixes indicate the styles of one walking. Then Two-Dimensional Principal Component Analysis(2DPCA) is used to transform these two matrixes to two vectors while at the same time keep the maximum separability. Finally the similarity of two human gait images is calculated by the Euclidean distance of the two vectors. The performance of our methods is illustrated using the CASIA Gait Database.
Online signature recognition using principal component analysis and artificial neural network
NASA Astrophysics Data System (ADS)
Hwang, Seung-Jun; Park, Seung-Je; Baek, Joong-Hwan
2016-12-01
In this paper, we propose an algorithm for on-line signature recognition using fingertip point in the air from the depth image acquired by Kinect. We extract 10 statistical features from X, Y, Z axis, which are invariant to changes in shifting and scaling of the signature trajectories in three-dimensional space. Artificial neural network is adopted to solve the complex signature classification problem. 30 dimensional features are converted into 10 principal components using principal component analysis, which is 99.02% of total variances. We implement the proposed algorithm and test to actual on-line signatures. In experiment, we verify the proposed method is successful to classify 15 different on-line signatures. Experimental result shows 98.47% of recognition rate when using only 10 feature vectors.
Psychometric Measurement Models and Artificial Neural Networks
ERIC Educational Resources Information Center
Sese, Albert; Palmer, Alfonso L.; Montano, Juan J.
2004-01-01
The study of measurement models in psychometrics by means of dimensionality reduction techniques such as Principal Components Analysis (PCA) is a very common practice. In recent times, an upsurge of interest in the study of artificial neural networks apt to computing a principal component extraction has been observed. Despite this interest, the…
Azevedo, C F; Nascimento, M; Silva, F F; Resende, M D V; Lopes, P S; Guimarães, S E F; Glória, L S
2015-10-09
A significant contribution of molecular genetics is the direct use of DNA information to identify genetically superior individuals. With this approach, genome-wide selection (GWS) can be used for this purpose. GWS consists of analyzing a large number of single nucleotide polymorphism markers widely distributed in the genome; however, because the number of markers is much larger than the number of genotyped individuals, and such markers are highly correlated, special statistical methods are widely required. Among these methods, independent component regression, principal component regression, partial least squares, and partial principal components stand out. Thus, the aim of this study was to propose an application of the methods of dimensionality reduction to GWS of carcass traits in an F2 (Piau x commercial line) pig population. The results show similarities between the principal and the independent component methods and provided the most accurate genomic breeding estimates for most carcass traits in pigs.
Dimensionality reduction of collective motion by principal manifolds
NASA Astrophysics Data System (ADS)
Gajamannage, Kelum; Butail, Sachit; Porfiri, Maurizio; Bollt, Erik M.
2015-01-01
While the existence of low-dimensional embedding manifolds has been shown in patterns of collective motion, the current battery of nonlinear dimensionality reduction methods is not amenable to the analysis of such manifolds. This is mainly due to the necessary spectral decomposition step, which limits control over the mapping from the original high-dimensional space to the embedding space. Here, we propose an alternative approach that demands a two-dimensional embedding which topologically summarizes the high-dimensional data. In this sense, our approach is closely related to the construction of one-dimensional principal curves that minimize orthogonal error to data points subject to smoothness constraints. Specifically, we construct a two-dimensional principal manifold directly in the high-dimensional space using cubic smoothing splines, and define the embedding coordinates in terms of geodesic distances. Thus, the mapping from the high-dimensional data to the manifold is defined in terms of local coordinates. Through representative examples, we show that compared to existing nonlinear dimensionality reduction methods, the principal manifold retains the original structure even in noisy and sparse datasets. The principal manifold finding algorithm is applied to configurations obtained from a dynamical system of multiple agents simulating a complex maneuver called predator mobbing, and the resulting two-dimensional embedding is compared with that of a well-established nonlinear dimensionality reduction method.
Three dimensional empirical mode decomposition analysis apparatus, method and article manufacture
NASA Technical Reports Server (NTRS)
Gloersen, Per (Inventor)
2004-01-01
An apparatus and method of analysis for three-dimensional (3D) physical phenomena. The physical phenomena may include any varying 3D phenomena such as time varying polar ice flows. A repesentation of the 3D phenomena is passed through a Hilbert transform to convert the data into complex form. A spatial variable is separated from the complex representation by producing a time based covariance matrix. The temporal parts of the principal components are produced by applying Singular Value Decomposition (SVD). Based on the rapidity with which the eigenvalues decay, the first 3-10 complex principal components (CPC) are selected for Empirical Mode Decomposition into intrinsic modes. The intrinsic modes produced are filtered in order to reconstruct the spatial part of the CPC. Finally, a filtered time series may be reconstructed from the first 3-10 filtered complex principal components.
Modified neural networks for rapid recovery of tokamak plasma parameters for real time control
NASA Astrophysics Data System (ADS)
Sengupta, A.; Ranjan, P.
2002-07-01
Two modified neural network techniques are used for the identification of the equilibrium plasma parameters of the Superconducting Steady State Tokamak I from external magnetic measurements. This is expected to ultimately assist in a real time plasma control. As different from the conventional network structure where a single network with the optimum number of processing elements calculates the outputs, a multinetwork system connected in parallel does the calculations here in one of the methods. This network is called the double neural network. The accuracy of the recovered parameters is clearly more than the conventional network. The other type of neural network used here is based on the statistical function parametrization combined with a neural network. The principal component transformation removes linear dependences from the measurements and a dimensional reduction process reduces the dimensionality of the input space. This reduced and transformed input set, rather than the entire set, is fed into the neural network input. This is known as the principal component transformation-based neural network. The accuracy of the recovered parameters in the latter type of modified network is found to be a further improvement over the accuracy of the double neural network. This result differs from that obtained in an earlier work where the double neural network showed better performance. The conventional network and the function parametrization methods have also been used for comparison. The conventional network has been used for an optimization of the set of magnetic diagnostics. The effective set of sensors, as assessed by this network, are compared with the principal component based network. Fault tolerance of the neural networks has been tested. The double neural network showed the maximum resistance to faults in the diagnostics, while the principal component based network performed poorly. Finally the processing times of the methods have been compared. The double network and the principal component network involve the minimum computation time, although the conventional network also performs well enough to be used in real time.
XCOM intrinsic dimensionality for low-Z elements at diagnostic energies
DOE Office of Scientific and Technical Information (OSTI.GOV)
Bornefalk, Hans
2012-02-15
Purpose: To determine the intrinsic dimensionality of linear attenuation coefficients (LACs) from XCOM for elements with low atomic number (Z = 1-20) at diagnostic x-ray energies (25-120 keV). H{sub 0}{sup q}, the hypothesis that the space of LACs is spanned by q bases, is tested for various q-values. Methods: Principal component analysis is first applied and the LACs are projected onto the first q principal component bases. The residuals of the model values vs XCOM data are determined for all energies and atomic numbers. Heteroscedasticity invalidates the prerequisite of i.i.d. errors necessary for bootstrapping residuals. Instead wild bootstrap is applied,more » which, by not mixing residuals, allows the effect of the non-i.i.d residuals to be reflected in the result. Credible regions for the eigenvalues of the correlation matrix for the bootstrapped LAC data are determined. If subsequent credible regions for the eigenvalues overlap, the corresponding principal component is not considered to represent true data structure but noise. If this happens for eigenvalues l and l + 1, for any l{<=}q, H{sub 0}{sup q} is rejected. Results: The largest value of q for which H{sub 0}{sup q} is nonrejectable at the 5%-level is q = 4. This indicates that the statistically significant intrinsic dimensionality of low-Z XCOM data at diagnostic energies is four. Conclusions: The method presented allows determination of the statistically significant dimensionality of any noisy linear subspace. Knowledge of such significant dimensionality is of interest for any method making assumptions on intrinsic dimensionality and evaluating results on noisy reference data. For LACs, knowledge of the low-Z dimensionality might be relevant when parametrization schemes are tuned to XCOM data. For x-ray imaging techniques based on the basis decomposition method (Alvarez and Macovski, Phys. Med. Biol. 21, 733-744, 1976), an underlying dimensionality of two is commonly assigned to the LAC of human tissue at diagnostic energies. The finding of a higher statistically significant dimensionality thus raises the question whether a higher assumed model dimensionality (now feasible with the advent of multibin x-ray systems) might also be practically relevant, i.e., if better tissue characterization results can be obtained.« less
Fang, Leyuan; Wang, Chong; Li, Shutao; Yan, Jun; Chen, Xiangdong; Rabbani, Hossein
2017-11-01
We present an automatic method, termed as the principal component analysis network with composite kernel (PCANet-CK), for the classification of three-dimensional (3-D) retinal optical coherence tomography (OCT) images. Specifically, the proposed PCANet-CK method first utilizes the PCANet to automatically learn features from each B-scan of the 3-D retinal OCT images. Then, multiple kernels are separately applied to a set of very important features of the B-scans and these kernels are fused together, which can jointly exploit the correlations among features of the 3-D OCT images. Finally, the fused (composite) kernel is incorporated into an extreme learning machine for the OCT image classification. We tested our proposed algorithm on two real 3-D spectral domain OCT (SD-OCT) datasets (of normal subjects and subjects with the macular edema and age-related macular degeneration), which demonstrated its effectiveness. (2017) COPYRIGHT Society of Photo-Optical Instrumentation Engineers (SPIE).
Dimensionality Reduction Through Classifier Ensembles
NASA Technical Reports Server (NTRS)
Oza, Nikunj C.; Tumer, Kagan; Norwig, Peter (Technical Monitor)
1999-01-01
In data mining, one often needs to analyze datasets with a very large number of attributes. Performing machine learning directly on such data sets is often impractical because of extensive run times, excessive complexity of the fitted model (often leading to overfitting), and the well-known "curse of dimensionality." In practice, to avoid such problems, feature selection and/or extraction are often used to reduce data dimensionality prior to the learning step. However, existing feature selection/extraction algorithms either evaluate features by their effectiveness across the entire data set or simply disregard class information altogether (e.g., principal component analysis). Furthermore, feature extraction algorithms such as principal components analysis create new features that are often meaningless to human users. In this article, we present input decimation, a method that provides "feature subsets" that are selected for their ability to discriminate among the classes. These features are subsequently used in ensembles of classifiers, yielding results superior to single classifiers, ensembles that use the full set of features, and ensembles based on principal component analysis on both real and synthetic datasets.
Exploring patterns enriched in a dataset with contrastive principal component analysis.
Abid, Abubakar; Zhang, Martin J; Bagaria, Vivek K; Zou, James
2018-05-30
Visualization and exploration of high-dimensional data is a ubiquitous challenge across disciplines. Widely used techniques such as principal component analysis (PCA) aim to identify dominant trends in one dataset. However, in many settings we have datasets collected under different conditions, e.g., a treatment and a control experiment, and we are interested in visualizing and exploring patterns that are specific to one dataset. This paper proposes a method, contrastive principal component analysis (cPCA), which identifies low-dimensional structures that are enriched in a dataset relative to comparison data. In a wide variety of experiments, we demonstrate that cPCA with a background dataset enables us to visualize dataset-specific patterns missed by PCA and other standard methods. We further provide a geometric interpretation of cPCA and strong mathematical guarantees. An implementation of cPCA is publicly available, and can be used for exploratory data analysis in many applications where PCA is currently used.
NASA Astrophysics Data System (ADS)
Polat, Esra; Gunay, Suleyman
2013-10-01
One of the problems encountered in Multiple Linear Regression (MLR) is multicollinearity, which causes the overestimation of the regression parameters and increase of the variance of these parameters. Hence, in case of multicollinearity presents, biased estimation procedures such as classical Principal Component Regression (CPCR) and Partial Least Squares Regression (PLSR) are then performed. SIMPLS algorithm is the leading PLSR algorithm because of its speed, efficiency and results are easier to interpret. However, both of the CPCR and SIMPLS yield very unreliable results when the data set contains outlying observations. Therefore, Hubert and Vanden Branden (2003) have been presented a robust PCR (RPCR) method and a robust PLSR (RPLSR) method called RSIMPLS. In RPCR, firstly, a robust Principal Component Analysis (PCA) method for high-dimensional data on the independent variables is applied, then, the dependent variables are regressed on the scores using a robust regression method. RSIMPLS has been constructed from a robust covariance matrix for high-dimensional data and robust linear regression. The purpose of this study is to show the usage of RPCR and RSIMPLS methods on an econometric data set, hence, making a comparison of two methods on an inflation model of Turkey. The considered methods have been compared in terms of predictive ability and goodness of fit by using a robust Root Mean Squared Error of Cross-validation (R-RMSECV), a robust R2 value and Robust Component Selection (RCS) statistic.
Lin, Jyh-Woei
2013-01-01
Two-dimensional principal component analysis (2DPCA) and principal component analysis (PCA) are used to examine the ionospheric total electron content (TEC) data during the time period from 00:00 on August 21 to 12: 45 on August 31 (UT), which are 10 days before the M = 7.6 Philippines earthquake at 12:47:34 on August 31, 2012 (UT) with the depth at 34.9 km. From the results by using 2DPCA, a TEC precursor of Philippines earthquake is found during the time period from 4:25 to 4:40 on August 28, 2012 (UT) with the duration time of at least 15 minutes. Another earthquake-related TEC anomaly is detectable for the time period from 04:35 to 04:40 on August 27, 2012 (UT) with the duration time of at least 5 minutes during the Puerto earthquake at 04: 37:20 on August 27, 2012 (UT) (M w = 7.3) with the depth at 20.3 km. The precursor of the Puerto earthquake is not detectable. TEC anomaly is not to be found related to the Jan Mayen Island earthquake (M w = 6.8) at 13:43:24 on August 30, 2012 (UT). These earthquake-related TEC anomalies are detectable by using 2DPCA rather than PCA. They are localized nearby the epicenters of the Philippines and Puerto earthquakes. PMID:23844386
Zuendorf, Gerhard; Kerrouche, Nacer; Herholz, Karl; Baron, Jean-Claude
2003-01-01
Principal component analysis (PCA) is a well-known technique for reduction of dimensionality of functional imaging data. PCA can be looked at as the projection of the original images onto a new orthogonal coordinate system with lower dimensions. The new axes explain the variance in the images in decreasing order of importance, showing correlations between brain regions. We used an efficient, stable and analytical method to work out the PCA of Positron Emission Tomography (PET) images of 74 normal subjects using [(18)F]fluoro-2-deoxy-D-glucose (FDG) as a tracer. Principal components (PCs) and their relation to age effects were investigated. Correlations between the projections of the images on the new axes and the age of the subjects were carried out. The first two PCs could be identified as being the only PCs significantly correlated to age. The first principal component, which explained 10% of the data set variance, was reduced only in subjects of age 55 or older and was related to loss of signal in and adjacent to ventricles and basal cisterns, reflecting expected age-related brain atrophy with enlarging CSF spaces. The second principal component, which accounted for 8% of the total variance, had high loadings from prefrontal, posterior parietal and posterior cingulate cortices and showed the strongest correlation with age (r = -0.56), entirely consistent with previously documented age-related declines in brain glucose utilization. Thus, our method showed that the effect of aging on brain metabolism has at least two independent dimensions. This method should have widespread applications in multivariate analysis of brain functional images. Copyright 2002 Wiley-Liss, Inc.
Dynamic competitive probabilistic principal components analysis.
López-Rubio, Ezequiel; Ortiz-DE-Lazcano-Lobato, Juan Miguel
2009-04-01
We present a new neural model which extends the classical competitive learning (CL) by performing a Probabilistic Principal Components Analysis (PPCA) at each neuron. The model also has the ability to learn the number of basis vectors required to represent the principal directions of each cluster, so it overcomes a drawback of most local PCA models, where the dimensionality of a cluster must be fixed a priori. Experimental results are presented to show the performance of the network with multispectral image data.
Zhang, Wanfeng; Zhu, Shukui; He, Sheng; Wang, Yanxin
2015-02-06
Using comprehensive two-dimensional gas chromatography coupled to time-of-flight mass spectrometry (GC×GC/TOFMS), volatile and semi-volatile organic compounds in crude oil samples from different reservoirs or regions were analyzed for the development of a molecular fingerprint database. Based on the GC×GC/TOFMS fingerprints of crude oils, principal component analysis (PCA) and cluster analysis were used to distinguish the oil sources and find biomarkers. As a supervised technique, the geological characteristics of crude oils, including thermal maturity, sedimentary environment etc., are assigned to the principal components. The results show that tri-aromatic steroid (TAS) series are the suitable marker compounds in crude oils for the oil screening, and the relative abundances of individual TAS compounds have excellent correlation with oil sources. In order to correct the effects of some other external factors except oil sources, the variables were defined as the content ratio of some target compounds and 13 parameters were proposed for the screening of oil sources. With the developed model, the crude oils were easily discriminated, and the result is in good agreement with the practical geological setting. Copyright © 2014 Elsevier B.V. All rights reserved.
NASA Astrophysics Data System (ADS)
Price-Jones, Natalie; Bovy, Jo
2018-03-01
Chemical tagging of stars based on their similar compositions can offer new insights about the star formation and dynamical history of the Milky Way. We investigate the feasibility of identifying groups of stars in chemical space by forgoing the use of model derived abundances in favour of direct analysis of spectra. This facilitates the propagation of measurement uncertainties and does not pre-suppose knowledge of which elements are important for distinguishing stars in chemical space. We use ˜16 000 red giant and red clump H-band spectra from the Apache Point Observatory Galactic Evolution Experiment (APOGEE) and perform polynomial fits to remove trends not due to abundance-ratio variations. Using expectation maximized principal component analysis, we find principal components with high signal in the wavelength regions most important for distinguishing between stars. Different subsamples of red giant and red clump stars are all consistent with needing about 10 principal components to accurately model the spectra above the level of the measurement uncertainties. The dimensionality of stellar chemical space that can be investigated in the H band is therefore ≲10. For APOGEE observations with typical signal-to-noise ratios of 100, the number of chemical space cells within which stars cannot be distinguished is approximately 1010±2 × (5 ± 2)n - 10 with n the number of principal components. This high dimensionality and the fine-grained sampling of chemical space are a promising first step towards chemical tagging based on spectra alone.
NASA Astrophysics Data System (ADS)
Li, Jiangtong; Luo, Yongdao; Dai, Honglin
2018-01-01
Water is the source of life and the essential foundation of all life. With the development of industrialization, the phenomenon of water pollution is becoming more and more frequent, which directly affects the survival and development of human. Water quality detection is one of the necessary measures to protect water resources. Ultraviolet (UV) spectral analysis is an important research method in the field of water quality detection, which partial least squares regression (PLSR) analysis method is becoming predominant technology, however, in some special cases, PLSR's analysis produce considerable errors. In order to solve this problem, the traditional principal component regression (PCR) analysis method was improved by using the principle of PLSR in this paper. The experimental results show that for some special experimental data set, improved PCR analysis method performance is better than PLSR. The PCR and PLSR is the focus of this paper. Firstly, the principal component analysis (PCA) is performed by MATLAB to reduce the dimensionality of the spectral data; on the basis of a large number of experiments, the optimized principal component is extracted by using the principle of PLSR, which carries most of the original data information. Secondly, the linear regression analysis of the principal component is carried out with statistic package for social science (SPSS), which the coefficients and relations of principal components can be obtained. Finally, calculating a same water spectral data set by PLSR and improved PCR, analyzing and comparing two results, improved PCR and PLSR is similar for most data, but improved PCR is better than PLSR for data near the detection limit. Both PLSR and improved PCR can be used in Ultraviolet spectral analysis of water, but for data near the detection limit, improved PCR's result better than PLSR.
Karasawa, N; Mitsutake, A; Takano, H
2017-12-01
Proteins implement their functionalities when folded into specific three-dimensional structures, and their functions are related to the protein structures and dynamics. Previously, we applied a relaxation mode analysis (RMA) method to protein systems; this method approximately estimates the slow relaxation modes and times via simulation and enables investigation of the dynamic properties underlying the protein structural fluctuations. Recently, two-step RMA with multiple evolution times has been proposed and applied to a slightly complex homopolymer system, i.e., a single [n]polycatenane. This method can be applied to more complex heteropolymer systems, i.e., protein systems, to estimate the relaxation modes and times more accurately. In two-step RMA, we first perform RMA and obtain rough estimates of the relaxation modes and times. Then, we apply RMA with multiple evolution times to a small number of the slowest relaxation modes obtained in the previous calculation. Herein, we apply this method to the results of principal component analysis (PCA). First, PCA is applied to a 2-μs molecular dynamics simulation of hen egg-white lysozyme in aqueous solution. Then, the two-step RMA method with multiple evolution times is applied to the obtained principal components. The slow relaxation modes and corresponding relaxation times for the principal components are much improved by the second RMA.
NASA Astrophysics Data System (ADS)
Karasawa, N.; Mitsutake, A.; Takano, H.
2017-12-01
Proteins implement their functionalities when folded into specific three-dimensional structures, and their functions are related to the protein structures and dynamics. Previously, we applied a relaxation mode analysis (RMA) method to protein systems; this method approximately estimates the slow relaxation modes and times via simulation and enables investigation of the dynamic properties underlying the protein structural fluctuations. Recently, two-step RMA with multiple evolution times has been proposed and applied to a slightly complex homopolymer system, i.e., a single [n ] polycatenane. This method can be applied to more complex heteropolymer systems, i.e., protein systems, to estimate the relaxation modes and times more accurately. In two-step RMA, we first perform RMA and obtain rough estimates of the relaxation modes and times. Then, we apply RMA with multiple evolution times to a small number of the slowest relaxation modes obtained in the previous calculation. Herein, we apply this method to the results of principal component analysis (PCA). First, PCA is applied to a 2-μ s molecular dynamics simulation of hen egg-white lysozyme in aqueous solution. Then, the two-step RMA method with multiple evolution times is applied to the obtained principal components. The slow relaxation modes and corresponding relaxation times for the principal components are much improved by the second RMA.
NASA Astrophysics Data System (ADS)
Abdulla, Hussain A. N.; Minor, Elizabeth C.; Dias, Robert F.; Hatcher, Patrick G.
2013-10-01
In a study of chemical transformations of estuarine high-molecular-weight (HMW, >1000 Da) dissolved organic matter (DOM) collected over a period of two years along a transect through the Elizabeth River/Chesapeake Bay system to the coastal Atlantic Ocean off Virginia, USA, δ13C values, N/C ratios, and principal component analysis (PCA) of the solid-state 13C NMR (nuclear magnetic resonance) spectra of HMW-DOM show an abrupt change in both its sources and chemical structural composition occurring around salinity 20. HMW-DOM in the lower salinity region had lighter isotopic values, higher aromatic and lower carbohydrate contents relative to that in the higher salinity region. These changes around a salinity of 20 are possibly due to introduction of a significant amount of new carbon (autotrophic DOM) to the transect. PC-1 loadings plot shows that spatially differing DOM components are similar to previously reported 13C NMR spectra of heteropolysaccharides (HPS) and carboxyl-rich alicyclic molecules (CRAM). Applying two dimensional correlation spectroscopy techniques to 1H NMR spectra from the same samples reveals increases in the contribution of N-acetyl amino sugars, 6-deoxy sugars, and sulfated polysaccharides to HPS components along the salinity transect, which suggests a transition from plant derived carbohydrates to marine produced carbohydrates within the HMW-DOM pool. In contrast to what has been suggested previously, our combined results from 13C NMR, 1H NMR, and FTIR indicate that CRAM consists of at least two different classes of compounds (aliphatic polycarboxyl compounds and lignin-like compounds).
[Discrimination of varieties of brake fluid using visual-near infrared spectra].
Jiang, Lu-lu; Tan, Li-hong; Qiu, Zheng-jun; Lu, Jiang-feng; He, Yong
2008-06-01
A new method was developed to fast discriminate brands of brake fluid by means of visual-near infrared spectroscopy. Five different brands of brake fluid were analyzed using a handheld near infrared spectrograph, manufactured by ASD Company, and 60 samples were gotten from each brand of brake fluid. The samples data were pretreated using average smoothing and standard normal variable method, and then analyzed using principal component analysis (PCA). A 2-dimensional plot was drawn based on the first and the second principal components, and the plot indicated that the clustering characteristic of different brake fluid is distinct. The foregoing 6 principal components were taken as input variable, and the band of brake fluid as output variable to build the discriminate model by stepwise discriminant analysis method. Two hundred twenty five samples selected randomly were used to create the model, and the rest 75 samples to verify the model. The result showed that the distinguishing rate was 94.67%, indicating that the method proposed in this paper has good performance in classification and discrimination. It provides a new way to fast discriminate different brands of brake fluid.
Turgeon, Maxime; Oualkacha, Karim; Ciampi, Antonio; Miftah, Hanane; Dehghan, Golsa; Zanke, Brent W; Benedet, Andréa L; Rosa-Neto, Pedro; Greenwood, Celia Mt; Labbe, Aurélie
2018-05-01
The genomics era has led to an increase in the dimensionality of data collected in the investigation of biological questions. In this context, dimension-reduction techniques can be used to summarise high-dimensional signals into low-dimensional ones, to further test for association with one or more covariates of interest. This paper revisits one such approach, previously known as principal component of heritability and renamed here as principal component of explained variance (PCEV). As its name suggests, the PCEV seeks a linear combination of outcomes in an optimal manner, by maximising the proportion of variance explained by one or several covariates of interest. By construction, this method optimises power; however, due to its computational complexity, it has unfortunately received little attention in the past. Here, we propose a general analytical PCEV framework that builds on the assets of the original method, i.e. conceptually simple and free of tuning parameters. Moreover, our framework extends the range of applications of the original procedure by providing a computationally simple strategy for high-dimensional outcomes, along with exact and asymptotic testing procedures that drastically reduce its computational cost. We investigate the merits of the PCEV using an extensive set of simulations. Furthermore, the use of the PCEV approach is illustrated using three examples taken from the fields of epigenetics and brain imaging.
Multivariate analysis of light scattering spectra of liquid dairy products
NASA Astrophysics Data System (ADS)
Khodasevich, M. A.
2010-05-01
Visible light scattering spectra from the surface layer of samples of commercial liquid dairy products are recorded with a colorimeter. The principal component method is used to analyze these spectra. Vectors representing the samples of dairy products in a multidimensional space of spectral counts are projected onto a three-dimensional subspace of principal components. The magnitudes of these projections are found to depend on the type of dairy product.
ERIC Educational Resources Information Center
Steinley, Douglas; Brusco, Michael J.; Henson, Robert
2012-01-01
A measure of "clusterability" serves as the basis of a new methodology designed to preserve cluster structure in a reduced dimensional space. Similar to principal component analysis, which finds the direction of maximal variance in multivariate space, principal cluster axes find the direction of maximum clusterability in multivariate space.…
Steingass, Christof Björn; Jutzi, Manfred; Müller, Jenny; Carle, Reinhold; Schmarr, Hans-Georg
2015-03-01
Ripening-dependent changes of pineapple volatiles were studied in a nontargeted profiling analysis. Volatiles were isolated via headspace solid phase microextraction and analyzed by comprehensive 2D gas chromatography and mass spectrometry (HS-SPME-GC×GC-qMS). Profile patterns presented in the contour plots were evaluated applying image processing techniques and subsequent multivariate statistical data analysis. Statistical methods comprised unsupervised hierarchical cluster analysis (HCA) and principal component analysis (PCA) to classify the samples. Supervised partial least squares discriminant analysis (PLS-DA) and partial least squares (PLS) regression were applied to discriminate different ripening stages and describe the development of volatiles during postharvest storage, respectively. Hereby, substantial chemical markers allowing for class separation were revealed. The workflow permitted the rapid distinction between premature green-ripe pineapples and postharvest-ripened sea-freighted fruits. Volatile profiles of fully ripe air-freighted pineapples were similar to those of green-ripe fruits postharvest ripened for 6 days after simulated sea freight export, after PCA with only two principal components. However, PCA considering also the third principal component allowed differentiation between air-freighted fruits and the four progressing postharvest maturity stages of sea-freighted pineapples.
Kurosumi, M; Mizukoshi, K
2018-05-01
The types of shape feature that constitutes a face have not been comprehensively established, and most previous studies of age-related changes in facial shape have focused on individual characteristics, such as wrinkle, sagging skin, etc. In this study, we quantitatively measured differences in face shape between individuals and investigated how shape features changed with age. We analyzed three-dimensionally the faces of 280 Japanese women aged 20-69 years and used principal component analysis to establish the shape features that characterized individual differences. We also evaluated the relationships between each feature and age, clarifying the shape features characteristic of different age groups. Changes in facial shape in middle age were a decreased volume of the upper face and increased volume of the whole cheeks and around the chin. Changes in older people were an increased volume of the lower cheeks and around the chin, sagging skin, and jaw distortion. Principal component analysis was effective for identifying facial shape features that represent individual and age-related differences. This method allowed straightforward measurements, such as the increase or decrease in cheeks caused by soft tissue changes or skeletal-based changes to the forehead or jaw, simply by acquiring three-dimensional facial images. © 2017 John Wiley & Sons A/S. Published by John Wiley & Sons Ltd.
Metsalu, Tauno; Vilo, Jaak
2015-01-01
The Principal Component Analysis (PCA) is a widely used method of reducing the dimensionality of high-dimensional data, often followed by visualizing two of the components on the scatterplot. Although widely used, the method is lacking an easy-to-use web interface that scientists with little programming skills could use to make plots of their own data. The same applies to creating heatmaps: it is possible to add conditional formatting for Excel cells to show colored heatmaps, but for more advanced features such as clustering and experimental annotations, more sophisticated analysis tools have to be used. We present a web tool called ClustVis that aims to have an intuitive user interface. Users can upload data from a simple delimited text file that can be created in a spreadsheet program. It is possible to modify data processing methods and the final appearance of the PCA and heatmap plots by using drop-down menus, text boxes, sliders etc. Appropriate defaults are given to reduce the time needed by the user to specify input parameters. As an output, users can download PCA plot and heatmap in one of the preferred file formats. This web server is freely available at http://biit.cs.ut.ee/clustvis/. PMID:25969447
Balsamo, Geisi M; Valentim-Neto, Pedro A; Mello, Carla S; Arisi, Ana C M
2015-12-09
The genetically modified (GM) common bean event Embrapa 5.1 was commercially approved in Brazil in 2011; it is resistant to golden mosaic virus infection. In the present work grain proteome profiles of two Embrapa 5.1 common bean varieties, Pérola and Pontal, and their non-GM counterparts were compared by two-dimensional gel electrophoresis (2-DE) followed by mass spectrometry (MS). Analyses detected 23 spots differentially accumulated between GM Pérola and non-GM Pérola and 21 spots between GM Pontal and non-GM Pontal, although they were not the same proteins in Pérola and Pontal varieties, indicating that the variability observed may not be due to the genetic transformation. Among them, eight proteins were identified in Pérola varieties, and four proteins were identified in Pontal. Moreover, we applied principal component analysis (PCA) on 2-DE data, and variation between varieties was explained in the first two principal components. This work provides a first 2-DE-MS/MS-based analysis of Embrapa 5.1 common bean grains.
Reconfigurable exciton-plasmon interconversion for nanophotonic circuits
Lee, Hyun Seok; Luong, Dinh Hoa; Kim, Min Su; Jin, Youngjo; Kim, Hyun; Yun, Seokjoon; Lee, Young Hee
2016-01-01
The recent challenges for improving the operation speed of nanoelectronics have motivated research on manipulating light in on-chip integrated circuits. Hybrid plasmonic waveguides with low-dimensional semiconductors, including quantum dots and quantum wells, are a promising platform for realizing sub-diffraction limited optical components. Meanwhile, two-dimensional transition metal dichalcogenides (TMDs) have received broad interest in optoelectronics owing to tightly bound excitons at room temperature, strong light-matter and exciton-plasmon interactions, available top-down wafer-scale integration, and band-gap tunability. Here, we demonstrate principal functionalities for on-chip optical communications via reconfigurable exciton-plasmon interconversions in ∼200-nm-diameter Ag-nanowires overlapping onto TMD transistors. By varying device configurations for each operation purpose, three active components for optical communications are realized: field-effect exciton transistors with a channel length of ∼32 μm, field-effect exciton multiplexers transmitting multiple signals through a single NW and electrical detectors of propagating plasmons with a high On/Off ratio of∼190. Our results illustrate the unique merits of two-dimensional semiconductors for constructing reconfigurable device architectures in integrated nanophotonic circuits. PMID:27892463
State and group dynamics of world stock market by principal component analysis
NASA Astrophysics Data System (ADS)
Nobi, Ashadun; Lee, Jae Woo
2016-05-01
We study the dynamic interactions and structural changes by a principal component analysis (PCA) to cross-correlation coefficients of global financial indices in the years 1998-2012. The variances explained by the first PC increase with time and show a drastic change during the crisis. A sharp change in PC coefficient implies a transition of market state, a situation which occurs frequently in the American and Asian indices. However, the European indices remain stable over time. Using the first two PC coefficients, we identify indices that are similar and more strongly correlated than the others. We observe that the European indices form a robust group over the observation period. The dynamics of the individual indices within the group increase in similarity with time, and the dynamics of indices are more similar during the crises. Furthermore, the group formation of indices changes position in two-dimensional spaces due to crises. Finally, after a financial crisis, the difference of PCs between the European and American indices narrows.
Hoelzle, James B; Nelson, Nathaniel W; Smith, Clifford A
2011-03-01
Dimensional structures underlying the Wechsler Memory Scale-Fourth Edition (WMS-IV) and Wechsler Memory Scale-Third Edition (WMS-III) were compared to determine whether the revised measure has a more coherent and clinically relevant factor structure. Principal component analyses were conducted in normative samples reported in the respective technical manuals. Empirically supported procedures guided retention of dimensions. An invariant two-dimensional WMS-IV structure reflecting constructs of auditory learning/memory and visual attention/memory (C1 = .97; C2 = .96) is more theoretically coherent than the replicable, heterogeneous WMS-III dimension (C1 = .97). This research suggests that the WMS-IV may have greater utility in identifying lateralized memory dysfunction.
Peleato, Nicolás M; Andrews, Robert C
2015-01-01
This work investigated the application of several fluorescence excitation-emission matrix analysis methods as natural organic matter (NOM) indicators for use in predicting the formation of trihalomethanes (THMs) and haloacetic acids (HAAs). Waters from four different sources (two rivers and two lakes) were subjected to jar testing followed by 24hr disinfection by-product formation tests using chlorine. NOM was quantified using three common measures: dissolved organic carbon, ultraviolet absorbance at 254 nm, and specific ultraviolet absorbance as well as by principal component analysis, peak picking, and parallel factor analysis of fluorescence spectra. Based on multi-linear modeling of THMs and HAAs, principle component (PC) scores resulted in the lowest mean squared prediction error of cross-folded test sets (THMs: 43.7 (μg/L)(2), HAAs: 233.3 (μg/L)(2)). Inclusion of principle components representative of protein-like material significantly decreased prediction error for both THMs and HAAs. Parallel factor analysis did not identify a protein-like component and resulted in prediction errors similar to traditional NOM surrogates as well as fluorescence peak picking. These results support the value of fluorescence excitation-emission matrix-principal component analysis as a suitable NOM indicator in predicting the formation of THMs and HAAs for the water sources studied. Copyright © 2014. Published by Elsevier B.V.
Azevedo, Mônia Stremel; Valentim-Neto, Pedro Alexandre; Seraglio, Siluana Katia Tischer; da Luz, Cynthia Fernandes Pinto; Arisi, Ana Carolina Maisonnave; Costa, Ana Carolina Oliveira
2017-10-01
Due to the increasing valuation and appreciation of honeydew honey in many European countries and also to existing contamination among different types of honeys, authentication is an important aspect of quality control with regard to guaranteeing the origin in terms of source (honeydew or floral) and needs to be determined. Furthermore, proteins are minor components of the honey, despite the importance of their physiological effects, and can differ according to the source of the honey. In this context, the aims of this study were to carry out protein extraction from honeydew and floral honeys and to discriminate these honeys from the same botanical species, Mimosa scabrella Bentham, through proteome comparison using two-dimensional gel electrophoresis and principal component analysis. The results showed that the proteome profile and principal component analysis can be a useful tool for discrimination between these types of honey using matched proteins (45 matched spots). Also, the proteome profile showed 160 protein spots in honeydew honey and 84 spots in the floral honey. The protein profile can be a differential characteristic of this type of honey, in view of the importance of proteins as bioactive compounds in honey. © 2017 Society of Chemical Industry. © 2017 Society of Chemical Industry.
Effect of noise in principal component analysis with an application to ozone pollution
NASA Astrophysics Data System (ADS)
Tsakiri, Katerina G.
This thesis analyzes the effect of independent noise in principal components of k normally distributed random variables defined by a covariance matrix. We prove that the principal components as well as the canonical variate pairs determined from joint distribution of original sample affected by noise can be essentially different in comparison with those determined from the original sample. However when the differences between the eigenvalues of the original covariance matrix are sufficiently large compared to the level of the noise, the effect of noise in principal components and canonical variate pairs proved to be negligible. The theoretical results are supported by simulation study and examples. Moreover, we compare our results about the eigenvalues and eigenvectors in the two dimensional case with other models examined before. This theory can be applied in any field for the decomposition of the components in multivariate analysis. One application is the detection and prediction of the main atmospheric factor of ozone concentrations on the example of Albany, New York. Using daily ozone, solar radiation, temperature, wind speed and precipitation data, we determine the main atmospheric factor for the explanation and prediction of ozone concentrations. A methodology is described for the decomposition of the time series of ozone and other atmospheric variables into the global term component which describes the long term trend and the seasonal variations, and the synoptic scale component which describes the short term variations. By using the Canonical Correlation Analysis, we show that solar radiation is the only main factor between the atmospheric variables considered here for the explanation and prediction of the global and synoptic scale component of ozone. The global term components are modeled by a linear regression model, while the synoptic scale components by a vector autoregressive model and the Kalman filter. The coefficient of determination, R2, for the prediction of the synoptic scale ozone component was found to be the highest when we consider the synoptic scale component of the time series for solar radiation and temperature. KEY WORDS: multivariate analysis; principal component; canonical variate pairs; eigenvalue; eigenvector; ozone; solar radiation; spectral decomposition; Kalman filter; time series prediction
NASA Astrophysics Data System (ADS)
Farsadnia, Farhad; Ghahreman, Bijan
2016-04-01
Hydrologic homogeneous group identification is considered both fundamental and applied research in hydrology. Clustering methods are among conventional methods to assess the hydrological homogeneous regions. Recently, Self-Organizing feature Map (SOM) method has been applied in some studies. However, the main problem of this method is the interpretation on the output map of this approach. Therefore, SOM is used as input to other clustering algorithms. The aim of this study is to apply a two-level Self-Organizing feature map and Ward hierarchical clustering method to determine the hydrologic homogenous regions in North and Razavi Khorasan provinces. At first by principal component analysis, we reduced SOM input matrix dimension, then the SOM was used to form a two-dimensional features map. To determine homogeneous regions for flood frequency analysis, SOM output nodes were used as input into the Ward method. Generally, the regions identified by the clustering algorithms are not statistically homogeneous. Consequently, they have to be adjusted to improve their homogeneity. After adjustment of the homogeneity regions by L-moment tests, five hydrologic homogeneous regions were identified. Finally, adjusted regions were created by a two-level SOM and then the best regional distribution function and associated parameters were selected by the L-moment approach. The results showed that the combination of self-organizing maps and Ward hierarchical clustering by principal components as input is more effective than the hierarchical method, by principal components or standardized inputs to achieve hydrologic homogeneous regions.
Jesse, Stephen; Kalinin, Sergei V
2009-02-25
An approach for the analysis of multi-dimensional, spectroscopic-imaging data based on principal component analysis (PCA) is explored. PCA selects and ranks relevant response components based on variance within the data. It is shown that for examples with small relative variations between spectra, the first few PCA components closely coincide with results obtained using model fitting, and this is achieved at rates approximately four orders of magnitude faster. For cases with strong response variations, PCA allows an effective approach to rapidly process, de-noise, and compress data. The prospects for PCA combined with correlation function analysis of component maps as a universal tool for data analysis and representation in microscopy are discussed.
NASA Technical Reports Server (NTRS)
Li, Z. K.
1985-01-01
A specialized program was developed for flow cytometric list-mode data using an heirarchical tree method for identifying and enumerating individual subpopulations, the method of principal components for a two-dimensional display of 6-parameter data array, and a standard sorting algorithm for characterizing subpopulations. The program was tested against a published data set subjected to cluster analysis and experimental data sets from controlled flow cytometry experiments using a Coulter Electronics EPICS V Cell Sorter. A version of the program in compiled BASIC is usable on a 16-bit microcomputer with the MS-DOS operating system. It is specialized for 6 parameters and up to 20,000 cells. Its two-dimensional display of Euclidean distances reveals clusters clearly, as does its 1-dimensional display. The identified subpopulations can, in suitable experiments, be related to functional subpopulations of cells.
Optical Aptasensors for Adenosine Triphosphate
Ng, Stella; Lim, Hui Si; Ma, Qian; Gao, Zhiqiang
2016-01-01
Nucleic acids are among the most researched and applied biomolecules. Their diverse two- and three-dimensional structures in conjunction with their robust chemistry and ease of manipulation provide a rare opportunity for sensor applications. Moreover, their high biocompatibility has seen them being used in the construction of in vivo assays. Various nucleic acid-based devices have been extensively studied as either the principal element in discrete molecule-like sensors or as the main component in the fabrication of sensing devices. The use of aptamers in sensors - aptasensors, in particular, has led to improvements in sensitivity, selectivity, and multiplexing capacity for a wide verity of analytes like proteins, nucleic acids, as well as small biomolecules such as glucose and adenosine triphosphate (ATP). This article reviews the progress in the use of aptamers as the principal component in sensors for optical detection of ATP with an emphasis on sensing mechanism, performance, and applications with some discussion on challenges and perspectives. PMID:27446501
Optical Aptasensors for Adenosine Triphosphate.
Ng, Stella; Lim, Hui Si; Ma, Qian; Gao, Zhiqiang
2016-01-01
Nucleic acids are among the most researched and applied biomolecules. Their diverse two- and three-dimensional structures in conjunction with their robust chemistry and ease of manipulation provide a rare opportunity for sensor applications. Moreover, their high biocompatibility has seen them being used in the construction of in vivo assays. Various nucleic acid-based devices have been extensively studied as either the principal element in discrete molecule-like sensors or as the main component in the fabrication of sensing devices. The use of aptamers in sensors - aptasensors, in particular, has led to improvements in sensitivity, selectivity, and multiplexing capacity for a wide verity of analytes like proteins, nucleic acids, as well as small biomolecules such as glucose and adenosine triphosphate (ATP). This article reviews the progress in the use of aptamers as the principal component in sensors for optical detection of ATP with an emphasis on sensing mechanism, performance, and applications with some discussion on challenges and perspectives.
Multivariate analysis for scanning tunneling spectroscopy data
NASA Astrophysics Data System (ADS)
Yamanishi, Junsuke; Iwase, Shigeru; Ishida, Nobuyuki; Fujita, Daisuke
2018-01-01
We applied principal component analysis (PCA) to two-dimensional tunneling spectroscopy (2DTS) data obtained on a Si(111)-(7 × 7) surface to explore the effectiveness of multivariate analysis for interpreting 2DTS data. We demonstrated that several components that originated mainly from specific atoms at the Si(111)-(7 × 7) surface can be extracted by PCA. Furthermore, we showed that hidden components in the tunneling spectra can be decomposed (peak separation), which is difficult to achieve with normal 2DTS analysis without the support of theoretical calculations. Our analysis showed that multivariate analysis can be an additional powerful way to analyze 2DTS data and extract hidden information from a large amount of spectroscopic data.
Bayesian estimation of Karhunen–Loève expansions; A random subspace approach
DOE Office of Scientific and Technical Information (OSTI.GOV)
Chowdhary, Kenny; Najm, Habib N.
One of the most widely-used statistical procedures for dimensionality reduction of high dimensional random fields is Principal Component Analysis (PCA), which is based on the Karhunen-Lo eve expansion (KLE) of a stochastic process with finite variance. The KLE is analogous to a Fourier series expansion for a random process, where the goal is to find an orthogonal transformation for the data such that the projection of the data onto this orthogonal subspace is optimal in the L 2 sense, i.e, which minimizes the mean square error. In practice, this orthogonal transformation is determined by performing an SVD (Singular Value Decomposition)more » on the sample covariance matrix or on the data matrix itself. Sampling error is typically ignored when quantifying the principal components, or, equivalently, basis functions of the KLE. Furthermore, it is exacerbated when the sample size is much smaller than the dimension of the random field. In this paper, we introduce a Bayesian KLE procedure, allowing one to obtain a probabilistic model on the principal components, which can account for inaccuracies due to limited sample size. The probabilistic model is built via Bayesian inference, from which the posterior becomes the matrix Bingham density over the space of orthonormal matrices. We use a modified Gibbs sampling procedure to sample on this space and then build a probabilistic Karhunen-Lo eve expansions over random subspaces to obtain a set of low-dimensional surrogates of the stochastic process. We illustrate this probabilistic procedure with a finite dimensional stochastic process inspired by Brownian motion.« less
Bayesian estimation of Karhunen–Loève expansions; A random subspace approach
Chowdhary, Kenny; Najm, Habib N.
2016-04-13
One of the most widely-used statistical procedures for dimensionality reduction of high dimensional random fields is Principal Component Analysis (PCA), which is based on the Karhunen-Lo eve expansion (KLE) of a stochastic process with finite variance. The KLE is analogous to a Fourier series expansion for a random process, where the goal is to find an orthogonal transformation for the data such that the projection of the data onto this orthogonal subspace is optimal in the L 2 sense, i.e, which minimizes the mean square error. In practice, this orthogonal transformation is determined by performing an SVD (Singular Value Decomposition)more » on the sample covariance matrix or on the data matrix itself. Sampling error is typically ignored when quantifying the principal components, or, equivalently, basis functions of the KLE. Furthermore, it is exacerbated when the sample size is much smaller than the dimension of the random field. In this paper, we introduce a Bayesian KLE procedure, allowing one to obtain a probabilistic model on the principal components, which can account for inaccuracies due to limited sample size. The probabilistic model is built via Bayesian inference, from which the posterior becomes the matrix Bingham density over the space of orthonormal matrices. We use a modified Gibbs sampling procedure to sample on this space and then build a probabilistic Karhunen-Lo eve expansions over random subspaces to obtain a set of low-dimensional surrogates of the stochastic process. We illustrate this probabilistic procedure with a finite dimensional stochastic process inspired by Brownian motion.« less
NASA Astrophysics Data System (ADS)
Lin, Pei-Chun; Yu, Chun-Chang; Chen, Charlie Chung-Ping
2015-01-01
As one of the critical stages of a very large scale integration fabrication process, postexposure bake (PEB) plays a crucial role in determining the final three-dimensional (3-D) profiles and lessening the standing wave effects. However, the full 3-D chemically amplified resist simulation is not widely adopted during the postlayout optimization due to the long run-time and huge memory usage. An efficient simulation method is proposed to simulate the PEB while considering standing wave effects and resolution enhancement techniques, such as source mask optimization and subresolution assist features based on the Sylvester equation and Abbe-principal component analysis method. Simulation results show that our algorithm is 20× faster than the conventional Gaussian convolution method.
Predicting Visual Disability in Glaucoma With Combinations of Vision Measures.
Lin, Stephanie; Mihailovic, Aleksandra; West, Sheila K; Johnson, Chris A; Friedman, David S; Kong, Xiangrong; Ramulu, Pradeep Y
2018-04-01
We characterized vision in glaucoma using seven visual measures, with the goals of determining the dimensionality of vision, and how many and which visual measures best model activity limitation. We analyzed cross-sectional data from 150 older adults with glaucoma, collecting seven visual measures: integrated visual field (VF) sensitivity, visual acuity, contrast sensitivity (CS), area under the log CS function, color vision, stereoacuity, and visual acuity with noise. Principal component analysis was used to examine the dimensionality of vision. Multivariable regression models using one, two, or three vision tests (and nonvisual predictors) were compared to determine which was best associated with Rasch-analyzed Glaucoma Quality of Life-15 (GQL-15) person measure scores. The participants had a mean age of 70.2 and IVF sensitivity of 26.6 dB, suggesting mild-to-moderate glaucoma. All seven vision measures loaded similarly onto the first principal component (eigenvectors, 0.220-0.442), which explained 56.9% of the variance in vision scores. In models for GQL scores, the maximum adjusted- R 2 values obtained were 0.263, 0.296, and 0.301 when using one, two, and three vision tests in the models, respectively, though several models in each category had similar adjusted- R 2 values. All three of the best-performing models contained CS. Vision in glaucoma is a multidimensional construct that can be described by several variably-correlated vision measures. Measuring more than two vision tests does not substantially improve models for activity limitation. A sufficient description of disability in glaucoma can be obtained using one to two vision tests, especially VF and CS.
Towards reconstruction of overlapping fingerprints using plasma spectroscopy
NASA Astrophysics Data System (ADS)
Yang, Jun-Ho; Choi, Soo-Jin; Yoh, Jack J.
2017-08-01
Chemical analysis is commonly used in the field of forensic science where the precise discrimination of primary evidence is of significant importance. Laser-Induced Breakdown Spectroscopy (LIBS) exceeds other spectroscopic methods in terms of the time required for pre- and post-sample preparation, the insensitivity to sample phase state be it solid, liquid, or gas, and the detection of two-dimensional spectral mapping from real time point measurements. In this research, fingerprint samples on various surface materials are considered in the chemical detection and reconstruction of fingerprints using the two-dimensional LIBS technique. Strong and distinct intensities of specific wavelengths represent visible ink, natural secretion of sweat, and contaminants from the environment, all of which can be present in latent fingerprints. The particular aim of the work presented here is to enhance the precision of the two-dimensional recreation of the fingerprints present on metal, plastic, and artificially prepared soil surface using LIBS with principal component analysis. By applying a distinct wavelength discrimination for two overlapping fingerprint samples, separation into two non-identical chemical fingerprints was successfully performed.
NASA Technical Reports Server (NTRS)
Revilock, Duane M., Jr.; Thesken, John C.; Schmidt, Timothy E.
2007-01-01
Ambient temperature hydrostatic pressurization tests were conducted on a composite overwrapped pressure vessel (COPV) to understand the fiber stresses in COPV components. Two three-dimensional digital image correlation systems with high speed cameras were used in the evaluation to provide full field displacement and strain data for each pressurization test. A few of the key findings will be discussed including how the principal strains provided better insight into system behavior than traditional gauges, a high localized strain that was measured where gages were not present and the challenges of measuring curved surfaces with the use of a 1.25 in. thick layered polycarbonate panel that protected the cameras.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Hua, Xin; Marshall, Matthew J.; Xiong, Yijia
2015-05-01
A vacuum compatible microfluidic reactor, SALVI (System for Analysis at the Liquid Vacuum Interface) was employed for in situ chemical imaging of live biofilms using time-of-flight secondary ion mass spectrometry (ToF-SIMS). Depth profiling by sputtering materials in sequential layers resulted in live biofilm spatial chemical mapping. 2D images were reconstructed to report the first 3D images of hydrated biofilm elucidating spatial and chemical heterogeneity. 2D image principal component analysis (PCA) was conducted among biofilms at different locations in the microchannel. Our approach directly visualized spatial and chemical heterogeneity within the living biofilm by dynamic liquid ToF-SIMS.
Grassi, Mario; Nucera, Andrea
2010-01-01
The objective of this study was twofold: 1) to confirm the hypothetical eight scales and two-component summaries of the questionnaire Short Form 36 Health Survey (SF-36), and 2) to evaluate the performance of two alternative measures to the original physical component summary (PCS) and mental component summary (MCS). We performed principal component analysis (PCA) based on 35 items, after optimal scaling via multiple correspondence analysis (MCA), and subsequently on eight scales, after standard summative scoring. Item-based summary measures were planned. Data from the European Community Respiratory Health Survey II follow-up of 8854 subjects from 25 centers were analyzed to cross-validate the original and the novel PCS and MCS. Overall, the scale- and item-based comparison indicated that the SF-36 scales and summaries meet the supposed dimensionality. However, vitality, social functioning, and general health items did not fit data optimally. The novel measures, derived a posteriori by unit-rule from an oblique (correlated) MCA/PCA solution, are simple item sums or weighted scale sums where the weights are the raw scale ranges. These item-based scores yielded consistent scale-summary results for outliers profiles, with an expected known-group differences validity. We were able to confirm the hypothesized dimensionality of eight scales and two summaries of the SF-36. The alternative scoring reaches at least the same required standards of the original scoring. In addition, it can reduce the item-scale inconsistencies without loss of predictive validity.
Candidate gene analyses of 3-dimensional dentoalveolar phenotypes in subjects with malocclusion
Weaver, Cole A.; Miller, Steven F.; da Fontoura, Clarissa S. G.; Wehby, George L.; Amendt, Brad A.; Holton, Nathan E.; Allareddy, Veeratrishul; Southard, Thomas E.; Moreno Uribe, Lina M.
2017-01-01
Introduction Genetic studies of malocclusion etiology have identified 4 deleterious mutations in genes, DUSP6, ARHGAP21, FGF23, and ADAMTS1 in familial Class III cases. Although these variants may have large impacts on Class III phenotypic expression, their low frequency (<1%) makes them unlikely to explain most malocclusions. Thus, much of the genetic variation underlying the dentofacial phenotypic variation associated with malocclusion remains unknown. In this study, we evaluated associations between common genetic variations in craniofacial candidate genes and 3-dimensional dentoalveolar phenotypes in patients with malocclusion. Methods Pretreatment dental casts or cone-beam computed tomographic images from 300 healthy subjects were digitized with 48 landmarks. The 3-dimensional coordinate data were submitted to a geometric morphometric approach along with principal component analysis to generate continuous phenotypes including symmetric and asymmetric components of dentoalveolar shape variation, fluctuating asymmetry, and size. The subjects were genotyped for 222 single-nucleotide polymorphisms in 82 genes/loci, and phenotpye-genotype associations were tested via multivariate linear regression. Results Principal component analysis of symmetric variation identified 4 components that explained 68% of the total variance and depicted anteroposterior, vertical, and transverse dentoalveolar discrepancies. Suggestive associations (P < 0.05) were identified with PITX2, SNAI3, 11q22.2-q22.3, 4p16.1, ISL1, and FGF8. Principal component analysis for asymmetric variations identified 4 components that explained 51% of the total variations and captured left-to-right discrepancies resulting in midline deviations, unilateral crossbites, and ectopic eruptions. Suggestive associations were found with TBX1 AJUBA, SNAI3 SATB2, TP63, and 1p22.1. Fluctuating asymmetry was associated with BMP3 and LATS1. Associations for SATB2 and BMP3 with asymmetric variations remained significant after the Bonferroni correction (P <0.00022). Suggestive associations were found for centroid size, a proxy for dentoalveolar size variation with 4p16.1 and SNAI1. Conclusions Specific genetic pathways associated with 3-dimensional dentoalveolar phenotypic variation in malocclusions were identified. PMID:28257739
How multi segmental patterns deviate in spastic diplegia from typical developed.
Zago, Matteo; Sforza, Chiarella; Bona, Alessia; Cimolin, Veronica; Costici, Pier Francesco; Condoluci, Claudia; Galli, Manuela
2017-10-01
The relationship between gait features and coordination in children with Cerebral Palsy is not sufficiently analyzed yet. Principal Component Analysis can help in understanding motion patterns decomposing movement into its fundamental components (Principal Movements). This study aims at quantitatively characterizing the functional connections between multi-joint gait patterns in Cerebral Palsy. 65 children with spastic diplegia aged 10.6 (SD 3.7) years participated in standardized gait analysis trials; 31 typically developing adolescents aged 13.6 (4.4) years were also tested. To determine if posture affects gait patterns, patients were split into Crouch and knee Hyperextension group according to knee flexion angle at standing. 3D coordinates of hips, knees, ankles, metatarsal joints, pelvis and shoulders were submitted to Principal Component Analysis. Four Principal Movements accounted for 99% of global variance; components 1-3 explained major sagittal patterns, components 4-5 referred to movements on frontal plane and component 6 to additional movement refinements. Dimensionality was higher in patients than in controls (p<0.01), and the Crouch group significantly differed from controls in the application of components 1 and 4-6 (p<0.05), while the knee Hyperextension group in components 1-2 and 5 (p<0.05). Compensatory strategies of children with Cerebral Palsy (interactions between main and secondary movement patterns), were objectively determined. Principal Movements can reduce the effort in interpreting gait reports, providing an immediate and quantitative picture of the connections between movement components. Copyright © 2017 Elsevier Ltd. All rights reserved.
Cocco, Simona; Monasson, Remi; Weigt, Martin
2013-01-01
Various approaches have explored the covariation of residues in multiple-sequence alignments of homologous proteins to extract functional and structural information. Among those are principal component analysis (PCA), which identifies the most correlated groups of residues, and direct coupling analysis (DCA), a global inference method based on the maximum entropy principle, which aims at predicting residue-residue contacts. In this paper, inspired by the statistical physics of disordered systems, we introduce the Hopfield-Potts model to naturally interpolate between these two approaches. The Hopfield-Potts model allows us to identify relevant ‘patterns’ of residues from the knowledge of the eigenmodes and eigenvalues of the residue-residue correlation matrix. We show how the computation of such statistical patterns makes it possible to accurately predict residue-residue contacts with a much smaller number of parameters than DCA. This dimensional reduction allows us to avoid overfitting and to extract contact information from multiple-sequence alignments of reduced size. In addition, we show that low-eigenvalue correlation modes, discarded by PCA, are important to recover structural information: the corresponding patterns are highly localized, that is, they are concentrated in few sites, which we find to be in close contact in the three-dimensional protein fold. PMID:23990764
Analysing Normal and Partial Glossectomee Tongues Using Ultrasound
ERIC Educational Resources Information Center
Bressmann, Tim; Uy, Catherine; Irish, Jonathan C.
2005-01-01
The present study aimed at identifying underlying parameters that govern the shape of the tongue. A functional topography of the tongue surface was developed based on three-dimensional ultrasound scans of sustained speech sounds in ten normal subjects. A principal component analysis extracted three components that explained 89.2% of the variance…
Missing data is a common problem in the application of statistical techniques. In principal component analysis (PCA), a technique for dimensionality reduction, incomplete data points are either discarded or imputed using interpolation methods. Such approaches are less valid when ...
NASA Astrophysics Data System (ADS)
Ueki, Kenta; Iwamori, Hikaru
2017-10-01
In this study, with a view of understanding the structure of high-dimensional geochemical data and discussing the chemical processes at work in the evolution of arc magmas, we employed principal component analysis (PCA) to evaluate the compositional variations of volcanic rocks from the Sengan volcanic cluster of the Northeastern Japan Arc. We analyzed the trace element compositions of various arc volcanic rocks, sampled from 17 different volcanoes in a volcanic cluster. The PCA results demonstrated that the first three principal components accounted for 86% of the geochemical variation in the magma of the Sengan region. Based on the relationships between the principal components and the major elements, the mass-balance relationships with respect to the contributions of minerals, the composition of plagioclase phenocrysts, geothermal gradient, and seismic velocity structure in the crust, the first, the second, and the third principal components appear to represent magma mixing, crystallizations of olivine/pyroxene, and crystallizations of plagioclase, respectively. These represented 59%, 20%, and 6%, respectively, of the variance in the entire compositional range, indicating that magma mixing accounted for the largest variance in the geochemical variation of the arc magma. Our result indicated that crustal processes dominate the geochemical variation of magma in the Sengan volcanic cluster.
Chen, Ting; Zhang, Miao; Jabbour, Salma; Wang, Hesheng; Barbee, David; Das, Indra J; Yue, Ning
2018-04-10
Through-plane motion introduces uncertainty in three-dimensional (3D) motion monitoring when using single-slice on-board imaging (OBI) modalities such as cine MRI. We propose a principal component analysis (PCA)-based framework to determine the optimal imaging plane to minimize the through-plane motion for single-slice imaging-based motion monitoring. Four-dimensional computed tomography (4DCT) images of eight thoracic cancer patients were retrospectively analyzed. The target volumes were manually delineated at different respiratory phases of 4DCT. We performed automated image registration to establish the 4D respiratory target motion trajectories for all patients. PCA was conducted using the motion information to define the three principal components of the respiratory motion trajectories. Two imaging planes were determined perpendicular to the second and third principal component, respectively, to avoid imaging with the primary principal component of the through-plane motion. Single-slice images were reconstructed from 4DCT in the PCA-derived orthogonal imaging planes and were compared against the traditional AP/Lateral image pairs on through-plane motion, residual error in motion monitoring, absolute motion amplitude error and the similarity between target segmentations at different phases. We evaluated the significance of the proposed motion monitoring improvement using paired t test analysis. The PCA-determined imaging planes had overall less through-plane motion compared against the AP/Lateral image pairs. For all patients, the average through-plane motion was 3.6 mm (range: 1.6-5.6 mm) for the AP view and 1.7 mm (range: 0.6-2.7 mm) for the Lateral view. With PCA optimization, the average through-plane motion was 2.5 mm (range: 1.3-3.9 mm) and 0.6 mm (range: 0.2-1.5 mm) for the two imaging planes, respectively. The absolute residual error of the reconstructed max-exhale-to-inhale motion averaged 0.7 mm (range: 0.4-1.3 mm, 95% CI: 0.4-1.1 mm) using optimized imaging planes, averaged 0.5 mm (range: 0.3-1.0 mm, 95% CI: 0.2-0.8 mm) using an imaging plane perpendicular to the minimal motion component only and averaged 1.3 mm (range: 0.4-2.8 mm, 95% CI: 0.4-2.3 mm) in AP/Lateral orthogonal image pairs. The root-mean-square error of reconstructed displacement was 0.8 mm for optimized imaging planes, 0.6 mm for imaging plane perpendicular to the minimal motion component only, and 1.6 mm for AP/Lateral orthogonal image pairs. When using the optimized imaging planes for motion monitoring, there was no significant absolute amplitude error of the reconstructed motion (P = 0.0988), while AP/Lateral images had significant error (P = 0.0097) with a paired t test. The average surface distance (ASD) between overlaid two-dimensional (2D) tumor segmentation at end-of-inhale and end-of-exhale for all eight patients was 0.6 ± 0.2 mm in optimized imaging planes and 1.4 ± 0.8 mm in AP/Lateral images. The Dice similarity coefficient (DSC) between overlaid 2D tumor segmentation at end-of-inhale and end-of-exhale for all eight patients was 0.96 ± 0.03 in optimized imaging planes and 0.89 ± 0.05 in AP/Lateral images. Both ASD (P = 0.034) and DSC (P = 0.022) were significantly improved in the optimized imaging planes. Motion monitoring using imaging planes determined by the proposed PCA-based framework had significantly improved performance. Single-slice image-based motion tracking can be used for clinical implementations such as MR image-guided radiation therapy (MR-IGRT). © 2018 American Association of Physicists in Medicine.
Transforming Graph Data for Statistical Relational Learning
2012-10-01
Jordan, 2003), PLSA (Hofmann, 1999), ? Classification via RMN (Taskar et al., 2003) or SVM (Hasan, Chaoji, Salem , & Zaki, 2006) ? Hierarchical...dimensionality reduction methods such as Principal 407 Rossi, McDowell, Aha, & Neville Component Analysis (PCA), Principal Factor Analysis ( PFA ), and...clustering algorithm. Journal of the Royal Statistical Society. Series C, Applied statistics, 28, 100–108. Hasan, M. A., Chaoji, V., Salem , S., & Zaki, M
Reduced order surrogate modelling (ROSM) of high dimensional deterministic simulations
NASA Astrophysics Data System (ADS)
Mitry, Mina
Often, computationally expensive engineering simulations can prohibit the engineering design process. As a result, designers may turn to a less computationally demanding approximate, or surrogate, model to facilitate their design process. However, owing to the the curse of dimensionality, classical surrogate models become too computationally expensive for high dimensional data. To address this limitation of classical methods, we develop linear and non-linear Reduced Order Surrogate Modelling (ROSM) techniques. Two algorithms are presented, which are based on a combination of linear/kernel principal component analysis and radial basis functions. These algorithms are applied to subsonic and transonic aerodynamic data, as well as a model for a chemical spill in a channel. The results of this thesis show that ROSM can provide a significant computational benefit over classical surrogate modelling, sometimes at the expense of a minor loss in accuracy.
The assessment of facial variation in 4747 British school children.
Toma, Arshed M; Zhurov, Alexei I; Playle, Rebecca; Marshall, David; Rosin, Paul L; Richmond, Stephen
2012-12-01
The aim of this study is to identify key components contributing to facial variation in a large population-based sample of 15.5-year-old children (2514 females and 2233 males). The subjects were recruited from the Avon Longitudinal Study of Parents and Children. Three-dimensional facial images were obtained for each subject using two high-resolution Konica Minolta laser scanners. Twenty-one reproducible facial landmarks were identified and their coordinates were recorded. The facial images were registered using Procrustes analysis. Principal component analysis was then employed to identify independent groups of correlated coordinates. For the total data set, 14 principal components (PCs) were identified which explained 82 per cent of the total variance, with the first three components accounting for 46 per cent of the variance. Similar results were obtained for males and females separately with only subtle gender differences in some PCs. Facial features may be treated as a multidimensional statistical continuum with respect to the PCs. The first three PCs characterize the face in terms of height, width, and prominence of the nose. The derived PCs may be useful to identify and classify faces according to a scale of normality.
Principal components of wrist circumduction from electromagnetic surgical tracking.
Rasquinha, Brian J; Rainbow, Michael J; Zec, Michelle L; Pichora, David R; Ellis, Randy E
2017-02-01
An electromagnetic (EM) surgical tracking system was used for a functionally calibrated kinematic analysis of wrist motion. Circumduction motions were tested for differences in subject gender and for differences in the sense of the circumduction as clockwise or counter-clockwise motion. Twenty subjects were instrumented for EM tracking. Flexion-extension motion was used to identify the functional axis. Subjects performed unconstrained wrist circumduction in a clockwise and counter-clockwise sense. Data were decomposed into orthogonal flexion-extension motions and radial-ulnar deviation motions. PCA was used to concisely represent motions. Nonparametric Wilcoxon tests were used to distinguish the groups. Flexion-extension motions were projected onto a direction axis with a root-mean-square error of [Formula: see text]. Using the first three principal components, there was no statistically significant difference in gender (all [Formula: see text]). For motion sense, radial-ulnar deviation distinguished the sense of circumduction in the first principal component ([Formula: see text]) and in the third principal component ([Formula: see text]); flexion-extension distinguished the sense in the second principal component ([Formula: see text]). The clockwise sense of circumduction could be distinguished by a multifactorial combination of components; there were no gender differences in this small population. These data constitute a baseline for normal wrist circumduction. The multifactorial PCA findings suggest that a higher-dimensional method, such as manifold analysis, may be a more concise way of representing circumduction in human joints.
Hybrid Feature Extraction-based Approach for Facial Parts Representation and Recognition
NASA Astrophysics Data System (ADS)
Rouabhia, C.; Tebbikh, H.
2008-06-01
Face recognition is a specialized image processing which has attracted a considerable attention in computer vision. In this article, we develop a new facial recognition system from video sequences images dedicated to person identification whose face is partly occulted. This system is based on a hybrid image feature extraction technique called ACPDL2D (Rouabhia et al. 2007), it combines two-dimensional principal component analysis and two-dimensional linear discriminant analysis with neural network. We performed the feature extraction task on the eyes and the nose images separately then a Multi-Layers Perceptron classifier is used. Compared to the whole face, the results of simulation are in favor of the facial parts in terms of memory capacity and recognition (99.41% for the eyes part, 98.16% for the nose part and 97.25 % for the whole face).
Dimensionality reduction for the quantitative evaluation of a smartphone-based Timed Up and Go test.
Palmerini, Luca; Mellone, Sabato; Rocchi, Laura; Chiari, Lorenzo
2011-01-01
The Timed Up and Go is a clinical test to assess mobility in the elderly and in Parkinson's disease. Lately instrumented versions of the test are being considered, where inertial sensors assess motion. To improve the pervasiveness, ease of use, and cost, we consider a smartphone's accelerometer as the measurement system. Several parameters (usually highly correlated) can be computed from the signals recorded during the test. To avoid redundancy and obtain the features that are most sensitive to the locomotor performance, a dimensionality reduction was performed through principal component analysis (PCA). Forty-nine healthy subjects of different ages were tested. PCA was performed to extract new features (principal components) which are not redundant combinations of the original parameters and account for most of the data variability. They can be useful for exploratory analysis and outlier detection. Then, a reduced set of the original parameters was selected through correlation analysis with the principal components. This set could be recommended for studies based on healthy adults. The proposed procedure could be used as a first-level feature selection in classification studies (i.e. healthy-Parkinson's disease, fallers-non fallers) and could allow, in the future, a complete system for movement analysis to be incorporated in a smartphone.
NASA Astrophysics Data System (ADS)
Ma, Mengli; Lei, En; Meng, Hengling; Wang, Tiantao; Xie, Linyan; Shen, Dong; Xianwang, Zhou; Lu, Bingyue
2017-08-01
Amomum tsao-ko is a commercial plant that used for various purposes in medicinal and food industries. For the present investigation, 44 germplasm samples were collected from Jinping County of Yunnan Province. Clusters analysis and 2-dimensional principal component analysis (PCA) was used to represent the genetic relations among Amomum tsao-ko by using simple sequence repeat (SSR) markers. Clustering analysis clearly distinguished the samples groups. Two major clusters were formed; first (Cluster I) consisted of 34 individuals, the second (Cluster II) consisted of 10 individuals, Cluster I as the main group contained multiple sub-clusters. PCA also showed 2 groups: PCA Group 1 included 29 individuals, PCA Group 2 included 12 individuals, consistent with the results of cluster analysis. The purpose of the present investigation was to provide information on genetic relationship of Amomum tsao-ko germplasm resources in main producing areas, also provide a theoretical basis for the protection and utilization of Amomum tsao-ko resources.
Internal friction and mode relaxation in a simple chain model.
Fugmann, S; Sokolov, I M
2009-12-21
We consider the equilibrium relaxation properties of the end-to-end distance and of the principal components in a one-dimensional polymer chain model with nonlinear interaction between the beads. While for the single-well potentials these properties are similar to the ones of a Rouse chain, for the double-well interaction potentials, modeling internal friction, they differ vastly from the ones of the harmonic chain at intermediate times and intermediate temperatures. This minimal description within a one-dimensional model mimics the relaxation properties found in much more complex polymer systems. Thus, the relaxation time of the end-to-end distance may grow by orders of magnitude at intermediate temperatures. The principal components (whose directions are shown to coincide with the normal modes of the harmonic chain, whatever interaction potential is assumed) not only display larger relaxation times but also subdiffusive scaling.
Fractal-like hierarchical organization of bone begins at the nanoscale
NASA Astrophysics Data System (ADS)
Reznikov, Natalie; Bilton, Matthew; Lari, Leonardo; Stevens, Molly M.; Kröger, Roland
2018-05-01
The components of bone assemble hierarchically to provide stiffness and toughness. However, the organization and relationship between bone’s principal components—mineral and collagen—has not been clearly elucidated. Using three-dimensional electron tomography imaging and high-resolution two-dimensional electron microscopy, we demonstrate that bone mineral is hierarchically assembled beginning at the nanoscale: Needle-shaped mineral units merge laterally to form platelets, and these are further organized into stacks of roughly parallel platelets. These stacks coalesce into aggregates that exceed the lateral dimensions of the collagen fibrils and span adjacent fibrils as continuous, cross-fibrillar mineralization. On the basis of these observations, we present a structural model of hierarchy and continuity for the mineral phase, which contributes to the structural integrity of bone.
Welke, Juliane Elisa; Zanus, Mauro; Lazzarotto, Marcelo; Pulgati, Fernando Hepp; Zini, Cláudia Alcaraz
2014-12-01
The main changes in the volatile profile of base wines and their corresponding sparkling wines produced by traditional method were evaluated and investigated for the first time using headspace solid-phase microextraction combined with comprehensive two-dimensional gas chromatography with time-of-flight mass spectrometry detection (GC×GC/TOFMS) and chemometric tools. Fisher ratios helped to find the 119 analytes that were responsible for the main differences between base and sparkling wines and principal component analysis explained 93.1% of the total variance related to the selected 78 compounds. It was also possible to observe five subclusters in base wines and four subclusters in sparkling wines samples through hierarchical cluster analysis, which seemed to have an organised distribution according to the regions where the wines came from. Twenty of the most important volatile compounds co-eluted with other components and separation of some of them was possible due to GC×GC/TOFMS performance. Copyright © 2014. Published by Elsevier Ltd.
EM in high-dimensional spaces.
Draper, Bruce A; Elliott, Daniel L; Hayes, Jeremy; Baek, Kyungim
2005-06-01
This paper considers fitting a mixture of Gaussians model to high-dimensional data in scenarios where there are fewer data samples than feature dimensions. Issues that arise when using principal component analysis (PCA) to represent Gaussian distributions inside Expectation-Maximization (EM) are addressed, and a practical algorithm results. Unlike other algorithms that have been proposed, this algorithm does not try to compress the data to fit low-dimensional models. Instead, it models Gaussian distributions in the (N - 1)-dimensional space spanned by the N data samples. We are able to show that this algorithm converges on data sets where low-dimensional techniques do not.
Choi, Ji Yeh; Hwang, Heungsun; Yamamoto, Michio; Jung, Kwanghee; Woodward, Todd S
2017-06-01
Functional principal component analysis (FPCA) and functional multiple-set canonical correlation analysis (FMCCA) are data reduction techniques for functional data that are collected in the form of smooth curves or functions over a continuum such as time or space. In FPCA, low-dimensional components are extracted from a single functional dataset such that they explain the most variance of the dataset, whereas in FMCCA, low-dimensional components are obtained from each of multiple functional datasets in such a way that the associations among the components are maximized across the different sets. In this paper, we propose a unified approach to FPCA and FMCCA. The proposed approach subsumes both techniques as special cases. Furthermore, it permits a compromise between the techniques, such that components are obtained from each set of functional data to maximize their associations across different datasets, while accounting for the variance of the data well. We propose a single optimization criterion for the proposed approach, and develop an alternating regularized least squares algorithm to minimize the criterion in combination with basis function approximations to functions. We conduct a simulation study to investigate the performance of the proposed approach based on synthetic data. We also apply the approach for the analysis of multiple-subject functional magnetic resonance imaging data to obtain low-dimensional components of blood-oxygen level-dependent signal changes of the brain over time, which are highly correlated across the subjects as well as representative of the data. The extracted components are used to identify networks of neural activity that are commonly activated across the subjects while carrying out a working memory task.
Villanueva, Cristina M; Castaño-Vinyals, Gemma; Moreno, Víctor; Carrasco-Turigas, Glòria; Aragonés, Nuria; Boldo, Elena; Ardanaz, Eva; Toledo, Estefanía; Altzibar, Jone M; Zaldua, Itziar; Azpiroz, Lourdes; Goñi, Fernando; Tardón, Adonina; Molina, Antonio J; Martín, Vicente; López-Rojo, Concepción; Jiménez-Moleón, José J; Capelo, Rocío; Gómez-Acebo, Inés; Peiró, Rosana; Ripoll, Mónica; Gracia-Lavedan, Esther; Nieuwenhujsen, Mark J; Rantakokko, Panu; Goslan, Emma H; Pollán, Marina; Kogevinas, Manolis
2012-04-01
Although disinfection by-products (DBPs) occur in complex mixtures, studies evaluating health risks have been focused in few chemicals. In the framework of an epidemiological study on cancer in 11 Spanish provinces, we describe the concentration of four trihalomethanes (THMs), nine haloacetic acids (HAA), 3-chloro-4-(dichloromethyl)-5-hydroxy-2(5H)-furanone (MX), four haloacetonitries, two haloketones, chloropicrin and chloral hydrate and estimate correlations. A total of 233 tap water samples were collected in 2010. Principal component analyses were conducted to reduce dimensionality of DBPs. Overall median (range) level of THMs and HAAs was 26.4 (0.8-98.1) and 26.4 (0.9-86.9) μg/l, respectively (N=217). MX analysed in a subset (N=36) showed a median (range) concentration of 16.7 (0.8-54.1)ng/l. Haloacetonitries, haloketones, chloropicrin and chloral hydrate were analysed in a subset (N=16), showing levels from unquantifiable (<1 μg/l) to 5.5 μg/l (dibromoacetonitrile). Spearman rank correlation coefficients between DBPs varied between species and across areas, being highest between dibromochloromethane and dibromochloroacetic acid (r(s)=0.87). Principal component analyses of 13 DBPs (4 THMs, 9 HAAs) led 3 components explaining more than 80% of variance. In conclusion, THMs and HAAs have limited value as predictors of other DBPs on a generalised basis. Principal component analysis provides a complementary tool to address the complex nature of the mixture. Copyright © 2012 Elsevier Inc. All rights reserved.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Pierce, Karisa M.; Wood, Lianna F.; Wright, Bob W.
2005-12-01
A comprehensive two-dimensional (2D) retention time alignment algorithm was developed using a novel indexing scheme. The algorithm is termed comprehensive because it functions to correct the entire chromatogram in both dimensions and it preserves the separation information in both dimensions. Although the algorithm is demonstrated by correcting comprehensive two-dimensional gas chromatography (GC x GC) data, the algorithm is designed to correct shifting in all forms of 2D separations, such as LC x LC, LC x CE, CE x CE, and LC x GC. This 2D alignment algorithm was applied to three different data sets composed of replicate GC x GCmore » separations of (1) three 22-component control mixtures, (2) three gasoline samples, and (3) three diesel samples. The three data sets were collected using slightly different temperature or pressure programs to engender significant retention time shifting in the raw data and then demonstrate subsequent corrections of that shifting upon comprehensive 2D alignment of the data sets. Thirty 12-min GC x GC separations from three 22-component control mixtures were used to evaluate the 2D alignment performance (10 runs/mixture). The average standard deviation of the first column retention time improved 5-fold from 0.020 min (before alignment) to 0.004 min (after alignment). Concurrently, the average standard deviation of second column retention time improved 4-fold from 3.5 ms (before alignment) to 0.8 ms (after alignment). Alignment of the 30 control mixture chromatograms took 20 min. The quantitative integrity of the GC x GC data following 2D alignment was also investigated. The mean integrated signal was determined for all components in the three 22-component mixtures for all 30 replicates. The average percent difference in the integrated signal for each component before and after alignment was 2.6%. Singular value decomposition (SVD) was applied to the 22-component control mixture data before and after alignment to show the restoration of trilinearity to the data, since trilinearity benefits chemometric analysis. By applying comprehensive 2D retention time alignment to all three data sets (control mixtures, gasoline samples, and diesel samples), classification by principal component analysis (PCA) substantially improved, resulting in 100% accurate scores clustering.« less
NASA Astrophysics Data System (ADS)
Lim, Hoong-Ta; Murukeshan, Vadakke Matham
2017-06-01
Hyperspectral imaging combines imaging and spectroscopy to provide detailed spectral information for each spatial point in the image. This gives a three-dimensional spatial-spatial-spectral datacube with hundreds of spectral images. Probe-based hyperspectral imaging systems have been developed so that they can be used in regions where conventional table-top platforms would find it difficult to access. A fiber bundle, which is made up of specially-arranged optical fibers, has recently been developed and integrated with a spectrograph-based hyperspectral imager. This forms a snapshot hyperspectral imaging probe, which is able to form a datacube using the information from each scan. Compared to the other configurations, which require sequential scanning to form a datacube, the snapshot configuration is preferred in real-time applications where motion artifacts and pixel misregistration can be minimized. Principal component analysis is a dimension-reducing technique that can be applied in hyperspectral imaging to convert the spectral information into uncorrelated variables known as principal components. A confidence ellipse can be used to define the region of each class in the principal component feature space and for classification. This paper demonstrates the use of the snapshot hyperspectral imaging probe to acquire data from samples of different colors. The spectral library of each sample was acquired and then analyzed using principal component analysis. Confidence ellipse was then applied to the principal components of each sample and used as the classification criteria. The results show that the applied analysis can be used to perform classification of the spectral data acquired using the snapshot hyperspectral imaging probe.
Estimation of surface curvature from full-field shape data using principal component analysis
NASA Astrophysics Data System (ADS)
Sharma, Sameer; Vinuchakravarthy, S.; Subramanian, S. J.
2017-01-01
Three-dimensional digital image correlation (3D-DIC) is a popular image-based experimental technique for estimating surface shape, displacements and strains of deforming objects. In this technique, a calibrated stereo rig is used to obtain and stereo-match pairs of images of the object of interest from which the shapes of the imaged surface are then computed using the calibration parameters of the rig. Displacements are obtained by performing an additional temporal correlation of the shapes obtained at various stages of deformation and strains by smoothing and numerically differentiating the displacement data. Since strains are of primary importance in solid mechanics, significant efforts have been put into computation of strains from the measured displacement fields; however, much less attention has been paid to date to computation of curvature from the measured 3D surfaces. In this work, we address this gap by proposing a new method of computing curvature from full-field shape measurements using principal component analysis (PCA) along the lines of a similar work recently proposed to measure strains (Grama and Subramanian 2014 Exp. Mech. 54 913-33). PCA is a multivariate analysis tool that is widely used to reveal relationships between a large number of variables, reduce dimensionality and achieve significant denoising. This technique is applied here to identify dominant principal components in the shape fields measured by 3D-DIC and these principal components are then differentiated systematically to obtain the first and second fundamental forms used in the curvature calculation. The proposed method is first verified using synthetically generated noisy surfaces and then validated experimentally on some real world objects with known ground-truth curvatures.
NASA Astrophysics Data System (ADS)
Shah, Syed Muhammad Saqlain; Batool, Safeera; Khan, Imran; Ashraf, Muhammad Usman; Abbas, Syed Hussnain; Hussain, Syed Adnan
2017-09-01
Automatic diagnosis of human diseases are mostly achieved through decision support systems. The performance of these systems is mainly dependent on the selection of the most relevant features. This becomes harder when the dataset contains missing values for the different features. Probabilistic Principal Component Analysis (PPCA) has reputation to deal with the problem of missing values of attributes. This research presents a methodology which uses the results of medical tests as input, extracts a reduced dimensional feature subset and provides diagnosis of heart disease. The proposed methodology extracts high impact features in new projection by using Probabilistic Principal Component Analysis (PPCA). PPCA extracts projection vectors which contribute in highest covariance and these projection vectors are used to reduce feature dimension. The selection of projection vectors is done through Parallel Analysis (PA). The feature subset with the reduced dimension is provided to radial basis function (RBF) kernel based Support Vector Machines (SVM). The RBF based SVM serves the purpose of classification into two categories i.e., Heart Patient (HP) and Normal Subject (NS). The proposed methodology is evaluated through accuracy, specificity and sensitivity over the three datasets of UCI i.e., Cleveland, Switzerland and Hungarian. The statistical results achieved through the proposed technique are presented in comparison to the existing research showing its impact. The proposed technique achieved an accuracy of 82.18%, 85.82% and 91.30% for Cleveland, Hungarian and Switzerland dataset respectively.
Principal components analysis based control of a multi-DoF underactuated prosthetic hand.
Matrone, Giulia C; Cipriani, Christian; Secco, Emanuele L; Magenes, Giovanni; Carrozza, Maria Chiara
2010-04-23
Functionality, controllability and cosmetics are the key issues to be addressed in order to accomplish a successful functional substitution of the human hand by means of a prosthesis. Not only the prosthesis should duplicate the human hand in shape, functionality, sensorization, perception and sense of body-belonging, but it should also be controlled as the natural one, in the most intuitive and undemanding way. At present, prosthetic hands are controlled by means of non-invasive interfaces based on electromyography (EMG). Driving a multi degrees of freedom (DoF) hand for achieving hand dexterity implies to selectively modulate many different EMG signals in order to make each joint move independently, and this could require significant cognitive effort to the user. A Principal Components Analysis (PCA) based algorithm is used to drive a 16 DoFs underactuated prosthetic hand prototype (called CyberHand) with a two dimensional control input, in order to perform the three prehensile forms mostly used in Activities of Daily Living (ADLs). Such Principal Components set has been derived directly from the artificial hand by collecting its sensory data while performing 50 different grasps, and subsequently used for control. Trials have shown that two independent input signals can be successfully used to control the posture of a real robotic hand and that correct grasps (in terms of involved fingers, stability and posture) may be achieved. This work demonstrates the effectiveness of a bio-inspired system successfully conjugating the advantages of an underactuated, anthropomorphic hand with a PCA-based control strategy, and opens up promising possibilities for the development of an intuitively controllable hand prosthesis.
Principal component analysis of dynamic fluorescence images for diagnosis of diabetic vasculopathy
NASA Astrophysics Data System (ADS)
Seo, Jihye; An, Yuri; Lee, Jungsul; Ku, Taeyun; Kang, Yujung; Ahn, Chulwoo; Choi, Chulhee
2016-04-01
Indocyanine green (ICG) fluorescence imaging has been clinically used for noninvasive visualizations of vascular structures. We have previously developed a diagnostic system based on dynamic ICG fluorescence imaging for sensitive detection of vascular disorders. However, because high-dimensional raw data were used, the analysis of the ICG dynamics proved difficult. We used principal component analysis (PCA) in this study to extract important elements without significant loss of information. We examined ICG spatiotemporal profiles and identified critical features related to vascular disorders. PCA time courses of the first three components showed a distinct pattern in diabetic patients. Among the major components, the second principal component (PC2) represented arterial-like features. The explained variance of PC2 in diabetic patients was significantly lower than in normal controls. To visualize the spatial pattern of PCs, pixels were mapped with red, green, and blue channels. The PC2 score showed an inverse pattern between normal controls and diabetic patients. We propose that PC2 can be used as a representative bioimaging marker for the screening of vascular diseases. It may also be useful in simple extractions of arterial-like features.
NASA Technical Reports Server (NTRS)
Huang, Norden E. (Inventor)
2001-01-01
A computer implemented method of processing two-dimensional physical signals includes five basic components and the associated presentation techniques of the results. The first component decomposes the two-dimensional signal into one-dimensional profiles. The second component is a computer implemented Empirical Mode Decomposition that extracts a collection of Intrinsic Mode Functions (IMF's) from each profile based on local extrema and/or curvature extrema. The decomposition is based on the direct extraction of the energy associated with various intrinsic time scales in the profiles. In the third component, the IMF's of each profile are then subjected to a Hilbert Transform. The fourth component collates the Hilbert transformed IMF's of the profiles to form a two-dimensional Hilbert Spectrum. A fifth component manipulates the IMF's by, for example, filtering the two-dimensional signal by reconstructing the two-dimensional signal from selected IMF(s).
Modulated Hebb-Oja learning rule--a method for principal subspace analysis.
Jankovic, Marko V; Ogawa, Hidemitsu
2006-03-01
This paper presents analysis of the recently proposed modulated Hebb-Oja (MHO) method that performs linear mapping to a lower-dimensional subspace. Principal component subspace is the method that will be analyzed. Comparing to some other well-known methods for yielding principal component subspace (e.g., Oja's Subspace Learning Algorithm), the proposed method has one feature that could be seen as desirable from the biological point of view--synaptic efficacy learning rule does not need the explicit information about the value of the other efficacies to make individual efficacy modification. Also, the simplicity of the "neural circuits" that perform global computations and a fact that their number does not depend on the number of input and output neurons, could be seen as good features of the proposed method.
Magagna, Federico; Guglielmetti, Alessandro; Liberto, Erica; Reichenbach, Stephen E; Allegrucci, Elena; Gobino, Guido; Bicchi, Carlo; Cordero, Chiara
2017-08-02
This study investigates chemical information of volatile fractions of high-quality cocoa (Theobroma cacao L. Malvaceae) from different origins (Mexico, Ecuador, Venezuela, Columbia, Java, Trinidad, and Sao Tomè) produced for fine chocolate. This study explores the evolution of the entire pattern of volatiles in relation to cocoa processing (raw, roasted, steamed, and ground beans). Advanced chemical fingerprinting (e.g., combined untargeted and targeted fingerprinting) with comprehensive two-dimensional gas chromatography coupled with mass spectrometry allows advanced pattern recognition for classification, discrimination, and sensory-quality characterization. The entire data set is analyzed for 595 reliable two-dimensional peak regions, including 130 known analytes and 13 potent odorants. Multivariate analysis with unsupervised exploration (principal component analysis) and simple supervised discrimination methods (Fisher ratios and linear regression trees) reveal informative patterns of similarities and differences and identify characteristic compounds related to sample origin and manufacturing step.
Topological patterns of mesh textures in serpentinites
NASA Astrophysics Data System (ADS)
Miyazawa, M.; Suzuki, A.; Shimizu, H.; Okamoto, A.; Hiraoka, Y.; Obayashi, I.; Tsuji, T.; Ito, T.
2017-12-01
Serpentinization is a hydration process that forms serpentine minerals and magnetite within the oceanic lithosphere. Microfractures crosscut these minerals during the reactions, and the structures look like mesh textures. It has been known that the patterns of microfractures and the system evolutions are affected by the hydration reaction and fluid transport in fractures and within matrices. This study aims at quantifying the topological patterns of the mesh textures and understanding possible conditions of fluid transport and reaction during serpentinization in the oceanic lithosphere. Two-dimensional simulation by the distinct element method (DEM) generates fracture patterns due to serpentinization. The microfracture patterns are evaluated by persistent homology, which measures features of connected components of a topological space and encodes multi-scale topological features in the persistence diagrams. The persistence diagrams of the different mesh textures are evaluated by principal component analysis to bring out the strong patterns of persistence diagrams. This approach help extract feature values of fracture patterns from high-dimensional and complex datasets.
Correspondence Analysis-Theory and Application in Management Accounting Research
NASA Astrophysics Data System (ADS)
Duller, Christine
2010-09-01
Correspondence analysis is an explanatory data analytic technique and is used to identify systematic relations between categorical variables. It is related to principal component analysis and the results provide information on the structure of categorical variables similar to the results given by a principal component analysis in case of metric variables. Classical correspondence analysis is designed two-dimensional, whereas multiple correspondence analysis is an extension to more than two variables. After an introductory overview of the idea and the implementation in standard software packages (PASW, SAS, R) an example in recent research is presented, which deals with strategic management accounting in family and non-family enterprises in Austria, where 70% to 80% of all enterprises can be classified as family firms. Although there is a growing body of literature focusing on various management issues in family firms, so far the state of the art of strategic management accounting in family firms is an empirically under-researched subject. In relevant literature only the (empirically untested) hypothesis can be found, that family firms tend to have less formalized management accounting systems than non-family enterprises. Creating a correspondence analysis will help to identify the underlying structure, which is responsible for differences in strategic management accounting.
Three-dimensional geometry of coronal loops inferred by the Principal Component Analysis
NASA Astrophysics Data System (ADS)
Nisticò, Giuseppe; Nakariakov, Valery
We propose a new method for the determination of the three dimensional (3D) shape of coronal loops from stereoscopy. The common approach requires to find a 1D geometric curve, as circumference or ellipse, that best-fits the 3D tie-points which sample the loop shape in a given coordinate system. This can be easily achieved by the Principal Component (PC) analysis. It mainly consists in calculating the eigenvalues and eigenvectors of the covariance matrix of the 3D tie-points: the eigenvalues give a measure of the variability of the distribution of the tie-points, and the corresponding eigenvectors define a new cartesian reference frame directly related to the loop. The eigenvector associated with the smallest eigenvalues defines the normal to the loop plane, while the other two determine the directions of the loop axes: the major axis is related to the largest eigenvalue, and the minor axis with the second one. The magnitude of the axes is directly proportional to the square roots of these eigenvalues. The technique is fast and easily implemented in some examples, returning best-fitting estimations of the loop parameters and 3D reconstruction with a reasonable small number of tie-points. The method is suitable for serial reconstruction of coronal loops in active regions, providing a useful tool for comparison between observations and theoretical magnetic field extrapolations from potential or force-free fields.
Evidence of tampering in watermark identification
NASA Astrophysics Data System (ADS)
McLauchlan, Lifford; Mehrübeoglu, Mehrübe
2009-08-01
In this work, watermarks are embedded in digital images in the discrete wavelet transform (DWT) domain. Principal component analysis (PCA) is performed on the DWT coefficients. Next higher order statistics based on the principal components and the eigenvalues are determined for different sets of images. Feature sets are analyzed for different types of attacks in m dimensional space. The results demonstrate the separability of the features for the tampered digital copies. Different feature sets are studied to determine more effective tamper evident feature sets. The digital forensics, the probable manipulation(s) or modification(s) performed on the digital information can be identified using the described technique.
Maximally reliable spatial filtering of steady state visual evoked potentials.
Dmochowski, Jacek P; Greaves, Alex S; Norcia, Anthony M
2015-04-01
Due to their high signal-to-noise ratio (SNR) and robustness to artifacts, steady state visual evoked potentials (SSVEPs) are a popular technique for studying neural processing in the human visual system. SSVEPs are conventionally analyzed at individual electrodes or linear combinations of electrodes which maximize some variant of the SNR. Here we exploit the fundamental assumption of evoked responses--reproducibility across trials--to develop a technique that extracts a small number of high SNR, maximally reliable SSVEP components. This novel spatial filtering method operates on an array of Fourier coefficients and projects the data into a low-dimensional space in which the trial-to-trial spectral covariance is maximized. When applied to two sample data sets, the resulting technique recovers physiologically plausible components (i.e., the recovered topographies match the lead fields of the underlying sources) while drastically reducing the dimensionality of the data (i.e., more than 90% of the trial-to-trial reliability is captured in the first four components). Moreover, the proposed technique achieves a higher SNR than that of the single-best electrode or the Principal Components. We provide a freely-available MATLAB implementation of the proposed technique, herein termed "Reliable Components Analysis". Copyright © 2015 Elsevier Inc. All rights reserved.
Liver DCE-MRI Registration in Manifold Space Based on Robust Principal Component Analysis.
Feng, Qianjin; Zhou, Yujia; Li, Xueli; Mei, Yingjie; Lu, Zhentai; Zhang, Yu; Feng, Yanqiu; Liu, Yaqin; Yang, Wei; Chen, Wufan
2016-09-29
A technical challenge in the registration of dynamic contrast-enhanced magnetic resonance (DCE-MR) imaging in the liver is intensity variations caused by contrast agents. Such variations lead to the failure of the traditional intensity-based registration method. To address this problem, a manifold-based registration framework for liver DCE-MR time series is proposed. We assume that liver DCE-MR time series are located on a low-dimensional manifold and determine intrinsic similarities between frames. Based on the obtained manifold, the large deformation of two dissimilar images can be decomposed into a series of small deformations between adjacent images on the manifold through gradual deformation of each frame to the template image along the geodesic path. Furthermore, manifold construction is important in automating the selection of the template image, which is an approximation of the geodesic mean. Robust principal component analysis is performed to separate motion components from intensity changes induced by contrast agents; the components caused by motion are used to guide registration in eliminating the effect of contrast enhancement. Visual inspection and quantitative assessment are further performed on clinical dataset registration. Experiments show that the proposed method effectively reduces movements while preserving the topology of contrast-enhancing structures and provides improved registration performance.
Identifying apple surface defects using principal components analysis and artifical neural networks
USDA-ARS?s Scientific Manuscript database
Artificial neural networks and principal components were used to detect surface defects on apples in near-infrared images. Neural networks were trained and tested on sets of principal components derived from columns of pixels from images of apples acquired at two wavelengths (740 nm and 950 nm). I...
Non-linear principal component analysis applied to Lorenz models and to North Atlantic SLP
NASA Astrophysics Data System (ADS)
Russo, A.; Trigo, R. M.
2003-04-01
A non-linear generalisation of Principal Component Analysis (PCA), denoted Non-Linear Principal Component Analysis (NLPCA), is introduced and applied to the analysis of three data sets. Non-Linear Principal Component Analysis allows for the detection and characterisation of low-dimensional non-linear structure in multivariate data sets. This method is implemented using a 5-layer feed-forward neural network introduced originally in the chemical engineering literature (Kramer, 1991). The method is described and details of its implementation are addressed. Non-Linear Principal Component Analysis is first applied to a data set sampled from the Lorenz attractor (1963). It is found that the NLPCA approximations are more representative of the data than are the corresponding PCA approximations. The same methodology was applied to the less known Lorenz attractor (1984). However, the results obtained weren't as good as those attained with the famous 'Butterfly' attractor. Further work with this model is underway in order to assess if NLPCA techniques can be more representative of the data characteristics than are the corresponding PCA approximations. The application of NLPCA to relatively 'simple' dynamical systems, such as those proposed by Lorenz, is well understood. However, the application of NLPCA to a large climatic data set is much more challenging. Here, we have applied NLPCA to the sea level pressure (SLP) field for the entire North Atlantic area and the results show a slight imcrement of explained variance associated. Finally, directions for future work are presented.%}
Tracking Equilibrium and Nonequilibrium Shifts in Data with TREND.
Xu, Jia; Van Doren, Steven R
2017-01-24
Principal component analysis (PCA) discovers patterns in multivariate data that include spectra, microscopy, and other biophysical measurements. Direct application of PCA to crowded spectra, images, and movies (without selecting peaks or features) was shown recently to identify their equilibrium or temporal changes. To enable the community to utilize these capabilities with a wide range of measurements, we have developed multiplatform software named TREND to Track Equilibrium and Nonequilibrium population shifts among two-dimensional Data frames. TREND can also carry this out by independent component analysis. We highlight a few examples of finding concurrent processes. TREND extracts dual phases of binding to two sites directly from the NMR spectra of the titrations. In a cardiac movie from magnetic resonance imaging, TREND resolves principal components (PCs) representing breathing and the cardiac cycle. TREND can also reconstruct the series of measurements from selected PCs, as illustrated for a biphasic, NMR-detected titration and the cardiac MRI movie. Fidelity of reconstruction of series of NMR spectra or images requires more PCs than needed to plot the largest population shifts. TREND reads spectra from many spectroscopies in the most common formats (JCAMP-DX and NMR) and multiple movie formats. The TREND package thus provides convenient tools to resolve the processes recorded by diverse biophysical methods. Copyright © 2017 Biophysical Society. Published by Elsevier Inc. All rights reserved.
Visual enhancement of images of natural resources: Applications in geology
NASA Technical Reports Server (NTRS)
Dejesusparada, N. (Principal Investigator); Neto, G.; Araujo, E. O.; Mascarenhas, N. D. A.; Desouza, R. C. M.
1980-01-01
The principal components technique for use in multispectral scanner LANDSAT data processing results in optimum dimensionality reduction. A powerful tool for MSS IMAGE enhancement, the method provides a maximum impression of terrain ruggedness; this fact makes the technique well suited for geological analysis.
Wang, Jie-sheng; Han, Shuang; Shen, Na-na
2014-01-01
For predicting the key technology indicators (concentrate grade and tailings recovery rate) of flotation process, an echo state network (ESN) based fusion soft-sensor model optimized by the improved glowworm swarm optimization (GSO) algorithm is proposed. Firstly, the color feature (saturation and brightness) and texture features (angular second moment, sum entropy, inertia moment, etc.) based on grey-level co-occurrence matrix (GLCM) are adopted to describe the visual characteristics of the flotation froth image. Then the kernel principal component analysis (KPCA) method is used to reduce the dimensionality of the high-dimensional input vector composed by the flotation froth image characteristics and process datum and extracts the nonlinear principal components in order to reduce the ESN dimension and network complex. The ESN soft-sensor model of flotation process is optimized by the GSO algorithm with congestion factor. Simulation results show that the model has better generalization and prediction accuracy to meet the online soft-sensor requirements of the real-time control in the flotation process. PMID:24982935
Kong, Jessica; Giridharagopal, Rajiv; Harrison, Jeffrey S; Ginger, David S
2018-05-31
Correlating nanoscale chemical specificity with operational physics is a long-standing goal of functional scanning probe microscopy (SPM). We employ a data analytic approach combining multiple microscopy modes, using compositional information in infrared vibrational excitation maps acquired via photoinduced force microscopy (PiFM) with electrical information from conductive atomic force microscopy. We study a model polymer blend comprising insulating poly(methyl methacrylate) (PMMA) and semiconducting poly(3-hexylthiophene) (P3HT). We show that PiFM spectra are different from FTIR spectra, but can still be used to identify local composition. We use principal component analysis to extract statistically significant principal components and principal component regression to predict local current and identify local polymer composition. In doing so, we observe evidence of semiconducting P3HT within PMMA aggregates. These methods are generalizable to correlated SPM data and provide a meaningful technique for extracting complex compositional information that are impossible to measure from any one technique.
Similar and contrasting dimensions of social cognition in schizophrenia and healthy subjects.
Mehta, Urvakhsh Meherwan; Thirthalli, Jagadisha; Bhagyavathi, H D; Keshav Kumar, J; Subbakrishna, D K; Gangadhar, Bangalore N; Eack, Shaun M; Keshavan, Matcheri S
2014-08-01
Schizophrenia patients experience substantial impairments in social cognition (SC) and these deficits are associated with their poor functional outcome. Though SC is consistently shown to emerge as a cognitive dimension distinct from neurocognition, the dimensionality of SC is poorly understood. Moreover, comparing the components of SC between schizophrenia patients and healthy comparison subjects would provide specific insights on the construct validity of SC. We conducted principal component analyses of eight SC test scores (representing four domains of SC, namely, theory of mind, emotion processing, social perception and attributional bias) independently in 170 remitted schizophrenia patients and 111 matched healthy comparison subjects. We also conducted regression analyses to evaluate the relative contribution of individual SC components to other symptom dimensions, which are important clinical determinants of functional outcome (i.e., neurocognition, negative symptoms, motivational deficits and insight) in schizophrenia. A three-factor solution representing socio-emotional processing, social-inferential ability and external attribution components emerged in the patient group that accounted for 64.43% of the variance. In contrast, a two-factor solution representing socio-emotional processing and social-inferential ability was derived in the healthy comparison group that explained 56.5% of the variance. In the patient group, the social-inferential component predicted negative symptoms and motivational deficits. Our results suggest the presence of a multidimensional SC construct. The dimensionality of SC observed across the two groups, though not identical, displayed important parallels. Individual components also demonstrated distinct patterns of association with other symptom dimensions, thus supporting their external validity. Copyright © 2014 Elsevier B.V. All rights reserved.
Multi-segmental movement patterns reflect juggling complexity and skill level.
Zago, Matteo; Pacifici, Ilaria; Lovecchio, Nicola; Galli, Manuela; Federolf, Peter Andreas; Sforza, Chiarella
2017-08-01
The juggling action of six experts and six intermediates jugglers was recorded with a motion capture system and decomposed into its fundamental components through Principal Component Analysis. The aim was to quantify trends in movement dimensionality, multi-segmental patterns and rhythmicity as a function of proficiency level and task complexity. Dimensionality was quantified in terms of Residual Variance, while the Relative Amplitude was introduced to account for individual differences in movement components. We observed that: experience-related modifications in multi-segmental actions exist, such as the progressive reduction of error-correction movements, especially in complex task condition. The systematic identification of motor patterns sensitive to the acquisition of specific experience could accelerate the learning process. Copyright © 2017 Elsevier B.V. All rights reserved.
NASA Astrophysics Data System (ADS)
Nagai, Toshiki; Mitsutake, Ayori; Takano, Hiroshi
2013-02-01
A new relaxation mode analysis method, which is referred to as the principal component relaxation mode analysis method, has been proposed to handle a large number of degrees of freedom of protein systems. In this method, principal component analysis is carried out first and then relaxation mode analysis is applied to a small number of principal components with large fluctuations. To reduce the contribution of fast relaxation modes in these principal components efficiently, we have also proposed a relaxation mode analysis method using multiple evolution times. The principal component relaxation mode analysis method using two evolution times has been applied to an all-atom molecular dynamics simulation of human lysozyme in aqueous solution. Slow relaxation modes and corresponding relaxation times have been appropriately estimated, demonstrating that the method is applicable to protein systems.
NASA Astrophysics Data System (ADS)
Fan, Qimeng; Chen, Chaoyin; Huang, Zaiqiang; Zhang, Chunmei; Liang, Pengjuan; Zhao, Shenglan
2015-02-01
Rhizoma Gastrodiae (Tianma) of different variants and different geographical origins has vital difference in quality and physiological efficacy. This paper focused on the classification and identification of Tianma of six types (two variants from three different geographical origins) using three dimensional synchronous fluorescence spectroscopy (3D-SFS) coupled with principal component analysis (PCA). 3D-SF spectra of aqueous extracts, which were obtained from Tianma of the six types, were measured by a LS-50B luminescence spectrofluorometer. The experimental results showed that the characteristic fluorescent spectral regions of the 3D-SF spectra were similar, while the intensities of characteristic regions are different significantly. Coupled these differences in peak intensities with PCA, Tianma of six types could be discriminated successfully. In conclusion, 3D-SFS coupled with PCA, which has such advantages as effective, specific, rapid, non-polluting, has an edge for discrimination of the similar Chinese herbal medicine. And the proposed methodology is a useful tool to classify and identify Tianma of different variants and different geographical origins.
NASA Astrophysics Data System (ADS)
Shirota, Yukari; Hashimoto, Takako; Fitri Sari, Riri
2018-03-01
It has been very significant to visualize time series big data. In the paper we shall discuss a new analysis method called “statistical shape analysis” or “geometry driven statistics” on time series statistical data in economics. In the paper, we analyse the agriculture, value added and industry, value added (percentage of GDP) changes from 2000 to 2010 in Asia. We handle the data as a set of landmarks on a two-dimensional image to see the deformation using the principal components. The point of the analysis method is the principal components of the given formation which are eigenvectors of its bending energy matrix. The local deformation can be expressed as the set of non-Affine transformations. The transformations give us information about the local differences between in 2000 and in 2010. Because the non-Affine transformation can be decomposed into a set of partial warps, we present the partial warps visually. The statistical shape analysis is widely used in biology but, in economics, no application can be found. In the paper, we investigate its potential to analyse the economic data.
Visual Exploration of Semantic Relationships in Neural Word Embeddings
Liu, Shusen; Bremer, Peer-Timo; Thiagarajan, Jayaraman J.; ...
2017-08-29
Constructing distributed representations for words through neural language models and using the resulting vector spaces for analysis has become a crucial component of natural language processing (NLP). But, despite their widespread application, little is known about the structure and properties of these spaces. To gain insights into the relationship between words, the NLP community has begun to adapt high-dimensional visualization techniques. Particularly, researchers commonly use t-distributed stochastic neighbor embeddings (t-SNE) and principal component analysis (PCA) to create two-dimensional embeddings for assessing the overall structure and exploring linear relationships (e.g., word analogies), respectively. Unfortunately, these techniques often produce mediocre or evenmore » misleading results and cannot address domain-specific visualization challenges that are crucial for understanding semantic relationships in word embeddings. We introduce new embedding techniques for visualizing semantic and syntactic analogies, and the corresponding tests to determine whether the resulting views capture salient structures. Additionally, we introduce two novel views for a comprehensive study of analogy relationships. Finally, we augment t-SNE embeddings to convey uncertainty information in order to allow a reliable interpretation. Combined, the different views address a number of domain-specific tasks difficult to solve with existing tools.« less
Visual Exploration of Semantic Relationships in Neural Word Embeddings
DOE Office of Scientific and Technical Information (OSTI.GOV)
Liu, Shusen; Bremer, Peer-Timo; Thiagarajan, Jayaraman J.
Constructing distributed representations for words through neural language models and using the resulting vector spaces for analysis has become a crucial component of natural language processing (NLP). But, despite their widespread application, little is known about the structure and properties of these spaces. To gain insights into the relationship between words, the NLP community has begun to adapt high-dimensional visualization techniques. Particularly, researchers commonly use t-distributed stochastic neighbor embeddings (t-SNE) and principal component analysis (PCA) to create two-dimensional embeddings for assessing the overall structure and exploring linear relationships (e.g., word analogies), respectively. Unfortunately, these techniques often produce mediocre or evenmore » misleading results and cannot address domain-specific visualization challenges that are crucial for understanding semantic relationships in word embeddings. We introduce new embedding techniques for visualizing semantic and syntactic analogies, and the corresponding tests to determine whether the resulting views capture salient structures. Additionally, we introduce two novel views for a comprehensive study of analogy relationships. Finally, we augment t-SNE embeddings to convey uncertainty information in order to allow a reliable interpretation. Combined, the different views address a number of domain-specific tasks difficult to solve with existing tools.« less
Hierarchical Regularity in Multi-Basin Dynamics on Protein Landscapes
NASA Astrophysics Data System (ADS)
Matsunaga, Yasuhiro; Kostov, Konstatin S.; Komatsuzaki, Tamiki
2004-04-01
We analyze time series of potential energy fluctuations and principal components at several temperatures for two kinds of off-lattice 46-bead models that have two distinctive energy landscapes. The less-frustrated "funnel" energy landscape brings about stronger nonstationary behavior of the potential energy fluctuations at the folding temperature than the other, rather frustrated energy landscape at the collapse temperature. By combining principal component analysis with an embedding nonlinear time-series analysis, it is shown that the fast fluctuations with small amplitudes of 70-80% of the principal components cause the time series to become almost "random" in only 100 simulation steps. However, the stochastic feature of the principal components tends to be suppressed through a wide range of degrees of freedom at the transition temperature.
NASA Astrophysics Data System (ADS)
Crivori, Patrizia; Zamora, Ismael; Speed, Bill; Orrenius, Christian; Poggesi, Italo
2004-03-01
A number of computational approaches are being proposed for an early optimization of ADME (absorption, distribution, metabolism and excretion) properties to increase the success rate in drug discovery. The present study describes the development of an in silico model able to estimate, from the three-dimensional structure of a molecule, the stability of a compound with respect to the human cytochrome P450 (CYP) 3A4 enzyme activity. Stability data were obtained by measuring the amount of unchanged compound remaining after a standardized incubation with human cDNA-expressed CYP3A4. The computational method transforms the three-dimensional molecular interaction fields (MIFs) generated from the molecular structure into descriptors (VolSurf and Almond procedures). The descriptors were correlated to the experimental metabolic stability classes by a partial least squares discriminant procedure. The model was trained using a set of 1800 compounds from the Pharmacia collection and was validated using two test sets: the first one including 825 compounds from the Pharmacia collection and the second one consisting of 20 known drugs. This model correctly predicted 75% of the first and 85% of the second test set and showed a precision above 86% to correctly select metabolically stable compounds. The model appears a valuable tool in the design of virtual libraries to bias the selection toward more stable compounds. Abbreviations: ADME - absorption, distribution, metabolism and excretion; CYP - cytochrome P450; MIFs - molecular interaction fields; HTS - high throughput screening; DDI - drug-drug interactions; 3D - three-dimensional; PCA - principal components analysis; CPCA - consensus principal components analysis; PLS - partial least squares; PLSD - partial least squares discriminant; GRIND - grid independent descriptors; GRID - software originally created and developed by Professor Peter Goodford.
NASA Technical Reports Server (NTRS)
Liu, Xu; Smith, William L.; Zhou, Daniel K.; Larar, Allen
2005-01-01
Modern infrared satellite sensors such as Atmospheric Infrared Sounder (AIRS), Cosmic Ray Isotope Spectrometer (CrIS), Thermal Emission Spectrometer (TES), Geosynchronous Imaging Fourier Transform Spectrometer (GIFTS) and Infrared Atmospheric Sounding Interferometer (IASI) are capable of providing high spatial and spectral resolution infrared spectra. To fully exploit the vast amount of spectral information from these instruments, super fast radiative transfer models are needed. This paper presents a novel radiative transfer model based on principal component analysis. Instead of predicting channel radiance or transmittance spectra directly, the Principal Component-based Radiative Transfer Model (PCRTM) predicts the Principal Component (PC) scores of these quantities. This prediction ability leads to significant savings in computational time. The parameterization of the PCRTM model is derived from properties of PC scores and instrument line shape functions. The PCRTM is very accurate and flexible. Due to its high speed and compressed spectral information format, it has great potential for super fast one-dimensional physical retrievals and for Numerical Weather Prediction (NWP) large volume radiance data assimilation applications. The model has been successfully developed for the National Polar-orbiting Operational Environmental Satellite System Airborne Sounder Testbed - Interferometer (NAST-I) and AIRS instruments. The PCRTM model performs monochromatic radiative transfer calculations and is able to include multiple scattering calculations to account for clouds and aerosols.
Quantum machine learning for quantum anomaly detection
NASA Astrophysics Data System (ADS)
Liu, Nana; Rebentrost, Patrick
2018-04-01
Anomaly detection is used for identifying data that deviate from "normal" data patterns. Its usage on classical data finds diverse applications in many important areas such as finance, fraud detection, medical diagnoses, data cleaning, and surveillance. With the advent of quantum technologies, anomaly detection of quantum data, in the form of quantum states, may become an important component of quantum applications. Machine-learning algorithms are playing pivotal roles in anomaly detection using classical data. Two widely used algorithms are the kernel principal component analysis and the one-class support vector machine. We find corresponding quantum algorithms to detect anomalies in quantum states. We show that these two quantum algorithms can be performed using resources that are logarithmic in the dimensionality of quantum states. For pure quantum states, these resources can also be logarithmic in the number of quantum states used for training the machine-learning algorithm. This makes these algorithms potentially applicable to big quantum data applications.
Feng, Xiao; Ding, Xuesong; Chen, Long; Wu, Yang; Liu, Lili; Addicoat, Matthew; Irle, Stephan; Dong, Yuping; Jiang, Donglin
2016-01-01
Highly ordered discrete assemblies of chlorophylls that are found in natural light-harvesting antennae are key to photosynthesis, which converts light energy to chemical energy and is the principal producer of organic matter on Earth. Porphyrins and phthalocyanines, which are analogues of chlorophylls, exhibit a strong absorbance of visible and near-infrared light, respectively. A highly ordered porphyrin-co-phthalocyanine antennae would harvest photons over the entire solar spectrum for chemical transformation. However, such a robust antennae has not yet been synthesised. Herein, we report a strategy that merges covalent bonds and noncovalent forces to produce highly ordered two-dimensional porphyrin-co-phthalocyanine antennae. This methodology enables control over the stoichiometry and order of the porphyrin and phthalocyanine units; more importantly, this approach is compatible with various metalloporphyrin and metallophthalocyanine derivatives and thus may lead to the generation of a broad structural diversity of two-dimensional artificial antennae. These ordered porphyrin-co-phthalocyanine two-dimensional antennae exhibit unique optical properties and catalytic functions that are not available with single-component or non-structured materials. These 2D artificial antennae exhibit exceptional light-harvesting capacity over the entire solar spectrum as a result of a synergistic light-absorption effect. In addition, they exhibit outstanding photosensitising activities in using both visible and near-infrared photons for producing singlet oxygen. PMID:27622274
Haid, Thomas H.; Doix, Aude-Clémence M.; Nigg, Benno M.; Federolf, Peter A.
2018-01-01
Optimal feedback control theory suggests that control of movement is focused on movement dimensions that are important for the task's success. The current study tested the hypotheses that age effects would emerge in the control of only specific movement components and that these components would be linked to the task relevance. Fifty healthy volunteers, 25 young and 25 older adults, performed a 80s-tandem stance while their postural movements were recorded using a standard motion capture system. The postural movements were decomposed by a principal component analysis into one-dimensional movement components, PMk, whose control was assessed through two variables, Nk and σk, which characterized the tightness and the regularity of the neuro-muscular control, respectively. The older volunteers showed less tight and more irregular control in PM2 (N2: −9.2%, p = 0.007; σ2: +14.3.0%, p = 0.017) but tighter control in PM8 and PM9 (N8: +4.7%, p = 0.020; N9: +2.5%, p = 0.043; σ9: −8.8%, p = 0.025). These results suggest that aging effects alter the postural control system not as a whole, but emerge in specific, task relevant components. The findings of the current study thus support the hypothesis that the minimal intervention principle, as described in the context of optimal feedback control (OFC), may be relevant when assessing aging effects on postural control. PMID:29459826
Hinkel-Lipsker, Jacob W; Hahn, Michael E
2018-06-01
Gait adaptation is a task that requires fine-tuned coordination of all degrees of freedom in the lower limbs by the central nervous system. However, when individuals change their gait it is unknown how this coordination is organized, and how it can be influenced by contextual interference during practice. Such knowledge could provide information about measurement of gait adaptation during rehabilitation. Able-bodied individuals completed an acute bout of asymmetric split-belt treadmill walking, where one limb was driven at a constant velocity and the other according to one of three designed practice paradigms: serial practice, where the variable limb belt velocity increased over time; random blocked practice, where every 20 strides the variable limb belt velocity changed randomly; random practice, where every stride the variable limb belt velocity changed randomly. On the second day, subjects completed one of two different transfer tests; one with a belt asymmetry close to that experienced on the acquisition day (transfer 1; 1.5:1), and one with a greater asymmetry (transfer 2; 2:1) . To reduce this inherently high-dimensional dataset, principal component analyses were used for kinematic data collected throughout the acquisition and transfer phases; resulting in extraction of the first two principal components (PCs). For acquisition, PC1 and PC2 were related to sagittal and frontal plane control. For transfer 1, PC1 and PC2 were related to frontal plane control of the base of support and whole-body center of mass. For transfer 2, PC1 did not have any variables with high enough coefficients deemed to be relevant, and PC2 was related to sagittal plane control. Observations of principal component scores indicate that variance structuring differs among practice groups during acquisition and transfer 1, but not transfer 2. These results demonstrate the main kinematic coordinative structures that exist during gait adaptation, and that control of sagittal plane and frontal plane motion are perhaps a trade-off during acquisition of a novel asymmetric gait pattern. Copyright © 2018 Elsevier B.V. All rights reserved.
Dall'Asta, Andrea; Schievano, Silvia; Bruse, Jan L; Paramasivam, Gowrishankar; Kaihura, Christine Tita; Dunaway, David; Lees, Christoph C
2017-07-01
The antenatal detection of facial dysmorphism using 3-dimensional ultrasound may raise the suspicion of an underlying genetic condition but infrequently leads to a definitive antenatal diagnosis. Despite advances in array and noninvasive prenatal testing, not all genetic conditions can be ascertained from such testing. The aim of this study was to investigate the feasibility of quantitative assessment of fetal face features using prenatal 3-dimensional ultrasound volumes and statistical shape modeling. STUDY DESIGN: Thirteen normal and 7 abnormal stored 3-dimensional ultrasound fetal face volumes were analyzed, at a median gestation of 29 +4 weeks (25 +0 to 36 +1 ). The 20 3-dimensional surface meshes generated were aligned and served as input for a statistical shape model, which computed the mean 3-dimensional face shape and 3-dimensional shape variations using principal component analysis. Ten shape modes explained more than 90% of the total shape variability in the population. While the first mode accounted for overall size differences, the second highlighted shape feature changes from an overall proportionate toward a more asymmetric face shape with a wide prominent forehead and an undersized, posteriorly positioned chin. Analysis of the Mahalanobis distance in principal component analysis shape space suggested differences between normal and abnormal fetuses (median and interquartile range distance values, 7.31 ± 5.54 for the normal group vs 13.27 ± 9.82 for the abnormal group) (P = .056). This feasibility study demonstrates that objective characterization and quantification of fetal facial morphology is possible from 3-dimensional ultrasound. This technique has the potential to assist in utero diagnosis, particularly of rare conditions in which facial dysmorphology is a feature. Copyright © 2017 Elsevier Inc. All rights reserved.
Principals' Perceptions Regarding Their Supervision and Evaluation
ERIC Educational Resources Information Center
Hvidston, David J.; Range, Bret G.; McKim, Courtney Ann
2015-01-01
This study examined the perceptions of principals concerning principal evaluation and supervisory feedback. Principals were asked two open-ended questions. Respondents included 82 principals in the Rocky Mountain region. The emerging themes were "Superintendent Performance," "Principal Evaluation Components," "Specific…
Differences in kinematic control of ankle joint motions in people with chronic ankle instability.
Kipp, Kristof; Palmieri-Smith, Riann M
2013-06-01
People with chronic ankle instability display different ankle joint motions compared to healthy people. The purpose of this study was to investigate the strategies used to control ankle joint motions between a group of people with chronic ankle instability and a group of healthy, matched controls. Kinematic data were collected from 11 people with chronic ankle instability and 11 matched control subjects as they performed a single-leg land-and-cut maneuver. Three-dimensional ankle joint angles were calculated from 100 ms before, to 200 ms after landing. Kinematic control of the three rotational ankle joint degrees of freedom was investigated by simultaneously examining the three-dimensional co-variation of plantarflexion/dorsiflexion, toe-in/toe-out rotation, and inversion/eversion motions with principal component analysis. Group differences in the variance proportions of the first two principal components indicated that the angular co-variation between ankle joint motions was more linear in the control group, but more planar in the chronic ankle instability group. Frontal and transverse plane motions, in particular, contributed to the group differences in the linearity and planarity of angular co-variation. People with chronic ankle instability use a different kinematic control strategy to coordinate ankle joint motions during a single-leg landing task. Compared to the healthy group, the chronic ankle instability group's control strategy appeared to be more complex and involved joint-specific contributions that would tend to predispose this group to recurring episodes of instability. Copyright © 2013 Elsevier Ltd. All rights reserved.
Jankovic, Marko; Ogawa, Hidemitsu
2003-08-01
This paper presents one possible implementation of a transformation that performs linear mapping to a lower-dimensional subspace. Principal component subspace will be the one that will be analyzed. Idea implemented in this paper represents generalization of the recently proposed infinity OH neural method for principal component extraction. The calculations in the newly proposed method are performed locally--a feature which is usually considered as desirable from the biological point of view. Comparing to some other wellknown methods, proposed synaptic efficacy learning rule requires less information about the value of the other efficacies to make single efficacy modification. Synaptic efficacies are modified by implementation of Modulated Hebb-type (MH) learning rule. Slightly modified MH algorithm named Modulated Hebb Oja (MHO) algorithm, will be also introduced. Structural similarity of the proposed network with part of the retinal circuit will be presented, too.
Multi-Dimensional Parental Involvement in Schools: A Principal's Guide
ERIC Educational Resources Information Center
Rapp, Nicole; Duncan, Heather
2012-01-01
Parental involvement is an important indicator of students' success in school. When schools engage families in a manner connected to improving learning, students do make greater gains. Creating and implementing an effective parental involvement model is an essential component in increasing student achievement in school. This article addresses the…
Beta Hebbian Learning as a New Method for Exploratory Projection Pursuit.
Quintián, Héctor; Corchado, Emilio
2017-09-01
In this research, a novel family of learning rules called Beta Hebbian Learning (BHL) is thoroughly investigated to extract information from high-dimensional datasets by projecting the data onto low-dimensional (typically two dimensional) subspaces, improving the existing exploratory methods by providing a clear representation of data's internal structure. BHL applies a family of learning rules derived from the Probability Density Function (PDF) of the residual based on the beta distribution. This family of rules may be called Hebbian in that all use a simple multiplication of the output of the neural network with some function of the residuals after feedback. The derived learning rules can be linked to an adaptive form of Exploratory Projection Pursuit and with artificial distributions, the networks perform as the theory suggests they should: the use of different learning rules derived from different PDFs allows the identification of "interesting" dimensions (as far from the Gaussian distribution as possible) in high-dimensional datasets. This novel algorithm, BHL, has been tested over seven artificial datasets to study the behavior of BHL parameters, and was later applied successfully over four real datasets, comparing its results, in terms of performance, with other well-known Exploratory and projection models such as Maximum Likelihood Hebbian Learning (MLHL), Locally-Linear Embedding (LLE), Curvilinear Component Analysis (CCA), Isomap and Neural Principal Component Analysis (Neural PCA).
Sparse modeling of spatial environmental variables associated with asthma
Chang, Timothy S.; Gangnon, Ronald E.; Page, C. David; Buckingham, William R.; Tandias, Aman; Cowan, Kelly J.; Tomasallo, Carrie D.; Arndt, Brian G.; Hanrahan, Lawrence P.; Guilbert, Theresa W.
2014-01-01
Geographically distributed environmental factors influence the burden of diseases such as asthma. Our objective was to identify sparse environmental variables associated with asthma diagnosis gathered from a large electronic health record (EHR) dataset while controlling for spatial variation. An EHR dataset from the University of Wisconsin’s Family Medicine, Internal Medicine and Pediatrics Departments was obtained for 199,220 patients aged 5–50 years over a three-year period. Each patient’s home address was geocoded to one of 3,456 geographic census block groups. Over one thousand block group variables were obtained from a commercial database. We developed a Sparse Spatial Environmental Analysis (SASEA). Using this method, the environmental variables were first dimensionally reduced with sparse principal component analysis. Logistic thin plate regression spline modeling was then used to identify block group variables associated with asthma from sparse principal components. The addresses of patients from the EHR dataset were distributed throughout the majority of Wisconsin’s geography. Logistic thin plate regression spline modeling captured spatial variation of asthma. Four sparse principal components identified via model selection consisted of food at home, dog ownership, household size, and disposable income variables. In rural areas, dog ownership and renter occupied housing units from significant sparse principal components were associated with asthma. Our main contribution is the incorporation of sparsity in spatial modeling. SASEA sequentially added sparse principal components to Logistic thin plate regression spline modeling. This method allowed association of geographically distributed environmental factors with asthma using EHR and environmental datasets. SASEA can be applied to other diseases with environmental risk factors. PMID:25533437
Sparse modeling of spatial environmental variables associated with asthma.
Chang, Timothy S; Gangnon, Ronald E; David Page, C; Buckingham, William R; Tandias, Aman; Cowan, Kelly J; Tomasallo, Carrie D; Arndt, Brian G; Hanrahan, Lawrence P; Guilbert, Theresa W
2015-02-01
Geographically distributed environmental factors influence the burden of diseases such as asthma. Our objective was to identify sparse environmental variables associated with asthma diagnosis gathered from a large electronic health record (EHR) dataset while controlling for spatial variation. An EHR dataset from the University of Wisconsin's Family Medicine, Internal Medicine and Pediatrics Departments was obtained for 199,220 patients aged 5-50years over a three-year period. Each patient's home address was geocoded to one of 3456 geographic census block groups. Over one thousand block group variables were obtained from a commercial database. We developed a Sparse Spatial Environmental Analysis (SASEA). Using this method, the environmental variables were first dimensionally reduced with sparse principal component analysis. Logistic thin plate regression spline modeling was then used to identify block group variables associated with asthma from sparse principal components. The addresses of patients from the EHR dataset were distributed throughout the majority of Wisconsin's geography. Logistic thin plate regression spline modeling captured spatial variation of asthma. Four sparse principal components identified via model selection consisted of food at home, dog ownership, household size, and disposable income variables. In rural areas, dog ownership and renter occupied housing units from significant sparse principal components were associated with asthma. Our main contribution is the incorporation of sparsity in spatial modeling. SASEA sequentially added sparse principal components to Logistic thin plate regression spline modeling. This method allowed association of geographically distributed environmental factors with asthma using EHR and environmental datasets. SASEA can be applied to other diseases with environmental risk factors. Copyright © 2014 Elsevier Inc. All rights reserved.
Large Covariance Estimation by Thresholding Principal Orthogonal Complements
Fan, Jianqing; Liao, Yuan; Mincheva, Martina
2012-01-01
This paper deals with the estimation of a high-dimensional covariance with a conditional sparsity structure and fast-diverging eigenvalues. By assuming sparse error covariance matrix in an approximate factor model, we allow for the presence of some cross-sectional correlation even after taking out common but unobservable factors. We introduce the Principal Orthogonal complEment Thresholding (POET) method to explore such an approximate factor structure with sparsity. The POET estimator includes the sample covariance matrix, the factor-based covariance matrix (Fan, Fan, and Lv, 2008), the thresholding estimator (Bickel and Levina, 2008) and the adaptive thresholding estimator (Cai and Liu, 2011) as specific examples. We provide mathematical insights when the factor analysis is approximately the same as the principal component analysis for high-dimensional data. The rates of convergence of the sparse residual covariance matrix and the conditional sparse covariance matrix are studied under various norms. It is shown that the impact of estimating the unknown factors vanishes as the dimensionality increases. The uniform rates of convergence for the unobserved factors and their factor loadings are derived. The asymptotic results are also verified by extensive simulation studies. Finally, a real data application on portfolio allocation is presented. PMID:24348088
Large Covariance Estimation by Thresholding Principal Orthogonal Complements.
Fan, Jianqing; Liao, Yuan; Mincheva, Martina
2013-09-01
This paper deals with the estimation of a high-dimensional covariance with a conditional sparsity structure and fast-diverging eigenvalues. By assuming sparse error covariance matrix in an approximate factor model, we allow for the presence of some cross-sectional correlation even after taking out common but unobservable factors. We introduce the Principal Orthogonal complEment Thresholding (POET) method to explore such an approximate factor structure with sparsity. The POET estimator includes the sample covariance matrix, the factor-based covariance matrix (Fan, Fan, and Lv, 2008), the thresholding estimator (Bickel and Levina, 2008) and the adaptive thresholding estimator (Cai and Liu, 2011) as specific examples. We provide mathematical insights when the factor analysis is approximately the same as the principal component analysis for high-dimensional data. The rates of convergence of the sparse residual covariance matrix and the conditional sparse covariance matrix are studied under various norms. It is shown that the impact of estimating the unknown factors vanishes as the dimensionality increases. The uniform rates of convergence for the unobserved factors and their factor loadings are derived. The asymptotic results are also verified by extensive simulation studies. Finally, a real data application on portfolio allocation is presented.
Encounter complexes and dimensionality reduction in protein-protein association.
Kozakov, Dima; Li, Keyong; Hall, David R; Beglov, Dmitri; Zheng, Jiefu; Vakili, Pirooz; Schueler-Furman, Ora; Paschalidis, Ioannis Ch; Clore, G Marius; Vajda, Sandor
2014-04-08
An outstanding challenge has been to understand the mechanism whereby proteins associate. We report here the results of exhaustively sampling the conformational space in protein-protein association using a physics-based energy function. The agreement between experimental intermolecular paramagnetic relaxation enhancement (PRE) data and the PRE profiles calculated from the docked structures shows that the method captures both specific and non-specific encounter complexes. To explore the energy landscape in the vicinity of the native structure, the nonlinear manifold describing the relative orientation of two solid bodies is projected onto a Euclidean space in which the shape of low energy regions is studied by principal component analysis. Results show that the energy surface is canyon-like, with a smooth funnel within a two dimensional subspace capturing over 75% of the total motion. Thus, proteins tend to associate along preferred pathways, similar to sliding of a protein along DNA in the process of protein-DNA recognition. DOI: http://dx.doi.org/10.7554/eLife.01370.001.
Encounter complexes and dimensionality reduction in protein–protein association
Kozakov, Dima; Li, Keyong; Hall, David R; Beglov, Dmitri; Zheng, Jiefu; Vakili, Pirooz; Schueler-Furman, Ora; Paschalidis, Ioannis Ch; Clore, G Marius; Vajda, Sandor
2014-01-01
An outstanding challenge has been to understand the mechanism whereby proteins associate. We report here the results of exhaustively sampling the conformational space in protein–protein association using a physics-based energy function. The agreement between experimental intermolecular paramagnetic relaxation enhancement (PRE) data and the PRE profiles calculated from the docked structures shows that the method captures both specific and non-specific encounter complexes. To explore the energy landscape in the vicinity of the native structure, the nonlinear manifold describing the relative orientation of two solid bodies is projected onto a Euclidean space in which the shape of low energy regions is studied by principal component analysis. Results show that the energy surface is canyon-like, with a smooth funnel within a two dimensional subspace capturing over 75% of the total motion. Thus, proteins tend to associate along preferred pathways, similar to sliding of a protein along DNA in the process of protein-DNA recognition. DOI: http://dx.doi.org/10.7554/eLife.01370.001 PMID:24714491
Unsupervised spike sorting based on discriminative subspace learning.
Keshtkaran, Mohammad Reza; Yang, Zhi
2014-01-01
Spike sorting is a fundamental preprocessing step for many neuroscience studies which rely on the analysis of spike trains. In this paper, we present two unsupervised spike sorting algorithms based on discriminative subspace learning. The first algorithm simultaneously learns the discriminative feature subspace and performs clustering. It uses histogram of features in the most discriminative projection to detect the number of neurons. The second algorithm performs hierarchical divisive clustering that learns a discriminative 1-dimensional subspace for clustering in each level of the hierarchy until achieving almost unimodal distribution in the subspace. The algorithms are tested on synthetic and in-vivo data, and are compared against two widely used spike sorting methods. The comparative results demonstrate that our spike sorting methods can achieve substantially higher accuracy in lower dimensional feature space, and they are highly robust to noise. Moreover, they provide significantly better cluster separability in the learned subspace than in the subspace obtained by principal component analysis or wavelet transform.
Compressive light field imaging
NASA Astrophysics Data System (ADS)
Ashok, Amit; Neifeld, Mark A.
2010-04-01
Light field imagers such as the plenoptic and the integral imagers inherently measure projections of the four dimensional (4D) light field scalar function onto a two dimensional sensor and therefore, suffer from a spatial vs. angular resolution trade-off. Programmable light field imagers, proposed recently, overcome this spatioangular resolution trade-off and allow high-resolution capture of the (4D) light field function with multiple measurements at the cost of a longer exposure time. However, these light field imagers do not exploit the spatio-angular correlations inherent in the light fields of natural scenes and thus result in photon-inefficient measurements. Here, we describe two architectures for compressive light field imaging that require relatively few photon-efficient measurements to obtain a high-resolution estimate of the light field while reducing the overall exposure time. Our simulation study shows that, compressive light field imagers using the principal component (PC) measurement basis require four times fewer measurements and three times shorter exposure time compared to a conventional light field imager in order to achieve an equivalent light field reconstruction quality.
A silicon microwire under a three-dimensional anisotropic tensile stress
DOE Office of Scientific and Technical Information (OSTI.GOV)
Ji, Xiaoyu; Poilvert, Nicolas; Liu, Wenjun
Three-dimensional tensile stress, or triaxial tensile stress, is difficult to achieve in a material. We present the investigation of an unusual three-dimensional anisotropic tensile stress field and its influence on the electronic properties of a single crystal silicon microwire. The microwire was created by laser heating an amorphous silicon wire deposited in a 1.7 μm silica glass capillary by high pressure chemical vapor deposition. Tensile strain arises due to the thermal expansion mismatch between silicon and silica. Synchrotron X-ray micro-beam Laue diffraction (μ-Laue) microscopy reveals that the three principal strain components are +0.47% (corresponding to a tensile stress of +0.7more » GPa) along the fiber axis and nearly isotropic +0.02% (corresponding to a tensile stress of +0.3 GPa) in the cross-sectional plane. This effect was accompanied with a reduction of 30 meV in the band gap energy of silicon, as predicted by the density-functional theory calculations and in close agreement with energy-dependent photoconductivity measurements. While silicon has been explored under many stress states, this study explores a stress state where all three principal stress components are tensile. Given the technological importance of silicon, the influence of such an unusual stress state on its electronic properties is of fundamental interest.« less
Joint evolution of multiple social traits: a kin selection analysis
Brown, Sam P.; Taylor, Peter D.
2010-01-01
General models of the evolution of cooperation, altruism and other social behaviours have focused almost entirely on single traits, whereas it is clear that social traits commonly interact. We develop a general kin-selection framework for the evolution of social behaviours in multiple dimensions. We show that whenever there are interactions among social traits new behaviours can emerge that are not predicted by one-dimensional analyses. For example, a prohibitively costly cooperative trait can ultimately be favoured owing to initial evolution in other (cheaper) social traits that in turn change the cost–benefit ratio of the original trait. To understand these behaviours, we use a two-dimensional stability criterion that can be viewed as an extension of Hamilton's rule. Our principal example is the social dilemma posed by, first, the construction and, second, the exploitation of a shared public good. We find that, contrary to the separate one-dimensional analyses, evolutionary feedback between the two traits can cause an increase in the equilibrium level of selfish exploitation with increasing relatedness, while both social (production plus exploitation) and asocial (neither) strategies can be locally stable. Our results demonstrate the importance of emergent stability properties of multidimensional social dilemmas, as one-dimensional stability in all component dimensions can conceal multidimensional instability. PMID:19828549
Constrained Principal Component Analysis: Various Applications.
ERIC Educational Resources Information Center
Hunter, Michael; Takane, Yoshio
2002-01-01
Provides example applications of constrained principal component analysis (CPCA) that illustrate the method on a variety of contexts common to psychological research. Two new analyses, decompositions into finer components and fitting higher order structures, are presented, followed by an illustration of CPCA on contingency tables and the CPCA of…
Alizadeh-Pasdar, Nooshin; Nakai, Shuryo; Li-Chan, Eunice C Y
2002-10-09
Raman spectroscopy was used to elucidate structural changes of beta-lactoglobulin (BLG), whey protein isolate (WPI), and bovine serum albumin (BSA), at 15% concentration, as a function of pH (5.0, 7.0, and 9.0), heating (80 degrees C, 30 min), and presence of 0.24% kappa-carrageenan. Three data-processing techniques were used to assist in identifying significant changes in Raman spectral data. Analysis of variance showed that of 12 characteristics examined in the Raman spectra, only a few were significantly affected by pH, heating, kappa-carrageenan, and their interactions. These included amide I (1658 cm(-1)) for WPI and BLG, alpha-helix for BLG and BSA, beta-sheet for BSA, CH stretching (2880 cm(-1)) for BLG and BSA, and CH stretching (2930 cm(-1)) for BSA. Principal component analysis reduced dimensionality of the characteristics. Heating and its interaction with kappa-carrageenan were identified as the most influential in overall structure of the whey proteins, using principal component similarity analysis.
Finger crease pattern recognition using Legendre moments and principal component analysis
NASA Astrophysics Data System (ADS)
Luo, Rongfang; Lin, Tusheng
2007-03-01
The finger joint lines defined as finger creases and its distribution can identify a person. In this paper, we propose a new finger crease pattern recognition method based on Legendre moments and principal component analysis (PCA). After obtaining the region of interest (ROI) for each finger image in the pre-processing stage, Legendre moments under Radon transform are applied to construct a moment feature matrix from the ROI, which greatly decreases the dimensionality of ROI and can represent principal components of the finger creases quite well. Then, an approach to finger crease pattern recognition is designed based on Karhunen-Loeve (K-L) transform. The method applies PCA to a moment feature matrix rather than the original image matrix to achieve the feature vector. The proposed method has been tested on a database of 824 images from 103 individuals using the nearest neighbor classifier. The accuracy up to 98.584% has been obtained when using 4 samples per class for training. The experimental results demonstrate that our proposed approach is feasible and effective in biometrics.
Nonparametric regression applied to quantitative structure-activity relationships
Constans; Hirst
2000-03-01
Several nonparametric regressors have been applied to modeling quantitative structure-activity relationship (QSAR) data. The simplest regressor, the Nadaraya-Watson, was assessed in a genuine multivariate setting. Other regressors, the local linear and the shifted Nadaraya-Watson, were implemented within additive models--a computationally more expedient approach, better suited for low-density designs. Performances were benchmarked against the nonlinear method of smoothing splines. A linear reference point was provided by multilinear regression (MLR). Variable selection was explored using systematic combinations of different variables and combinations of principal components. For the data set examined, 47 inhibitors of dopamine beta-hydroxylase, the additive nonparametric regressors have greater predictive accuracy (as measured by the mean absolute error of the predictions or the Pearson correlation in cross-validation trails) than MLR. The use of principal components did not improve the performance of the nonparametric regressors over use of the original descriptors, since the original descriptors are not strongly correlated. It remains to be seen if the nonparametric regressors can be successfully coupled with better variable selection and dimensionality reduction in the context of high-dimensional QSARs.
NASA Technical Reports Server (NTRS)
Ioup, G. E.
1985-01-01
Appendix 5 of the Study of One- and Two-Dimensional Filtering and Deconvolution Algorithms for a Streaming Array Computer includes a resume of the professional background of the Principal Investigator on the project, lists of this publications and research papers, graduate thesis supervised, and grants received.
Visualizing Hyolaryngeal Mechanics in Swallowing Using Dynamic MRI
Pearson, William G.; Zumwalt, Ann C.
2013-01-01
Introduction Coordinates of anatomical landmarks are captured using dynamic MRI to explore whether a proposed two-sling mechanism underlies hyolaryngeal elevation in pharyngeal swallowing. A principal components analysis (PCA) is applied to coordinates to determine the covariant function of the proposed mechanism. Methods Dynamic MRI (dMRI) data were acquired from eleven healthy subjects during a repeated swallows task. Coordinates mapping the proposed mechanism are collected from each dynamic (frame) of a dynamic MRI swallowing series of a randomly selected subject in order to demonstrate shape changes in a single subject. Coordinates representing minimum and maximum hyolaryngeal elevation of all 11 subjects were also mapped to demonstrate shape changes of the system among all subjects. MophoJ software was used to perform PCA and determine vectors of shape change (eigenvectors) for elements of the two-sling mechanism of hyolaryngeal elevation. Results For both single subject and group PCAs, hyolaryngeal elevation accounted for the first principal component of variation. For the single subject PCA, the first principal component accounted for 81.5% of the variance. For the between subjects PCA, the first principal component accounted for 58.5% of the variance. Eigenvectors and shape changes associated with this first principal component are reported. Discussion Eigenvectors indicate that two-muscle slings and associated skeletal elements function as components of a covariant mechanism to elevate the hyolaryngeal complex. Morphological analysis is useful to model shape changes in the two-sling mechanism of hyolaryngeal elevation. PMID:25090608
Discriminative components of data.
Peltonen, Jaakko; Kaski, Samuel
2005-01-01
A simple probabilistic model is introduced to generalize classical linear discriminant analysis (LDA) in finding components that are informative of or relevant for data classes. The components maximize the predictability of the class distribution which is asymptotically equivalent to 1) maximizing mutual information with the classes, and 2) finding principal components in the so-called learning or Fisher metrics. The Fisher metric measures only distances that are relevant to the classes, that is, distances that cause changes in the class distribution. The components have applications in data exploration, visualization, and dimensionality reduction. In empirical experiments, the method outperformed, in addition to more classical methods, a Renyi entropy-based alternative while having essentially equivalent computational cost.
Evaluation of Deep Learning Representations of Spatial Storm Data
NASA Astrophysics Data System (ADS)
Gagne, D. J., II; Haupt, S. E.; Nychka, D. W.
2017-12-01
The spatial structure of a severe thunderstorm and its surrounding environment provide useful information about the potential for severe weather hazards, including tornadoes, hail, and high winds. Statistics computed over the area of a storm or from the pre-storm environment can provide descriptive information but fail to capture structural information. Because the storm environment is a complex, high-dimensional space, identifying methods to encode important spatial storm information in a low-dimensional form should aid analysis and prediction of storms by statistical and machine learning models. Principal component analysis (PCA), a more traditional approach, transforms high-dimensional data into a set of linearly uncorrelated, orthogonal components ordered by the amount of variance explained by each component. The burgeoning field of deep learning offers two potential approaches to this problem. Convolutional Neural Networks are a supervised learning method for transforming spatial data into a hierarchical set of feature maps that correspond with relevant combinations of spatial structures in the data. Generative Adversarial Networks (GANs) are an unsupervised deep learning model that uses two neural networks trained against each other to produce encoded representations of spatial data. These different spatial encoding methods were evaluated on the prediction of severe hail for a large set of storm patches extracted from the NCAR convection-allowing ensemble. Each storm patch contains information about storm structure and the near-storm environment. Logistic regression and random forest models were trained using the PCA and GAN encodings of the storm data and were compared against the predictions from a convolutional neural network. All methods showed skill over climatology at predicting the probability of severe hail. However, the verification scores among the methods were very similar and the predictions were highly correlated. Further evaluations are being performed to determine how the choice of input variables affects the results.
Numerical, analytical, experimental study of fluid dynamic forces in seals
NASA Technical Reports Server (NTRS)
Shapiro, William; Artiles, Antonio; Aggarwal, Bharat; Walowit, Jed; Athavale, Mahesh M.; Preskwas, Andrzej J.
1992-01-01
NASA/Lewis Research Center is sponsoring a program for providing computer codes for analyzing and designing turbomachinery seals for future aerospace and engine systems. The program is made up of three principal components: (1) the development of advanced three dimensional (3-D) computational fluid dynamics codes, (2) the production of simpler two dimensional (2-D) industrial codes, and (3) the development of a knowledge based system (KBS) that contains an expert system to assist in seal selection and design. The first task has been to concentrate on cylindrical geometries with straight, tapered, and stepped bores. Improvements have been made by adoption of a colocated grid formulation, incorporation of higher order, time accurate schemes for transient analysis and high order discretization schemes for spatial derivatives. This report describes the mathematical formulations and presents a variety of 2-D results, including labyrinth and brush seal flows. Extensions of 3-D are presently in progress.
Sufficient Forecasting Using Factor Models
Fan, Jianqing; Xue, Lingzhou; Yao, Jiawei
2017-01-01
We consider forecasting a single time series when there is a large number of predictors and a possible nonlinear effect. The dimensionality was first reduced via a high-dimensional (approximate) factor model implemented by the principal component analysis. Using the extracted factors, we develop a novel forecasting method called the sufficient forecasting, which provides a set of sufficient predictive indices, inferred from high-dimensional predictors, to deliver additional predictive power. The projected principal component analysis will be employed to enhance the accuracy of inferred factors when a semi-parametric (approximate) factor model is assumed. Our method is also applicable to cross-sectional sufficient regression using extracted factors. The connection between the sufficient forecasting and the deep learning architecture is explicitly stated. The sufficient forecasting correctly estimates projection indices of the underlying factors even in the presence of a nonparametric forecasting function. The proposed method extends the sufficient dimension reduction to high-dimensional regimes by condensing the cross-sectional information through factor models. We derive asymptotic properties for the estimate of the central subspace spanned by these projection directions as well as the estimates of the sufficient predictive indices. We further show that the natural method of running multiple regression of target on estimated factors yields a linear estimate that actually falls into this central subspace. Our method and theory allow the number of predictors to be larger than the number of observations. We finally demonstrate that the sufficient forecasting improves upon the linear forecasting in both simulation studies and an empirical study of forecasting macroeconomic variables. PMID:29731537
Demixed principal component analysis of neural population data.
Kobak, Dmitry; Brendel, Wieland; Constantinidis, Christos; Feierstein, Claudia E; Kepecs, Adam; Mainen, Zachary F; Qi, Xue-Lian; Romo, Ranulfo; Uchida, Naoshige; Machens, Christian K
2016-04-12
Neurons in higher cortical areas, such as the prefrontal cortex, are often tuned to a variety of sensory and motor variables, and are therefore said to display mixed selectivity. This complexity of single neuron responses can obscure what information these areas represent and how it is represented. Here we demonstrate the advantages of a new dimensionality reduction technique, demixed principal component analysis (dPCA), that decomposes population activity into a few components. In addition to systematically capturing the majority of the variance of the data, dPCA also exposes the dependence of the neural representation on task parameters such as stimuli, decisions, or rewards. To illustrate our method we reanalyze population data from four datasets comprising different species, different cortical areas and different experimental tasks. In each case, dPCA provides a concise way of visualizing the data that summarizes the task-dependent features of the population response in a single figure.
Noise-induced drift in two-dimensional anisotropic systems
NASA Astrophysics Data System (ADS)
Farago, Oded
2017-10-01
We study the isothermal Brownian dynamics of a particle in a system with spatially varying diffusivity. Due to the heterogeneity of the system, the particle's mean displacement does not vanish even if it does not experience any physical force. This phenomenon has been termed "noise-induced drift," and has been extensively studied for one-dimensional systems. Here, we examine the noise-induced drift in a two-dimensional anisotropic system, characterized by a symmetric diffusion tensor with unequal diagonal elements. A general expression for the mean displacement vector is derived and presented as a sum of two vectors, depicting two distinct drifting effects. The first vector describes the tendency of the particle to drift toward the high diffusivity side in each orthogonal principal diffusion direction. This is a generalization of the well-known expression for the noise-induced drift in one-dimensional systems. The second vector represents a novel drifting effect, not found in one-dimensional systems, originating from the spatial rotation in the directions of the principal axes. The validity of the derived expressions is verified by using Langevin dynamics simulations. As a specific example, we consider the relative diffusion of two transmembrane proteins, and demonstrate that the average distance between them increases at a surprisingly fast rate of several tens of micrometers per second.
Spatial and spectral analysis of corneal epithelium injury using hyperspectral images
NASA Astrophysics Data System (ADS)
Md Noor, Siti Salwa; Michael, Kaleena; Marshall, Stephen; Ren, Jinchang
2017-12-01
Eye assessment is essential in preventing blindness. Currently, the existing methods to assess corneal epithelium injury are complex and require expert knowledge. Hence, we have introduced a non-invasive technique using hyperspectral imaging (HSI) and an image analysis algorithm of corneal epithelium injury. Three groups of images were compared and analyzed, namely healthy eyes, injured eyes, and injured eyes with stain. Dimensionality reduction using principal component analysis (PCA) was applied to reduce massive data and redundancies. The first 10 principal components (PCs) were selected for further processing. The mean vector of 10 PCs with 45 pairs of all combinations was computed and sent to two classifiers. A quadratic Bayes normal classifier (QDC) and a support vector classifier (SVC) were used in this study to discriminate the eleven eyes into three groups. As a result, the combined classifier of QDC and SVC showed optimal performance with 2D PCA features (2DPCA-QDSVC) and was utilized to classify normal and abnormal tissues, using color image segmentation. The result was compared with human segmentation. The outcome showed that the proposed algorithm produced extremely promising results to assist the clinician in quantifying a cornea injury.
Maurer, Christian; Federolf, Peter; von Tscharner, Vinzenz; Stirling, Lisa; Nigg, Benno M
2012-05-01
Changes in gait kinematics have often been analyzed using pattern recognition methods such as principal component analysis (PCA). It is usually just the first few principal components that are analyzed, because they describe the main variability within a dataset and thus represent the main movement patterns. However, while subtle changes in gait pattern (for instance, due to different footwear) may not change main movement patterns, they may affect movements represented by higher principal components. This study was designed to test two hypotheses: (1) speed and gender differences can be observed in the first principal components, and (2) small interventions such as changing footwear change the gait characteristics of higher principal components. Kinematic changes due to different running conditions (speed - 3.1m/s and 4.9 m/s, gender, and footwear - control shoe and adidas MicroBounce shoe) were investigated by applying PCA and support vector machine (SVM) to a full-body reflective marker setup. Differences in speed changed the basic movement pattern, as was reflected by a change in the time-dependent coefficient derived from the first principal. Gender was differentiated by using the time-dependent coefficient derived from intermediate principal components. (Intermediate principal components are characterized by limb rotations of the thigh and shank.) Different shoe conditions were identified in higher principal components. This study showed that different interventions can be analyzed using a full-body kinematic approach. Within the well-defined vector space spanned by the data of all subjects, higher principal components should also be considered because these components show the differences that result from small interventions such as footwear changes. Crown Copyright © 2012. Published by Elsevier B.V. All rights reserved.
Foch, Eric; Milner, Clare E
2014-01-03
Iliotibial band syndrome (ITBS) is a common knee overuse injury among female runners. Atypical discrete trunk and lower extremity biomechanics during running may be associated with the etiology of ITBS. Examining discrete data points limits the interpretation of a waveform to a single value. Characterizing entire kinematic and kinetic waveforms may provide additional insight into biomechanical factors associated with ITBS. Therefore, the purpose of this cross-sectional investigation was to determine whether female runners with previous ITBS exhibited differences in kinematics and kinetics compared to controls using a principal components analysis (PCA) approach. Forty participants comprised two groups: previous ITBS and controls. Principal component scores were retained for the first three principal components and were analyzed using independent t-tests. The retained principal components accounted for 93-99% of the total variance within each waveform. Runners with previous ITBS exhibited low principal component one scores for frontal plane hip angle. Principal component one accounted for the overall magnitude in hip adduction which indicated that runners with previous ITBS assumed less hip adduction throughout stance. No differences in the remaining retained principal component scores for the waveforms were detected among groups. A smaller hip adduction angle throughout the stance phase of running may be a compensatory strategy to limit iliotibial band strain. This running strategy may have persisted after ITBS symptoms subsided. © 2013 Published by Elsevier Ltd.
2L-PCA: a two-level principal component analyzer for quantitative drug design and its applications.
Du, Qi-Shi; Wang, Shu-Qing; Xie, Neng-Zhong; Wang, Qing-Yan; Huang, Ri-Bo; Chou, Kuo-Chen
2017-09-19
A two-level principal component predictor (2L-PCA) was proposed based on the principal component analysis (PCA) approach. It can be used to quantitatively analyze various compounds and peptides about their functions or potentials to become useful drugs. One level is for dealing with the physicochemical properties of drug molecules, while the other level is for dealing with their structural fragments. The predictor has the self-learning and feedback features to automatically improve its accuracy. It is anticipated that 2L-PCA will become a very useful tool for timely providing various useful clues during the process of drug development.
Relationships between NIR spectra and sensory attributes of Thai commercial fish sauces.
Ritthiruangdej, Pitiporn; Suwonsichon, Thongchai
2007-07-01
Twenty Thai commercial fish sauces were characterized by sensory descriptive analysis and near-infrared (NIR) spectroscopy. The main objectives were i) to investigate the relationships between sensory attributes and NIR spectra of samples and ii) to characterize the sensory characteristics of fish sauces based on NIR data. A generic descriptive analysis with 12 trained panels was used to characterize the sensory attributes. These attributes consisted of 15 descriptors: brown color, 5 aromatics (sweet, caramelized, fermented, fishy, and musty), 4 tastes (sweet, salty, bitter, and umami), 3 aftertastes (sweet, salty and bitter) and 2 flavors (caramelized and fishy). The results showed that Thai fish sauce samples exhibited significant differences in all of sensory attribute values (p < 0.05). NIR transflectance spectra were obtained from 1100 to 2500 nm. Prior to investigation of the relationships between sensory attributes and NIR spectra, principal component analysis (PCA) was applied to reduce the dimensionality of the spectral data from 622 wavelengths to two uncorrelated components (NIR1 and NIR2) which explained 92 and 7% of the total variation, respectively. NIR1 was highly correlated with the wavelength regions of 1100 - 1544, 1774 - 2062, 2092 - 2308, and 2358 - 2440 nm, while NIR2 was highly correlated with the wavelength regions of 1742 - 1764, 2066 - 2088, and 2312 - 2354 nm. Subsequently, the relationships among these two components and all sensory attributes were also investigated by PCA. The results showed that the first three principal components (PCs) named as fishy flavor component (PC1), sweet component (PC2) and bitterness component (PC3), respectively, explained a total of 66.86% of the variation. NIR1 was mainly correlated to the sensory attributes of fishy aromatic, fishy flavor and sweet aftertaste on PC1. In addition, the PCA using only the factor loadings of NIR1 and NIR2 could be used to classify samples into three groups which showed high, medium and low degrees of fishy aromatic, fishy flavor and sweet aftertaste.
Onsager vortex formation in two-component Bose–Einstein condensates in two-dimensional traps
NASA Astrophysics Data System (ADS)
Han, Junsik; Tsubota, Makoto
2018-03-01
We study numerically the dynamics of quantized vortices in two-dimensional one-component and two-component Bose–Einstein condensates (BECs) trapped by a harmonic and box potentials. In two-component miscible BECs, we confirmed the tendency of the formation of Onsager vortices in both traps. The vortices in one component separate spatially from those in the other component, which comes from their intercomponent-coupling. We also discuss the decay of the number of vortices.
Maisuradze, Gia G; Leitner, David M
2007-05-15
Dihedral principal component analysis (dPCA) has recently been developed and shown to display complex features of the free energy landscape of a biomolecule that may be absent in the free energy landscape plotted in principal component space due to mixing of internal and overall rotational motion that can occur in principal component analysis (PCA) [Mu et al., Proteins: Struct Funct Bioinfo 2005;58:45-52]. Another difficulty in the implementation of PCA is sampling convergence, which we address here for both dPCA and PCA using a tetrapeptide as an example. We find that for both methods the sampling convergence can be reached over a similar time. Minima in the free energy landscape in the space of the two largest dihedral principal components often correspond to unique structures, though we also find some distinct minima to correspond to the same structure. 2007 Wiley-Liss, Inc.
Volumetric three-component velocimetry measurements of the turbulent flow around a Rushton turbine
NASA Astrophysics Data System (ADS)
Sharp, Kendra V.; Hill, David; Troolin, Daniel; Walters, Geoffrey; Lai, Wing
2010-01-01
Volumetric three-component velocimetry measurements have been taken of the flow field near a Rushton turbine in a stirred tank reactor. This particular flow field is highly unsteady and three-dimensional, and is characterized by a strong radial jet, large tank-scale ring vortices, and small-scale blade tip vortices. The experimental technique uses a single camera head with three apertures to obtain approximately 15,000 three-dimensional vectors in a cubic volume. These velocity data offer the most comprehensive view to date of this flow field, especially since they are acquired at three Reynolds numbers (15,000, 107,000, and 137,000). Mean velocity fields and turbulent kinetic energy quantities are calculated. The volumetric nature of the data enables tip vortex identification, vortex trajectory analysis, and calculation of vortex strength. Three identification methods for the vortices are compared based on: the calculation of circumferential vorticity; the calculation of local pressure minima via an eigenvalue approach; and the calculation of swirling strength again via an eigenvalue approach. The use of two-dimensional data and three-dimensional data is compared for vortex identification; a `swirl strength' criterion is less sensitive to completeness of the velocity gradient tensor and overall provides clearer identification of the tip vortices. The principal components of the strain rate tensor are also calculated for one Reynolds number case as these measures of stretching and compression have recently been associated with tip vortex characterization. Vortex trajectories and strength compare favorably with those in the literature. No clear dependence of trajectory on Reynolds number is deduced. The visualization of tip vortices up to 140° past blade passage in the highest Reynolds number case is notable and has not previously been shown.
Clustering high dimensional data using RIA
DOE Office of Scientific and Technical Information (OSTI.GOV)
Aziz, Nazrina
2015-05-15
Clustering may simply represent a convenient method for organizing a large data set so that it can easily be understood and information can efficiently be retrieved. However, identifying cluster in high dimensionality data sets is a difficult task because of the curse of dimensionality. Another challenge in clustering is some traditional functions cannot capture the pattern dissimilarity among objects. In this article, we used an alternative dissimilarity measurement called Robust Influence Angle (RIA) in the partitioning method. RIA is developed using eigenstructure of the covariance matrix and robust principal component score. We notice that, it can obtain cluster easily andmore » hence avoid the curse of dimensionality. It is also manage to cluster large data sets with mixed numeric and categorical value.« less
Automated Analysis, Classification, and Display of Waveforms
NASA Technical Reports Server (NTRS)
Kwan, Chiman; Xu, Roger; Mayhew, David; Zhang, Frank; Zide, Alan; Bonggren, Jeff
2004-01-01
A computer program partly automates the analysis, classification, and display of waveforms represented by digital samples. In the original application for which the program was developed, the raw waveform data to be analyzed by the program are acquired from space-shuttle auxiliary power units (APUs) at a sampling rate of 100 Hz. The program could also be modified for application to other waveforms -- for example, electrocardiograms. The program begins by performing principal-component analysis (PCA) of 50 normal-mode APU waveforms. Each waveform is segmented. A covariance matrix is formed by use of the segmented waveforms. Three eigenvectors corresponding to three principal components are calculated. To generate features, each waveform is then projected onto the eigenvectors. These features are displayed on a three-dimensional diagram, facilitating the visualization of the trend of APU operations.
Detection of ochratoxin A contamination in stored wheat using near-infrared hyperspectral imaging
NASA Astrophysics Data System (ADS)
Senthilkumar, T.; Jayas, D. S.; White, N. D. G.; Fields, P. G.; Gräfenhan, T.
2017-03-01
Near-infrared (NIR) hyperspectral imaging system was used to detect five concentration levels of ochratoxin A (OTA) in contaminated wheat kernels. The wheat kernels artificially inoculated with two different OTA producing Penicillium verrucosum strains, two different non-toxigenic P. verrucosum strains, and sterile control wheat kernels were subjected to NIR hyperspectral imaging. The acquired three-dimensional data were reshaped into readable two-dimensional data. Principal Component Analysis (PCA) was applied to the two dimensional data to identify the key wavelengths which had greater significance in detecting OTA contamination in wheat. Statistical and histogram features extracted at the key wavelengths were used in the linear, quadratic and Mahalanobis statistical discriminant models to differentiate between sterile control, five concentration levels of OTA contamination in wheat kernels, and five infection levels of non-OTA producing P. verrucosum inoculated wheat kernels. The classification models differentiated sterile control samples from OTA contaminated wheat kernels and non-OTA producing P. verrucosum inoculated wheat kernels with a 100% accuracy. The classification models also differentiated between five concentration levels of OTA contaminated wheat kernels and between five infection levels of non-OTA producing P. verrucosum inoculated wheat kernels with a correct classification of more than 98%. The non-OTA producing P. verrucosum inoculated wheat kernels and OTA contaminated wheat kernels subjected to hyperspectral imaging provided different spectral patterns.
Wanda, Elijah M M; Nyoni, Hlengilizwe; Mamba, Bhekie B; Msagati, Titus A M
2017-01-13
The ubiquitous occurrence of emerging micropollutants (EMPs) in water is an issue of growing environmental-health concern worldwide. However, there remains a paucity of data regarding their levels and occurrence in water. This study determined the occurrence of EMPs namely: carbamazepine (CBZ), galaxolide (HHCB), caffeine (CAF), tonalide (AHTN), 4-nonylphenol (NP), and bisphenol A (BPA) in water from Gauteng, Mpumalanga, and North West provinces, South Africa using comprehensive two-dimensional gas chromatography coupled to high resolution time-of-flight mass spectrometry (GCxGC-HRTOFMS). Kruskal-Wallis test and ANOVA were performed to determine temporal variations in occurrence of the EMPs. Principal component analysis (PCA) and Surfer Golden Graphics software for surface mapping were used to determine spatial variations in levels and occurrence of the EMPs. The mean levels ranged from 11.22 ± 18.8 ng/L for CAF to 158.49 ± 662 ng/L for HHCB. There was no evidence of statistically significant temporal variations in occurrence of EMPs in water. Nevertheless, their levels and occurrence vary spatially and are a function of two principal components (PCs, PC1 and PC2) which controlled 89.99% of the variance. BPA was the most widely distributed EMP, which was present in 62% of the water samples. The detected EMPs pose ecotoxicological risks in water samples, especially those from Mpumalanga province.
Wanda, Elijah M. M.; Nyoni, Hlengilizwe; Mamba, Bhekie B.; Msagati, Titus A. M.
2017-01-01
The ubiquitous occurrence of emerging micropollutants (EMPs) in water is an issue of growing environmental-health concern worldwide. However, there remains a paucity of data regarding their levels and occurrence in water. This study determined the occurrence of EMPs namely: carbamazepine (CBZ), galaxolide (HHCB), caffeine (CAF), tonalide (AHTN), 4-nonylphenol (NP), and bisphenol A (BPA) in water from Gauteng, Mpumalanga, and North West provinces, South Africa using comprehensive two-dimensional gas chromatography coupled to high resolution time-of-flight mass spectrometry (GCxGC-HRTOFMS). Kruskal-Wallis test and ANOVA were performed to determine temporal variations in occurrence of the EMPs. Principal component analysis (PCA) and Surfer Golden Graphics software for surface mapping were used to determine spatial variations in levels and occurrence of the EMPs. The mean levels ranged from 11.22 ± 18.8 ng/L for CAF to 158.49 ± 662 ng/L for HHCB. There was no evidence of statistically significant temporal variations in occurrence of EMPs in water. Nevertheless, their levels and occurrence vary spatially and are a function of two principal components (PCs, PC1 and PC2) which controlled 89.99% of the variance. BPA was the most widely distributed EMP, which was present in 62% of the water samples. The detected EMPs pose ecotoxicological risks in water samples, especially those from Mpumalanga province. PMID:28098799
NASA Astrophysics Data System (ADS)
Chaa, Mourad; Boukezzoula, Naceur-Eddine; Attia, Abdelouahab
2017-01-01
Two types of scores extracted from two-dimensional (2-D) and three-dimensional (3-D) palmprint for personal recognition systems are merged, introducing a local image descriptor for 2-D palmprint-based recognition systems, named bank of binarized statistical image features (B-BSIF). The main idea of B-BSIF is that the extracted histograms from the binarized statistical image features (BSIF) code images (the results of applying the different BSIF descriptor size with the length 12) are concatenated into one to produce a large feature vector. 3-D palmprint contains the depth information of the palm surface. The self-quotient image (SQI) algorithm is applied for reconstructing illumination-invariant 3-D palmprint images. To extract discriminative Gabor features from SQI images, Gabor wavelets are defined and used. Indeed, the dimensionality reduction methods have shown their ability in biometrics systems. Given this, a principal component analysis (PCA)+linear discriminant analysis (LDA) technique is employed. For the matching process, the cosine Mahalanobis distance is applied. Extensive experiments were conducted on a 2-D and 3-D palmprint database with 10,400 range images from 260 individuals. Then, a comparison was made between the proposed algorithm and other existing methods in the literature. Results clearly show that the proposed framework provides a higher correct recognition rate. Furthermore, the best results were obtained by merging the score of B-BSIF descriptor with the score of the SQI+Gabor wavelets+PCA+LDA method, yielding an equal error rate of 0.00% and a recognition rate of rank-1=100.00%.
Performance Evaluation of Multimodal Multifeature Authentication System Using KNN Classification.
Rajagopal, Gayathri; Palaniswamy, Ramamoorthy
2015-01-01
This research proposes a multimodal multifeature biometric system for human recognition using two traits, that is, palmprint and iris. The purpose of this research is to analyse integration of multimodal and multifeature biometric system using feature level fusion to achieve better performance. The main aim of the proposed system is to increase the recognition accuracy using feature level fusion. The features at the feature level fusion are raw biometric data which contains rich information when compared to decision and matching score level fusion. Hence information fused at the feature level is expected to obtain improved recognition accuracy. However, information fused at feature level has the problem of curse in dimensionality; here PCA (principal component analysis) is used to diminish the dimensionality of the feature sets as they are high dimensional. The proposed multimodal results were compared with other multimodal and monomodal approaches. Out of these comparisons, the multimodal multifeature palmprint iris fusion offers significant improvements in the accuracy of the suggested multimodal biometric system. The proposed algorithm is tested using created virtual multimodal database using UPOL iris database and PolyU palmprint database.
Performance Evaluation of Multimodal Multifeature Authentication System Using KNN Classification
Rajagopal, Gayathri; Palaniswamy, Ramamoorthy
2015-01-01
This research proposes a multimodal multifeature biometric system for human recognition using two traits, that is, palmprint and iris. The purpose of this research is to analyse integration of multimodal and multifeature biometric system using feature level fusion to achieve better performance. The main aim of the proposed system is to increase the recognition accuracy using feature level fusion. The features at the feature level fusion are raw biometric data which contains rich information when compared to decision and matching score level fusion. Hence information fused at the feature level is expected to obtain improved recognition accuracy. However, information fused at feature level has the problem of curse in dimensionality; here PCA (principal component analysis) is used to diminish the dimensionality of the feature sets as they are high dimensional. The proposed multimodal results were compared with other multimodal and monomodal approaches. Out of these comparisons, the multimodal multifeature palmprint iris fusion offers significant improvements in the accuracy of the suggested multimodal biometric system. The proposed algorithm is tested using created virtual multimodal database using UPOL iris database and PolyU palmprint database. PMID:26640813
Geographic variation in the black bear (Ursus americanus) in the eastern United States and Canada
Kennedy, M.L.; Kennedy, P.K.; Bogan, M.A.; Waits, J.L.
2002-01-01
The pattern of geographic variation in morphologic characters of the black bear (Ursus americanus) was assessed at 13 sites in the eastern United States and Canada. Thirty measurements from 206 males and 207 females were recorded to the nearest 0.01 mm using digital calipers and subjected to principal components analysis. A matrix of correlations among skull characters was computed, and the first 3 principal components were extracted. These accounted for 90.5% of the variation in the character set for males and 87.1% for females. Three-dimensional projection of localities onto principal components showed that, for males and females, largest individuals occurred in the more southern localities (e.g., males--Louisiana-Mississippi, eastern Texas; females--Louisiana-eastern Texas) and the smallest animals occurred in the northernmost locality (Quebec). Generally, bears were similar morphologically to those in nearby geographic areas. For males, correlations between morphologic variation and environmental factors indicated a significant relationship between size variation and mean January temperature, mean July temperature, mean annual precipitation, latitude, and actual evapotranspiration; for females, a significant relationship was observed between morphologic variation and mean annual temperature, mean January temperature, mean July temperature, latitude, and actual evapotranspiration. There was no significant correlation for either sex between environmental factors and projections onto components II and III.
Mackrill, J B; Jennings, P A; Cain, R
2013-01-01
Work on the perception of urban soundscapes has generated a number of perceptual models which are proposed as tools to test and evaluate soundscape interventions. However, despite the excessive sound levels and noise within hospital environments, perceptual models have not been developed for these spaces. To address this, a two-stage approach was developed by the authors to create such a model. First, semantics were obtained from listening evaluations which captured the feelings of individuals from hearing hospital sounds. Then, 30 participants rated a range of sound clips representative of a ward soundscape based on these semantics. Principal component analysis extracted a two-dimensional space representing an emotional-cognitive response. The framework enables soundscape interventions to be tested which may improve the perception of these hospital environments.
Genome Data Exploration Using Correspondence Analysis
Tekaia, Fredj
2016-01-01
Recent developments of sequencing technologies that allow the production of massive amounts of genomic and genotyping data have highlighted the need for synthetic data representation and pattern recognition methods that can mine and help discovering biologically meaningful knowledge included in such large data sets. Correspondence analysis (CA) is an exploratory descriptive method designed to analyze two-way data tables, including some measure of association between rows and columns. It constructs linear combinations of variables, known as factors. CA has been used for decades to study high-dimensional data, and remarkable inferences from large data tables were obtained by reducing the dimensionality to a few orthogonal factors that correspond to the largest amount of variability in the data. Herein, I review CA and highlight its use by considering examples in handling high-dimensional data that can be constructed from genomic and genetic studies. Examples in amino acid compositions of large sets of species (viruses, phages, yeast, and fungi) as well as an example related to pairwise shared orthologs in a set of yeast and fungal species, as obtained from their proteome comparisons, are considered. For the first time, results show striking segregations between yeasts and fungi as well as between viruses and phages. Distributions obtained from shared orthologs show clusters of yeast and fungal species corresponding to their phylogenetic relationships. A direct comparison with the principal component analysis method is discussed using a recently published example of genotyping data related to newly discovered traces of an ancient hominid that was compared to modern human populations in the search for ancestral similarities. CA offers more detailed results highlighting links between modern humans and the ancient hominid and their characterizations. Compared to the popular principal component analysis method, CA allows easier and more effective interpretation of results, particularly by the ability of relating individual patterns with their corresponding characteristic variables. PMID:27279736
DOE Office of Scientific and Technical Information (OSTI.GOV)
Zhang, Le; Timbie, Peter T.; Bunn, Emory F.
In this paper, we present a new Bayesian semi-blind approach for foreground removal in observations of the 21 cm signal measured by interferometers. The technique, which we call H i Expectation–Maximization Independent Component Analysis (HIEMICA), is an extension of the Independent Component Analysis technique developed for two-dimensional (2D) cosmic microwave background maps to three-dimensional (3D) 21 cm cosmological signals measured by interferometers. This technique provides a fully Bayesian inference of power spectra and maps and separates the foregrounds from the signal based on the diversity of their power spectra. Relying only on the statistical independence of the components, this approachmore » can jointly estimate the 3D power spectrum of the 21 cm signal, as well as the 2D angular power spectrum and the frequency dependence of each foreground component, without any prior assumptions about the foregrounds. This approach has been tested extensively by applying it to mock data from interferometric 21 cm intensity mapping observations under idealized assumptions of instrumental effects. We also discuss the impact when the noise properties are not known completely. As a first step toward solving the 21 cm power spectrum analysis problem, we compare the semi-blind HIEMICA technique to the commonly used Principal Component Analysis. Under the same idealized circumstances, the proposed technique provides significantly improved recovery of the power spectrum. This technique can be applied in a straightforward manner to all 21 cm interferometric observations, including epoch of reionization measurements, and can be extended to single-dish observations as well.« less
Farrell, John M.; Kapuscinski, Kevin L.; Underwood, Harold
2014-01-01
Radio telemetry of stocked muskellunge (n = 6) and wild northern pike (n = 6) was used to track late summer and fall movements from a common release point in a known shared nursery bay to test the hypothesis that age-1 northern pike and stocked muskellunge segregate and have different habitat affinities. Water depth, temperature, substrate and aquatic vegetation variables were estimated for each muskellunge (n = 103) and northern pike (n = 131) position and nested ANOVA comparisons by species indicated differences in habitat use. Muskellunge exhibited a greater displacement from the release point and used habitat in shallower water depths (mean = 0.85 m, SE = 0.10) than northern pike (mean = 1.45 m, SE = 0.08). Both principal components analysis (PCA) and principal components ordination (PCO) were used to interpret underlying gradients relative to fish positions in two-dimensional space. Our analysis indicated that a separation of age-1 northern pike and muskellunge occurred 7 d post-release. This first principal component explained 48% of the variation in habitat use. Northern pike locations were associated with deeper habitats that generally had softer silt substrates and dense submersed vegetation. Muskellunge locations post-acclimation showed greater association with shallower habitats containing firmer sandy and clay substrates and emergent vegetation. The observed differences in habitat use suggest that fine-scale ecological separation occurred between these stocked muskellunge and wild northern pike, but small sample sizes and potential for individual variation limit extension of these conclusions. Further research is needed to determine if these patterns exist between larger samples of fishes over a greater range of habitats.
Gaudreault, Nathaly; Mezghani, Neila; Turcot, Katia; Hagemeister, Nicola; Boivin, Karine; de Guise, Jacques A
2011-03-01
Interpreting gait data is challenging due to intersubject variability observed in the gait pattern of both normal and pathological populations. The objective of this study was to investigate the impact of using principal component analysis for grouping knee osteoarthritis (OA) patients' gait data in more homogeneous groups when studying the effect of a physiotherapy treatment. Three-dimensional (3D) knee kinematic and kinetic data were recorded during the gait of 29 participants diagnosed with knee OA before and after they received 12 weeks of physiotherapy treatment. Principal component analysis was applied to extract groups of knee flexion/extension, adduction/abduction and internal/external rotation angle and moment data. The treatment's effect on parameters of interest was assessed using paired t-tests performed before and after grouping the knee kinematic data. Increased quadriceps and hamstring strength was observed following treatment (P<0.05). Except for the knee flexion/extension angle, two different groups (G(1) and G(2)) were extracted from the angle and moment data. When pre- and post-treatment analyses were performed considering the groups, participants exhibiting a G(2) knee moment pattern demonstrated a greater first peak flexion moment, lower adduction moment impulse and smaller rotation angle range post-treatment (P<0.05). When pre- and post-treatment comparisons were performed without grouping, the data showed no treatment effect. The results of the present study suggest that the effect of physiotherapy on gait mechanics of knee osteoarthritis patients may be masked or underestimated if kinematic data are not separated into more homogeneous groups when performing pre- and post-treatment comparisons. Copyright © 2010 Elsevier Ltd. All rights reserved.
ASCS online fault detection and isolation based on an improved MPCA
NASA Astrophysics Data System (ADS)
Peng, Jianxin; Liu, Haiou; Hu, Yuhui; Xi, Junqiang; Chen, Huiyan
2014-09-01
Multi-way principal component analysis (MPCA) has received considerable attention and been widely used in process monitoring. A traditional MPCA algorithm unfolds multiple batches of historical data into a two-dimensional matrix and cut the matrix along the time axis to form subspaces. However, low efficiency of subspaces and difficult fault isolation are the common disadvantages for the principal component model. This paper presents a new subspace construction method based on kernel density estimation function that can effectively reduce the storage amount of the subspace information. The MPCA model and the knowledge base are built based on the new subspace. Then, fault detection and isolation with the squared prediction error (SPE) statistic and the Hotelling ( T 2) statistic are also realized in process monitoring. When a fault occurs, fault isolation based on the SPE statistic is achieved by residual contribution analysis of different variables. For fault isolation of subspace based on the T 2 statistic, the relationship between the statistic indicator and state variables is constructed, and the constraint conditions are presented to check the validity of fault isolation. Then, to improve the robustness of fault isolation to unexpected disturbances, the statistic method is adopted to set the relation between single subspace and multiple subspaces to increase the corrective rate of fault isolation. Finally fault detection and isolation based on the improved MPCA is used to monitor the automatic shift control system (ASCS) to prove the correctness and effectiveness of the algorithm. The research proposes a new subspace construction method to reduce the required storage capacity and to prove the robustness of the principal component model, and sets the relationship between the state variables and fault detection indicators for fault isolation.
A stochastic model for correlated protein motions
NASA Astrophysics Data System (ADS)
Karain, Wael I.; Qaraeen, Nael I.; Ajarmah, Basem
2006-06-01
A one-dimensional Langevin-type stochastic difference equation is used to find the deterministic and Gaussian contributions of time series representing the projections of a Bovine Pancreatic Trypsin Inhibitor (BPTI) protein molecular dynamics simulation along different eigenvector directions determined using principal component analysis. The deterministic part shows a distinct nonlinear behavior only for eigenvectors contributing significantly to the collective protein motion.
ERIC Educational Resources Information Center
Fernandez-Sainz, A.; García-Merino, J. D.; Urionabarrenetxea, S.
2016-01-01
This paper seeks to discover whether the performance of university students has improved in the wake of the changes in higher education introduced by the Bologna Declaration of 1999 and the construction of the European Higher Education Area. A principal component analysis is used to construct a multi-dimensional performance variable called the…
The Analysis of Dimensionality Reduction Techniques in Cryptographic Object Code Classification
DOE Office of Scientific and Technical Information (OSTI.GOV)
Jason L. Wright; Milos Manic
2010-05-01
This paper compares the application of three different dimension reduction techniques to the problem of locating cryptography in compiled object code. A simple classi?er is used to compare dimension reduction via sorted covariance, principal component analysis, and correlation-based feature subset selection. The analysis concentrates on the classi?cation accuracy as the number of dimensions is increased.
Kinematics of swimming of the manta ray: three-dimensional analysis of open-water maneuverability.
Fish, Frank E; Kolpas, Allison; Crossett, Andrew; Dudas, Michael A; Moored, Keith W; Bart-Smith, Hilary
2018-03-22
For aquatic animals, turning maneuvers represent a locomotor activity that may not be confined to a single coordinate plane, making analysis difficult, particularly in the field. To measure turning performance in a three-dimensional space for the manta ray ( Mobula birostris ), a large open-water swimmer, scaled stereo video recordings were collected. Movements of the cephalic lobes, eye and tail base were tracked to obtain three-dimensional coordinates. A mathematical analysis was performed on the coordinate data to calculate the turning rate and curvature (1/turning radius) as a function of time by numerically estimating the derivative of manta trajectories through three-dimensional space. Principal component analysis was used to project the three-dimensional trajectory onto the two-dimensional turn. Smoothing splines were applied to these turns. These are flexible models that minimize a cost function with a parameter controlling the balance between data fidelity and regularity of the derivative. Data for 30 sequences of rays performing slow, steady turns showed the highest 20% of values for the turning rate and smallest 20% of turn radii were 42.65±16.66 deg s -1 and 2.05±1.26 m, respectively. Such turning maneuvers fall within the range of performance exhibited by swimmers with rigid bodies. © 2018. Published by The Company of Biologists Ltd.
Maslia, M.L.; Randolph, R.B.
1986-01-01
The theory of anisotropic aquifer hydraulic properties and a computer program, written in Fortran 77, developed to compute the components of the anisotropic transmissivity tensor of two-dimensional groundwater flow are described. To determine the tensor components using one pumping well and three observation wells, the type-curve and straight-line approximation methods are developed. These methods are based on the equation of drawdown developed for two-dimensional nonsteady flow in an infinite anisotropic aquifer. To determine tensor components using more than three observation wells, a weighted least squares optimization procedure is described for use with the type-curve and straight-line approximation methods. The computer program described in this report allows the type-curve, straight-line approximation, and weighted least squares optimization methods to be used in conjunction with data from observation and pumping wells. Three example applications using the computer program and field data gathered during geohydrologic investigations at a site near Dawsonville, Georgia , are provided to illustrate the use of the computer program. The example applications demonstrate the use of the type-curve method using three observation wells, the weighted least squares optimization method using eight observation wells and equal weighting, and the weighted least squares optimization method using eight observation wells and unequal weighting. Results obtained using the computer program indicate major transmissivity in the range of 347-296 sq ft/day, minor transmissivity in the range of 139-99 sq ft/day, aquifer anisotropy in the range of 3.54 to 2.14, principal direction of flow in the range of N. 45.9 degrees E. to N. 58.7 degrees E., and storage coefficient in the range of 0.0063 to 0.0037. The numerical results are in good agreement with field data gathered on the weathered crystalline rocks underlying the investigation site. Supplemental material provides definitions of variables, data requirements and corresponding formats, input data and output results for the example applications, and a listing of the Fortran 77 computer code. (Author 's abstract)
Defining a Tri-Dimensional Role for Leadership in Further Education Colleges
ERIC Educational Resources Information Center
Lambert, Steve
2013-01-01
This article presents a review of current leadership practices of principals in further education colleges and suggests that principalship is more than a two-dimensional functional model comprising internal or externally focused activities. During the past 20 years further education leadership has become more demanding, with greater accountability…
Surface representations of two- and three-dimensional fluid flow topology
NASA Technical Reports Server (NTRS)
Helman, James L.; Hesselink, Lambertus
1990-01-01
We discuss our work using critical point analysis to generate representations of the vector field topology of numerical flow data sets. Critical points are located and characterized in a two-dimensional domain, which may be either a two-dimensional flow field or the tangential velocity field near a three-dimensional body. Tangent curves are then integrated out along the principal directions of certain classes of critical points. The points and curves are linked to form a skeleton representing the two-dimensional vector field topology. When generated from the tangential velocity field near a body in a three-dimensional flow, the skeleton includes the critical points and curves which provide a basis for analyzing the three-dimensional structure of the flow separation. The points along the separation curves in the skeleton are used to start tangent curve integrations to generate surfaces representing the topology of the associated flow separations.
Considering Horn's Parallel Analysis from a Random Matrix Theory Point of View.
Saccenti, Edoardo; Timmerman, Marieke E
2017-03-01
Horn's parallel analysis is a widely used method for assessing the number of principal components and common factors. We discuss the theoretical foundations of parallel analysis for principal components based on a covariance matrix by making use of arguments from random matrix theory. In particular, we show that (i) for the first component, parallel analysis is an inferential method equivalent to the Tracy-Widom test, (ii) its use to test high-order eigenvalues is equivalent to the use of the joint distribution of the eigenvalues, and thus should be discouraged, and (iii) a formal test for higher-order components can be obtained based on a Tracy-Widom approximation. We illustrate the performance of the two testing procedures using simulated data generated under both a principal component model and a common factors model. For the principal component model, the Tracy-Widom test performs consistently in all conditions, while parallel analysis shows unpredictable behavior for higher-order components. For the common factor model, including major and minor factors, both procedures are heuristic approaches, with variable performance. We conclude that the Tracy-Widom procedure is preferred over parallel analysis for statistically testing the number of principal components based on a covariance matrix.
PCA leverage: outlier detection for high-dimensional functional magnetic resonance imaging data.
Mejia, Amanda F; Nebel, Mary Beth; Eloyan, Ani; Caffo, Brian; Lindquist, Martin A
2017-07-01
Outlier detection for high-dimensional (HD) data is a popular topic in modern statistical research. However, one source of HD data that has received relatively little attention is functional magnetic resonance images (fMRI), which consists of hundreds of thousands of measurements sampled at hundreds of time points. At a time when the availability of fMRI data is rapidly growing-primarily through large, publicly available grassroots datasets-automated quality control and outlier detection methods are greatly needed. We propose principal components analysis (PCA) leverage and demonstrate how it can be used to identify outlying time points in an fMRI run. Furthermore, PCA leverage is a measure of the influence of each observation on the estimation of principal components, which are often of interest in fMRI data. We also propose an alternative measure, PCA robust distance, which is less sensitive to outliers and has controllable statistical properties. The proposed methods are validated through simulation studies and are shown to be highly accurate. We also conduct a reliability study using resting-state fMRI data from the Autism Brain Imaging Data Exchange and find that removal of outliers using the proposed methods results in more reliable estimation of subject-level resting-state networks using independent components analysis. © The Author 2017. Published by Oxford University Press. All rights reserved. For permissions, please e-mail: journals.permissions@oup.com.
Matsen IV, Frederick A.; Evans, Steven N.
2013-01-01
Principal components analysis (PCA) and hierarchical clustering are two of the most heavily used techniques for analyzing the differences between nucleic acid sequence samples taken from a given environment. They have led to many insights regarding the structure of microbial communities. We have developed two new complementary methods that leverage how this microbial community data sits on a phylogenetic tree. Edge principal components analysis enables the detection of important differences between samples that contain closely related taxa. Each principal component axis is a collection of signed weights on the edges of the phylogenetic tree, and these weights are easily visualized by a suitable thickening and coloring of the edges. Squash clustering outputs a (rooted) clustering tree in which each internal node corresponds to an appropriate “average” of the original samples at the leaves below the node. Moreover, the length of an edge is a suitably defined distance between the averaged samples associated with the two incident nodes, rather than the less interpretable average of distances produced by UPGMA, the most widely used hierarchical clustering method in this context. We present these methods and illustrate their use with data from the human microbiome. PMID:23505415
Demixed principal component analysis of neural population data
Kobak, Dmitry; Brendel, Wieland; Constantinidis, Christos; Feierstein, Claudia E; Kepecs, Adam; Mainen, Zachary F; Qi, Xue-Lian; Romo, Ranulfo; Uchida, Naoshige; Machens, Christian K
2016-01-01
Neurons in higher cortical areas, such as the prefrontal cortex, are often tuned to a variety of sensory and motor variables, and are therefore said to display mixed selectivity. This complexity of single neuron responses can obscure what information these areas represent and how it is represented. Here we demonstrate the advantages of a new dimensionality reduction technique, demixed principal component analysis (dPCA), that decomposes population activity into a few components. In addition to systematically capturing the majority of the variance of the data, dPCA also exposes the dependence of the neural representation on task parameters such as stimuli, decisions, or rewards. To illustrate our method we reanalyze population data from four datasets comprising different species, different cortical areas and different experimental tasks. In each case, dPCA provides a concise way of visualizing the data that summarizes the task-dependent features of the population response in a single figure. DOI: http://dx.doi.org/10.7554/eLife.10989.001 PMID:27067378
Joint principal trend analysis for longitudinal high-dimensional data.
Zhang, Yuping; Ouyang, Zhengqing
2018-06-01
We consider a research scenario motivated by integrating multiple sources of information for better knowledge discovery in diverse dynamic biological processes. Given two longitudinal high-dimensional datasets for a group of subjects, we want to extract shared latent trends and identify relevant features. To solve this problem, we present a new statistical method named as joint principal trend analysis (JPTA). We demonstrate the utility of JPTA through simulations and applications to gene expression data of the mammalian cell cycle and longitudinal transcriptional profiling data in response to influenza viral infections. © 2017, The International Biometric Society.
Multivariate Analysis of Solar Spectral Irradiance Measurements
NASA Technical Reports Server (NTRS)
Pilewskie, P.; Rabbette, M.
2001-01-01
Principal component analysis is used to characterize approximately 7000 downwelling solar irradiance spectra retrieved at the Southern Great Plains site during an Atmospheric Radiation Measurement (ARM) shortwave intensive operating period. This analysis technique has proven to be very effective in reducing a large set of variables into a much smaller set of independent variables while retaining the information content. It is used to determine the minimum number of parameters necessary to characterize atmospheric spectral irradiance or the dimensionality of atmospheric variability. It was found that well over 99% of the spectral information was contained in the first six mutually orthogonal linear combinations of the observed variables (flux at various wavelengths). Rotation of the principal components was effective in separating various components by their independent physical influences. The majority of the variability in the downwelling solar irradiance (380-1000 nm) was explained by the following fundamental atmospheric parameters (in order of their importance): cloud scattering, water vapor absorption, molecular scattering, and ozone absorption. In contrast to what has been proposed as a resolution to a clear-sky absorption anomaly, no unexpected gaseous absorption signature was found in any of the significant components.
Describing temperament in an ungulate: a multidimensional approach.
Graunke, Katharina L; Nürnberg, Gerd; Repsilber, Dirk; Puppe, Birger; Langbein, Jan
2013-01-01
Studies on animal temperament have often described temperament using a one-dimensional scale, whereas theoretical framework has recently suggested two or more dimensions using terms like "valence" or "arousal" to describe these dimensions. Yet, the valence or assessment of a situation is highly individual. The aim of this study was to provide support for the multidimensional framework with experimental data originating from an economically important species (Bos taurus). We tested 361 calves at 90 days post natum (dpn) in a novel-object test. Using a principal component analysis (PCA), we condensed numerous behaviours into fewer variables to describe temperament and correlated these variables with simultaneously measured heart rate variability (HRV) data. The PCA resulted in two behavioural dimensions (principal components, PC): novel-object-related (PC 1) and exploration-activity-related (PC 2). These PCs explained 58% of the variability in our data. The animals were distributed evenly within the two behavioural dimensions independent of their sex. Calves with different scores in these PCs differed significantly in HRV, and thus in the autonomous nervous system's activity. Based on these combined behavioural and physiological data we described four distinct temperament types resulting from two behavioural dimensions: "neophobic/fearful--alert", "interested--stressed", "subdued/uninterested--calm", and "neoophilic/outgoing--alert". Additionally, 38 calves were tested at 90 and 197 dpn. Using the same PCA-model, they correlated significantly in PC 1 and tended to correlate in PC 2 between the two test ages. Of these calves, 42% expressed a similar behaviour pattern in both dimensions and 47% in one. No differences in temperament scores were found between sexes or breeds. In conclusion, we described distinct temperament types in calves based on behavioural and physiological measures emphasising the benefits of a multidimensional approach.
Contact- and distance-based principal component analysis of protein dynamics.
Ernst, Matthias; Sittel, Florian; Stock, Gerhard
2015-12-28
To interpret molecular dynamics simulations of complex systems, systematic dimensionality reduction methods such as principal component analysis (PCA) represent a well-established and popular approach. Apart from Cartesian coordinates, internal coordinates, e.g., backbone dihedral angles or various kinds of distances, may be used as input data in a PCA. Adopting two well-known model problems, folding of villin headpiece and the functional dynamics of BPTI, a systematic study of PCA using distance-based measures is presented which employs distances between Cα-atoms as well as distances between inter-residue contacts including side chains. While this approach seems prohibitive for larger systems due to the quadratic scaling of the number of distances with the size of the molecule, it is shown that it is sufficient (and sometimes even better) to include only relatively few selected distances in the analysis. The quality of the PCA is assessed by considering the resolution of the resulting free energy landscape (to identify metastable conformational states and barriers) and the decay behavior of the corresponding autocorrelation functions (to test the time scale separation of the PCA). By comparing results obtained with distance-based, dihedral angle, and Cartesian coordinates, the study shows that the choice of input variables may drastically influence the outcome of a PCA.
Contact- and distance-based principal component analysis of protein dynamics
NASA Astrophysics Data System (ADS)
Ernst, Matthias; Sittel, Florian; Stock, Gerhard
2015-12-01
To interpret molecular dynamics simulations of complex systems, systematic dimensionality reduction methods such as principal component analysis (PCA) represent a well-established and popular approach. Apart from Cartesian coordinates, internal coordinates, e.g., backbone dihedral angles or various kinds of distances, may be used as input data in a PCA. Adopting two well-known model problems, folding of villin headpiece and the functional dynamics of BPTI, a systematic study of PCA using distance-based measures is presented which employs distances between Cα-atoms as well as distances between inter-residue contacts including side chains. While this approach seems prohibitive for larger systems due to the quadratic scaling of the number of distances with the size of the molecule, it is shown that it is sufficient (and sometimes even better) to include only relatively few selected distances in the analysis. The quality of the PCA is assessed by considering the resolution of the resulting free energy landscape (to identify metastable conformational states and barriers) and the decay behavior of the corresponding autocorrelation functions (to test the time scale separation of the PCA). By comparing results obtained with distance-based, dihedral angle, and Cartesian coordinates, the study shows that the choice of input variables may drastically influence the outcome of a PCA.
Systematic comparison of the behaviors produced by computational models of epileptic neocortex.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Warlaumont, A. S.; Lee, H. C.; Benayoun, M.
2010-12-01
Two existing models of brain dynamics in epilepsy, one detailed (i.e., realistic) and one abstract (i.e., simplified) are compared in terms of behavioral range and match to in vitro mouse recordings. A new method is introduced for comparing across computational models that may have very different forms. First, high-level metrics were extracted from model and in vitro output time series. A principal components analysis was then performed over these metrics to obtain a reduced set of derived features. These features define a low-dimensional behavior space in which quantitative measures of behavioral range and degree of match to real data canmore » be obtained. The detailed and abstract models and the mouse recordings overlapped considerably in behavior space. Both the range of behaviors and similarity to mouse data were similar between the detailed and abstract models. When no high-level metrics were used and principal components analysis was computed over raw time series, the models overlapped minimally with the mouse recordings. The method introduced here is suitable for comparing across different kinds of model data and across real brain recordings. It appears that, despite differences in form and computational expense, detailed and abstract models do not necessarily differ in their behaviors.« less
Semantic word impressions expressed by hue.
Shinomori, Keizo; Komatsu, Honami
2018-04-01
We investigated the possibility of whether impressions of semantic words showing complex concepts could be stably expressed by hues. Using a paired comparison method, we asked ten subjects to select from a pair of hues the one that more suitably matched a word impression. We employed nine Japanese semantic words and used twelve hues from vivid tones in the practical color coordinate system. As examples of the results, for the word "vigorous" the most frequently selected color was yellow and the least selected was blue to purple; for "tranquil" the most selected was yellow to green and the least selected was red. Principal component analysis of the selection data indicated that the cumulative contribution rate of the first two components was 94.6%, and in the two-dimensional space of the components, all hues were distributed as a hue-circle shape. In addition, comparison with additional data of color impressions measured by a semantic differential method suggested that most semantic word impressions can be stably expressed by hue, but the impression of some words, such as "magnificent" cannot. These results suggest that semantic word impression can be expressed reasonably well by color, and that hues are treated as impressions from the hue circle, not from color categories.
An application of principal component analysis to the clavicle and clavicle fixation devices.
Daruwalla, Zubin J; Courtis, Patrick; Fitzpatrick, Clare; Fitzpatrick, David; Mullett, Hannan
2010-03-26
Principal component analysis (PCA) enables the building of statistical shape models of bones and joints. This has been used in conjunction with computer assisted surgery in the past. However, PCA of the clavicle has not been performed. Using PCA, we present a novel method that examines the major modes of size and three-dimensional shape variation in male and female clavicles and suggests a method of grouping the clavicle into size and shape categories. Twenty-one high-resolution computerized tomography scans of the clavicle were reconstructed and analyzed using a specifically developed statistical software package. After performing statistical shape analysis, PCA was applied to study the factors that account for anatomical variation. The first principal component representing size accounted for 70.5 percent of anatomical variation. The addition of a further three principal components accounted for almost 87 percent. Using statistical shape analysis, clavicles in males have a greater lateral depth and are longer, wider and thicker than in females. However, the sternal angle in females is larger than in males. PCA confirmed these differences between genders but also noted that men exhibit greater variance and classified clavicles into five morphological groups. This unique approach is the first that standardizes a clavicular orientation. It provides information that is useful to both, the biomedical engineer and clinician. Other applications include implant design with regard to modifying current or designing future clavicle fixation devices. Our findings support the need for further development of clavicle fixation devices and the questioning of whether gender-specific devices are necessary.
Sullivan, Karen A; Lurie, Janine K
2017-01-01
The study examined the component structure of the Neurobehavioral Symptom Inventory (NSI) under five different models. The evaluated models comprised the full NSI (NSI-22) and the NSI-20 (NSI minus two orphan items). A civilian nonclinical sample was used. The 575 volunteers were predominantly university students who screened negative for mild TBI. The study design was cross-sectional, with questionnaires administered online. The main measure was the Neurobehavioral Symptom Inventory. Subscale, total and embedded validity scores were derived (the Validity-10, the LOW6, and the NIM5). In both models, the principal components analysis yielded two intercorrelated components (psychological and somatic/sensory) with acceptable internal consistency (alphas > 0.80). In this civilian nonclinical sample, the NSI had two underlying components. These components represent psychological and somatic/sensory neurobehavioral symptoms.
Comparing Networks from a Data Analysis Perspective
NASA Astrophysics Data System (ADS)
Li, Wei; Yang, Jing-Yu
To probe network characteristics, two predominant ways of network comparison are global property statistics and subgraph enumeration. However, they suffer from limited information and exhaustible computing. Here, we present an approach to compare networks from the perspective of data analysis. Initially, the approach projects each node of original network as a high-dimensional data point, and the network is seen as clouds of data points. Then the dispersion information of the principal component analysis (PCA) projection of the generated data clouds can be used to distinguish networks. We applied this node projection method to the yeast protein-protein interaction networks and the Internet Autonomous System networks, two types of networks with several similar higher properties. The method can efficiently distinguish one from the other. The identical result of different datasets from independent sources also indicated that the method is a robust and universal framework.
ERIC Educational Resources Information Center
Hendrix, Dean
2010-01-01
This study analyzed 2005-2006 Web of Science bibliometric data from institutions belonging to the Association of Research Libraries (ARL) and corresponding ARL statistics to find any associations between indicators from the two data sets. Principal components analysis on 36 variables from 103 universities revealed obvious associations between…
NASA Astrophysics Data System (ADS)
Khaleghi, Morteza; Lu, Weina; Dobrev, Ivo; Cheng, Jeffrey Tao; Furlong, Cosme; Rosowski, John J.
2013-10-01
Acoustically induced vibrations of the tympanic membrane (TM) play a primary role in the hearing process, in that these motions are the initial mechanical response of the ear to airborne sound. Characterization of the shape and three-dimensional (3-D) displacement patterns of the TM is a crucial step to a better understanding of the complicated mechanics of sound reception by the ear. Sound-induced 3-D displacements of the TM are estimated from shape and one-dimensional displacements measured in cadaveric chinchillas using a lensless dual-wavelength digital holography system (DWDHS). The DWDHS consists of laser delivery, optical head, and computing platform subsystems. Shape measurements are performed in double-exposure mode with the use of two wavelengths of a tunable laser, while nanometer-scale displacements are measured along a single sensitivity direction with a constant wavelength. Taking into consideration the geometrical and dimensional constrains imposed by the anatomy of the TM, we combine principles of thin-shell theory together with displacement measurements along a single sensitivity vector and TM surface shape to extract the three principal components of displacement in the full-field-of-view. We test, validate, and identify limitations of this approach via the application of finite element method to artificial geometries.
Locally linear embedding: dimension reduction of massive protostellar spectra
NASA Astrophysics Data System (ADS)
Ward, J. L.; Lumsden, S. L.
2016-09-01
We present the results of the application of locally linear embedding (LLE) to reduce the dimensionality of dereddened and continuum subtracted near-infrared spectra using a combination of models and real spectra of massive protostars selected from the Red MSX Source survey data base. A brief comparison is also made with two other dimension reduction techniques; principal component analysis (PCA) and Isomap using the same set of spectra as well as a more advanced form of LLE, Hessian locally linear embedding. We find that whilst LLE certainly has its limitations, it significantly outperforms both PCA and Isomap in classification of spectra based on the presence/absence of emission lines and provides a valuable tool for classification and analysis of large spectral data sets.
Patwary, Nurmohammed; Preza, Chrysanthe
2015-01-01
A depth-variant (DV) image restoration algorithm for wide field fluorescence microscopy, using an orthonormal basis decomposition of DV point-spread functions (PSFs), is investigated in this study. The efficient PSF representation is based on a previously developed principal component analysis (PCA), which is computationally intensive. We present an approach developed to reduce the number of DV PSFs required for the PCA computation, thereby making the PCA-based approach computationally tractable for thick samples. Restoration results from both synthetic and experimental images show consistency and that the proposed algorithm addresses efficiently depth-induced aberration using a small number of principal components. Comparison of the PCA-based algorithm with a previously-developed strata-based DV restoration algorithm demonstrates that the proposed method improves performance by 50% in terms of accuracy and simultaneously reduces the processing time by 64% using comparable computational resources. PMID:26504634
Personal authentication through dorsal hand vein patterns
NASA Astrophysics Data System (ADS)
Hsu, Chih-Bin; Hao, Shu-Sheng; Lee, Jen-Chun
2011-08-01
Biometric identification is an emerging technology that can solve security problems in our networked society. A reliable and robust personal verification approach using dorsal hand vein patterns is proposed in this paper. The characteristic of the approach needs less computational and memory requirements and has a higher recognition accuracy. In our work, the near-infrared charge-coupled device (CCD) camera is adopted as an input device for capturing dorsal hand vein images, it has the advantages of the low-cost and noncontact imaging. In the proposed approach, two finger-peaks are automatically selected as the datum points to define the region of interest (ROI) in the dorsal hand vein images. The modified two-directional two-dimensional principal component analysis, which performs an alternate two-dimensional PCA (2DPCA) in the column direction of images in the 2DPCA subspace, is proposed to exploit the correlation of vein features inside the ROI between images. The major advantage of the proposed method is that it requires fewer coefficients for efficient dorsal hand vein image representation and recognition. The experimental results on our large dorsal hand vein database show that the presented schema achieves promising performance (false reject rate: 0.97% and false acceptance rate: 0.05%) and is feasible for dorsal hand vein recognition.
Principal component analysis of molecular dynamics: On the use of Cartesian vs. internal coordinates
NASA Astrophysics Data System (ADS)
Sittel, Florian; Jain, Abhinav; Stock, Gerhard
2014-07-01
Principal component analysis of molecular dynamics simulations is a popular method to account for the essential dynamics of the system on a low-dimensional free energy landscape. Using Cartesian coordinates, first the translation and overall rotation need to be removed from the trajectory. Since the rotation depends via the moment of inertia on the molecule's structure, this separation is only straightforward for relatively rigid systems. Adopting millisecond molecular dynamics simulations of the folding of villin headpiece and the functional dynamics of BPTI provided by D. E. Shaw Research, it is demonstrated via a comparison of local and global rotational fitting that the structural dynamics of flexible molecules necessarily results in a mixing of overall and internal motion. Even for the small-amplitude functional motion of BPTI, the conformational distribution obtained from a Cartesian principal component analysis therefore reflects to some extend the dominant overall motion rather than the much smaller internal motion of the protein. Internal coordinates such as backbone dihedral angles, on the other hand, are found to yield correct and well-resolved energy landscapes for both examples. The virtues and shortcomings of the choice of various fitting schemes and coordinate sets as well as the generality of these results are discussed in some detail.
Sittel, Florian; Jain, Abhinav; Stock, Gerhard
2014-07-07
Principal component analysis of molecular dynamics simulations is a popular method to account for the essential dynamics of the system on a low-dimensional free energy landscape. Using Cartesian coordinates, first the translation and overall rotation need to be removed from the trajectory. Since the rotation depends via the moment of inertia on the molecule's structure, this separation is only straightforward for relatively rigid systems. Adopting millisecond molecular dynamics simulations of the folding of villin headpiece and the functional dynamics of BPTI provided by D. E. Shaw Research, it is demonstrated via a comparison of local and global rotational fitting that the structural dynamics of flexible molecules necessarily results in a mixing of overall and internal motion. Even for the small-amplitude functional motion of BPTI, the conformational distribution obtained from a Cartesian principal component analysis therefore reflects to some extend the dominant overall motion rather than the much smaller internal motion of the protein. Internal coordinates such as backbone dihedral angles, on the other hand, are found to yield correct and well-resolved energy landscapes for both examples. The virtues and shortcomings of the choice of various fitting schemes and coordinate sets as well as the generality of these results are discussed in some detail.
Preliminary Comparisons of the Information Content and Utility of TM Versus MSS Data
NASA Technical Reports Server (NTRS)
Markham, B. L.
1984-01-01
Comparisons were made between subscenes from the first TM scene acquired of the Washington, D.C. area and a MSS scene acquired approximately one year earlier. Three types of analyses were conducted to compare TM and MSS data: a water body analysis, a principal components analysis and a spectral clustering analysis. The water body analysis compared the capability of the TM to the MSS for detecting small uniform targets. Of the 59 ponds located on aerial photographs 34 (58%) were detected by the TM with six commission errors (15%) and 13 (22%) were detected by the MSS with three commission errors (19%). The smallest water body detected by the TM was 16 meters; the smallest detected by the MSS was 40 meters. For the principal components analysis, means and covariance matrices were calculated for each subscene, and principal components images generated and characterized. In the spectral clustering comparison each scene was independently clustered and the clusters were assigned to informational classes. The preliminary comparison indicated that TM data provides enhancements over MSS in terms of (1) small target detection and (2) data dimensionality (even with 4-band data). The extra dimension, partially resultant from TM band 1, appears useful for built-up/non-built-up area separation.
Kalegowda, Yogesh; Harmer, Sarah L
2013-01-08
Artificial neural network (ANN) and a hybrid principal component analysis-artificial neural network (PCA-ANN) classifiers have been successfully implemented for classification of static time-of-flight secondary ion mass spectrometry (ToF-SIMS) mass spectra collected from complex Cu-Fe sulphides (chalcopyrite, bornite, chalcocite and pyrite) at different flotation conditions. ANNs are very good pattern classifiers because of: their ability to learn and generalise patterns that are not linearly separable; their fault and noise tolerance capability; and high parallelism. In the first approach, fragments from the whole ToF-SIMS spectrum were used as input to the ANN, the model yielded high overall correct classification rates of 100% for feed samples, 88% for conditioned feed samples and 91% for Eh modified samples. In the second approach, the hybrid pattern classifier PCA-ANN was integrated. PCA is a very effective multivariate data analysis tool applied to enhance species features and reduce data dimensionality. Principal component (PC) scores which accounted for 95% of the raw spectral data variance, were used as input to the ANN, the model yielded high overall correct classification rates of 88% for conditioned feed samples and 95% for Eh modified samples. Copyright © 2012 Elsevier B.V. All rights reserved.
Scalable Robust Principal Component Analysis Using Grassmann Averages.
Hauberg, Sren; Feragen, Aasa; Enficiaud, Raffi; Black, Michael J
2016-11-01
In large datasets, manual data verification is impossible, and we must expect the number of outliers to increase with data size. While principal component analysis (PCA) can reduce data size, and scalable solutions exist, it is well-known that outliers can arbitrarily corrupt the results. Unfortunately, state-of-the-art approaches for robust PCA are not scalable. We note that in a zero-mean dataset, each observation spans a one-dimensional subspace, giving a point on the Grassmann manifold. We show that the average subspace corresponds to the leading principal component for Gaussian data. We provide a simple algorithm for computing this Grassmann Average ( GA), and show that the subspace estimate is less sensitive to outliers than PCA for general distributions. Because averages can be efficiently computed, we immediately gain scalability. We exploit robust averaging to formulate the Robust Grassmann Average (RGA) as a form of robust PCA. The resulting Trimmed Grassmann Average ( TGA) is appropriate for computer vision because it is robust to pixel outliers. The algorithm has linear computational complexity and minimal memory requirements. We demonstrate TGA for background modeling, video restoration, and shadow removal. We show scalability by performing robust PCA on the entire Star Wars IV movie; a task beyond any current method. Source code is available online.
Product competitiveness analysis for e-commerce platform of special agricultural products
NASA Astrophysics Data System (ADS)
Wan, Fucheng; Ma, Ning; Yang, Dongwei; Xiong, Zhangyuan
2017-09-01
On the basis of analyzing the influence factors of the product competitiveness of the e-commerce platform of the special agricultural products and the characteristics of the analytical methods for the competitiveness of the special agricultural products, the price, the sales volume, the postage included service, the store reputation, the popularity, etc. were selected in this paper as the dimensionality for analyzing the competitiveness of the agricultural products, and the principal component factor analysis was taken as the competitiveness analysis method. Specifically, the web crawler was adopted to capture the information of various special agricultural products in the e-commerce platform ---- chi.taobao.com. Then, the original data captured thereby were preprocessed and MYSQL database was adopted to establish the information library for the special agricultural products. Then, the principal component factor analysis method was adopted to establish the analysis model for the competitiveness of the special agricultural products, and SPSS was adopted in the principal component factor analysis process to obtain the competitiveness evaluation factor system (support degree factor, price factor, service factor and evaluation factor) of the special agricultural products. Then, the linear regression method was adopted to establish the competitiveness index equation of the special agricultural products for estimating the competitiveness of the special agricultural products.
Ho, Hau My; Lin, Binhua; Rice, Stuart A
2006-11-14
We report the results of experimental determinations of the triplet correlation functions of quasi-two-dimensional one-component and binary colloid suspensions in which the colloid-colloid interaction is short ranged. The suspensions studied range in density from modestly dilute to solid. The triplet correlation function of the one-component colloid system reveals extensive ordering deep in the liquid phase. At the same density the ordering of the larger diameter component in a binary colloid system is greatly diminished by a very small amount of the smaller diameter component. The possible utilization of information contained in the triplet correlation function in the theory of melting of a quasi-two-dimensional system is briefly discussed.
Sparks, Rachel; Madabhushi, Anant
2016-01-01
Content-based image retrieval (CBIR) retrieves database images most similar to the query image by (1) extracting quantitative image descriptors and (2) calculating similarity between database and query image descriptors. Recently, manifold learning (ML) has been used to perform CBIR in a low dimensional representation of the high dimensional image descriptor space to avoid the curse of dimensionality. ML schemes are computationally expensive, requiring an eigenvalue decomposition (EVD) for every new query image to learn its low dimensional representation. We present out-of-sample extrapolation utilizing semi-supervised ML (OSE-SSL) to learn the low dimensional representation without recomputing the EVD for each query image. OSE-SSL incorporates semantic information, partial class label, into a ML scheme such that the low dimensional representation co-localizes semantically similar images. In the context of prostate histopathology, gland morphology is an integral component of the Gleason score which enables discrimination between prostate cancer aggressiveness. Images are represented by shape features extracted from the prostate gland. CBIR with OSE-SSL for prostate histology obtained from 58 patient studies, yielded an area under the precision recall curve (AUPRC) of 0.53 ± 0.03 comparatively a CBIR with Principal Component Analysis (PCA) to learn a low dimensional space yielded an AUPRC of 0.44 ± 0.01. PMID:27264985
Liu, Hui-lin; Wan, Xia; Yang, Gong-huan
2013-02-01
To explore the relationship between the strength of tobacco control and the effectiveness of creating smoke-free hospital, and summarize the main factors that affect the program of creating smoke-free hospitals. A total of 210 hospitals from 7 provinces/municipalities directly under the central government were enrolled in this study using stratified random sampling method. Principle component analysis and regression analysis were conducted to analyze the strength of tobacco control and the effectiveness of creating smoke-free hospitals. Two principal components were extracted in the strength of tobacco control index, which respectively reflected the tobacco control policies and efforts, and the willingness and leadership of hospital managers regarding tobacco control. The regression analysis indicated that only the first principal component was significantly correlated with the progression in creating smoke-free hospital (P<0.001), i.e. hospitals with higher scores on the first principal component had better achievements in smoke-free environment creation. Tobacco control policies and efforts are critical in creating smoke-free hospitals. The principal component analysis provides a comprehensive and objective tool for evaluating the creation of smoke-free hospitals.
Falkum, Erik; Pedersen, Geir; Karterud, Sigmund
2009-01-01
This article examines reliability and validity aspects of the Diagnostic and Statistical Manual of Mental Disorders, Fourth Edition (DSM-IV) paranoid personality disorder (PPD) diagnosis. Patients with personality disorders (n = 930) from the Norwegian network of psychotherapeutic day hospitals, of which 114 had PPD, were included in the study. Frequency distribution, chi(2), correlations, reliability statistics, exploratory, and confirmatory factor analyses were performed. The distribution of PPD criteria revealed no distinct boundary between patients with and without PPD. Diagnostic category membership was obtained in 37 of 64 theoretically possible ways. The PPD criteria formed a separate factor in a principal component analysis, whereas a confirmatory factor analysis indicated that the DSM-IV PPD construct consists of 2 separate dimensions as follows: suspiciousness and hostility. The reliability of the unitary PPD scale was only 0.70, probably partly due to the apparent 2-dimensionality of the construct. Persistent unwarranted doubts about the loyalty of friends had the highest diagnostic efficiency, whereas unwarranted accusations of infidelity of partner had particularly poor indicator properties. The reliability and validity of the unitary PPD construct may be questioned. The 2-dimensional PPD model should be further explored.
Stochastic analysis of transverse dispersion in density‐coupled transport in aquifers
Welty, Claire; Kane, Allen C.; Kauffman, Leon J.
2003-01-01
Spectral perturbation techniques have been used previously to derive integral expressions for dispersive mixing in concentration‐dependent transport in three‐dimensional, heterogeneous porous media, where fluid density and viscosity are functions of solute concentration. Whereas earlier work focused on evaluating longitudinal dispersivity in isotropic media and incorporating the result in a mean one‐dimensional transport model, the emphasis of this paper is on evaluation of the complete dispersion tensor, including the more general case of anisotropic media. Approximate analytic expressions for all components of the macroscopic dispersivity tensor are derived, and the tensor is shown to be asymmetric. The tensor is separated into its symmetric and antisymmetric parts, where the symmetric part is used to calculate the principal components and principal directions of dispersivity, and the antisymmetric part of the tensor is shown to modify the velocity of the solute body compared to that of the background fluid. An example set of numerical simulations incorporating the tensor illustrates the effect of density‐coupled dispersivity on a sinking plume in an aquifer. The simulations show that the effective transverse vertical spreading in a sinking plume to be significantly greater than would be predicted by a standard density‐coupled transport model that does not incorporate the coupling in the dispersivity tensor.
Caggiano, Alessandra
2018-03-09
Machining of titanium alloys is characterised by extremely rapid tool wear due to the high cutting temperature and the strong adhesion at the tool-chip and tool-workpiece interface, caused by the low thermal conductivity and high chemical reactivity of Ti alloys. With the aim to monitor the tool conditions during dry turning of Ti-6Al-4V alloy, a machine learning procedure based on the acquisition and processing of cutting force, acoustic emission and vibration sensor signals during turning is implemented. A number of sensorial features are extracted from the acquired sensor signals in order to feed machine learning paradigms based on artificial neural networks. To reduce the large dimensionality of the sensorial features, an advanced feature extraction methodology based on Principal Component Analysis (PCA) is proposed. PCA allowed to identify a smaller number of features ( k = 2 features), the principal component scores, obtained through linear projection of the original d features into a new space with reduced dimensionality k = 2, sufficient to describe the variance of the data. By feeding artificial neural networks with the PCA features, an accurate diagnosis of tool flank wear ( VB max ) was achieved, with predicted values very close to the measured tool wear values.
2018-01-01
Machining of titanium alloys is characterised by extremely rapid tool wear due to the high cutting temperature and the strong adhesion at the tool-chip and tool-workpiece interface, caused by the low thermal conductivity and high chemical reactivity of Ti alloys. With the aim to monitor the tool conditions during dry turning of Ti-6Al-4V alloy, a machine learning procedure based on the acquisition and processing of cutting force, acoustic emission and vibration sensor signals during turning is implemented. A number of sensorial features are extracted from the acquired sensor signals in order to feed machine learning paradigms based on artificial neural networks. To reduce the large dimensionality of the sensorial features, an advanced feature extraction methodology based on Principal Component Analysis (PCA) is proposed. PCA allowed to identify a smaller number of features (k = 2 features), the principal component scores, obtained through linear projection of the original d features into a new space with reduced dimensionality k = 2, sufficient to describe the variance of the data. By feeding artificial neural networks with the PCA features, an accurate diagnosis of tool flank wear (VBmax) was achieved, with predicted values very close to the measured tool wear values. PMID:29522443
Scaling between Wind Tunnels-Results Accuracy in Two-Dimensional Testing
NASA Astrophysics Data System (ADS)
Rasuo, Bosko
The establishment of exact two-dimensional flow conditions in wind tunnels is a very difficult problem. This has been evident for wind tunnels of all types and scales. In this paper, the principal factors that influence the accuracy of two-dimensional wind tunnel test results are analyzed. The influences of the Reynolds number, Mach number and wall interference with reference to solid and flow blockage (blockage of wake) as well as the influence of side-wall boundary layer control are analyzed. Interesting results are brought to light regarding the Reynolds number effects of the test model versus the Reynolds number effects of the facility in subsonic and transonic flow.
Vorticity and helicity decompositions and dynamics with real Schur form of the velocity gradient
NASA Astrophysics Data System (ADS)
Zhu, Jian-Zhou
2018-03-01
The real Schur form (RSF) of a generic velocity gradient field ∇u is exploited to expose the structures of flows, in particular, our field decomposition resulting in two vorticities with only mutual linkage as the topological content of the global helicity (accordingly decomposed into two equal parts). The local transformation to the RSF may indicate alternative (co)rotating frame(s) for specifying the objective argument(s) of the constitutive equation. When ∇u is uniformly of RSF in a fixed Cartesian coordinate frame, i.e., ux = ux(x, y) and uy = uy(x, y), but uz = uz(x, y, z), the model, with the decomposed vorticities both frozen-in to u, is for two-component-two-dimensional-coupled-with-one-component-three-dimensional flows in between two-dimensional-three-component (2D3C) and fully three-dimensional-three-component ones and may help curing the pathology in the helical 2D3C absolute equilibrium, making the latter effectively work in more realistic situations.
NASA Astrophysics Data System (ADS)
Sakaguchi, Hidetsugu; Malomed, Boris A.
2017-10-01
We analyze the possibility of macroscopic quantum effects in the form of coupled structural oscillations and shuttle motion of bright two-component spin-orbit-coupled striped (one-dimensional, 1D) and semivortex (two-dimensional, 2D) matter-wave solitons, under the action of linear mixing (Rabi coupling) between the components. In 1D, the intrinsic oscillations manifest themselves as flippings between spatially even and odd components of striped solitons, while in 2D the system features periodic transitions between zero-vorticity and vortical components of semivortex solitons. The consideration is performed by means of a combination of analytical and numerical methods.
NASA Astrophysics Data System (ADS)
Takeuchi, Hiromitsu; Kasamatsu, Kenichi; Tsubota, Makoto; Nitta, Muneto
2013-05-01
In brane cosmology, the Big Bang is hypothesized to occur by the annihilation of the brane-anti-brane pair in a collision, where the branes are three-dimensional objects in a higher-dimensional Universe. Spontaneous symmetry breaking accompanied by the formation of lower-dimensional topological defects, e.g. cosmic strings, is triggered by the so-called `tachyon condensation', where the existence of tachyons is attributable to the instability of the brane-anti-brane system. Here, we discuss the closest analogue of the tachyon condensation in atomic Bose-Einstein condensates. We consider annihilation of domain walls, namely branes, in strongly segregated two-component condensates, where one component is sandwiched by two domains of the other component. In this system, the process of the brane annihilation can be projected effectively as ferromagnetic ordering dynamics onto a two-dimensional space. Based on this correspondence, three-dimensional formation of vortices from a domain-wall annihilation is considered to be a kink formation due to spontaneous symmetry breaking in the two-dimensional space. We also discuss a mechanism to create a `vorton' when the sandwiched component has a vortex string bridged between the branes. We hope that this study motivates experimental researches to realize this exotic phenomenon of spontaneous symmetry breaking in superfluid systems.
Colorimetric Sensor Array for White Wine Tasting.
Chung, Soo; Park, Tu San; Park, Soo Hyun; Kim, Joon Yong; Park, Seongmin; Son, Daesik; Bae, Young Min; Cho, Seong In
2015-07-24
A colorimetric sensor array was developed to characterize and quantify the taste of white wines. A charge-coupled device (CCD) camera captured images of the sensor array from 23 different white wine samples, and the change in the R, G, B color components from the control were analyzed by principal component analysis. Additionally, high performance liquid chromatography (HPLC) was used to analyze the chemical components of each wine sample responsible for its taste. A two-dimensional score plot was created with 23 data points. It revealed clusters created from the same type of grape, and trends of sweetness, sourness, and astringency were mapped. An artificial neural network model was developed to predict the degree of sweetness, sourness, and astringency of the white wines. The coefficients of determination (R2) for the HPLC results and the sweetness, sourness, and astringency were 0.96, 0.95, and 0.83, respectively. This research could provide a simple and low-cost but sensitive taste prediction system, and, by helping consumer selection, will be able to have a positive effect on the wine industry.
Colorimetric Sensor Array for White Wine Tasting
Chung, Soo; Park, Tu San; Park, Soo Hyun; Kim, Joon Yong; Park, Seongmin; Son, Daesik; Bae, Young Min; Cho, Seong In
2015-01-01
A colorimetric sensor array was developed to characterize and quantify the taste of white wines. A charge-coupled device (CCD) camera captured images of the sensor array from 23 different white wine samples, and the change in the R, G, B color components from the control were analyzed by principal component analysis. Additionally, high performance liquid chromatography (HPLC) was used to analyze the chemical components of each wine sample responsible for its taste. A two-dimensional score plot was created with 23 data points. It revealed clusters created from the same type of grape, and trends of sweetness, sourness, and astringency were mapped. An artificial neural network model was developed to predict the degree of sweetness, sourness, and astringency of the white wines. The coefficients of determination (R2) for the HPLC results and the sweetness, sourness, and astringency were 0.96, 0.95, and 0.83, respectively. This research could provide a simple and low-cost but sensitive taste prediction system, and, by helping consumer selection, will be able to have a positive effect on the wine industry. PMID:26213946
Differential principal component analysis of ChIP-seq.
Ji, Hongkai; Li, Xia; Wang, Qian-fei; Ning, Yang
2013-04-23
We propose differential principal component analysis (dPCA) for analyzing multiple ChIP-sequencing datasets to identify differential protein-DNA interactions between two biological conditions. dPCA integrates unsupervised pattern discovery, dimension reduction, and statistical inference into a single framework. It uses a small number of principal components to summarize concisely the major multiprotein synergistic differential patterns between the two conditions. For each pattern, it detects and prioritizes differential genomic loci by comparing the between-condition differences with the within-condition variation among replicate samples. dPCA provides a unique tool for efficiently analyzing large amounts of ChIP-sequencing data to study dynamic changes of gene regulation across different biological conditions. We demonstrate this approach through analyses of differential chromatin patterns at transcription factor binding sites and promoters as well as allele-specific protein-DNA interactions.
Principal components analysis in clinical studies.
Zhang, Zhongheng; Castelló, Adela
2017-09-01
In multivariate analysis, independent variables are usually correlated to each other which can introduce multicollinearity in the regression models. One approach to solve this problem is to apply principal components analysis (PCA) over these variables. This method uses orthogonal transformation to represent sets of potentially correlated variables with principal components (PC) that are linearly uncorrelated. PCs are ordered so that the first PC has the largest possible variance and only some components are selected to represent the correlated variables. As a result, the dimension of the variable space is reduced. This tutorial illustrates how to perform PCA in R environment, the example is a simulated dataset in which two PCs are responsible for the majority of the variance in the data. Furthermore, the visualization of PCA is highlighted.
NASA Technical Reports Server (NTRS)
Generazio, E. R.
1986-01-01
Microstructural images may be tone pulse encoded and subsequently Fourier transformed to determine the two-dimensional density of frequency components. A theory is developed relating the density of frequency components to the density of length components. The density of length components corresponds directly to the actual grain size distribution function from which the mean grain shape, size, and orientation can be obtained.
Deep Adaptive Log-Demons: Diffeomorphic Image Registration with Very Large Deformations
Jia, Kebin
2015-01-01
This paper proposes a new framework for capturing large and complex deformation in image registration. Traditionally, this challenging problem relies firstly on a preregistration, usually an affine matrix containing rotation, scale, and translation and afterwards on a nonrigid transformation. According to preregistration, the directly calculated affine matrix, which is obtained by limited pixel information, may misregistrate when large biases exist, thus misleading following registration subversively. To address this problem, for two-dimensional (2D) images, the two-layer deep adaptive registration framework proposed in this paper firstly accurately classifies the rotation parameter through multilayer convolutional neural networks (CNNs) and then identifies scale and translation parameters separately. For three-dimensional (3D) images, affine matrix is located through feature correspondences by a triplanar 2D CNNs. Then deformation removal is done iteratively through preregistration and demons registration. By comparison with the state-of-the-art registration framework, our method gains more accurate registration results on both synthetic and real datasets. Besides, principal component analysis (PCA) is combined with correlation like Pearson and Spearman to form new similarity standards in 2D and 3D registration. Experiment results also show faster convergence speed. PMID:26120356
Deep Adaptive Log-Demons: Diffeomorphic Image Registration with Very Large Deformations.
Zhao, Liya; Jia, Kebin
2015-01-01
This paper proposes a new framework for capturing large and complex deformation in image registration. Traditionally, this challenging problem relies firstly on a preregistration, usually an affine matrix containing rotation, scale, and translation and afterwards on a nonrigid transformation. According to preregistration, the directly calculated affine matrix, which is obtained by limited pixel information, may misregistrate when large biases exist, thus misleading following registration subversively. To address this problem, for two-dimensional (2D) images, the two-layer deep adaptive registration framework proposed in this paper firstly accurately classifies the rotation parameter through multilayer convolutional neural networks (CNNs) and then identifies scale and translation parameters separately. For three-dimensional (3D) images, affine matrix is located through feature correspondences by a triplanar 2D CNNs. Then deformation removal is done iteratively through preregistration and demons registration. By comparison with the state-of-the-art registration framework, our method gains more accurate registration results on both synthetic and real datasets. Besides, principal component analysis (PCA) is combined with correlation like Pearson and Spearman to form new similarity standards in 2D and 3D registration. Experiment results also show faster convergence speed.
Missouri University Multi-Plane Imager (MUMPI): A high sensitivity rapid dynamic ECT brain imager
DOE Office of Scientific and Technical Information (OSTI.GOV)
Logan, K.W.; Holmes, R.A.
1984-01-01
The authors have designed a unique ECT imaging device that can record rapid dynamic images of brain perfusion. The Missouri University Multi-Plane Imager (MUMPI) uses a single crystal detector that produces four orthogonal two-dimensional images simultaneously. Multiple slice images are reconstructed from counts recorded from stepwise or continuous collimator rotation. Four simultaneous 2-d image fields may also be recorded and reviewed. The cylindrical sodium iodide crystal and the rotating collimator concentrically surround the source volume being imaged with the collimator the only moving part. The design and function parameters of MUMPI have been compared to other competitive tomographic head imagingmore » devices. MUMPI's principal advantages are: 1) simultaneous direct acquisition of four two-dimensional images; 2) extremely rapid project set acquisition for ECT reconstruction; and 3) instrument practicality and economy due to single detector design and the absence of heavy mechanical moving components (only collimator rotation is required). MUMPI should be ideal for imaging neutral lipophilic chelates such as Tc-99m-PnAO which passively diffuses across the intact blood-brain-barrier and rapidly clears from brain tissue.« less
QSAR modeling of flotation collectors using principal components extracted from topological indices.
Natarajan, R; Nirdosh, Inderjit; Basak, Subhash C; Mills, Denise R
2002-01-01
Several topological indices were calculated for substituted-cupferrons that were tested as collectors for the froth flotation of uranium. The principal component analysis (PCA) was used for data reduction. Seven principal components (PC) were found to account for 98.6% of the variance among the computed indices. The principal components thus extracted were used in stepwise regression analyses to construct regression models for the prediction of separation efficiencies (Es) of the collectors. A two-parameter model with a correlation coefficient of 0.889 and a three-parameter model with a correlation coefficient of 0.913 were formed. PCs were found to be better than partition coefficient to form regression equations, and inclusion of an electronic parameter such as Hammett sigma or quantum mechanically derived electronic charges on the chelating atoms did not improve the correlation coefficient significantly. The method was extended to model the separation efficiencies of mercaptobenzothiazoles (MBT) and aminothiophenols (ATP) used in the flotation of lead and zinc ores, respectively. Five principal components were found to explain 99% of the data variability in each series. A three-parameter equation with correlation coefficient of 0.985 and a two-parameter equation with correlation coefficient of 0.926 were obtained for MBT and ATP, respectively. The amenability of separation efficiencies of chelating collectors to QSAR modeling using PCs based on topological indices might lead to the selection of collectors for synthesis and testing from a virtual database.
Target oriented dimensionality reduction of hyperspectral data by Kernel Fukunaga-Koontz Transform
NASA Astrophysics Data System (ADS)
Binol, Hamidullah; Ochilov, Shuhrat; Alam, Mohammad S.; Bal, Abdullah
2017-02-01
Principal component analysis (PCA) is a popular technique in remote sensing for dimensionality reduction. While PCA is suitable for data compression, it is not necessarily an optimal technique for feature extraction, particularly when the features are exploited in supervised learning applications (Cheriyadat and Bruce, 2003) [1]. Preserving features belonging to the target is very crucial to the performance of target detection/recognition techniques. Fukunaga-Koontz Transform (FKT) based supervised band reduction technique can be used to provide this requirement. FKT achieves feature selection by transforming into a new space in where feature classes have complimentary eigenvectors. Analysis of these eigenvectors under two classes, target and background clutter, can be utilized for target oriented band reduction since each basis functions best represent target class while carrying least information of the background class. By selecting few eigenvectors which are the most relevant to the target class, dimension of hyperspectral data can be reduced and thus, it presents significant advantages for near real time target detection applications. The nonlinear properties of the data can be extracted by kernel approach which provides better target features. Thus, we propose constructing kernel FKT (KFKT) to present target oriented band reduction. The performance of the proposed KFKT based target oriented dimensionality reduction algorithm has been tested employing two real-world hyperspectral data and results have been reported consequently.
Static vs. dynamic decoding algorithms in a non-invasive body-machine interface
Seáñez-González, Ismael; Pierella, Camilla; Farshchiansadegh, Ali; Thorp, Elias B.; Abdollahi, Farnaz; Pedersen, Jessica; Mussa-Ivaldi, Ferdinando A.
2017-01-01
In this study, we consider a non-invasive body-machine interface that captures body motions still available to people with spinal cord injury (SCI) and maps them into a set of signals for controlling a computer user interface while engaging in a sustained level of mobility and exercise. We compare the effectiveness of two decoding algorithms that transform a high-dimensional body-signal vector into a lower dimensional control vector on 6 subjects with high-level SCI and 8 controls. One algorithm is based on a static map from current body signals to the current value of the control vector set through principal component analysis (PCA), the other on dynamic mapping a segment of body signals to the value and the temporal derivatives of the control vector set through a Kalman filter. SCI and control participants performed straighter and smoother cursor movements with the Kalman algorithm during center-out reaching, but their movements were faster and more precise when using PCA. All participants were able to use the BMI’s continuous, two-dimensional control to type on a virtual keyboard and play pong, and performance with both algorithms was comparable. However, seven of eight control participants preferred PCA as their method of virtual wheelchair control. The unsupervised PCA algorithm was easier to train and seemed sufficient to achieve a higher degree of learnability and perceived ease of use. PMID:28092564
Rossi, Gabriela Barbosa; Valentim-Neto, Pedro Alexandre; Blank, Martina; Faria, Josias Correa de; Arisi, Ana Carolina Maisonnave
2017-08-30
Common bean (Phaseolus vulgaris L.) is a source of proteins for about one billion people worldwide. In Brazil, 'BRS Sublime', 'BRS Vereda', 'BRS Esteio', and 'BRS Estilo' cultivars were developed by Embrapa to offer high yield to farmers and excellent quality to final consumers. In this work, grain proteomes of these common bean cultivars were compared based on two-dimensional gel electrophoresis (2-DE) and tandem mass spectrometry (MS/MS). Principal component analysis (PCA) was applied to compare 349 matched spots in these cultivars proteomes, and all cultivars were clearly separated in PCA plot. Thirty-two differentially accumulated proteins were identified by MS. Storage proteins such as phaseolins, legumins, and lectins were the most abundant, and novel proteins were also identified. We have built a useful platform that could be used to analyze other Brazilian cultivars and genotypes of common beans.
Exploring the CAESAR database using dimensionality reduction techniques
NASA Astrophysics Data System (ADS)
Mendoza-Schrock, Olga; Raymer, Michael L.
2012-06-01
The Civilian American and European Surface Anthropometry Resource (CAESAR) database containing over 40 anthropometric measurements on over 4000 humans has been extensively explored for pattern recognition and classification purposes using the raw, original data [1-4]. However, some of the anthropometric variables would be impossible to collect in an uncontrolled environment. Here, we explore the use of dimensionality reduction methods in concert with a variety of classification algorithms for gender classification using only those variables that are readily observable in an uncontrolled environment. Several dimensionality reduction techniques are employed to learn the underlining structure of the data. These techniques include linear projections such as the classical Principal Components Analysis (PCA) and non-linear (manifold learning) techniques, such as Diffusion Maps and the Isomap technique. This paper briefly describes all three techniques, and compares three different classifiers, Naïve Bayes, Adaboost, and Support Vector Machines (SVM), for gender classification in conjunction with each of these three dimensionality reduction approaches.
A DATA-DRIVEN MODEL FOR SPECTRA: FINDING DOUBLE REDSHIFTS IN THE SLOAN DIGITAL SKY SURVEY
DOE Office of Scientific and Technical Information (OSTI.GOV)
Tsalmantza, P.; Hogg, David W., E-mail: vivitsal@mpia.de
2012-07-10
We present a data-driven method-heteroscedastic matrix factorization, a kind of probabilistic factor analysis-for modeling or performing dimensionality reduction on observed spectra or other high-dimensional data with known but non-uniform observational uncertainties. The method uses an iterative inverse-variance-weighted least-squares minimization procedure to generate a best set of basis functions. The method is similar to principal components analysis (PCA), but with the substantial advantage that it uses measurement uncertainties in a responsible way and accounts naturally for poorly measured and missing data; it models the variance in the noise-deconvolved data space. A regularization can be applied, in the form of a smoothnessmore » prior (inspired by Gaussian processes) or a non-negative constraint, without making the method prohibitively slow. Because the method optimizes a justified scalar (related to the likelihood), the basis provides a better fit to the data in a probabilistic sense than any PCA basis. We test the method on Sloan Digital Sky Survey (SDSS) spectra, concentrating on spectra known to contain two redshift components: these are spectra of gravitational lens candidates and massive black hole binaries. We apply a hypothesis test to compare one-redshift and two-redshift models for these spectra, utilizing the data-driven model trained on a random subset of all SDSS spectra. This test confirms 129 of the 131 lens candidates in our sample and all of the known binary candidates, and turns up very few false positives.« less
NASA Astrophysics Data System (ADS)
Ginanjar, Irlandia; Pasaribu, Udjianna S.; Indratno, Sapto W.
2017-03-01
This article presents the application of the principal component analysis (PCA) biplot for the needs of data mining. This article aims to simplify and objectify the methods for objects clustering in PCA biplot. The novelty of this paper is to get a measure that can be used to objectify the objects clustering in PCA biplot. Orthonormal eigenvectors, which are the coefficients of a principal component model representing an association between principal components and initial variables. The existence of the association is a valid ground to objects clustering based on principal axes value, thus if m principal axes used in the PCA, then the objects can be classified into 2m clusters. The inter-city buses are clustered based on maintenance costs data by using two principal axes PCA biplot. The buses are clustered into four groups. The first group is the buses with high maintenance costs, especially for lube, and brake canvass. The second group is the buses with high maintenance costs, especially for tire, and filter. The third group is the buses with low maintenance costs, especially for lube, and brake canvass. The fourth group is buses with low maintenance costs, especially for tire, and filter.
Lax representations for matrix short pulse equations
NASA Astrophysics Data System (ADS)
Popowicz, Z.
2017-10-01
The Lax representation for different matrix generalizations of Short Pulse Equations (SPEs) is considered. The four-dimensional Lax representations of four-component Matsuno, Feng, and Dimakis-Müller-Hoissen-Matsuno equations are obtained. The four-component Feng system is defined by generalization of the two-dimensional Lax representation to the four-component case. This system reduces to the original Feng equation, to the two-component Matsuno equation, or to the Yao-Zang equation. The three-component version of the Feng equation is presented. The four-component version of the Matsuno equation with its Lax representation is given. This equation reduces the new two-component Feng system. The two-component Dimakis-Müller-Hoissen-Matsuno equations are generalized to the four-parameter family of the four-component SPE. The bi-Hamiltonian structure of this generalization, for special values of parameters, is defined. This four-component SPE in special cases reduces to the new two-component SPE.
Coarse Point Cloud Registration by Egi Matching of Voxel Clusters
NASA Astrophysics Data System (ADS)
Wang, Jinhu; Lindenbergh, Roderik; Shen, Yueqian; Menenti, Massimo
2016-06-01
Laser scanning samples the surface geometry of objects efficiently and records versatile information as point clouds. However, often more scans are required to fully cover a scene. Therefore, a registration step is required that transforms the different scans into a common coordinate system. The registration of point clouds is usually conducted in two steps, i.e. coarse registration followed by fine registration. In this study an automatic marker-free coarse registration method for pair-wise scans is presented. First the two input point clouds are re-sampled as voxels and dimensionality features of the voxels are determined by principal component analysis (PCA). Then voxel cells with the same dimensionality are clustered. Next, the Extended Gaussian Image (EGI) descriptor of those voxel clusters are constructed using significant eigenvectors of each voxel in the cluster. Correspondences between clusters in source and target data are obtained according to the similarity between their EGI descriptors. The random sampling consensus (RANSAC) algorithm is employed to remove outlying correspondences until a coarse alignment is obtained. If necessary, a fine registration is performed in a final step. This new method is illustrated on scan data sampling two indoor scenarios. The results of the tests are evaluated by computing the point to point distance between the two input point clouds. The presented two tests resulted in mean distances of 7.6 mm and 9.5 mm respectively, which are adequate for fine registration.
Schwartz, Ted R.; Stalling, David L.
1991-01-01
The separation and characterization of complex mixtures of polychlorinated biphenyls (PCBs) is approached from the perspective of a problem in chemometrics. A technique for quantitative determination of PCB congeners is described as well as an enrichment technique designed to isolate only those congener residues which induce mixed aryl hydrocarbon hydroxylase enzyme activity. A congener-specific procedure is utilized for the determination of PCBs in whichn-alkyl trichloroacetates are used as retention index marker compounds. Retention indices are reproducible in the range of ±0.05 to ±0.7 depending on the specific congener. A laboratory data base system developed to aid in the editing and quantitation of data generated from capillary gas chromatography was employed to quantitate chromatographic data. Data base management was provided by computer programs written in VAX-DSM (Digital Standard MUMPS) for the VAX-DEC (Digital Equipment Corp.) family of computers.In the chemometric evaluation of these complex chromatographic profiles, data are viewed from a single analysis as a point in multi-dimensional space. Principal Components Analysis was used to obtain a representation of the data in a lower dimensional space. Two-and three-dimensional proections based on sample scores from the principal components models were used to visualize the behavior of Aroclor® mixtures. These models can be used to determine if new sample profiles may be represented by Aroclor profiles. Concentrations of individual congeners of a given chlorine substitution may be summed to form homologue concentration. However, the use of homologue concentrations in classification studies with environmental samples can lead to erroneous conclusions about sample similarity. Chemometric applications are discussed for evaluation of Aroclor mixture analysis and compositional description of environmental residues of PCBs in eggs of Forster's terns (Sterna fosteri) collected from colonies near Lake Poygan and Green Bay, Wisconsin. The application of chemometrics is extended to the comparison of: a) Aroclors and PCB-containing environmental samples; to b) fractions of Aroclors and of environmental samples that have been enriched in congeners which induce mixed aryl hydrocarbon hydroxylase enzyme activity.
Finding Imaging Patterns of Structural Covariance via Non-Negative Matrix Factorization
Sotiras, Aristeidis; Resnick, Susan M.; Davatzikos, Christos
2015-01-01
In this paper, we investigate the use of Non-Negative Matrix Factorization (NNMF) for the analysis of structural neuroimaging data. The goal is to identify the brain regions that co-vary across individuals in a consistent way, hence potentially being part of underlying brain networks or otherwise influenced by underlying common mechanisms such as genetics and pathologies. NNMF offers a directly data-driven way of extracting relatively localized co-varying structural regions, thereby transcending limitations of Principal Component Analysis (PCA), Independent Component Analysis (ICA) and other related methods that tend to produce dispersed components of positive and negative loadings. In particular, leveraging upon the well known ability of NNMF to produce parts-based representations of image data, we derive decompositions that partition the brain into regions that vary in consistent ways across individuals. Importantly, these decompositions achieve dimensionality reduction via highly interpretable ways and generalize well to new data as shown via split-sample experiments. We empirically validate NNMF in two data sets: i) a Diffusion Tensor (DT) mouse brain development study, and ii) a structural Magnetic Resonance (sMR) study of human brain aging. We demonstrate the ability of NNMF to produce sparse parts-based representations of the data at various resolutions. These representations seem to follow what we know about the underlying functional organization of the brain and also capture some pathological processes. Moreover, we show that these low dimensional representations favorably compare to descriptions obtained with more commonly used matrix factorization methods like PCA and ICA. PMID:25497684
Tsai, Tsung-Yuan; Li, Jing-Sheng; Wang, Shaobai; Li, Pingyue; Kwon, Young-Min; Li, Guoan
2015-01-01
The statistical shape model (SSM) method that uses 2D images of the knee joint to predict the three-dimensional (3D) joint surface model has been reported in the literature. In this study, we constructed a SSM database using 152 human computed tomography (CT) knee joint models, including the femur, tibia and patella and analysed the characteristics of each principal component of the SSM. The surface models of two in vivo knees were predicted using the SSM and their 2D bi-plane fluoroscopic images. The predicted models were compared to their CT joint models. The differences between the predicted 3D knee joint surfaces and the CT image-based surfaces were 0.30 ± 0.81 mm, 0.34 ± 0.79 mm and 0.36 ± 0.59 mm for the femur, tibia and patella, respectively (average ± standard deviation). The computational time for each bone of the knee joint was within 30 s using a personal computer. The analysis of this study indicated that the SSM method could be a useful tool to construct 3D surface models of the knee with sub-millimeter accuracy in real time. Thus, it may have a broad application in computer-assisted knee surgeries that require 3D surface models of the knee.
Lee, Seung Ho; Lee, Sang Hwa; Shin, Jae-Ho; Choi, Samjin
2018-06-01
Although the confirmation of inflammatory changes within tissues at the onset of various diseases is critical for the early detection of disease and selection of appropriate treatment, most therapies are based on complex and time-consuming diagnostic procedures. Raman spectroscopy has the ability to provide non-invasive, real-time, chemical bonding analysis through the inelastic scattering of photons. In this study, we evaluate the feasibility of Raman spectroscopy as a new, easy, fast, and accurate diagnostic method to support diagnostic decisions. The molecular changes in carrageenan-induced acute inflammation rat tissues were assessed by Raman spectroscopy. Volumes of 0 (control), 100, 150, and 200 µL of 1% carrageenan were administered to rat hind paws to control the degree of inflammation. The prominent peaks at [1,062, 1,131] cm -1 and [2,847, 2,881] cm -1 were selected as characteristic measurements corresponding to the C-C stretching vibrational modes and the symmetric and asymmetric C-H (CH 2 ) stretching vibrational modes, respectively. Principal component analysis of the inflammatory Raman spectra enabled graphical representation of the degree of inflammation through principal component loading profiles of inflammatory tissues on a two-dimensional plot. Therefore, Raman spectroscopy with multivariate statistical analysis represents a promising method for detecting biomolecular responses based on different types of inflammatory tissues. © 2018 Wiley Periodicals, Inc.
NASA Astrophysics Data System (ADS)
Hu, Wenjian; Singh, Rajiv R. P.; Scalettar, Richard T.
2017-06-01
We apply unsupervised machine learning techniques, mainly principal component analysis (PCA), to compare and contrast the phase behavior and phase transitions in several classical spin models—the square- and triangular-lattice Ising models, the Blume-Capel model, a highly degenerate biquadratic-exchange spin-1 Ising (BSI) model, and the two-dimensional X Y model—and we examine critically what machine learning is teaching us. We find that quantified principal components from PCA not only allow the exploration of different phases and symmetry-breaking, but they can distinguish phase-transition types and locate critical points. We show that the corresponding weight vectors have a clear physical interpretation, which is particularly interesting in the frustrated models such as the triangular antiferromagnet, where they can point to incipient orders. Unlike the other well-studied models, the properties of the BSI model are less well known. Using both PCA and conventional Monte Carlo analysis, we demonstrate that the BSI model shows an absence of phase transition and macroscopic ground-state degeneracy. The failure to capture the "charge" correlations (vorticity) in the BSI model (X Y model) from raw spin configurations points to some of the limitations of PCA. Finally, we employ a nonlinear unsupervised machine learning procedure, the "autoencoder method," and we demonstrate that it too can be trained to capture phase transitions and critical points.
Yang, Lei; Wei, Ran; Shen, Henggen
2017-01-01
New principal component analysis (PCA) respirator fit test panels had been developed for current American and Chinese civilian workers based on anthropometric surveys. The PCA panels used the first two principal components (PCs) obtained from a set of 10 facial dimensions. Although the PCA panels for American and Chinese subjects adopted the bivairate framework with two PCs, the number of the PCs retained in the PCA analysis was different between Chinese subjects and Americans. For the Chinese youth group, the third PC should be retained in the PCA analysis for developing new fit test panels. In this article, an additional number label (ANL) is used to explain the third PC in PCA analysis when the first two PCs are used to construct the PCA half-facepiece respirator fit test panel for Chinese group. The three-dimensional box-counting method is proposed to estimate the ANLs by calculating fractal dimensions of the facial anthropometric data of the Chinese youth. The linear regression coefficients of scale-free range R 2 are all over 0.960, which demonstrates that the facial anthropometric data of the Chinese youth has fractal characteristic. The youth subjects born in Henan province has an ANL of 2.002, which is lower than the composite facial anthropometric data of Chinese subjects born in many provinces. Hence, Henan youth subjects have the self-similar facial anthropometric characteristic and should use the particular ANL (2.002) as the important tool along with using the PCA panel. The ANL method proposed in this article not only provides a new methodology in quantifying the characteristics of facial anthropometric dimensions for any ethnic/racial group, but also extends the scope of PCA panel studies to higher dimensions.
Berg Soto, Alvaro; Marsh, Helene; Everingham, Yvette; Smith, Joshua N; Parra, Guido J; Noad, Michael
2014-08-01
Australian snubfin and Indo-Pacific humpback dolphins co-occur throughout most of their range in coastal waters of tropical Australia. Little is known of their ecology or acoustic repertoires. Vocalizations from humpback and snubfin dolphins were recorded in two locations along the Queensland coast during 2008 and 2010 to describe their vocalizations and evaluate the acoustic differences between these two species. Broad vocalization types were categorized qualitatively. Both species produced click trains burst pulses and whistles. Principal component analysis of the nine acoustic variables extracted from the whistles produced nine principal components that were input into discriminant function analyses to classify 96% of humpback dolphin whistles and about 78% of snubfin dolphin calls correctly. Results indicate clear acoustic differences between the vocal whistle repertoires of these two species. A stepwise routine identified two principal components as significantly distinguishable between whistles of each species: frequency parameters and frequency trend ratio. The capacity to identify these species using acoustic monitoring techniques has the potential to provide information on presence/absence, habitat use and relative abundance for each species.
Perturbation analyses of intermolecular interactions
NASA Astrophysics Data System (ADS)
Koyama, Yohei M.; Kobayashi, Tetsuya J.; Ueda, Hiroki R.
2011-08-01
Conformational fluctuations of a protein molecule are important to its function, and it is known that environmental molecules, such as water molecules, ions, and ligand molecules, significantly affect the function by changing the conformational fluctuations. However, it is difficult to systematically understand the role of environmental molecules because intermolecular interactions related to the conformational fluctuations are complicated. To identify important intermolecular interactions with regard to the conformational fluctuations, we develop herein (i) distance-independent and (ii) distance-dependent perturbation analyses of the intermolecular interactions. We show that these perturbation analyses can be realized by performing (i) a principal component analysis using conditional expectations of truncated and shifted intermolecular potential energy terms and (ii) a functional principal component analysis using products of intermolecular forces and conditional cumulative densities. We refer to these analyses as intermolecular perturbation analysis (IPA) and distance-dependent intermolecular perturbation analysis (DIPA), respectively. For comparison of the IPA and the DIPA, we apply them to the alanine dipeptide isomerization in explicit water. Although the first IPA principal components discriminate two states (the α state and PPII (polyproline II) + β states) for larger cutoff length, the separation between the PPII state and the β state is unclear in the second IPA principal components. On the other hand, in the large cutoff value, DIPA eigenvalues converge faster than that for IPA and the top two DIPA principal components clearly identify the three states. By using the DIPA biplot, the contributions of the dipeptide-water interactions to each state are analyzed systematically. Since the DIPA improves the state identification and the convergence rate with retaining distance information, we conclude that the DIPA is a more practical method compared with the IPA. To test the feasibility of the DIPA for larger molecules, we apply the DIPA to the ten-residue chignolin folding in explicit water. The top three principal components identify the four states (native state, two misfolded states, and unfolded state) and their corresponding eigenfunctions identify important chignolin-water interactions to each state. Thus, the DIPA provides the practical method to identify conformational states and their corresponding important intermolecular interactions with distance information.
Perturbation analyses of intermolecular interactions.
Koyama, Yohei M; Kobayashi, Tetsuya J; Ueda, Hiroki R
2011-08-01
Conformational fluctuations of a protein molecule are important to its function, and it is known that environmental molecules, such as water molecules, ions, and ligand molecules, significantly affect the function by changing the conformational fluctuations. However, it is difficult to systematically understand the role of environmental molecules because intermolecular interactions related to the conformational fluctuations are complicated. To identify important intermolecular interactions with regard to the conformational fluctuations, we develop herein (i) distance-independent and (ii) distance-dependent perturbation analyses of the intermolecular interactions. We show that these perturbation analyses can be realized by performing (i) a principal component analysis using conditional expectations of truncated and shifted intermolecular potential energy terms and (ii) a functional principal component analysis using products of intermolecular forces and conditional cumulative densities. We refer to these analyses as intermolecular perturbation analysis (IPA) and distance-dependent intermolecular perturbation analysis (DIPA), respectively. For comparison of the IPA and the DIPA, we apply them to the alanine dipeptide isomerization in explicit water. Although the first IPA principal components discriminate two states (the α state and PPII (polyproline II) + β states) for larger cutoff length, the separation between the PPII state and the β state is unclear in the second IPA principal components. On the other hand, in the large cutoff value, DIPA eigenvalues converge faster than that for IPA and the top two DIPA principal components clearly identify the three states. By using the DIPA biplot, the contributions of the dipeptide-water interactions to each state are analyzed systematically. Since the DIPA improves the state identification and the convergence rate with retaining distance information, we conclude that the DIPA is a more practical method compared with the IPA. To test the feasibility of the DIPA for larger molecules, we apply the DIPA to the ten-residue chignolin folding in explicit water. The top three principal components identify the four states (native state, two misfolded states, and unfolded state) and their corresponding eigenfunctions identify important chignolin-water interactions to each state. Thus, the DIPA provides the practical method to identify conformational states and their corresponding important intermolecular interactions with distance information.
Using Structural Equation Modeling To Fit Models Incorporating Principal Components.
ERIC Educational Resources Information Center
Dolan, Conor; Bechger, Timo; Molenaar, Peter
1999-01-01
Considers models incorporating principal components from the perspectives of structural-equation modeling. These models include the following: (1) the principal-component analysis of patterned matrices; (2) multiple analysis of variance based on principal components; and (3) multigroup principal-components analysis. Discusses fitting these models…
Use of Principal Components Analysis to Explain Controls on Nutrient Fluxes to the Chesapeake Bay
NASA Astrophysics Data System (ADS)
Rice, K. C.; Mills, A. L.
2017-12-01
The Chesapeake Bay watershed, on the east coast of the United States, encompasses about 166,000-square kilometers (km2) of diverse land use, which includes a mixture of forested, agricultural, and developed land. The watershed is now managed under a Total Daily Maximum Load (TMDL), which requires implementation of management actions by 2025 that are sufficient to reduce nitrogen, phosphorus, and suspended-sediment fluxes to the Chesapeake Bay and restore the bay's water quality. We analyzed nutrient and sediment data along with land-use and climatic variables in nine sub watersheds to better understand the drivers of flux within the watershed and to provide relevant management implications. The nine sub watersheds range in area from 300 to 30,000 km2, and the analysis period was 1985-2014. The 31 variables specific to each sub watershed were highly statistically significantly correlated, so Principal Components Analysis was used to reduce the dimensionality of the dataset. The analysis revealed that about 80% of the variability in the whole dataset can be explained by discharge, flux, and concentration of nutrients and sediment. The first two principal components (PCs) explained about 68% of the total variance. PC1 loaded strongly on discharge and flux, and PC2 loaded on concentration. The PC scores of both PC1 and PC2 varied by season. Subsequent analysis of PC1 scores versus PC2 scores, broken out by sub watershed, revealed management implications. Some of the largest sub watersheds are largely driven by discharge, and consequently large fluxes. In contrast, some of the smaller sub watersheds are more variable in nutrient concentrations than discharge and flux. Our results suggest that, given no change in discharge, a reduction in nutrient flux to the streams in the smaller watersheds could result in a proportionately larger decrease in fluxes of nutrients down the river to the bay, than in the larger watersheds.
Q-mode versus R-mode principal component analysis for linear discriminant analysis (LDA)
NASA Astrophysics Data System (ADS)
Lee, Loong Chuen; Liong, Choong-Yeun; Jemain, Abdul Aziz
2017-05-01
Many literature apply Principal Component Analysis (PCA) as either preliminary visualization or variable con-struction methods or both. Focus of PCA can be on the samples (R-mode PCA) or variables (Q-mode PCA). Traditionally, R-mode PCA has been the usual approach to reduce high-dimensionality data before the application of Linear Discriminant Analysis (LDA), to solve classification problems. Output from PCA composed of two new matrices known as loadings and scores matrices. Each matrix can then be used to produce a plot, i.e. loadings plot aids identification of important variables whereas scores plot presents spatial distribution of samples on new axes that are also known as Principal Components (PCs). Fundamentally, the scores matrix always be the input variables for building classification model. A recent paper uses Q-mode PCA but the focus of analysis was not on the variables but instead on the samples. As a result, the authors have exchanged the use of both loadings and scores plots in which clustering of samples was studied using loadings plot whereas scores plot has been used to identify important manifest variables. Therefore, the aim of this study is to statistically validate the proposed practice. Evaluation is based on performance of external error obtained from LDA models according to number of PCs. On top of that, bootstrapping was also conducted to evaluate the external error of each of the LDA models. Results show that LDA models produced by PCs from R-mode PCA give logical performance and the matched external error are also unbiased whereas the ones produced with Q-mode PCA show the opposites. With that, we concluded that PCs produced from Q-mode is not statistically stable and thus should not be applied to problems of classifying samples, but variables. We hope this paper will provide some insights on the disputable issues.
Stress changes ahead of an advancing tunnel
Abel, J.F.; Lee, F.T.
1973-01-01
Instrumentation placed ahead of three model tunnels in the laboratory and ahead of a crosscut driven in a metamorphic rock mass detected stress changes several tunnel diameters ahead of the tunnel face. Stress changes were detected 4 diameters ahead of a model tunnel drilled into nearly elastic acrylic, 2??50 diameters ahead of a model tunnel drilled into concrete, and 2 diameters ahead of a model tunnel drilled into Silver Plume Granite. Stress changes were detected 7??50 diameters ahead of a crosscut driven in jointed, closely foliated gneisses and gneissic granites in an experimental mine at Idaho Springs, Colorado. These results contrast markedly with a theoretical elastic estimate of the onset of detectable stress changes at 1 tunnel diameter ahead of the tunnel face. A small compressive stress concentration was detected 2 diameters ahead of the model tunnel in acrylic, 1.25 diameters ahead of the model tunnel in concrete, and 1 diameter ahead of the model tunnel in granite. A similar stress peak was detected about 6 diameters ahead of the crosscut. No such stress peak is predicted from elastic theory. The 3-dimensional in situ stress determined in the field demonstrate that geologic structure controls stress orientations in the metamorphic rock mass. Two of the computed principal stresses are parallel to the foliation and the other principal stress is normal to it. The principal stress orientations vary approximately as the foliation attitude varies. The average horizontal stress components and the average vertical stress component are three times and twice as large, respectively, as those predicted from the overburden load. An understanding of the measured stress field appears to require the application of either tectonic or residual stress components, or both. Laboratory studies indicate the presence of proportionately large residual stresses. Mining may have triggered the release of strain energy, which is controlled by geologic structure. ?? 1973.
Lithium Causes G2 Arrest of Renal Principal Cells
de Groot, Theun; Alsady, Mohammad; Jaklofsky, Marcel; Otte-Höller, Irene; Baumgarten, Ruben; Giles, Rachel H.
2014-01-01
Vasopressin-regulated expression and insertion of aquaporin-2 channels in the luminal membrane of renal principal cells is essential for urine concentration. Lithium affects urine concentrating ability, and approximately 20% of patients treated with lithium develop nephrogenic diabetes insipidus (NDI), a disorder characterized by polyuria and polydipsia. Lithium-induced NDI is caused by aquaporin-2 downregulation and a reduced ratio of principal/intercalated cells, yet lithium induces principal cell proliferation. Here, we studied how lithium-induced principal cell proliferation can lead to a reduced ratio of principal/intercalated cells using two-dimensional and three-dimensional polarized cultures of mouse renal collecting duct cells and mice treated with clinically relevant lithium concentrations. DNA image cytometry and immunoblotting revealed that lithium initiated proliferation of mouse renal collecting duct cells but also increased the G2/S ratio, indicating G2/M phase arrest. In mice, treatment with lithium for 4, 7, 10, or 13 days led to features of NDI and an increase in the number of principal cells expressing PCNA in the papilla. Remarkably, 30%–40% of the PCNA-positive principal cells also expressed pHistone-H3, a late G2/M phase marker detected in approximately 20% of cells during undisturbed proliferation. Our data reveal that lithium treatment initiates proliferation of renal principal cells but that a significant percentage of these cells are arrested in the late G2 phase, which explains the reduced principal/intercalated cell ratio and may identify the molecular pathway underlying the development of lithium-induced renal fibrosis. PMID:24408872
Influential Observations in Principal Factor Analysis.
ERIC Educational Resources Information Center
Tanaka, Yutaka; Odaka, Yoshimasa
1989-01-01
A method is proposed for detecting influential observations in iterative principal factor analysis. Theoretical influence functions are derived for two components of the common variance decomposition. The major mathematical tool is the influence function derived by Tanaka (1988). (SLD)
Rosacea assessment by erythema index and principal component analysis segmentation maps
NASA Astrophysics Data System (ADS)
Kuzmina, Ilona; Rubins, Uldis; Saknite, Inga; Spigulis, Janis
2017-12-01
RGB images of rosacea were analyzed using segmentation maps of principal component analysis (PCA) and erythema index (EI). Areas of segmented clusters were compared to Clinician's Erythema Assessment (CEA) values given by two dermatologists. The results show that visible blood vessels are segmented more precisely on maps of the erythema index and the third principal component (PC3). In many cases, a distribution of clusters on EI and PC3 maps are very similar. Mean values of clusters' areas on these maps show a decrease of the area of blood vessels and erythema and an increase of lighter skin area after the therapy for the patients with diagnosis CEA = 2 on the first visit and CEA=1 on the second visit. This study shows that EI and PC3 maps are more useful than the maps of the first (PC1) and second (PC2) principal components for indicating vascular structures and erythema on the skin of rosacea patients and therapy monitoring.
Multilevel sparse functional principal component analysis.
Di, Chongzhi; Crainiceanu, Ciprian M; Jank, Wolfgang S
2014-01-29
We consider analysis of sparsely sampled multilevel functional data, where the basic observational unit is a function and data have a natural hierarchy of basic units. An example is when functions are recorded at multiple visits for each subject. Multilevel functional principal component analysis (MFPCA; Di et al. 2009) was proposed for such data when functions are densely recorded. Here we consider the case when functions are sparsely sampled and may contain only a few observations per function. We exploit the multilevel structure of covariance operators and achieve data reduction by principal component decompositions at both between and within subject levels. We address inherent methodological differences in the sparse sampling context to: 1) estimate the covariance operators; 2) estimate the functional principal component scores; 3) predict the underlying curves. Through simulations the proposed method is able to discover dominating modes of variations and reconstruct underlying curves well even in sparse settings. Our approach is illustrated by two applications, the Sleep Heart Health Study and eBay auctions.
Kakio, Tomoko; Nagase, Hitomi; Takaoka, Takashi; Yoshida, Naoko; Hirakawa, Junichi; Macha, Susan; Hiroshima, Takashi; Ikeda, Yukihiro; Tsuboi, Hirohito; Kimura, Kazuko
2018-06-01
The World Health Organization has warned that substandard and falsified medical products (SFs) can harm patients and fail to treat the diseases for which they were intended, and they affect every region of the world, leading to loss of confidence in medicines, health-care providers, and health systems. Therefore, development of analytical procedures to detect SFs is extremely important. In this study, we investigated the quality of pharmaceutical tablets containing the antihypertensive candesartan cilexetil, collected in China, Indonesia, Japan, and Myanmar, using the Japanese pharmacopeial analytical procedures for quality control, together with principal component analysis (PCA) of Raman spectrum obtained with handheld Raman spectrometer. Some samples showed delayed dissolution and failed to meet the pharmacopeial specification, whereas others failed the assay test. These products appeared to be substandard. Principal component analysis showed that all Raman spectra could be explained in terms of two components: the amount of the active pharmaceutical ingredient and the kinds of excipients. Principal component analysis score plot indicated one substandard, and the falsified tablets have similar principal components in Raman spectra, in contrast to authentic products. The locations of samples within the PCA score plot varied according to the source country, suggesting that manufacturers in different countries use different excipients. Our results indicate that the handheld Raman device will be useful for detection of SFs in the field. Principal component analysis of that Raman data clarify the difference in chemical properties between good quality products and SFs that circulate in the Asian market.
Karpuzcu, M Ekrem; Fairbairn, David; Arnold, William A; Barber, Brian L; Kaufenberg, Elizabeth; Koskinen, William C; Novak, Paige J; Rice, Pamela J; Swackhamer, Deborah L
2014-01-01
Principal components analysis (PCA) was used to identify sources of emerging organic contaminants in the Zumbro River watershed in Southeastern Minnesota. Two main principal components (PCs) were identified, which together explained more than 50% of the variance in the data. Principal Component 1 (PC1) was attributed to urban wastewater-derived sources, including municipal wastewater and residential septic tank effluents, while Principal Component 2 (PC2) was attributed to agricultural sources. The variances of the concentrations of cotinine, DEET and the prescription drugs carbamazepine, erythromycin and sulfamethoxazole were best explained by PC1, while the variances of the concentrations of the agricultural pesticides atrazine, metolachlor and acetochlor were best explained by PC2. Mixed use compounds carbaryl, iprodione and daidzein did not specifically group with either PC1 or PC2. Furthermore, despite the fact that caffeine and acetaminophen have been historically associated with human use, they could not be attributed to a single dominant land use category (e.g., urban/residential or agricultural). Contributions from septic systems did not clarify the source for these two compounds, suggesting that additional sources, such as runoff from biosolid-amended soils, may exist. Based on these results, PCA may be a useful way to broadly categorize the sources of new and previously uncharacterized emerging contaminants or may help to clarify transport pathways in a given area. Acetaminophen and caffeine were not ideal markers for urban/residential contamination sources in the study area and may need to be reconsidered as such in other areas as well.
JOINT AND INDIVIDUAL VARIATION EXPLAINED (JIVE) FOR INTEGRATED ANALYSIS OF MULTIPLE DATA TYPES.
Lock, Eric F; Hoadley, Katherine A; Marron, J S; Nobel, Andrew B
2013-03-01
Research in several fields now requires the analysis of datasets in which multiple high-dimensional types of data are available for a common set of objects. In particular, The Cancer Genome Atlas (TCGA) includes data from several diverse genomic technologies on the same cancerous tumor samples. In this paper we introduce Joint and Individual Variation Explained (JIVE), a general decomposition of variation for the integrated analysis of such datasets. The decomposition consists of three terms: a low-rank approximation capturing joint variation across data types, low-rank approximations for structured variation individual to each data type, and residual noise. JIVE quantifies the amount of joint variation between data types, reduces the dimensionality of the data, and provides new directions for the visual exploration of joint and individual structure. The proposed method represents an extension of Principal Component Analysis and has clear advantages over popular two-block methods such as Canonical Correlation Analysis and Partial Least Squares. A JIVE analysis of gene expression and miRNA data on Glioblastoma Multiforme tumor samples reveals gene-miRNA associations and provides better characterization of tumor types.
Buonaccorsi, G A; Rose, C J; O'Connor, J P B; Roberts, C; Watson, Y; Jackson, A; Jayson, G C; Parker, G J M
2010-01-01
Clinical trials of anti-angiogenic and vascular-disrupting agents often use biomarkers derived from DCE-MRI, typically reporting whole-tumor summary statistics and so overlooking spatial parameter variations caused by tissue heterogeneity. We present a data-driven segmentation method comprising tracer-kinetic model-driven registration for motion correction, conversion from MR signal intensity to contrast agent concentration for cross-visit normalization, iterative principal components analysis for imputation of missing data and dimensionality reduction, and statistical outlier detection using the minimum covariance determinant to obtain a robust Mahalanobis distance. After applying these techniques we cluster in the principal components space using k-means. We present results from a clinical trial of a VEGF inhibitor, using time-series data selected because of problems due to motion and outlier time series. We obtained spatially-contiguous clusters that map to regions with distinct microvascular characteristics. This methodology has the potential to uncover localized effects in trials using DCE-MRI-based biomarkers.
Šašiċ, Slobodan; Ojakovo, Peter; Warman, Martin; Sanghvi, Tapan
2013-09-01
Raman chemical mapping was used to determine the distribution of magnesium stearate, a lubricant, on the surface of tablets. The lubrication was carried out via a punch-face lubrication system with different spraying rates applied on placebo and active-containing tablets. Principal component analysis was used for decomposing the matrix of Raman mapping spectra. Some of the loadings associated with minuscule variation in the data significantly overlap with the Raman spectrum of magnesium stearate in placebo tablets and allow for imaging the domains of magnesium stearate via corresponding scores. Despite the negligible variation accounted for by respective principal components, the score images seem reliable as demonstrated through thresholding the one-dimensional representation and the spectra of the hot pixels that show a weak but perceivable magnesium stearate band at 1295 cm(-1). The same approach was applied on the active formulation, but no magnesium stearate was identified, presumably due to overwhelming concentration and spectral contribution of the active pharmaceutical ingredient.
Characterizing Time Series Data Diversity for Wind Forecasting: Preprint
DOE Office of Scientific and Technical Information (OSTI.GOV)
Hodge, Brian S; Chartan, Erol Kevin; Feng, Cong
Wind forecasting plays an important role in integrating variable and uncertain wind power into the power grid. Various forecasting models have been developed to improve the forecasting accuracy. However, it is challenging to accurately compare the true forecasting performances from different methods and forecasters due to the lack of diversity in forecasting test datasets. This paper proposes a time series characteristic analysis approach to visualize and quantify wind time series diversity. The developed method first calculates six time series characteristic indices from various perspectives. Then the principal component analysis is performed to reduce the data dimension while preserving the importantmore » information. The diversity of the time series dataset is visualized by the geometric distribution of the newly constructed principal component space. The volume of the 3-dimensional (3D) convex polytope (or the length of 1D number axis, or the area of the 2D convex polygon) is used to quantify the time series data diversity. The method is tested with five datasets with various degrees of diversity.« less
A Principal Component Analysis of 39 Scientific Impact Measures
Bollen, Johan; Van de Sompel, Herbert
2009-01-01
Background The impact of scientific publications has traditionally been expressed in terms of citation counts. However, scientific activity has moved online over the past decade. To better capture scientific impact in the digital era, a variety of new impact measures has been proposed on the basis of social network analysis and usage log data. Here we investigate how these new measures relate to each other, and how accurately and completely they express scientific impact. Methodology We performed a principal component analysis of the rankings produced by 39 existing and proposed measures of scholarly impact that were calculated on the basis of both citation and usage log data. Conclusions Our results indicate that the notion of scientific impact is a multi-dimensional construct that can not be adequately measured by any single indicator, although some measures are more suitable than others. The commonly used citation Impact Factor is not positioned at the core of this construct, but at its periphery, and should thus be used with caution. PMID:19562078
Total Electron Content forecast model over Australia
NASA Astrophysics Data System (ADS)
Bouya, Zahra; Terkildsen, Michael; Francis, Matthew
Ionospheric perturbations can cause serious propagation errors in modern radio systems such as Global Navigation Satellite Systems (GNSS). Forecasting ionospheric parameters is helpful to estimate potential degradation of the performance of these systems. Our purpose is to establish an Australian Regional Total Electron Content (TEC) forecast model at IPS. In this work we present an approach based on the combined use of the Principal Component Analysis (PCA) and Artificial Neural Network (ANN) to predict future TEC values. PCA is used to reduce the dimensionality of the original TEC data by mapping it into its eigen-space. In this process the top- 5 eigenvectors are chosen to reflect the directions of the maximum variability. An ANN approach was then used for the multicomponent prediction. We outline the design of the ANN model with its parameters. A number of activation functions along with different spectral ranges and different numbers of Principal Components (PCs) were tested to find the PCA-ANN models reaching the best results. Keywords: GNSS, Space Weather, Regional, Forecast, PCA, ANN.
Anomalous magnetic structure and spin dynamics in magnetoelectric LiFePO 4
Toft-Petersen, Rasmus; Reehuis, Manfred; Jensen, Thomas B. S.; ...
2015-07-06
We report significant details of the magnetic structure and spin dynamics of LiFePO 4 obtained by single-crystal neutron scattering. Our results confirm a previously reported collinear rotation of the spins away from the principal b axis, and they determine that the rotation is toward the a axis. In addition, we find a significant spin-canting component along c. Furthermore, the possible causes of these components are discussed, and their significance for the magnetoelectric effect is analyzed. Inelastic neutron scattering along the three principal directions reveals a highly anisotropic hard plane consistent with earlier susceptibility measurements. While using a spin Hamiltonian, wemore » show that the spin dimensionality is intermediate between XY- and Ising-like, with an easy b axis and a hard c axis. As a result, it is shown that both next-nearest neighbor exchange couplings in the bc plane are in competition with the strongest nearest neighbor coupling.« less
Wang, Jing-Jing; Wu, Hai-Feng; Sun, Tao; Li, Xia; Wang, Wei; Tao, Li-Xin; Huo, Da; Lv, Ping-Xin; He, Wen; Guo, Xiu-Hua
2013-01-01
Lung cancer, one of the leading causes of cancer-related deaths, usually appears as solitary pulmonary nodules (SPNs) which are hard to diagnose using the naked eye. In this paper, curvelet-based textural features and clinical parameters are used with three prediction models [a multilevel model, a least absolute shrinkage and selection operator (LASSO) regression method, and a support vector machine (SVM)] to improve the diagnosis of benign and malignant SPNs. Dimensionality reduction of the original curvelet-based textural features was achieved using principal component analysis. In addition, non-conditional logistical regression was used to find clinical predictors among demographic parameters and morphological features. The results showed that, combined with 11 clinical predictors, the accuracy rates using 12 principal components were higher than those using the original curvelet-based textural features. To evaluate the models, 10-fold cross validation and back substitution were applied. The results obtained, respectively, were 0.8549 and 0.9221 for the LASSO method, 0.9443 and 0.9831 for SVM, and 0.8722 and 0.9722 for the multilevel model. All in all, it was found that using curvelet-based textural features after dimensionality reduction and using clinical predictors, the highest accuracy rate was achieved with SVM. The method may be used as an auxiliary tool to differentiate between benign and malignant SPNs in CT images.
The use of experimental structures to model protein dynamics.
Katebi, Ataur R; Sankar, Kannan; Jia, Kejue; Jernigan, Robert L
2015-01-01
The number of solved protein structures submitted in the Protein Data Bank (PDB) has increased dramatically in recent years. For some specific proteins, this number is very high-for example, there are over 550 solved structures for HIV-1 protease, one protein that is essential for the life cycle of human immunodeficiency virus (HIV) which causes acquired immunodeficiency syndrome (AIDS) in humans. The large number of structures for the same protein and its variants include a sample of different conformational states of the protein. A rich set of structures solved experimentally for the same protein has information buried within the dataset that can explain the functional dynamics and structural mechanism of the protein. To extract the dynamics information and functional mechanism from the experimental structures, this chapter focuses on two methods-Principal Component Analysis (PCA) and Elastic Network Models (ENM). PCA is a widely used statistical dimensionality reduction technique to classify and visualize high-dimensional data. On the other hand, ENMs are well-established simple biophysical method for modeling the functionally important global motions of proteins. This chapter covers the basics of these two. Moreover, an improved ENM version that utilizes the variations found within a given set of structures for a protein is described. As a practical example, we have extracted the functional dynamics and mechanism of HIV-1 protease dimeric structure by using a set of 329 PDB structures of this protein. We have described, step by step, how to select a set of protein structures, how to extract the needed information from the PDB files for PCA, how to extract the dynamics information using PCA, how to calculate ENM modes, how to measure the congruency between the dynamics computed from the principal components (PCs) and the ENM modes, and how to compute entropies using the PCs. We provide the computer programs or references to software tools to accomplish each step and show how to use these programs and tools. We also include computer programs to generate movies based on PCs and ENM modes and describe how to visualize them.
The Use of Experimental Structures to Model Protein Dynamics
Katebi, Ataur R.; Sankar, Kannan; Jia, Kejue; Jernigan, Robert L.
2014-01-01
Summary The number of solved protein structures submitted in the Protein Data Bank (PDB) has increased dramatically in recent years. For some specific proteins, this number is very high – for example, there are over 550 solved structures for HIV-1 protease, one protein that is essential for the life cycle of human immunodeficiency virus (HIV) which causes acquired immunodeficiency syndrome (AIDS) in humans. The large number of structures for the same protein and its variants include a sample of different conformational states of the protein. A rich set of structures solved experimentally for the same protein has information buried within the dataset that can explain the functional dynamics and structural mechanism of the protein. To extract the dynamics information and functional mechanism from the experimental structures, this chapter focuses on two methods – Principal Component Analysis (PCA) and Elastic Network Models (ENM). PCA is a widely used statistical dimensionality reduction technique to classify and visualize high-dimensional data. On the other hand, ENMs are well-established simple biophysical method for modeling the functionally important global motions of proteins. This chapter covers the basics of these two. Moreover, an improved ENM version that utilizes the variations found within a given set of structures for a protein is described. As a practical example, we have extracted the functional dynamics and mechanism of HIV-1 protease dimeric structure by using a set of 329 PDB structures of this protein. We have described, step by step, how to select a set of protein structures, how to extract the needed information from the PDB files for PCA, how to extract the dynamics information using PCA, how to calculate ENM modes, how to measure the congruency between the dynamics computed from the principal components (PCs) and the ENM modes, and how to compute entropies using the PCs. We provide the computer programs or references to software tools to accomplish each step and show how to use these programs and tools. We also include computer programs to generate movies based on PCs and ENM modes and describe how to visualize them. PMID:25330965
Model Reduction via Principe Component Analysis and Markov Chain Monte Carlo (MCMC) Methods
NASA Astrophysics Data System (ADS)
Gong, R.; Chen, J.; Hoversten, M. G.; Luo, J.
2011-12-01
Geophysical and hydrogeological inverse problems often include a large number of unknown parameters, ranging from hundreds to millions, depending on parameterization and problems undertaking. This makes inverse estimation and uncertainty quantification very challenging, especially for those problems in two- or three-dimensional spatial domains. Model reduction technique has the potential of mitigating the curse of dimensionality by reducing total numbers of unknowns while describing the complex subsurface systems adequately. In this study, we explore the use of principal component analysis (PCA) and Markov chain Monte Carlo (MCMC) sampling methods for model reduction through the use of synthetic datasets. We compare the performances of three different but closely related model reduction approaches: (1) PCA methods with geometric sampling (referred to as 'Method 1'), (2) PCA methods with MCMC sampling (referred to as 'Method 2'), and (3) PCA methods with MCMC sampling and inclusion of random effects (referred to as 'Method 3'). We consider a simple convolution model with five unknown parameters as our goal is to understand and visualize the advantages and disadvantages of each method by comparing their inversion results with the corresponding analytical solutions. We generated synthetic data with noise added and invert them under two different situations: (1) the noised data and the covariance matrix for PCA analysis are consistent (referred to as the unbiased case), and (2) the noise data and the covariance matrix are inconsistent (referred to as biased case). In the unbiased case, comparison between the analytical solutions and the inversion results show that all three methods provide good estimates of the true values and Method 1 is computationally more efficient. In terms of uncertainty quantification, Method 1 performs poorly because of relatively small number of samples obtained, Method 2 performs best, and Method 3 overestimates uncertainty due to inclusion of random effects. However, in the biased case, only Method 3 correctly estimates all the unknown parameters, and both Methods 1 and 2 provide wrong values for the biased parameters. The synthetic case study demonstrates that if the covariance matrix for PCA analysis is inconsistent with true models, the PCA methods with geometric or MCMC sampling will provide incorrect estimates.
Fast principal component analysis for stacking seismic data
NASA Astrophysics Data System (ADS)
Wu, Juan; Bai, Min
2018-04-01
Stacking seismic data plays an indispensable role in many steps of the seismic data processing and imaging workflow. Optimal stacking of seismic data can help mitigate seismic noise and enhance the principal components to a great extent. Traditional average-based seismic stacking methods cannot obtain optimal performance when the ambient noise is extremely strong. We propose a principal component analysis (PCA) algorithm for stacking seismic data without being sensitive to noise level. Considering the computational bottleneck of the classic PCA algorithm in processing massive seismic data, we propose an efficient PCA algorithm to make the proposed method readily applicable for industrial applications. Two numerically designed examples and one real seismic data are used to demonstrate the performance of the presented method.
Online dimensionality reduction using competitive learning and Radial Basis Function network.
Tomenko, Vladimir
2011-06-01
The general purpose dimensionality reduction method should preserve data interrelations at all scales. Additional desired features include online projection of new data, processing nonlinearly embedded manifolds and large amounts of data. The proposed method, called RBF-NDR, combines these features. RBF-NDR is comprised of two modules. The first module learns manifolds by utilizing modified topology representing networks and geodesic distance in data space and approximates sampled or streaming data with a finite set of reference patterns, thus achieving scalability. Using input from the first module, the dimensionality reduction module constructs mappings between observation and target spaces. Introduction of specific loss function and synthesis of the training algorithm for Radial Basis Function network results in global preservation of data structures and online processing of new patterns. The RBF-NDR was applied for feature extraction and visualization and compared with Principal Component Analysis (PCA), neural network for Sammon's projection (SAMANN) and Isomap. With respect to feature extraction, the method outperformed PCA and yielded increased performance of the model describing wastewater treatment process. As for visualization, RBF-NDR produced superior results compared to PCA and SAMANN and matched Isomap. For the Topic Detection and Tracking corpus, the method successfully separated semantically different topics. Copyright © 2011 Elsevier Ltd. All rights reserved.
TESTING HIGH-DIMENSIONAL COVARIANCE MATRICES, WITH APPLICATION TO DETECTING SCHIZOPHRENIA RISK GENES
Zhu, Lingxue; Lei, Jing; Devlin, Bernie; Roeder, Kathryn
2017-01-01
Scientists routinely compare gene expression levels in cases versus controls in part to determine genes associated with a disease. Similarly, detecting case-control differences in co-expression among genes can be critical to understanding complex human diseases; however statistical methods have been limited by the high dimensional nature of this problem. In this paper, we construct a sparse-Leading-Eigenvalue-Driven (sLED) test for comparing two high-dimensional covariance matrices. By focusing on the spectrum of the differential matrix, sLED provides a novel perspective that accommodates what we assume to be common, namely sparse and weak signals in gene expression data, and it is closely related with Sparse Principal Component Analysis. We prove that sLED achieves full power asymptotically under mild assumptions, and simulation studies verify that it outperforms other existing procedures under many biologically plausible scenarios. Applying sLED to the largest gene-expression dataset obtained from post-mortem brain tissue from Schizophrenia patients and controls, we provide a novel list of genes implicated in Schizophrenia and reveal intriguing patterns in gene co-expression change for Schizophrenia subjects. We also illustrate that sLED can be generalized to compare other gene-gene “relationship” matrices that are of practical interest, such as the weighted adjacency matrices. PMID:29081874
Zhu, Lingxue; Lei, Jing; Devlin, Bernie; Roeder, Kathryn
2017-09-01
Scientists routinely compare gene expression levels in cases versus controls in part to determine genes associated with a disease. Similarly, detecting case-control differences in co-expression among genes can be critical to understanding complex human diseases; however statistical methods have been limited by the high dimensional nature of this problem. In this paper, we construct a sparse-Leading-Eigenvalue-Driven (sLED) test for comparing two high-dimensional covariance matrices. By focusing on the spectrum of the differential matrix, sLED provides a novel perspective that accommodates what we assume to be common, namely sparse and weak signals in gene expression data, and it is closely related with Sparse Principal Component Analysis. We prove that sLED achieves full power asymptotically under mild assumptions, and simulation studies verify that it outperforms other existing procedures under many biologically plausible scenarios. Applying sLED to the largest gene-expression dataset obtained from post-mortem brain tissue from Schizophrenia patients and controls, we provide a novel list of genes implicated in Schizophrenia and reveal intriguing patterns in gene co-expression change for Schizophrenia subjects. We also illustrate that sLED can be generalized to compare other gene-gene "relationship" matrices that are of practical interest, such as the weighted adjacency matrices.
Micromechanical Aspects of Hydraulic Fracturing Processes
NASA Astrophysics Data System (ADS)
Galindo-torres, S. A.; Behraftar, S.; Scheuermann, A.; Li, L.; Williams, D.
2014-12-01
A micromechanical model is developed to simulate the hydraulic fracturing process. The model comprises two key components. Firstly, the solid matrix, assumed as a rock mass with pre-fabricated cracks, is represented by an array of bonded particles simulated by the Discrete Element Model (DEM)[1]. The interaction is ruled by the spheropolyhedra method, which was introduced by the authors previously and has been shown to realistically represent many of the features found in fracturing and communition processes. The second component is the fluid, which is modelled by the Lattice Boltzmann Method (LBM). It was recently coupled with the spheropolyhedra by the authors and validated. An advantage of this coupled LBM-DEM model is the control of many of the parameters of the fracturing fluid, such as its viscosity and the injection rate. To the best of the authors' knowledge this is the first application of such a coupled scheme for studying hydraulic fracturing[2]. In this first implementation, results are presented for a two-dimensional situation. Fig. 1 shows one snapshot of the LBM-DEM coupled simulation for the hydraulic fracturing where the elements with broken bonds can be identified and the fracture geometry quantified. The simulation involves a variation of the underground stress, particularly the difference between the two principal components of the stress tensor, to explore the effect on the fracture path. A second study focuses on the fluid viscosity to examine the effect of the time scales of different injection plans on the fracture geometry. The developed tool and the presented results have important implications for future studies of the hydraulic fracturing process and technology. references 1. Galindo-Torres, S.A., et al., Breaking processes in three-dimensional bonded granular materials with general shapes. Computer Physics Communications, 2012. 183(2): p. 266-277. 2. Galindo-Torres, S.A., A coupled Discrete Element Lattice Boltzmann Method for the simulation of fluid-solid interaction with particles of general shapes. Computer Methods in Applied Mechanics and Engineering, 2013. 265(0): p. 107-119.
Machine learning of frustrated classical spin models. I. Principal component analysis
NASA Astrophysics Data System (ADS)
Wang, Ce; Zhai, Hui
2017-10-01
This work aims at determining whether artificial intelligence can recognize a phase transition without prior human knowledge. If this were successful, it could be applied to, for instance, analyzing data from the quantum simulation of unsolved physical models. Toward this goal, we first need to apply the machine learning algorithm to well-understood models and see whether the outputs are consistent with our prior knowledge, which serves as the benchmark for this approach. In this work, we feed the computer data generated by the classical Monte Carlo simulation for the X Y model in frustrated triangular and union jack lattices, which has two order parameters and exhibits two phase transitions. We show that the outputs of the principal component analysis agree very well with our understanding of different orders in different phases, and the temperature dependences of the major components detect the nature and the locations of the phase transitions. Our work offers promise for using machine learning techniques to study sophisticated statistical models, and our results can be further improved by using principal component analysis with kernel tricks and the neural network method.
Sun, Jin; Kelbert, Anna; Egbert, G.D.
2015-01-01
Long-period global-scale electromagnetic induction studies of deep Earth conductivity are based almost exclusively on magnetovariational methods and require accurate models of external source spatial structure. We describe approaches to inverting for both the external sources and three-dimensional (3-D) conductivity variations and apply these methods to long-period (T≥1.2 days) geomagnetic observatory data. Our scheme involves three steps: (1) Observatory data from 60 years (only partly overlapping and with many large gaps) are reduced and merged into dominant spatial modes using a scheme based on frequency domain principal components. (2) Resulting modes are inverted for corresponding external source spatial structure, using a simplified conductivity model with radial variations overlain by a two-dimensional thin sheet. The source inversion is regularized using a physically based source covariance, generated through superposition of correlated tilted zonal (quasi-dipole) current loops, representing ionospheric source complexity smoothed by Earth rotation. Free parameters in the source covariance model are tuned by a leave-one-out cross-validation scheme. (3) The estimated data modes are inverted for 3-D Earth conductivity, assuming the source excitation estimated in step 2. Together, these developments constitute key components in a practical scheme for simultaneous inversion of the catalogue of historical and modern observatory data for external source spatial structure and 3-D Earth conductivity.
Alakent, Burak; Doruker, Pemra; Camurdan, Mehmet C
2004-09-08
Time series analysis is applied on the collective coordinates obtained from principal component analysis of independent molecular dynamics simulations of alpha-amylase inhibitor tendamistat and immunity protein of colicin E7 based on the Calpha coordinates history. Even though the principal component directions obtained for each run are considerably different, the dynamics information obtained from these runs are surprisingly similar in terms of time series models and parameters. There are two main differences in the dynamics of the two proteins: the higher density of low frequencies and the larger step sizes for the interminima motions of colicin E7 than those of alpha-amylase inhibitor, which may be attributed to the higher number of residues of colicin E7 and/or the structural differences of the two proteins. The cumulative density function of the low frequencies in each run conforms to the expectations from the normal mode analysis. When different runs of alpha-amylase inhibitor are projected on the same set of eigenvectors, it is found that principal components obtained from a certain conformational region of a protein has a moderate explanation power in other conformational regions and the local minima are similar to a certain extent, while the height of the energy barriers in between the minima significantly change. As a final remark, time series analysis tools are further exploited in this study with the motive of explaining the equilibrium fluctuations of proteins. Copyright 2004 American Institute of Physics
NASA Astrophysics Data System (ADS)
Alakent, Burak; Doruker, Pemra; Camurdan, Mehmet C.
2004-09-01
Time series analysis is applied on the collective coordinates obtained from principal component analysis of independent molecular dynamics simulations of α-amylase inhibitor tendamistat and immunity protein of colicin E7 based on the Cα coordinates history. Even though the principal component directions obtained for each run are considerably different, the dynamics information obtained from these runs are surprisingly similar in terms of time series models and parameters. There are two main differences in the dynamics of the two proteins: the higher density of low frequencies and the larger step sizes for the interminima motions of colicin E7 than those of α-amylase inhibitor, which may be attributed to the higher number of residues of colicin E7 and/or the structural differences of the two proteins. The cumulative density function of the low frequencies in each run conforms to the expectations from the normal mode analysis. When different runs of α-amylase inhibitor are projected on the same set of eigenvectors, it is found that principal components obtained from a certain conformational region of a protein has a moderate explanation power in other conformational regions and the local minima are similar to a certain extent, while the height of the energy barriers in between the minima significantly change. As a final remark, time series analysis tools are further exploited in this study with the motive of explaining the equilibrium fluctuations of proteins.
Stratmann, Philipp; Lakatos, Dominic; Albu-Schäffer, Alin
2016-01-01
There are multiple indications that the nervous system of animals tunes muscle output to exploit natural dynamics of the elastic locomotor system and the environment. This is an advantageous strategy especially in fast periodic movements, since the elastic elements store energy and increase energy efficiency and movement speed. Experimental evidence suggests that coordination among joints involves proprioceptive input and neuromodulatory influence originating in the brain stem. However, the neural strategies underlying the coordination of fast periodic movements remain poorly understood. Based on robotics control theory, we suggest that the nervous system implements a mechanism to accomplish coordination between joints by a linear coordinate transformation from the multi-dimensional space representing proprioceptive input at the joint level into a one-dimensional controller space. In this one-dimensional subspace, the movements of a whole limb can be driven by a single oscillating unit as simple as a reflex interneuron. The output of the oscillating unit is transformed back to joint space via the same transformation. The transformation weights correspond to the dominant principal component of the movement. In this study, we propose a biologically plausible neural network to exemplify that the central nervous system (CNS) may encode our controller design. Using theoretical considerations and computer simulations, we demonstrate that spike-timing-dependent plasticity (STDP) for the input mapping and serotonergic neuromodulation for the output mapping can extract the dominant principal component of sensory signals. Our simulations show that our network can reliably control mechanical systems of different complexity and increase the energy efficiency of ongoing cyclic movements. The proposed network is simple and consistent with previous biologic experiments. Thus, our controller could serve as a candidate to describe the neural control of fast, energy-efficient, periodic movements involving multiple coupled joints.
Stratmann, Philipp; Lakatos, Dominic; Albu-Schäffer, Alin
2016-01-01
There are multiple indications that the nervous system of animals tunes muscle output to exploit natural dynamics of the elastic locomotor system and the environment. This is an advantageous strategy especially in fast periodic movements, since the elastic elements store energy and increase energy efficiency and movement speed. Experimental evidence suggests that coordination among joints involves proprioceptive input and neuromodulatory influence originating in the brain stem. However, the neural strategies underlying the coordination of fast periodic movements remain poorly understood. Based on robotics control theory, we suggest that the nervous system implements a mechanism to accomplish coordination between joints by a linear coordinate transformation from the multi-dimensional space representing proprioceptive input at the joint level into a one-dimensional controller space. In this one-dimensional subspace, the movements of a whole limb can be driven by a single oscillating unit as simple as a reflex interneuron. The output of the oscillating unit is transformed back to joint space via the same transformation. The transformation weights correspond to the dominant principal component of the movement. In this study, we propose a biologically plausible neural network to exemplify that the central nervous system (CNS) may encode our controller design. Using theoretical considerations and computer simulations, we demonstrate that spike-timing-dependent plasticity (STDP) for the input mapping and serotonergic neuromodulation for the output mapping can extract the dominant principal component of sensory signals. Our simulations show that our network can reliably control mechanical systems of different complexity and increase the energy efficiency of ongoing cyclic movements. The proposed network is simple and consistent with previous biologic experiments. Thus, our controller could serve as a candidate to describe the neural control of fast, energy-efficient, periodic movements involving multiple coupled joints. PMID:27014051
Pouch, Alison M; Vergnat, Mathieu; McGarvey, Jeremy R; Ferrari, Giovanni; Jackson, Benjamin M; Sehgal, Chandra M; Yushkevich, Paul A; Gorman, Robert C; Gorman, Joseph H
2014-01-01
The basis of mitral annuloplasty ring design has progressed from qualitative surgical intuition to experimental and theoretical analysis of annular geometry with quantitative imaging techniques. In this work, we present an automated three-dimensional (3D) echocardiographic image analysis method that can be used to statistically assess variability in normal mitral annular geometry to support advancement in annuloplasty ring design. Three-dimensional patient-specific models of the mitral annulus were automatically generated from 3D echocardiographic images acquired from subjects with normal mitral valve structure and function. Geometric annular measurements including annular circumference, annular height, septolateral diameter, intercommissural width, and the annular height to intercommissural width ratio were automatically calculated. A mean 3D annular contour was computed, and principal component analysis was used to evaluate variability in normal annular shape. The following mean ± standard deviations were obtained from 3D echocardiographic image analysis: annular circumference, 107.0 ± 14.6 mm; annular height, 7.6 ± 2.8 mm; septolateral diameter, 28.5 ± 3.7 mm; intercommissural width, 33.0 ± 5.3 mm; and annular height to intercommissural width ratio, 22.7% ± 6.9%. Principal component analysis indicated that shape variability was primarily related to overall annular size, with more subtle variation in the skewness and height of the anterior annular peak, independent of annular diameter. Patient-specific 3D echocardiographic-based modeling of the human mitral valve enables statistical analysis of physiologically normal mitral annular geometry. The tool can potentially lead to the development of a new generation of annuloplasty rings that restore the diseased mitral valve annulus back to a truly normal geometry. Copyright © 2014 The Society of Thoracic Surgeons. Published by Elsevier Inc. All rights reserved.
Grimbergen, M C M; van Swol, C F P; Kendall, C; Verdaasdonk, R M; Stone, N; Bosch, J L H R
2010-01-01
The overall quality of Raman spectra in the near-infrared region, where biological samples are often studied, has benefited from various improvements to optical instrumentation over the past decade. However, obtaining ample spectral quality for analysis is still challenging due to device requirements and short integration times required for (in vivo) clinical applications of Raman spectroscopy. Multivariate analytical methods, such as principal component analysis (PCA) and linear discriminant analysis (LDA), are routinely applied to Raman spectral datasets to develop classification models. Data compression is necessary prior to discriminant analysis to prevent or decrease the degree of over-fitting. The logical threshold for the selection of principal components (PCs) to be used in discriminant analysis is likely to be at a point before the PCs begin to introduce equivalent signal and noise and, hence, include no additional value. Assessment of the signal-to-noise ratio (SNR) at a certain peak or over a specific spectral region will depend on the sample measured. Therefore, the mean SNR over the whole spectral region (SNR(msr)) is determined in the original spectrum as well as for spectra reconstructed from an increasing number of principal components. This paper introduces a method of assessing the influence of signal and noise from individual PC loads and indicates a method of selection of PCs for LDA. To evaluate this method, two data sets with different SNRs were used. The sets were obtained with the same Raman system and the same measurement parameters on bladder tissue collected during white light cystoscopy (set A) and fluorescence-guided cystoscopy (set B). This method shows that the mean SNR over the spectral range in the original Raman spectra of these two data sets is related to the signal and noise contribution of principal component loads. The difference in mean SNR over the spectral range can also be appreciated since fewer principal components can reliably be used in the low SNR data set (set B) compared to the high SNR data set (set A). Despite the fact that no definitive threshold could be found, this method may help to determine the cutoff for the number of principal components used in discriminant analysis. Future analysis of a selection of spectral databases using this technique will allow optimum thresholds to be selected for different applications and spectral data quality levels.
NASA Astrophysics Data System (ADS)
Ono, Junichi; Takada, Shoji; Saito, Shinji
2015-06-01
An analytical method based on a three-time correlation function and the corresponding two-dimensional (2D) lifetime spectrum is developed to elucidate the time-dependent couplings between the multi-timescale (i.e., hierarchical) conformational dynamics in heterogeneous systems such as proteins. In analogy with 2D NMR, IR, electronic, and fluorescence spectroscopies, the waiting-time dependence of the off-diagonal peaks in the 2D lifetime spectra can provide a quantitative description of the dynamical correlations between the conformational motions with different lifetimes. The present method is applied to intrinsic conformational changes of substrate-free adenylate kinase (AKE) using long-time coarse-grained molecular dynamics simulations. It is found that the hierarchical conformational dynamics arise from the intra-domain structural transitions among conformational substates of AKE by analyzing the one-time correlation functions and one-dimensional lifetime spectra for the donor-acceptor distances corresponding to single-molecule Förster resonance energy transfer experiments with the use of the principal component analysis. In addition, the complicated waiting-time dependence of the off-diagonal peaks in the 2D lifetime spectra for the donor-acceptor distances is attributed to the fact that the time evolution of the couplings between the conformational dynamics depends upon both the spatial and temporal characters of the system. The present method is expected to shed light on the biological relationship among the structure, dynamics, and function.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Ono, Junichi; Takada, Shoji; Department of Biophysics, Graduate School of Science, Kyoto University, Kyoto 606-8502
2015-06-07
An analytical method based on a three-time correlation function and the corresponding two-dimensional (2D) lifetime spectrum is developed to elucidate the time-dependent couplings between the multi-timescale (i.e., hierarchical) conformational dynamics in heterogeneous systems such as proteins. In analogy with 2D NMR, IR, electronic, and fluorescence spectroscopies, the waiting-time dependence of the off-diagonal peaks in the 2D lifetime spectra can provide a quantitative description of the dynamical correlations between the conformational motions with different lifetimes. The present method is applied to intrinsic conformational changes of substrate-free adenylate kinase (AKE) using long-time coarse-grained molecular dynamics simulations. It is found that the hierarchicalmore » conformational dynamics arise from the intra-domain structural transitions among conformational substates of AKE by analyzing the one-time correlation functions and one-dimensional lifetime spectra for the donor-acceptor distances corresponding to single-molecule Förster resonance energy transfer experiments with the use of the principal component analysis. In addition, the complicated waiting-time dependence of the off-diagonal peaks in the 2D lifetime spectra for the donor-acceptor distances is attributed to the fact that the time evolution of the couplings between the conformational dynamics depends upon both the spatial and temporal characters of the system. The present method is expected to shed light on the biological relationship among the structure, dynamics, and function.« less
Ibrahim, George M; Morgan, Benjamin R; Macdonald, R Loch
2014-03-01
Predictors of outcome after aneurysmal subarachnoid hemorrhage have been determined previously through hypothesis-driven methods that often exclude putative covariates and require a priori knowledge of potential confounders. Here, we apply a data-driven approach, principal component analysis, to identify baseline patient phenotypes that may predict neurological outcomes. Principal component analysis was performed on 120 subjects enrolled in a prospective randomized trial of clazosentan for the prevention of angiographic vasospasm. Correlation matrices were created using a combination of Pearson, polyserial, and polychoric regressions among 46 variables. Scores of significant components (with eigenvalues>1) were included in multivariate logistic regression models with incidence of severe angiographic vasospasm, delayed ischemic neurological deficit, and long-term outcome as outcomes of interest. Sixteen significant principal components accounting for 74.6% of the variance were identified. A single component dominated by the patients' initial hemodynamic status, World Federation of Neurosurgical Societies score, neurological injury, and initial neutrophil/leukocyte counts was significantly associated with poor outcome. Two additional components were associated with angiographic vasospasm, of which one was also associated with delayed ischemic neurological deficit. The first was dominated by the aneurysm-securing procedure, subarachnoid clot clearance, and intracerebral hemorrhage, whereas the second had high contributions from markers of anemia and albumin levels. Principal component analysis, a data-driven approach, identified patient phenotypes that are associated with worse neurological outcomes. Such data reduction methods may provide a better approximation of unique patient phenotypes and may inform clinical care as well as patient recruitment into clinical trials. http://www.clinicaltrials.gov. Unique identifier: NCT00111085.
The use of multidate multichannel radiance data in urban feature analysis
NASA Technical Reports Server (NTRS)
Duggin, M. J.; Rowntree, R.; Emmons, M.; Hubbard, N.; Odell, A. W.
1986-01-01
Two images were obtained from thematic mappers on Landsats 4 and 5 over the Washington, DC area during November 1982 and March 1984. Selected training areas containing different types of urban land use were examined,one area consisting entirely of forest. Mean digital radiance values for each bandpass in each image were examined, and variances, standard deviations, and covariances between bandpasses were calculated. It has been found that two bandpasses caused forested areas to stand out from other land use types, especially for the November 1982 image. In order to evaluate quantitatively the possible utility of the principal components analysis in selected feature extraction, the eigenvectors were evaluated for principal axes rotations which rendered each selected land use type most separable from all other land use types. The evaluated eigenvectors were plotted as a function of land use type, whose order was decided by considering anticipated shadow component and by examining the relative loadings indicative of vegetation for each of the principal components for the different features considered. The analysis was performed for each seven-band image separately and for the two combined images. It was found that by combining the two images, more dramatic land use type separation could be obtained.
Krohn, M.D.; Milton, N.M.; Segal, D.; Enland, A.
1981-01-01
A principal component image enhancement has been effective in applying Landsat data to geologic mapping in a heavily forested area of E Virginia. The image enhancement procedure consists of a principal component transformation, a histogram normalization, and the inverse principal componnet transformation. The enhancement preserves the independence of the principal components, yet produces a more readily interpretable image than does a single principal component transformation. -from Authors
Techniques for interpretation of geoid anomalies
NASA Technical Reports Server (NTRS)
Chapman, M. E.
1979-01-01
For purposes of geological interpretation, techniques are developed to compute directly the geoid anomaly over models of density within the earth. Ideal bodies such as line segments, vertical sheets, and rectangles are first used to calculate the geoid anomaly. Realistic bodies are modeled with formulas for two-dimensional polygons and three-dimensional polyhedra. By using Fourier transform methods the two-dimensional geoid is seen to be a filtered version of the gravity field, in which the long-wavelength components are magnified and the short-wavelength components diminished.
Finding imaging patterns of structural covariance via Non-Negative Matrix Factorization.
Sotiras, Aristeidis; Resnick, Susan M; Davatzikos, Christos
2015-03-01
In this paper, we investigate the use of Non-Negative Matrix Factorization (NNMF) for the analysis of structural neuroimaging data. The goal is to identify the brain regions that co-vary across individuals in a consistent way, hence potentially being part of underlying brain networks or otherwise influenced by underlying common mechanisms such as genetics and pathologies. NNMF offers a directly data-driven way of extracting relatively localized co-varying structural regions, thereby transcending limitations of Principal Component Analysis (PCA), Independent Component Analysis (ICA) and other related methods that tend to produce dispersed components of positive and negative loadings. In particular, leveraging upon the well known ability of NNMF to produce parts-based representations of image data, we derive decompositions that partition the brain into regions that vary in consistent ways across individuals. Importantly, these decompositions achieve dimensionality reduction via highly interpretable ways and generalize well to new data as shown via split-sample experiments. We empirically validate NNMF in two data sets: i) a Diffusion Tensor (DT) mouse brain development study, and ii) a structural Magnetic Resonance (sMR) study of human brain aging. We demonstrate the ability of NNMF to produce sparse parts-based representations of the data at various resolutions. These representations seem to follow what we know about the underlying functional organization of the brain and also capture some pathological processes. Moreover, we show that these low dimensional representations favorably compare to descriptions obtained with more commonly used matrix factorization methods like PCA and ICA. Copyright © 2014 Elsevier Inc. All rights reserved.
Multispectral histogram normalization contrast enhancement
NASA Technical Reports Server (NTRS)
Soha, J. M.; Schwartz, A. A.
1979-01-01
A multispectral histogram normalization or decorrelation enhancement which achieves effective color composites by removing interband correlation is described. The enhancement procedure employs either linear or nonlinear transformations to equalize principal component variances. An additional rotation to any set of orthogonal coordinates is thus possible, while full histogram utilization is maintained by avoiding the reintroduction of correlation. For the three-dimensional case, the enhancement procedure may be implemented with a lookup table. An application of the enhancement to Landsat multispectral scanning imagery is presented.
A three-dimensional simulation of the equatorial quasi-biennial oscillation
DOE Office of Scientific and Technical Information (OSTI.GOV)
Takahashi, M.; Boville, B.A.
1992-06-15
A simulation of the equatorial quasi-biennial oscillation (QBO) has been obtained using a three-dimensional mechanistic model of the stratosphere. The model is a simplified form of the NCAR CCM (Community Climate Model) in which the troposphere has been replaced with a specified geopotential distribution near the tropical tropopause and most of the physical parameterizations have been removed. A Kelvin wave and a Rossby-gravity wave are forced at the bottom boundary as in previous one- and two-dimensional models. The model reproduces most of the principal features of the observed QBO, as do previous models with lower dimensionality. The principal difference betweenmore » the present model and previous QBO models is that the wave propagation is explicitly represented, allowing wave-wave interactions to take place. It is found that these interactions significantly affect the simulated oscillation. The interaction of the Rossby-gravity waves with the Kelvin waves results in about twice as much easterly compared to westerly forcing being required in order to obtain a QBO. 26 refs., 12 figs.« less
NASA Technical Reports Server (NTRS)
Harp, J. L., Jr.
1977-01-01
A two-dimensional time-dependent computer code was utilized to calculate the three-dimensional steady flow within the impeller blading. The numerical method is an explicit time marching scheme in two spatial dimensions. Initially, an inviscid solution is generated on the hub blade-to-blade surface by the method of Katsanis and McNally (1973). Starting with the known inviscid solution, the viscous effects are calculated through iteration. The approach makes it possible to take into account principal impeller fluid-mechanical effects. It is pointed out that the second iterate provides a complete solution to the three-dimensional, compressible, Navier-Stokes equations for flow in a centrifugal impeller. The problems investigated are related to the study of a radial impeller and a backswept impeller.
NASA Astrophysics Data System (ADS)
Ketcham, Richard A.
2017-04-01
Anisotropy in three-dimensional quantities such as geometric shape and orientation is commonly quantified using principal components analysis, in which a second order tensor determines the orientations of orthogonal components and their relative magnitudes. This approach has many advantages, such as simplicity and ability to accommodate many forms of data, and resilience to data sparsity. However, when data are sufficiently plentiful and precise, they sometimes show that aspects of the principal components approach are oversimplifications that may affect how the data are interpreted or extrapolated for mathematical or physical modeling. High-resolution X-ray computed tomography (CT) can effectively extract thousands of measurements from a single sample, providing a data density sufficient to examine the ways in which anisotropy on the hand-sample scale and smaller can be quantified, and the extent to which the ways the data are simplified are faithful to the underlying distributions. Features within CT data can be considered as discrete objects or continuum fabrics; the latter can be characterized using a variety of metrics, such as the most commonly used mean intercept length, and also the more specialized star length and star volume distributions. Each method posits a different scaling among components that affects the measured degree of anisotropy. The star volume distribution is the most sensitive to anisotropy, and commonly distinguishes strong fabric components that are not orthogonal. Although these data are well-presented using a stereoplot, 3D rose diagrams are another visualization option that can often help identify these components. This talk presents examples from a number of cases, starting with trabecular bone and extending to geological features such as fractures and brittle and ductile fabrics, in which non-orthogonal principal components identified using CT provide some insight into the origin of the underlying structures, and how they should be interpreted and potentially up-scaled.
Principal component regression analysis with SPSS.
Liu, R X; Kuang, J; Gong, Q; Hou, X L
2003-06-01
The paper introduces all indices of multicollinearity diagnoses, the basic principle of principal component regression and determination of 'best' equation method. The paper uses an example to describe how to do principal component regression analysis with SPSS 10.0: including all calculating processes of the principal component regression and all operations of linear regression, factor analysis, descriptives, compute variable and bivariate correlations procedures in SPSS 10.0. The principal component regression analysis can be used to overcome disturbance of the multicollinearity. The simplified, speeded up and accurate statistical effect is reached through the principal component regression analysis with SPSS.
NASA Astrophysics Data System (ADS)
Wang, Zhibiao; Wang, Xu; Pei, Wenxuan; Li, Sen; Sun, Suqin; Zhou, Qun; Chen, Jianbo
2018-03-01
Areca semen is a common herb used in traditional Chinese medicine, but alkaloids in this herb are categorized as Group I carcinogens by IARC. It has been proven that the stir-baking process can reduce alkaloids in Areca semen while keep the activity for promoting digestion. However, the changes of compositions other than alkaloids during the thermal processing are unclear. Understanding the thermal chemical transitions of Areca semen is necessary to explore the processing mechanisms and optimize the procedures. In this research, FTIR spectroscopy with a temperature-controlled ATR accessory is employed to study the heating process of Areca semen. Principal component analysis and two-dimensional correlation spectroscopy are used to interpret the spectra to reveal the chemical transitions of Areca semen in different temperature ranges. The loss of a few volatile compounds in the testa and sperm happens below 105 °C, while some esters in the sperm decreases above 105 °C. As the heating temperature is close to 210 °C, Areca semen begins to be scorched and the decomposition of many compounds can be observed. This research shows the potential of the temperature-resolved ATR-FTIR spectroscopy in exploring the chemical transitions of the thermal processing of herbal materials.
Perrault, Katelynn A; Stefanuto, Pierre-Hugues; Stuart, Barbara H; Rai, Tapan; Focant, Jean-François; Forbes, Shari L
2015-01-01
Challenges in decomposition odour profiling have led to variation in the documented odour profile by different research groups worldwide. Background subtraction and use of controls are important considerations given the variation introduced by decomposition studies conducted in different geographical environments. The collection of volatile organic compounds (VOCs) from soil beneath decomposing remains is challenging due to the high levels of inherent soil VOCs, further confounded by the use of highly sensitive instrumentation. This study presents a method that provides suitable chromatographic resolution for profiling decomposition odour in soil by comprehensive two-dimensional gas chromatography coupled with time-of-flight mass spectrometry using appropriate controls and field blanks. Logarithmic transformation and t-testing of compounds permitted the generation of a compound list of decomposition VOCs in soil. Principal component analysis demonstrated the improved discrimination between experimental and control soil, verifying the value of the data handling method. Data handling procedures have not been well documented in this field and standardisation would thereby reduce misidentification of VOCs present in the surrounding environment as decomposition byproducts. Uniformity of data handling and instrumental procedures will reduce analytical variation, increasing confidence in the future when investigating the effect of taphonomic variables on the decomposition VOC profile. © 2014 WILEY-VCH Verlag GmbH & Co. KGaA, Weinheim.
Cocco, S; Monasson, R; Sessak, V
2011-05-01
We consider the problem of inferring the interactions between a set of N binary variables from the knowledge of their frequencies and pairwise correlations. The inference framework is based on the Hopfield model, a special case of the Ising model where the interaction matrix is defined through a set of patterns in the variable space, and is of rank much smaller than N. We show that maximum likelihood inference is deeply related to principal component analysis when the amplitude of the pattern components ξ is negligible compared to √N. Using techniques from statistical mechanics, we calculate the corrections to the patterns to the first order in ξ/√N. We stress the need to generalize the Hopfield model and include both attractive and repulsive patterns in order to correctly infer networks with sparse and strong interactions. We present a simple geometrical criterion to decide how many attractive and repulsive patterns should be considered as a function of the sampling noise. We moreover discuss how many sampled configurations are required for a good inference, as a function of the system size N and of the amplitude ξ. The inference approach is illustrated on synthetic and biological data.
Derivation of Boundary Manikins: A Principal Component Analysis
NASA Technical Reports Server (NTRS)
Young, Karen; Margerum, Sarah; Barr, Abbe; Ferrer, Mike A.; Rajulu, Sudhakar
2008-01-01
When designing any human-system interface, it is critical to provide realistic anthropometry to properly represent how a person fits within a given space. This study aimed to identify a minimum number of boundary manikins or representative models of subjects anthropometry from a target population, which would realistically represent the population. The boundary manikin anthropometry was derived using, Principal Component Analysis (PCA). PCA is a statistical approach to reduce a multi-dimensional dataset using eigenvectors and eigenvalues. The measurements used in the PCA were identified as those measurements critical for suit and cockpit design. The PCA yielded a total of 26 manikins per gender, as well as their anthropometry from the target population. Reduction techniques were implemented to reduce this number further with a final result of 20 female and 22 male subjects. The anthropometry of the boundary manikins was then be used to create 3D digital models (to be discussed in subsequent papers) intended for use by designers to test components of their space suit design, to verify that the requirements specified in the Human Systems Integration Requirements (HSIR) document are met. The end-goal is to allow for designers to generate suits which accommodate the diverse anthropometry of the user population.
NASA Technical Reports Server (NTRS)
Hyer, M. W.
1980-01-01
The determination of the stress distribution in the inner lap of double-lap, double-bolt joints using photoelastic models of the joint is discussed. The principal idea is to fabricate the inner lap of a photoelastic material and to use a photoelastically sensitive material for the two outer laps. With this setup, polarized light transmitted through the stressed model responds principally to the stressed inner lap. The model geometry, the procedures for making and testing the model, and test results are described.
Phenotypic and genetic structure of traits delineating personality disorder.
Livesley, W J; Jang, K L; Vernon, P A
1998-10-01
The evidence suggests that personality traits are hierarchically organized with more specific or lower-order traits combining to form more generalized higher-order traits. Agreement exists across studies regarding the lower-order traits that delineate personality disorder but not the higher-order traits. This study seeks to identify the higher-order structure of personality disorder by examining the phenotypic and genetic structures underlying lower-order traits. Eighteen lower-order traits were assessed using the Dimensional Assessment of Personality Disorder-Basic Questionnaire in samples of 656 personality disordered patients, 939 general population subjects, and a volunteer sample of 686 twin pairs. Principal components analysis yielded 4 components, labeled Emotional Dysregulation, Dissocial Behavior, Inhibitedness, and Compulsivity, that were similar across the 3 samples. Multivariate genetic analyses also yielded 4 genetic and environmental factors that were remarkably similar to the phenotypic factors. Analysis of the residual heritability of the lower-order traits when the effects of the higher-order factors were removed revealed a substantial residual heritable component for 12 of the 18 traits. The results support the following conclusions. First, the stable structure of traits across clinical and nonclinical samples is consistent with dimensional representations of personality disorders. Second, the higher-order traits of personality disorder strongly resemble dimensions of normal personality. This implies that a dimensional classification should be compatible with normative personality. Third, the residual heritability of the lower-order traits suggests that the personality phenotypes are based on a large number of specific genetic components.
Principal Component 2-D Long Short-Term Memory for Font Recognition on Single Chinese Characters.
Tao, Dapeng; Lin, Xu; Jin, Lianwen; Li, Xuelong
2016-03-01
Chinese character font recognition (CCFR) has received increasing attention as the intelligent applications based on optical character recognition becomes popular. However, traditional CCFR systems do not handle noisy data effectively. By analyzing in detail the basic strokes of Chinese characters, we propose that font recognition on a single Chinese character is a sequence classification problem, which can be effectively solved by recurrent neural networks. For robust CCFR, we integrate a principal component convolution layer with the 2-D long short-term memory (2DLSTM) and develop principal component 2DLSTM (PC-2DLSTM) algorithm. PC-2DLSTM considers two aspects: 1) the principal component layer convolution operation helps remove the noise and get a rational and complete font information and 2) simultaneously, 2DLSTM deals with the long-range contextual processing along scan directions that can contribute to capture the contrast between character trajectory and background. Experiments using the frequently used CCFR dataset suggest the effectiveness of PC-2DLSTM compared with other state-of-the-art font recognition methods.
Further two-dimensional code development for Stirling space engine components
NASA Technical Reports Server (NTRS)
Ibrahim, Mounir; Tew, Roy C.; Dudenhoefer, James E.
1990-01-01
The development of multidimensional models of Stirling engine components is described. Two-dimensional parallel plate models of an engine regenerator and a cooler were used to study heat transfer under conditions of laminar, incompressible oscillating flow. Substantial differences in the nature of the temperature variations in time over the cycle were observed for the cooler as contrasted with the regenerator. When the two-dimensional cooler model was used to calculate a heat transfer coefficient, it yields a very different result from that calculated using steady-flow correlations. Simulation results for the regenerator and the cooler are presented.
Clement, Cristina C.; Aphkhazava, David; Nieves, Edward; Callaway, Myrasol; Olszewski, Waldemar; Rotzschke, Olaf; Santambrogio, Laura
2013-01-01
In this study a proteomic approach was used to define the protein content of matched samples of afferent prenodal lymph and plasma derived from healthy volunteers. The analysis was performed using two analytical methodologies coupled with nanoliquid chromatography-tandem mass spectrometry: one-dimensional gel electrophoresis (1DEF nanoLC Orbitrap–ESI–MS/MS), and two-dimensional fluorescence difference-in-gel electrophoresis (2D-DIGE nanoLC–ESI–MS/MS). The 253 significantly identified proteins (p<0.05), obtained from the tandem mass spectrometry data, were further analyzed with pathway analysis (IPA) to define the functional signature of prenodal lymph and matched plasma. The 1DEF coupled with nanoLC–MS–MS revealed that the common proteome between the two biological fluids (144 out of 253 proteins) was dominated by complement activation and blood coagulation components, transporters and protease inhibitors. The enriched proteome of human lymph (72 proteins) consisted of products derived from the extracellular matrix, apoptosis and cellular catabolism. In contrast, the enriched proteome of human plasma (37 proteins) consisted of soluble molecules of the coagulation system and cell–cell signaling factors. The functional networks associated with both common and source-distinctive proteomes highlight the principal biological activity of these immunologically relevant body fluids. PMID:23202415
Results of an Oncology Clinical Trial Nurse Role Delineation Study.
Purdom, Michelle A; Petersen, Sandra; Haas, Barbara K
2017-09-01
To evaluate the relevance of a five-dimensional model of clinical trial nursing practice in an oncology clinical trial nurse population. . Web-based cross-sectional survey. . Online via Qualtrics. . 167 oncology nurses throughout the United States, including 41 study coordinators, 35 direct care providers, and 91 dual-role nurses who provide direct patient care and trial coordination. . Principal components analysis was used to determine the dimensions of oncology clinical trial nursing practice. . Self-reported frequency of 59 activities. . The results did not support the original five-dimensional model of nursing care but revealed a more multidimensional model. . An analysis of frequency data revealed an eight-dimensional model of oncology research nursing, including care, manage study, expert, lead, prepare, data, advance science, and ethics. . This evidence-based model expands understanding of the multidimensional roles of oncology nurses caring for patients with cancer enrolled in clinical trials.
Sheppard, P S; Stevenson, J M; Graham, R B
2016-05-01
The objective of the present study was to determine if there is a sex-based difference in lifting technique across increasing-load conditions. Eleven male and 14 female participants (n = 25) with no previous history of low back disorder participated in the study. Participants completed freestyle, symmetric lifts of a box with handles from the floor to a table positioned at 50% of their height for five trials under three load conditions (10%, 20%, and 30% of their individual maximum isometric back strength). Joint kinematic data for the ankle, knee, hip, and lumbar and thoracic spine were collected using a two-camera Optotrak motion capture system. Joint angles were calculated using a three-dimensional Euler rotation sequence. Principal component analysis (PCA) and single component reconstruction were applied to assess differences in lifting technique across the entire waveforms. Thirty-two PCs were retained from the five joints and three axes in accordance with the 90% trace criterion. Repeated-measures ANOVA with a mixed design revealed no significant effect of sex for any of the PCs. This is contrary to previous research that used discrete points on the lifting curve to analyze sex-based differences, but agrees with more recent research using more complex analysis techniques. There was a significant effect of load on lifting technique for five PCs of the lower limb (PC1 of ankle flexion, knee flexion, and knee adduction, as well as PC2 and PC3 of hip flexion) (p < 0.005). However, there was no significant effect of load on the thoracic and lumbar spine. It was concluded that when load is standardized to individual back strength characteristics, males and females adopted a similar lifting technique. In addition, as load increased male and female participants changed their lifting technique in a similar manner. Copyright © 2016. Published by Elsevier Ltd.
Psychophysical evaluation of three-dimensional auditory displays
NASA Technical Reports Server (NTRS)
Wightman, Frederic L.
1991-01-01
Work during this reporting period included the completion of our research on the use of principal components analysis (PCA) to model the acoustical head related transfer functions (HRTFs) that are used to synthesize virtual sources for three dimensional auditory displays. In addition, a series of studies was initiated on the perceptual errors made by listeners when localizing free-field and virtual sources. Previous research has revealed that under certain conditions these perceptual errors, often called 'confusions' or 'reversals', are both large and frequent, thus seriously comprising the utility of a 3-D virtual auditory display. The long-range goal of our work in this area is to elucidate the sources of the confusions and to develop signal-processing strategies to reduce or eliminate them.
Sun, Xiaochun; Ma, Ping; Mumm, Rita H
2012-01-01
Genomic selection (GS) procedures have proven useful in estimating breeding value and predicting phenotype with genome-wide molecular marker information. However, issues of high dimensionality, multicollinearity, and the inability to deal effectively with epistasis can jeopardize accuracy and predictive ability. We, therefore, propose a new nonparametric method, pRKHS, which combines the features of supervised principal component analysis (SPCA) and reproducing kernel Hilbert spaces (RKHS) regression, with versions for traits with no/low epistasis, pRKHS-NE, to high epistasis, pRKHS-E. Instead of assigning a specific relationship to represent the underlying epistasis, the method maps genotype to phenotype in a nonparametric way, thus requiring fewer genetic assumptions. SPCA decreases the number of markers needed for prediction by filtering out low-signal markers with the optimal marker set determined by cross-validation. Principal components are computed from reduced marker matrix (called supervised principal components, SPC) and included in the smoothing spline ANOVA model as independent variables to fit the data. The new method was evaluated in comparison with current popular methods for practicing GS, specifically RR-BLUP, BayesA, BayesB, as well as a newer method by Crossa et al., RKHS-M, using both simulated and real data. Results demonstrate that pRKHS generally delivers greater predictive ability, particularly when epistasis impacts trait expression. Beyond prediction, the new method also facilitates inferences about the extent to which epistasis influences trait expression.
Sun, Xiaochun; Ma, Ping; Mumm, Rita H.
2012-01-01
Genomic selection (GS) procedures have proven useful in estimating breeding value and predicting phenotype with genome-wide molecular marker information. However, issues of high dimensionality, multicollinearity, and the inability to deal effectively with epistasis can jeopardize accuracy and predictive ability. We, therefore, propose a new nonparametric method, pRKHS, which combines the features of supervised principal component analysis (SPCA) and reproducing kernel Hilbert spaces (RKHS) regression, with versions for traits with no/low epistasis, pRKHS-NE, to high epistasis, pRKHS-E. Instead of assigning a specific relationship to represent the underlying epistasis, the method maps genotype to phenotype in a nonparametric way, thus requiring fewer genetic assumptions. SPCA decreases the number of markers needed for prediction by filtering out low-signal markers with the optimal marker set determined by cross-validation. Principal components are computed from reduced marker matrix (called supervised principal components, SPC) and included in the smoothing spline ANOVA model as independent variables to fit the data. The new method was evaluated in comparison with current popular methods for practicing GS, specifically RR-BLUP, BayesA, BayesB, as well as a newer method by Crossa et al., RKHS-M, using both simulated and real data. Results demonstrate that pRKHS generally delivers greater predictive ability, particularly when epistasis impacts trait expression. Beyond prediction, the new method also facilitates inferences about the extent to which epistasis influences trait expression. PMID:23226325
1990-08-01
the spectral domain is extended to include the effects of two-dimensional, two-component current flow in planar transmission line discontinuities 6n...PROFESSOR: Tatsuo Itoh A deterministic formulation of the method of moments carried out in the spectral domain is extended to include the effects of...two-dimensional, two- component current flow in planar transmission line discontinuities on open substrates. The method includes the effects of space
NASA Astrophysics Data System (ADS)
Bradde, Serena; Bialek, William
A system with many degrees of freedom can be characterized by a covariance matrix; principal components analysis (PCA) focuses on the eigenvalues of this matrix, hoping to find a lower dimensional description. But when the spectrum is nearly continuous, any distinction between components that we keep and those that we ignore becomes arbitrary; it then is natural to ask what happens as we vary this arbitrary cutoff. We argue that this problem is analogous to the momentum shell renormalization group (RG). Following this analogy, we can define relevant and irrelevant operators, where the role of dimensionality is played by properties of the eigenvalue density. These results also suggest an approach to the analysis of real data. As an example, we study neural activity in the vertebrate retina as it responds to naturalistic movies, and find evidence of behavior controlled by a nontrivial fixed point. Applied to financial data, our analysis separates modes dominated by sampling noise from a smaller but still macroscopic number of modes described by a non-Gaussian distribution.
Machine learning and data science in soft materials engineering
NASA Astrophysics Data System (ADS)
Ferguson, Andrew L.
2018-01-01
In many branches of materials science it is now routine to generate data sets of such large size and dimensionality that conventional methods of analysis fail. Paradigms and tools from data science and machine learning can provide scalable approaches to identify and extract trends and patterns within voluminous data sets, perform guided traversals of high-dimensional phase spaces, and furnish data-driven strategies for inverse materials design. This topical review provides an accessible introduction to machine learning tools in the context of soft and biological materials by ‘de-jargonizing’ data science terminology, presenting a taxonomy of machine learning techniques, and surveying the mathematical underpinnings and software implementations of popular tools, including principal component analysis, independent component analysis, diffusion maps, support vector machines, and relative entropy. We present illustrative examples of machine learning applications in soft matter, including inverse design of self-assembling materials, nonlinear learning of protein folding landscapes, high-throughput antimicrobial peptide design, and data-driven materials design engines. We close with an outlook on the challenges and opportunities for the field.
Machine learning and data science in soft materials engineering.
Ferguson, Andrew L
2018-01-31
In many branches of materials science it is now routine to generate data sets of such large size and dimensionality that conventional methods of analysis fail. Paradigms and tools from data science and machine learning can provide scalable approaches to identify and extract trends and patterns within voluminous data sets, perform guided traversals of high-dimensional phase spaces, and furnish data-driven strategies for inverse materials design. This topical review provides an accessible introduction to machine learning tools in the context of soft and biological materials by 'de-jargonizing' data science terminology, presenting a taxonomy of machine learning techniques, and surveying the mathematical underpinnings and software implementations of popular tools, including principal component analysis, independent component analysis, diffusion maps, support vector machines, and relative entropy. We present illustrative examples of machine learning applications in soft matter, including inverse design of self-assembling materials, nonlinear learning of protein folding landscapes, high-throughput antimicrobial peptide design, and data-driven materials design engines. We close with an outlook on the challenges and opportunities for the field.
NASA Astrophysics Data System (ADS)
Bradde, Serena; Bialek, William
2017-05-01
A system with many degrees of freedom can be characterized by a covariance matrix; principal components analysis focuses on the eigenvalues of this matrix, hoping to find a lower dimensional description. But when the spectrum is nearly continuous, any distinction between components that we keep and those that we ignore becomes arbitrary; it then is natural to ask what happens as we vary this arbitrary cutoff. We argue that this problem is analogous to the momentum shell renormalization group. Following this analogy, we can define relevant and irrelevant operators, where the role of dimensionality is played by properties of the eigenvalue density. These results also suggest an approach to the analysis of real data. As an example, we study neural activity in the vertebrate retina as it responds to naturalistic movies, and find evidence of behavior controlled by a nontrivial fixed point. Applied to financial data, our analysis separates modes dominated by sampling noise from a smaller but still macroscopic number of modes described by a non-Gaussian distribution.
Analysis of 3-D Tongue Motion From Tagged and Cine Magnetic Resonance Images
Woo, Jonghye; Lee, Junghoon; Murano, Emi Z.; Stone, Maureen; Prince, Jerry L.
2016-01-01
Purpose Measuring tongue deformation and internal muscle motion during speech has been a challenging task because the tongue deforms in 3 dimensions, contains interdigitated muscles, and is largely hidden within the vocal tract. In this article, a new method is proposed to analyze tagged and cine magnetic resonance images of the tongue during speech in order to estimate 3-dimensional tissue displacement and deformation over time. Method The method involves computing 2-dimensional motion components using a standard tag-processing method called harmonic phase, constructing superresolution tongue volumes using cine magnetic resonance images, segmenting the tongue region using a random-walker algorithm, and estimating 3-dimensional tongue motion using an incompressible deformation estimation algorithm. Results Evaluation of the method is presented with a control group and a group of people who had received a glossectomy carrying out a speech task. A 2-step principal-components analysis is then used to reveal the unique motion patterns of the subjects. Azimuth motion angles and motion on the mirrored hemi-tongues are analyzed. Conclusion Tests of the method with a various collection of subjects show its capability of capturing patient motion patterns and indicate its potential value in future speech studies. PMID:27295428
Multivariate image analysis of laser-induced photothermal imaging used for detection of caries tooth
NASA Astrophysics Data System (ADS)
El-Sherif, Ashraf F.; Abdel Aziz, Wessam M.; El-Sharkawy, Yasser H.
2010-08-01
Time-resolved photothermal imaging has been investigated to characterize tooth for the purpose of discriminating between normal and caries areas of the hard tissue using thermal camera. Ultrasonic thermoelastic waves were generated in hard tissue by the absorption of fiber-coupled Q-switched Nd:YAG laser pulses operating at 1064 nm in conjunction with a laser-induced photothermal technique used to detect the thermal radiation waves for diagnosis of human tooth. The concepts behind the use of photo-thermal techniques for off-line detection of caries tooth features were presented by our group in earlier work. This paper illustrates the application of multivariate image analysis (MIA) techniques to detect the presence of caries tooth. MIA is used to rapidly detect the presence and quantity of common caries tooth features as they scanned by the high resolution color (RGB) thermal cameras. Multivariate principal component analysis is used to decompose the acquired three-channel tooth images into a two dimensional principal components (PC) space. Masking score point clusters in the score space and highlighting corresponding pixels in the image space of the two dominant PCs enables isolation of caries defect pixels based on contrast and color information. The technique provides a qualitative result that can be used for early stage caries tooth detection. The proposed technique can potentially be used on-line or real-time resolved to prescreen the existence of caries through vision based systems like real-time thermal camera. Experimental results on the large number of extracted teeth as well as one of the thermal image panoramas of the human teeth voltanteer are investigated and presented.
Time series analysis of collective motions in proteins
NASA Astrophysics Data System (ADS)
Alakent, Burak; Doruker, Pemra; ćamurdan, Mehmet C.
2004-01-01
The dynamics of α-amylase inhibitor tendamistat around its native state is investigated using time series analysis of the principal components of the Cα atomic displacements obtained from molecular dynamics trajectories. Collective motion along a principal component is modeled as a homogeneous nonstationary process, which is the result of the damped oscillations in local minima superimposed on a random walk. The motion in local minima is described by a stationary autoregressive moving average model, consisting of the frequency, damping factor, moving average parameters and random shock terms. Frequencies for the first 50 principal components are found to be in the 3-25 cm-1 range, which are well correlated with the principal component indices and also with atomistic normal mode analysis results. Damping factors, though their correlation is less pronounced, decrease as principal component indices increase, indicating that low frequency motions are less affected by friction. The existence of a positive moving average parameter indicates that the stochastic force term is likely to disturb the mode in opposite directions for two successive sampling times, showing the modes tendency to stay close to minimum. All these four parameters affect the mean square fluctuations of a principal mode within a single minimum. The inter-minima transitions are described by a random walk model, which is driven by a random shock term considerably smaller than that for the intra-minimum motion. The principal modes are classified into three subspaces based on their dynamics: essential, semiconstrained, and constrained, at least in partial consistency with previous studies. The Gaussian-type distributions of the intermediate modes, called "semiconstrained" modes, are explained by asserting that this random walk behavior is not completely free but between energy barriers.
Development of an integrated BEM for hot fluid-structure interaction
NASA Technical Reports Server (NTRS)
Banerjee, P. K.; Dargush, G. F.
1989-01-01
The Boundary Element Method (BEM) is chosen as a basic analysis tool principally because the definition of quantities like fluxes, temperature, displacements, and velocities is very precise on a boundary base discretization scheme. One fundamental difficulty is, of course, that the entire analysis requires a very considerable amount of analytical work which is not present in other numerical methods. During the last 18 months all of this analytical work was completed and a two-dimensional, general purpose code was written. Some of the early results are described. It is anticipated that within the next two to three months almost all two-dimensional idealizations will be examined. It should be noted that the analytical work for the three-dimensional case has also been done and numerical implementation will begin next year.
Kesharaju, Manasa; Nagarajah, Romesh
2015-09-01
The motivation for this research stems from a need for providing a non-destructive testing method capable of detecting and locating any defects and microstructural variations within armour ceramic components before issuing them to the soldiers who rely on them for their survival. The development of an automated ultrasonic inspection based classification system would make possible the checking of each ceramic component and immediately alert the operator about the presence of defects. Generally, in many classification problems a choice of features or dimensionality reduction is significant and simultaneously very difficult, as a substantial computational effort is required to evaluate possible feature subsets. In this research, a combination of artificial neural networks and genetic algorithms are used to optimize the feature subset used in classification of various defects in reaction-sintered silicon carbide ceramic components. Initially wavelet based feature extraction is implemented from the region of interest. An Artificial Neural Network classifier is employed to evaluate the performance of these features. Genetic Algorithm based feature selection is performed. Principal Component Analysis is a popular technique used for feature selection and is compared with the genetic algorithm based technique in terms of classification accuracy and selection of optimal number of features. The experimental results confirm that features identified by Principal Component Analysis lead to improved performance in terms of classification percentage with 96% than Genetic algorithm with 94%. Copyright © 2015 Elsevier B.V. All rights reserved.
Putilov, Arcady A; Donskaya, Olga G
2016-01-01
Age-associated changes in different bandwidths of the human electroencephalographic (EEG) spectrum are well documented, but their functional significance is poorly understood. This spectrum seems to represent summation of simultaneous influences of several sleep-wake regulatory processes. Scoring of its orthogonal (uncorrelated) principal components can help in separation of the brain signatures of these processes. In particular, the opposite age-associated changes were documented for scores on the two largest (1st and 2nd) principal components of the sleep EEG spectrum. A decrease of the first score and an increase of the second score can reflect, respectively, the weakening of the sleep drive and disinhibition of the opposing wake drive with age. In order to support the suggestion of age-associated disinhibition of the wake drive from the antagonistic influence of the sleep drive, we analyzed principal component scores of the resting EEG spectra obtained in sleep deprivation experiments with 81 healthy young adults aged between 19 and 26 and 40 healthy older adults aged between 45 and 66 years. At the second day of the sleep deprivation experiments, frontal scores on the 1st principal component of the EEG spectrum demonstrated an age-associated reduction of response to eyes closed relaxation. Scores on the 2nd principal component were either initially increased during wakefulness or less responsive to such sleep-provoking conditions (frontal and occipital scores, respectively). These results are in line with the suggestion of disinhibition of the wake drive with age. They provide an explanation of why older adults are less vulnerable to sleep deprivation than young adults.
Chung, Ill-Min; Kim, Jae-Kwang; Jin, Yong-Ik; Oh, Yong-Taek; Prabakaran, Mayakrishnan; Youn, Kyoung-Jin; Kim, Seung-Hyun
2016-12-01
Compared to other foods, the use of common bio-elements to identify the geographical origin of potato remains limited. Thus, this study aimed to verify whether the cultivation regions of raw potato tubers could be determined by the stable isotope composition analysis of bio-elements. δ(13)CVPDB and δ(15)NAIR in potato were influenced by region and cultivar, whereas δ(18)OVSMOW and δ(34)SVCDT were only influenced by region (p<0.0001). A two-dimensional plot of δ(18)OVSMOW and δ(34)SVCDT effectively distinguished between high and low altitude regions, and also reliably discriminated Wanju, Haenam, and Boseong cultivars in low altitude regions. δ(34)SVCDT was the main component that was responsible for the separation of samples in the principal component analysis (eigenvector of -0.6209) and orthogonal projection to latent structure-discriminant analysis (VIP value of 1.0566). In conclusion, this study improves our understanding of how the isotope composition of potato tubers varies with respect to cultivation regions and cultivars. Copyright © 2016 Elsevier Ltd. All rights reserved.
A Direct Approach to In-Plane Stress Separation using Photoelastic Ptychography
NASA Astrophysics Data System (ADS)
Anthony, Nicholas; Cadenazzi, Guido; Kirkwood, Henry; Huwald, Eric; Nugent, Keith; Abbey, Brian
2016-08-01
The elastic properties of materials, either under external load or in a relaxed state, influence their mechanical behaviour. Conventional optical approaches based on techniques such as photoelasticity or thermoelasticity can be used for full-field analysis of the stress distribution within a specimen. The circular polariscope in combination with holographic photoelasticity allows the sum and difference of principal stress components to be determined by exploiting the temporary birefringent properties of materials under load. Phase stepping and interferometric techniques have been proposed as a method for separating the in-plane stress components in two-dimensional photoelasticity experiments. In this paper we describe and demonstrate an alternative approach based on photoelastic ptychography which is able to obtain quantitative stress information from far fewer measurements than is required for interferometric based approaches. The complex light intensity equations based on Jones calculus for this setup are derived. We then apply this approach to the problem of a disc under diametrical compression. The experimental results are validated against the analytical solution derived by Hertz for the theoretical displacement fields for an elastic disc subject to point loading.
A Direct Approach to In-Plane Stress Separation using Photoelastic Ptychography
Anthony, Nicholas; Cadenazzi, Guido; Kirkwood, Henry; Huwald, Eric; Nugent, Keith; Abbey, Brian
2016-01-01
The elastic properties of materials, either under external load or in a relaxed state, influence their mechanical behaviour. Conventional optical approaches based on techniques such as photoelasticity or thermoelasticity can be used for full-field analysis of the stress distribution within a specimen. The circular polariscope in combination with holographic photoelasticity allows the sum and difference of principal stress components to be determined by exploiting the temporary birefringent properties of materials under load. Phase stepping and interferometric techniques have been proposed as a method for separating the in-plane stress components in two-dimensional photoelasticity experiments. In this paper we describe and demonstrate an alternative approach based on photoelastic ptychography which is able to obtain quantitative stress information from far fewer measurements than is required for interferometric based approaches. The complex light intensity equations based on Jones calculus for this setup are derived. We then apply this approach to the problem of a disc under diametrical compression. The experimental results are validated against the analytical solution derived by Hertz for the theoretical displacement fields for an elastic disc subject to point loading. PMID:27488605
Gellért, Akos; Balázs, Ervin
2010-02-26
The three-dimensional structures of two cucumovirus coat proteins (CP), namely Cucumber mosaic virus (CMV) and Tomato aspermy virus (TAV), were explored by molecular dynamics (MD) simulations. The N-terminal domain and the C-terminal tail of the CPs proved to be intrinsically unstructured protein regions in aqueous solution. The N-terminal alpha-helix had a partially unrolled conformation. The thermal factor analysis of the CP loop regions demonstrated that the CMV CP had more flexible loop regions than the TAV CP. The principal component analysis (PCA) of the MD trajectories showed that the first three eigenvectors represented the three main conformational motions in the CPs. The first motion components with the highest variance contribution described an opening movement between the hinge and the N-terminal domain of both CPs. The second eigenvector showed a closing motion, while the third eigenvector represented crosswise conformational fluctuations. These new findings, together with previous results, suggest that the hinge region of CPs plays a central role in the recognition and binding of viral RNA. Copyright 2009 Elsevier Inc. All rights reserved.
Vergara, Fredd; Shino, Amiu; Kikuchi, Jun
2016-09-02
Cannibalism is known in many insect species, yet its impact on insect metabolism has not been investigated in detail. This study assessed the effects of cannibalism on the metabolism of fourth-instar larvae of the non-predatory insect Helicoverpa armigera (Lepidotera: Noctuidea). Two groups of larvae were analyzed: one group fed with fourth-instar larvae of H. armigera (cannibal), the other group fed with an artificial plant diet. Water-soluble small organic compounds present in the larvae were analyzed using two-dimensional nuclear magnetic resonance (NMR) and principal component analysis (PCA). Cannibalism negatively affected larval growth. PCA of NMR spectra showed that the metabolic profiles of cannibal and herbivore larvae were statistically different with monomeric sugars, fatty acid- and amino acid-related metabolites as the most variable compounds. Quantitation of ¹H-(13)C HSQC (Heteronuclear Single Quantum Coherence) signals revealed that the concentrations of glucose, glucono-1,5-lactone, glycerol phosphate, glutamine, glycine, leucine, isoleucine, lysine, ornithine, proline, threonine and valine were higher in the herbivore larvae.
THE CORRELATION BIAS FOR TWO-DIMENSIONAL SIMULATIONS BY TURNING BANDS. (R825173)
The perspectives, information and conclusions conveyed in research project abstracts, progress reports, final reports, journal abstracts and journal publications convey the viewpoints of the principal investigator and may not represent the views and policies of ORD and EPA. Concl...
Shading of a computer-generated hologram by zone plate modulation.
Kurihara, Takayuki; Takaki, Yasuhiro
2012-02-13
We propose a hologram calculation technique that enables reconstructing a shaded three-dimensional (3D) image. The amplitude distributions of zone plates, which generate the object points that constitute a 3D object, were two-dimensionally modulated. Two-dimensional (2D) amplitude modulation was determined on the basis of the Phong reflection model developed for computer graphics, which considers the specular, diffuse, and ambient reflection light components. The 2D amplitude modulation added variable and constant modulations: the former controlled the specular light component and the latter controlled the diffuse and ambient components. The proposed calculation technique was experimentally verified. The reconstructed image showed specular reflection that varied depending on the viewing position.
Structured Sparse Principal Components Analysis With the TV-Elastic Net Penalty.
de Pierrefeu, Amicie; Lofstedt, Tommy; Hadj-Selem, Fouad; Dubois, Mathieu; Jardri, Renaud; Fovet, Thomas; Ciuciu, Philippe; Frouin, Vincent; Duchesnay, Edouard
2018-02-01
Principal component analysis (PCA) is an exploratory tool widely used in data analysis to uncover the dominant patterns of variability within a population. Despite its ability to represent a data set in a low-dimensional space, PCA's interpretability remains limited. Indeed, the components produced by PCA are often noisy or exhibit no visually meaningful patterns. Furthermore, the fact that the components are usually non-sparse may also impede interpretation, unless arbitrary thresholding is applied. However, in neuroimaging, it is essential to uncover clinically interpretable phenotypic markers that would account for the main variability in the brain images of a population. Recently, some alternatives to the standard PCA approach, such as sparse PCA (SPCA), have been proposed, their aim being to limit the density of the components. Nonetheless, sparsity alone does not entirely solve the interpretability problem in neuroimaging, since it may yield scattered and unstable components. We hypothesized that the incorporation of prior information regarding the structure of the data may lead to improved relevance and interpretability of brain patterns. We therefore present a simple extension of the popular PCA framework that adds structured sparsity penalties on the loading vectors in order to identify the few stable regions in the brain images that capture most of the variability. Such structured sparsity can be obtained by combining, e.g., and total variation (TV) penalties, where the TV regularization encodes information on the underlying structure of the data. This paper presents the structured SPCA (denoted SPCA-TV) optimization framework and its resolution. We demonstrate SPCA-TV's effectiveness and versatility on three different data sets. It can be applied to any kind of structured data, such as, e.g., -dimensional array images or meshes of cortical surfaces. The gains of SPCA-TV over unstructured approaches (such as SPCA and ElasticNet PCA) or structured approach (such as GraphNet PCA) are significant, since SPCA-TV reveals the variability within a data set in the form of intelligible brain patterns that are easier to interpret and more stable across different samples.
Peleato, Nicolas M; Legge, Raymond L; Andrews, Robert C
2018-06-01
The use of fluorescence data coupled with neural networks for improved predictability of drinking water disinfection by-products (DBPs) was investigated. Novel application of autoencoders to process high-dimensional fluorescence data was related to common dimensionality reduction techniques of parallel factors analysis (PARAFAC) and principal component analysis (PCA). The proposed method was assessed based on component interpretability as well as for prediction of organic matter reactivity to formation of DBPs. Optimal prediction accuracies on a validation dataset were observed with an autoencoder-neural network approach or by utilizing the full spectrum without pre-processing. Latent representation by an autoencoder appeared to mitigate overfitting when compared to other methods. Although DBP prediction error was minimized by other pre-processing techniques, PARAFAC yielded interpretable components which resemble fluorescence expected from individual organic fluorophores. Through analysis of the network weights, fluorescence regions associated with DBP formation can be identified, representing a potential method to distinguish reactivity between fluorophore groupings. However, distinct results due to the applied dimensionality reduction approaches were observed, dictating a need for considering the role of data pre-processing in the interpretability of the results. In comparison to common organic measures currently used for DBP formation prediction, fluorescence was shown to improve prediction accuracies, with improvements to DBP prediction best realized when appropriate pre-processing and regression techniques were applied. The results of this study show promise for the potential application of neural networks to best utilize fluorescence EEM data for prediction of organic matter reactivity. Copyright © 2018 Elsevier Ltd. All rights reserved.
Thermodynamics of Yukawa fluids near the one-component-plasma limit
DOE Office of Scientific and Technical Information (OSTI.GOV)
Khrapak, Sergey A.; Aix-Marseille-Université, CNRS, Laboratoire PIIM, UMR 7345, 13397 Marseille Cedex 20; Semenov, Igor L.
Thermodynamics of weakly screened (near the one-component-plasma limit) Yukawa fluids in two and three dimensions is analyzed in detail. It is shown that the thermal component of the excess internal energy of these fluids, when expressed in terms of the properly normalized coupling strength, exhibits the scaling pertinent to the corresponding one-component-plasma limit (the scalings differ considerably between the two- and three-dimensional situations). This provides us with a simple and accurate practical tool to estimate thermodynamic properties of weakly screened Yukawa fluids. Particular attention is paid to the two-dimensional fluids, for which several important thermodynamic quantities are calculated to illustratemore » the application of the approach.« less
Dynamic Fracture Simulations of Explosively Loaded Cylinders
DOE Office of Scientific and Technical Information (OSTI.GOV)
Arthur, Carly W.; Goto, D. M.
2015-11-30
This report documents the modeling results of high explosive experiments investigating dynamic fracture of steel (AerMet® 100 alloy) cylinders. The experiments were conducted at Lawrence Livermore National Laboratory (LLNL) during 2007 to 2008 [10]. A principal objective of this study was to gain an understanding of dynamic material failure through the analysis of hydrodynamic computer code simulations. Two-dimensional and three-dimensional computational cylinder models were analyzed using the ALE3D multi-physics computer code.
Bayesian wavelet PCA methodology for turbomachinery damage diagnosis under uncertainty
NASA Astrophysics Data System (ADS)
Xu, Shengli; Jiang, Xiaomo; Huang, Jinzhi; Yang, Shuhua; Wang, Xiaofang
2016-12-01
Centrifugal compressor often suffers various defects such as impeller cracking, resulting in forced outage of the total plant. Damage diagnostics and condition monitoring of such a turbomachinery system has become an increasingly important and powerful tool to prevent potential failure in components and reduce unplanned forced outage and further maintenance costs, while improving reliability, availability and maintainability of a turbomachinery system. This paper presents a probabilistic signal processing methodology for damage diagnostics using multiple time history data collected from different locations of a turbomachine, considering data uncertainty and multivariate correlation. The proposed methodology is based on the integration of three advanced state-of-the-art data mining techniques: discrete wavelet packet transform, Bayesian hypothesis testing, and probabilistic principal component analysis. The multiresolution wavelet analysis approach is employed to decompose a time series signal into different levels of wavelet coefficients. These coefficients represent multiple time-frequency resolutions of a signal. Bayesian hypothesis testing is then applied to each level of wavelet coefficient to remove possible imperfections. The ratio of posterior odds Bayesian approach provides a direct means to assess whether there is imperfection in the decomposed coefficients, thus avoiding over-denoising. Power spectral density estimated by the Welch method is utilized to evaluate the effectiveness of Bayesian wavelet cleansing method. Furthermore, the probabilistic principal component analysis approach is developed to reduce dimensionality of multiple time series and to address multivariate correlation and data uncertainty for damage diagnostics. The proposed methodology and generalized framework is demonstrated with a set of sensor data collected from a real-world centrifugal compressor with impeller cracks, through both time series and contour analyses of vibration signal and principal components.
2015-01-01
Color is one of the most prominent features of an image and used in many skin and face detection applications. Color space transformation is widely used by researchers to improve face and skin detection performance. Despite the substantial research efforts in this area, choosing a proper color space in terms of skin and face classification performance which can address issues like illumination variations, various camera characteristics and diversity in skin color tones has remained an open issue. This research proposes a new three-dimensional hybrid color space termed SKN by employing the Genetic Algorithm heuristic and Principal Component Analysis to find the optimal representation of human skin color in over seventeen existing color spaces. Genetic Algorithm heuristic is used to find the optimal color component combination setup in terms of skin detection accuracy while the Principal Component Analysis projects the optimal Genetic Algorithm solution to a less complex dimension. Pixel wise skin detection was used to evaluate the performance of the proposed color space. We have employed four classifiers including Random Forest, Naïve Bayes, Support Vector Machine and Multilayer Perceptron in order to generate the human skin color predictive model. The proposed color space was compared to some existing color spaces and shows superior results in terms of pixel-wise skin detection accuracy. Experimental results show that by using Random Forest classifier, the proposed SKN color space obtained an average F-score and True Positive Rate of 0.953 and False Positive Rate of 0.0482 which outperformed the existing color spaces in terms of pixel wise skin detection accuracy. The results also indicate that among the classifiers used in this study, Random Forest is the most suitable classifier for pixel wise skin detection applications. PMID:26267377
Principal semantic components of language and the measurement of meaning.
Samsonovich, Alexei V; Samsonovic, Alexei V; Ascoli, Giorgio A
2010-06-11
Metric systems for semantics, or semantic cognitive maps, are allocations of words or other representations in a metric space based on their meaning. Existing methods for semantic mapping, such as Latent Semantic Analysis and Latent Dirichlet Allocation, are based on paradigms involving dissimilarity metrics. They typically do not take into account relations of antonymy and yield a large number of domain-specific semantic dimensions. Here, using a novel self-organization approach, we construct a low-dimensional, context-independent semantic map of natural language that represents simultaneously synonymy and antonymy. Emergent semantics of the map principal components are clearly identifiable: the first three correspond to the meanings of "good/bad" (valence), "calm/excited" (arousal), and "open/closed" (freedom), respectively. The semantic map is sufficiently robust to allow the automated extraction of synonyms and antonyms not originally in the dictionaries used to construct the map and to predict connotation from their coordinates. The map geometric characteristics include a limited number ( approximately 4) of statistically significant dimensions, a bimodal distribution of the first component, increasing kurtosis of subsequent (unimodal) components, and a U-shaped maximum-spread planar projection. Both the semantic content and the main geometric features of the map are consistent between dictionaries (Microsoft Word and Princeton's WordNet), among Western languages (English, French, German, and Spanish), and with previously established psychometric measures. By defining the semantics of its dimensions, the constructed map provides a foundational metric system for the quantitative analysis of word meaning. Language can be viewed as a cumulative product of human experiences. Therefore, the extracted principal semantic dimensions may be useful to characterize the general semantic dimensions of the content of mental states. This is a fundamental step toward a universal metric system for semantics of human experiences, which is necessary for developing a rigorous science of the mind.
Pan, Hongwei; Lei, Hongjun; Liu, Xin; Wei, Huaibin; Liu, Shufang
2017-09-01
A large number of simple and informal landfills exist in developing countries, which pose as tremendous soil and groundwater pollution threats. Early warning and monitoring of landfill leachate pollution status is of great importance. However, there is a shortage of affordable and effective tools and methods. In this study, a soil column experiment was performed to simulate the pollution status of leachate using three-dimensional excitation-emission fluorescence (3D-EEMF) and parallel factor analysis (PARAFAC) models. Sum of squared residuals (SSR) and principal component analysis (PCA) were used to determine the optimal components for PARAFAC. A one-way analysis of variance showed that the component scores of the soil column leachate were significant influenced by landfill leachate (p<0.05). Therefore, the ratio of the component scores of the soil under the landfill to that of natural soil could be used to evaluate the leakage status of landfill leachate. Furthermore, a hazard index (HI) and a hazard evaluation standard were established. A case study of Kaifeng landfill indicated a low hazard (level 5) by the use of HI. In summation, HI is presented as a tool to evaluate landfill pollution status and for the guidance of municipal solid waste management. Copyright © 2017 Elsevier Ltd. All rights reserved.
Reduced nonlinear prognostic model construction from high-dimensional data
NASA Astrophysics Data System (ADS)
Gavrilov, Andrey; Mukhin, Dmitry; Loskutov, Evgeny; Feigin, Alexander
2017-04-01
Construction of a data-driven model of evolution operator using universal approximating functions can only be statistically justified when the dimension of its phase space is small enough, especially in the case of short time series. At the same time in many applications real-measured data is high-dimensional, e.g. it is space-distributed and multivariate in climate science. Therefore it is necessary to use efficient dimensionality reduction methods which are also able to capture key dynamical properties of the system from observed data. To address this problem we present a Bayesian approach to an evolution operator construction which incorporates two key reduction steps. First, the data is decomposed into a set of certain empirical modes, such as standard empirical orthogonal functions or recently suggested nonlinear dynamical modes (NDMs) [1], and the reduced space of corresponding principal components (PCs) is obtained. Then, the model of evolution operator for PCs is constructed which maps a number of states in the past to the current state. The second step is to reduce this time-extended space in the past using appropriate decomposition methods. Such a reduction allows us to capture only the most significant spatio-temporal couplings. The functional form of the evolution operator includes separately linear, nonlinear (based on artificial neural networks) and stochastic terms. Explicit separation of the linear term from the nonlinear one allows us to more easily interpret degree of nonlinearity as well as to deal better with smooth PCs which can naturally occur in the decompositions like NDM, as they provide a time scale separation. Results of application of the proposed method to climate data are demonstrated and discussed. The study is supported by Government of Russian Federation (agreement #14.Z50.31.0033 with the Institute of Applied Physics of RAS). 1. Mukhin, D., Gavrilov, A., Feigin, A., Loskutov, E., & Kurths, J. (2015). Principal nonlinear dynamical modes of climate variability. Scientific Reports, 5, 15510. http://doi.org/10.1038/srep15510
Shear localization in three-dimensional amorphous solids.
Dasgupta, Ratul; Gendelman, Oleg; Mishra, Pankaj; Procaccia, Itamar; Shor, Carmel A B Z
2013-09-01
In this paper we extend the recent theory of shear localization in two-dimensional (2D) amorphous solids to three dimensions. In two dimensions the fundamental instability of shear localization is related to the appearance of a line of displacement quadrupoles that makes an angle of 45^{∘} with the principal stress axis. In three dimensions the fundamental plastic instability is also explained by the formation of a lattice of anisotropic elastic inclusions. In the case of pure external shear stress, we demonstrate that this is a 2D triangular lattice of similar elementary events. It is shown that this lattice is arranged on a plane that, similarly to the 2D case, makes an angle of 45^{∘} with the principal stress axis. This solution is energetically favorable only if the external strain exceeds a yield-strain value that is determined by the strain parameters of the elementary events and the Poisson ratio. The predictions of the theory are compared to numerical simulations and very good agreement is observed.
NASA Technical Reports Server (NTRS)
Lahti, G. P.
1971-01-01
The method of steepest descent used in optimizing one-dimensional layered radiation shields is extended to multidimensional, multiconstraint situations. The multidimensional optimization algorithm and equations are developed for the case of a dose constraint in any one direction being dependent only on the shield thicknesses in that direction and independent of shield thicknesses in other directions. Expressions are derived for one-, two-, and three-dimensional cases (one, two, and three constraints). The precedure is applicable to the optimization of shields where there are different dose constraints and layering arrangements in the principal directions.
NASA Technical Reports Server (NTRS)
Stabe, R. G.
1971-01-01
A jet-flap blade was designed for a velocity diagram typical of the first-stage stator of a jet engine turbine and was tested in a simple two-dimensional cascade of six blades. The principal measurements were blade surface static pressure and cross-channel surveys of exit total pressure, static pressure, and flow angle. The results of the experimental investigation include blade loading, exit angle, flow, and loss data for a range of exit critical velocity ratios and three jet flow conditions.
Muhammad, Saqib; Han, Shengli; Xie, Xiaoyu; Wang, Sicen; Aziz, Muhammad Majid
2017-01-01
Cell membrane chromatography is a simple, specific, and time-saving technique for studying drug-receptor interactions, screening of active components from complex mixtures, and quality control of traditional Chinese medicines. However, the short column life, low sensitivity, low column efficiency (so cannot resolve satisfactorily mixture of compounds), low peak capacity, and inefficient in structure identification were bottleneck in its application. Combinations of cell membrane chromatography with multidimensional chromatography such as two-dimensional liquid chromatography and high sensitivity detectors like mass have significantly reduced many of the above-mentioned shortcomings. This paper provides an overview of the current advances in online two-dimensional-based cell membrane chromatography for screening target components from traditional Chinese medicines with particular emphasis on the instrumentation, preparation of cell membrane stationary phase, advantages, and disadvantages compared to alternative approaches. The last section of the review summarizes the applications of the online two-dimensional high-performance liquid chromatography based cell membrane chromatography reported since its emergence to date (2010-June 2016). © 2016 WILEY-VCH Verlag GmbH & Co. KGaA, Weinheim.
Understanding the pattern of the BSE Sensex
NASA Astrophysics Data System (ADS)
Mukherjee, I.; Chatterjee, Soumya; Giri, A.; Barat, P.
2017-09-01
An attempt is made to understand the pattern of behaviour of the BSE Sensex by analysing the tick-by-tick Sensex data for the years 2006 to 2012 on yearly as well as cumulative basis using Principal Component Analysis (PCA) and its nonlinear variant Kernel Principal Component Analysis (KPCA). The latter technique ensures that the nonlinear character of the interactions present in the system gets captured in the analysis. The analysis is carried out by constructing vector spaces of varying dimensions. The size of the data set ranges from a minimum of 360,000 for one year to a maximum of 2,520,000 for seven years. In all cases the prices appear to be highly correlated and restricted to a very low dimensional subspace of the original vector space. An external perturbation is added to the system in the form of noise. It is observed that while standard PCA is unable to distinguish the behaviour of the noise-mixed data from that of the original, KPCA clearly identifies the effect of the noise. The exercise is extended in case of daily data of other stock markets and similar results are obtained.
On the Fallibility of Principal Components in Research
ERIC Educational Resources Information Center
Raykov, Tenko; Marcoulides, George A.; Li, Tenglong
2017-01-01
The measurement error in principal components extracted from a set of fallible measures is discussed and evaluated. It is shown that as long as one or more measures in a given set of observed variables contains error of measurement, so also does any principal component obtained from the set. The error variance in any principal component is shown…
Guan, Ming
2017-01-01
Since 1978, rural-urban migrants mainly contribute Chinese urbanization. The purpose of this paper is to examine the effects of socioeconomic factors on mental health of them. Their mental health was measured by 12-item general health questionnaire (GHQ-12). The study sample comprised 5925 migrants obtained from the 2009 rural-to-urban migrants survey (RUMiC). The relationships among the instruments were assessed by the correlation analysis. The one-factor (overall items), two-factor (positive vs. negative items), and model conducted by principal component analysis were tested in the confirmatory factor analysis (CFA). On the basis of three CFA models, the three multiple indicators multiple causes (MIMIC) models with age, gender, marriage, ethnicity, and employment were constructed to investigate the concurrent associations between socioeconomic factors and GHQ-12. Of the sample, only 1.94% were of ethnic origin and mean age was 31.63 (SD = ±10.43) years. The one-factor, two-factor, and three-factor structure (i.e. semi-positive/negative/independent usefulness) had good model fits in the CFA analysis and gave order (i.e. 2 factor>3 factor>1 factor), which suggests that the three models can be used to assess psychological symptoms of migrants in urban China. All MIMIC models had acceptable fit and gave order (i.e. one-dimensional model>two-dimensional model>three-dimensional model). There were weak associations of socioeconomic factors with mental health among migrants in urban China. Policy discussion suggested that improvement of socioeconomic status of rural-urban migrants and mental health systems in urban China should be highlighted and strengthened.
Exploring the Factor Structure of Neurocognitive Measures in Older Individuals
Santos, Nadine Correia; Costa, Patrício Soares; Amorim, Liliana; Moreira, Pedro Silva; Cunha, Pedro; Cotter, Jorge; Sousa, Nuno
2015-01-01
Here we focus on factor analysis from a best practices point of view, by investigating the factor structure of neuropsychological tests and using the results obtained to illustrate on choosing a reasonable solution. The sample (n=1051 individuals) was randomly divided into two groups: one for exploratory factor analysis (EFA) and principal component analysis (PCA), to investigate the number of factors underlying the neurocognitive variables; the second to test the “best fit” model via confirmatory factor analysis (CFA). For the exploratory step, three extraction (maximum likelihood, principal axis factoring and principal components) and two rotation (orthogonal and oblique) methods were used. The analysis methodology allowed exploring how different cognitive/psychological tests correlated/discriminated between dimensions, indicating that to capture latent structures in similar sample sizes and measures, with approximately normal data distribution, reflective models with oblimin rotation might prove the most adequate. PMID:25880732
Han, Te; Jiang, Dongxiang; Zhang, Xiaochen; Sun, Yankui
2017-03-27
Rotating machinery is widely used in industrial applications. With the trend towards more precise and more critical operating conditions, mechanical failures may easily occur. Condition monitoring and fault diagnosis (CMFD) technology is an effective tool to enhance the reliability and security of rotating machinery. In this paper, an intelligent fault diagnosis method based on dictionary learning and singular value decomposition (SVD) is proposed. First, the dictionary learning scheme is capable of generating an adaptive dictionary whose atoms reveal the underlying structure of raw signals. Essentially, dictionary learning is employed as an adaptive feature extraction method regardless of any prior knowledge. Second, the singular value sequence of learned dictionary matrix is served to extract feature vector. Generally, since the vector is of high dimensionality, a simple and practical principal component analysis (PCA) is applied to reduce dimensionality. Finally, the K -nearest neighbor (KNN) algorithm is adopted for identification and classification of fault patterns automatically. Two experimental case studies are investigated to corroborate the effectiveness of the proposed method in intelligent diagnosis of rotating machinery faults. The comparison analysis validates that the dictionary learning-based matrix construction approach outperforms the mode decomposition-based methods in terms of capacity and adaptability for feature extraction.
Xiao, Keke; Chen, Yun; Jiang, Xie; Zhou, Yan
2017-03-01
An investigation was conducted for 20 different types of sludge in order to identify the key organic compounds in extracellular polymeric substances (EPS) that are important in assessing variations of sludge filterability. The different types of sludge varied in initial total solids (TS) content, organic composition and pre-treatment methods. For instance, some of the sludges were pre-treated by acid, ultrasonic, thermal, alkaline, or advanced oxidation technique. The Pearson's correlation results showed significant correlations between sludge filterability and zeta potential, pH, dissolved organic carbon, protein and polysaccharide in soluble EPS (SB EPS), loosely bound EPS (LB EPS) and tightly bound EPS (TB EPS). The principal component analysis (PCA) method was used to further explore correlations between variables and similarities among EPS fractions of different types of sludge. Two principal components were extracted: principal component 1 accounted for 59.24% of total EPS variations, while principal component 2 accounted for 25.46% of total EPS variations. Dissolved organic carbon, protein and polysaccharide in LB EPS showed higher eigenvector projection values than the corresponding compounds in SB EPS and TB EPS in principal component 1. Further characterization of fractionized key organic compounds in LB EPS was conducted with size-exclusion chromatography-organic carbon detection-organic nitrogen detection (LC-OCD-OND). A numerical multiple linear regression model was established to describe relationship between organic compounds in LB EPS and sludge filterability. Copyright © 2016 Elsevier Ltd. All rights reserved.
Vaníčková, Lucie; Břízová, Radka; Pompeiano, Antonio; Ferreira, Luana Lima; de Aquino, Nathaly Costa; Tavares, Raphael de Farias; Rodriguez, Laura D.; Mendonça, Adriana de Lima; Canal, Nelson Augusto; do Nascimento, Ruth Rufino
2015-01-01
Abstract Fruit fly sexual behaviour is directly influenced by chemical and non-chemical cues that play important roles in reproductive isolation. The chemical profiles of pheromones and cuticular hydrocarbons (CHs) of eight fruit fly populations of the Andean, Brazilian-1 and Brazilian-3 morphotypes of the Anastrepha fraterculus cryptic species complex originating from Colombia (four populations) and Brazil (four populations) were analysed using two-dimensional gas chromatography with mass spectrometric detection. The resulting chemical diversity data were studied using principal component analyses. Andean morphotypes could be discriminated from the Brazilian-1 and Brazilian-3 morphotypes by means of male-borne pheromones and/or male and female CH profiles. The Brazilian-1 and Brazilian-3 morphotypes were found to be monophyletic. The use of chemical profiles as species- and sex-specific signatures for cryptic species separations is discussed. PMID:26798260
Oberg, Tomas
2004-01-01
Halogenated aliphatic compounds have many technical uses, but substances within this group are also ubiquitous environmental pollutants that can affect the ozone layer and contribute to global warming. The establishment of quantitative structure-property relationships is of interest not only to fill in gaps in the available database but also to validate experimental data already acquired. The three-dimensional structures of 240 compounds were modeled with molecular mechanics prior to the generation of empirical descriptors. Two bilinear projection methods, principal component analysis (PCA) and partial-least-squares regression (PLSR), were used to identify outliers. PLSR was subsequently used to build a multivariate calibration model by extracting the latent variables that describe most of the covariation between the molecular structure and the boiling point. Boiling points were also estimated with an extension of the group contribution method of Stein and Brown.
Unsupervised Feature Learning for Heart Sounds Classification Using Autoencoder
NASA Astrophysics Data System (ADS)
Hu, Wei; Lv, Jiancheng; Liu, Dongbo; Chen, Yao
2018-04-01
Cardiovascular disease seriously threatens the health of many people. It is usually diagnosed during cardiac auscultation, which is a fast and efficient method of cardiovascular disease diagnosis. In recent years, deep learning approach using unsupervised learning has made significant breakthroughs in many fields. However, to our knowledge, deep learning has not yet been used for heart sound classification. In this paper, we first use the average Shannon energy to extract the envelope of the heart sounds, then find the highest point of S1 to extract the cardiac cycle. We convert the time-domain signals of the cardiac cycle into spectrograms and apply principal component analysis whitening to reduce the dimensionality of the spectrogram. Finally, we apply a two-layer autoencoder to extract the features of the spectrogram. The experimental results demonstrate that the features from the autoencoder are suitable for heart sound classification.
Sigmundsson, H; Polman, R C J; Lorås, H
2013-08-01
Individual differences in mathematical skills are typically explained by an innate capability to solve mathematical tasks. At the behavioural level this implies a consistent level of mathematical achievement that can be captured by strong relationships between tasks, as well as by a single statistical dimension that underlies performance on all mathematical tasks. To investigate this general assumption, the present study explored interrelations and dimensions of mathematical skills. For this purpose, 68 ten-year-old children from two schools were tested using nine mathematics tasks from the Basic Knowledge in Mathematics Test. Relatively low-to-moderate correlations between the mathematics tasks indicated most tasks shared less than 25% of their variance. There were four principal components, accounting for 70% of the variance in mathematical skill across tasks and participants. The high specificity in mathematical skills was discussed in relation to the principle of task specificity of learning.
imDEV: a graphical user interface to R multivariate analysis tools in Microsoft Excel.
Grapov, Dmitry; Newman, John W
2012-09-01
Interactive modules for Data Exploration and Visualization (imDEV) is a Microsoft Excel spreadsheet embedded application providing an integrated environment for the analysis of omics data through a user-friendly interface. Individual modules enables interactive and dynamic analyses of large data by interfacing R's multivariate statistics and highly customizable visualizations with the spreadsheet environment, aiding robust inferences and generating information-rich data visualizations. This tool provides access to multiple comparisons with false discovery correction, hierarchical clustering, principal and independent component analyses, partial least squares regression and discriminant analysis, through an intuitive interface for creating high-quality two- and a three-dimensional visualizations including scatter plot matrices, distribution plots, dendrograms, heat maps, biplots, trellis biplots and correlation networks. Freely available for download at http://sourceforge.net/projects/imdev/. Implemented in R and VBA and supported by Microsoft Excel (2003, 2007 and 2010).
Surface-enhanced Raman spectra of hemoglobin for esophageal cancer diagnosis
NASA Astrophysics Data System (ADS)
Zhou, Xue; Diao, Zhenqi; Fan, Chunzhen; Guo, Huiqiang; Xiong, Yang; Tang, Weiyue
2014-03-01
Surface-enhanced Raman scattering (SERS) spectra of hemoglobin from 30 esophageal cancer patients and 30 healthy persons have been detected and analyzed. The results indicate that, there are more iron ions in low spin state and less in high for the hemoglobin of esophageal cancer patients than normal persons, which is consistent with the fact that it is easier to hemolyze for the blood of cancer patients. By using principal component analysis (PCA) and discriminate analysis, we can get a three-dimensional scatter plot of PC scores from the SERS spectra of healthy persons and cancer patients, from which the two groups can be discriminated. The total accuracy of this method is 90%, while the diagnostic specificity is 93.3% and sensitivity is 86.7%. Thus SERS spectra of hemoglobin analysis combined with PCA may be a new technique for the early diagnose of esophageal cancer.
Optimal wavelength band clustering for multispectral iris recognition.
Gong, Yazhuo; Zhang, David; Shi, Pengfei; Yan, Jingqi
2012-07-01
This work explores the possibility of clustering spectral wavelengths based on the maximum dissimilarity of iris textures. The eventual goal is to determine how many bands of spectral wavelengths will be enough for iris multispectral fusion and to find these bands that will provide higher performance of iris multispectral recognition. A multispectral acquisition system was first designed for imaging the iris at narrow spectral bands in the range of 420 to 940 nm. Next, a set of 60 human iris images that correspond to the right and left eyes of 30 different subjects were acquired for an analysis. Finally, we determined that 3 clusters were enough to represent the 10 feature bands of spectral wavelengths using the agglomerative clustering based on two-dimensional principal component analysis. The experimental results suggest (1) the number, center, and composition of clusters of spectral wavelengths and (2) the higher performance of iris multispectral recognition based on a three wavelengths-bands fusion.
An Exospheric Temperature Model Based On CHAMP Observations and TIEGCM Simulations
NASA Astrophysics Data System (ADS)
Ruan, Haibing; Lei, Jiuhou; Dou, Xiankang; Liu, Siqing; Aa, Ercha
2018-02-01
In this work, thermospheric densities from the accelerometer measurement on board the CHAMP satellite during 2002-2009 and the simulations from the National Center for Atmospheric Research Thermosphere Ionosphere Electrodynamics General Circulation Model (NCAR-TIEGCM) are employed to develop an empirical exospheric temperature model (ETM). The two-dimensional basis functions of the ETM are first provided from the principal component analysis of the TIEGCM simulations. Based on the exospheric temperatures derived from CHAMP thermospheric densities, a global distribution of the exospheric temperatures is reconstructed. A parameterization is conducted for each basis function amplitude as a function of solar-geophysical and seasonal conditions. Thus, the ETM can be utilized to model the thermospheric temperature and mass density under a specified condition. Our results showed that the averaged standard deviation of the ETM is generally less than 10% than approximately 30% in the MSIS model. Besides, the ETM reproduces the global thermospheric evolutions including the equatorial thermosphere anomaly.
Research on assessment and improvement method of remote sensing image reconstruction
NASA Astrophysics Data System (ADS)
Sun, Li; Hua, Nian; Yu, Yanbo; Zhao, Zhanping
2018-01-01
Remote sensing image quality assessment and improvement is an important part of image processing. Generally, the use of compressive sampling theory in remote sensing imaging system can compress images while sampling which can improve efficiency. A method of two-dimensional principal component analysis (2DPCA) is proposed to reconstruct the remote sensing image to improve the quality of the compressed image in this paper, which contain the useful information of image and can restrain the noise. Then, remote sensing image quality influence factors are analyzed, and the evaluation parameters for quantitative evaluation are introduced. On this basis, the quality of the reconstructed images is evaluated and the different factors influence on the reconstruction is analyzed, providing meaningful referential data for enhancing the quality of remote sensing images. The experiment results show that evaluation results fit human visual feature, and the method proposed have good application value in the field of remote sensing image processing.
Factor structure and dimensionality of the two depression scales in STAR*D using level 1 datasets.
Bech, P; Fava, M; Trivedi, M H; Wisniewski, S R; Rush, A J
2011-08-01
The factor structure and dimensionality of the HAM-D(17) and the IDS-C(30) are as yet uncertain, because psychometric analyses of these scales have been performed without a clear separation between factor structure profile and dimensionality (total scores being a sufficient statistic). The first treatment step (Level 1) in the STAR*D study provided a dataset of 4041 outpatients with DSM-IV nonpsychotic major depression. The HAM-D(17) and IDS-C(30) were evaluated by principal component analysis (PCA) without rotation. Mokken analysis tested the unidimensionality of the IDS-C(6), which corresponds to the unidimensional HAM-D(6.) For both the HAM-D(17) and IDS-C(30), PCA identified a bi-directional factor contrasting the depressive symptoms versus the neurovegetative symptoms. The HAM-D(6) and the corresponding IDS-C(6) symptoms all emerged in the depression factor. Both the HAM-D(6) and IDS-C(6) were found to be unidimensional scales, i.e., their total scores are each a sufficient statistic for the measurement of depressive states. STAR*D used only one medication in Level 1. The unidimensional HAM-D(6) and IDS-C(6) should be used when evaluating the pure clinical effect of antidepressive treatment, whereas the multidimensional HAM-D(17) and IDS-C(30) should be considered when selecting antidepressant treatment. Copyright © 2011 Elsevier B.V. All rights reserved.
Fazal, Md Abul; Palmer, Vanessa R; Dovichi, Norman J
2006-10-20
Differential detergent fractionation was used to sequentially extract cytosolic, membrane, nuclear, and cytoskeletal fractions from AtT-20 cells. Extracted components were denatured by sodium dodecyl sulfate (SDS) and then labeled with the fluorogenic reagent 3-(2-furoyl) quinoline-1-carboxaldehyde. Both capillary sieving electrophoresis (CSE) and micellar electrokinetic capillary chromatography (MECC) were used to separate labeled components by one-dimensional (1D) electrophoresis. Labeled components were also separated by two-dimensional (2D) capillary electrophoresis; CSE was employed in the first dimension and MECC in the second dimension. Roughly 150 fractions were transferred from the first to the second capillary for this comprehensive analysis in 2.5 h.
Dimensions and profiles of the generalized health-related self-concept.
Wiesmann, Ulrich; Niehörster, Gabriele; Hannich, Hans-Joachim; Hartmann, Ute
2008-11-01
We explore the significance of health as a potentially self-relevant category from the perspective of dynamic self-concept theory. Our intention was to describe the dimensional structure of the generalized health-related self-concept, to identify particular prototypes of health-related self-definition, and to see if these prototypes would differ with respect to appraisals of health behaviour and subjective health. We conducted a cross-sectional questionnaire study involving 545 college students (23.3% male) at the mean age of 22 years. The self-administered questionnaire assessed a relevant spectrum of health-related cognitions denoting their generalized declarative knowledge about their health (the generalized health-related self-concept). Additionally, participants rated their multiple health behaviour, their perceived health, and their anticipated vulnerability. A principal components analysis of the health-related cognitions revealed the following five dimensions: health-protective dispositions, health-protective motivation, vulnerability, health-risky habits, and external, avoidant motivation. A two-step cluster analysis of the five components identified six profiles of health-related self-concept: careless/carefree, omnipotents, risk-takers, mentally affected, reluctant-avoidant, and medically fragile. These prototypes could be successfully reclassified (97.6%). The six profiles differed with respect to their health behaviour and subjective health appraisals. The dimensional structure represents both resources and deficits with respect to an individual's health-related self-concept. An individual's profile of these dimensions might correspond to a characteristic set of particular health needs and motivations. Successful health communications should follow a complementary strategy of affirming the self-concept.
Fault Detection of Bearing Systems through EEMD and Optimization Algorithm
Lee, Dong-Han; Ahn, Jong-Hyo; Koh, Bong-Hwan
2017-01-01
This study proposes a fault detection and diagnosis method for bearing systems using ensemble empirical mode decomposition (EEMD) based feature extraction, in conjunction with particle swarm optimization (PSO), principal component analysis (PCA), and Isomap. First, a mathematical model is assumed to generate vibration signals from damaged bearing components, such as the inner-race, outer-race, and rolling elements. The process of decomposing vibration signals into intrinsic mode functions (IMFs) and extracting statistical features is introduced to develop a damage-sensitive parameter vector. Finally, PCA and Isomap algorithm are used to classify and visualize this parameter vector, to separate damage characteristics from healthy bearing components. Moreover, the PSO-based optimization algorithm improves the classification performance by selecting proper weightings for the parameter vector, to maximize the visualization effect of separating and grouping of parameter vectors in three-dimensional space. PMID:29143772
Strale, Mathieu; Krysinska, Karolina; Overmeiren, Gaëtan Van; Andriessen, Karl
2017-06-01
This study investigated the geographic distribution of suicide and railway suicide in Belgium over 2008--2013 on local (i.e., district or arrondissement) level. There were differences in the regional distribution of suicide and railway suicides in Belgium over the study period. Principal component analysis identified three groups of correlations among population variables and socio-economic indicators, such as population density, unemployment, and age group distribution, on two components that helped explaining the variance of railway suicide at a local (arrondissement) level. This information is of particular importance to prevent suicides in high-risk areas on the Belgian railway network.
A numerical study of blood flow using mixture theory
Wu, Wei-Tao; Aubry, Nadine; Massoudi, Mehrdad; Kim, Jeongho; Antaki, James F.
2014-01-01
In this paper, we consider the two dimensional flow of blood in a rectangular microfluidic channel. We use Mixture Theory to treat this problem as a two-component system: One component is the red blood cells (RBCs) modeled as a generalized Reiner–Rivlin type fluid, which considers the effects of volume fraction (hematocrit) and influence of shear rate upon viscosity. The other component, plasma, is assumed to behave as a linear viscous fluid. A CFD solver based on OpenFOAM® was developed and employed to simulate a specific problem, namely blood flow in a two dimensional micro-channel, is studied. Finally to better understand this two-component flow system and the effects of the different parameters, the equations are made dimensionless and a parametric study is performed. PMID:24791016
A numerical study of blood flow using mixture theory.
Wu, Wei-Tao; Aubry, Nadine; Massoudi, Mehrdad; Kim, Jeongho; Antaki, James F
2014-03-01
In this paper, we consider the two dimensional flow of blood in a rectangular microfluidic channel. We use Mixture Theory to treat this problem as a two-component system: One component is the red blood cells (RBCs) modeled as a generalized Reiner-Rivlin type fluid, which considers the effects of volume fraction (hematocrit) and influence of shear rate upon viscosity. The other component, plasma, is assumed to behave as a linear viscous fluid. A CFD solver based on OpenFOAM ® was developed and employed to simulate a specific problem, namely blood flow in a two dimensional micro-channel, is studied. Finally to better understand this two-component flow system and the effects of the different parameters, the equations are made dimensionless and a parametric study is performed.
Chang, Chi-Ying; Chang, Chia-Chi; Hsiao, Tzu-Chien
2013-01-01
Excitation-emission matrix (EEM) fluorescence spectroscopy is a noninvasive method for tissue diagnosis and has become important in clinical use. However, the intrinsic characterization of EEM fluorescence remains unclear. Photobleaching and the complexity of the chemical compounds make it difficult to distinguish individual compounds due to overlapping features. Conventional studies use principal component analysis (PCA) for EEM fluorescence analysis, and the relationship between the EEM features extracted by PCA and diseases has been examined. The spectral features of different tissue constituents are not fully separable or clearly defined. Recently, a non-stationary method called multi-dimensional ensemble empirical mode decomposition (MEEMD) was introduced; this method can extract the intrinsic oscillations on multiple spatial scales without loss of information. The aim of this study was to propose a fluorescence spectroscopy system for EEM measurements and to describe a method for extracting the intrinsic characteristics of EEM by MEEMD. The results indicate that, although PCA provides the principal factor for the spectral features associated with chemical compounds, MEEMD can provide additional intrinsic features with more reliable mapping of the chemical compounds. MEEMD has the potential to extract intrinsic fluorescence features and improve the detection of biochemical changes. PMID:24240806
Real-Time Detection of In-flight Aircraft Damage
Blair, Brenton; Lee, Herbert K. H.; Davies, Misty
2017-10-02
When there is damage to an aircraft, it is critical to be able to quickly detect and diagnose the problem so that the pilot can attempt to maintain control of the aircraft and land it safely. We develop methodology for real-time classification of flight trajectories to be able to distinguish between an undamaged aircraft and five different damage scenarios. Principal components analysis allows a lower-dimensional representation of multi-dimensional trajectory information in time. Random Forests provide a computationally efficient approach with sufficient accuracy to be able to detect and classify the different scenarios in real-time. We demonstrate our approach by classifyingmore » realizations of a 45 degree bank angle generated from the Generic Transport Model flight simulator in collaboration with NASA.« less
Real-Time Detection of In-flight Aircraft Damage
DOE Office of Scientific and Technical Information (OSTI.GOV)
Blair, Brenton; Lee, Herbert K. H.; Davies, Misty
When there is damage to an aircraft, it is critical to be able to quickly detect and diagnose the problem so that the pilot can attempt to maintain control of the aircraft and land it safely. We develop methodology for real-time classification of flight trajectories to be able to distinguish between an undamaged aircraft and five different damage scenarios. Principal components analysis allows a lower-dimensional representation of multi-dimensional trajectory information in time. Random Forests provide a computationally efficient approach with sufficient accuracy to be able to detect and classify the different scenarios in real-time. We demonstrate our approach by classifyingmore » realizations of a 45 degree bank angle generated from the Generic Transport Model flight simulator in collaboration with NASA.« less
NASA Astrophysics Data System (ADS)
Santarius, John; Navarro, Marcos; Michalak, Matthew; Fancher, Aaron; Kulcinski, Gerald; Bonomo, Richard
2016-10-01
A newly initiated research project will be described that investigates methods for detecting shielded special nuclear materials by combining multi-dimensional neutron sources, forward/adjoint calculations modeling neutron and gamma transport, and sparse data analysis of detector signals. The key tasks for this project are: (1) developing a radiation transport capability for use in optimizing adaptive-geometry, inertial-electrostatic confinement (IEC) neutron source/detector configurations for neutron pulses distributed in space and/or phased in time; (2) creating distributed-geometry, gas-target, IEC fusion neutron sources; (3) applying sparse data and noise reduction algorithms, such as principal component analysis (PCA) and wavelet transform analysis, to enhance detection fidelity; and (4) educating graduate and undergraduate students. Funded by DHS DNDO Project 2015-DN-077-ARI095.
Dynamic contact guidance of migrating cells
NASA Astrophysics Data System (ADS)
Losert, Wolfgang; Sun, Xiaoyu; Guven, Can; Driscoll, Meghan; Fourkas, John
2014-03-01
We investigate the effects of nanotopographical surfaces on the cell migration and cell shape dynamics of the amoeba Dictyostelium discoideum. Amoeboid motion exhibits significant contact guidance along surfaces with nanoscale ridges or grooves. We show quantitatively that nanoridges spaced 1.5 μm apart exhibit the greatest contact guidance efficiency. Using principal component analysis, we characterize the dynamics of the cell shape modulated by the coupling between the cell membrane and ridges. We show that motion parallel to the ridges is enhanced, while the turning, at the largest spatial scales, is suppressed. Since protrusion dynamics are principally governed by actin dynamics, we imaged the actin polymerization of cells on ridges. We found that actin polymerization occurs preferentially along nanoridges in a ``monorail'' like fashion. The ridges then provide us with a tool to study actin dynamics in an effectively reduced dimensional system.
NASA Astrophysics Data System (ADS)
Dafu, Shen; Leihong, Zhang; Dong, Liang; Bei, Li; Yi, Kang
2017-07-01
The purpose of this study is to improve the reconstruction precision and better copy the color of spectral image surfaces. A new spectral reflectance reconstruction algorithm based on an iterative threshold combined with weighted principal component space is presented in this paper, and the principal component with weighted visual features is the sparse basis. Different numbers of color cards are selected as the training samples, a multispectral image is the testing sample, and the color differences in the reconstructions are compared. The channel response value is obtained by a Mega Vision high-accuracy, multi-channel imaging system. The results show that spectral reconstruction based on weighted principal component space is superior in performance to that based on traditional principal component space. Therefore, the color difference obtained using the compressive-sensing algorithm with weighted principal component analysis is less than that obtained using the algorithm with traditional principal component analysis, and better reconstructed color consistency with human eye vision is achieved.
Kernel PLS-SVC for Linear and Nonlinear Discrimination
NASA Technical Reports Server (NTRS)
Rosipal, Roman; Trejo, Leonard J.; Matthews, Bryan
2003-01-01
A new methodology for discrimination is proposed. This is based on kernel orthonormalized partial least squares (PLS) dimensionality reduction of the original data space followed by support vector machines for classification. Close connection of orthonormalized PLS and Fisher's approach to linear discrimination or equivalently with canonical correlation analysis is described. This gives preference to use orthonormalized PLS over principal component analysis. Good behavior of the proposed method is demonstrated on 13 different benchmark data sets and on the real world problem of the classification finger movement periods versus non-movement periods based on electroencephalogram.
A Principle Component Analysis of Galaxy Properties from a Large, Gas-Selected Sample
Chang, Yu-Yen; Chao, Rikon; Wang, Wei-Hao; ...
2012-01-01
Disney emore » t al. (2008) have found a striking correlation among global parameters of H i -selected galaxies and concluded that this is in conflict with the CDM model. Considering the importance of the issue, we reinvestigate the problem using the principal component analysis on a fivefold larger sample and additional near-infrared data. We use databases from the Arecibo Legacy Fast Arecibo L -band Feed Array Survey for the gas properties, the Sloan Digital Sky Survey for the optical properties, and the Two Micron All Sky Survey for the near-infrared properties. We confirm that the parameters are indeed correlated where a single physical parameter can explain 83% of the variations. When color ( g - i ) is included, the first component still dominates but it develops a second principal component. In addition, the near-infrared color ( i - J ) shows an obvious second principal component that might provide evidence of the complex old star formation. Based on our data, we suggest that it is premature to pronounce the failure of the CDM model and it motivates more theoretical work.« less
Psychometric properties of the Image of God Scale in breast cancer survivors.
Schreiber, Judy A
2012-07-01
To examine the psychometric properties of the Image of God Scale (IGS) in a clinical population. Descriptive, cross-sectional. University and community oncology practices in the southeastern United States. 123 breast cancer survivors no more than two years from completion of treatment. Scale reliability was determined with the coefficient alpha. Instrument dimensionality was examined using principal component analysis. Construct validity was evaluated by examining correlations with other instruments used in the study. An individual's image of God. Internal consistency was strong (anger subscale = 0.8; engagement subscale = 0.89). The principle component analysis resulted in a two-factor solution with items loading uniquely on Factor 1-Engagement (8) and Factor 2-Anger (6). Significant correlations between the IGS and religious coping support convergence on a God concept. Correlations with psychological well-being, psychological distress, and concern about recurrence were nonsignificant (engagement) or inverse (anger), supporting discrimination between concepts of God and psychological adjustment. The IGS is a unique measure of how God is viewed by the depth and character of His involvement with the individual and the world. The IGS may be a measure that can transcend sects, denominations, and religions by identifying the image of God that underlies and defines an individuals' worldview, which influences their attitudes and behaviors.
NASA Astrophysics Data System (ADS)
Rohmer, Jeremy
2016-04-01
Predicting the temporal evolution of landslides is typically supported by numerical modelling. Dynamic sensitivity analysis aims at assessing the influence of the landslide properties on the time-dependent predictions (e.g., time series of landslide displacements). Yet two major difficulties arise: 1. Global sensitivity analysis require running the landslide model a high number of times (> 1000), which may become impracticable when the landslide model has a high computation time cost (> several hours); 2. Landslide model outputs are not scalar, but function of time, i.e. they are n-dimensional vectors with n usually ranging from 100 to 1000. In this article, I explore the use of a basis set expansion, such as principal component analysis, to reduce the output dimensionality to a few components, each of them being interpreted as a dominant mode of variation in the overall structure of the temporal evolution. The computationally intensive calculation of the Sobol' indices for each of these components are then achieved through meta-modelling, i.e. by replacing the landslide model by a "costless-to-evaluate" approximation (e.g., a projection pursuit regression model). The methodology combining "basis set expansion - meta-model - Sobol' indices" is then applied to the La Frasse landslide to investigate the dynamic sensitivity analysis of the surface horizontal displacements to the slip surface properties during the pore pressure changes. I show how to extract information on the sensitivity of each main modes of temporal behaviour using a limited number (a few tens) of long running simulations. In particular, I identify the parameters, which trigger the occurrence of a turning point marking a shift between a regime of low values of landslide displacements and one of high values.
An analysis of curvature effects for the control of wall-bounded shear flows
NASA Technical Reports Server (NTRS)
Gatski, T. B.; Savill, A. M.
1989-01-01
The Reynolds stress transport equations are used to predict the effects of simultaneous and sequential combinations of distortions on turbulent boundary layers. The equations are written in general orthogonal curvilinear coordinates, with the curvature terms expressed in terms of the principal radii of curvature of the respective coordinate surfaces. Results are obtained for the cases of two-dimensional and three-dimensional flows in the limit where production and pressure-strain redistribution dominate over diffusion effects.
Classical Testing in Functional Linear Models.
Kong, Dehan; Staicu, Ana-Maria; Maity, Arnab
2016-01-01
We extend four tests common in classical regression - Wald, score, likelihood ratio and F tests - to functional linear regression, for testing the null hypothesis, that there is no association between a scalar response and a functional covariate. Using functional principal component analysis, we re-express the functional linear model as a standard linear model, where the effect of the functional covariate can be approximated by a finite linear combination of the functional principal component scores. In this setting, we consider application of the four traditional tests. The proposed testing procedures are investigated theoretically for densely observed functional covariates when the number of principal components diverges. Using the theoretical distribution of the tests under the alternative hypothesis, we develop a procedure for sample size calculation in the context of functional linear regression. The four tests are further compared numerically for both densely and sparsely observed noisy functional data in simulation experiments and using two real data applications.
Classical Testing in Functional Linear Models
Kong, Dehan; Staicu, Ana-Maria; Maity, Arnab
2016-01-01
We extend four tests common in classical regression - Wald, score, likelihood ratio and F tests - to functional linear regression, for testing the null hypothesis, that there is no association between a scalar response and a functional covariate. Using functional principal component analysis, we re-express the functional linear model as a standard linear model, where the effect of the functional covariate can be approximated by a finite linear combination of the functional principal component scores. In this setting, we consider application of the four traditional tests. The proposed testing procedures are investigated theoretically for densely observed functional covariates when the number of principal components diverges. Using the theoretical distribution of the tests under the alternative hypothesis, we develop a procedure for sample size calculation in the context of functional linear regression. The four tests are further compared numerically for both densely and sparsely observed noisy functional data in simulation experiments and using two real data applications. PMID:28955155
Rapid 3D Reconstruction for Image Sequence Acquired from UAV Camera
Qu, Yufu; Huang, Jianyu; Zhang, Xuan
2018-01-01
In order to reconstruct three-dimensional (3D) structures from an image sequence captured by unmanned aerial vehicles’ camera (UAVs) and improve the processing speed, we propose a rapid 3D reconstruction method that is based on an image queue, considering the continuity and relevance of UAV camera images. The proposed approach first compresses the feature points of each image into three principal component points by using the principal component analysis method. In order to select the key images suitable for 3D reconstruction, the principal component points are used to estimate the interrelationships between images. Second, these key images are inserted into a fixed-length image queue. The positions and orientations of the images are calculated, and the 3D coordinates of the feature points are estimated using weighted bundle adjustment. With this structural information, the depth maps of these images can be calculated. Next, we update the image queue by deleting some of the old images and inserting some new images into the queue, and a structural calculation of all the images can be performed by repeating the previous steps. Finally, a dense 3D point cloud can be obtained using the depth–map fusion method. The experimental results indicate that when the texture of the images is complex and the number of images exceeds 100, the proposed method can improve the calculation speed by more than a factor of four with almost no loss of precision. Furthermore, as the number of images increases, the improvement in the calculation speed will become more noticeable. PMID:29342908
Ding, Haiquan; Lu, Qipeng; Gao, Hongzhi; Peng, Zhongqi
2014-01-01
To facilitate non-invasive diagnosis of anemia, specific equipment was developed, and non-invasive hemoglobin (HB) detection method based on back propagation artificial neural network (BP-ANN) was studied. In this paper, we combined a broadband light source composed of 9 LEDs with grating spectrograph and Si photodiode array, and then developed a high-performance spectrophotometric system. By using this equipment, fingertip spectra of 109 volunteers were measured. In order to deduct the interference of redundant data, principal component analysis (PCA) was applied to reduce the dimensionality of collected spectra. Then the principal components of the spectra were taken as input of BP-ANN model. On this basis we obtained the optimal network structure, in which node numbers of input layer, hidden layer, and output layer was 9, 11, and 1. Calibration and correction sample sets were used for analyzing the accuracy of non-invasive hemoglobin measurement, and prediction sample set was used for testing the adaptability of the model. The correlation coefficient of network model established by this method is 0.94, standard error of calibration, correction, and prediction are 11.29g/L, 11.47g/L, and 11.01g/L respectively. The result proves that there exist good correlations between spectra of three sample sets and actual hemoglobin level, and the model has a good robustness. It is indicated that the developed spectrophotometric system has potential for the non-invasive detection of HB levels with the method of BP-ANN combined with PCA. PMID:24761296
Rapid 3D Reconstruction for Image Sequence Acquired from UAV Camera.
Qu, Yufu; Huang, Jianyu; Zhang, Xuan
2018-01-14
In order to reconstruct three-dimensional (3D) structures from an image sequence captured by unmanned aerial vehicles' camera (UAVs) and improve the processing speed, we propose a rapid 3D reconstruction method that is based on an image queue, considering the continuity and relevance of UAV camera images. The proposed approach first compresses the feature points of each image into three principal component points by using the principal component analysis method. In order to select the key images suitable for 3D reconstruction, the principal component points are used to estimate the interrelationships between images. Second, these key images are inserted into a fixed-length image queue. The positions and orientations of the images are calculated, and the 3D coordinates of the feature points are estimated using weighted bundle adjustment. With this structural information, the depth maps of these images can be calculated. Next, we update the image queue by deleting some of the old images and inserting some new images into the queue, and a structural calculation of all the images can be performed by repeating the previous steps. Finally, a dense 3D point cloud can be obtained using the depth-map fusion method. The experimental results indicate that when the texture of the images is complex and the number of images exceeds 100, the proposed method can improve the calculation speed by more than a factor of four with almost no loss of precision. Furthermore, as the number of images increases, the improvement in the calculation speed will become more noticeable.
Principal Component and Linkage Analysis of Cardiovascular Risk Traits in the Norfolk Isolate
Cox, Hannah C.; Bellis, Claire; Lea, Rod A.; Quinlan, Sharon; Hughes, Roger; Dyer, Thomas; Charlesworth, Jac; Blangero, John; Griffiths, Lyn R.
2009-01-01
Objective(s) An individual's risk of developing cardiovascular disease (CVD) is influenced by genetic factors. This study focussed on mapping genetic loci for CVD-risk traits in a unique population isolate derived from Norfolk Island. Methods This investigation focussed on 377 individuals descended from the population founders. Principal component analysis was used to extract orthogonal components from 11 cardiovascular risk traits. Multipoint variance component methods were used to assess genome-wide linkage using SOLAR to the derived factors. A total of 285 of the 377 related individuals were informative for linkage analysis. Results A total of 4 principal components accounting for 83% of the total variance were derived. Principal component 1 was loaded with body size indicators; principal component 2 with body size, cholesterol and triglyceride levels; principal component 3 with the blood pressures; and principal component 4 with LDL-cholesterol and total cholesterol levels. Suggestive evidence of linkage for principal component 2 (h2 = 0.35) was observed on chromosome 5q35 (LOD = 1.85; p = 0.0008). While peak regions on chromosome 10p11.2 (LOD = 1.27; p = 0.005) and 12q13 (LOD = 1.63; p = 0.003) were observed to segregate with principal components 1 (h2 = 0.33) and 4 (h2 = 0.42), respectively. Conclusion(s): This study investigated a number of CVD risk traits in a unique isolated population. Findings support the clustering of CVD risk traits and provide interesting evidence of a region on chromosome 5q35 segregating with weight, waist circumference, HDL-c and total triglyceride levels. PMID:19339786
Shape Analysis of 3D Head Scan Data for U.S. Respirator Users
NASA Astrophysics Data System (ADS)
Zhuang, Ziqing; Slice, DennisE; Benson, Stacey; Lynch, Stephanie; Viscusi, DennisJ
2010-12-01
In 2003, the National Institute for Occupational Safety and Health (NIOSH) conducted a head-and-face anthropometric survey of diverse, civilian respirator users. Of the 3,997 subjects measured using traditional anthropometric techniques, surface scans and 26 three-dimensional (3D) landmark locations were collected for 947 subjects. The objective of this study was to report the size and shape variation of the survey participants using the 3D data. Generalized Procrustes Analysis (GPA) was conducted to standardize configurations of landmarks associated with individuals into a common coordinate system. The superimposed coordinates for each individual were used as commensurate variables that describe individual shape and were analyzed using Principal Component Analysis (PCA) to identify population variation. The first four principal components (PC) account for 49% of the total sample variation. The first PC indicates that overall size is an important component of facial variability. The second PC accounts for long and narrow or short and wide faces. Longer narrow orbits versus shorter wider orbits can be described by PC3, and PC4 represents variation in the degree of ortho/prognathism. Geometric Morphometrics provides a detailed and interpretable assessment of morphological variation that may be useful in assessing respirators and devising new test and certification standards.
Colniță, Alia; Dina, Nicoleta Elena; Leopold, Nicolae; Vodnar, Dan Cristian; Bogdan, Diana; Porav, Sebastian Alin; David, Leontin
2017-09-01
Raman scattering and its particular effect, surface-enhanced Raman scattering (SERS), are whole-organism fingerprinting spectroscopic techniques that gain more and more popularity in bacterial detection. In this work, two relevant Gram-positive bacteria species, Lactobacillus casei ( L. casei ) and Listeria monocytogenes ( L. monocytogenes ) were characterized based on their Raman and SERS spectral fingerprints. The SERS spectra were used to identify the biochemical structures of the bacterial cell wall. Two synthesis methods of the SERS-active nanomaterials were used and the recorded spectra were analyzed. L. casei and L. monocytogenes were successfully discriminated by applying Principal Component Analysis (PCA) to their specific spectral data.
Leopold, Nicolae; Vodnar, Dan Cristian; Bogdan, Diana; Porav, Sebastian Alin; David, Leontin
2017-01-01
Raman scattering and its particular effect, surface-enhanced Raman scattering (SERS), are whole-organism fingerprinting spectroscopic techniques that gain more and more popularity in bacterial detection. In this work, two relevant Gram-positive bacteria species, Lactobacillus casei (L. casei) and Listeria monocytogenes (L. monocytogenes) were characterized based on their Raman and SERS spectral fingerprints. The SERS spectra were used to identify the biochemical structures of the bacterial cell wall. Two synthesis methods of the SERS-active nanomaterials were used and the recorded spectra were analyzed. L. casei and L. monocytogenes were successfully discriminated by applying Principal Component Analysis (PCA) to their specific spectral data. PMID:28862655
Chen, Jian-bo; Sun, Su-qin; Zhou, Qun
2015-07-01
The nondestructive and label-free infrared (IR) spectroscopy is a direct tool to characterize the spatial distribution of organic and inorganic compounds in plant. Since plant samples are usually complex mixtures, signal-resolving methods are necessary to find the spectral features of compounds of interest in the signal-overlapped IR spectra. In this research, two approaches using existing data-driven signal-resolving methods are proposed to interpret the IR spectra of plant samples. If the number of spectra is small, "tri-step identification" can enhance the spectral resolution to separate and identify the overlapped bands. First, the envelope bands of the original spectrum are interpreted according to the spectra-structure correlations. Then the spectrum is differentiated to resolve the underlying peaks in each envelope band. Finally, two-dimensional correlation spectroscopy is used to enhance the spectral resolution further. For a large number of spectra, "tri-step decomposition" can resolve the spectra by multivariate methods to obtain the structural and semi-quantitative information about the chemical components. Principal component analysis is used first to explore the existing signal types without any prior knowledge. Then the spectra are decomposed by self-modeling curve resolution methods to estimate the spectra and contents of significant chemical components. At last, targeted methods such as partial least squares target can explore the content profiles of specific components sensitively. As an example, the macroscopic and microscopic distribution of eugenol and calcium oxalate in the bud of clove is studied.
Low-Dimensional Feature Representation for Instrument Identification
NASA Astrophysics Data System (ADS)
Ihara, Mizuki; Maeda, Shin-Ichi; Ikeda, Kazushi; Ishii, Shin
For monophonic music instrument identification, various feature extraction and selection methods have been proposed. One of the issues toward instrument identification is that the same spectrum is not always observed even in the same instrument due to the difference of the recording condition. Therefore, it is important to find non-redundant instrument-specific features that maintain information essential for high-quality instrument identification to apply them to various instrumental music analyses. For such a dimensionality reduction method, the authors propose the utilization of linear projection methods: local Fisher discriminant analysis (LFDA) and LFDA combined with principal component analysis (PCA). After experimentally clarifying that raw power spectra are actually good for instrument classification, the authors reduced the feature dimensionality by LFDA or by PCA followed by LFDA (PCA-LFDA). The reduced features achieved reasonably high identification performance that was comparable or higher than those by the power spectra and those achieved by other existing studies. These results demonstrated that our LFDA and PCA-LFDA can successfully extract low-dimensional instrument features that maintain the characteristic information of the instruments.
Efficient Stochastic Inversion Using Adjoint Models and Kernel-PCA
DOE Office of Scientific and Technical Information (OSTI.GOV)
Thimmisetty, Charanraj A.; Zhao, Wenju; Chen, Xiao
2017-10-18
Performing stochastic inversion on a computationally expensive forward simulation model with a high-dimensional uncertain parameter space (e.g. a spatial random field) is computationally prohibitive even when gradient information can be computed efficiently. Moreover, the ‘nonlinear’ mapping from parameters to observables generally gives rise to non-Gaussian posteriors even with Gaussian priors, thus hampering the use of efficient inversion algorithms designed for models with Gaussian assumptions. In this paper, we propose a novel Bayesian stochastic inversion methodology, which is characterized by a tight coupling between the gradient-based Langevin Markov Chain Monte Carlo (LMCMC) method and a kernel principal component analysis (KPCA). Thismore » approach addresses the ‘curse-of-dimensionality’ via KPCA to identify a low-dimensional feature space within the high-dimensional and nonlinearly correlated parameter space. In addition, non-Gaussian posterior distributions are estimated via an efficient LMCMC method on the projected low-dimensional feature space. We will demonstrate this computational framework by integrating and adapting our recent data-driven statistics-on-manifolds constructions and reduction-through-projection techniques to a linear elasticity model.« less
Lin, Ying-he; Man, Yi; Qu, Yi-li; Guan, Dong-hua; Lu, Xuan; Wei, Na
2006-01-01
To study the movement of long axis and the distribution of principal stress in the abutment teeth in removable partial denture which is retained by use of conical telescope. An ideal three dimensional finite element model was constructed by using SCT image reconstruction technique, self-programming and ANSYS software. The static loads were applied. The displacement of the long axis and the distribution of the principal stress in the abutment teeth was analyzed. There is no statistic difference of displacenat and stress distribution among different three-dimensional finite element models. Generally, the abutment teeth move along the long axis itself. Similar stress distribution was observed in each three-dimensional finite element model. The maximal principal compressive stress was observed at the distal cervix of the second premolar. The abutment teeth can be well protected by use of conical telescope.
Arbogast, Luke W; Delaglio, Frank; Schiel, John E; Marino, John P
2017-11-07
Two-dimensional (2D) 1 H- 13 C methyl NMR provides a powerful tool to probe the higher order structure (HOS) of monoclonal antibodies (mAbs), since spectra can readily be acquired on intact mAbs at natural isotopic abundance, and small changes in chemical environment and structure give rise to observable changes in corresponding spectra, which can be interpreted at atomic resolution. This makes it possible to apply 2D NMR spectral fingerprinting approaches directly to drug products in order to systematically characterize structure and excipient effects. Systematic collections of NMR spectra are often analyzed in terms of the changes in specifically identified peak positions, as well as changes in peak height and line widths. A complementary approach is to apply principal component analysis (PCA) directly to the matrix of spectral data, correlating spectra according to similarities and differences in their overall shapes, rather than according to parameters of individually identified peaks. This is particularly well-suited for spectra of mAbs, where some of the individual peaks might not be well resolved. Here we demonstrate the performance of the PCA method for discriminating structural variation among systematic sets of 2D NMR fingerprint spectra using the NISTmAb and illustrate how spectral variability identified by PCA may be correlated to structure.
Ntlhokwe, Gaalebalwe; Tredoux, Andreas G J; Górecki, Tadeusz; Edwards, Matthew; Vestner, Jochen; Muller, Magdalena; Erasmus, Lené; Joubert, Elizabeth; Christel Cronje, J; de Villiers, André
2017-07-01
The applicability of comprehensive two-dimensional gas chromatography (GC×GC) using a single-stage thermal modulator was explored for the analysis of honeybush tea (Cyclopia spp.) volatile compounds. Headspace solid phase micro-extraction (HS-SPME) was used in combination with GC×GC separation on a non-polar × polar column set with flame ionisation (FID) detection for the analysis of fermented Cyclopia maculata, Cyclopia subternata and Cyclopia genistoides tea infusions of a single harvest season. Method optimisation entailed evaluation of the effects of several experimental parameters on the performance of the modulator, the choice of columns in both dimensions, as well as the HS-SPME extraction fibre. Eighty-four volatile compounds were identified by co-injection of reference standards. Principal component analysis (PCA) showed clear differentiation between the species based on their volatile profiles. Due to the highly reproducible separations obtained using the single-stage thermal modulator, multivariate data analysis was simplified. The results demonstrate both the complexity of honeybush volatile profiles and the potential of GC×GC separation in combination with suitable data analysis techniques for the investigation of the relationship between sensory properties and volatile composition of these products. The developed method therefore offers a fast and inexpensive methodology for the profiling of honeybush tea volatiles. Graphical abstract Surface plot obtained for the GC×GC-FID analysis of honeybush tea volatiles.
NASA Astrophysics Data System (ADS)
Singh, Dharmendra; Kumar, Harish
Earth observation satellites provide data that covers different portions of the electromagnetic spectrum at different spatial and spectral resolutions. The increasing availability of information products generated from satellite images are extending the ability to understand the patterns and dynamics of the earth resource systems at all scales of inquiry. In which one of the most important application is the generation of land cover classification from satellite images for understanding the actual status of various land cover classes. The prospect for the use of satel-lite images in land cover classification is an extremely promising one. The quality of satellite images available for land-use mapping is improving rapidly by development of advanced sensor technology. Particularly noteworthy in this regard is the improved spatial and spectral reso-lution of the images captured by new satellite sensors like MODIS, ASTER, Landsat 7, and SPOT 5. For the full exploitation of increasingly sophisticated multisource data, fusion tech-niques are being developed. Fused images may enhance the interpretation capabilities. The images used for fusion have different temporal, and spatial resolution. Therefore, the fused image provides a more complete view of the observed objects. It is one of the main aim of image fusion to integrate different data in order to obtain more information that can be de-rived from each of the single sensor data alone. A good example of this is the fusion of images acquired by different sensors having a different spatial resolution and of different spectral res-olution. Researchers are applying the fusion technique since from three decades and propose various useful methods and techniques. The importance of high-quality synthesis of spectral information is well suited and implemented for land cover classification. More recently, an underlying multiresolution analysis employing the discrete wavelet transform has been used in image fusion. It was found that multisensor image fusion is a tradeoff between the spectral information from a low resolution multi-spectral images and the spatial information from a high resolution multi-spectral images. With the wavelet transform based fusion method, it is easy to control this tradeoff. A new transform, the curvelet transform was used in recent years by Starck. A ridgelet transform is applied to square blocks of detail frames of undecimated wavelet decomposition, consequently the curvelet transform is obtained. Since the ridgelet transform possesses basis functions matching directional straight lines therefore, the curvelet transform is capable of representing piecewise linear contours on multiple scales through few significant coefficients. This property leads to a better separation between geometric details and background noise, which may be easily reduced by thresholding curvelet coefficients before they are used for fusion. The Terra and Aqua Moderate Resolution Imaging Spectroradiometer (MODIS) instrument provides high radiometric sensitivity (12 bit) in 36 spectral bands ranging in wavelength from 0.4 m to 14.4 m and also it is freely available. Two bands are imaged at a nominal resolution of 250 m at nadir, with five bands at 500 m, and the remaining 29 bands at 1 km. In this paper, the band 1 of spatial resolution 250 m and bandwidth 620-670 nm, and band 2, of spatial resolution of 250m and bandwidth 842-876 nm is considered as these bands has special features to identify the agriculture and other land covers. In January 2006, the Advanced Land Observing Satellite (ALOS) was successfully launched by the Japan Aerospace Exploration Agency (JAXA). The Phased Arraytype L-band SAR (PALSAR) sensor onboard the satellite acquires SAR imagery at a wavelength of 23.5 cm (frequency 1.27 GHz) with capabilities of multimode and multipolarization observation. PALSAR can operate in several modes: the fine-beam single (FBS) polarization mode (HH), fine-beam dual (FBD) polariza-tion mode (HH/HV or VV/VH), polarimetric (PLR) mode (HH/HV/VH/VV), and ScanSAR (WB) mode (HH/VV) [15]. These makes PALSAR imagery very attractive for spatially and temporally consistent monitoring system. The Overview of Principal Component Analysis is that the most of the information within all the bands can be compressed into a much smaller number of bands with little loss of information. It allows us to extract the low-dimensional subspaces that capture the main linear correlation among the high-dimensional image data. This facilitates viewing the explained variance or signal in the available imagery, allowing both gross and more subtle features in the imagery to be seen. In this paper we have explored the fusion technique for enhancing the land cover classification of low resolution satellite data espe-cially freely available satellite data. For this purpose, we have considered to fuse the PALSAR principal component data with MODIS principal component data. Initially, the MODIS band 1 and band 2 is considered, its principal component is computed. Similarly the PALSAR HH, HV and VV polarized data are considered, and there principal component is also computed. con-sequently, the PALSAR principal component image is fused with MODIS principal component image. The aim of this paper is to analyze the effect of classification accuracy on major type of land cover types like agriculture, water and urban bodies with fusion of PALSAR data to MODIS data. Curvelet transformation has been applied for fusion of these two satellite images and Minimum Distance classification technique has been applied for the resultant fused image. It is qualitatively and visually observed that the overall classification accuracy of MODIS image after fusion is enhanced. This type of fusion technique may be quite helpful in near future to use freely available satellite data to develop monitoring system for different land cover classes on the earth.
Exhausted Parents: Development and Preliminary Validation of the Parental Burnout Inventory
Roskam, Isabelle; Raes, Marie-Emilie; Mikolajczak, Moïra
2017-01-01
Can parents burn out? The aim of this research was to examine the construct validity of the concept of parental burnout and to provide researchers which an instrument to measure it. We conducted two successive questionnaire-based online studies, the first with a community-sample of 379 parents using principal component analyses and the second with a community- sample of 1,723 parents using both principal component analyses and confirmatory factor analyses. We investigated whether the tridimensional structure of the burnout syndrome (i.e., exhaustion, inefficacy, and depersonalization) held in the parental context. We then examined the specificity of parental burnout vis-à-vis professional burnout assessed with the Maslach Burnout Inventory, parental stress assessed with the Parental Stress Questionnaire and depression assessed with the Beck Depression Inventory. The results support the validity of a tri-dimensional burnout syndrome including exhaustion, inefficacy and emotional distancing with, respectively, 53.96 and 55.76% variance explained in study 1 and study 2, and reliability ranging from 0.89 to 0.94. The final version of the Parental Burnout Inventory (PBI) consists of 22 items and displays strong psychometric properties (CFI = 0.95, RMSEA = 0.06). Low to moderate correlations between parental burnout and professional burnout, parental stress and depression suggests that parental burnout is not just burnout, stress or depression. The prevalence of parental burnout confirms that some parents are so exhausted that the term “burnout” is appropriate. The proportion of burnout parents lies somewhere between 2 and 12%. The results are discussed in light of their implications at the micro-, meso- and macro-levels. PMID:28232811
Altitude-temporal behaviour of atmospheric ozone, temperature and wind velocity observed at Svalbard
NASA Astrophysics Data System (ADS)
Petkov, Boyan H.; Vitale, Vito; Svendby, Tove M.; Hansen, Georg H.; Sobolewski, Piotr S.; Láska, Kamil; Elster, Josef; Pavlova, Kseniya; Viola, Angelo; Mazzola, Mauro; Lupi, Angelo; Solomatnikova, Anna
2018-07-01
The vertical features of the variations in the atmospheric ozone density, temperature and wind velocity observed at Ny-Ålesund, Svalbard were studied by applying the principal component analysis to the ozonesounding data collected during the 1992-2016 period. Two data sets corresponding to intra-seasonal (IS) variations, which are composed by harmonics with lower than 1 year periods and inter-annual (IA) variations, characterised by larger periods, were extracted and analysed separately. The IS variations in all the three parameters were found to be composed mainly by harmonics typical for the Madden-Julian Oscillation (from 30- to 60-day periods) and, while the first four principal components (PCs) associated with the temperature and wind contributed about 90% to the IS variations, the ozone IS oscillations appeared to be a higher dimensional object for which the first 15 PCs presented almost the same extent of contribution. The IA variations in the three parameters were consisted of harmonics that correspond to widely registered over the globe Quasi-Biennial, El Niño-Southern, North Atlantic and Arctic Oscillations respectively, and the IA variations turned out to be negligible below the tropopause that characterises the Svalbard troposphere as comparatively closed system with respect to the long-period global variations. The behaviour of the first and second PCs associated with IS ozone variations in the time of particular events, like the strong ozone depletion over Arctic in the spring 2011 and solar eclipses was discussed and the changes in the amplitude-frequency features of these PCs were assumed as signs of the atmosphere response to the considered phenomena.
NASA Astrophysics Data System (ADS)
Hoell, Simon; Omenzetter, Piotr
2015-03-01
The development of large wind turbines that enable to harvest energy more efficiently is a consequence of the increasing demand for renewables in the world. To optimize the potential energy output, light and flexible wind turbine blades (WTBs) are designed. However, the higher flexibilities and lower buckling capacities adversely affect the long-term safety and reliability of WTBs, and thus the increased operation and maintenance costs reduce the expected revenue. Effective structural health monitoring techniques can help to counteract this by limiting inspection efforts and avoiding unplanned maintenance actions. Vibration-based methods deserve high attention due to the moderate instrumentation efforts and the applicability for in-service measurements. The present paper proposes the use of cross-correlations (CCs) of acceleration responses between sensors at different locations for structural damage detection in WTBs. CCs were in the past successfully applied for damage detection in numerical and experimental beam structures while utilizing only single lags between the signals. The present approach uses vectors of CC coefficients for multiple lags between measurements of two selected sensors taken from multiple possible combinations of sensors. To reduce the dimensionality of the damage sensitive feature (DSF) vectors, principal component analysis is performed. The optimal number of principal components (PCs) is chosen with respect to a statistical threshold. Finally, the detection phase uses the selected PCs of the healthy structure to calculate scores from a current DSF vector, where statistical hypothesis testing is performed for making a decision about the current structural state. The method is applied to laboratory experiments conducted on a small WTB with non-destructive damage scenarios.
Pérez-Mora, Walter; Jorrin-Novo, Jesús V; Melgarejo, Luz Marina
2018-02-01
Substantial equivalence studies were performed in three Theobroma spp., cacao, bicolor and grandiflorum through chemical composition analysis and protein profiling of fruit (pulp juice and seeds). Principal component analysis of sugar, organic acid, and phenol content in pulp juice revealed equivalence among the three species, with differences in some of the compounds that may result in different organoleptic properties. Proteins were extracted from seeds and pulp juice, resolved by two dimensional electrophoresis and major spots subjected to mass spectrometry analysis and identification. The protein profile, as revealed by principal component analysis, was variable among the three species in both seed and pulp, with qualitative and quantitative differences in some of protein species. The functional grouping of the identified proteins correlated with the biological role of each organ. Some of the identified proteins are of interest, being minimally discussed, including vicilin, a protease inhibitor, and a flavonol synthase/flavanone 3-hydroxylase. Theobroma grandiflorum and Theobroma bicolor are endemic Amazonian plants that are poorly traded at the local level. As close relatives of Theobroma cacao, they may provide a good alternative for human consumption and industrial purposes. In this regard, we performed equivalence studies by conducting a comparative biochemical and proteomics analysis of the fruit, pulp juice and seeds of these three species. The results indicated equivalent chemical compositions and variable protein profiles with some differences in the content of the specific compounds or protein species that may result in variable organoleptic properties between the species and can be exploited for traceability purposes. Copyright © 2017 Elsevier Ltd. All rights reserved.
Paschou, Peristera
2010-01-01
Recent large-scale studies of European populations have demonstrated the existence of population genetic structure within Europe and the potential to accurately infer individual ancestry when information from hundreds of thousands of genetic markers is used. In fact, when genomewide genetic variation of European populations is projected down to a two-dimensional Principal Components Analysis plot, a surprising correlation with actual geographic coordinates of self-reported ancestry has been reported. This substructure can hamper the search of susceptibility genes for common complex disorders leading to spurious correlations. The identification of genetic markers that can correct for population stratification becomes therefore of paramount importance. Analyzing 1,200 individuals from 11 populations genotyped for more than 500,000 SNPs (Population Reference Sample), we present a systematic exploration of the extent to which geographic coordinates of origin within Europe can be predicted, with small panels of SNPs. Markers are selected to correlate with the top principal components of the dataset, as we have previously demonstrated. Performing thorough cross-validation experiments we show that it is indeed possible to predict individual ancestry within Europe down to a few hundred kilometers from actual individual origin, using information from carefully selected panels of 500 or 1,000 SNPs. Furthermore, we show that these panels can be used to correctly assign the HapMap Phase 3 European populations to their geographic origin. The SNPs that we propose can prove extremely useful in a variety of different settings, such as stratification correction or genetic ancestry testing, and the study of the history of European populations. PMID:20805874
DOE Office of Scientific and Technical Information (OSTI.GOV)
Wiengarten, T.; Fichtner, H.; Kleimann, J.
2016-12-10
We extend a two-component model for the evolution of fluctuations in the solar wind plasma so that it is fully three-dimensional (3D) and also coupled self-consistently to the large-scale magnetohydrodynamic equations describing the background solar wind. The two classes of fluctuations considered are a high-frequency parallel-propagating wave-like piece and a low-frequency quasi-two-dimensional component. For both components, the nonlinear dynamics is dominanted by quasi-perpendicular spectral cascades of energy. Driving of the fluctuations by, for example, velocity shear and pickup ions is included. Numerical solutions to the new model are obtained using the Cronos framework, and validated against previous simpler models. Comparing results frommore » the new model with spacecraft measurements, we find improved agreement relative to earlier models that employ prescribed background solar wind fields. Finally, the new results for the wave-like and quasi-two-dimensional fluctuations are used to calculate ab initio diffusion mean-free paths and drift lengthscales for the transport of cosmic rays in the turbulent solar wind.« less
Corilo, Yuri E; Podgorski, David C; McKenna, Amy M; Lemkau, Karin L; Reddy, Christopher M; Marshall, Alan G; Rodgers, Ryan P
2013-10-01
One fundamental challenge with either acute or chronic oil spills is to identify the source, especially in highly polluted areas, near natural oil seeps, when the source contains more than one petroleum product or when extensive weathering has occurred. Here we focus on heavy fuel oil that spilled (~200,000 L) from two suspected fuel tanks that were ruptured on the motor vessel (M/V) Cosco Busan when it struck the San Francisco-Oakland Bay Bridge in November 2007. We highlight the utility of principal component analysis (PCA) of elemental composition data obtained by high resolution FT-ICR mass spectrometry to correctly identify the source of environmental contamination caused by the unintended release of heavy fuel oil (HFO). Using ultrahigh resolution electrospray ionization (ESI) Fourier transform ion cyclotron resonance mass spectrometry, we uniquely assigned thousands of elemental compositions of heteroatom-containing species in neat samples from both tanks and then applied principal component analysis. The components were based on double bond equivalents for constituents of elemental composition, CcHhN1S1. To determine if the fidelity of our source identification was affected by weathering, field samples were collected at various intervals up to two years after the spill. We are able to identify a suite of polar petroleum markers that are environmentally persistent, enabling us to confidently identify that only one tank was the source of the spilled oil: in fact, a single principal component could account for 98% of the variance. Although identification is unaffected by the presence of higher polarity, petrogenic oxidation (weathering) products, future studies may require removal of such species by anion exchange chromatography prior to mass spectral analysis due to their preferential ionization by ESI.
Kalgin, Igor V; Caflisch, Amedeo; Chekmarev, Sergei F; Karplus, Martin
2013-05-23
A new analysis of the 20 μs equilibrium folding/unfolding molecular dynamics simulations of the three-stranded antiparallel β-sheet miniprotein (beta3s) in implicit solvent is presented. The conformation space is reduced in dimensionality by introduction of linear combinations of hydrogen bond distances as the collective variables making use of a specially adapted principal component analysis (PCA); i.e., to make structured conformations more pronounced, only the formed bonds are included in determining the principal components. It is shown that a three-dimensional (3D) subspace gives a meaningful representation of the folding behavior. The first component, to which eight native hydrogen bonds make the major contribution (four in each beta hairpin), is found to play the role of the reaction coordinate for the overall folding process, while the second and third components distinguish the structured conformations. The representative points of the trajectory in the 3D space are grouped into conformational clusters that correspond to locally stable conformations of beta3s identified in earlier work. A simplified kinetic network based on the three components is constructed, and it is complemented by a hydrodynamic analysis. The latter, making use of "passive tracers" in 3D space, indicates that the folding flow is much more complex than suggested by the kinetic network. A 2D representation of streamlines shows there are vortices which correspond to repeated local rearrangement, not only around minima of the free energy surface but also in flat regions between minima. The vortices revealed by the hydrodynamic analysis are apparently not evident in folding pathways generated by transition-path sampling. Making use of the fact that the values of the collective hydrogen bond variables are linearly related to the Cartesian coordinate space, the RMSD between clusters is determined. Interestingly, the transition rates show an approximate exponential correlation with distance in the hydrogen bond subspace. Comparison with the many published studies shows good agreement with the present analysis for the parts that can be compared, supporting the robust character of our understanding of this "hydrogen atom" of protein folding.
Structural aspects of face recognition and the other-race effect.
O'Toole, A J; Deffenbacher, K A; Valentin, D; Abdi, H
1994-03-01
The other-race effect was examined in a series of experiments and simulations that looked at the relationships among observer ratings of typicality, familiarity, attractiveness, memorability, and the performance variables of d' and criterion. Experiment 1 replicated the other-race effect with our Caucasian and Japanese stimuli for both Caucasian and Asian observers. In Experiment 2, we collected ratings from Caucasian observers on the faces used in the recognition task. A Varimax-rotated principal components analysis on the rating and performance data for the Caucasian faces replicated Vokey and Read's (1992) finding that typicality is composed of two orthogonal components, dissociable via their independent relationships to: (1) attractiveness and familiarity ratings and (2) memorability ratings. For Japanese faces, however, we found that typicality was related only to memorability. Where performance measures were concerned, two additional principal components dominated by criterion and by d' emerged for Caucasian faces. For the Japanese faces, however, the performance measures of d' and criterion merged into a single component that represented a second component of typicality, one orthogonal to the memorability-dominated component. A measure of face representation quality extracted from an autoassociative neural network trained with a majority of Caucasian faces and a minority of Japanese faces was incorporated into the principal components analysis. For both Caucasian and Japanese faces, the neural network measure related both to memorability ratings and to human accuracy measures. Combined, the human data and simulation results indicate that the memorability component of typicality may be related to small, local, distinctive features, whereas the attractiveness/familiarity component may be more related to the global, shape-based properties of the face.
Jia, Youmei; Cai, Jianfeng; Xin, Huaxia; Feng, Jiatao; Fu, Yanhui; Fu, Qing; Jin, Yu
2017-06-08
A preparative two dimensional hydrophilic interaction liquid chromatography/reversed-phase liquid chromatography (Pre-2D-HILIC/RPLC) method was established to separate and purify the components in Trachelospermum jasminoides . The pigments and strongly polar components were removed from the crude extract after the active carbon decolorization and solid phase extraction processes. A Click XIon column (250 mm×20 mm, 10 μm) was selected as stationary phase and water-acetonitrile as mobile phases in the first dimensional HILIC. Finally, 15 fractions were collected under UV-triggered mode. In the second dimensional RPLC, a C18 column (250 mm×20 mm, 5 μm) was selected and water-acetonitrile was used as mobile phases. As a result, 14 compounds with high purity were obtained, which were further identified by mass spectrometry (MS) and nuclear magnetic resonance (NMR). Finally, 11 lignan compounds and three flavonoid compounds were obtained. The method has a good orthogonality, and can improve the resolution and the peak capacity. It is significant for the separation of complex components from Trachelospermum jasminoides .
Measurement of 13C chemical shift tensor principal values with a magic-angle turning experiment.
Hu, J Z; Orendt, A M; Alderman, D W; Pugmire, R J; Ye, C; Grant, D M
1994-08-01
The magic-angle turning (MAT) experiment introduced by Gan is developed into a powerful and routine method for measuring the principal values of 13C chemical shift tensors in powdered solids. A large-volume MAT probe with stable rotation frequencies down to 22 Hz is described. A triple-echo MAT pulse sequence is introduced to improve the quality of the two-dimensional baseplane. It is shown that measurements of the principal values of chemical shift tensors in complex compounds can be enhanced by using either short contact times or dipolar dephasing pulse sequences to isolate the powder patterns from protonated or non-protonated carbons, respectively. A model compound, 1,2,3-trimethoxybenzene, is used to demonstrate these techniques, and the 13C principal values in 2,3-dimethylnaphthalene and Pocahontas coal are reported as typical examples.
Dong, Jianghu J; Wang, Liangliang; Gill, Jagbir; Cao, Jiguo
2017-01-01
This article is motivated by some longitudinal clinical data of kidney transplant recipients, where kidney function progression is recorded as the estimated glomerular filtration rates at multiple time points post kidney transplantation. We propose to use the functional principal component analysis method to explore the major source of variations of glomerular filtration rate curves. We find that the estimated functional principal component scores can be used to cluster glomerular filtration rate curves. Ordering functional principal component scores can detect abnormal glomerular filtration rate curves. Finally, functional principal component analysis can effectively estimate missing glomerular filtration rate values and predict future glomerular filtration rate values.
Comparative multivariate analysis of biometric traits of West African Dwarf and Red Sokoto goats.
Yakubu, Abdulmojeed; Salako, Adebowale E; Imumorin, Ikhide G
2011-03-01
The population structure of 302 randomly selected West African Dwarf (WAD) and Red Sokoto (RS) goats was examined using multivariate morphometric analyses. This was to make the case for conservation, rational management and genetic improvement of these two most important Nigerian goat breeds. Fifteen morphometric measurements were made on each individual animal. RS goats were superior (P<0.05) to the WAD for the body size and skeletal proportions investigated. The phenotypic variability between the two breeds was revealed by their mutual responses in the principal components. While four principal components were extracted for WAD goats, three components were obtained for their RS counterparts with variation in the loading traits of each component for each breed. The Mahalanobis distance of 72.28 indicated a high degree of spatial racial separation in morphology between the genotypes. The Ward's option of the cluster analysis consolidated the morphometric distinctness of the two breeds. Application of selective breeding to genetic improvement would benefit from the detected phenotypic differentiation. Other implications for management and conservation of the goats are highlighted.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Shalchi, A., E-mail: andreasm4@yahoo.com
2016-10-20
We explore the transport of energetic particles in two-component turbulence in which the stochastic magnetic field is assumed to be a superposition of slab and two-dimensional modes. It is known that in magnetostatic slab turbulence, the motion of particles across the mean magnetic field is subdiffusive. If a two-dimensional component is added, diffusion is recovered. It was also shown before that in two-component turbulence, the slab modes do not explicitly contribute to the perpendicular diffusion coefficient. In the current paper, the implicit contribution of slab modes is explored and it is shown that this contribution leads to a reduction ofmore » the perpendicular diffusion coefficient. This effect improves the agreement between simulations and analytical theory. Furthermore, the obtained results are relevant for investigations of diffusive shock acceleration.« less
ERIC Educational Resources Information Center
Roberts, Maria Banda; Hernandez, Rosalinda
2012-01-01
This article explains two components in the review process of a university's principal preparation program. A superintendents' focus group session and an analysis of other universities' program profiles revealed a need to eliminate dated courses, include certification in the degree requirements, update the program curriculum with best practice,…
GAC: Gene Associations with Clinical, a web based application.
Zhang, Xinyan; Rupji, Manali; Kowalski, Jeanne
2017-01-01
We present GAC, a shiny R based tool for interactive visualization of clinical associations based on high-dimensional data. The tool provides a web-based suite to perform supervised principal component analysis (SuperPC), an approach that uses both high-dimensional data, such as gene expression, combined with clinical data to infer clinical associations. We extended the approach to address binary outcomes, in addition to continuous and time-to-event data in our package, thereby increasing the use and flexibility of SuperPC. Additionally, the tool provides an interactive visualization for summarizing results based on a forest plot for both binary and time-to-event data. In summary, the GAC suite of tools provide a one stop shop for conducting statistical analysis to identify and visualize the association between a clinical outcome of interest and high-dimensional data types, such as genomic data. Our GAC package has been implemented in R and is available via http://shinygispa.winship.emory.edu/GAC/. The developmental repository is available at https://github.com/manalirupji/GAC.
Wavelet decomposition based principal component analysis for face recognition using MATLAB
NASA Astrophysics Data System (ADS)
Sharma, Mahesh Kumar; Sharma, Shashikant; Leeprechanon, Nopbhorn; Ranjan, Aashish
2016-03-01
For the realization of face recognition systems in the static as well as in the real time frame, algorithms such as principal component analysis, independent component analysis, linear discriminate analysis, neural networks and genetic algorithms are used for decades. This paper discusses an approach which is a wavelet decomposition based principal component analysis for face recognition. Principal component analysis is chosen over other algorithms due to its relative simplicity, efficiency, and robustness features. The term face recognition stands for identifying a person from his facial gestures and having resemblance with factor analysis in some sense, i.e. extraction of the principal component of an image. Principal component analysis is subjected to some drawbacks, mainly the poor discriminatory power and the large computational load in finding eigenvectors, in particular. These drawbacks can be greatly reduced by combining both wavelet transform decomposition for feature extraction and principal component analysis for pattern representation and classification together, by analyzing the facial gestures into space and time domain, where, frequency and time are used interchangeably. From the experimental results, it is envisaged that this face recognition method has made a significant percentage improvement in recognition rate as well as having a better computational efficiency.
Sei, Ryosuke; Kitani, Suguru; Fukumura, Tomoteru; Kawaji, Hitoshi; Hasegawa, Tetsuya
2016-09-07
Discovery of layered superconductors such as cuprates and iron-based compounds has unveiled new science and compounds. In these superconductors, quasi-two-dimensional layers including transition metal cations play principal role in the superconductivity via carrier doping by means of aliovalent-ion substitution. Here, we report on a two-dimensional superconductivity at 2 K in ThCr2Si2-type layered oxide Y2O2Bi possessing conducting monatomic Bi(2-) square net, possibly associated with an exotic superconductivity. The superconductivity emerges only in excessively oxygen-incorporated Y2O2Bi with expanded inter-net distance, in stark contrast to nonsuperconducting pristine Y2O2Bi reported previously. This result suggests that the element incorporation into hidden interstitial site could be an alternative approach to conventional substitution and intercalation methods for search of novel superconductors.
The Relation between Factor Score Estimates, Image Scores, and Principal Component Scores
ERIC Educational Resources Information Center
Velicer, Wayne F.
1976-01-01
Investigates the relation between factor score estimates, principal component scores, and image scores. The three methods compared are maximum likelihood factor analysis, principal component analysis, and a variant of rescaled image analysis. (RC)
DOE Office of Scientific and Technical Information (OSTI.GOV)
Holden, H.; LeDrew, E.
1997-06-01
Remote discrimination of substrate types in relatively shallow coastal waters has been limited by the spatial and spectral resolution of available sensors. An additional limiting factor is the strong attenuating influence of the water column over the substrate. As a result, there have been limited attempts to map submerged ecosystems such as coral reefs based on spectral characteristics. Both healthy and bleached corals were measured at depth with a hand-held spectroradiometer, and their spectra compared. Two separate principal components analyses (PCA) were performed on two sets of spectral data. The PCA revealed that there is indeed a spectral difference basedmore » on health. In the first data set, the first component (healthy coral) explains 46.82%, while the second component (bleached coral) explains 46.35% of the variance. In the second data set, the first component (bleached coral) explained 46.99%; the second component (healthy coral) explained 36.55%; and the third component (healthy coral) explained 15.44 % of the total variance in the original data. These results are encouraging with respect to using an airborne spectroradiometer to identify areas of bleached corals thus enabling accurate monitoring over time.« less
The Butterflies of Principal Components: A Case of Ultrafine-Grained Polyphase Units
NASA Astrophysics Data System (ADS)
Rietmeijer, F. J. M.
1996-03-01
Dusts in the accretion regions of chondritic interplanetary dust particles [IDPs] consisted of three principal components: carbonaceous units [CUs], carbon-bearing chondritic units [GUs] and carbon-free silicate units [PUs]. Among others, differences among chondritic IDP morphologies and variable bulk C/Si ratios reflect variable mixtures of principal components. The spherical shapes of the initially amorphous principal components remain visible in many chondritic porous IDPs but fusion was documented for CUs, GUs and PUs. The PUs occur as coarse- and ultrafine-grained units that include so called GEMS. Spherical principal components preserved in an IDP as recognisable textural units have unique proporties with important implications for their petrological evolution from pre-accretion processing to protoplanet alteration and dynamic pyrometamorphism. Throughout their lifetime the units behaved as closed-systems without chemical exchange with other units. This behaviour is reflected in their mineralogies while the bulk compositions of principal components define the environments wherein they were formed.
Climate and atmospheric modeling studies
NASA Technical Reports Server (NTRS)
1992-01-01
The climate and atmosphere modeling research programs have concentrated on the development of appropriate atmospheric and upper ocean models, and preliminary applications of these models. Principal models are a one-dimensional radiative-convective model, a three-dimensional global model, and an upper ocean model. Principal applications were the study of the impact of CO2, aerosols, and the solar 'constant' on climate.
2011-01-01
Background Hemorrhagic fever with renal syndrome (HFRS) is an important infectious disease caused by different species of hantaviruses. As a rodent-borne disease with a seasonal distribution, external environmental factors including climate factors may play a significant role in its transmission. The city of Shenyang is one of the most seriously endemic areas for HFRS. Here, we characterized the dynamic temporal trend of HFRS, and identified climate-related risk factors and their roles in HFRS transmission in Shenyang, China. Methods The annual and monthly cumulative numbers of HFRS cases from 2004 to 2009 were calculated and plotted to show the annual and seasonal fluctuation in Shenyang. Cross-correlation and autocorrelation analyses were performed to detect the lagged effect of climate factors on HFRS transmission and the autocorrelation of monthly HFRS cases. Principal component analysis was constructed by using climate data from 2004 to 2009 to extract principal components of climate factors to reduce co-linearity. The extracted principal components and autocorrelation terms of monthly HFRS cases were added into a multiple regression model called principal components regression model (PCR) to quantify the relationship between climate factors, autocorrelation terms and transmission of HFRS. The PCR model was compared to a general multiple regression model conducted only with climate factors as independent variables. Results A distinctly declining temporal trend of annual HFRS incidence was identified. HFRS cases were reported every month, and the two peak periods occurred in spring (March to May) and winter (November to January), during which, nearly 75% of the HFRS cases were reported. Three principal components were extracted with a cumulative contribution rate of 86.06%. Component 1 represented MinRH0, MT1, RH1, and MWV1; component 2 represented RH2, MaxT3, and MAP3; and component 3 represented MaxT2, MAP2, and MWV2. The PCR model was composed of three principal components and two autocorrelation terms. The association between HFRS epidemics and climate factors was better explained in the PCR model (F = 446.452, P < 0.001, adjusted R2 = 0.75) than in the general multiple regression model (F = 223.670, P < 0.000, adjusted R2 = 0.51). Conclusion The temporal distribution of HFRS in Shenyang varied in different years with a distinctly declining trend. The monthly trends of HFRS were significantly associated with local temperature, relative humidity, precipitation, air pressure, and wind velocity of the different previous months. The model conducted in this study will make HFRS surveillance simpler and the control of HFRS more targeted in Shenyang. PMID:22133347
Fabrication of micromechanical and microoptical systems by two-photon polymerization
NASA Astrophysics Data System (ADS)
Reinhardt, Carsten; Ovsianikov, A.; Passinger, Sven; Chichkov, Boris N.
2007-01-01
The recently developed two-photon polymerisation technique is used for the fabrication of two- and three-dimensional structures in photosensitive inorganic-organic hybrid material (ORMOCER), in SU8 , and in positive tone resist with resolutions down to 100nm. In this contribution we present applications of this powerful technology for the realization of micromechanical systems and microoptical components. We will demonstrate results on the fabrication of complex movable three-dimensional micromechanical systems and microfluidic components which cannot be realized by other technologies. This approach of structuring photosensitive materials also provides unique possibilities for the fabrication of different microoptical components such as arbitrary shaped microlenses, microprisms, and 3D-photonic crystals with high optical quality.
High-Dimensional Intrinsic Interpolation Using Gaussian Process Regression and Diffusion Maps
Thimmisetty, Charanraj A.; Ghanem, Roger G.; White, Joshua A.; ...
2017-10-10
This article considers the challenging task of estimating geologic properties of interest using a suite of proxy measurements. The current work recast this task as a manifold learning problem. In this process, this article introduces a novel regression procedure for intrinsic variables constrained onto a manifold embedded in an ambient space. The procedure is meant to sharpen high-dimensional interpolation by inferring non-linear correlations from the data being interpolated. The proposed approach augments manifold learning procedures with a Gaussian process regression. It first identifies, using diffusion maps, a low-dimensional manifold embedded in an ambient high-dimensional space associated with the data. Itmore » relies on the diffusion distance associated with this construction to define a distance function with which the data model is equipped. This distance metric function is then used to compute the correlation structure of a Gaussian process that describes the statistical dependence of quantities of interest in the high-dimensional ambient space. The proposed method is applicable to arbitrarily high-dimensional data sets. Here, it is applied to subsurface characterization using a suite of well log measurements. The predictions obtained in original, principal component, and diffusion space are compared using both qualitative and quantitative metrics. Considerable improvement in the prediction of the geological structural properties is observed with the proposed method.« less
High-Dimensional Intrinsic Interpolation Using Gaussian Process Regression and Diffusion Maps
DOE Office of Scientific and Technical Information (OSTI.GOV)
Thimmisetty, Charanraj A.; Ghanem, Roger G.; White, Joshua A.
This article considers the challenging task of estimating geologic properties of interest using a suite of proxy measurements. The current work recast this task as a manifold learning problem. In this process, this article introduces a novel regression procedure for intrinsic variables constrained onto a manifold embedded in an ambient space. The procedure is meant to sharpen high-dimensional interpolation by inferring non-linear correlations from the data being interpolated. The proposed approach augments manifold learning procedures with a Gaussian process regression. It first identifies, using diffusion maps, a low-dimensional manifold embedded in an ambient high-dimensional space associated with the data. Itmore » relies on the diffusion distance associated with this construction to define a distance function with which the data model is equipped. This distance metric function is then used to compute the correlation structure of a Gaussian process that describes the statistical dependence of quantities of interest in the high-dimensional ambient space. The proposed method is applicable to arbitrarily high-dimensional data sets. Here, it is applied to subsurface characterization using a suite of well log measurements. The predictions obtained in original, principal component, and diffusion space are compared using both qualitative and quantitative metrics. Considerable improvement in the prediction of the geological structural properties is observed with the proposed method.« less
Development of an Input Suite for an Orthotropic Composite Material Model
NASA Technical Reports Server (NTRS)
Hoffarth, Canio; Shyamsunder, Loukham; Khaled, Bilal; Rajan, Subramaniam; Goldberg, Robert K.; Carney, Kelly S.; Dubois, Paul; Blankenhorn, Gunther
2017-01-01
An orthotropic three-dimensional material model suitable for use in modeling impact tests has been developed that has three major components elastic and inelastic deformations, damage and failure. The material model has been implemented as MAT213 into a special version of LS-DYNA and uses tabulated data obtained from experiments. The prominent features of the constitutive model are illustrated using a widely-used aerospace composite the T800S3900-2B[P2352W-19] BMS8-276 Rev-H-Unitape fiber resin unidirectional composite. The input for the deformation model consists of experimental data from 12 distinct experiments at a known temperature and strain rate: tension and compression along all three principal directions, shear in all three principal planes, and off axis tension or compression tests in all three principal planes, along with other material constants. There are additional input associated with the damage and failure models. The steps in using this model are illustrated composite characterization tests, verification tests and a validation test. The results show that the developed and implemented model is stable and yields acceptably accurate results.
Sekiguchi, Yuki; Oroguchi, Tomotaka; Nakasako, Masayoshi
2016-01-01
Coherent X-ray diffraction imaging (CXDI) is one of the techniques used to visualize structures of non-crystalline particles of micrometer to submicrometer size from materials and biological science. In the structural analysis of CXDI, the electron density map of a sample particle can theoretically be reconstructed from a diffraction pattern by using phase-retrieval (PR) algorithms. However, in practice, the reconstruction is difficult because diffraction patterns are affected by Poisson noise and miss data in small-angle regions due to the beam stop and the saturation of detector pixels. In contrast to X-ray protein crystallography, in which the phases of diffracted waves are experimentally estimated, phase retrieval in CXDI relies entirely on the computational procedure driven by the PR algorithms. Thus, objective criteria and methods to assess the accuracy of retrieved electron density maps are necessary in addition to conventional parameters monitoring the convergence of PR calculations. Here, a data analysis scheme, named ASURA, is proposed which selects the most probable electron density maps from a set of maps retrieved from 1000 different random seeds for a diffraction pattern. Each electron density map composed of J pixels is expressed as a point in a J-dimensional space. Principal component analysis is applied to describe characteristics in the distribution of the maps in the J-dimensional space. When the distribution is characterized by a small number of principal components, the distribution is classified using the k-means clustering method. The classified maps are evaluated by several parameters to assess the quality of the maps. Using the proposed scheme, structure analysis of a diffraction pattern from a non-crystalline particle is conducted in two stages: estimation of the overall shape and determination of the fine structure inside the support shape. In each stage, the most accurate and probable density maps are objectively selected. The validity of the proposed scheme is examined by application to diffraction data that were obtained from an aggregate of metal particles and a biological specimen at the XFEL facility SACLA using custom-made diffraction apparatus.
Fleming, Brandon J.; LaMotte, Andrew E.; Sekellick, Andrew J.
2013-01-01
Hydrogeologic regions in the fractured rock area of Maryland were classified using geographic information system tools with principal components and cluster analyses. A study area consisting of the 8-digit Hydrologic Unit Code (HUC) watersheds with rivers that flow through the fractured rock area of Maryland and bounded by the Fall Line was further subdivided into 21,431 catchments from the National Hydrography Dataset Plus. The catchments were then used as a common hydrologic unit to compile relevant climatic, topographic, and geologic variables. A principal components analysis was performed on 10 input variables, and 4 principal components that accounted for 83 percent of the variability in the original data were identified. A subsequent cluster analysis grouped the catchments based on four principal component scores into six hydrogeologic regions. Two crystalline rock hydrogeologic regions, including large parts of the Washington, D.C. and Baltimore metropolitan regions that represent over 50 percent of the fractured rock area of Maryland, are distinguished by differences in recharge, Precipitation minus Potential Evapotranspiration, sand content in soils, and groundwater contributions to streams. This classification system will provide a georeferenced digital hydrogeologic framework for future investigations of groundwater availability in the fractured rock area of Maryland.
Relationship between regional population and healthcare delivery in Japan.
Niga, Takeo; Mori, Maiko; Kawahara, Kazuo
2016-01-01
In order to address regional inequality in healthcare delivery in Japan, healthcare districts were established in 1985. However, regional healthcare delivery has now become a national issue because of population migration and the aging population. In this study, the state of healthcare delivery at the district level is examined by analyzing population, the number of physicians, and the number of hospital beds. The results indicate a continuing disparity in healthcare delivery among districts. We find that the rate of change in population has a strong positive correlation with that in the number of physicians and a weak positive correlation with that in the number of hospital beds. In addition, principal component analysis is performed on three variables: the rate of change in population, the number of physicians per capita, and the number of hospital beds per capita. This analysis suggests that the two principal components contribute 90.1% of the information. The first principal component is thought to show the effect of the regulations on hospital beds. The second principal component is thought to show the capacity to recruit physicians. This study indicates that an adjustment to the regulations on hospital beds as well as physician allocation by public funds may be key to resolving the impending issue of regionally disproportionate healthcare delivery.
Performance evaluation of PCA-based spike sorting algorithms.
Adamos, Dimitrios A; Kosmidis, Efstratios K; Theophilidis, George
2008-09-01
Deciphering the electrical activity of individual neurons from multi-unit noisy recordings is critical for understanding complex neural systems. A widely used spike sorting algorithm is being evaluated for single-electrode nerve trunk recordings. The algorithm is based on principal component analysis (PCA) for spike feature extraction. In the neuroscience literature it is generally assumed that the use of the first two or most commonly three principal components is sufficient. We estimate the optimum PCA-based feature space by evaluating the algorithm's performance on simulated series of action potentials. A number of modifications are made to the open source nev2lkit software to enable systematic investigation of the parameter space. We introduce a new metric to define clustering error considering over-clustering more favorable than under-clustering as proposed by experimentalists for our data. Both the program patch and the metric are available online. Correlated and white Gaussian noise processes are superimposed to account for biological and artificial jitter in the recordings. We report that the employment of more than three principal components is in general beneficial for all noise cases considered. Finally, we apply our results to experimental data and verify that the sorting process with four principal components is in agreement with a panel of electrophysiology experts.
Fluorescence fingerprint as an instrumental assessment of the sensory quality of tomato juices.
Trivittayasil, Vipavee; Tsuta, Mizuki; Imamura, Yoshinori; Sato, Tsuneo; Otagiri, Yuji; Obata, Akio; Otomo, Hiroe; Kokawa, Mito; Sugiyama, Junichi; Fujita, Kaori; Yoshimura, Masatoshi
2016-03-15
Sensory analysis is an important standard for evaluating food products. However, as trained panelists and time are required for the process, the potential of using fluorescence fingerprint as a rapid instrumental method to approximate sensory characteristics was explored in this study. Thirty-five out of 44 descriptive sensory attributes were found to show a significant difference between samples (analysis of variance test). Principal component analysis revealed that principal component 1 could capture 73.84 and 75.28% variance for aroma category and combined flavor and taste category respectively. Fluorescence fingerprints of tomato juices consisted of two visible peaks at excitation/emission wavelengths of 290/350 and 315/425 nm and a long narrow emission peak at 680 nm. The 680 nm peak was only clearly observed in juices obtained from tomatoes cultivated to be eaten raw. The ability to predict overall sensory profiles was investigated by using principal component 1 as a regression target. Fluorescence fingerprint could predict principal component 1 of both aroma and combined flavor and taste with a coefficient of determination above 0.8. The results obtained in this study indicate the potential of using fluorescence fingerprint as an instrumental method for assessing sensory characteristics of tomato juices. © 2015 Society of Chemical Industry.
Photoelastic Analysis of Three-dimensional Stress Systems Using Scattered Light
NASA Technical Reports Server (NTRS)
Weller, R; Bussey, J K
1939-01-01
A method has been developed for making photoelastic analyses of three-dimensional stress systems by utilizing the polarization phenomena associated with the scattering of light. By this method, the maximum shear and the directions of the three principal stresses at any point within a model can be determined, and the two principal stresses at a free-bounding surface can be separately evaluated. Polarized light is projected into the model through a slit so that it illuminates a plane section. The light is continuously analyzed along its path by scattering and the state of stress in the illuminated section is obtained. By means of a series of such sections, the entire stress field may be explored. The method was used to analyze the stress system of a simple beam in bending. The results were found to be in good agreement with those expected from elementary theory.
Lubaba, Caesar H; Hidano, Arata; Welburn, Susan C; Revie, Crawford W; Eisler, Mark C
2015-01-01
Two-dimensional motion sensors use electronic accelerometers to record the lying, standing and walking activity of cattle. Movement behaviour data collected automatically using these sensors over prolonged periods of time could be of use to stakeholders making management and disease control decisions in rural sub-Saharan Africa leading to potential improvements in animal health and production. Motion sensors were used in this study with the aim of monitoring and quantifying the movement behaviour of traditionally managed Angoni cattle in Petauke District in the Eastern Province of Zambia. This study was designed to assess whether motion sensors were suitable for use on traditionally managed cattle in two veterinary camps in Petauke District in the Eastern Province of Zambia. In each veterinary camp, twenty cattle were selected for study. Each animal had a motion sensor placed on its hind leg to continuously measure and record its movement behaviour over a two week period. Analysing the sensor data using principal components analysis (PCA) revealed that the majority of variability in behaviour among studied cattle could be attributed to their behaviour at night and in the morning. The behaviour at night was markedly different between veterinary camps; while differences in the morning appeared to reflect varying behaviour across all animals. The study results validate the use of such motion sensors in the chosen setting and highlight the importance of appropriate data summarisation techniques to adequately describe and compare animal movement behaviours if association to other factors, such as location, breed or health status are to be assessed.
Combination of PCA and LORETA for sources analysis of ERP data: an emotional processing study
NASA Astrophysics Data System (ADS)
Hu, Jin; Tian, Jie; Yang, Lei; Pan, Xiaohong; Liu, Jiangang
2006-03-01
The purpose of this paper is to study spatiotemporal patterns of neuronal activity in emotional processing by analysis of ERP data. 108 pictures (categorized as positive, negative and neutral) were presented to 24 healthy, right-handed subjects while 128-channel EEG data were recorded. An analysis of two steps was applied to the ERP data. First, principal component analysis was performed to obtain significant ERP components. Then LORETA was applied to each component to localize their brain sources. The first six principal components were extracted, each of which showed different spatiotemporal patterns of neuronal activity. The results agree with other emotional study by fMRI or PET. The combination of PCA and LORETA can be used to analyze spatiotemporal patterns of ERP data in emotional processing.
Gould, Gillian S; Watt, Kerrianne; Stevenson, Leah; McEwen, Andy; Cadet-James, Yvonne; Clough, Alan R
2014-03-13
Smoking rates in Australian Aboriginal and Torres Strait Islander peoples remain high, with limited impact of government measures for many subgroups. The aim of this cross-sectional study was to investigate differences in organisational practice for developing anti-tobacco messages for these target populations. Telephone interviews were conducted with 47 organisation representatives using a structured questionnaire based on health communication and health promotion frameworks. Responses were coded into phases of message development, message types (educational, threat, positive or advocacy), target groups, message recommendations, and evaluations undertaken. Cultural sensitivity for message development was divided into surface structure (use of images, language, demographics) and deep structure (use of socio-cultural values). A categorical principal component analysis explored the key dimensions of the findings and their component relationships. Among organisations interviewed, a community-orientated, bottom-up approach for developing anti-tobacco messages was reported by 47% (n=24); 55% based message development on a theoretical framework; 87% used a positive benefit appeal; 38% used threat messages. More Aboriginal Medical Services (AMSs) targeted youth (p<0.005) and advised smokers to quit (p<0.05) than other types of organisations. AMSs were significantly more likely to report using deep structure in tailoring messages compared with non-government (p<0.05) and government organisations (p<0.05). Organisations that were oriented to the general population were more likely to evaluate their programs (p<0.05). A two-dimensional non-linear principal component analysis extracted components interpreted as "cultural understanding" (bottom-up, community-based approaches, deep structures) and "rigour" (theoretical frameworks, and planned/completed evaluations), and accounted for 53% of the variability in the data. Message features, associated with successful campaigns in other populations, are starting to be used for Aboriginal and Torres Strait Islander peoples. A model is proposed to facilitate the development of targeted anti-tobacco messages for Aboriginal and Torres Strait Islander peoples. Organisations could consider incorporating both components of cultural understanding-rigour to enable the growth of evidence-based practice.
2014-01-01
Background Smoking rates in Australian Aboriginal and Torres Strait Islander peoples remain high, with limited impact of government measures for many subgroups. The aim of this cross-sectional study was to investigate differences in organisational practice for developing anti-tobacco messages for these target populations. Methods Telephone interviews were conducted with 47 organisation representatives using a structured questionnaire based on health communication and health promotion frameworks. Responses were coded into phases of message development, message types (educational, threat, positive or advocacy), target groups, message recommendations, and evaluations undertaken. Cultural sensitivity for message development was divided into surface structure (use of images, language, demographics) and deep structure (use of socio-cultural values). A categorical principal component analysis explored the key dimensions of the findings and their component relationships. Results Among organisations interviewed, a community-orientated, bottom-up approach for developing anti-tobacco messages was reported by 47% (n = 24); 55% based message development on a theoretical framework; 87% used a positive benefit appeal; 38% used threat messages. More Aboriginal Medical Services (AMSs) targeted youth (p < 0.005) and advised smokers to quit (p < 0.05) than other types of organisations. AMSs were significantly more likely to report using deep structure in tailoring messages compared with non-government (p < 0.05) and government organisations (p < 0.05). Organisations that were oriented to the general population were more likely to evaluate their programs (p < 0.05). A two-dimensional non-linear principal component analysis extracted components interpreted as “cultural understanding” (bottom-up, community-based approaches, deep structures) and “rigour” (theoretical frameworks, and planned/completed evaluations), and accounted for 53% of the variability in the data. Conclusion Message features, associated with successful campaigns in other populations, are starting to be used for Aboriginal and Torres Strait Islander peoples. A model is proposed to facilitate the development of targeted anti-tobacco messages for Aboriginal and Torres Strait Islander peoples. Organisations could consider incorporating both components of cultural understanding-rigour to enable the growth of evidence-based practice. PMID:24625235
Pereira, R J; Ayres, D R; El Faro, L; Verneque, R S; Vercesi Filho, A E; Albuquerque, L G
2013-09-27
We analyzed 46,161 monthly test-day records of milk production from 7453 first lactations of crossbred dairy Gyr (Bos indicus) x Holstein cows. The following seven models were compared: standard multivariate model (M10), three reduced rank models fitting the first 2, 3, or 4 genetic principal components, and three models considering a 2-, 3-, or 4-factor structure for the genetic covariance matrix. Full rank residual covariance matrices were considered for all models. The model fitting the first two principal components (PC2) was the best according to the model selection criteria. Similar phenotypic, genetic, and residual variances were obtained with models M10 and PC2. The heritability estimates ranged from 0.14 to 0.21 and from 0.13 to 0.21 for models M10 and PC2, respectively. The genetic correlations obtained with model PC2 were slightly higher than those estimated with model M10. PC2 markedly reduced the number of parameters estimated and the time spent to reach convergence. We concluded that two principal components are sufficient to model the structure of genetic covariances between test-day milk yields.
Modeling vertebrate diversity in Oregon using satellite imagery
NASA Astrophysics Data System (ADS)
Cablk, Mary Elizabeth
Vertebrate diversity was modeled for the state of Oregon using a parametric approach to regression tree analysis. This exploratory data analysis effectively modeled the non-linear relationships between vertebrate richness and phenology, terrain, and climate. Phenology was derived from time-series NOAA-AVHRR satellite imagery for the year 1992 using two methods: principal component analysis and derivation of EROS data center greenness metrics. These two measures of spatial and temporal vegetation condition incorporated the critical temporal element in this analysis. The first three principal components were shown to contain spatial and temporal information about the landscape and discriminated phenologically distinct regions in Oregon. Principal components 2 and 3, 6 greenness metrics, elevation, slope, aspect, annual precipitation, and annual seasonal temperature difference were investigated as correlates to amphibians, birds, all vertebrates, reptiles, and mammals. Variation explained for each regression tree by taxa were: amphibians (91%), birds (67%), all vertebrates (66%), reptiles (57%), and mammals (55%). Spatial statistics were used to quantify the pattern of each taxa and assess validity of resulting predictions from regression tree models. Regression tree analysis was relatively robust against spatial autocorrelation in the response data and graphical results indicated models were well fit to the data.
Multivariate Analysis of Seismic Field Data
DOE Office of Scientific and Technical Information (OSTI.GOV)
Alam, M. Kathleen
1999-06-01
This report includes the details of the model building procedure and prediction of seismic field data. Principal Components Regression, a multivariate analysis technique, was used to model seismic data collected as two pieces of equipment were cycled on and off. Models built that included only the two pieces of equipment of interest had trouble predicting data containing signals not included in the model. Evidence for poor predictions came from the prediction curves as well as spectral F-ratio plots. Once the extraneous signals were included in the model, predictions improved dramatically. While Principal Components Regression performed well for the present datamore » sets, the present data analysis suggests further work will be needed to develop more robust modeling methods as the data become more complex.« less
The perspectives, information and conclusions conveyed in research project abstracts, progress reports, final reports, journal abstracts and journal publications convey the viewpoints of the principal investigator and may not represent the views and policies of ORD and EPA. Concl...
The perspectives, information and conclusions conveyed in research project abstracts, progress reports, final reports, journal abstracts and journal publications convey the viewpoints of the principal investigator and may not represent the views and policies of ORD and EPA. Concl...
Two-Dimensional, Supersonic, Linearized Flow with Heat Addition
NASA Technical Reports Server (NTRS)
Lomax, Harvard
1959-01-01
Calculations are presented for the forces on a thin supersonic wing underneath which the air is heated. The analysis is limited principally to linearized theory but nonlinear effects are considered. It is shown that significant advantages to external heating would exist if the heat were added well below and ahead of the wing.
Principal Killing strings in higher-dimensional Kerr-NUT-(A)dS spacetimes
NASA Astrophysics Data System (ADS)
Boos, Jens; Frolov, Valeri P.
2018-04-01
We construct special solutions of the Nambu-Goto equations for stationary strings in a general Kerr-NUT-(A)dS spacetime in any number of dimensions. This construction is based on the existence of explicit and hidden symmetries generated by the principal tensor which exists for these metrics. The characteristic property of these string configurations, which we call "principal Killing strings," is that they are stretched out from "infinity" to the horizon of the Kerr-NUT-(A)dS black hole and remain regular at the latter. We also demonstrate that principal Killing strings extract angular momentum from higher-dimensional rotating black holes and interpret this as the action of an asymptotic torque.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Yamazaki, Kazuo
2014-03-15
We study the three-dimensional magnetohydrodynamics system and obtain its regularity criteria in terms of only two velocity vector field components eliminating the condition on the third component completely. The proof consists of a new decomposition of the four nonlinear terms of the system and estimating a component of the magnetic vector field in terms of the same component of the velocity vector field. This result may be seen as a component reduction result of many previous works [C. He and Z. Xin, “On the regularity of weak solutions to the magnetohydrodynamic equations,” J. Differ. Equ. 213(2), 234–254 (2005); Y. Zhou,more » “Remarks on regularities for the 3D MHD equations,” Discrete Contin. Dyn. Syst. 12(5), 881–886 (2005)].« less
A multifactor approach to forecasting Romanian gross domestic product (GDP) in the short run.
Armeanu, Daniel; Andrei, Jean Vasile; Lache, Leonard; Panait, Mirela
2017-01-01
The purpose of this paper is to investigate the application of a generalized dynamic factor model (GDFM) based on dynamic principal components analysis to forecasting short-term economic growth in Romania. We have used a generalized principal components approach to estimate a dynamic model based on a dataset comprising 86 economic and non-economic variables that are linked to economic output. The model exploits the dynamic correlations between these variables and uses three common components that account for roughly 72% of the information contained in the original space. We show that it is possible to generate reliable forecasts of quarterly real gross domestic product (GDP) using just the common components while also assessing the contribution of the individual variables to the dynamics of real GDP. In order to assess the relative performance of the GDFM to standard models based on principal components analysis, we have also estimated two Stock-Watson (SW) models that were used to perform the same out-of-sample forecasts as the GDFM. The results indicate significantly better performance of the GDFM compared with the competing SW models, which empirically confirms our expectations that the GDFM produces more accurate forecasts when dealing with large datasets.
A multifactor approach to forecasting Romanian gross domestic product (GDP) in the short run
Armeanu, Daniel; Lache, Leonard; Panait, Mirela
2017-01-01
The purpose of this paper is to investigate the application of a generalized dynamic factor model (GDFM) based on dynamic principal components analysis to forecasting short-term economic growth in Romania. We have used a generalized principal components approach to estimate a dynamic model based on a dataset comprising 86 economic and non-economic variables that are linked to economic output. The model exploits the dynamic correlations between these variables and uses three common components that account for roughly 72% of the information contained in the original space. We show that it is possible to generate reliable forecasts of quarterly real gross domestic product (GDP) using just the common components while also assessing the contribution of the individual variables to the dynamics of real GDP. In order to assess the relative performance of the GDFM to standard models based on principal components analysis, we have also estimated two Stock-Watson (SW) models that were used to perform the same out-of-sample forecasts as the GDFM. The results indicate significantly better performance of the GDFM compared with the competing SW models, which empirically confirms our expectations that the GDFM produces more accurate forecasts when dealing with large datasets. PMID:28742100
NASA Astrophysics Data System (ADS)
Qian, Lin-Feng; Shi, Guo-Dong; Huang, Yong; Xing, Yu-Ming
2017-10-01
In vector radiative transfer, backward ray tracing is seldom used. We present a backward and forward Monte Carlo method to simulate vector radiative transfer in a two-dimensional graded index medium, which is new and different from the conventional Monte Carlo method. The backward and forward Monte Carlo method involves dividing the ray tracing into two processes backward tracing and forward tracing. In multidimensional graded index media, the trajectory of a ray is usually a three-dimensional curve. During the transport of a polarization ellipse, the curved ray trajectory will induce geometrical effects and cause Stokes parameters to continuously change. The solution processes for a non-scattering medium and an anisotropic scattering medium are analysed. We also analyse some parameters that influence the Stokes vector in two-dimensional graded index media. The research shows that the Q component of the Stokes vector cannot be ignored. However, the U and V components of the Stokes vector are very small.
Two-Dimensional One-Component Plasma on Flamm's Paraboloid
NASA Astrophysics Data System (ADS)
Fantoni, Riccardo; Téllez, Gabriel
2008-11-01
We study the classical non-relativistic two-dimensional one-component plasma at Coulomb coupling Γ=2 on the Riemannian surface known as Flamm's paraboloid which is obtained from the spatial part of the Schwarzschild metric. At this special value of the coupling constant, the statistical mechanics of the system are exactly solvable analytically. The Helmholtz free energy asymptotic expansion for the large system has been found. The density of the plasma, in the thermodynamic limit, has been carefully studied in various situations.
ERIC Educational Resources Information Center
Leithwood, K. A.
The Centre for Principal Development at the Ontario Institute for Studies in Education is a new organization devoted to research and professional development. This paper describes the framework for the center's research program and reviews research reported since 1985 that is relevant to each of the components. The review serves two purposes: (1)…
A two-dimensional, finite-difference model of the high plains aquifer in southern South Dakota
Kolm, K.E.; Case, H. L.
1983-01-01
The High Plains aquifer is the principal source of water for irrigation, industry, municipalities, and domestic use in south-central South Dakota. The aquifer, composed of upper sandstone units of the Arikaree Formation, and the overlying Ogallala and Sand Hills Formations, was simulated using a two-dimensional, finite-difference computer model. The maximum difference between simulated and measured potentiometric heads was less than 60 feet (1- to 4-percent error). Two-thirds of the simulated potentiometric heads were within 26 feet of the measured values (3-percent error). The estimated saturated thickness, computed from simulated potentiometric heads, was within 25-percent error of the known saturated thickness for 95 percent of the study area. (USGS)
NASA Technical Reports Server (NTRS)
Ghosh, Sanjoy; Goldstein, Melvyn L.
2011-01-01
Recent analysis of the magnetic correlation function of solar wind fluctuations at 1 AU suggests the existence of two-component structure near the proton-cyclotron scale. Here we use two-and-one-half dimensional and three-dimensional compressible MHD models to look for two-component structure adjacent the proton-cyclotron scale. Our MHD system incorporates both Hall and Finite Larmor Radius (FLR) terms. We find that strong spectral anisotropies appear adjacent the proton-cyclotron scales depending on selections of initial condition and plasma beta. These anisotropies are enhancements on top of related anisotropies that appear in standard MHD turbulence in the presence of a mean magnetic field and are suggestive of one turbulence component along the inertial scales and another component adjacent the dissipative scales. We compute the relative strengths of linear and nonlinear accelerations on the velocity and magnetic fields to gauge the relative influence of terms that drive the system with wave-like (linear) versus turbulent (nonlinear) dynamics.
3D Texture Analysis in Renal Cell Carcinoma Tissue Image Grading
Cho, Nam-Hoon; Choi, Heung-Kook
2014-01-01
One of the most significant processes in cancer cell and tissue image analysis is the efficient extraction of features for grading purposes. This research applied two types of three-dimensional texture analysis methods to the extraction of feature values from renal cell carcinoma tissue images, and then evaluated the validity of the methods statistically through grade classification. First, we used a confocal laser scanning microscope to obtain image slices of four grades of renal cell carcinoma, which were then reconstructed into 3D volumes. Next, we extracted quantitative values using a 3D gray level cooccurrence matrix (GLCM) and a 3D wavelet based on two types of basis functions. To evaluate their validity, we predefined 6 different statistical classifiers and applied these to the extracted feature sets. In the grade classification results, 3D Haar wavelet texture features combined with principal component analysis showed the best discrimination results. Classification using 3D wavelet texture features was significantly better than 3D GLCM, suggesting that the former has potential for use in a computer-based grading system. PMID:25371701
Vergara, Fredd; Shino, Amiu; Kikuchi, Jun
2016-01-01
Cannibalism is known in many insect species, yet its impact on insect metabolism has not been investigated in detail. This study assessed the effects of cannibalism on the metabolism of fourth-instar larvae of the non-predatory insect Helicoverpa armigera (Lepidotera: Noctuidea). Two groups of larvae were analyzed: one group fed with fourth-instar larvae of H. armigera (cannibal), the other group fed with an artificial plant diet. Water-soluble small organic compounds present in the larvae were analyzed using two-dimensional nuclear magnetic resonance (NMR) and principal component analysis (PCA). Cannibalism negatively affected larval growth. PCA of NMR spectra showed that the metabolic profiles of cannibal and herbivore larvae were statistically different with monomeric sugars, fatty acid- and amino acid-related metabolites as the most variable compounds. Quantitation of 1H-13C HSQC (Heteronuclear Single Quantum Coherence) signals revealed that the concentrations of glucose, glucono-1,5-lactone, glycerol phosphate, glutamine, glycine, leucine, isoleucine, lysine, ornithine, proline, threonine and valine were higher in the herbivore larvae. PMID:27598144
Ramli, Saifullah; Ismail, Noryati; Alkarkhi, Abbas Fadhl Mubarek; Easa, Azhar Mat
2010-08-01
Banana peel flour (BPF) prepared from green or ripe Cavendish and Dream banana fruits were assessed for their total starch (TS), digestible starch (DS), resistant starch (RS), total dietary fibre (TDF), soluble dietary fibre (SDF) and insoluble dietary fibre (IDF). Principal component analysis (PCA) identified that only 1 component was responsible for 93.74% of the total variance in the starch and dietary fibre components that differentiated ripe and green banana flours. Cluster analysis (CA) applied to similar data obtained two statistically significant clusters (green and ripe bananas) to indicate difference in behaviours according to the stages of ripeness based on starch and dietary fibre components. We concluded that the starch and dietary fibre components could be used to discriminate between flours prepared from peels obtained from fruits of different ripeness. The results were also suggestive of the potential of green and ripe BPF as functional ingredients in food.
Ramli, Saifullah; Ismail, Noryati; Alkarkhi, Abbas Fadhl Mubarek; Easa, Azhar Mat
2010-01-01
Banana peel flour (BPF) prepared from green or ripe Cavendish and Dream banana fruits were assessed for their total starch (TS), digestible starch (DS), resistant starch (RS), total dietary fibre (TDF), soluble dietary fibre (SDF) and insoluble dietary fibre (IDF). Principal component analysis (PCA) identified that only 1 component was responsible for 93.74% of the total variance in the starch and dietary fibre components that differentiated ripe and green banana flours. Cluster analysis (CA) applied to similar data obtained two statistically significant clusters (green and ripe bananas) to indicate difference in behaviours according to the stages of ripeness based on starch and dietary fibre components. We concluded that the starch and dietary fibre components could be used to discriminate between flours prepared from peels obtained from fruits of different ripeness. The results were also suggestive of the potential of green and ripe BPF as functional ingredients in food. PMID:24575193
NASA Astrophysics Data System (ADS)
Sicard, François; Senet, Patrick
2013-06-01
Well-Tempered Metadynamics (WTmetaD) is an efficient method to enhance the reconstruction of the free-energy surface of proteins. WTmetaD guarantees a faster convergence in the long time limit in comparison with the standard metadynamics. It still suffers, however, from the same limitation, i.e., the non-trivial choice of pertinent collective variables (CVs). To circumvent this problem, we couple WTmetaD with a set of CVs generated from a dihedral Principal Component Analysis (dPCA) on the Ramachandran dihedral angles describing the backbone structure of the protein. The dPCA provides a generic method to extract relevant CVs built from internal coordinates, and does not depend on the alignment to an arbitrarily chosen reference structure as usual in Cartesian PCA. We illustrate the robustness of this method in the case of a reference model protein, the small and very diffusive Met-enkephalin pentapeptide. We propose a justification a posteriori of the considered number of CVs necessary to bias the metadynamics simulation in terms of the one-dimensional free-energy profiles associated with Ramachandran dihedral angles along the amino-acid sequence.
Sicard, François; Senet, Patrick
2013-06-21
Well-Tempered Metadynamics (WTmetaD) is an efficient method to enhance the reconstruction of the free-energy surface of proteins. WTmetaD guarantees a faster convergence in the long time limit in comparison with the standard metadynamics. It still suffers, however, from the same limitation, i.e., the non-trivial choice of pertinent collective variables (CVs). To circumvent this problem, we couple WTmetaD with a set of CVs generated from a dihedral Principal Component Analysis (dPCA) on the Ramachandran dihedral angles describing the backbone structure of the protein. The dPCA provides a generic method to extract relevant CVs built from internal coordinates, and does not depend on the alignment to an arbitrarily chosen reference structure as usual in Cartesian PCA. We illustrate the robustness of this method in the case of a reference model protein, the small and very diffusive Met-enkephalin pentapeptide. We propose a justification a posteriori of the considered number of CVs necessary to bias the metadynamics simulation in terms of the one-dimensional free-energy profiles associated with Ramachandran dihedral angles along the amino-acid sequence.
Hirayama, Jun-ichiro; Hyvärinen, Aapo; Kiviniemi, Vesa; Kawanabe, Motoaki; Yamashita, Okito
2016-01-01
Characterizing the variability of resting-state functional brain connectivity across subjects and/or over time has recently attracted much attention. Principal component analysis (PCA) serves as a fundamental statistical technique for such analyses. However, performing PCA on high-dimensional connectivity matrices yields complicated “eigenconnectivity” patterns, for which systematic interpretation is a challenging issue. Here, we overcome this issue with a novel constrained PCA method for connectivity matrices by extending the idea of the previously proposed orthogonal connectivity factorization method. Our new method, modular connectivity factorization (MCF), explicitly introduces the modularity of brain networks as a parametric constraint on eigenconnectivity matrices. In particular, MCF analyzes the variability in both intra- and inter-module connectivities, simultaneously finding network modules in a principled, data-driven manner. The parametric constraint provides a compact module-based visualization scheme with which the result can be intuitively interpreted. We develop an optimization algorithm to solve the constrained PCA problem and validate our method in simulation studies and with a resting-state functional connectivity MRI dataset of 986 subjects. The results show that the proposed MCF method successfully reveals the underlying modular eigenconnectivity patterns in more general situations and is a promising alternative to existing methods. PMID:28002474
Integrative sparse principal component analysis of gene expression data.
Liu, Mengque; Fan, Xinyan; Fang, Kuangnan; Zhang, Qingzhao; Ma, Shuangge
2017-12-01
In the analysis of gene expression data, dimension reduction techniques have been extensively adopted. The most popular one is perhaps the PCA (principal component analysis). To generate more reliable and more interpretable results, the SPCA (sparse PCA) technique has been developed. With the "small sample size, high dimensionality" characteristic of gene expression data, the analysis results generated from a single dataset are often unsatisfactory. Under contexts other than dimension reduction, integrative analysis techniques, which jointly analyze the raw data of multiple independent datasets, have been developed and shown to outperform "classic" meta-analysis and other multidatasets techniques and single-dataset analysis. In this study, we conduct integrative analysis by developing the iSPCA (integrative SPCA) method. iSPCA achieves the selection and estimation of sparse loadings using a group penalty. To take advantage of the similarity across datasets and generate more accurate results, we further impose contrasted penalties. Different penalties are proposed to accommodate different data conditions. Extensive simulations show that iSPCA outperforms the alternatives under a wide spectrum of settings. The analysis of breast cancer and pancreatic cancer data further shows iSPCA's satisfactory performance. © 2017 WILEY PERIODICALS, INC.
Detection and tracking of gas plumes in LWIR hyperspectral video sequence data
NASA Astrophysics Data System (ADS)
Gerhart, Torin; Sunu, Justin; Lieu, Lauren; Merkurjev, Ekaterina; Chang, Jen-Mei; Gilles, Jérôme; Bertozzi, Andrea L.
2013-05-01
Automated detection of chemical plumes presents a segmentation challenge. The segmentation problem for gas plumes is difficult due to the diffusive nature of the cloud. The advantage of considering hyperspectral images in the gas plume detection problem over the conventional RGB imagery is the presence of non-visual data, allowing for a richer representation of information. In this paper we present an effective method of visualizing hyperspectral video sequences containing chemical plumes and investigate the effectiveness of segmentation techniques on these post-processed videos. Our approach uses a combination of dimension reduction and histogram equalization to prepare the hyperspectral videos for segmentation. First, Principal Components Analysis (PCA) is used to reduce the dimension of the entire video sequence. This is done by projecting each pixel onto the first few Principal Components resulting in a type of spectral filter. Next, a Midway method for histogram equalization is used. These methods redistribute the intensity values in order to reduce icker between frames. This properly prepares these high-dimensional video sequences for more traditional segmentation techniques. We compare the ability of various clustering techniques to properly segment the chemical plume. These include K-means, spectral clustering, and the Ginzburg-Landau functional.
Latifoğlu, Fatma; Polat, Kemal; Kara, Sadik; Güneş, Salih
2008-02-01
In this study, we proposed a new medical diagnosis system based on principal component analysis (PCA), k-NN based weighting pre-processing, and Artificial Immune Recognition System (AIRS) for diagnosis of atherosclerosis from Carotid Artery Doppler Signals. The suggested system consists of four stages. First, in the feature extraction stage, we have obtained the features related with atherosclerosis disease using Fast Fourier Transformation (FFT) modeling and by calculating of maximum frequency envelope of sonograms. Second, in the dimensionality reduction stage, the 61 features of atherosclerosis disease have been reduced to 4 features using PCA. Third, in the pre-processing stage, we have weighted these 4 features using different values of k in a new weighting scheme based on k-NN based weighting pre-processing. Finally, in the classification stage, AIRS classifier has been used to classify subjects as healthy or having atherosclerosis. Hundred percent of classification accuracy has been obtained by the proposed system using 10-fold cross validation. This success shows that the proposed system is a robust and effective system in diagnosis of atherosclerosis disease.
Robust 2DPCA with non-greedy l1 -norm maximization for image analysis.
Wang, Rong; Nie, Feiping; Yang, Xiaojun; Gao, Feifei; Yao, Minli
2015-05-01
2-D principal component analysis based on l1 -norm (2DPCA-L1) is a recently developed approach for robust dimensionality reduction and feature extraction in image domain. Normally, a greedy strategy is applied due to the difficulty of directly solving the l1 -norm maximization problem, which is, however, easy to get stuck in local solution. In this paper, we propose a robust 2DPCA with non-greedy l1 -norm maximization in which all projection directions are optimized simultaneously. Experimental results on face and other datasets confirm the effectiveness of the proposed approach.
Increasing the Coverage of Medicinal Chemistry-Relevant Space in Commercial Fragments Screening
2014-01-01
Analyzing the chemical space coverage in commercial fragment screening collections revealed the overlap between bioactive medicinal chemistry substructures and rule-of-three compliant fragments is only ∼25%. We recommend including these fragments in fragment screening libraries to maximize confidence in discovering hit matter within known bioactive chemical space, while incorporation of nonoverlapping substructures could offer novel hits in screening libraries. Using principal component analysis, polar and three-dimensional substructures display a higher-than-average enrichment of bioactive compounds, indicating increasing representation of these substructures may be beneficial in fragment screening. PMID:24405118
Kula, Katherine; Hale, Lindsay N; Ghoneima, Ahmed; Tholpady, Sunil; Starbuck, John M
2016-11-01
To compare maxillary mucosal thickening and sinus volumes of unilateral cleft lip and palate subjects (UCLP) with noncleft (nonCLP) controls. Randomized, retrospective study of cone-beam computed tomographs (CBCT). University. Fifteen UCLP subjects and 15 sex- and age-matched non-CLP controls, aged 8 to 14 years. Following institutional review board approval and reliability tests, Dolphin three-dimensional imaging software was used to segment and slice maxillary sinuses on randomly selected CBCTs. The surface area (SA) of bony sinus and airspace on all sinus slices was determined using Dolphin and multiplied by slice thickness (0.4 mm) to calculate volume. Mucosal thickening was the difference between bony sinus and airspace volumes. The number of slices with bony sinus and airspace outlines was totaled. Right and left sinus values for each group were pooled (t tests, P > .05; n = 30 each group). All measures were compared (principal components analysis, multivariate analysis of variance, analysis of variance) by group and age (P ≤ .016 was considered significant). Principal components analysis axis 1 and 2 explained 89.6% of sample variance. Principal components analysis showed complete separation based on the sample on axis 1 only. Age groups showed some separation on axis 2. Unilateral cleft lip and palate subjects had significantly smaller bony sinus and airspace volumes, fewer bony and airspace slices, and greater mucosal thickening and percentage mucosal thickening when compared with controls. Older subjects had significantly greater bony sinus and airspace volumes than younger subjects. Children with UCLP have significantly more maxillary sinus mucosal thickening and smaller sinuses than controls.
Onsager Vortex Formation in Two-component Bose-Einstein Condensates
NASA Astrophysics Data System (ADS)
Han, Junsik; Tsubota, Makoto
2018-06-01
We numerically study the dynamics of quantized vortices in two-dimensional two-component Bose-Einstein condensates (BECs) trapped by a box potential. For one-component BECs in a box potential, it is known that quantized vortices form Onsager vortices, which are clusters of same-sign vortices. We confirm that the vortices of the two components spatially separate from each other — even for miscible two-component BECs — suppressing the formation of Onsager vortices. This phenomenon is caused by the repulsive interaction between vortices belonging to different components, hence, suggesting a new possibility for vortex phase separation.
The Kirkendall and Frenkel effects during 2D diffusion process
NASA Astrophysics Data System (ADS)
Wierzba, Bartek
2014-11-01
The two-dimensional approach for inter-diffusion and voids generation is presented. The voids evolution and growth is discussed. This approach is based on the bi-velocity (Darken) method which combines the Darken and Brenner concepts that the volume velocity is essential in defining the local material velocity in multi-component mixture at non-equilibrium. The model is formulated for arbitrary multi-component two-dimensional systems. It is shown that the voids growth is due to the drift velocity and vacancy migration. The radius of the void can be easily estimated. The distributions of (1) components, (2) vacancy and (3) voids radius over the distance is presented.
Lu, Ake Tzu-Hui; Austin, Erin; Bonner, Ashley; Huang, Hsin-Hsiung; Cantor, Rita M
2014-09-01
Machine learning methods (MLMs), designed to develop models using high-dimensional predictors, have been used to analyze genome-wide genetic and genomic data to predict risks for complex traits. We summarize the results from six contributions to our Genetic Analysis Workshop 18 working group; these investigators applied MLMs and data mining to analyses of rare and common genetic variants measured in pedigrees. To develop risk profiles, group members analyzed blood pressure traits along with single-nucleotide polymorphisms and rare variant genotypes derived from sequence and imputation analyses in large Mexican American pedigrees. Supervised MLMs included penalized regression with varying penalties, support vector machines, and permanental classification. Unsupervised MLMs included sparse principal components analysis and sparse graphical models. Entropy-based components analyses were also used to mine these data. None of the investigators fully capitalized on the genetic information provided by the complete pedigrees. Their approaches either corrected for the nonindependence of the individuals within the pedigrees or analyzed only those who were independent. Some methods allowed for covariate adjustment, whereas others did not. We evaluated these methods using a variety of metrics. Four contributors conducted primary analyses on the real data, and the other two research groups used the simulated data with and without knowledge of the underlying simulation model. One group used the answers to the simulated data to assess power and type I errors. Although the MLMs applied were substantially different, each research group concluded that MLMs have advantages over standard statistical approaches with these high-dimensional data. © 2014 WILEY PERIODICALS, INC.
Coastal modification of a scene employing multispectral images and vector operators.
Lira, Jorge
2017-05-01
Changes in sea level, wind patterns, sea current patterns, and tide patterns have produced morphologic transformations in the coastline area of Tamaulipas Sate in North East Mexico. Such changes generated a modification of the coastline and variations of the texture-relief and texture of the continental area of Tamaulipas. Two high-resolution multispectral satellite Satellites Pour l'Observation de la Terre images were employed to quantify the morphologic change of such continental area. The images cover a time span close to 10 years. A variant of the principal component analysis was used to delineate the modification of the land-water line. To quantify changes in texture-relief and texture, principal component analysis was applied to the multispectral images. The first principal components of each image were modeled as a discrete bidimensional vector field. The divergence and Laplacian vector operators were applied to the discrete vector field. The divergence provided the change of texture, while the Laplacian produced the change of texture-relief in the area of study.
Dean, Jamie A; Wong, Kee H; Gay, Hiram; Welsh, Liam C; Jones, Ann-Britt; Schick, Ulrike; Oh, Jung Hun; Apte, Aditya; Newbold, Kate L; Bhide, Shreerang A; Harrington, Kevin J; Deasy, Joseph O; Nutting, Christopher M; Gulliford, Sarah L
2016-11-15
Current normal tissue complication probability modeling using logistic regression suffers from bias and high uncertainty in the presence of highly correlated radiation therapy (RT) dose data. This hinders robust estimates of dose-response associations and, hence, optimal normal tissue-sparing strategies from being elucidated. Using functional data analysis (FDA) to reduce the dimensionality of the dose data could overcome this limitation. FDA was applied to modeling of severe acute mucositis and dysphagia resulting from head and neck RT. Functional partial least squares regression (FPLS) and functional principal component analysis were used for dimensionality reduction of the dose-volume histogram data. The reduced dose data were input into functional logistic regression models (functional partial least squares-logistic regression [FPLS-LR] and functional principal component-logistic regression [FPC-LR]) along with clinical data. This approach was compared with penalized logistic regression (PLR) in terms of predictive performance and the significance of treatment covariate-response associations, assessed using bootstrapping. The area under the receiver operating characteristic curve for the PLR, FPC-LR, and FPLS-LR models was 0.65, 0.69, and 0.67, respectively, for mucositis (internal validation) and 0.81, 0.83, and 0.83, respectively, for dysphagia (external validation). The calibration slopes/intercepts for the PLR, FPC-LR, and FPLS-LR models were 1.6/-0.67, 0.45/0.47, and 0.40/0.49, respectively, for mucositis (internal validation) and 2.5/-0.96, 0.79/-0.04, and 0.79/0.00, respectively, for dysphagia (external validation). The bootstrapped odds ratios indicated significant associations between RT dose and severe toxicity in the mucositis and dysphagia FDA models. Cisplatin was significantly associated with severe dysphagia in the FDA models. None of the covariates was significantly associated with severe toxicity in the PLR models. Dose levels greater than approximately 1.0 Gy/fraction were most strongly associated with severe acute mucositis and dysphagia in the FDA models. FPLS and functional principal component analysis marginally improved predictive performance compared with PLR and provided robust dose-response associations. FDA is recommended for use in normal tissue complication probability modeling. Copyright © 2016 The Author(s). Published by Elsevier Inc. All rights reserved.
NASA Astrophysics Data System (ADS)
Kevrekidis, P. G.; Wang, Wenlong; Carretero-González, R.; Frantzeskakis, D. J.
2018-06-01
In the present work, we develop an adiabatic invariant approach for the evolution of quasi-one-dimensional (stripe) solitons embedded in a two-dimensional Bose-Einstein condensate. The results of the theory are obtained both for the one-component case of dark soliton stripes, as well as for the considerably more involved case of the two-component dark-bright (alias "filled dark") soliton stripes. In both cases, analytical predictions regarding the stability and dynamics of these structures are obtained. One of our main findings is the determination of the instability modes of the waves as a function of the parameters of the system (such as the trap strength and the chemical potential). Our analytical predictions are favorably compared with results of direct numerical simulations.
Finch, Aureliano Paolo; Brazier, John Edward; Mukuria, Clara; Bjorner, Jakob Bue
2017-12-01
Generic preference-based measures such as the EuroQol five-dimensional questionnaire (EQ-5D) are used in economic evaluation, but may not be appropriate for all conditions. When this happens, a possible solution is adding bolt-ons to expand their descriptive systems. Using review-based methods, studies published to date claimed the relevance of bolt-ons in the presence of poor psychometric results. This approach does not identify the specific dimensions missing from the Generic preference-based measure core descriptive system, and is inappropriate for identifying dimensions that might improve the measure generically. This study explores the use of principal-component analysis (PCA) and confirmatory factor analysis (CFA) for bolt-on identification in the EQ-5D. Data were drawn from the international Multi-Instrument Comparison study, which is an online survey on health and well-being measures in five countries. Analysis was based on a pool of 92 items from nine instruments. Initial content analysis provided a theoretical framework for PCA results interpretation and CFA model development. PCA was used to investigate the underlining dimensional structure and whether EQ-5D items were represented in the identified constructs. CFA was used to confirm the structure. CFA was cross-validated in random halves of the sample. PCA suggested a nine-component solution, which was confirmed by CFA. This included psychological symptoms, physical functioning, and pain, which were covered by the EQ-5D, and satisfaction, speech/cognition,relationships, hearing, vision, and energy/sleep which were not. These latter factors may represent relevant candidate bolt-ons. PCA and CFA appear useful methods for identifying potential bolt-ons dimensions for an instrument such as the EQ-5D. Copyright © 2017 International Society for Pharmacoeconomics and Outcomes Research (ISPOR). Published by Elsevier Inc. All rights reserved.
Borges, Chad R
2007-07-01
A chemometrics-based data analysis concept has been developed as a substitute for manual inspection of extracted ion chromatograms (XICs), which facilitates rapid, analyst-mediated interpretation of GC- and LC/MS(n) data sets from samples undergoing qualitative batchwise screening for prespecified sets of analytes. Automatic preparation of data into two-dimensional row space-derived scatter plots (row space plots) eliminates the need to manually interpret hundreds to thousands of XICs per batch of samples while keeping all interpretation of raw data directly in the hands of the analyst-saving great quantities of human time without loss of integrity in the data analysis process. For a given analyte, two analyte-specific variables are automatically collected by a computer algorithm and placed into a data matrix (i.e., placed into row space): the first variable is the ion abundance corresponding to scan number x and analyte-specific m/z value y, and the second variable is the ion abundance corresponding to scan number x and analyte-specific m/z value z (a second ion). These two variables serve as the two axes of the aforementioned row space plots. In order to collect appropriate scan number (retention time) information, it is necessary to analyze, as part of every batch, a sample containing a mixture of all analytes to be tested. When pure standard materials of tested analytes are unavailable, but representative ion m/z values are known and retention time can be approximated, data are evaluated based on two-dimensional scores plots from principal component analysis of small time range(s) of mass spectral data. The time-saving efficiency of this concept is directly proportional to the percentage of negative samples and to the total number of samples processed simultaneously.
Wang, Qiang; Xu, Junnan; Li, Xiang; Zhang, Dawei; Han, Yong; Zhang, Xu
2017-07-01
Radix Sophorae flavescentis is generally used for the treatment of different stages of prostate cancer in China. It has ideal effects when combined with surgical treatment and chemotherapy. However, its active components are still ambiguous. We devised a comprehensive two-dimensional PC-3 prostate cancer cell membrane chromatography system for screening anti-prostate cancer components in Radix Sophorae flavescentis. Gefitinib and dexamethasone were chosen as positive and negative drugs respectively for validation and optimization the selectivity and suitability of the comprehensive two-dimensional chromatographic system. Five compounds, sophocarpine, matrine, oxymatrine, oxysophocarpine, and xanthohumol were found to have significant retention behaviors on the PC-3 cell membrane chromatography and were unambiguously identified by time-of-flight mass spectrometry. Cell proliferation and apoptosis assays confirmed that all five compounds had anti-prostate cancer effects. Matrine and xanthohumol had good inhibitory effects, with half maximal inhibitory concentration values of 0.893 and 0.137 mg/mL, respectively. Our comprehensive two-dimensional PC-3 prostate cancer cell membrane chromatographic system promotes the efficient recognition and rapid analysis of drug candidates, and it will be practical for the discovery of prostate cancer drugs from complex traditional Chinese medicines. © 2017 WILEY-VCH Verlag GmbH & Co. KGaA, Weinheim.
Slowdown of Interhelical Motions Induces a Glass Transition in RNA
Frank, Aaron T.; Zhang, Qi; Al-Hashimi, Hashim M.; Andricioaei, Ioan
2015-01-01
RNA function depends crucially on the details of its dynamics. The simplest RNA dynamical unit is a two-way interhelical junction. Here, for such a unit—the transactivation response RNA element—we present evidence from molecular dynamics simulations, supported by nuclear magnetic resonance relaxation experiments, for a dynamical transition near 230 K. This glass transition arises from the freezing out of collective interhelical motional modes. The motions, resolved with site-specificity, are dynamically heterogeneous and exhibit non-Arrhenius relaxation. The microscopic origin of the glass transition is a low-dimensional, slow manifold consisting largely of the Euler angles describing interhelical reorientation. Principal component analysis over a range of temperatures covering the glass transition shows that the abrupt slowdown of motion finds its explanation in a localization transition that traps probability density into several disconnected conformational pools over the low-dimensional energy landscape. Upon temperature increase, the probability density pools then flood a larger basin, akin to a lakes-to-sea transition. Simulations on transactivation response RNA are also used to backcalculate inelastic neutron scattering data that match previous inelastic neutron scattering measurements on larger and more complex RNA structures and which, upon normalization, give temperature-dependent fluctuation profiles that overlap onto a glass transition curve that is quasi-universal over a range of systems and techniques. PMID:26083927
Negative Refraction Angular Characterization in One-Dimensional Photonic Crystals
Lugo, Jesus Eduardo; Doti, Rafael; Faubert, Jocelyn
2011-01-01
Background Photonic crystals are artificial structures that have periodic dielectric components with different refractive indices. Under certain conditions, they abnormally refract the light, a phenomenon called negative refraction. Here we experimentally characterize negative refraction in a one dimensional photonic crystal structure; near the low frequency edge of the fourth photonic bandgap. We compare the experimental results with current theory and a theory based on the group velocity developed here. We also analytically derived the negative refraction correctness condition that gives the angular region where negative refraction occurs. Methodology/Principal Findings By using standard photonic techniques we experimentally determined the relationship between incidence and negative refraction angles and found the negative refraction range by applying the correctness condition. In order to compare both theories with experimental results an output refraction correction was utilized. The correction uses Snell's law and an effective refractive index based on two effective dielectric constants. We found good agreement between experiment and both theories in the negative refraction zone. Conclusions/Significance Since both theories and the experimental observations agreed well in the negative refraction region, we can use both negative refraction theories plus the output correction to predict negative refraction angles. This can be very useful from a practical point of view for space filtering applications such as a photonic demultiplexer or for sensing applications. PMID:21494332
Mahmud, Iqbal; Thapaliya, Monica; Boroujerdi, Arezue; Chowdhury, Kamal
2014-01-01
The culture of sugarcane leaf explant onto culture induction medium triggers the stimulation of cell metabolism into both embryogenic and non-embryogenic callus tissues. Previous analyses demonstrated that embryogenic and nonembryogenic callus tissues have distinct metabolic profiles. This study is the follow-up to understand the biochemical relationship between the nutrient media and callus tissues using one-dimensional (1D 1H) and two-dimensional (2D 1H–13C) NMR spectroscopy followed by principal component analysis (PCA). 1D 1H spectral comparisons of fresh unspent media (FM), embryogenic callus media (ECM), non-embryogenic callus media (NECM), embryogenic callus (EC), and non-embryogenic callus (NEC), showed different metabolic relationships between callus tissues and media. Based on metabolite fold change analysis, significantly changing sugar compounds such as glucose, fructose, sucrose, and maltose were maintained in large quantities by EC only. Significantly different amino acid compounds such as valine, leucine, alanine, threonine, asparagine, and glutamine and different organic acid derivatives such as lactate, 2-hydroxyisobutyrate, 4-aminobutyrate, malonate, and choline were present in EC, NEC, and NECM, which indicates that EC maintained these nutrients, while NEC either maintained or secreted the metabolites. These media and callus-specific results suggest that EC and NEC utilize and/or secrete media nutrients differently. PMID:25012359
Aggregate Measures of Watershed Health from Reconstructed ...
Risk-based indices such as reliability, resilience, and vulnerability (R-R-V), have the potential to serve as watershed health assessment tools. Recent research has demonstrated the applicability of such indices for water quality (WQ) constituents such as total suspended solids and nutrients on an individual basis. However, the calculations can become tedious when time-series data for several WQ constituents have to be evaluated individually. Also, comparisons between locations with different sets of constituent data can prove difficult. In this study, data reconstruction using relevance vector machine algorithm was combined with dimensionality reduction via variational Bayesian noisy principal component analysis to reconstruct and condense sparse multidimensional WQ data sets into a single time series. The methodology allows incorporation of uncertainty in both the reconstruction and dimensionality-reduction steps. The R-R-V values were calculated using the aggregate time series at multiple locations within two Indiana watersheds. Results showed that uncertainty present in the reconstructed WQ data set propagates to the aggregate time series and subsequently to the aggregate R-R-V values as well. serving as motivating examples. Locations with different WQ constituents and different standards for impairment were successfully combined to provide aggregate measures of R-R-V values. Comparisons with individual constituent R-R-V values showed that v
Han, Te; Jiang, Dongxiang; Zhang, Xiaochen; Sun, Yankui
2017-01-01
Rotating machinery is widely used in industrial applications. With the trend towards more precise and more critical operating conditions, mechanical failures may easily occur. Condition monitoring and fault diagnosis (CMFD) technology is an effective tool to enhance the reliability and security of rotating machinery. In this paper, an intelligent fault diagnosis method based on dictionary learning and singular value decomposition (SVD) is proposed. First, the dictionary learning scheme is capable of generating an adaptive dictionary whose atoms reveal the underlying structure of raw signals. Essentially, dictionary learning is employed as an adaptive feature extraction method regardless of any prior knowledge. Second, the singular value sequence of learned dictionary matrix is served to extract feature vector. Generally, since the vector is of high dimensionality, a simple and practical principal component analysis (PCA) is applied to reduce dimensionality. Finally, the K-nearest neighbor (KNN) algorithm is adopted for identification and classification of fault patterns automatically. Two experimental case studies are investigated to corroborate the effectiveness of the proposed method in intelligent diagnosis of rotating machinery faults. The comparison analysis validates that the dictionary learning-based matrix construction approach outperforms the mode decomposition-based methods in terms of capacity and adaptability for feature extraction. PMID:28346385
Wang, W; Lopez, V; Thompson, D R
2006-09-01
To evaluate the validity, reliability, and cultural relevance of the Chinese Mandarin version of Myocardial Infarction Dimensional Assessment Scale (MIDAS) as a disease-specific quality of life measure. The cultural relevance and content validity of the Chinese Mandarin version of the MIDAS (CM-MIDAS) was evaluated by an expert panel. Measurement performance was tested on 180 randomly selected Chinese MI patents. Thirty participants from the primary group completed the CM-MIDAS for test-retest reliability after 2 weeks. Reliability, validity and discriminatory power of the CM-MIDAS were calculated. Two items were modified as suggested by the expert panel. The overall CM-MIDAS had acceptable internal consistency with Cronbach's alpha coefficient 0.93 for the scale and 0.71-0.94 for the seven domains. Test-retest reliability by intraclass correlations was 0.85 for the overall scale and 0.74-0.94 for the seven domains. There was acceptable concurrent validity with significant (p < 0.05) correlations between the CM-MDAS and the Chinese Version of the Short Form 36. The principal components analysis extracted seven factors that explained 67.18% of the variance with high factor loading indicating good construct validity. Empirical data support CM-MIDAS as a valid and reliable disease-specific quality of life measure for Chinese Mandarin speaking patients with myocardial infarction.
Colihueque, Nelson; Corrales, Olga; Yáñez, Miguel
2017-01-01
Trichomycterus areolatus Valenciennes, 1846 is a small endemic catfish inhabiting the Andean river basins of Chile. In this study, the morphological variability of three T. areolatus populations, collected in two river basins from southern Chile, was assessed with multivariate analyses, including principal component analysis (PCA) and discriminant function analysis (DFA). It is hypothesized that populations must segregate morphologically from each other based on the river basin that they were sampled from, since each basin presents relatively particular hydrological characteristics. Significant morphological differences among the three populations were found with PCA (ANOSIM test, r = 0.552, p < 0.0001) and DFA (Wilks's λ = 0.036, p < 0.01). PCA accounted for a total variation of 56.16% by the first two principal components. The first Principal Component (PC1) and PC2 explained 34.72 and 21.44% of the total variation, respectively. The scatter-plot of the first two discriminant functions (DF1 on DF2) also validated the existence of three different populations. In group classification using DFA, 93.3% of the specimens were correctly-classified into their original populations. Of the total of 22 transformed truss measurements, 17 exhibited highly significant ( p < 0.01) differences among populations. The data support the existence of T. areolatus morphological variation across different rivers in southern Chile, likely reflecting the geographic isolation underlying population structure of the species.
Sub-Nyquist Sampling and Moire-Like Waveform Distortions
NASA Technical Reports Server (NTRS)
Williams, Glenn L.
2000-01-01
Investigations of aliasing effects in digital waveform sampling have revealed the existence of a mathematical field and a pseudo-alias domain lying to the left of a "Nyquist line" in a plane defining the boundary between two domains of sampling. To the right of the line lies the classic alias domain. For signals band-limited below the Nyquist limit, displayed output may show a false modulation envelope. The effect occurs whenever the sample rate and the signal frequency are related by ratios of mutually prime integers. Belying the principal of a 10:1 sampling ratio being "good enough", this distortion easily occurs in graphed one-dimensional waveforms and two-dimensional images and occurs daily on television.
Zhang, Miaomiao; Wells, William M; Golland, Polina
2017-10-01
We present an efficient probabilistic model of anatomical variability in a linear space of initial velocities of diffeomorphic transformations and demonstrate its benefits in clinical studies of brain anatomy. To overcome the computational challenges of the high dimensional deformation-based descriptors, we develop a latent variable model for principal geodesic analysis (PGA) based on a low dimensional shape descriptor that effectively captures the intrinsic variability in a population. We define a novel shape prior that explicitly represents principal modes as a multivariate complex Gaussian distribution on the initial velocities in a bandlimited space. We demonstrate the performance of our model on a set of 3D brain MRI scans from the Alzheimer's Disease Neuroimaging Initiative (ADNI) database. Our model yields a more compact representation of group variation at substantially lower computational cost than the state-of-the-art method such as tangent space PCA (TPCA) and probabilistic principal geodesic analysis (PPGA) that operate in the high dimensional image space. Copyright © 2017 Elsevier B.V. All rights reserved.
Alphatome--Enhancing Spatial Reasoning: A Simulation in Two and Three Dimensions
ERIC Educational Resources Information Center
LeClair, Elizabeth E.
2003-01-01
Using refrigerator magnets, foam blocks, ink pads, and modeling clay, students manipulate the letters of the alphabet at multiple angles, reconstructing three-dimensional forms from two-dimensional data. This exercise increases students' spatial reasoning ability, an important component in many scientific disciplines. (Contains 5 figures.)
Recognizing human activities using appearance metric feature and kinematics feature
NASA Astrophysics Data System (ADS)
Qian, Huimin; Zhou, Jun; Lu, Xinbiao; Wu, Xinye
2017-05-01
The problem of automatically recognizing human activities from videos through the fusion of the two most important cues, appearance metric feature and kinematics feature, is considered. And a system of two-dimensional (2-D) Poisson equations is introduced to extract the more discriminative appearance metric feature. Specifically, the moving human blobs are first detected out from the video by background subtraction technique to form a binary image sequence, from which the appearance feature designated as the motion accumulation image and the kinematics feature termed as centroid instantaneous velocity are extracted. Second, 2-D discrete Poisson equations are employed to reinterpret the motion accumulation image to produce a more differentiated Poisson silhouette image, from which the appearance feature vector is created through the dimension reduction technique called bidirectional 2-D principal component analysis, considering the balance between classification accuracy and time consumption. Finally, a cascaded classifier based on the nearest neighbor classifier and two directed acyclic graph support vector machine classifiers, integrated with the fusion of the appearance feature vector and centroid instantaneous velocity vector, is applied to recognize the human activities. Experimental results on the open databases and a homemade one confirm the recognition performance of the proposed algorithm.
Nonlinear Principal Components Analysis: Introduction and Application
ERIC Educational Resources Information Center
Linting, Marielle; Meulman, Jacqueline J.; Groenen, Patrick J. F.; van der Koojj, Anita J.
2007-01-01
The authors provide a didactic treatment of nonlinear (categorical) principal components analysis (PCA). This method is the nonlinear equivalent of standard PCA and reduces the observed variables to a number of uncorrelated principal components. The most important advantages of nonlinear over linear PCA are that it incorporates nominal and ordinal…
USDA-ARS?s Scientific Manuscript database
Selective principal component regression analysis (SPCR) uses a subset of the original image bands for principal component transformation and regression. For optimal band selection before the transformation, this paper used genetic algorithms (GA). In this case, the GA process used the regression co...
Rank estimation and the multivariate analysis of in vivo fast-scan cyclic voltammetric data
Keithley, Richard B.; Carelli, Regina M.; Wightman, R. Mark
2010-01-01
Principal component regression has been used in the past to separate current contributions from different neuromodulators measured with in vivo fast-scan cyclic voltammetry. Traditionally, a percent cumulative variance approach has been used to determine the rank of the training set voltammetric matrix during model development, however this approach suffers from several disadvantages including the use of arbitrary percentages and the requirement of extreme precision of training sets. Here we propose that Malinowski’s F-test, a method based on a statistical analysis of the variance contained within the training set, can be used to improve factor selection for the analysis of in vivo fast-scan cyclic voltammetric data. These two methods of rank estimation were compared at all steps in the calibration protocol including the number of principal components retained, overall noise levels, model validation as determined using a residual analysis procedure, and predicted concentration information. By analyzing 119 training sets from two different laboratories amassed over several years, we were able to gain insight into the heterogeneity of in vivo fast-scan cyclic voltammetric data and study how differences in factor selection propagate throughout the entire principal component regression analysis procedure. Visualizing cyclic voltammetric representations of the data contained in the retained and discarded principal components showed that using Malinowski’s F-test for rank estimation of in vivo training sets allowed for noise to be more accurately removed. Malinowski’s F-test also improved the robustness of our criterion for judging multivariate model validity, even though signal-to-noise ratios of the data varied. In addition, pH change was the majority noise carrier of in vivo training sets while dopamine prediction was more sensitive to noise. PMID:20527815
Kernel Principal Component Analysis for dimensionality reduction in fMRI-based diagnosis of ADHD.
Sidhu, Gagan S; Asgarian, Nasimeh; Greiner, Russell; Brown, Matthew R G
2012-01-01
This study explored various feature extraction methods for use in automated diagnosis of Attention-Deficit Hyperactivity Disorder (ADHD) from functional Magnetic Resonance Image (fMRI) data. Each participant's data consisted of a resting state fMRI scan as well as phenotypic data (age, gender, handedness, IQ, and site of scanning) from the ADHD-200 dataset. We used machine learning techniques to produce support vector machine (SVM) classifiers that attempted to differentiate between (1) all ADHD patients vs. healthy controls and (2) ADHD combined (ADHD-c) type vs. ADHD inattentive (ADHD-i) type vs. controls. In different tests, we used only the phenotypic data, only the imaging data, or else both the phenotypic and imaging data. For feature extraction on fMRI data, we tested the Fast Fourier Transform (FFT), different variants of Principal Component Analysis (PCA), and combinations of FFT and PCA. PCA variants included PCA over time (PCA-t), PCA over space and time (PCA-st), and kernelized PCA (kPCA-st). Baseline chance accuracy was 64.2% produced by guessing healthy control (the majority class) for all participants. Using only phenotypic data produced 72.9% accuracy on two class diagnosis and 66.8% on three class diagnosis. Diagnosis using only imaging data did not perform as well as phenotypic-only approaches. Using both phenotypic and imaging data with combined FFT and kPCA-st feature extraction yielded accuracies of 76.0% on two class diagnosis and 68.6% on three class diagnosis-better than phenotypic-only approaches. Our results demonstrate the potential of using FFT and kPCA-st with resting-state fMRI data as well as phenotypic data for automated diagnosis of ADHD. These results are encouraging given known challenges of learning ADHD diagnostic classifiers using the ADHD-200 dataset (see Brown et al., 2012).
Ritter, Philip L; Lorig, Kate
2014-11-01
Self-efficacy theory, as developed by Bandura, suggests that self-efficacy is an important predictor of future behavior. The Chronic Disease Self-Management Program was designed to enhance self-efficacy as one approach to improving health behaviors and outcomes for people with varying chronic diseases. The six-item Self-Efficacy to Manage Chronic Disease Scale (SEMCD) and the four-item Spanish-language version (SEMCD-S) were developed to measure changes in self-efficacy in program participants and have been used in a numerous evaluations of chronic disease self-management programs. This study describes the development of the scales and their psychometric properties. Secondary analyses of questionnaire data from 2,866 participants in six studies are used to quantify and evaluate the SEMCD. Data from 868 participants in two studies are used for the SEMCD-S. Subjects consisted of individuals with various chronic conditions, who enrolled in chronic disease self-management programs (either small group or Internet based). Subjects came from United States, England, Canada, Mexico, and Australia. Descriptive statistics are summarized, reliability tested (Cronbach alpha), and principal component analyses applied to items. Baseline and change scores are correlated with baseline and change scores for five medical outcome variables that have been shown to be associated with self-efficacy measures in past studies. Principal component analyses confirmed the one-dimensional structure of the scales. The SEMCD had means ranging from 4.9 to 6.1 and the SEMCD-S 6.1 and 6.2. Internal consistency was high (Cronbach alpha, 0.88-0.95). The scales were sensitive to change and significantly correlated with health outcomes. The SEMCD and SEMCD-S are reliable and appear to be valid instruments for assessing self-efficacy for managing chronic disease. There was remarkable consistency across a range of studies from varying countries using two languages. Copyright © 2014 Elsevier Inc. All rights reserved.
Predicting degree of benefit from adjuvant trastuzumab in NSABP trial B-31.
Pogue-Geile, Katherine L; Kim, Chungyeul; Jeong, Jong-Hyeon; Tanaka, Noriko; Bandos, Hanna; Gavin, Patrick G; Fumagalli, Debora; Goldstein, Lynn C; Sneige, Nour; Burandt, Eike; Taniyama, Yusuke; Bohn, Olga L; Lee, Ahwon; Kim, Seung-Il; Reilly, Megan L; Remillard, Matthew Y; Blackmon, Nicole L; Kim, Seong-Rim; Horne, Zachary D; Rastogi, Priya; Fehrenbacher, Louis; Romond, Edward H; Swain, Sandra M; Mamounas, Eleftherios P; Wickerham, D Lawrence; Geyer, Charles E; Costantino, Joseph P; Wolmark, Norman; Paik, Soonmyung
2013-12-04
National Surgical Adjuvant Breast and Bowel Project (NSABP) trial B-31 suggested the efficacy of adjuvant trastuzumab, even in HER2-negative breast cancer. This finding prompted us to develop a predictive model for degree of benefit from trastuzumab using archived tumor blocks from B-31. Case subjects with tumor blocks were randomly divided into discovery (n = 588) and confirmation cohorts (n = 991). A predictive model was built from the discovery cohort through gene expression profiling of 462 genes with nCounter assay. A predefined cut point for the predictive model was tested in the confirmation cohort. Gene-by-treatment interaction was tested with Cox models, and correlations between variables were assessed with Spearman correlation. Principal component analysis was performed on the final set of selected genes. All statistical tests were two-sided. Eight predictive genes associated with HER2 (ERBB2, c17orf37, GRB7) or ER (ESR1, NAT1, GATA3, CA12, IGF1R) were selected for model building. Three-dimensional subset treatment effect pattern plot using two principal components of these genes was used to identify a subset with no benefit from trastuzumab, characterized by intermediate-level ERBB2 and high-level ESR1 mRNA expression. In the confirmation set, the predefined cut points for this model classified patients into three subsets with differential benefit from trastuzumab with hazard ratios of 1.58 (95% confidence interval [CI] = 0.67 to 3.69; P = .29; n = 100), 0.60 (95% CI = 0.41 to 0.89; P = .01; n = 449), and 0.28 (95% CI = 0.20 to 0.41; P < .001; n = 442; P(interaction) between the model and trastuzumab < .001). We developed a gene expression-based predictive model for degree of benefit from trastuzumab and demonstrated that HER2-negative tumors belong to the moderate benefit group, thus providing justification for testing trastuzumab in HER2-negative patients (NSABP B-47).
Predicting Degree of Benefit From Adjuvant Trastuzumab in NSABP Trial B-31
Pogue-Geile, Katherine L.; Kim, Chungyeul; Jeong, Jong-Hyeon; Tanaka, Noriko; Bandos, Hanna; Gavin, Patrick G.; Fumagalli, Debora; Goldstein, Lynn C.; Sneige, Nour; Burandt, Eike; Taniyama, Yusuke; Bohn, Olga L.; Lee, Ahwon; Kim, Seung-Il; Reilly, Megan L.; Remillard, Matthew Y.; Blackmon, Nicole L.; Kim, Seong-Rim; Horne, Zachary D.; Rastogi, Priya; Fehrenbacher, Louis; Romond, Edward H.; Swain, Sandra M.; Mamounas, Eleftherios P.; Wickerham, D. Lawrence; Geyer, Charles E.; Costantino, Joseph P.; Wolmark, Norman
2013-01-01
Background National Surgical Adjuvant Breast and Bowel Project (NSABP) trial B-31 suggested the efficacy of adjuvant trastuzumab, even in HER2-negative breast cancer. This finding prompted us to develop a predictive model for degree of benefit from trastuzumab using archived tumor blocks from B-31. Methods Case subjects with tumor blocks were randomly divided into discovery (n = 588) and confirmation cohorts (n = 991). A predictive model was built from the discovery cohort through gene expression profiling of 462 genes with nCounter assay. A predefined cut point for the predictive model was tested in the confirmation cohort. Gene-by-treatment interaction was tested with Cox models, and correlations between variables were assessed with Spearman correlation. Principal component analysis was performed on the final set of selected genes. All statistical tests were two-sided. Results Eight predictive genes associated with HER2 (ERBB2, c17orf37, GRB7) or ER (ESR1, NAT1, GATA3, CA12, IGF1R) were selected for model building. Three-dimensional subset treatment effect pattern plot using two principal components of these genes was used to identify a subset with no benefit from trastuzumab, characterized by intermediate-level ERBB2 and high-level ESR1 mRNA expression. In the confirmation set, the predefined cut points for this model classified patients into three subsets with differential benefit from trastuzumab with hazard ratios of 1.58 (95% confidence interval [CI] = 0.67 to 3.69; P = .29; n = 100), 0.60 (95% CI = 0.41 to 0.89; P = .01; n = 449), and 0.28 (95% CI = 0.20 to 0.41; P < .001; n = 442; P interaction between the model and trastuzumab < .001). Conclusions We developed a gene expression–based predictive model for degree of benefit from trastuzumab and demonstrated that HER2-negative tumors belong to the moderate benefit group, thus providing justification for testing trastuzumab in HER2-negative patients (NSABP B-47). PMID:24262440
ERIC Educational Resources Information Center
Li, Lijuan; Hallinger, Philip; Walker, Allan
2016-01-01
This study attempted to identify effects of trust between principal leadership and teacher professional learning in Hong Kong primary schools. To verify the potential mediating effects of trust as a component of school capacity, survey data with a sample of 970 teachers from 32 local primary schools was used. Two questionnaires were combined to…
Otsuka, Makoto; Fukui, Yuya; Ozaki, Yukihiro
2009-03-01
The purpose of this study was to evaluate the enzymatic stability of colloidal trypsin powder during heating in a solid-state by using Fourier transform infrared (FT-IR) spectra with chemoinformatics and generalized two-dimensional (2D) correlation spectroscopy. Colloidal crystalline trypsin powders were heated using differential scanning calorimetry. The enzymatic activity of trypsin was assayed by the kinetic degradation method. Spectra of 10 calibration sample sets were recorded three times with a FT-IR spectrometer. The maximum intensity at 1634cm(-1) of FT-IR spectra and enzymatic activity of trypsin decreased as the temperature increased. The FT-IR spectra of trypsin samples were analyzed by a principal component regression analysis (PCR). A plot of the calibration data obtained was made between the actual and predicted trypsin activity based on a two-component model with gamma(2)=0.962. On the other hand, a 2D method was applied to FT-IR spectra of heat-treated trypsin. The result was consistent with that of the chemoinformetrical method. The results for deactivation of colloidal trypsin powder by heat-treatment indicated that nano-structure of crystalline trypsin changed by heating reflecting that the beta-sheet was mainly transformed, since the peak at 1634cm(-1) decreased with dehydration. The FT-IR chemoinformetrical method allows for a solid-state quantitative analysis of the bioactivity of the bulk powder of trypsin during drying.
The perspectives, information and conclusions conveyed in research project abstracts, progress reports, final reports, journal abstracts and journal publications convey the viewpoints of the principal investigator and may not represent the views and policies of ORD and EPA. Concl...
Development of Science Anxiety Scale for Primary School Students
ERIC Educational Resources Information Center
Guzeller, Cem Oktay; Dogru, Mustafa
2012-01-01
The principal aim of the study is to develop a new scale Science Anxiety Scale and to examine its the psychometric properties and construct validity of the Science Anxiety Scale in a sample of 797 primary school students. Exploratory factor analysis was applied and found to have a two-dimensional structure. Confirmatory factor analyses provide…
NASA Astrophysics Data System (ADS)
Fan, Qingju; Wu, Yonghong
2015-08-01
In this paper, we develop a new method for the multifractal characterization of two-dimensional nonstationary signal, which is based on the detrended fluctuation analysis (DFA). By applying to two artificially generated signals of two-component ARFIMA process and binomial multifractal model, we show that the new method can reliably determine the multifractal scaling behavior of two-dimensional signal. We also illustrate the applications of this method in finance and physiology. The analyzing results exhibit that the two-dimensional signals under investigation are power-law correlations, and the electricity market consists of electricity price and trading volume is multifractal, while the two-dimensional EEG signal in sleep recorded for a single patient is weak multifractal. The new method based on the detrended fluctuation analysis may add diagnostic power to existing statistical methods.
Comparison of centre of gravity and centre of pressure patterns in the golf swing.
Smith, Aimée C; Roberts, Jonathan R; Kong, Pui Wah; Forrester, Stephanie E
2017-03-01
Analysing the centre of pressure (COP) and centre of gravity (COG) could reveal stabilising strategies used by golfers throughout the golf swing. This study identified and compared golfers' COP and COG patterns throughout the golf swing in medial-lateral (ML) and anterior-posterior (AP) directions using principal component analysis (PCA) and examined their relationship to clubhead velocity. Three-dimensional marker trajectories were collected using Vicon motion analysis and force plate data from two Kistler force plates for 22 low-handicap golfers during drives. Golfers' COG and COP were expressed as a percentage distance between their feet. PCA was performed on COG and COP in ML and AP directions. Relationships between principal component (PC) scores were examined using Pearson correlation and regression analysis used to examine the relationship with clubhead velocity. ML COP movements varied in magnitude (PC1), rate of change and timing (PC2 and PC3). The COP and COG PC1 scores were strongly correlated in both directions (ML: r = 0.90, P < .05; AP: r = 0.81, P < .05). Clubhead velocity, explained by three PCs (74%), related to timing and rate of change in COP ML near downswing (PC2 and PC3) and timing of COG ML late backswing (PC2). The relationship between COP ML and COG ML PC1 scores identified extremes of COP and COG patterns in golfers and could indicate a golfer's dynamic balance. Golfers with earlier movement of COP to the front foot (PC2) and rate of change (PC3) patterns in ML COP, prior to the downswing, may be more likely to generate higher clubhead velocity.
NASA Astrophysics Data System (ADS)
Valle, R. R.; Carvalho, F. M.; Muniz, J. A. P. C.; Leal, C. L. V.; García-Herreros, M.
2013-10-01
The aim of this study was to develop an objective method to determine the incidence of pleiomorphisms and its influence on the distribution of sperm morphometric subpopulations in ejaculates of howling monkeys ( Alouatta caraya) by using a combination of computerized analysis system (ASMA) and principal component analysis (PCA) methods. Ejaculates were collected by electroejaculation methods on a regular basis from five individuals maintained under identical captive environmental, nutritional, and management conditions. Each sperm head was measured for dimensional parameters (Area [ A, (square micrometers)], Perimeter [ P, (micrometers)], Length [ L, (micrometers)], and Width [ W, (micrometers)]) and shape-derived parameters (Ellipticity [( L/ W)], Elongation [( L - W)/( L + W)], and Rugosity [(4л A/ P 2)]). PCA revealed two principal components explaining more than the 96 % of the variance. Clustering methods and discriminant analyzes were performed and seven separate subpopulations were identified. There were differences ( P < 0.001) in the distribution of the seven subpopulations as well as in the incidence of abnormal pleiomorphisms (58.6 %, 49.8 %, 35.1 %, 66.4 %, and 55.1 %, P < 0.05) among the five donors tested. Our results indicated that differences among individuals related to the incidence of pleiomorphisms, and sperm subpopulational structure was not related to the captivity conditions or the sperm collection method, since all individuals were studied under identical conditions. In conclusion, the combination of ASMA and PCA is a useful clinical diagnostic resource for detecting deficiencies in sperm morphology and sperm subpopulations in A. caraya ejaculates that could be used in ex situ conservation programs of threatened species in Alouatta genus or even other endangered neotropical primate species.
GPU-based simulation of optical propagation through turbulence for active and passive imaging
NASA Astrophysics Data System (ADS)
Monnier, Goulven; Duval, François-Régis; Amram, Solène
2014-10-01
IMOTEP is a GPU-based (Graphical Processing Units) software relying on a fast parallel implementation of Fresnel diffraction through successive phase screens. Its applications include active imaging, laser telemetry and passive imaging through turbulence with anisoplanatic spatial and temporal fluctuations. Thanks to parallel implementation on GPU, speedups ranging from 40X to 70X are achieved. The present paper gives a brief overview of IMOTEP models, algorithms, implementation and user interface. It then focuses on major improvements recently brought to the anisoplanatic imaging simulation method. Previously, we took advantage of the computational power offered by the GPU to develop a simulation method based on large series of deterministic realisations of the PSF distorted by turbulence. The phase screen propagation algorithm, by reproducing higher moments of the incident wavefront distortion, provides realistic PSFs. However, we first used a coarse gaussian model to fit the numerical PSFs and characterise there spatial statistics through only 3 parameters (two-dimensional displacements of centroid and width). Meanwhile, this approach was unable to reproduce the effects related to the details of the PSF structure, especially the "speckles" leading to prominent high-frequency content in short-exposure images. To overcome this limitation, we recently implemented a new empirical model of the PSF, based on Principal Components Analysis (PCA), ought to catch most of the PSF complexity. The GPU implementation allows estimating and handling efficiently the numerous (up to several hundreds) principal components typically required under the strong turbulence regime. A first demanding computational step involves PCA, phase screen propagation and covariance estimates. In a second step, realistic instantaneous images, fully accounting for anisoplanatic effects, are quickly generated. Preliminary results are presented.