Sample records for stepwise feature selection

  1. Diagnosis of Chronic Kidney Disease Based on Support Vector Machine by Feature Selection Methods.

    PubMed

    Polat, Huseyin; Danaei Mehr, Homay; Cetin, Aydin

    2017-04-01

    As Chronic Kidney Disease progresses slowly, early detection and effective treatment are the only cure to reduce the mortality rate. Machine learning techniques are gaining significance in medical diagnosis because of their classification ability with high accuracy rates. The accuracy of classification algorithms depend on the use of correct feature selection algorithms to reduce the dimension of datasets. In this study, Support Vector Machine classification algorithm was used to diagnose Chronic Kidney Disease. To diagnose the Chronic Kidney Disease, two essential types of feature selection methods namely, wrapper and filter approaches were chosen to reduce the dimension of Chronic Kidney Disease dataset. In wrapper approach, classifier subset evaluator with greedy stepwise search engine and wrapper subset evaluator with the Best First search engine were used. In filter approach, correlation feature selection subset evaluator with greedy stepwise search engine and filtered subset evaluator with the Best First search engine were used. The results showed that the Support Vector Machine classifier by using filtered subset evaluator with the Best First search engine feature selection method has higher accuracy rate (98.5%) in the diagnosis of Chronic Kidney Disease compared to other selected methods.

  2. Sex determination based on a thoracic vertebra and ribs evaluation using clinical chest radiography.

    PubMed

    Tsubaki, Shun; Morishita, Junji; Usumoto, Yosuke; Sakaguchi, Kyoko; Matsunobu, Yusuke; Kawazoe, Yusuke; Okumura, Miki; Ikeda, Noriaki

    2017-07-01

    Our aim was to investigate whether sex can be determined from a combination of geometric features obtained from the 10th thoracic vertebra, 6th rib, and 7th rib. Six hundred chest radiographs (300 males and 300 females) were randomly selected to include patients of six age groups (20s, 30s, 40s, 50s, 60s, and 70s). Each group included 100 images (50 males and 50 females). A total of 14 features, including 7 lengths, 5 indices for the vertebra, and 2 types of widths for ribs, were utilized and analyzed for sex determination. Dominant features contributing to sex determination were selected by stepwise discriminant analysis after checking the variance inflation factors for multicollinearity. The accuracy of sex determination using a combination of the vertebra and ribs was evaluated from the selected features by the stepwise discriminant analysis. The accuracies in each age group were also evaluated in this study. The accuracy of sex determination based on a combination of features of the vertebra and ribs was 88.8% (533/600). This performance was superior to that of the vertebra or ribs only. Moreover, sex determination of subjects in their 20s demonstrated the highest accuracy (96.0%, 96/100). The features selected in the stepwise discriminant analysis included some features in both the vertebra and ribs. These results indicate the usefulness of combined information obtained from the vertebra and ribs for sex determination. We conclude that a combination of geometric characteristics obtained from the vertebra and ribs could be useful for determining sex. Copyright © 2017 Elsevier B.V. All rights reserved.

  3. Max-AUC Feature Selection in Computer-Aided Detection of Polyps in CT Colonography

    PubMed Central

    Xu, Jian-Wu; Suzuki, Kenji

    2014-01-01

    We propose a feature selection method based on a sequential forward floating selection (SFFS) procedure to improve the performance of a classifier in computerized detection of polyps in CT colonography (CTC). The feature selection method is coupled with a nonlinear support vector machine (SVM) classifier. Unlike the conventional linear method based on Wilks' lambda, the proposed method selected the most relevant features that would maximize the area under the receiver operating characteristic curve (AUC), which directly maximizes classification performance, evaluated based on AUC value, in the computer-aided detection (CADe) scheme. We presented two variants of the proposed method with different stopping criteria used in the SFFS procedure. The first variant searched all feature combinations allowed in the SFFS procedure and selected the subsets that maximize the AUC values. The second variant performed a statistical test at each step during the SFFS procedure, and it was terminated if the increase in the AUC value was not statistically significant. The advantage of the second variant is its lower computational cost. To test the performance of the proposed method, we compared it against the popular stepwise feature selection method based on Wilks' lambda for a colonic-polyp database (25 polyps and 2624 nonpolyps). We extracted 75 morphologic, gray-level-based, and texture features from the segmented lesion candidate regions. The two variants of the proposed feature selection method chose 29 and 7 features, respectively. Two SVM classifiers trained with these selected features yielded a 96% by-polyp sensitivity at false-positive (FP) rates of 4.1 and 6.5 per patient, respectively. Experiments showed a significant improvement in the performance of the classifier with the proposed feature selection method over that with the popular stepwise feature selection based on Wilks' lambda that yielded 18.0 FPs per patient at the same sensitivity level. PMID:24608058

  4. Max-AUC feature selection in computer-aided detection of polyps in CT colonography.

    PubMed

    Xu, Jian-Wu; Suzuki, Kenji

    2014-03-01

    We propose a feature selection method based on a sequential forward floating selection (SFFS) procedure to improve the performance of a classifier in computerized detection of polyps in CT colonography (CTC). The feature selection method is coupled with a nonlinear support vector machine (SVM) classifier. Unlike the conventional linear method based on Wilks' lambda, the proposed method selected the most relevant features that would maximize the area under the receiver operating characteristic curve (AUC), which directly maximizes classification performance, evaluated based on AUC value, in the computer-aided detection (CADe) scheme. We presented two variants of the proposed method with different stopping criteria used in the SFFS procedure. The first variant searched all feature combinations allowed in the SFFS procedure and selected the subsets that maximize the AUC values. The second variant performed a statistical test at each step during the SFFS procedure, and it was terminated if the increase in the AUC value was not statistically significant. The advantage of the second variant is its lower computational cost. To test the performance of the proposed method, we compared it against the popular stepwise feature selection method based on Wilks' lambda for a colonic-polyp database (25 polyps and 2624 nonpolyps). We extracted 75 morphologic, gray-level-based, and texture features from the segmented lesion candidate regions. The two variants of the proposed feature selection method chose 29 and 7 features, respectively. Two SVM classifiers trained with these selected features yielded a 96% by-polyp sensitivity at false-positive (FP) rates of 4.1 and 6.5 per patient, respectively. Experiments showed a significant improvement in the performance of the classifier with the proposed feature selection method over that with the popular stepwise feature selection based on Wilks' lambda that yielded 18.0 FPs per patient at the same sensitivity level.

  5. QSPR models for half-wave reduction potential of steroids: a comparative study between feature selection and feature extraction from subsets of or entire set of descriptors.

    PubMed

    Hemmateenejad, Bahram; Yazdani, Mahdieh

    2009-02-16

    Steroids are widely distributed in nature and are found in plants, animals, and fungi in abundance. A data set consists of a diverse set of steroids have been used to develop quantitative structure-electrochemistry relationship (QSER) models for their half-wave reduction potential. Modeling was established by means of multiple linear regression (MLR) and principle component regression (PCR) analyses. In MLR analysis, the QSPR models were constructed by first grouping descriptors and then stepwise selection of variables from each group (MLR1) and stepwise selection of predictor variables from the pool of all calculated descriptors (MLR2). Similar procedure was used in PCR analysis so that the principal components (or features) were extracted from different group of descriptors (PCR1) and from entire set of descriptors (PCR2). The resulted models were evaluated using cross-validation, chance correlation, application to prediction reduction potential of some test samples and accessing applicability domain. Both MLR approaches represented accurate results however the QSPR model found by MLR1 was statistically more significant. PCR1 approach produced a model as accurate as MLR approaches whereas less accurate results were obtained by PCR2 approach. In overall, the correlation coefficients of cross-validation and prediction of the QSPR models resulted from MLR1, MLR2 and PCR1 approaches were higher than 90%, which show the high ability of the models to predict reduction potential of the studied steroids.

  6. Prediction of troponin-T degradation using color image texture features in 10d aged beef longissimus steaks.

    PubMed

    Sun, X; Chen, K J; Berg, E P; Newman, D J; Schwartz, C A; Keller, W L; Maddock Carlin, K R

    2014-02-01

    The objective was to use digital color image texture features to predict troponin-T degradation in beef. Image texture features, including 88 gray level co-occurrence texture features, 81 two-dimension fast Fourier transformation texture features, and 48 Gabor wavelet filter texture features, were extracted from color images of beef strip steaks (longissimus dorsi, n = 102) aged for 10d obtained using a digital camera and additional lighting. Steaks were designated degraded or not-degraded based on troponin-T degradation determined on d 3 and d 10 postmortem by immunoblotting. Statistical analysis (STEPWISE regression model) and artificial neural network (support vector machine model, SVM) methods were designed to classify protein degradation. The d 3 and d 10 STEPWISE models were 94% and 86% accurate, respectively, while the d 3 and d 10 SVM models were 63% and 71%, respectively, in predicting protein degradation in aged meat. STEPWISE and SVM models based on image texture features show potential to predict troponin-T degradation in meat. © 2013.

  7. QSRR modeling for diverse drugs using different feature selection methods coupled with linear and nonlinear regressions.

    PubMed

    Goodarzi, Mohammad; Jensen, Richard; Vander Heyden, Yvan

    2012-12-01

    A Quantitative Structure-Retention Relationship (QSRR) is proposed to estimate the chromatographic retention of 83 diverse drugs on a Unisphere poly butadiene (PBD) column, using isocratic elutions at pH 11.7. Previous work has generated QSRR models for them using Classification And Regression Trees (CART). In this work, Ant Colony Optimization is used as a feature selection method to find the best molecular descriptors from a large pool. In addition, several other selection methods have been applied, such as Genetic Algorithms, Stepwise Regression and the Relief method, not only to evaluate Ant Colony Optimization as a feature selection method but also to investigate its ability to find the important descriptors in QSRR. Multiple Linear Regression (MLR) and Support Vector Machines (SVMs) were applied as linear and nonlinear regression methods, respectively, giving excellent correlation between the experimental, i.e. extrapolated to a mobile phase consisting of pure water, and predicted logarithms of the retention factors of the drugs (logk(w)). The overall best model was the SVM one built using descriptors selected by ACO. Copyright © 2012 Elsevier B.V. All rights reserved.

  8. Classification of early-stage non-small cell lung cancer by weighing gene expression profiles with connectivity information.

    PubMed

    Zhang, Ao; Tian, Suyan

    2018-05-01

    Pathway-based feature selection algorithms, which utilize biological information contained in pathways to guide which features/genes should be selected, have evolved quickly and become widespread in the field of bioinformatics. Based on how the pathway information is incorporated, we classify pathway-based feature selection algorithms into three major categories-penalty, stepwise forward, and weighting. Compared to the first two categories, the weighting methods have been underutilized even though they are usually the simplest ones. In this article, we constructed three different genes' connectivity information-based weights for each gene and then conducted feature selection upon the resulting weighted gene expression profiles. Using both simulations and a real-world application, we have demonstrated that when the data-driven connectivity information constructed from the data of specific disease under study is considered, the resulting weighted gene expression profiles slightly outperform the original expression profiles. In summary, a big challenge faced by the weighting method is how to estimate pathway knowledge-based weights more accurately and precisely. Only until the issue is conquered successfully will wide utilization of the weighting methods be impossible. © 2017 WILEY-VCH Verlag GmbH & Co. KGaA, Weinheim.

  9. Automated texture-based identification of ovarian cancer in confocal microendoscope images

    NASA Astrophysics Data System (ADS)

    Srivastava, Saurabh; Rodriguez, Jeffrey J.; Rouse, Andrew R.; Brewer, Molly A.; Gmitro, Arthur F.

    2005-03-01

    The fluorescence confocal microendoscope provides high-resolution, in-vivo imaging of cellular pathology during optical biopsy. There are indications that the examination of human ovaries with this instrument has diagnostic implications for the early detection of ovarian cancer. The purpose of this study was to develop a computer-aided system to facilitate the identification of ovarian cancer from digital images captured with the confocal microendoscope system. To achieve this goal, we modeled the cellular-level structure present in these images as texture and extracted features based on first-order statistics, spatial gray-level dependence matrices, and spatial-frequency content. Selection of the best features for classification was performed using traditional feature selection techniques including stepwise discriminant analysis, forward sequential search, a non-parametric method, principal component analysis, and a heuristic technique that combines the results of these methods. The best set of features selected was used for classification, and performance of various machine classifiers was compared by analyzing the areas under their receiver operating characteristic curves. The results show that it is possible to automatically identify patients with ovarian cancer based on texture features extracted from confocal microendoscope images and that the machine performance is superior to that of the human observer.

  10. Cytopathologic differential diagnosis of low-grade urothelial carcinoma and reactive urothelial proliferation in bladder washings: a logistic regression analysis.

    PubMed

    Cakir, Ebru; Kucuk, Ulku; Pala, Emel Ebru; Sezer, Ozlem; Ekin, Rahmi Gokhan; Cakmak, Ozgur

    2017-05-01

    Conventional cytomorphologic assessment is the first step to establish an accurate diagnosis in urinary cytology. In cytologic preparations, the separation of low-grade urothelial carcinoma (LGUC) from reactive urothelial proliferation (RUP) can be exceedingly difficult. The bladder washing cytologies of 32 LGUC and 29 RUP were reviewed. The cytologic slides were examined for the presence or absence of the 28 cytologic features. The cytologic criteria showing statistical significance in LGUC were increased numbers of monotonous single (non-umbrella) cells, three-dimensional cellular papillary clusters without fibrovascular cores, irregular bordered clusters, atypical single cells, irregular nuclear overlap, cytoplasmic homogeneity, increased N/C ratio, pleomorphism, nuclear border irregularity, nuclear eccentricity, elongated nuclei, and hyperchromasia (p ˂ 0.05), and the cytologic criteria showing statistical significance in RUP were inflammatory background, mixture of small and large urothelial cells, loose monolayer aggregates, and vacuolated cytoplasm (p ˂ 0.05). When these variables were subjected to a stepwise logistic regression analysis, four features were selected to distinguish LGUC from RUP: increased numbers of monotonous single (non-umbrella) cells, increased nuclear cytoplasmic ratio, hyperchromasia, and presence of small and large urothelial cells (p = 0.0001). By this logistic model of the 32 cases with proven LGUC, the stepwise logistic regression analysis correctly predicted 31 (96.9%) patients with this diagnosis, and of the 29 patients with RUP, the logistic model correctly predicted 26 (89.7%) patients as having this disease. There are several cytologic features to separate LGUC from RUP. Stepwise logistic regression analysis is a valuable tool for determining the most useful cytologic criteria to distinguish these entities. © 2017 APMIS. Published by John Wiley & Sons Ltd.

  11. Stepwise and stagewise approaches for spatial cluster detection

    PubMed Central

    Xu, Jiale

    2016-01-01

    Spatial cluster detection is an important tool in many areas such as sociology, botany and public health. Previous work has mostly taken either hypothesis testing framework or Bayesian framework. In this paper, we propose a few approaches under a frequentist variable selection framework for spatial cluster detection. The forward stepwise methods search for multiple clusters by iteratively adding currently most likely cluster while adjusting for the effects of previously identified clusters. The stagewise methods also consist of a series of steps, but with tiny step size in each iteration. We study the features and performances of our proposed methods using simulations on idealized grids or real geographic area. From the simulations, we compare the performance of the proposed methods in terms of estimation accuracy and power of detections. These methods are applied to the the well-known New York leukemia data as well as Indiana poverty data. PMID:27246273

  12. Stepwise and stagewise approaches for spatial cluster detection.

    PubMed

    Xu, Jiale; Gangnon, Ronald E

    2016-05-01

    Spatial cluster detection is an important tool in many areas such as sociology, botany and public health. Previous work has mostly taken either a hypothesis testing framework or a Bayesian framework. In this paper, we propose a few approaches under a frequentist variable selection framework for spatial cluster detection. The forward stepwise methods search for multiple clusters by iteratively adding currently most likely cluster while adjusting for the effects of previously identified clusters. The stagewise methods also consist of a series of steps, but with a tiny step size in each iteration. We study the features and performances of our proposed methods using simulations on idealized grids or real geographic areas. From the simulations, we compare the performance of the proposed methods in terms of estimation accuracy and power. These methods are applied to the the well-known New York leukemia data as well as Indiana poverty data. Copyright © 2016 Elsevier Ltd. All rights reserved.

  13. Guest-Induced Switchable Breathing Behavior in a Flexible Metal-Organic Framework with Pronounced Negative Gas Pressure.

    PubMed

    Shi, Yi-Xiang; Li, Wu-Xiang; Zhang, Wen-Hua; Lang, Jian-Ping

    2018-06-29

    Flexible metal-organic frameworks (MOFs) have attracted great interest for their dynamically structural transformability in response to external stimuli. Herein, we report a switchable "breathing" or "gate-opening" behavior associated with the phase transformation between a narrow pore (np) and a large pore (lp) in a flexible pillared-layered MOF, denoted as MOF-1 as, which is also confirmed by SCXRD and PXRD. The desolvated phase (MOF-1 des) features a unique stepwise adsorption isotherm for N 2 coupled with a pronounced negative gas adsorption pressure. For comparison, however, no appreciable CO 2 adsorption and gate-opening phenomenon with stepwise sorption can be observed. Furthermore, the polar micropore walls decorated with thiophene groups in MOF-1 des reveals the selective sorption of toluene over benzene and p-xylene associated with self-structural adjustment in spite of the markedly similar physicochemical properties of these vapor molecules.

  14. Impact of statistical learning methods on the predictive power of multivariate normal tissue complication probability models.

    PubMed

    Xu, Cheng-Jian; van der Schaaf, Arjen; Schilstra, Cornelis; Langendijk, Johannes A; van't Veld, Aart A

    2012-03-15

    To study the impact of different statistical learning methods on the prediction performance of multivariate normal tissue complication probability (NTCP) models. In this study, three learning methods, stepwise selection, least absolute shrinkage and selection operator (LASSO), and Bayesian model averaging (BMA), were used to build NTCP models of xerostomia following radiotherapy treatment for head and neck cancer. Performance of each learning method was evaluated by a repeated cross-validation scheme in order to obtain a fair comparison among methods. It was found that the LASSO and BMA methods produced models with significantly better predictive power than that of the stepwise selection method. Furthermore, the LASSO method yields an easily interpretable model as the stepwise method does, in contrast to the less intuitive BMA method. The commonly used stepwise selection method, which is simple to execute, may be insufficient for NTCP modeling. The LASSO method is recommended. Copyright © 2012 Elsevier Inc. All rights reserved.

  15. Transforming paper-based assessment forms to a digital format: Exemplified by the Housing Enabler prototype app.

    PubMed

    Svarre, Tanja; Lunn, Tine Bieber Kirkegaard; Helle, Tina

    2017-11-01

    The aim of this paper is to provide the reader with an overall impression of the stepwise user-centred design approach including the specific methods used and lessons learned when transforming paper-based assessment forms into a prototype app, taking the Housing Enabler as an example. Four design iterations were performed, building on a domain study, workshops, expert evaluation and controlled and realistic usability tests. The user-centred design process involved purposefully selected participants with different Housing Enabler knowledge and housing adaptation experience. The design iterations resulted in the development of a Housing Enabler prototype app. The prototype app has several features and options that are new compared with the original paper-based Housing Enabler assessment form. These new features include a user friendly overview of the assessment form; easy navigation by swiping back and forth between items; onsite data analysis; and ranking of the accessibility score, photo documentation and a data export facility. Based on the presented stepwise approach, a high-fidelity Housing Enabler prototype app was successfully developed. The development process has emphasized the importance of combining design participants' knowledge and experiences, and has shown that methods should seem relevant to participants to increase their engagement.

  16. Causal correlation of foliar biochemical concentrations with AVIRIS spectra using forced entry linear regression

    NASA Technical Reports Server (NTRS)

    Dawson, Terence P.; Curran, Paul J.; Kupiec, John A.

    1995-01-01

    A major goal of airborne imaging spectrometry is to estimate the biochemical composition of vegetation canopies from reflectance spectra. Remotely-sensed estimates of foliar biochemical concentrations of forests would provide valuable indicators of ecosystem function at regional and eventually global scales. Empirical research has shown a relationship exists between the amount of radiation reflected from absorption features and the concentration of given biochemicals in leaves and canopies (Matson et al., 1994, Johnson et al., 1994). A technique commonly used to determine which wavelengths have the strongest correlation with the biochemical of interest is unguided (stepwise) multiple regression. Wavelengths are entered into a multivariate regression equation, in their order of importance, each contributing to the reduction of the variance in the measured biochemical concentration. A significant problem with the use of stepwise regression for determining the correlation between biochemical concentration and spectra is that of 'overfitting' as there are significantly more wavebands than biochemical measurements. This could result in the selection of wavebands which may be more accurately attributable to noise or canopy effects. In addition, there is a real problem of collinearity in that the individual biochemical concentrations may covary. A strong correlation between the reflectance at a given wavelength and the concentration of a biochemical of interest, therefore, may be due to the effect of another biochemical which is closely related. Furthermore, it is not always possible to account for potentially suitable waveband omissions in the stepwise selection procedure. This concern about the suitability of stepwise regression has been identified and acknowledged in a number of recent studies (Wessman et al., 1988, Curran, 1989, Curran et al., 1992, Peterson and Hubbard, 1992, Martine and Aber, 1994, Kupiec, 1994). These studies have pointed to the lack of a physical link between wavelengths chosen by stepwise regression and the biochemical of interest, and this in turn has cast doubts on the use of imaging spectrometry for the estimation of foliar biochemical concentrations at sites distant from the training sites. To investigate this problem, an analysis was conducted on the variation in canopy biochemical concentrations and reflectance spectra using forced entry linear regression.

  17. Testing Different Model Building Procedures Using Multiple Regression.

    ERIC Educational Resources Information Center

    Thayer, Jerome D.

    The stepwise regression method of selecting predictors for computer assisted multiple regression analysis was compared with forward, backward, and best subsets regression, using 16 data sets. The results indicated the stepwise method was preferred because of its practical nature, when the models chosen by different selection methods were similar…

  18. The building blocks of a 'Liveable Neighbourhood': Identifying the key performance indicators for walking of an operational planning policy in Perth, Western Australia.

    PubMed

    Hooper, Paula; Knuiman, Matthew; Foster, Sarah; Giles-Corti, Billie

    2015-11-01

    Planning policy makers are requesting clearer guidance on the key design features required to build neighbourhoods that promote active living. Using a backwards stepwise elimination procedure (logistic regression with generalised estimating equations adjusting for demographic characteristics, self-selection factors, stage of construction and scale of development) this study identified specific design features (n=16) from an operational planning policy ("Liveable Neighbourhoods") that showed the strongest associations with walking behaviours (measured using the Neighbourhood Physical Activity Questionnaire). The interacting effects of design features on walking behaviours were also investigated. The urban design features identified were grouped into the "building blocks of a Liveable Neighbourhood", reflecting the scale, importance and sequencing of the design and implementation phases required to create walkable, pedestrian friendly developments. Copyright © 2015 Elsevier Ltd. All rights reserved.

  19. Selecting predictors for discriminant analysis of species performance: an example from an amphibious softwater plant.

    PubMed

    Vanderhaeghe, F; Smolders, A J P; Roelofs, J G M; Hoffmann, M

    2012-03-01

    Selecting an appropriate variable subset in linear multivariate methods is an important methodological issue for ecologists. Interest often exists in obtaining general predictive capacity or in finding causal inferences from predictor variables. Because of a lack of solid knowledge on a studied phenomenon, scientists explore predictor variables in order to find the most meaningful (i.e. discriminating) ones. As an example, we modelled the response of the amphibious softwater plant Eleocharis multicaulis using canonical discriminant function analysis. We asked how variables can be selected through comparison of several methods: univariate Pearson chi-square screening, principal components analysis (PCA) and step-wise analysis, as well as combinations of some methods. We expected PCA to perform best. The selected methods were evaluated through fit and stability of the resulting discriminant functions and through correlations between these functions and the predictor variables. The chi-square subset, at P < 0.05, followed by a step-wise sub-selection, gave the best results. In contrast to expectations, PCA performed poorly, as so did step-wise analysis. The different chi-square subset methods all yielded ecologically meaningful variables, while probable noise variables were also selected by PCA and step-wise analysis. We advise against the simple use of PCA or step-wise discriminant analysis to obtain an ecologically meaningful variable subset; the former because it does not take into account the response variable, the latter because noise variables are likely to be selected. We suggest that univariate screening techniques are a worthwhile alternative for variable selection in ecology. © 2011 German Botanical Society and The Royal Botanical Society of the Netherlands.

  20. Variable selection with stepwise and best subset approaches

    PubMed Central

    2016-01-01

    While purposeful selection is performed partly by software and partly by hand, the stepwise and best subset approaches are automatically performed by software. Two R functions stepAIC() and bestglm() are well designed for stepwise and best subset regression, respectively. The stepAIC() function begins with a full or null model, and methods for stepwise regression can be specified in the direction argument with character values “forward”, “backward” and “both”. The bestglm() function begins with a data frame containing explanatory variables and response variables. The response variable should be in the last column. Varieties of goodness-of-fit criteria can be specified in the IC argument. The Bayesian information criterion (BIC) usually results in more parsimonious model than the Akaike information criterion. PMID:27162786

  1. MULGRES: a computer program for stepwise multiple regression analysis

    Treesearch

    A. Jeff Martin

    1971-01-01

    MULGRES is a computer program source deck that is designed for multiple regression analysis employing the technique of stepwise deletion in the search for most significant variables. The features of the program, along with inputs and outputs, are briefly described, with a note on machine compatibility.

  2. Method of selective reduction of polyhalosilanes with alkyltin hydrides

    DOEpatents

    Sharp, Kenneth G.; D'Errico, John J.

    1989-01-01

    The invention relates to the selective and stepwise reduction of polyhalosilanes by reacting at room temperature or below with alkyltin hydrides without the use of free radical intermediates. Alkyltin hydrides selectively and stepwise reduce the Si--Br, Si--Cl, or Si--I bonds while leaving intact any Si--F bonds. When two or more different halogens are present on the polyhalosilane, the halogen with the highest atomic weight is preferentially reduced.

  3. Differential privacy-based evaporative cooling feature selection and classification with relief-F and random forests.

    PubMed

    Le, Trang T; Simmons, W Kyle; Misaki, Masaya; Bodurka, Jerzy; White, Bill C; Savitz, Jonathan; McKinney, Brett A

    2017-09-15

    Classification of individuals into disease or clinical categories from high-dimensional biological data with low prediction error is an important challenge of statistical learning in bioinformatics. Feature selection can improve classification accuracy but must be incorporated carefully into cross-validation to avoid overfitting. Recently, feature selection methods based on differential privacy, such as differentially private random forests and reusable holdout sets, have been proposed. However, for domains such as bioinformatics, where the number of features is much larger than the number of observations p≫n , these differential privacy methods are susceptible to overfitting. We introduce private Evaporative Cooling, a stochastic privacy-preserving machine learning algorithm that uses Relief-F for feature selection and random forest for privacy preserving classification that also prevents overfitting. We relate the privacy-preserving threshold mechanism to a thermodynamic Maxwell-Boltzmann distribution, where the temperature represents the privacy threshold. We use the thermal statistical physics concept of Evaporative Cooling of atomic gases to perform backward stepwise privacy-preserving feature selection. On simulated data with main effects and statistical interactions, we compare accuracies on holdout and validation sets for three privacy-preserving methods: the reusable holdout, reusable holdout with random forest, and private Evaporative Cooling, which uses Relief-F feature selection and random forest classification. In simulations where interactions exist between attributes, private Evaporative Cooling provides higher classification accuracy without overfitting based on an independent validation set. In simulations without interactions, thresholdout with random forest and private Evaporative Cooling give comparable accuracies. We also apply these privacy methods to human brain resting-state fMRI data from a study of major depressive disorder. Code available at http://insilico.utulsa.edu/software/privateEC . brett-mckinney@utulsa.edu. Supplementary data are available at Bioinformatics online. © The Author (2017). Published by Oxford University Press. All rights reserved. For Permissions, please email: journals.permissions@oup.com

  4. Variable selection in near-infrared spectroscopy: benchmarking of feature selection methods on biodiesel data.

    PubMed

    Balabin, Roman M; Smirnov, Sergey V

    2011-04-29

    During the past several years, near-infrared (near-IR/NIR) spectroscopy has increasingly been adopted as an analytical tool in various fields from petroleum to biomedical sectors. The NIR spectrum (above 4000 cm(-1)) of a sample is typically measured by modern instruments at a few hundred of wavelengths. Recently, considerable effort has been directed towards developing procedures to identify variables (wavelengths) that contribute useful information. Variable selection (VS) or feature selection, also called frequency selection or wavelength selection, is a critical step in data analysis for vibrational spectroscopy (infrared, Raman, or NIRS). In this paper, we compare the performance of 16 different feature selection methods for the prediction of properties of biodiesel fuel, including density, viscosity, methanol content, and water concentration. The feature selection algorithms tested include stepwise multiple linear regression (MLR-step), interval partial least squares regression (iPLS), backward iPLS (BiPLS), forward iPLS (FiPLS), moving window partial least squares regression (MWPLS), (modified) changeable size moving window partial least squares (CSMWPLS/MCSMWPLSR), searching combination moving window partial least squares (SCMWPLS), successive projections algorithm (SPA), uninformative variable elimination (UVE, including UVE-SPA), simulated annealing (SA), back-propagation artificial neural networks (BP-ANN), Kohonen artificial neural network (K-ANN), and genetic algorithms (GAs, including GA-iPLS). Two linear techniques for calibration model building, namely multiple linear regression (MLR) and partial least squares regression/projection to latent structures (PLS/PLSR), are used for the evaluation of biofuel properties. A comparison with a non-linear calibration model, artificial neural networks (ANN-MLP), is also provided. Discussion of gasoline, ethanol-gasoline (bioethanol), and diesel fuel data is presented. The results of other spectroscopic techniques application, such as Raman, ultraviolet-visible (UV-vis), or nuclear magnetic resonance (NMR) spectroscopies, can be greatly improved by an appropriate feature selection choice. Copyright © 2011 Elsevier B.V. All rights reserved.

  5. CORRELATION PURSUIT: FORWARD STEPWISE VARIABLE SELECTION FOR INDEX MODELS

    PubMed Central

    Zhong, Wenxuan; Zhang, Tingting; Zhu, Yu; Liu, Jun S.

    2012-01-01

    In this article, a stepwise procedure, correlation pursuit (COP), is developed for variable selection under the sufficient dimension reduction framework, in which the response variable Y is influenced by the predictors X1, X2, …, Xp through an unknown function of a few linear combinations of them. Unlike linear stepwise regression, COP does not impose a special form of relationship (such as linear) between the response variable and the predictor variables. The COP procedure selects variables that attain the maximum correlation between the transformed response and the linear combination of the variables. Various asymptotic properties of the COP procedure are established, and in particular, its variable selection performance under diverging number of predictors and sample size has been investigated. The excellent empirical performance of the COP procedure in comparison with existing methods are demonstrated by both extensive simulation studies and a real example in functional genomics. PMID:23243388

  6. Stationary-phase optimized selectivity liquid chromatography: development of a linear gradient prediction algorithm.

    PubMed

    De Beer, Maarten; Lynen, Fréderic; Chen, Kai; Ferguson, Paul; Hanna-Brown, Melissa; Sandra, Pat

    2010-03-01

    Stationary-phase optimized selectivity liquid chromatography (SOS-LC) is a tool in reversed-phase LC (RP-LC) to optimize the selectivity for a given separation by combining stationary phases in a multisegment column. The presently (commercially) available SOS-LC optimization procedure and algorithm are only applicable to isocratic analyses. Step gradient SOS-LC has been developed, but this is still not very elegant for the analysis of complex mixtures composed of components covering a broad hydrophobicity range. A linear gradient prediction algorithm has been developed allowing one to apply SOS-LC as a generic RP-LC optimization method. The algorithm allows operation in isocratic, stepwise, and linear gradient run modes. The features of SOS-LC in the linear gradient mode are demonstrated by means of a mixture of 13 steroids, whereby baseline separation is predicted and experimentally demonstrated.

  7. Learning templates for artistic portrait lighting analysis.

    PubMed

    Chen, Xiaowu; Jin, Xin; Wu, Hongyu; Zhao, Qinping

    2015-02-01

    Lighting is a key factor in creating impressive artistic portraits. In this paper, we propose to analyze portrait lighting by learning templates of lighting styles. Inspired by the experience of artists, we first define several novel features that describe the local contrasts in various face regions. The most informative features are then selected with a stepwise feature pursuit algorithm to derive the templates of various lighting styles. After that, the matching scores that measure the similarity between a testing portrait and those templates are calculated for lighting style classification. Furthermore, we train a regression model by the subjective scores and the feature responses of a template to predict the score of a portrait lighting quality. Based on the templates, a novel face illumination descriptor is defined to measure the difference between two portrait lightings. Experimental results show that the learned templates can well describe the lighting styles, whereas the proposed approach can assess the lighting quality of artistic portraits as human being does.

  8. A survey of variable selection methods in two Chinese epidemiology journals

    PubMed Central

    2010-01-01

    Background Although much has been written on developing better procedures for variable selection, there is little research on how it is practiced in actual studies. This review surveys the variable selection methods reported in two high-ranking Chinese epidemiology journals. Methods Articles published in 2004, 2006, and 2008 in the Chinese Journal of Epidemiology and the Chinese Journal of Preventive Medicine were reviewed. Five categories of methods were identified whereby variables were selected using: A - bivariate analyses; B - multivariable analysis; e.g. stepwise or individual significance testing of model coefficients; C - first bivariate analyses, followed by multivariable analysis; D - bivariate analyses or multivariable analysis; and E - other criteria like prior knowledge or personal judgment. Results Among the 287 articles that reported using variable selection methods, 6%, 26%, 30%, 21%, and 17% were in categories A through E, respectively. One hundred sixty-three studies selected variables using bivariate analyses, 80% (130/163) via multiple significance testing at the 5% alpha-level. Of the 219 multivariable analyses, 97 (44%) used stepwise procedures, 89 (41%) tested individual regression coefficients, but 33 (15%) did not mention how variables were selected. Sixty percent (58/97) of the stepwise routines also did not specify the algorithm and/or significance levels. Conclusions The variable selection methods reported in the two journals were limited in variety, and details were often missing. Many studies still relied on problematic techniques like stepwise procedures and/or multiple testing of bivariate associations at the 0.05 alpha-level. These deficiencies should be rectified to safeguard the scientific validity of articles published in Chinese epidemiology journals. PMID:20920252

  9. Predicting Retention Times of Naturally Occurring Phenolic Compounds in Reversed-Phase Liquid Chromatography: A Quantitative Structure-Retention Relationship (QSRR) Approach

    PubMed Central

    Akbar, Jamshed; Iqbal, Shahid; Batool, Fozia; Karim, Abdul; Chan, Kim Wei

    2012-01-01

    Quantitative structure-retention relationships (QSRRs) have successfully been developed for naturally occurring phenolic compounds in a reversed-phase liquid chromatographic (RPLC) system. A total of 1519 descriptors were calculated from the optimized structures of the molecules using MOPAC2009 and DRAGON softwares. The data set of 39 molecules was divided into training and external validation sets. For feature selection and mapping we used step-wise multiple linear regression (SMLR), unsupervised forward selection followed by step-wise multiple linear regression (UFS-SMLR) and artificial neural networks (ANN). Stable and robust models with significant predictive abilities in terms of validation statistics were obtained with negation of any chance correlation. ANN models were found better than remaining two approaches. HNar, IDM, Mp, GATS2v, DISP and 3D-MoRSE (signals 22, 28 and 32) descriptors based on van der Waals volume, electronegativity, mass and polarizability, at atomic level, were found to have significant effects on the retention times. The possible implications of these descriptors in RPLC have been discussed. All the models are proven to be quite able to predict the retention times of phenolic compounds and have shown remarkable validation, robustness, stability and predictive performance. PMID:23203132

  10. Joint effect of unlinked genotypes: application to type 2 diabetes in the EPIC-Potsdam case-cohort study.

    PubMed

    Knüppel, Sven; Meidtner, Karina; Arregui, Maria; Holzhütter, Hermann-Georg; Boeing, Heiner

    2015-07-01

    Analyzing multiple single nucleotide polymorphisms (SNPs) is a promising approach to finding genetic effects beyond single-locus associations. We proposed the use of multilocus stepwise regression (MSR) to screen for allele combinations as a method to model joint effects, and compared the results with the often used genetic risk score (GRS), conventional stepwise selection, and the shrinkage method LASSO. In contrast to MSR, the GRS, conventional stepwise selection, and LASSO model each genotype by the risk allele doses. We reanalyzed 20 unlinked SNPs related to type 2 diabetes (T2D) in the EPIC-Potsdam case-cohort study (760 cases, 2193 noncases). No SNP-SNP interactions and no nonlinear effects were found. Two SNP combinations selected by MSR (Nagelkerke's R² = 0.050 and 0.048) included eight SNPs with mean allele combination frequency of 2%. GRS and stepwise selection selected nearly the same SNP combinations consisting of 12 and 13 SNPs (Nagelkerke's R² ranged from 0.020 to 0.029). LASSO showed similar results. The MSR method showed the best model fit measured by Nagelkerke's R² suggesting that further improvement may render this method a useful tool in genetic research. However, our comparison suggests that the GRS is a simple way to model genetic effects since it does not consider linkage, SNP-SNP interactions, and no non-linear effects. © 2015 John Wiley & Sons Ltd/University College London.

  11. EEG-based mild depressive detection using feature selection methods and classifiers.

    PubMed

    Li, Xiaowei; Hu, Bin; Sun, Shuting; Cai, Hanshu

    2016-11-01

    Depression has become a major health burden worldwide, and effectively detection of such disorder is a great challenge which requires latest technological tool, such as Electroencephalography (EEG). This EEG-based research seeks to find prominent frequency band and brain regions that are most related to mild depression, as well as an optimal combination of classification algorithms and feature selection methods which can be used in future mild depression detection. An experiment based on facial expression viewing task (Emo_block and Neu_block) was conducted, and EEG data of 37 university students were collected using a 128 channel HydroCel Geodesic Sensor Net (HCGSN). For discriminating mild depressive patients and normal controls, BayesNet (BN), Support Vector Machine (SVM), Logistic Regression (LR), k-nearest neighbor (KNN) and RandomForest (RF) classifiers were used. And BestFirst (BF), GreedyStepwise (GSW), GeneticSearch (GS), LinearForwordSelection (LFS) and RankSearch (RS) based on Correlation Features Selection (CFS) were applied for linear and non-linear EEG features selection. Independent Samples T-test with Bonferroni correction was used to find the significantly discriminant electrodes and features. Data mining results indicate that optimal performance is achieved using a combination of feature selection method GSW based on CFS and classifier KNN for beta frequency band. Accuracies achieved 92.00% and 98.00%, and AUC achieved 0.957 and 0.997, for Emo_block and Neu_block beta band data respectively. T-test results validate the effectiveness of selected features by search method GSW. Simplified EEG system with only FP1, FP2, F3, O2, T3 electrodes was also explored with linear features, which yielded accuracies of 91.70% and 96.00%, AUC of 0.952 and 0.972, for Emo_block and Neu_block respectively. Classification results obtained by GSW + KNN are encouraging and better than previously published results. In the spatial distribution of features, we find that left parietotemporal lobe in beta EEG frequency band has greater effect on mild depression detection. And fewer EEG channels (FP1, FP2, F3, O2 and T3) combined with linear features may be good candidates for usage in portable systems for mild depression detection. Copyright © 2016 Elsevier Ireland Ltd. All rights reserved.

  12. The role of multicollinearity in landslide susceptibility assessment by means of Binary Logistic Regression: comparison between VIF and AIC stepwise selection

    NASA Astrophysics Data System (ADS)

    Cama, Mariaelena; Cristi Nicu, Ionut; Conoscenti, Christian; Quénéhervé, Geraldine; Maerker, Michael

    2016-04-01

    Landslide susceptibility can be defined as the likelihood of a landslide occurring in a given area on the basis of local terrain conditions. In the last decades many research focused on its evaluation by means of stochastic approaches under the assumption that 'the past is the key to the future' which means that if a model is able to reproduce a known landslide spatial distribution, it will be able to predict the future locations of new (i.e. unknown) slope failures. Among the various stochastic approaches, Binary Logistic Regression (BLR) is one of the most used because it calculates the susceptibility in probabilistic terms and its results are easily interpretable from a geomorphological point of view. However, very often not much importance is given to multicollinearity assessment whose effect is that the coefficient estimates are unstable, with opposite sign and therefore difficult to interpret. Therefore, it should be evaluated every time in order to make a model whose results are geomorphologically correct. In this study the effects of multicollinearity in the predictive performance and robustness of landslide susceptibility models are analyzed. In particular, the multicollinearity is estimated by means of Variation Inflation Index (VIF) which is also used as selection criterion for the independent variables (VIF Stepwise Selection) and compared to the more commonly used AIC Stepwise Selection. The robustness of the results is evaluated through 100 replicates of the dataset. The study area selected to perform this analysis is the Moldavian Plateau where landslides are among the most frequent geomorphological processes. This area has an increasing trend of urbanization and a very high potential regarding the cultural heritage, being the place of discovery of the largest settlement belonging to the Cucuteni Culture from Eastern Europe (that led to the development of the great complex Cucuteni-Tripyllia). Therefore, identifying the areas susceptible to landslides may lead to a better understanding and mitigation for government, local authorities and stakeholders to plan the economic activities, minimize the damages costs, environmental and cultural heritage protection. The results show that although the VIF Stepwise selection allows a more stable selection of the controlling factors, the AIC Stepwise selection produces better predictive performance. Moreover, when working with replicates the effect of multicollinearity are statistically reduced by the application of the AIC stepwise selection and the results are easily interpretable in geomorphologic terms.

  13. Diagnostic features of Alzheimer's disease extracted from PET sinograms

    NASA Astrophysics Data System (ADS)

    Sayeed, A.; Petrou, M.; Spyrou, N.; Kadyrov, A.; Spinks, T.

    2002-01-01

    Texture analysis of positron emission tomography (PET) images of the brain is a very difficult task, due to the poor signal to noise ratio. As a consequence, very few techniques can be implemented successfully. We use a new global analysis technique known as the Trace transform triple features. This technique can be applied directly to the raw sinograms to distinguish patients with Alzheimer's disease (AD) from normal volunteers. FDG-PET images of 18 AD and 10 normal controls obtained from the same CTI ECAT-953 scanner were used in this study. The Trace transform triple feature technique was used to extract features that were invariant to scaling, translation and rotation, referred to as invariant features, as well as features that were sensitive to rotation but invariant to scaling and translation, referred to as sensitive features in this study. The features were used to classify the groups using discriminant function analysis. Cross-validation tests using stepwise discriminant function analysis showed that combining both sensitive and invariant features produced the best results, when compared with the clinical diagnosis. Selecting the five best features produces an overall accuracy of 93% with sensitivity of 94% and specificity of 90%. This is comparable with the classification accuracy achieved by Kippenhan et al (1992), using regional metabolic activity.

  14. Evaluation of alternative model selection criteria in the analysis of unimodal response curves using CART

    USGS Publications Warehouse

    Ribic, C.A.; Miller, T.W.

    1998-01-01

    We investigated CART performance with a unimodal response curve for one continuous response and four continuous explanatory variables, where two variables were important (ie directly related to the response) and the other two were not. We explored performance under three relationship strengths and two explanatory variable conditions: equal importance and one variable four times as important as the other. We compared CART variable selection performance using three tree-selection rules ('minimum risk', 'minimum risk complexity', 'one standard error') to stepwise polynomial ordinary least squares (OLS) under four sample size conditions. The one-standard-error and minimum-risk-complexity methods performed about as well as stepwise OLS with large sample sizes when the relationship was strong. With weaker relationships, equally important explanatory variables and larger sample sizes, the one-standard-error and minimum-risk-complexity rules performed better than stepwise OLS. With weaker relationships and explanatory variables of unequal importance, tree-structured methods did not perform as well as stepwise OLS. Comparing performance within tree-structured methods, with a strong relationship and equally important explanatory variables, the one-standard-error-rule was more likely to choose the correct model than were the other tree-selection rules 1) with weaker relationships and equally important explanatory variables; and 2) under all relationship strengths when explanatory variables were of unequal importance and sample sizes were lower.

  15. Polar cloud and surface classification using AVHRR imagery - An intercomparison of methods

    NASA Technical Reports Server (NTRS)

    Welch, R. M.; Sengupta, S. K.; Goroch, A. K.; Rabindra, P.; Rangaraj, N.; Navar, M. S.

    1992-01-01

    Six Advanced Very High-Resolution Radiometer local area coverage (AVHRR LAC) arctic scenes are classified into ten classes. Three different classifiers are examined: (1) the traditional stepwise discriminant analysis (SDA) method; (2) the feed-forward back-propagation (FFBP) neural network; and (3) the probabilistic neural network (PNN). More than 200 spectral and textural measures are computed. These are reduced to 20 features using sequential forward selection. Theoretical accuracy of the classifiers is determined using the bootstrap approach. Overall accuracy is 85.6 percent, 87.6 percent, and 87.0 percent for the SDA, FFBP, and PNN classifiers, respectively, with standard deviations of approximately 1 percent.

  16. Parameter optimization of parenchymal texture analysis for prediction of false-positive recalls from screening mammography

    NASA Astrophysics Data System (ADS)

    Ray, Shonket; Keller, Brad M.; Chen, Jinbo; Conant, Emily F.; Kontos, Despina

    2016-03-01

    This work details a methodology to obtain optimal parameter values for a locally-adaptive texture analysis algorithm that extracts mammographic texture features representative of breast parenchymal complexity for predicting falsepositive (FP) recalls from breast cancer screening with digital mammography. The algorithm has two components: (1) adaptive selection of localized regions of interest (ROIs) and (2) Haralick texture feature extraction via Gray- Level Co-Occurrence Matrices (GLCM). The following parameters were systematically varied: mammographic views used, upper limit of the ROI window size used for adaptive ROI selection, GLCM distance offsets, and gray levels (binning) used for feature extraction. Each iteration per parameter set had logistic regression with stepwise feature selection performed on a clinical screening cohort of 474 non-recalled women and 68 FP recalled women; FP recall prediction was evaluated using area under the curve (AUC) of the receiver operating characteristic (ROC) and associations between the extracted features and FP recall were assessed via odds ratios (OR). A default instance of mediolateral (MLO) view, upper ROI size limit of 143.36 mm (2048 pixels2), GLCM distance offset combination range of 0.07 to 0.84 mm (1 to 12 pixels) and 16 GLCM gray levels was set. The highest ROC performance value of AUC=0.77 [95% confidence intervals: 0.71-0.83] was obtained at three specific instances: the default instance, upper ROI window equal to 17.92 mm (256 pixels2), and gray levels set to 128. The texture feature of sum average was chosen as a statistically significant (p<0.05) predictor and associated with higher odds of FP recall for 12 out of 14 total instances.

  17. Neighborhood Structural Similarity Mapping for the Classification of Masses in Mammograms.

    PubMed

    Rabidas, Rinku; Midya, Abhishek; Chakraborty, Jayasree

    2018-05-01

    In this paper, two novel feature extraction methods, using neighborhood structural similarity (NSS), are proposed for the characterization of mammographic masses as benign or malignant. Since gray-level distribution of pixels is different in benign and malignant masses, more regular and homogeneous patterns are visible in benign masses compared to malignant masses; the proposed method exploits the similarity between neighboring regions of masses by designing two new features, namely, NSS-I and NSS-II, which capture global similarity at different scales. Complementary to these global features, uniform local binary patterns are computed to enhance the classification efficiency by combining with the proposed features. The performance of the features are evaluated using the images from the mini-mammographic image analysis society (mini-MIAS) and digital database for screening mammography (DDSM) databases, where a tenfold cross-validation technique is incorporated with Fisher linear discriminant analysis, after selecting the optimal set of features using stepwise logistic regression method. The best area under the receiver operating characteristic curve of 0.98 with an accuracy of is achieved with the mini-MIAS database, while the same for the DDSM database is 0.93 with accuracy .

  18. Effect of finite sample size on feature selection and classification: a simulation study.

    PubMed

    Way, Ted W; Sahiner, Berkman; Hadjiiski, Lubomir M; Chan, Heang-Ping

    2010-02-01

    The small number of samples available for training and testing is often the limiting factor in finding the most effective features and designing an optimal computer-aided diagnosis (CAD) system. Training on a limited set of samples introduces bias and variance in the performance of a CAD system relative to that trained with an infinite sample size. In this work, the authors conducted a simulation study to evaluate the performances of various combinations of classifiers and feature selection techniques and their dependence on the class distribution, dimensionality, and the training sample size. The understanding of these relationships will facilitate development of effective CAD systems under the constraint of limited available samples. Three feature selection techniques, the stepwise feature selection (SFS), sequential floating forward search (SFFS), and principal component analysis (PCA), and two commonly used classifiers, Fisher's linear discriminant analysis (LDA) and support vector machine (SVM), were investigated. Samples were drawn from multidimensional feature spaces of multivariate Gaussian distributions with equal or unequal covariance matrices and unequal means, and with equal covariance matrices and unequal means estimated from a clinical data set. Classifier performance was quantified by the area under the receiver operating characteristic curve Az. The mean Az values obtained by resubstitution and hold-out methods were evaluated for training sample sizes ranging from 15 to 100 per class. The number of simulated features available for selection was chosen to be 50, 100, and 200. It was found that the relative performance of the different combinations of classifier and feature selection method depends on the feature space distributions, the dimensionality, and the available training sample sizes. The LDA and SVM with radial kernel performed similarly for most of the conditions evaluated in this study, although the SVM classifier showed a slightly higher hold-out performance than LDA for some conditions and vice versa for other conditions. PCA was comparable to or better than SFS and SFFS for LDA at small samples sizes, but inferior for SVM with polynomial kernel. For the class distributions simulated from clinical data, PCA did not show advantages over the other two feature selection methods. Under this condition, the SVM with radial kernel performed better than the LDA when few training samples were available, while LDA performed better when a large number of training samples were available. None of the investigated feature selection-classifier combinations provided consistently superior performance under the studied conditions for different sample sizes and feature space distributions. In general, the SFFS method was comparable to the SFS method while PCA may have an advantage for Gaussian feature spaces with unequal covariance matrices. The performance of the SVM with radial kernel was better than, or comparable to, that of the SVM with polynomial kernel under most conditions studied.

  19. Identification of eggs from different production systems based on hyperspectra and CS-SVM.

    PubMed

    Sun, J; Cong, S L; Mao, H P; Zhou, X; Wu, X H; Zhang, X D

    2017-06-01

    1. To identify the origin of table eggs more accurately, a method based on hyperspectral imaging technology was studied. 2. The hyperspectral data of 200 samples of intensive and extensive eggs were collected. Standard normalised variables combined with a Savitzky-Golay were used to eliminate noise, then stepwise regression (SWR) was used for feature selection. Grid search algorithm (GS), genetic search algorithm (GA), particle swarm optimisation algorithm (PSO) and cuckoo search algorithm (CS) were applied by support vector machine (SVM) methods to establish an SVM identification model with the optimal parameters. The full spectrum data and the data after feature selection were the input of the model, while egg category was the output. 3. The SWR-CS-SVM model performed better than the other models, including SWR-GS-SVM, SWR-GA-SVM, SWR-PSO-SVM and others based on full spectral data. The training and test classification accuracy of the SWR-CS-SVM model were respectively 99.3% and 96%. 4. SWR-CS-SVM proved effective for identifying egg varieties and could also be useful for the non-destructive identification of other types of egg.

  20. Application of the laguerre deconvolution method for time-resolved fluorescence spectroscopy to the characterization of atherosclerotic plaques.

    PubMed

    Jo, J A; Fang, Q; Papaioannou, T; Qiao, J H; Fishbein, M C; Beseth, B; Dorafshar, A H; Reil, T; Baker, D; Freischlag, J; Marcu, L

    2005-01-01

    This study investigates the ability of time-resolved laser-induced fluorescence spectroscopy (TR-LIFS) to detect inflammation in atherosclerotic lesion, a key feature of plaque vulnerability. A total of 348 TR-LIFS measurements were taken from carotid plaques of 30 patients, and subsequently analyzed using the Laguerre deconvolution technique. The investigated spots were classified as Early, Fibrotic/Calcified or Inflamed lesions. A stepwise linear discriminant analysis algorithm was developed using spectral and TR features (normalized intensity values and Laguerre expansion coefficients at discrete emission wavelengths, respectively). Features from only three emission wavelengths (390, 450 and 500 nm) were used in the classifier. The Inflamed lesions were discriminated with sensitivity > 80% and specificity > 90 %, when the Laguerre expansion coefficients were included in the feature space. These results indicate that TR-LIFS information derived from the Laguerre expansion coefficients at few selected emission wavelengths can discriminate inflammation in atherosclerotic plaques. We believe that TR-LIFS derived Laguerre expansion coefficients can provide a valuable additional dimension for the detection of vulnerable plaques.

  1. Benign-malignant mass classification in mammogram using edge weighted local texture features

    NASA Astrophysics Data System (ADS)

    Rabidas, Rinku; Midya, Abhishek; Sadhu, Anup; Chakraborty, Jayasree

    2016-03-01

    This paper introduces novel Discriminative Robust Local Binary Pattern (DRLBP) and Discriminative Robust Local Ternary Pattern (DRLTP) for the classification of mammographic masses as benign or malignant. Mass is one of the common, however, challenging evidence of breast cancer in mammography and diagnosis of masses is a difficult task. Since DRLBP and DRLTP overcome the drawbacks of Local Binary Pattern (LBP) and Local Ternary Pattern (LTP) by discriminating a brighter object against the dark background and vice-versa, in addition to the preservation of the edge information along with the texture information, several edge-preserving texture features are extracted, in this study, from DRLBP and DRLTP. Finally, a Fisher Linear Discriminant Analysis method is incorporated with discriminating features, selected by stepwise logistic regression method, for the classification of benign and malignant masses. The performance characteristics of DRLBP and DRLTP features are evaluated using a ten-fold cross-validation technique with 58 masses from the mini-MIAS database, and the best result is observed with DRLBP having an area under the receiver operating characteristic curve of 0.982.

  2. Improved pulmonary nodule classification utilizing quantitative lung parenchyma features.

    PubMed

    Dilger, Samantha K N; Uthoff, Johanna; Judisch, Alexandra; Hammond, Emily; Mott, Sarah L; Smith, Brian J; Newell, John D; Hoffman, Eric A; Sieren, Jessica C

    2015-10-01

    Current computer-aided diagnosis (CAD) models for determining pulmonary nodule malignancy characterize nodule shape, density, and border in computed tomography (CT) data. Analyzing the lung parenchyma surrounding the nodule has been minimally explored. We hypothesize that improved nodule classification is achievable by including features quantified from the surrounding lung tissue. To explore this hypothesis, we have developed expanded quantitative CT feature extraction techniques, including volumetric Laws texture energy measures for the parenchyma and nodule, border descriptors using ray-casting and rubber-band straightening, histogram features characterizing densities, and global lung measurements. Using stepwise forward selection and leave-one-case-out cross-validation, a neural network was used for classification. When applied to 50 nodules (22 malignant and 28 benign) from high-resolution CT scans, 52 features (8 nodule, 39 parenchymal, and 5 global) were statistically significant. Nodule-only features yielded an area under the ROC curve of 0.918 (including nodule size) and 0.872 (excluding nodule size). Performance was improved through inclusion of parenchymal (0.938) and global features (0.932). These results show a trend toward increased performance when the parenchyma is included, coupled with the large number of significant parenchymal features that support our hypothesis: the pulmonary parenchyma is influenced differentially by malignant versus benign nodules, assisting CAD-based nodule characterizations.

  3. Mobile Phone Apps to Improve Medication Adherence: A Systematic Stepwise Process to Identify High-Quality Apps.

    PubMed

    Santo, Karla; Richtering, Sarah S; Chalmers, John; Thiagalingam, Aravinda; Chow, Clara K; Redfern, Julie

    2016-12-02

    There are a growing number of mobile phone apps available to support people in taking their medications and to improve medication adherence. However, little is known about how these apps differ in terms of features, quality, and effectiveness. We aimed to systematically review the medication reminder apps available in the Australian iTunes store and Google Play to assess their features and their quality in order to identify high-quality apps. This review was conducted in a similar manner to a systematic review by using a stepwise approach that included (1) a search strategy; (2) eligibility assessment; (3) app selection process through an initial screening of all retrieved apps and full app review of the included apps; (4) data extraction using a predefined set of features considered important or desirable in medication reminder apps; (5) analysis by classifying the apps as basic and advanced medication reminder apps and scoring and ranking them; and (6) a quality assessment by using the Mobile App Rating Scale (MARS), a reliable tool to assess mobile health apps. We identified 272 medication reminder apps, of which 152 were found only in Google Play, 87 only in iTunes, and 33 in both app stores. Apps found in Google Play had more customer reviews, higher star ratings, and lower cost compared with apps in iTunes. Only 109 apps were available for free and 124 were recently updated in 2015 or 2016. Overall, the median number of features per app was 3.0 (interquartile range 4.0) and only 18 apps had ≥9 of the 17 desirable features. The most common features were flexible scheduling that was present in 56.3% (153/272) of the included apps, medication tracking history in 54.8% (149/272), snooze option in 34.9% (95/272), and visual aids in 32.4% (88/272). We classified 54.8% (149/272) of the included apps as advanced medication reminder apps and 45.2% (123/272) as basic medication reminder apps. The advanced apps had a higher number of features per app compared with the basic apps. Using the MARS instrument, we were able to identify high-quality apps that were rated as being very interesting and entertaining, highly interactive and customizable, intuitive, and easy to use and to navigate as well as having a high level of visual appeal and good-quality information. Many medication reminder apps are available in the app stores; however, the majority of them did not have many of the desirable features and were, therefore, considered low quality. Through a systematic stepwise process, we were able to identify high-quality apps to be tested in a future study that will provide evidence on the use of medication reminder apps to improve medication adherence. ©Karla Santo, Sarah S Richtering, John Chalmers, Aravinda Thiagalingam, Clara K Chow, Julie Redfern. Originally published in JMIR Mhealth and Uhealth (http://mhealth.jmir.org), 02.12.2016.

  4. Mobile Phone Apps to Improve Medication Adherence: A Systematic Stepwise Process to Identify High-Quality Apps

    PubMed Central

    Richtering, Sarah S; Chalmers, John; Thiagalingam, Aravinda; Chow, Clara K; Redfern, Julie

    2016-01-01

    Background There are a growing number of mobile phone apps available to support people in taking their medications and to improve medication adherence. However, little is known about how these apps differ in terms of features, quality, and effectiveness. Objective We aimed to systematically review the medication reminder apps available in the Australian iTunes store and Google Play to assess their features and their quality in order to identify high-quality apps. Methods This review was conducted in a similar manner to a systematic review by using a stepwise approach that included (1) a search strategy; (2) eligibility assessment; (3) app selection process through an initial screening of all retrieved apps and full app review of the included apps; (4) data extraction using a predefined set of features considered important or desirable in medication reminder apps; (5) analysis by classifying the apps as basic and advanced medication reminder apps and scoring and ranking them; and (6) a quality assessment by using the Mobile App Rating Scale (MARS), a reliable tool to assess mobile health apps. Results We identified 272 medication reminder apps, of which 152 were found only in Google Play, 87 only in iTunes, and 33 in both app stores. Apps found in Google Play had more customer reviews, higher star ratings, and lower cost compared with apps in iTunes. Only 109 apps were available for free and 124 were recently updated in 2015 or 2016. Overall, the median number of features per app was 3.0 (interquartile range 4.0) and only 18 apps had ≥9 of the 17 desirable features. The most common features were flexible scheduling that was present in 56.3% (153/272) of the included apps, medication tracking history in 54.8% (149/272), snooze option in 34.9% (95/272), and visual aids in 32.4% (88/272). We classified 54.8% (149/272) of the included apps as advanced medication reminder apps and 45.2% (123/272) as basic medication reminder apps. The advanced apps had a higher number of features per app compared with the basic apps. Using the MARS instrument, we were able to identify high-quality apps that were rated as being very interesting and entertaining, highly interactive and customizable, intuitive, and easy to use and to navigate as well as having a high level of visual appeal and good-quality information. Conclusions Many medication reminder apps are available in the app stores; however, the majority of them did not have many of the desirable features and were, therefore, considered low quality. Through a systematic stepwise process, we were able to identify high-quality apps to be tested in a future study that will provide evidence on the use of medication reminder apps to improve medication adherence. PMID:27913373

  5. Feature selection and recognition from nonspecific volatile profiles for discrimination of apple juices according to variety and geographical origin.

    PubMed

    Guo, Jing; Yue, Tianli; Yuan, Yahong

    2012-10-01

    Apple juice is a complex mixture of volatile and nonvolatile components. To develop discrimination models on the basis of the volatile composition for an efficient classification of apple juices according to apple variety and geographical origin, chromatography volatile profiles of 50 apple juice samples belonging to 6 varieties and from 5 counties of Shaanxi (China) were obtained by headspace solid-phase microextraction coupled with gas chromatography. The volatile profiles were processed as continuous and nonspecific signals through multivariate analysis techniques. Different preprocessing methods were applied to raw chromatographic data. The blind chemometric analysis of the preprocessed chromatographic profiles was carried out. Stepwise linear discriminant analysis (SLDA) revealed satisfactory discriminations of apple juices according to variety and geographical origin, provided respectively 100% and 89.8% success rate in terms of prediction ability. Finally, the discriminant volatile compounds selected by SLDA were identified by gas chromatography-mass spectrometry. The proposed strategy was able to verify the variety and geographical origin of apple juices involving only a reduced number of discriminate retention times selected by the stepwise procedure. This result encourages the similar procedures to be considered in quality control of apple juices. This work presented a method for an efficient discrimination of apple juices according to apple variety and geographical origin using HS-SPME-GC-MS together with chemometric tools. Discrimination models developed could help to achieve greater control over the quality of the juice and to detect possible adulteration of the product. © 2012 Institute of Food Technologists®

  6. Computer-aided detection of bladder mass within non-contrast-enhanced region of CT Urography (CTU)

    NASA Astrophysics Data System (ADS)

    Cha, Kenny H.; Hadjiiski, Lubomir M.; Chan, Heang-Ping; Caoili, Elaine M.; Cohan, Richard H.; Weizer, Alon; Zhou, Chuan

    2016-03-01

    We are developing a computer-aided detection system for bladder cancer in CT urography (CTU). We have previously developed methods for detection of bladder masses within the contrast-enhanced region of the bladder. In this study, we investigated methods for detection of bladder masses within the non-contrast enhanced region. The bladder was first segmented using a newly developed deep-learning convolutional neural network in combination with level sets. The non-contrast-enhanced region was separated from the contrast-enhanced region with a maximum-intensityprojection- based method. The non-contrast region was smoothed and a gray level threshold was employed to segment the bladder wall and potential masses. The bladder wall was transformed into a straightened thickness profile, which was analyzed to identify lesion candidates as a prescreening step. The lesion candidates were segmented using our autoinitialized cascaded level set (AI-CALS) segmentation method, and 27 morphological features were extracted for each candidate. Stepwise feature selection with simplex optimization and leave-one-case-out resampling were used for training and validation of a false positive (FP) classifier. In each leave-one-case-out cycle, features were selected from the training cases and a linear discriminant analysis (LDA) classifier was designed to merge the selected features into a single score for classification of the left-out test case. A data set of 33 cases with 42 biopsy-proven lesions in the noncontrast enhanced region was collected. During prescreening, the system obtained 83.3% sensitivity at an average of 2.4 FPs/case. After feature extraction and FP reduction by LDA, the system achieved 81.0% sensitivity at 2.0 FPs/case, and 73.8% sensitivity at 1.5 FPs/case.

  7. Fast H-DROP: A thirty times accelerated version of H-DROP for interactive SVM-based prediction of helical domain linkers

    NASA Astrophysics Data System (ADS)

    Richa, Tambi; Ide, Soichiro; Suzuki, Ryosuke; Ebina, Teppei; Kuroda, Yutaka

    2017-02-01

    Efficient and rapid prediction of domain regions from amino acid sequence information alone is often required for swift structural and functional characterization of large multi-domain proteins. Here we introduce Fast H-DROP, a thirty times accelerated version of our previously reported H-DROP (Helical Domain linker pRediction using OPtimal features), which is unique in specifically predicting helical domain linkers (boundaries). Fast H-DROP, analogously to H-DROP, uses optimum features selected from a set of 3000 ones by combining a random forest and a stepwise feature selection protocol. We reduced the computational time from 8.5 min per sequence in H-DROP to 14 s per sequence in Fast H-DROP on an 8 Xeon processor Linux server by using SWISS-PROT instead of Genbank non-redundant (nr) database for generating the PSSMs. The sensitivity and precision of Fast H-DROP assessed by cross-validation were 33.7 and 36.2%, which were merely 2% lower than that of H-DROP. The reduced computational time of Fast H-DROP, without affecting prediction performances, makes it more interactive and user-friendly. Fast H-DROP and H-DROP are freely available from http://domserv.lab.tuat.ac.jp/.

  8. Quantitative analysis of arterial flow properties for detection of non-calcified plaques in ECG-gated coronary CT angiography

    NASA Astrophysics Data System (ADS)

    Wei, Jun; Zhou, Chuan; Chan, Heang-Ping; Chughtai, Aamer; Agarwal, Prachi; Kuriakose, Jean; Hadjiiski, Lubomir; Patel, Smita; Kazerooni, Ella

    2015-03-01

    We are developing a computer-aided detection system to assist radiologists in detection of non-calcified plaques (NCPs) in coronary CT angiograms (cCTA). In this study, we performed quantitative analysis of arterial flow properties in each vessel branch and extracted flow information to differentiate the presence and absence of stenosis in a vessel segment. Under rest conditions, blood flow in a single vessel branch was assumed to follow Poiseuille's law. For a uniform pressure distribution, two quantitative flow features, the normalized arterial compliance per unit length (Cu) and the normalized volumetric flow (Q) along the vessel centerline, were calculated based on the parabolic Poiseuille solution. The flow features were evaluated for a two-class classification task to differentiate NCP candidates obtained by prescreening as true NCPs and false positives (FPs) in cCTA. For evaluation, a data set of 83 cCTA scans was retrospectively collected from 83 patient files with IRB approval. A total of 118 NCPs were identified by experienced cardiothoracic radiologists. The correlation between the two flow features was 0.32. The discriminatory ability of the flow features evaluated as the area under the ROC curve (AUC) was 0.65 for Cu and 0.63 for Q in comparison with AUCs of 0.56-0.69 from our previous luminal features. With stepwise LDA feature selection, volumetric flow (Q) was selected in addition to three other luminal features. With FROC analysis, the test results indicated a reduction of the FP rates to 3.14, 1.98, and 1.32 FPs/scan at sensitivities of 90%, 80%, and 70%, respectively. The study indicated that quantitative blood flow analysis has the potential to provide useful features for the detection of NCPs in cCTA.

  9. Stepwise magnetic-geochemical approach for efficient assessment of heavy metal polluted sites

    NASA Astrophysics Data System (ADS)

    Appel, E.; Rösler, W.; Ojha, G.

    2012-04-01

    Previous studies have shown that magnetometry can outline the distribution of fly ash deposition in the surroundings of coal-burning power plants and steel industries. Especially the easy-to-measure magnetic susceptibility (MS) is capable to act as a proxy for heavy metal (HM) pollution caused by such kind of point source pollution. Here we present a demonstration project around the coal-burning power plant complex "Schwarze Pumpe" in eastern Germany. Before reunification of West and East Germany huge amounts of HM pollutants were emitted from the "Schwarze Pumpe" into the environment by both fly ash emission and dumped clinker. The project has been conducted as part of the TASK Centre of Competence which aims at bringing new innovative techniques closer to the market. Our project combines in situ and laboratory MS measurements and HM analyses in order to demonstrate the efficiency of a stepwise approach for site assessment of HM pollution around point sources of fly-ash emission and deposition into soil. The following scenario is played through: We assume that the "true" spatial distribution of HM pollution (given by the pollution load index PLI comprising Fe, Zn, Pb, and Cu) is represented by our entire set of 85 measured samples (XRF analyses) from forest sites around the "Schwarze Pumpe". Surface MS data (collected with a Bartington MS2D) and in situ vertical MS sections (logged by an SM400 instrument) are used to determine a qualitative overview of potentially higher and lower polluted areas. A suite of spatial HM distribution maps obtained by random selections of 30 out of the 85 analysed sites is compared to the HM map obtained from a targeted 30-sites-selection based on pre-information from the MS results. The PLI distribution map obtained from the targeted 30-sites-selection shows all essential details of the "true" pollution map, while the different random 30-sites-selections miss important features. This comparison shows that, for the same cost investment, a stepwise combined magnetic-geochemical site assessment leads to a clearly more significant characterization of soil pollution than by a common approach with exclusively random sampling for geochemical analysis, or alternatively to an equal quality result for lower costs.

  10. Manganese-catalysed benzylic C(sp3)-H amination for late-stage functionalization

    NASA Astrophysics Data System (ADS)

    Clark, Joseph R.; Feng, Kaibo; Sookezian, Anasheh; White, M. Christina

    2018-06-01

    Reactions that directly install nitrogen into C-H bonds of complex molecules are significant because of their potential to change the chemical and biological properties of a given compound. Although selective intramolecular C-H amination reactions are known, achieving high levels of reactivity while maintaining excellent site selectivity and functional-group tolerance remains a challenge for intermolecular C-H amination. Here, we report a manganese perchlorophthalocyanine catalyst [MnIII(ClPc)] for intermolecular benzylic C-H amination of bioactive molecules and natural products that proceeds with unprecedented levels of reactivity and site selectivity. In the presence of a Brønsted or Lewis acid, the [MnIII(ClPc)]-catalysed C-H amination demonstrates unique tolerance for tertiary amine, pyridine and benzimidazole functionalities. Mechanistic studies suggest that C-H amination likely proceeds through an electrophilic metallonitrene intermediate via a stepwise pathway where C-H cleavage is the rate-determining step of the reaction. Collectively, these mechanistic features contrast with previous base-metal-catalysed C-H aminations and provide new opportunities for tunable selectivities.

  11. Manganese-catalysed benzylic C(sp3)-H amination for late-stage functionalization.

    PubMed

    Clark, Joseph R; Feng, Kaibo; Sookezian, Anasheh; White, M Christina

    2018-06-01

    Reactions that directly install nitrogen into C-H bonds of complex molecules are significant because of their potential to change the chemical and biological properties of a given compound. Although selective intramolecular C-H amination reactions are known, achieving high levels of reactivity while maintaining excellent site selectivity and functional-group tolerance remains a challenge for intermolecular C-H amination. Here, we report a manganese perchlorophthalocyanine catalyst [MnIII(ClPc)] for intermolecular benzylic C-H amination of bioactive molecules and natural products that proceeds with unprecedented levels of reactivity and site selectivity. In the presence of a Brønsted or Lewis acid, the [MnIII(ClPc)]-catalysed C-H amination demonstrates unique tolerance for tertiary amine, pyridine and benzimidazole functionalities. Mechanistic studies suggest that C-H amination likely proceeds through an electrophilic metallonitrene intermediate via a stepwise pathway where C-H cleavage is the rate-determining step of the reaction. Collectively, these mechanistic features contrast with previous base-metal-catalysed C-H aminations and provide new opportunities for tunable selectivities.

  12. Fine tuning of transmission features in nanoporous anodic alumina distributed Bragg reflectors

    NASA Astrophysics Data System (ADS)

    Lim, Siew Yee; Law, Cheryl Suwen; Santos, Abel

    2018-01-01

    This study introduces an innovative apodisation strategy to tune the filtering features of distributed Bragg reflectors based on nanoporous anodic alumina (NAA-DBRs). The effective medium of NAA-DBRs, which is modulated in a stepwise fashion by a pulse-like anodisation approach, is apodised following a logarithmic negative function to engineer the transmission features of NAA-DBRs. We investigate the effect of various apodisation parameters such as apodisation amplitude difference, anodisation period, current density offset and pore widening time, to tune and optimise the optical properties of NAA-DBRs in terms of central wavelength position, full width at half maximum and quality of photonic stop band. The transmission features of NAA-DBRs are shown to be fully controllable with precision across the spectral regions by means of the apodisation parameters. Our study demonstrates that an apodisation strategy can significantly narrow the width and enhance the quality of the characteristic photonic stop band of NAA-DBRs. This rationally designed anodisation approach based on the combination of apodisation and stepwise pulse anodisation enables the development of optical filters with tuneable filtering features to be integrated into optical technologies acting as essential photonic elements in devices such as optical sensors and biosensors.

  13. Structure-based predictions of 13C-NMR chemical shifts for a series of 2-functionalized 5-(methylsulfonyl)-1-phenyl-1H-indoles derivatives using GA-based MLR method

    NASA Astrophysics Data System (ADS)

    Ghavami, Raouf; Sadeghi, Faridoon; Rasouli, Zolikha; Djannati, Farhad

    2012-12-01

    Experimental values for the 13C NMR chemical shifts (ppm, TMS = 0) at 300 K ranging from 96.28 ppm (C4' of indole derivative 17) to 159.93 ppm (C4' of indole derivative 23) relative to deuteride chloroform (CDCl3, 77.0 ppm) or dimethylsulfoxide (DMSO, 39.50 ppm) as internal reference in CDCl3 or DMSO-d6 solutions have been collected from literature for thirty 2-functionalized 5-(methylsulfonyl)-1-phenyl-1H-indole derivatives containing different substituted groups. An effective quantitative structure-property relationship (QSPR) models were built using hybrid method combining genetic algorithm (GA) based on stepwise selection multiple linear regression (SWS-MLR) as feature-selection tools and correlation models between each carbon atom of indole derivative and calculated descriptors. Each compound was depicted by molecular structural descriptors that encode constitutional, topological, geometrical, electrostatic, and quantum chemical features. The accuracy of all developed models were confirmed using different types of internal and external procedures and various statistical tests. Furthermore, the domain of applicability for each model which indicates the area of reliable predictions was defined.

  14. Stepwise molding, etching, and imprinting to form libraries of nanopatterned substrates.

    PubMed

    Zhao, Zhi; Cai, Yangjun; Liao, Wei-Ssu; Cremer, Paul S

    2013-06-04

    Herein, we describe a novel colloidal lithographic strategy for the stepwise patterning of planar substrates with numerous complex and unique designs. In conjunction with colloidal self-assembly, imprint molding, and capillary force lithography, reactive ion etching was used to create complex libraries of nanoscale features. This combinatorial strategy affords the ability to develop an exponentially increasing number of two-dimensional nanoscale patterns with each sequential step in the process. Specifically, dots, triangles, circles, and lines could be assembled on the surface separately and in combination with each other. Numerous architectures are obtained for the first time with high uniformity and reproducibility. These hexagonal arrays were made from polystyrene and gold features, whereby each surface element could be tuned from the micrometer size scale down to line widths of ~35 nm. The patterned area could be 1 cm(2) or even larger. The techniques described herein can be combined with further steps to make even larger libraries. Moreover, these polymer and metal features may prove useful in optical, sensing, and electronic applications.

  15. Selecting risk factors: a comparison of discriminant analysis, logistic regression and Cox's regression model using data from the Tromsø Heart Study.

    PubMed

    Brenn, T; Arnesen, E

    1985-01-01

    For comparative evaluation, discriminant analysis, logistic regression and Cox's model were used to select risk factors for total and coronary deaths among 6595 men aged 20-49 followed for 9 years. Groups with mortality between 5 and 93 per 1000 were considered. Discriminant analysis selected variable sets only marginally different from the logistic and Cox methods which always selected the same sets. A time-saving option, offered for both the logistic and Cox selection, showed no advantage compared with discriminant analysis. Analysing more than 3800 subjects, the logistic and Cox methods consumed, respectively, 80 and 10 times more computer time than discriminant analysis. When including the same set of variables in non-stepwise analyses, all methods estimated coefficients that in most cases were almost identical. In conclusion, discriminant analysis is advocated for preliminary or stepwise analysis, otherwise Cox's method should be used.

  16. Integrative Bayesian variable selection with gene-based informative priors for genome-wide association studies.

    PubMed

    Zhang, Xiaoshuai; Xue, Fuzhong; Liu, Hong; Zhu, Dianwen; Peng, Bin; Wiemels, Joseph L; Yang, Xiaowei

    2014-12-10

    Genome-wide Association Studies (GWAS) are typically designed to identify phenotype-associated single nucleotide polymorphisms (SNPs) individually using univariate analysis methods. Though providing valuable insights into genetic risks of common diseases, the genetic variants identified by GWAS generally account for only a small proportion of the total heritability for complex diseases. To solve this "missing heritability" problem, we implemented a strategy called integrative Bayesian Variable Selection (iBVS), which is based on a hierarchical model that incorporates an informative prior by considering the gene interrelationship as a network. It was applied here to both simulated and real data sets. Simulation studies indicated that the iBVS method was advantageous in its performance with highest AUC in both variable selection and outcome prediction, when compared to Stepwise and LASSO based strategies. In an analysis of a leprosy case-control study, iBVS selected 94 SNPs as predictors, while LASSO selected 100 SNPs. The Stepwise regression yielded a more parsimonious model with only 3 SNPs. The prediction results demonstrated that the iBVS method had comparable performance with that of LASSO, but better than Stepwise strategies. The proposed iBVS strategy is a novel and valid method for Genome-wide Association Studies, with the additional advantage in that it produces more interpretable posterior probabilities for each variable unlike LASSO and other penalized regression methods.

  17. Vapor permeation-stepwise injection simultaneous determination of methanol and ethanol in biodiesel with voltammetric detection.

    PubMed

    Shishov, Andrey; Penkova, Anastasia; Zabrodin, Andrey; Nikolaev, Konstantin; Dmitrenko, Maria; Ermakov, Sergey; Bulatov, Andrey

    2016-02-01

    A novel vapor permeation-stepwise injection (VP-SWI) method for the determination of methanol and ethanol in biodiesel samples is discussed. In the current study, stepwise injection analysis was successfully combined with voltammetric detection and vapor permeation. This method is based on the separation of methanol and ethanol from a sample using a vapor permeation module (VPM) with a selective polymer membrane based on poly(phenylene isophtalamide) (PA) containing high amounts of a residual solvent. After the evaporation into the headspace of the VPM, methanol and ethanol were transported, by gas bubbling, through a PA membrane to a mixing chamber equipped with a voltammetric detector. Ethanol was selectively detected at +0.19 V, and both compounds were detected at +1.20 V. Current subtractions (using a correction factor) were used for the selective determination of methanol. A linear range between 0.05 and 0.5% (m/m) was established for each analyte. The limits of detection were estimated at 0.02% (m/m) for ethanol and methanol. The sample throughput was 5 samples h(-1). The method was successfully applied to the analysis of biodiesel samples. Copyright © 2015 Elsevier B.V. All rights reserved.

  18. Invasive placenta previa: Placental bulge with distorted uterine outline and uterine serosal hypervascularity at 1.5T MRI - useful features for differentiating placenta percreta from placenta accreta.

    PubMed

    Chen, Xin; Shan, Ruiqin; Zhao, Lianxin; Song, Qingxu; Zuo, Changting; Zhang, Xinjuan; Wang, Shanshan; Shi, Honglu; Gao, Fei; Qian, Tianyi; Wang, Guangbin; Limperopoulos, Catherine

    2018-02-01

    To characterise MRI features of invasive placenta previa and to identify specific features for differentiating placenta percreta (PP) from placenta accreta (PA). Forty-five women with PP and 93 women with PA who underwent 1.5T placental MRI were included. Two radiologists independently evaluated the MRI features of invasive placenta previa, including our novel type of placental bulge (i.e. placental bulge type-II, characterized by placental bulge with distorted uterine outline). Pearson's chi-squared or Fisher's two-sided exact test was performed to compare the MRI features between PP and PA. Logistic stepwise regression analysis and the area under the receiver operating characteristic curve (AUC) were performed to select the optimal features for differentiating PP from PA. Significant differences were found in nine MRI features between women with PP and those with PA (P <0.05). Placental bulge type-II and uterine serosal hypervascularity were independently associated with PP (odds ratio = 48.618, P < 0.001; odds ratio = 4.165, P = 0.018 respectively), and the combination of the two MRI features to distinguish PP from PA yielded an AUC of 0.92 for its predictive performance. Placental bulge type-II and uterine serosal hypervascularity are useful MRI features for differentiating PP from PA. • Placental bulge type-II demonstrated the strongest independent association with PP. • Uterine serosal hypervascularity is a useful feature for differentiating PP from PA. • MRI features associated with abnormal vessels increase the risk of massive haemorrhage.

  19. Quantification of CT images for the classification of high- and low-risk pancreatic cysts

    NASA Astrophysics Data System (ADS)

    Gazit, Lior; Chakraborty, Jayasree; Attiyeh, Marc; Langdon-Embry, Liana; Allen, Peter J.; Do, Richard K. G.; Simpson, Amber L.

    2017-03-01

    Pancreatic cancer is the most lethal cancer with an overall 5-year survival rate of 7%1 due to the late stage at diagnosis and the ineffectiveness of current therapeutic strategies. Given the poor prognosis, early detection at a pre-cancerous stage is the best tool for preventing this disease. Intraductal papillary mucinous neoplasms (IPMN), cystic tumors of the pancreas, represent the only radiographically identifiable precursor lesion of pancreatic cancer and are known to evolve stepwise from low-to-high-grade dysplasia before progressing into an invasive carcinoma. Observation is usually recommended for low-risk (low- and intermediate-grade dysplasia) patients, while high-risk (high-grade dysplasia and invasive carcinoma) patients undergo resection; hence, patient selection is critically important in the management of pancreatic cysts.2 Radiologists use standard criteria such as main pancreatic duct size, cyst size, or presence of a solid enhancing component in the cyst to optimally select patients for surgery.3 However, these findings are subject to a radiologist's interpretation and have been shown to be inconsistent with regards to the presence of a mural nodule or solid component.4 We propose objective classification of risk groups based on quantitative imaging features extracted from CT scans. We apply new features that represent the solid component (i.e. areas of high intensity) within the cyst and extract standard texture features. An adaptive boost classifier5 achieves the best performance with area under receiver operating characteristic curve (AUC) of 0.73 and accuracy of 77.3% for texture features. The random forest classifier achieves the best performance with AUC of 0.71 and accuracy of 70.8% with the solid component features.

  20. Using discrete choice modeling to generate resource selection functions for female polar bears in the Beaufort Sea

    USGS Publications Warehouse

    Durner, George M.; Amstrup, Steven C.; Nielson, Ryan M.; McDonald, Trent; Huzurbazar, Snehalata

    2004-01-01

    Polar bears (Ursus maritimus) depend on ice-covered seas to satisfy life history requirements. Modern threats to polar bears include oil spills in the marine environment and changes in ice composition resulting from climate change. Managers need practical models that explain the distribution of bears in order to assess the impacts of these threats. We explored the use of discrete choice models to describe habitat selection by female polar bears in the Beaufort Sea. Using stepwise procedures we generated resource selection models of habitat use. Sea ice characteristics and ocean depths at known polar bear locations were compared to the same features at randomly selected locations. Models generated for each of four seasons confirmed complexities of habitat use by polar bears and their response to numerous factors. Bears preferred shallow water areas where different ice types intersected. Variation among seasons was reflected mainly in differential selection of total ice concentration, ice stages, floe sizes, and their interactions. Distance to the nearest ice interface was a significant term in models for three seasons. Water depth was selected as a significant term in all seasons, possibly reflecting higher productivity in shallow water areas. Preliminary tests indicate seasonal models can predict polar bear distribution based on prior sea ice data.

  1. Exploring nonlinear feature space dimension reduction and data representation in breast Cadx with Laplacian eigenmaps and t-SNE.

    PubMed

    Jamieson, Andrew R; Giger, Maryellen L; Drukker, Karen; Li, Hui; Yuan, Yading; Bhooshan, Neha

    2010-01-01

    In this preliminary study, recently developed unsupervised nonlinear dimension reduction (DR) and data representation techniques were applied to computer-extracted breast lesion feature spaces across three separate imaging modalities: Ultrasound (U.S.) with 1126 cases, dynamic contrast enhanced magnetic resonance imaging with 356 cases, and full-field digital mammography with 245 cases. Two methods for nonlinear DR were explored: Laplacian eigenmaps [M. Belkin and P. Niyogi, "Laplacian eigenmaps for dimensionality reduction and data representation," Neural Comput. 15, 1373-1396 (2003)] and t-distributed stochastic neighbor embedding (t-SNE) [L. van der Maaten and G. Hinton, "Visualizing data using t-SNE," J. Mach. Learn. Res. 9, 2579-2605 (2008)]. These methods attempt to map originally high dimensional feature spaces to more human interpretable lower dimensional spaces while preserving both local and global information. The properties of these methods as applied to breast computer-aided diagnosis (CADx) were evaluated in the context of malignancy classification performance as well as in the visual inspection of the sparseness within the two-dimensional and three-dimensional mappings. Classification performance was estimated by using the reduced dimension mapped feature output as input into both linear and nonlinear classifiers: Markov chain Monte Carlo based Bayesian artificial neural network (MCMC-BANN) and linear discriminant analysis. The new techniques were compared to previously developed breast CADx methodologies, including automatic relevance determination and linear stepwise (LSW) feature selection, as well as a linear DR method based on principal component analysis. Using ROC analysis and 0.632+bootstrap validation, 95% empirical confidence intervals were computed for the each classifier's AUC performance. In the large U.S. data set, sample high performance results include, AUC0.632+ = 0.88 with 95% empirical bootstrap interval [0.787;0.895] for 13 ARD selected features and AUC0.632+ = 0.87 with interval [0.817;0.906] for four LSW selected features compared to 4D t-SNE mapping (from the original 81D feature space) giving AUC0.632+ = 0.90 with interval [0.847;0.919], all using the MCMC-BANN. Preliminary results appear to indicate capability for the new methods to match or exceed classification performance of current advanced breast lesion CADx algorithms. While not appropriate as a complete replacement of feature selection in CADx problems, DR techniques offer a complementary approach, which can aid elucidation of additional properties associated with the data. Specifically, the new techniques were shown to possess the added benefit of delivering sparse lower dimensional representations for visual interpretation, revealing intricate data structure of the feature space.

  2. Breast-Lesion Characterization using Textural Features of Quantitative Ultrasound Parametric Maps.

    PubMed

    Sadeghi-Naini, Ali; Suraweera, Harini; Tran, William Tyler; Hadizad, Farnoosh; Bruni, Giancarlo; Rastegar, Rashin Fallah; Curpen, Belinda; Czarnota, Gregory J

    2017-10-20

    This study evaluated, for the first time, the efficacy of quantitative ultrasound (QUS) spectral parametric maps in conjunction with texture-analysis techniques to differentiate non-invasively benign versus malignant breast lesions. Ultrasound B-mode images and radiofrequency data were acquired from 78 patients with suspicious breast lesions. QUS spectral-analysis techniques were performed on radiofrequency data to generate parametric maps of mid-band fit, spectral slope, spectral intercept, spacing among scatterers, average scatterer diameter, and average acoustic concentration. Texture-analysis techniques were applied to determine imaging biomarkers consisting of mean, contrast, correlation, energy and homogeneity features of parametric maps. These biomarkers were utilized to classify benign versus malignant lesions with leave-one-patient-out cross-validation. Results were compared to histopathology findings from biopsy specimens and radiology reports on MR images to evaluate the accuracy of technique. Among the biomarkers investigated, one mean-value parameter and 14 textural features demonstrated statistically significant differences (p < 0.05) between the two lesion types. A hybrid biomarker developed using a stepwise feature selection method could classify the legions with a sensitivity of 96%, a specificity of 84%, and an AUC of 0.97. Findings from this study pave the way towards adapting novel QUS-based frameworks for breast cancer screening and rapid diagnosis in clinic.

  3. Constructive neutral evolution: exploring evolutionary theory's curious disconnect.

    PubMed

    Stoltzfus, Arlin

    2012-10-13

    Constructive neutral evolution (CNE) suggests that neutral evolution may follow a stepwise path to extravagance. Whether or not CNE is common, the mere possibility raises provocative questions about causation: in classical neo-Darwinian thinking, selection is the sole source of creativity and direction, the only force that can cause trends or build complex features. However, much of contemporary evolutionary genetics departs from the conception of evolution underlying neo-Darwinism, resulting in a widening gap between what formal models allow, and what the prevailing view of the causes of evolution suggests. In particular, a mutationist conception of evolution as a 2-step origin-fixation process has been a source of theoretical innovation for 40 years, appearing not only in the Neutral Theory, but also in recent breakthroughs in modeling adaptation (the "mutational landscape" model), and in practical software for sequence analysis. In this conception, mutation is not a source of raw materials, but an agent that introduces novelty, while selection is not an agent that shapes features, but a stochastic sieve. This view, which now lays claim to important theoretical, experimental, and practical results, demands our attention. CNE provides a way to explore its most significant implications about the role of variation in evolution. Alex Kondrashov, Eugene Koonin and Johann Peter Gogarten reviewed this article.

  4. Constructive neutral evolution: exploring evolutionary theory’s curious disconnect

    PubMed Central

    2012-01-01

    Abstract Constructive neutral evolution (CNE) suggests that neutral evolution may follow a stepwise path to extravagance. Whether or not CNE is common, the mere possibility raises provocative questions about causation: in classical neo-Darwinian thinking, selection is the sole source of creativity and direction, the only force that can cause trends or build complex features. However, much of contemporary evolutionary genetics departs from the conception of evolution underlying neo-Darwinism, resulting in a widening gap between what formal models allow, and what the prevailing view of the causes of evolution suggests. In particular, a mutationist conception of evolution as a 2-step origin-fixation process has been a source of theoretical innovation for 40 years, appearing not only in the Neutral Theory, but also in recent breakthroughs in modeling adaptation (the “mutational landscape” model), and in practical software for sequence analysis. In this conception, mutation is not a source of raw materials, but an agent that introduces novelty, while selection is not an agent that shapes features, but a stochastic sieve. This view, which now lays claim to important theoretical, experimental, and practical results, demands our attention. CNE provides a way to explore its most significant implications about the role of variation in evolution. Reviewers Alex Kondrashov, Eugene Koonin and Johann Peter Gogarten reviewed this article. PMID:23062217

  5. Order Selection for General Expression of Nonlinear Autoregressive Model Based on Multivariate Stepwise Regression

    NASA Astrophysics Data System (ADS)

    Shi, Jinfei; Zhu, Songqing; Chen, Ruwen

    2017-12-01

    An order selection method based on multiple stepwise regressions is proposed for General Expression of Nonlinear Autoregressive model which converts the model order problem into the variable selection of multiple linear regression equation. The partial autocorrelation function is adopted to define the linear term in GNAR model. The result is set as the initial model, and then the nonlinear terms are introduced gradually. Statistics are chosen to study the improvements of both the new introduced and originally existed variables for the model characteristics, which are adopted to determine the model variables to retain or eliminate. So the optimal model is obtained through data fitting effect measurement or significance test. The simulation and classic time-series data experiment results show that the method proposed is simple, reliable and can be applied to practical engineering.

  6. Prediction of near-term breast cancer risk using local region-based bilateral asymmetry features in mammography

    NASA Astrophysics Data System (ADS)

    Li, Yane; Fan, Ming; Li, Lihua; Zheng, Bin

    2017-03-01

    This study proposed a near-term breast cancer risk assessment model based on local region bilateral asymmetry features in Mammography. The database includes 566 cases who underwent at least two sequential FFDM examinations. The `prior' examination in the two series all interpreted as negative (not recalled). In the "current" examination, 283 women were diagnosed cancers and 283 remained negative. Age of cancers and negative cases completely matched. These cases were divided into three subgroups according to age: 152 cases among the 37-49 age-bracket, 220 cases in the age-bracket 50- 60, and 194 cases with the 61-86 age-bracket. For each image, two local regions including strip-based regions and difference-of-Gaussian basic element regions were segmented. After that, structural variation features among pixel values and structural similarity features were computed for strip regions. Meanwhile, positional features were extracted for basic element regions. The absolute subtraction value was computed between each feature of the left and right local-regions. Next, a multi-layer perception classifier was implemented to assess performance of features for prediction. Features were then selected according stepwise regression analysis. The AUC achieved 0.72, 0.75 and 0.71 for these 3 age-based subgroups, respectively. The maximum adjustable odds ratios were 12.4, 20.56 and 4.91 for these three groups, respectively. This study demonstrate that the local region-based bilateral asymmetry features extracted from CC-view mammography could provide useful information to predict near-term breast cancer risk.

  7. Hyperspectral Features of Oil-Polluted Sea Ice and the Response to the Contamination Area Fraction

    PubMed Central

    Li, Ying; Liu, Chengyu; Xie, Feng

    2018-01-01

    Researchers have studied oil spills in open waters using remote sensors, but few have focused on extracting reflectance features of oil pollution on sea ice. An experiment was conducted on natural sea ice in Bohai Bay, China, to obtain the spectral reflectance of oil-contaminated sea ice. The spectral absorption index (SAI), spectral peak height (SPH), and wavelet detail coefficient (DWT d5) were calculated using stepwise multiple linear regression. The reflectances of some false targets were measured and analysed. The simulated false targets were sediment, iron ore fines, coal dust, and the melt pool. The measured reflectances were resampled using five common sensors (GF-2, Landsat8-OLI, Sentinel3-OLCI, MODIS, and AVIRIS). Some significant spectral features could discriminate between oil-polluted and clean sea ice. The indices correlated well with the oil area fractions. All of the adjusted R2 values exceeded 0.9. The SPH model1, based on spectral features at 507–670 and 1627–1746 nm, displayed the best fitting. The resampled data indicated that these multi-spectral and hyper-spectral sensors could be used to detect crude oil on the sea ice if the effect of noise and spatial resolution are neglected. The spectral features and their identified changes may provide reference on sensor design and band selection. PMID:29342945

  8. A high surface area Zr(IV)-based metal–organic framework showing stepwise gas adsorption and selective dye uptake

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Lv, Xiu-Liang; Tong, Minman; Huang, Hongliang

    2015-03-15

    Exploitation of new metal–organic framework (MOF) materials with high surface areas has been attracting great attention in related research communities due to their broad potential applications. In this work, a new Zr(IV)-based MOF, [Zr{sub 6}O{sub 4}(OH){sub 4}(eddb){sub 6}] (BUT-30, H{sub 2}eddb=4,4′-(ethyne-1,2-diyl)dibenzoic acid) has been solvothermally synthesized, characterized, and explored for gases and dyes adsorptions. Single-crystal X-ray diffraction analysis demonstrates a three-dimensional cubic framework structure of this MOF, in which each Zr{sub 6}O{sub 4}(OH){sub 4} building unit is linked by 12 linear eddb ligands. BUT-30 has been found stable up to 400 °C and has a Brunauer–Emmett–Teller (BET) surface area asmore » high as 3940.6 m{sup 2} g{sup −1} (based on the N{sub 2} adsorption at 77 K) and total pore volume of 1.55 cm{sup 3} g{sup −1}. It is more interesting that this MOF exhibits stepwise adsorption behaviors for Ar, N{sub 2}, and CO{sub 2} at low temperatures, and selective uptakes towards different ionic dyes. - Graphical abstract: A new Zr(IV)-based MOF with high surface area has been synthesized and structurally characterized, which shows stepwise gas adsorption at low temperature and selective dye uptake from solution. - Highlights: • A new Zr-based MOF was synthesized and structurally characterized. • This MOF shows a higher surface area compared with its analogous UiO-67 and 68. • This MOF shows a rare stepwise adsorption towards light gases at low temperature. • This MOF performs selective uptakes towards cationic dyes over anionic ones. • Using triple-bond spacer is confirmed feasible in enhancing MOF surface areas.« less

  9. Shuffling cross-validation-bee algorithm as a new descriptor selection method for retention studies of pesticides in biopartitioning micellar chromatography.

    PubMed

    Zarei, Kobra; Atabati, Morteza; Ahmadi, Monire

    2017-05-04

    Bee algorithm (BA) is an optimization algorithm inspired by the natural foraging behaviour of honey bees to find the optimal solution which can be proposed to feature selection. In this paper, shuffling cross-validation-BA (CV-BA) was applied to select the best descriptors that could describe the retention factor (log k) in the biopartitioning micellar chromatography (BMC) of 79 heterogeneous pesticides. Six descriptors were obtained using BA and then the selected descriptors were applied for model development using multiple linear regression (MLR). The descriptor selection was also performed using stepwise, genetic algorithm and simulated annealing methods and MLR was applied to model development and then the results were compared with those obtained from shuffling CV-BA. The results showed that shuffling CV-BA can be applied as a powerful descriptor selection method. Support vector machine (SVM) was also applied for model development using six selected descriptors by BA. The obtained statistical results using SVM were better than those obtained using MLR, as the root mean square error (RMSE) and correlation coefficient (R) for whole data set (training and test), using shuffling CV-BA-MLR, were obtained as 0.1863 and 0.9426, respectively, while these amounts for the shuffling CV-BA-SVM method were obtained as 0.0704 and 0.9922, respectively.

  10. Stochastic model search with binary outcomes for genome-wide association studies.

    PubMed

    Russu, Alberto; Malovini, Alberto; Puca, Annibale A; Bellazzi, Riccardo

    2012-06-01

    The spread of case-control genome-wide association studies (GWASs) has stimulated the development of new variable selection methods and predictive models. We introduce a novel Bayesian model search algorithm, Binary Outcome Stochastic Search (BOSS), which addresses the model selection problem when the number of predictors far exceeds the number of binary responses. Our method is based on a latent variable model that links the observed outcomes to the underlying genetic variables. A Markov Chain Monte Carlo approach is used for model search and to evaluate the posterior probability of each predictor. BOSS is compared with three established methods (stepwise regression, logistic lasso, and elastic net) in a simulated benchmark. Two real case studies are also investigated: a GWAS on the genetic bases of longevity, and the type 2 diabetes study from the Wellcome Trust Case Control Consortium. Simulations show that BOSS achieves higher precisions than the reference methods while preserving good recall rates. In both experimental studies, BOSS successfully detects genetic polymorphisms previously reported to be associated with the analyzed phenotypes. BOSS outperforms the other methods in terms of F-measure on simulated data. In the two real studies, BOSS successfully detects biologically relevant features, some of which are missed by univariate analysis and the three reference techniques. The proposed algorithm is an advance in the methodology for model selection with a large number of features. Our simulated and experimental results showed that BOSS proves effective in detecting relevant markers while providing a parsimonious model.

  11. A method for the selection of a functional form for a thermodynamic equation of state using weighted linear least squares stepwise regression

    NASA Technical Reports Server (NTRS)

    Jacobsen, R. T.; Stewart, R. B.; Crain, R. W., Jr.; Rose, G. L.; Myers, A. F.

    1976-01-01

    A method was developed for establishing a rational choice of the terms to be included in an equation of state with a large number of adjustable coefficients. The methods presented were developed for use in the determination of an equation of state for oxygen and nitrogen. However, a general application of the methods is possible in studies involving the determination of an optimum polynomial equation for fitting a large number of data points. The data considered in the least squares problem are experimental thermodynamic pressure-density-temperature data. Attention is given to a description of stepwise multiple regression and the use of stepwise regression in the determination of an equation of state for oxygen and nitrogen.

  12. Laguerre-based method for analysis of time-resolved fluorescence data: application to in-vivo characterization and diagnosis of atherosclerotic lesions.

    PubMed

    Jo, Javier A; Fang, Qiyin; Papaioannou, Thanassis; Baker, J Dennis; Dorafshar, Amir H; Reil, Todd; Qiao, Jian-Hua; Fishbein, Michael C; Freischlag, Julie A; Marcu, Laura

    2006-01-01

    We report the application of the Laguerre deconvolution technique (LDT) to the analysis of in-vivo time-resolved laser-induced fluorescence spectroscopy (TR-LIFS) data and the diagnosis of atherosclerotic plaques. TR-LIFS measurements were obtained in vivo from normal and atherosclerotic aortas (eight rabbits, 73 areas), and subsequently analyzed using LDT. Spectral and time-resolved features were used to develop four classification algorithms: linear discriminant analysis (LDA), stepwise LDA (SLDA), principal component analysis (PCA), and artificial neural network (ANN). Accurate deconvolution of TR-LIFS in-vivo measurements from normal and atherosclerotic arteries was provided by LDT. The derived Laguerre expansion coefficients reflected changes in the arterial biochemical composition, and provided a means to discriminate lesions rich in macrophages with high sensitivity (>85%) and specificity (>95%). Classification algorithms (SLDA and PCA) using a selected number of features with maximum discriminating power provided the best performance. This study demonstrates the potential of the LDT for in-vivo tissue diagnosis, and specifically for the detection of macrophages infiltration in atherosclerotic lesions, a key marker of plaque vulnerability.

  13. Laguerre-based method for analysis of time-resolved fluorescence data: application to in-vivo characterization and diagnosis of atherosclerotic lesions

    NASA Astrophysics Data System (ADS)

    Jo, Javier A.; Fang, Qiyin; Papaioannou, Thanassis; Baker, J. Dennis; Dorafshar, Amir; Reil, Todd; Qiao, Jianhua; Fishbein, Michael C.; Freischlag, Julie A.; Marcu, Laura

    2006-03-01

    We report the application of the Laguerre deconvolution technique (LDT) to the analysis of in-vivo time-resolved laser-induced fluorescence spectroscopy (TR-LIFS) data and the diagnosis of atherosclerotic plaques. TR-LIFS measurements were obtained in vivo from normal and atherosclerotic aortas (eight rabbits, 73 areas), and subsequently analyzed using LDT. Spectral and time-resolved features were used to develop four classification algorithms: linear discriminant analysis (LDA), stepwise LDA (SLDA), principal component analysis (PCA), and artificial neural network (ANN). Accurate deconvolution of TR-LIFS in-vivo measurements from normal and atherosclerotic arteries was provided by LDT. The derived Laguerre expansion coefficients reflected changes in the arterial biochemical composition, and provided a means to discriminate lesions rich in macrophages with high sensitivity (>85%) and specificity (>95%). Classification algorithms (SLDA and PCA) using a selected number of features with maximum discriminating power provided the best performance. This study demonstrates the potential of the LDT for in-vivo tissue diagnosis, and specifically for the detection of macrophages infiltration in atherosclerotic lesions, a key marker of plaque vulnerability.

  14. Laguerre-based method for analysis of time-resolved fluorescence data: application to in-vivo characterization and diagnosis of atherosclerotic lesions

    PubMed Central

    Jo, Javier A.; Fang, Qiyin; Papaioannou, Thanassis; Baker, J. Dennis; Dorafshar, Amir H.; Reil, Todd; Qiao, Jian-Hua; Fishbein, Michael C.; Freischlag, Julie A.; Marcu, Laura

    2007-01-01

    We report the application of the Laguerre deconvolution technique (LDT) to the analysis of in-vivo time-resolved laser-induced fluorescence spectroscopy (TR-LIFS) data and the diagnosis of atherosclerotic plaques. TR-LIFS measurements were obtained in vivo from normal and atherosclerotic aortas (eight rabbits, 73 areas), and subsequently analyzed using LDT. Spectral and time-resolved features were used to develop four classification algorithms: linear discriminant analysis (LDA), stepwise LDA (SLDA), principal component analysis (PCA), and artificial neural network (ANN). Accurate deconvolution of TR-LIFS in-vivo measurements from normal and atherosclerotic arteries was provided by LDT. The derived Laguerre expansion coefficients reflected changes in the arterial biochemical composition, and provided a means to discriminate lesions rich in macrophages with high sensitivity (>85%) and specificity (>95%). Classification algorithms (SLDA and PCA) using a selected number of features with maximum discriminating power provided the best performance. This study demonstrates the potential of the LDT for in-vivo tissue diagnosis, and specifically for the detection of macrophages infiltration in atherosclerotic lesions, a key marker of plaque vulnerability. PMID:16674179

  15. Respiratory trace feature analysis for the prediction of respiratory-gated PET quantification.

    PubMed

    Wang, Shouyi; Bowen, Stephen R; Chaovalitwongse, W Art; Sandison, George A; Grabowski, Thomas J; Kinahan, Paul E

    2014-02-21

    The benefits of respiratory gating in quantitative PET/CT vary tremendously between individual patients. Respiratory pattern is among many patient-specific characteristics that are thought to play an important role in gating-induced imaging improvements. However, the quantitative relationship between patient-specific characteristics of respiratory pattern and improvements in quantitative accuracy from respiratory-gated PET/CT has not been well established. If such a relationship could be estimated, then patient-specific respiratory patterns could be used to prospectively select appropriate motion compensation during image acquisition on a per-patient basis. This study was undertaken to develop a novel statistical model that predicts quantitative changes in PET/CT imaging due to respiratory gating. Free-breathing static FDG-PET images without gating and respiratory-gated FDG-PET images were collected from 22 lung and liver cancer patients on a PET/CT scanner. PET imaging quality was quantified with peak standardized uptake value (SUV(peak)) over lesions of interest. Relative differences in SUV(peak) between static and gated PET images were calculated to indicate quantitative imaging changes due to gating. A comprehensive multidimensional extraction of the morphological and statistical characteristics of respiratory patterns was conducted, resulting in 16 features that characterize representative patterns of a single respiratory trace. The six most informative features were subsequently extracted using a stepwise feature selection approach. The multiple-regression model was trained and tested based on a leave-one-subject-out cross-validation. The predicted quantitative improvements in PET imaging achieved an accuracy higher than 90% using a criterion with a dynamic error-tolerance range for SUV(peak) values. The results of this study suggest that our prediction framework could be applied to determine which patients would likely benefit from respiratory motion compensation when clinicians quantitatively assess PET/CT for therapy target definition and response assessment.

  16. Respiratory trace feature analysis for the prediction of respiratory-gated PET quantification

    NASA Astrophysics Data System (ADS)

    Wang, Shouyi; Bowen, Stephen R.; Chaovalitwongse, W. Art; Sandison, George A.; Grabowski, Thomas J.; Kinahan, Paul E.

    2014-02-01

    The benefits of respiratory gating in quantitative PET/CT vary tremendously between individual patients. Respiratory pattern is among many patient-specific characteristics that are thought to play an important role in gating-induced imaging improvements. However, the quantitative relationship between patient-specific characteristics of respiratory pattern and improvements in quantitative accuracy from respiratory-gated PET/CT has not been well established. If such a relationship could be estimated, then patient-specific respiratory patterns could be used to prospectively select appropriate motion compensation during image acquisition on a per-patient basis. This study was undertaken to develop a novel statistical model that predicts quantitative changes in PET/CT imaging due to respiratory gating. Free-breathing static FDG-PET images without gating and respiratory-gated FDG-PET images were collected from 22 lung and liver cancer patients on a PET/CT scanner. PET imaging quality was quantified with peak standardized uptake value (SUVpeak) over lesions of interest. Relative differences in SUVpeak between static and gated PET images were calculated to indicate quantitative imaging changes due to gating. A comprehensive multidimensional extraction of the morphological and statistical characteristics of respiratory patterns was conducted, resulting in 16 features that characterize representative patterns of a single respiratory trace. The six most informative features were subsequently extracted using a stepwise feature selection approach. The multiple-regression model was trained and tested based on a leave-one-subject-out cross-validation. The predicted quantitative improvements in PET imaging achieved an accuracy higher than 90% using a criterion with a dynamic error-tolerance range for SUVpeak values. The results of this study suggest that our prediction framework could be applied to determine which patients would likely benefit from respiratory motion compensation when clinicians quantitatively assess PET/CT for therapy target definition and response assessment.

  17. [Clinical and anatomical features of congenital microphthalmia and anophthalmia in children and conservative methods of rehabilitation].

    PubMed

    Sudovskaya, T V; Filatova, I A; Kiseleva, T N; Bobrovskaya, Yu A; Kokoeva, N Sh

    2016-01-01

    To develop a comprehensive classification system of distinctive clinical and anatomical features of congenital microphthalmia and anophthalmia in children and to specify indications, contraindications, and optimal timing of the primary and subsequent prosthetic treatment. A total of 70 patients with congenital micro- or anophthalmia aged from 1 month to 12 years were examined. Besides the routine ophthalmic examination, all patients underwent eye and orbit ultrasound (axial length measurement and B-scan), computed tomography of the orbits and skull, and immunological tests for infectious diseases (enzyme-linked immunosorbent assays). Basing on the examination we have determined the common types of congenital micro- and anophthalmia in children. We have also developed a stepwise prosthetic treatment aimed at better cosmetic rehabilitation. Indications and contraindications for the use of ocular prostheses in children with congenital micro- and anophthalmia have been identified. The proposed method of stepwise prosthetics is the principal option for conservative rehabilitation of children with congenital micro- or anophthalmia.

  18. Predicting pork loin intramuscular fat using computer vision system.

    PubMed

    Liu, J-H; Sun, X; Young, J M; Bachmeier, L A; Newman, D J

    2018-09-01

    The objective of this study was to investigate the ability of computer vision system to predict pork intramuscular fat percentage (IMF%). Center-cut loin samples (n = 85) were trimmed of subcutaneous fat and connective tissue. Images were acquired and pixels were segregated to estimate image IMF% and 18 image color features for each image. Subjective IMF% was determined by a trained grader. Ether extract IMF% was calculated using ether extract method. Image color features and image IMF% were used as predictors for stepwise regression and support vector machine models. Results showed that subjective IMF% had a correlation of 0.81 with ether extract IMF% while the image IMF% had a 0.66 correlation with ether extract IMF%. Accuracy rates for regression models were 0.63 for stepwise and 0.75 for support vector machine. Although subjective IMF% has shown to have better prediction, results from computer vision system demonstrates the potential of being used as a tool in predicting pork IMF% in the future. Copyright © 2018 Elsevier Ltd. All rights reserved.

  19. Assessing the accuracy and stability of variable selection ...

    EPA Pesticide Factsheets

    Random forest (RF) modeling has emerged as an important statistical learning method in ecology due to its exceptional predictive performance. However, for large and complex ecological datasets there is limited guidance on variable selection methods for RF modeling. Typically, either a preselected set of predictor variables are used, or stepwise procedures are employed which iteratively add/remove variables according to their importance measures. This paper investigates the application of variable selection methods to RF models for predicting probable biological stream condition. Our motivating dataset consists of the good/poor condition of n=1365 stream survey sites from the 2008/2009 National Rivers and Stream Assessment, and a large set (p=212) of landscape features from the StreamCat dataset. Two types of RF models are compared: a full variable set model with all 212 predictors, and a reduced variable set model selected using a backwards elimination approach. We assess model accuracy using RF's internal out-of-bag estimate, and a cross-validation procedure with validation folds external to the variable selection process. We also assess the stability of the spatial predictions generated by the RF models to changes in the number of predictors, and argue that model selection needs to consider both accuracy and stability. The results suggest that RF modeling is robust to the inclusion of many variables of moderate to low importance. We found no substanti

  20. Computer-aided classification of breast microcalcification clusters: merging of features from image processing and radiologists

    NASA Astrophysics Data System (ADS)

    Lo, Joseph Y.; Gavrielides, Marios A.; Markey, Mia K.; Jesneck, Jonathan L.

    2003-05-01

    We developed an ensemble classifier for the task of computer-aided diagnosis of breast microcalcification clusters,which are very challenging to characterize for radiologists and computer models alike. The purpose of this study is to help radiologists identify whether suspicious calcification clusters are benign vs. malignant, such that they may potentially recommend fewer unnecessary biopsies for actually benign lesions. The data consists of mammographic features extracted by automated image processing algorithms as well as manually interpreted by radiologists according to a standardized lexicon. We used 292 cases from a publicly available mammography database. From each cases, we extracted 22 image processing features pertaining to lesion morphology, 5 radiologist features also pertaining to morphology, and the patient age. Linear discriminant analysis (LDA) models were designed using each of the three data types. Each local model performed poorly; the best was one based upon image processing features which yielded ROC area index AZ of 0.59 +/- 0.03 and partial AZ above 90% sensitivity of 0.08 +/- 0.03. We then developed ensemble models using different combinations of those data types, and these models all improved performance compared to the local models. The final ensemble model was based upon 5 features selected by stepwise LDA from all 28 available features. This ensemble performed with AZ of 0.69 +/- 0.03 and partial AZ of 0.21 +/- 0.04, which was statistically significantly better than the model based on the image processing features alone (p<0.001 and p=0.01 for full and partial AZ respectively). This demonstrated the value of the radiologist-extracted features as a source of information for this task. It also suggested there is potential for improved performance using this ensemble classifier approach to combine different sources of currently available data.

  1. Exploring nonlinear feature space dimension reduction and data representation in breast CADx with Laplacian eigenmaps and t-SNE

    PubMed Central

    Jamieson, Andrew R.; Giger, Maryellen L.; Drukker, Karen; Li, Hui; Yuan, Yading; Bhooshan, Neha

    2010-01-01

    Purpose: In this preliminary study, recently developed unsupervised nonlinear dimension reduction (DR) and data representation techniques were applied to computer-extracted breast lesion feature spaces across three separate imaging modalities: Ultrasound (U.S.) with 1126 cases, dynamic contrast enhanced magnetic resonance imaging with 356 cases, and full-field digital mammography with 245 cases. Two methods for nonlinear DR were explored: Laplacian eigenmaps [M. Belkin and P. Niyogi, “Laplacian eigenmaps for dimensionality reduction and data representation,” Neural Comput. 15, 1373–1396 (2003)] and t-distributed stochastic neighbor embedding (t-SNE) [L. van der Maaten and G. Hinton, “Visualizing data using t-SNE,” J. Mach. Learn. Res. 9, 2579–2605 (2008)]. Methods: These methods attempt to map originally high dimensional feature spaces to more human interpretable lower dimensional spaces while preserving both local and global information. The properties of these methods as applied to breast computer-aided diagnosis (CADx) were evaluated in the context of malignancy classification performance as well as in the visual inspection of the sparseness within the two-dimensional and three-dimensional mappings. Classification performance was estimated by using the reduced dimension mapped feature output as input into both linear and nonlinear classifiers: Markov chain Monte Carlo based Bayesian artificial neural network (MCMC-BANN) and linear discriminant analysis. The new techniques were compared to previously developed breast CADx methodologies, including automatic relevance determination and linear stepwise (LSW) feature selection, as well as a linear DR method based on principal component analysis. Using ROC analysis and 0.632+bootstrap validation, 95% empirical confidence intervals were computed for the each classifier’s AUC performance. Results: In the large U.S. data set, sample high performance results include, AUC0.632+=0.88 with 95% empirical bootstrap interval [0.787;0.895] for 13 ARD selected features and AUC0.632+=0.87 with interval [0.817;0.906] for four LSW selected features compared to 4D t-SNE mapping (from the original 81D feature space) giving AUC0.632+=0.90 with interval [0.847;0.919], all using the MCMC-BANN. Conclusions: Preliminary results appear to indicate capability for the new methods to match or exceed classification performance of current advanced breast lesion CADx algorithms. While not appropriate as a complete replacement of feature selection in CADx problems, DR techniques offer a complementary approach, which can aid elucidation of additional properties associated with the data. Specifically, the new techniques were shown to possess the added benefit of delivering sparse lower dimensional representations for visual interpretation, revealing intricate data structure of the feature space. PMID:20175497

  2. Inertial Response of Wind Power Plants: A Comparison of Frequency-Based Inertial Control and Stepwise Inertial Control

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Wang, Xiao; Gao, Wenzhong; Wang, Jianhui

    The frequency regulation capability of a wind power plant plays an important role in enhancing frequency reliability especially in an isolated power system with high wind power penetration levels. A comparison of two types of inertial control methods, namely frequency-based inertial control (FBIC) and stepwise inertial control (SIC), is presented in this paper. Comprehensive case studies are carried out to reveal features of the different inertial control methods, simulated in a modified Western System Coordination Council (WSCC) nine-bus power grid using real-time digital simulator (RTDS) platform. The simulation results provide an insight into the inertial control methods under various scenarios.

  3. [Discriminant analysis to predict the clinical diagnosis of primary immunodeficiencies: a preliminary report].

    PubMed

    Murata, Chiharu; Ramírez, Ana Belén; Ramírez, Guadalupe; Cruz, Alonso; Morales, José Luis; Lugo-Reyes, Saul Oswaldo

    2015-01-01

    The features in a clinical history from a patient with suspected primary immunodeficiency (PID) direct the differential diagnosis through pattern recognition. PIDs are a heterogeneous group of more than 250 congenital diseases with increased susceptibility to infection, inflammation, autoimmunity, allergy and malignancy. Linear discriminant analysis (LDA) is a multivariate supervised classification method to sort objects of study into groups by finding linear combinations of a number of variables. To identify the features that best explain membership of pediatric PID patients to a group of defect or disease. An analytic cross-sectional study was done with a pre-existing database with clinical and laboratory records from 168 patients with PID, followed at the National Institute of Pediatrics during 1991-2012, it was used to build linear discriminant models that would explain membership of each patient to the different group defects and to the most prevalent PIDs in our registry. After a preliminary run only 30 features were included (4 demographic, 10 clinical, 10 laboratory, 6 germs), with which the training models were developed through a stepwise regression algorithm. We compared the automatic feature selection with a selection made by a human expert, and then assessed the diagnostic usefulness of the resulting models (sensitivity, specificity, prediction accuracy and kappa coefficient), with 95% confidence intervals. The models incorporated 6 to 14 features to explain membership of PID patients to the five most abundant defect groups (combined, antibody, well-defined, dysregulation and phagocytosis), and to the four most prevalent PID diseases (X-linked agammaglobulinemia, chronic granulomatous disease, common variable immunodeficiency and ataxiatelangiectasia). In practically all cases of feature selection the machine outperformed the human expert. Diagnosis prediction using the equations created had a global accuracy of 83 to 94%, with sensitivity of 60 to 100%, specificity of 83 to 95% and kappa coefficient of 0.37 to 0.76. In general, the selection of features has clinical plausibility, and the practical advantage of utilizing only clinical attributes, infecting germs and routine lab results (blood cell counts and serum immunoglobulins). The performance of the model as a diagnostic tool was acceptable. The study's main limitations are a limited sample size and a lack of cross validation. This is only the first step in the construction of a machine learning system, with a wider approach that includes a larger database and different methodologies, to assist the clinical diagnosis of primary immunodeficiencies.

  4. Spectroscopic determination of leaf biochemistry using band-depth analysis of absorption features and stepwise multiple linear regression

    USGS Publications Warehouse

    Kokaly, R.F.; Clark, R.N.

    1999-01-01

    We develop a new method for estimating the biochemistry of plant material using spectroscopy. Normalized band depths calculated from the continuum-removed reflectance spectra of dried and ground leaves were used to estimate their concentrations of nitrogen, lignin, and cellulose. Stepwise multiple linear regression was used to select wavelengths in the broad absorption features centered at 1.73 ??m, 2.10 ??m, and 2.30 ??m that were highly correlated with the chemistry of samples from eastern U.S. forests. Band depths of absorption features at these wavelengths were found to also be highly correlated with the chemistry of four other sites. A subset of data from the eastern U.S. forest sites was used to derive linear equations that were applied to the remaining data to successfully estimate their nitrogen, lignin, and cellulose concentrations. Correlations were highest for nitrogen (R2 from 0.75 to 0.94). The consistent results indicate the possibility of establishing a single equation capable of estimating the chemical concentrations in a wide variety of species from the reflectance spectra of dried leaves. The extension of this method to remote sensing was investigated. The effects of leaf water content, sensor signal-to-noise and bandpass, atmospheric effects, and background soil exposure were examined. Leaf water was found to be the greatest challenge to extending this empirical method to the analysis of fresh whole leaves and complete vegetation canopies. The influence of leaf water on reflectance spectra must be removed to within 10%. Other effects were reduced by continuum removal and normalization of band depths. If the effects of leaf water can be compensated for, it might be possible to extend this method to remote sensing data acquired by imaging spectrometers to give estimates of nitrogen, lignin, and cellulose concentrations over large areas for use in ecosystem studies.We develop a new method for estimating the biochemistry of plant material using spectroscopy. Normalized band depths calculated from the continuum-removed reflectance spectra of dried and ground leaves were used to estimate their concentrations of nitrogen, lignin, and cellulose. Stepwise multiple linear regression was used to select wavelengths in the broad absorption features centered at 1.73 ??m, 2.10 ??m, and 2.301 ??m that were highly correlated with the chemistry of samples from eastern U.S. forests. Band depths of absorption features at these wavelengths were found to also be highly correlated with the chemistry of four other sites. A subset of data from the eastern U.S. forest sites was used to derive linear equations that were applied to the remaining data to successfully estimate their nitrogen, lignin, and cellulose concentrations. Correlations were highest for nitrogen (R2 from 0.75 to 0.94). The consistent results indicate the possibility of establishing a single equation capable of estimating the chemical concentrations in a wide variety of species from the reflectance spectra of dried leaves. The extension of this method to remote sensing was investigated. The effects of leaf water content, sensor signal-to-noise and bandpass, atmospheric effects, and background soil exposure were examined. Leaf water was found to be the greatest challenge to extending this empirical method to the analysis of fresh whole leaves and complete vegetation canopies. The influence of leaf water on reflectance spectra must be removed to within 10%. Other effects were reduced by continuum removal and normalization of band depths. If the effects of leaf water can be compensated for, it might be possible to extend this method to remote sensing data acquired by imaging spectrometers to give estimates of nitrogen, lignin, and cellulose concentrations over large areas for use in ecosystem studies.

  5. Preliminary experiments on quantification of skin condition

    NASA Astrophysics Data System (ADS)

    Kitajima, Kenzo; Iyatomi, Hitoshi

    2014-03-01

    In this study, we investigated a preliminary assessment method for skin conditions such as a moisturizing property and its fineness of the skin with an image analysis only. We captured a facial images from volunteer subjects aged between 30s and 60s by Pocket Micro (R) device (Scalar Co., Japan). This device has two image capturing modes; the normal mode and the non-reflection mode with the aid of the equipped polarization filter. We captured skin images from a total of 68 spots from subjects' face using both modes (i.e. total of 136 skin images). The moisture-retaining property of the skin and subjective evaluation score of the skin fineness in 5-point scale for each case were also obtained in advance as a gold standard (their mean and SD were 35.15 +/- 3.22 (μS) and 3.45 +/- 1.17, respectively). We extracted a total of 107 image features from each image and built linear regression models for estimating abovementioned criteria with a stepwise feature selection. The developed model for estimating the skin moisture achieved the MSE of 1.92 (μS) with 6 selected parameters, while the model for skin fineness achieved that of 0.51 scales with 7 parameters under the leave-one-out cross validation. We confirmed the developed models predicted the moisture-retaining property and fineness of the skin appropriately with only captured image.

  6. Near-infrared spectral image analysis of pork marbling based on Gabor filter and wide line detector techniques.

    PubMed

    Huang, Hui; Liu, Li; Ngadi, Michael O; Gariépy, Claude; Prasher, Shiv O

    2014-01-01

    Marbling is an important quality attribute of pork. Detection of pork marbling usually involves subjective scoring, which raises the efficiency costs to the processor. In this study, the ability to predict pork marbling using near-infrared (NIR) hyperspectral imaging (900-1700 nm) and the proper image processing techniques were studied. Near-infrared images were collected from pork after marbling evaluation according to current standard chart from the National Pork Producers Council. Image analysis techniques-Gabor filter, wide line detector, and spectral averaging-were applied to extract texture, line, and spectral features, respectively, from NIR images of pork. Samples were grouped into calibration and validation sets. Wavelength selection was performed on calibration set by stepwise regression procedure. Prediction models of pork marbling scores were built using multiple linear regressions based on derivatives of mean spectra and line features at key wavelengths. The results showed that the derivatives of both texture and spectral features produced good results, with correlation coefficients of validation of 0.90 and 0.86, respectively, using wavelengths of 961, 1186, and 1220 nm. The results revealed the great potential of the Gabor filter for analyzing NIR images of pork for the effective and efficient objective evaluation of pork marbling.

  7. Quantitative structure-activity relationship of the curcumin-related compounds using various regression methods

    NASA Astrophysics Data System (ADS)

    Khazaei, Ardeshir; Sarmasti, Negin; Seyf, Jaber Yousefi

    2016-03-01

    Quantitative structure activity relationship were used to study a series of curcumin-related compounds with inhibitory effect on prostate cancer PC-3 cells, pancreas cancer Panc-1 cells, and colon cancer HT-29 cells. Sphere exclusion method was used to split data set in two categories of train and test set. Multiple linear regression, principal component regression and partial least squares were used as the regression methods. In other hand, to investigate the effect of feature selection methods, stepwise, Genetic algorithm, and simulated annealing were used. In two cases (PC-3 cells and Panc-1 cells), the best models were generated by a combination of multiple linear regression and stepwise (PC-3 cells: r2 = 0.86, q2 = 0.82, pred_r2 = 0.93, and r2m (test) = 0.43, Panc-1 cells: r2 = 0.85, q2 = 0.80, pred_r2 = 0.71, and r2m (test) = 0.68). For the HT-29 cells, principal component regression with stepwise (r2 = 0.69, q2 = 0.62, pred_r2 = 0.54, and r2m (test) = 0.41) is the best method. The QSAR study reveals descriptors which have crucial role in the inhibitory property of curcumin-like compounds. 6ChainCount, T_C_C_1, and T_O_O_7 are the most important descriptors that have the greatest effect. With a specific end goal to design and optimization of novel efficient curcumin-related compounds it is useful to introduce heteroatoms such as nitrogen, oxygen, and sulfur atoms in the chemical structure (reduce the contribution of T_C_C_1 descriptor) and increase the contribution of 6ChainCount and T_O_O_7 descriptors. Models can be useful in the better design of some novel curcumin-related compounds that can be used in the treatment of prostate, pancreas, and colon cancers.

  8. A simple approach to quantitative analysis using three-dimensional spectra based on selected Zernike moments.

    PubMed

    Zhai, Hong Lin; Zhai, Yue Yuan; Li, Pei Zhen; Tian, Yue Li

    2013-01-21

    A very simple approach to quantitative analysis is proposed based on the technology of digital image processing using three-dimensional (3D) spectra obtained by high-performance liquid chromatography coupled with a diode array detector (HPLC-DAD). As the region-based shape features of a grayscale image, Zernike moments with inherently invariance property were employed to establish the linear quantitative models. This approach was applied to the quantitative analysis of three compounds in mixed samples using 3D HPLC-DAD spectra, and three linear models were obtained, respectively. The correlation coefficients (R(2)) for training and test sets were more than 0.999, and the statistical parameters and strict validation supported the reliability of established models. The analytical results suggest that the Zernike moment selected by stepwise regression can be used in the quantitative analysis of target compounds. Our study provides a new idea for quantitative analysis using 3D spectra, which can be extended to the analysis of other 3D spectra obtained by different methods or instruments.

  9. Clinical and cytological features predictive of malignancy in thyroid follicular neoplasms.

    PubMed

    Lubitz, Carrie C; Faquin, William C; Yang, Jingyun; Mekel, Michal; Gaz, Randall D; Parangi, Sareh; Randolph, Gregory W; Hodin, Richard A; Stephen, Antonia E

    2010-01-01

    The preoperative diagnosis of malignancy in nodules suspicious for a follicular neoplasm remains challenging. A number of clinical and cytological parameters have been previously studied; however, none have significantly impacted clinical practice. The aim of this study was to determine predictive characteristics of follicular neoplasms useful for clinical application. Four clinical (age, sex, nodule size, solitary nodule) and 17 cytological variables were retrospectively reviewed for 144 patients with a nodule suspicious for follicular neoplasm, diagnosed preoperatively by fine-needle aspiration (FNA), from a single institution over a 2-year period (January 2006 to December 2007). The FNAs were examined by a single, blinded pathologist and compared with final surgical pathology. Significance of clinical and cytological variables was determined by univariate analysis and backward stepwise logistic regression. Odds ratios (ORs) for malignancy, a receiver operating characteristic curve, and predicted probabilities of combined features were determined. There was an 11% incidence of malignancy (16/144). On univariate analysis, nodule size >OR=4.0 cm nears significance (p = 0.054) and 9 of 17 cytological features examined were significantly associated with malignancy. Three variables stay in the final model after performing backward stepwise selection in logistic regression: nodule size (OR = 0.25, p = 0.05), presence of a transgressing vessel (OR = 23, p < 0.0001), and nuclear grooves (OR = 4.3, p = 0.03). The predicted probability of malignancy was 88.4% with the presence of all three variables on preoperative FNA. When the two papillary carcinomas were excluded from the analysis, the presence of nuclear grooves was no longer significant, and anisokaryosis (OR = 12.74, p = 0.005) and presence of nucleolus (OR = 0.11, p = 0.04) were significantly associated with malignancy. Excluding the two papillary thyroid carcinomas, a nodule size >or=4 cm, with a transgressing vessel and anisokaryosis and lacking a nucleolus, has a predicted probability of malignancy of 96.5%. A combination of larger nodule size, transgressing vessels, and specific nuclear features are predictive of malignancy in patients with follicular neoplasms. These findings enhance our current limited predictive armamentarium and can be used to guide surgical decision making. Further study may result in the inclusion of these variables to the systematic evaluation of follicular neoplasms.

  10. Feature Relevance Assessment of Multispectral Airborne LIDAR Data for Tree Species Classification

    NASA Astrophysics Data System (ADS)

    Amiri, N.; Heurich, M.; Krzystek, P.; Skidmore, A. K.

    2018-04-01

    The presented experiment investigates the potential of Multispectral Laser Scanning (MLS) point clouds for single tree species classification. The basic idea is to simulate a MLS sensor by combining two different Lidar sensors providing three different wavelngthes. The available data were acquired in the summer 2016 at the same date in a leaf-on condition with an average point density of 37 points/m2. For the purpose of classification, we segmented the combined 3D point clouds consisiting of three different spectral channels into 3D clusters using Normalized Cut segmentation approach. Then, we extracted four group of features from the 3D point cloud space. Once a varity of features has been extracted, we applied forward stepwise feature selection in order to reduce the number of irrelevant or redundant features. For the classification, we used multinomial logestic regression with L1 regularization. Our study is conducted using 586 ground measured single trees from 20 sample plots in the Bavarian Forest National Park, in Germany. Due to lack of reference data for some rare species, we focused on four classes of species. The results show an improvement between 4-10 pp for the tree species classification by using MLS data in comparison to a single wavelength based approach. A cross validated (15-fold) accuracy of 0.75 can be achieved when all feature sets from three different spectral channels are used. Our results cleary indicates that the use of MLS point clouds has great potential to improve detailed forest species mapping.

  11. Stochastic model search with binary outcomes for genome-wide association studies

    PubMed Central

    Malovini, Alberto; Puca, Annibale A; Bellazzi, Riccardo

    2012-01-01

    Objective The spread of case–control genome-wide association studies (GWASs) has stimulated the development of new variable selection methods and predictive models. We introduce a novel Bayesian model search algorithm, Binary Outcome Stochastic Search (BOSS), which addresses the model selection problem when the number of predictors far exceeds the number of binary responses. Materials and methods Our method is based on a latent variable model that links the observed outcomes to the underlying genetic variables. A Markov Chain Monte Carlo approach is used for model search and to evaluate the posterior probability of each predictor. Results BOSS is compared with three established methods (stepwise regression, logistic lasso, and elastic net) in a simulated benchmark. Two real case studies are also investigated: a GWAS on the genetic bases of longevity, and the type 2 diabetes study from the Wellcome Trust Case Control Consortium. Simulations show that BOSS achieves higher precisions than the reference methods while preserving good recall rates. In both experimental studies, BOSS successfully detects genetic polymorphisms previously reported to be associated with the analyzed phenotypes. Discussion BOSS outperforms the other methods in terms of F-measure on simulated data. In the two real studies, BOSS successfully detects biologically relevant features, some of which are missed by univariate analysis and the three reference techniques. Conclusion The proposed algorithm is an advance in the methodology for model selection with a large number of features. Our simulated and experimental results showed that BOSS proves effective in detecting relevant markers while providing a parsimonious model. PMID:22534080

  12. Stepwise photochromism of bisnaphthopyrans exhibiting an excitation intensity-dependent color change.

    PubMed

    Inagaki, Yuki; Mutoh, Katsuya; Abe, Jiro

    2018-06-07

    Non-linear photoresponses against excitation light intensity are important for the development of attractive photofunctional materials exhibiting high spatial selective photoswitching that is not affected by weak background light. Biphotochromic systems composed of two fast photochromic units have the potential to show a stepwise two-photon absorption process in which the optical properties can be non-linearly controlled by changing the excitation light conditions. Herein, we designed and synthesized novel bisnaphthopyran derivatives containing fast photoswitchable naphthopyran units. The bisnaphthopyran derivatives show a stepwise two-photon-induced photochromic reaction upon UV light irradiation accompanied by a drastic color change due to a large change in the molecular structure between the one-photon product and the two-photon product. Consequently, the color of the bisnaphthopyran derivatives can be non-linearly controlled by changing the excitation intensity. This characteristic photochromic property of the biphotochromic system provides important insight into advanced photoresponsive materials.

  13. Design and Use of a Low Cost, Automated Morbidostat for Adaptive Evolution of Bacteria Under Antibiotic Drug Selection.

    PubMed

    Liu, Po C; Lee, Yi T; Wang, Chun Y; Yang, Ya-Tang

    2016-09-27

    We describe a low cost, configurable morbidostat for characterizing the evolutionary pathway of antibiotic resistance. The morbidostat is a bacterial culture device that continuously monitors bacterial growth and dynamically adjusts the drug concentration to constantly challenge the bacteria as they evolve to acquire drug resistance. The device features a working volume of ~10 ml and is fully automated and equipped with optical density measurement and micro-pumps for medium and drug delivery. To validate the platform, we measured the stepwise acquisition of trimethoprim resistance in Escherichia coli MG 1655, and integrated the device with a multiplexed microfluidic platform to investigate cell morphology and antibiotic susceptibility. The approach can be up-scaled to laboratory studies of antibiotic drug resistance, and is extendible to adaptive evolution for strain improvements in metabolic engineering and other bacterial culture experiments.

  14. The Sherlock Holmes approach to diagnosing fetal syndromes by ultrasound.

    PubMed

    Benacerraf, Beryl B

    2012-03-01

    Prenatal detection of fetal anomalies is one of the major goals of obstetrical ultrasound. The primary reason is the options that are often offered to the family and caregivers from therapy in selected cases to special care at delivery to termination of the pregnancy. An important aspect of the diagnosis is to determine whether the anomaly is expected to be lethal or associated with severe physical or mental impediments. This goal is often difficult to accomplish without a clear diagnosis. A systematic approach is essential when an abnormality is first identified sonographically to help the practitioner discover certain patterns of associated defects. The use of this logical and stepwise strategy facilitates arriving at the correct diagnosis of specific syndrome by taking all anatomic findings into account. This process focuses on first pinpointing a key or sentinel feature specific to each syndrome and which can anchor the diagnosis.

  15. TH-E-BRF-05: Comparison of Survival-Time Prediction Models After Radiotherapy for High-Grade Glioma Patients Based On Clinical and DVH Features

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Magome, T; Haga, A; Igaki, H

    Purpose: Although many outcome prediction models based on dose-volume information have been proposed, it is well known that the prognosis may be affected also by multiple clinical factors. The purpose of this study is to predict the survival time after radiotherapy for high-grade glioma patients based on features including clinical and dose-volume histogram (DVH) information. Methods: A total of 35 patients with high-grade glioma (oligodendroglioma: 2, anaplastic astrocytoma: 3, glioblastoma: 30) were selected in this study. All patients were treated with prescribed dose of 30–80 Gy after surgical resection or biopsy from 2006 to 2013 at The University of Tokyomore » Hospital. All cases were randomly separated into training dataset (30 cases) and test dataset (5 cases). The survival time after radiotherapy was predicted based on a multiple linear regression analysis and artificial neural network (ANN) by using 204 candidate features. The candidate features included the 12 clinical features (tumor location, extent of surgical resection, treatment duration of radiotherapy, etc.), and the 192 DVH features (maximum dose, minimum dose, D95, V60, etc.). The effective features for the prediction were selected according to a step-wise method by using 30 training cases. The prediction accuracy was evaluated by a coefficient of determination (R{sup 2}) between the predicted and actual survival time for the training and test dataset. Results: In the multiple regression analysis, the value of R{sup 2} between the predicted and actual survival time was 0.460 for the training dataset and 0.375 for the test dataset. On the other hand, in the ANN analysis, the value of R{sup 2} was 0.806 for the training dataset and 0.811 for the test dataset. Conclusion: Although a large number of patients would be needed for more accurate and robust prediction, our preliminary Result showed the potential to predict the outcome in the patients with high-grade glioma. This work was partly supported by the JSPS Core-to-Core Program(No. 23003) and Grant-in-aid from the JSPS Fellows.« less

  16. Ankylosaurid dinosaur tail clubs evolved through stepwise acquisition of key features.

    PubMed

    Arbour, Victoria M; Currie, Philip J

    2015-10-01

    Ankylosaurid ankylosaurs were quadrupedal, herbivorous dinosaurs with abundant dermal ossifications. They are best known for their distinctive tail club composed of stiff, interlocking vertebrae (the handle) and large, bulbous osteoderms (the knob), which may have been used as a weapon. However, tail clubs appear relatively late in the evolution of ankylosaurids, and seemed to have been present only in a derived clade of ankylosaurids during the last 20 million years of the Mesozoic Era. New evidence from mid Cretaceous fossils from China suggests that the evolution of the tail club occurred at least 40 million years earlier, and in a stepwise manner, with early ankylosaurids evolving handle-like vertebrae before the distal osteoderms enlarged and coossified to form a knob. © 2015 Anatomical Society.

  17. Stepwise Bay Annulation of Indigo for the Synthesis of Desymmetrized Electron Acceptors and Donor–Acceptor Constructs

    DOE PAGES

    Kolaczkowski, Matthew A.; He, Bo; Liu, Yi

    2016-10-10

    In this work, a selective stepwise annulation of indigo has been demonstrated as a means of providing both monoannulated and differentially double-annulated indigo derivatives. Disparate substitution of the electron accepting bay-annulated indigo system allows for fine control over both the electronic properties as well as donor-acceptor structural architectures. Optical and electronic properties were characterized computationally as well as through UV-vis absorption spectroscopy and cyclic voltammetry. Finally, this straightforward method provides a modular approach for the design of indigo-based materials with tailored optoelectronic properties.

  18. Biometric parameters in different stages of primary angle closure using low-coherence interferometry.

    PubMed

    Yazdani, Shahin; Akbarian, Shadi; Pakravan, Mohammad; Doozandeh, Azadeh; Afrouzifar, Mohsen

    2015-03-01

    To compare ocular biometric parameters using low-coherence interferometry among siblings affected with different degrees of primary angle closure (PAC). In this cross-sectional comparative study, a total of 170 eyes of 86 siblings from 47 families underwent low-coherence interferometry (LenStar 900; Haag-Streit, Koeniz, Switzerland) to determine central corneal thickness, anterior chamber depth (ACD), aqueous depth (AD), lens thickness (LT), vitreous depth, and axial length (AL). Regression coefficients were applied to show the trend of the measured variables in different stages of angle closure. To evaluate the discriminative power of the parameters, receiver operating characteristic curves were used. Best cutoff points were selected based on the Youden index. Sensitivity, specificity, positive and negative predicative values, positive and negative likelihood ratios, and diagnostic accuracy were determined for each variable. All biometric parameters changed significantly from normal eyes to PAC suspects, PAC, and PAC glaucoma; there was a significant stepwise decrease in central corneal thickness, ACD, AD, vitreous depth, and AL, and an increase in LT and LT/AL. Anterior chamber depth and AD had the best diagnostic power for detecting angle closure; best levels of sensitivity and specificity were obtained with cutoff values of 3.11 mm for ACD and 2.57 mm for AD. Biometric parameters measured by low-coherence interferometry demonstrated a significant and stepwise change among eyes affected with various degrees of angle closure. Although the current classification scheme for angle closure is based on anatomical features, it has excellent correlation with biometric parameters.

  19. ISECG Global Exploration Roadmap: A Stepwise Approach to Deep Space Exploration

    NASA Technical Reports Server (NTRS)

    Martinez, Roland; Goodliff, Kandyce; Whitley, Ryan

    2013-01-01

    In 2011, ISECG released the Global Exploration Roadmap (GER), advancing the "Global Exploration Strategy: The Framework for Coordination" by articulating the perspectives of participating agencies on exploration goals and objectives, mission scenarios, and coordination of exploration preparatory activities. The GER featured a stepwise development and demonstration of capabilities ultimately required for human exploration of Mars. In 2013 the GER was updated to reflect the ongoing evolution of agency's exploration policies and plans, informed by individual agency and coordinated analysis activities that are relevant to various elements of the GER framework as well as coordinated stakeholder engagement activities. For this release of version 2 of the GER in the mid 2013 timeframe, a modified mission scenario is presented, more firmly reflecting the importance of a stepwise evolution of critical capabilities provided by multiple partners necessary for executing increasingly complex missions to multiple destinations and leading to human exploration of Mars. This paper will describe the updated mission scenario, the changes since the release of version 1, the mission themes incorporated into the scenario, and risk reduction for Mars missions provided by exploration at various destinations.

  20. Identification of molecular markers associated with mite resistance in coconut (Cocos nucifera L.).

    PubMed

    Shalini, K V; Manjunatha, S; Lebrun, P; Berger, A; Baudouin, L; Pirany, N; Ranganath, R M; Prasad, D Theertha

    2007-01-01

    Coconut mite (Aceria guerreronis 'Keifer') has become a major threat to Indian coconut (Coçcos nucifera L.) cultivators and the processing industry. Chemical and biological control measures have proved to be costly, ineffective, and ecologically undesirable. Planting mite-resistant coconut cultivars is the most effective method of preventing yield loss and should form a major component of any integrated pest management stratagem. Coconut genotypes, and mite-resistant and -susceptible accessions were collected from different parts of South India. Thirty-two simple sequence repeat (SSR) and 7 RAPD primers were used for molecular analyses. In single-marker analysis, 9 SSR and 4 RAPD markers associated with mite resistance were identified. In stepwise multiple regression analysis of SSRs, a combination of 6 markers showed 100% association with mite infestation. Stepwise multiple regression analysis for RAPD data revealed that a combination of 3 markers accounted for 83.86% of mite resistance in the selected materials. Combined stepwise multiple regression analysis of RAPD and SSR data showed that a combination of 5 markers explained 100% of the association with mite resistance in coconut. Markers associated with mite resistance are important in coconut breeding programs and will facilitate the selection of mite-resistant plants at an early stage as well as mother plants for breeding programs.

  1. Advances in variable selection methods I: Causal selection methods versus stepwise regression and principal component analysis on data of known and unknown functional relationships

    EPA Science Inventory

    Hydrological predictions at a watershed scale are commonly based on extrapolation and upscaling of hydrological behavior at plot and hillslope scales. Yet, dominant hydrological drivers at a hillslope may not be as dominant at the watershed scale because of the heterogeneity of w...

  2. In vitro selection and amplification protocols for isolation of aptameric sensors for small molecules

    PubMed Central

    Yang, Kyung-Ae; Pei, Renjun; Stojanovic, Milan N.

    2016-01-01

    We recently optimized a procedure that directly yields aptameric sensors for small molecules in so-called structure-switching format. The protocol has a high success rate, short time, and is sufficiently simple to be readily implemented in a non-specialist laboratory. We provide a stepwise guide to this selection protocol. PMID:27155227

  3. Advancing a Model-Validated Statistical Method for Decomposing the Key Oceanic Drivers of Regional Climate: Focus on Northern and Tropical African Climate Variability in the Community Earth System Model (CESM)

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Wang, Fuyao; Yu, Yan; Notaro, Michael

    This study advances the practicality and stability of the traditional multivariate statistical method, generalized equilibrium feedback assessment (GEFA), for decomposing the key oceanic drivers of regional atmospheric variability, especially when available data records are short. An advanced stepwise GEFA methodology is introduced, in which unimportant forcings within the forcing matrix are eliminated through stepwise selection. Method validation of stepwise GEFA is performed using the CESM, with a focused application to northern and tropical Africa (NTA). First, a statistical assessment of the atmospheric response to each primary oceanic forcing is carried out by applying stepwise GEFA to a fully coupled controlmore » run. Then, a dynamical assessment of the atmospheric response to individual oceanic forcings is performed through ensemble experiments by imposing sea surface temperature anomalies over focal ocean basins. Finally, to quantify the reliability of stepwise GEFA, the statistical assessment is evaluated against the dynamical assessment in terms of four metrics: the percentage of grid cells with consistent response sign, the spatial correlation of atmospheric response patterns, the area-averaged seasonal cycle of response magnitude, and consistency in associated mechanisms between assessments. In CESM, tropical modes, namely El Niño–Southern Oscillation and the tropical Indian Ocean Basin, tropical Indian Ocean dipole, and tropical Atlantic Niño modes, are the dominant oceanic controls of NTA climate. In complementary studies, stepwise GEFA is validated in terms of isolating terrestrial forcings on the atmosphere, and observed oceanic and terrestrial drivers of NTA climate are extracted to establish an observational benchmark for subsequent coupled model evaluation and development of process-based weights for regional climate projections.« less

  4. Advancing a Model-Validated Statistical Method for Decomposing the Key Oceanic Drivers of Regional Climate: Focus on Northern and Tropical African Climate Variability in the Community Earth System Model (CESM)

    DOE PAGES

    Wang, Fuyao; Yu, Yan; Notaro, Michael; ...

    2017-09-27

    This study advances the practicality and stability of the traditional multivariate statistical method, generalized equilibrium feedback assessment (GEFA), for decomposing the key oceanic drivers of regional atmospheric variability, especially when available data records are short. An advanced stepwise GEFA methodology is introduced, in which unimportant forcings within the forcing matrix are eliminated through stepwise selection. Method validation of stepwise GEFA is performed using the CESM, with a focused application to northern and tropical Africa (NTA). First, a statistical assessment of the atmospheric response to each primary oceanic forcing is carried out by applying stepwise GEFA to a fully coupled controlmore » run. Then, a dynamical assessment of the atmospheric response to individual oceanic forcings is performed through ensemble experiments by imposing sea surface temperature anomalies over focal ocean basins. Finally, to quantify the reliability of stepwise GEFA, the statistical assessment is evaluated against the dynamical assessment in terms of four metrics: the percentage of grid cells with consistent response sign, the spatial correlation of atmospheric response patterns, the area-averaged seasonal cycle of response magnitude, and consistency in associated mechanisms between assessments. In CESM, tropical modes, namely El Niño–Southern Oscillation and the tropical Indian Ocean Basin, tropical Indian Ocean dipole, and tropical Atlantic Niño modes, are the dominant oceanic controls of NTA climate. In complementary studies, stepwise GEFA is validated in terms of isolating terrestrial forcings on the atmosphere, and observed oceanic and terrestrial drivers of NTA climate are extracted to establish an observational benchmark for subsequent coupled model evaluation and development of process-based weights for regional climate projections.« less

  5. The Effect of College Selection Factors on Persistence: An Examination of Black and Latino Males in the Community College

    ERIC Educational Resources Information Center

    Wood, J. Luke; Harris, Frank, III

    2015-01-01

    The purpose of this study was to understand the relationship (if any) between college selection factors and persistence for Black and Latino males in the community college. Using data derived from the Educational Longitudinal Study, backwards stepwise logistic regression models were developed for both groups. Findings are contextualized in light…

  6. Application of ERTS-1 imagery to the study of caribou movements and winter dispersal in relation to prevailing snowcover

    NASA Technical Reports Server (NTRS)

    Lent, P. C. (Principal Investigator)

    1973-01-01

    The author has identified the following significant results. Step-wise discriminate analysis has demonstrated the feasibility of feature identification using linear discriminate functions of ERTS-1 MSS band densities and their ratios. The analysis indicated that features such as small streams can be detected even when they are in dark mountain shadow. The potential utility of this and similar analytic techniques appears considerable, and the limits it can be applied to analysis of ERTS-1 imagery are not yet fully known.

  7. Cloud field classification based upon high spatial resolution textural features. I - Gray level co-occurrence matrix approach

    NASA Technical Reports Server (NTRS)

    Welch, R. M.; Sengupta, S. K.; Chen, D. W.

    1988-01-01

    Stratocumulus, cumulus, and cirrus clouds were identified on the basis of cloud textural features which were derived from a single high-resolution Landsat MSS NIR channel using a stepwise linear discriminant analysis. It is shown that, using this method, it is possible to distinguish high cirrus clouds from low clouds with high accuracy on the basis of spatial brightness patterns. The largest probability of misclassification is associated with confusion between the stratocumulus breakup regions and the fair-weather cumulus.

  8. COPD: A stepwise or a hit hard approach?

    PubMed

    Ferreira, A J; Reis, A; Marçal, N; Pinto, P; Bárbara, C

    2016-01-01

    Current guidelines differ slightly on the recommendations for treatment of Chronic Obstructive Pulmonary Disease (COPD) patients, and although there are some undisputed recommendations, there is still debate regarding the management of COPD. One of the hindrances to deciding which therapeutic approach to choose is late diagnosis or misdiagnosis of COPD. After a proper diagnosis is achieved and severity assessed, the choice between a stepwise or "hit hard" approach has to be made. For GOLD A patients the stepwise approach is recommended, whilst for B, C and D patients this remains debatable. Moreover, in patients for whom inhaled corticosteroids (ICS) are recommended, a step-up or "hit hard" approach with triple therapy will depend on the patient's characteristics and, for patients who are being over-treated with ICS, ICS withdrawal should be performed, in order to optimize therapy and reduce excessive medications. This paper discusses and proposes stepwise, "hit hard", step-up and ICS withdrawal therapeutic approaches for COPD patients based on their GOLD group. We conclude that all approaches have benefits, and only a careful patient selection will determine which approach is better, and which patients will benefit the most from each approach. Copyright © 2016 Sociedade Portuguesa de Pneumologia. Published by Elsevier España, S.L.U. All rights reserved.

  9. Application of a single-objective, hybrid genetic algorithm approach to pharmacokinetic model building.

    PubMed

    Sherer, Eric A; Sale, Mark E; Pollock, Bruce G; Belani, Chandra P; Egorin, Merrill J; Ivy, Percy S; Lieberman, Jeffrey A; Manuck, Stephen B; Marder, Stephen R; Muldoon, Matthew F; Scher, Howard I; Solit, David B; Bies, Robert R

    2012-08-01

    A limitation in traditional stepwise population pharmacokinetic model building is the difficulty in handling interactions between model components. To address this issue, a method was previously introduced which couples NONMEM parameter estimation and model fitness evaluation to a single-objective, hybrid genetic algorithm for global optimization of the model structure. In this study, the generalizability of this approach for pharmacokinetic model building is evaluated by comparing (1) correct and spurious covariate relationships in a simulated dataset resulting from automated stepwise covariate modeling, Lasso methods, and single-objective hybrid genetic algorithm approaches to covariate identification and (2) information criteria values, model structures, convergence, and model parameter values resulting from manual stepwise versus single-objective, hybrid genetic algorithm approaches to model building for seven compounds. Both manual stepwise and single-objective, hybrid genetic algorithm approaches to model building were applied, blinded to the results of the other approach, for selection of the compartment structure as well as inclusion and model form of inter-individual and inter-occasion variability, residual error, and covariates from a common set of model options. For the simulated dataset, stepwise covariate modeling identified three of four true covariates and two spurious covariates; Lasso identified two of four true and 0 spurious covariates; and the single-objective, hybrid genetic algorithm identified three of four true covariates and one spurious covariate. For the clinical datasets, the Akaike information criterion was a median of 22.3 points lower (range of 470.5 point decrease to 0.1 point decrease) for the best single-objective hybrid genetic-algorithm candidate model versus the final manual stepwise model: the Akaike information criterion was lower by greater than 10 points for four compounds and differed by less than 10 points for three compounds. The root mean squared error and absolute mean prediction error of the best single-objective hybrid genetic algorithm candidates were a median of 0.2 points higher (range of 38.9 point decrease to 27.3 point increase) and 0.02 points lower (range of 0.98 point decrease to 0.74 point increase), respectively, than that of the final stepwise models. In addition, the best single-objective, hybrid genetic algorithm candidate models had successful convergence and covariance steps for each compound, used the same compartment structure as the manual stepwise approach for 6 of 7 (86 %) compounds, and identified 54 % (7 of 13) of covariates included by the manual stepwise approach and 16 covariate relationships not included by manual stepwise models. The model parameter values between the final manual stepwise and best single-objective, hybrid genetic algorithm models differed by a median of 26.7 % (q₁ = 4.9 % and q₃ = 57.1 %). Finally, the single-objective, hybrid genetic algorithm approach was able to identify models capable of estimating absorption rate parameters for four compounds that the manual stepwise approach did not identify. The single-objective, hybrid genetic algorithm represents a general pharmacokinetic model building methodology whose ability to rapidly search the feasible solution space leads to nearly equivalent or superior model fits to pharmacokinetic data.

  10. Associations between dietary and lifestyle risk factors and colorectal cancer in the Scottish population.

    PubMed

    Theodoratou, Evropi; Farrington, Susan M; Tenesa, Albert; McNeill, Geraldine; Cetnarskyj, Roseanne; Korakakis, Emmanouil; Din, Farhat V N; Porteous, Mary E; Dunlop, Malcolm G; Campbell, Harry

    2014-01-01

    Colorectal cancer (CRC) accounts for 9.7% of all cancer cases and for 8% of all cancer-related deaths. Established risk factors include personal or family history of CRC as well as lifestyle and dietary factors. We investigated the relationship between CRC and demographic, lifestyle, food and nutrient risk factors through a case-control study that included 2062 patients and 2776 controls from Scotland. Forward and backward stepwise regression was applied and the stability of the models was assessed in 1000 bootstrap samples. The variables that were automatically selected to be included by the forward or backward stepwise regression and whose selection was verified by bootstrap sampling in the current study were family history, dietary energy, 'high-energy snack foods', eggs, juice, sugar-sweetened beverages and white fish (associated with an increased CRC risk) and NSAIDs, coffee and magnesium (associated with a decreased CRC risk). Application of forward and backward stepwise regression in this CRC study identified some already established as well as some novel potential risk factors. Bootstrap findings suggest that examination of the stability of regression models by bootstrap sampling is useful in the interpretation of study findings. 'High-energy snack foods' and high-energy drinks (including sugar-sweetened beverages and fruit juices) as risk factors for CRC have not been reported previously and merit further investigation as such snacks and beverages are important contributors in European and North American diets.

  11. Artificial nose, NIR and UV-visible spectroscopy for the characterisation of the PDO Chianti Classico olive oil.

    PubMed

    Forina, M; Oliveri, P; Bagnasco, L; Simonetti, R; Casolino, M C; Nizzi Grifi, F; Casale, M

    2015-11-01

    An authentication study of the Italian PDO (Protected Designation of Origin) olive oil Chianti Classico, based on artificial nose, near-infrared and UV-visible spectroscopy, with a set of samples representative of the whole Chianti Classico production area and a considerable number of samples from other Italian PDO regions was performed. The signals provided by the three analytical techniques were used both individually and jointly, after fusion of the respective variables, in order to build a model for the Chianti Classico PDO olive oil. Different signal pre-treatments were performed in order to investigate their importance and their effects in enhancing and extracting information from experimental data, correcting backgrounds or removing baseline variations. Stepwise-Linear Discriminant Analysis (STEP-LDA) was used as a feature selection technique and, afterward, Linear Discriminant Analysis (LDA) and the class-modelling technique Quadratic Discriminant Analysis-UNEQual dispersed classes (QDA-UNEQ) were applied to sub-sets of selected variables, in order to obtain efficient models capable of characterising the extra virgin olive oils produced in the Chianti Classico PDO area. Copyright © 2015 Elsevier B.V. All rights reserved.

  12. Discriminant Analysis as a Tool for Admission Selection to Special Academic Programs. AIR 1986 Annual Forum Paper.

    ERIC Educational Resources Information Center

    Kissel, Mary Ann

    The use of stepwise discriminant analysis as a means to select entering students who would benefit from a special program for the disadvantaged was studied. In fall 1984, 278 full-time black students were admitted as first-time students to a large urban university. Of the total, 200 entered a special program for the disadvantaged and 78 entered…

  13. Cation Selectivity in Biological Cation Channels Using Experimental Structural Information and Statistical Mechanical Simulation.

    PubMed

    Finnerty, Justin John; Peyser, Alexander; Carloni, Paolo

    2015-01-01

    Cation selective channels constitute the gate for ion currents through the cell membrane. Here we present an improved statistical mechanical model based on atomistic structural information, cation hydration state and without tuned parameters that reproduces the selectivity of biological Na+ and Ca2+ ion channels. The importance of the inclusion of step-wise cation hydration in these results confirms the essential role partial dehydration plays in the bacterial Na+ channels. The model, proven reliable against experimental data, could be straightforwardly used for designing Na+ and Ca2+ selective nanopores.

  14. An analysis of the relationship of seven selected variables to State Board Test Pool Examination performance of the University of Tennessee, Knoxville, College of Nursing.

    PubMed

    Sharp, T G

    1984-02-01

    The study was designed to determine whether any one of seven selected variables or a combination of the variables is predictive of performance on the State Board Test Pool Examination. The selected variables studied were: high school grade point average (HSGPA), The University of Tennessee, Knoxville, College of Nursing grade point average (GPA), and American College Test Assessment (ACT) standard scores (English, ENG; mathematics, MA; social studies, SS; natural sciences, NSC; composite, COMP). Data utilized were from graduates of the baccalaureate program of The University of Tennessee, Knoxville, College of Nursing from 1974 through 1979. The sample of 322 was selected from a total population of 572. The Statistical Analysis System (SAS) was designed to accomplish analysis of the predictive relationship of each of the seven selected variables to State Board Test Pool Examination performance (result of pass or fail), a stepwise discriminant analysis was designed for determining the predictive relationship of the strongest combination of the independent variables to overall State Board Test Pool Examination performance (result of pass or fail), and stepwise multiple regression analysis was designed to determine the strongest predictive combination of selected variables for each of the five subexams of the State Board Test Pool Examination. The selected variables were each found to be predictive of SBTPE performance (result of pass or fail). The strongest combination for predicting SBTPE performance (result of pass or fail) was found to be GPA, MA, and NSC.

  15. Mission requirements for a manned earth observatory. Volume 1, task 1: Experiment selection, definition, and documentation

    NASA Technical Reports Server (NTRS)

    1973-01-01

    Information related to proposed earth observation experiments for shuttle sortie missions (SSM) in the 1980's is presented. The step-wise progression of study activities and the development of the rationale that led to the identification, selection, and description of earth observation experiments for SSM are listed. The selected experiments are described, defined, and documented by individual disciplines. These disciplines include: oceanography; meteorology; agriculture, forestry, and rangeland; geology; hydrology; and environmental impact.

  16. An evaluation of supervised classifiers for indirectly detecting salt-affected areas at irrigation scheme level

    NASA Astrophysics Data System (ADS)

    Muller, Sybrand Jacobus; van Niekerk, Adriaan

    2016-07-01

    Soil salinity often leads to reduced crop yield and quality and can render soils barren. Irrigated areas are particularly at risk due to intensive cultivation and secondary salinization caused by waterlogging. Regular monitoring of salt accumulation in irrigation schemes is needed to keep its negative effects under control. The dynamic spatial and temporal characteristics of remote sensing can provide a cost-effective solution for monitoring salt accumulation at irrigation scheme level. This study evaluated a range of pan-fused SPOT-5 derived features (spectral bands, vegetation indices, image textures and image transformations) for classifying salt-affected areas in two distinctly different irrigation schemes in South Africa, namely Vaalharts and Breede River. The relationship between the input features and electro conductivity measurements were investigated using regression modelling (stepwise linear regression, partial least squares regression, curve fit regression modelling) and supervised classification (maximum likelihood, nearest neighbour, decision tree analysis, support vector machine and random forests). Classification and regression trees and random forest were used to select the most important features for differentiating salt-affected and unaffected areas. The results showed that the regression analyses produced weak models (<0.4 R squared). Better results were achieved using the supervised classifiers, but the algorithms tend to over-estimate salt-affected areas. A key finding was that none of the feature sets or classification algorithms stood out as being superior for monitoring salt accumulation at irrigation scheme level. This was attributed to the large variations in the spectral responses of different crops types at different growing stages, coupled with their individual tolerances to saline conditions.

  17. A new computer aided diagnosis system for evaluation of chronic liver disease with ultrasound shear wave elastography imaging.

    PubMed

    Gatos, Ilias; Tsantis, Stavros; Spiliopoulos, Stavros; Karnabatidis, Dimitris; Theotokas, Ioannis; Zoumpoulis, Pavlos; Loupas, Thanasis; Hazle, John D; Kagadis, George C

    2016-03-01

    Classify chronic liver disease (CLD) from ultrasound shear-wave elastography (SWE) imaging by means of a computer aided diagnosis (CAD) system. The proposed algorithm employs an inverse mapping technique (red-green-blue to stiffness) to quantify 85 SWE images (54 healthy and 31 with CLD). Texture analysis is then applied involving the automatic calculation of 330 first and second order textural features from every transformed stiffness value map to determine functional features that characterize liver elasticity and describe liver condition for all available stages. Consequently, a stepwise regression analysis feature selection procedure is utilized toward a reduced feature subset that is fed into the support vector machines (SVMs) classification algorithm in the design of the CAD system. With regard to the mapping procedure accuracy, the stiffness map values had an average difference of 0.01 ± 0.001 kPa compared to the quantification results derived from the color-box provided by the built-in software of the ultrasound system. Highest classification accuracy from the SVM model was 87.0% with sensitivity and specificity values of 83.3% and 89.1%, respectively. Receiver operating characteristic curves analysis gave an area under the curve value of 0.85 with [0.77-0.89] confidence interval. The proposed CAD system employing color to stiffness mapping and classification algorithms offered superior results, comparing the already published clinical studies. It could prove to be of value to physicians improving the diagnostic accuracy of CLD and can be employed as a second opinion tool for avoiding unnecessary invasive procedures.

  18. Simple models for estimating local removals of timber in the northeast

    Treesearch

    David N. Larsen; David A. Gansner

    1975-01-01

    Provides a practical method of estimating subregional removals of timber and demonstrates its application to a typical problem. Stepwise multiple regression analysis is used to develop equations for estimating removals of softwood, hardwood, and all timber from selected characteristics of socioeconomic structure.

  19. Use of principal-component, correlation, and stepwise multiple-regression analyses to investigate selected physical and hydraulic properties of carbonate-rock aquifers

    USGS Publications Warehouse

    Brown, C. Erwin

    1993-01-01

    Correlation analysis in conjunction with principal-component and multiple-regression analyses were applied to laboratory chemical and petrographic data to assess the usefulness of these techniques in evaluating selected physical and hydraulic properties of carbonate-rock aquifers in central Pennsylvania. Correlation and principal-component analyses were used to establish relations and associations among variables, to determine dimensions of property variation of samples, and to filter the variables containing similar information. Principal-component and correlation analyses showed that porosity is related to other measured variables and that permeability is most related to porosity and grain size. Four principal components are found to be significant in explaining the variance of data. Stepwise multiple-regression analysis was used to see how well the measured variables could predict porosity and (or) permeability for this suite of rocks. The variation in permeability and porosity is not totally predicted by the other variables, but the regression is significant at the 5% significance level. ?? 1993.

  20. The use of sea ice habitat by female polar bears in the Beaufort Sea

    USGS Publications Warehouse

    Durner, George M.; Amstrup, Steven C.; Nielson, Ryan M.; McDonald, Trent

    2003-01-01

    Polar bears (Ursus maritimus) depend on ice-covered seas to satisfy life history requirements. Modern threats to polar bears include oil spills in the marine environment and changes in ice composition resulting from climate change. Managers need practical models that explain the distribution of bears in order to assess the impacts of these threats. We used stepwise procedures to create resource selection models of habitat use for radio-collared female polar bears in the Beaufort Sea. Sea ice characteristics and ocean depths at known polar bear locations were compared to the same features at randomly selected locations. Models generated for each of four seasons confirmed complexities of habitat use by polar bears and their response to numerous factors. Bears preferred shallow water areas where ice concentrations were > 80 % and different ice types intersected. Variation among seasons was reflected mainly in differential selection of ice stages, floe sizes, and their interactions. Water depth, total ice concentration and distance to the nearest interface between different ice types were significant terms in models for most seasons. Variation in ice stage and form also appeared in three models, and several interaction effects were identified. Habitat selection by polar bears is likely related to prey abundance and availability. Use of habitats in shallow water possibly reflects higher productivity in those areas. Habitat use in close proximity to ice edges is probably related to greater access of prey in those habitats.

  1. Developing a spatial-statistical model and map of historical malaria prevalence in Botswana using a staged variable selection procedure

    PubMed Central

    Craig, Marlies H; Sharp, Brian L; Mabaso, Musawenkosi LH; Kleinschmidt, Immo

    2007-01-01

    Background Several malaria risk maps have been developed in recent years, many from the prevalence of infection data collated by the MARA (Mapping Malaria Risk in Africa) project, and using various environmental data sets as predictors. Variable selection is a major obstacle due to analytical problems caused by over-fitting, confounding and non-independence in the data. Testing and comparing every combination of explanatory variables in a Bayesian spatial framework remains unfeasible for most researchers. The aim of this study was to develop a malaria risk map using a systematic and practicable variable selection process for spatial analysis and mapping of historical malaria risk in Botswana. Results Of 50 potential explanatory variables from eight environmental data themes, 42 were significantly associated with malaria prevalence in univariate logistic regression and were ranked by the Akaike Information Criterion. Those correlated with higher-ranking relatives of the same environmental theme, were temporarily excluded. The remaining 14 candidates were ranked by selection frequency after running automated step-wise selection procedures on 1000 bootstrap samples drawn from the data. A non-spatial multiple-variable model was developed through step-wise inclusion in order of selection frequency. Previously excluded variables were then re-evaluated for inclusion, using further step-wise bootstrap procedures, resulting in the exclusion of another variable. Finally a Bayesian geo-statistical model using Markov Chain Monte Carlo simulation was fitted to the data, resulting in a final model of three predictor variables, namely summer rainfall, mean annual temperature and altitude. Each was independently and significantly associated with malaria prevalence after allowing for spatial correlation. This model was used to predict malaria prevalence at unobserved locations, producing a smooth risk map for the whole country. Conclusion We have produced a highly plausible and parsimonious model of historical malaria risk for Botswana from point-referenced data from a 1961/2 prevalence survey of malaria infection in 1–14 year old children. After starting with a list of 50 potential variables we ended with three highly plausible predictors, by applying a systematic and repeatable staged variable selection procedure that included a spatial analysis, which has application for other environmentally determined infectious diseases. All this was accomplished using general-purpose statistical software. PMID:17892584

  2. Association between ICP pulse waveform morphology and ICP B waves.

    PubMed

    Kasprowicz, Magdalena; Bergsneider, Marvin; Czosnyka, Marek; Hu, Xiao

    2012-01-01

    The study aimed to investigate changes in the shape of ICP pulses associated with different patterns of the ICP slow waves (0.5-2.0 cycles/min) during ICP overnight monitoring in hydrocephalus. Four patterns of ICP slow waves were characterized in 44 overnight ICP recordings (no waves - NW, slow symmetrical waves - SW, slow asymmetrical waves - AS, slow waves with plateau phase - PW). The morphological clustering and analysis of ICP pulse (MOCAIP) algorithm was utilized to calculate a set of metrics describing ICP pulse morphology based on the location of three sub-peaks in an ICP pulse: systolic peak (P(1)), tidal peak (P(2)) and dicrotic peak (P(3)). Step-wise discriminant analysis was applied to select the most characteristic morphological features to distinguish between different ICP slow waves. Based on relative changes in variability of amplitudes of P(2) and P(3) we were able to distinguish between the combined groups NW + SW and AS + PW (p < 0.000001). The AS pattern can be differentiated from PW based on respective changes in the mean curvature of P(2) and P(3) (p < 0.000001); however, none of the MOCAIP feature separates between NW and SW. The investigation of ICP pulse morphology associated with different ICP B waves may provide additional information for analysing recordings of overnight ICP.

  3. Lymphoma diagnosis in histopathology using a multi-stage visual learning approach

    NASA Astrophysics Data System (ADS)

    Codella, Noel; Moradi, Mehdi; Matasar, Matt; Sveda-Mahmood, Tanveer; Smith, John R.

    2016-03-01

    This work evaluates the performance of a multi-stage image enhancement, segmentation, and classification approach for lymphoma recognition in hematoxylin and eosin (H and E) stained histopathology slides of excised human lymph node tissue. In the first stage, the original histology slide undergoes various image enhancement and segmentation operations, creating an additional 5 images for every slide. These new images emphasize unique aspects of the original slide, including dominant staining, staining segmentations, non-cellular groupings, and cellular groupings. For the resulting 6 total images, a collection of visual features are extracted from 3 different spatial configurations. Visual features include the first fully connected layer (4096 dimensions) of the Caffe convolutional neural network trained from ImageNet data. In total, over 200 resultant visual descriptors are extracted for each slide. Non-linear SVMs are trained over each of the over 200 descriptors, which are then input to a forward stepwise ensemble selection that optimizes a late fusion sum of logistically normalized model outputs using local hill climbing. The approach is evaluated on a public NIH dataset containing 374 images representing 3 lymphoma conditions: chronic lymphocytic leukemia (CLL), follicular lymphoma (FL), and mantle cell lymphoma (MCL). Results demonstrate a 38.4% reduction in residual error over the current state-of-art on this dataset.

  4. Vocal individuality cues in the African penguin (Spheniscus demersus): a source-filter theory approach.

    PubMed

    Favaro, Livio; Gamba, Marco; Alfieri, Chiara; Pessani, Daniela; McElligott, Alan G

    2015-11-25

    The African penguin is a nesting seabird endemic to southern Africa. In penguins of the genus Spheniscus vocalisations are important for social recognition. However, it is not clear which acoustic features of calls can encode individual identity information. We recorded contact calls and ecstatic display songs of 12 adult birds from a captive colony. For each vocalisation, we measured 31 spectral and temporal acoustic parameters related to both source and filter components of calls. For each parameter, we calculated the Potential of Individual Coding (PIC). The acoustic parameters showing PIC ≥ 1.1 were used to perform a stepwise cross-validated discriminant function analysis (DFA). The DFA correctly classified 66.1% of the contact calls and 62.5% of display songs to the correct individual. The DFA also resulted in the further selection of 10 acoustic features for contact calls and 9 for display songs that were important for vocal individuality. Our results suggest that studying the anatomical constraints that influence nesting penguin vocalisations from a source-filter perspective, can lead to a much better understanding of the acoustic cues of individuality contained in their calls. This approach could be further extended to study and understand vocal communication in other bird species.

  5. Vocal individuality cues in the African penguin (Spheniscus demersus): a source-filter theory approach

    PubMed Central

    Favaro, Livio; Gamba, Marco; Alfieri, Chiara; Pessani, Daniela; McElligott, Alan G.

    2015-01-01

    The African penguin is a nesting seabird endemic to southern Africa. In penguins of the genus Spheniscus vocalisations are important for social recognition. However, it is not clear which acoustic features of calls can encode individual identity information. We recorded contact calls and ecstatic display songs of 12 adult birds from a captive colony. For each vocalisation, we measured 31 spectral and temporal acoustic parameters related to both source and filter components of calls. For each parameter, we calculated the Potential of Individual Coding (PIC). The acoustic parameters showing PIC ≥ 1.1 were used to perform a stepwise cross-validated discriminant function analysis (DFA). The DFA correctly classified 66.1% of the contact calls and 62.5% of display songs to the correct individual. The DFA also resulted in the further selection of 10 acoustic features for contact calls and 9 for display songs that were important for vocal individuality. Our results suggest that studying the anatomical constraints that influence nesting penguin vocalisations from a source-filter perspective, can lead to a much better understanding of the acoustic cues of individuality contained in their calls. This approach could be further extended to study and understand vocal communication in other bird species. PMID:26602001

  6. Stepwise Evolution of Nonliving to Living Chemical Systems

    NASA Astrophysics Data System (ADS)

    Lindahl, Paul A.

    2004-08-01

    Steps by which a nonliving chemical system could have transformed into a living system are described and discussed, assuming general features of Wächtershäuser's chemo-autotrophic surface theory of the origin of life. Environmental species such as CO2 and H2S are proposed to have reacted to form a quasi-steady state metal-bound intermediate (CH3-M) that slowly decayed into waste (CH4). Unpredictable dispersive reactions expanded the system to include surface-bound forms of the citric acid cycle intermediates (oxaloacetate --> citrate). Further reaction yielded an autocatalytic system in which raw materials are converted into the system at exponential rates. Combinatorial dispersive reactions that improved the performance of this system were automatically selected and incorporated into it. Systems evolved critical features of living systems (proteins, membranes, proteins, nucleic acids, etc.) using two related mechanisms called grafting and waste-conversion. Such living systems were transformed from less recognizable types (characterized by autocatalytic spreading, decentralization, poorly defined boundaries, etc.) into more recognizable ones (encapsulated by membranes, controlled by single-molecule genomes, etc.) that self-replicated by a cell division cycle and could evolve by the standard gene-based Darwinian mechanism. The resulting systems are viewed as having an autocatalytic network composed of three linked autocatalytic subreactions.

  7. Environmental influences on alcohol consumption practices of alcoholic beverage servers.

    PubMed

    Nusbaumer, Michael R; Reiling, Denise M

    2002-11-01

    Public drinking establishments have long been associated with heavy drinking among both their patrons and servers. Whether these environments represent locations where heavy drinking is learned (learning hypothesis) or simply places where already-heavy drinkers gather in a supportive environment (selection hypothesis) remains an important question. A sample of licensed alcoholic beverage servers in the state of Indiana, USA, was surveyed to better understand the drinking behaviors of servers within the alcohol service industry. Responses (N = 938) to a mailed questionnaire were analyzed to assess the relative influence of environmental and demographic factors on the drinking behavior of servers. Stepwise regression revealed "drinking on the job" as the most influential environmental factor on heavy drinking behaviors, followed by age and gender as influential demographic factors. Support was found for the selection hypothesis, but not for the learning hypothesis. Policy implications are discussed. factors on the drinking behavior of servers. Stepwise regression revealed "drinking on the job" as the most influential environmental factor on heavy drinking behaviors, followed by age and gender as influential demographic factors. Support was found for the selection hypothesis, but not for the learning hypothesis. Policy implications are discussed.

  8. Dealing with office emergencies. Stepwise approach for family physicians.

    PubMed Central

    Sempowski, Ian P.; Brison, Robert J.

    2002-01-01

    OBJECTIVE: To develop a simple stepwise approach to initial management of emergencies in family physicians' offices; to review how to prepare health care teams and equipment; and to illustrate a general approach to three of the most common office emergencies. QUALITY OF EVIDENCE: MEDLINE was searched from January 1980 to December 2001. Articles were selected based on their clinical relevance, quality of evidence, and date of publication. We reviewed American family medicine, pediatric, dental, and dermatologic articles, but found that the area has not been well studied from a Canadian family medicine perspective. Consensus statements by specialty professional groups were used to identify accepted emergency medical treatments. MAIN MESSAGE: Family medicine offices are frequently poorly equipped and inadequately prepared to deal with emergencies. Straightforward emergency response plans can be designed and tailored to an office's risk profile. A systematic team approach and effective use of skills, support staff, and equipment is important. The general approach can be modified for specific patients or conditions. CONCLUSION: Family physicians can plan ahead and use a team approach to develop a simple stepwise response to emergency situations in the office. PMID:12371305

  9. One-Step and Stepwise Magnification of a BOBBED LETHAL Chromosome in DROSOPHILA MELANOGASTER

    PubMed Central

    Endow, Sharyn A.; Komma, Donald J.

    1986-01-01

    Bobbed lethal (bbl) chromosomes carry too few ribosomal genes for homozygous flies to be viable. Reversion of bbl chromosomes to bb or nearly bb + occurs under magnifying conditions at a low frequency in a single generation. These reversions occur too rapidly to be accounted for by single unequal sister chromatid exchanges and seem unlikely to be due to multiple sister strand exchanges within a given cell lineage. Analysis of several one-step revertants indicates that they are X-Y recombinant chromosomes which probably arise from X-Y recombination at bb. The addition of ribosomal genes from the Y chromosome to the bbl chromosome explains the more rapid reversion of the bbl chromosome than is permitted by single events of unequal sister chromatid exchange. Analysis of stepwise bbl magnified chromosomes, which were selected over a period of 4–9 magnifying generations, shows ribosomal gene patterns that are closely similar to each other. Similarity in rDNA pattern among stepwise magnified products of the same parental chromosome is consistent with reversion by a mechanism of unequal sister strand exchange. PMID:3095184

  10. Three-dimensional texture analysis of contrast enhanced CT images for treatment response assessment in Hodgkin lymphoma: Comparison with F-18-FDG PET

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Knogler, Thomas; El-Rabadi, Karem; Weber, Michael

    2014-12-15

    Purpose: To determine the diagnostic performance of three-dimensional (3D) texture analysis (TA) of contrast-enhanced computed tomography (CE-CT) images for treatment response assessment in patients with Hodgkin lymphoma (HL), compared with F-18-fludeoxyglucose (FDG) positron emission tomography/CT. Methods: 3D TA of 48 lymph nodes in 29 patients was performed on venous-phase CE-CT images before and after chemotherapy. All lymph nodes showed pathologically elevated FDG uptake at baseline. A stepwise logistic regression with forward selection was performed to identify classic CT parameters and texture features (TF) that enable the separation of complete response (CR) and persistent disease. Results: The TF fraction of imagemore » in runs, calculated for the 45° direction, was able to correctly identify CR with an accuracy of 75%, a sensitivity of 79.3%, and a specificity of 68.4%. Classical CT features achieved an accuracy of 75%, a sensitivity of 86.2%, and a specificity of 57.9%, whereas the combination of TF and CT imaging achieved an accuracy of 83.3%, a sensitivity of 86.2%, and a specificity of 78.9%. Conclusions: 3D TA of CE-CT images is potentially useful to identify nodal residual disease in HL, with a performance comparable to that of classical CT parameters. Best results are achieved when TA and classical CT features are combined.« less

  11. Neural network classification of sweet potato embryos

    NASA Astrophysics Data System (ADS)

    Molto, Enrique; Harrell, Roy C.

    1993-05-01

    Somatic embryogenesis is a process that allows for the in vitro propagation of thousands of plants in sub-liter size vessels and has been successfully applied to many significant species. The heterogeneity of maturity and quality of embryos produced with this technique requires sorting to obtain a uniform product. An automated harvester is being developed at the University of Florida to sort embryos in vitro at different stages of maturation in a suspension culture. The system utilizes machine vision to characterize embryo morphology and a fluidic based separation device to isolate embryos associated with a pre-defined, targeted morphology. Two different backpropagation neural networks (BNN) were used to classify embryos based on information extracted from the vision system. One network utilized geometric features such as embryo area, length, and symmetry as inputs. The alternative network utilized polar coordinates of an embryo's perimeter with respect to its centroid as inputs. The performances of both techniques were compared with each other and with an embryo classification method based on linear discriminant analysis (LDA). Similar results were obtained with all three techniques. Classification efficiency was improved by reducing the dimension of the feature vector trough a forward stepwise analysis by LDA. In order to enhance the purity of the sample selected as harvestable, a reject to classify option was introduced in the model and analyzed. The best classifier performances (76% overall correct classifications, 75% harvestable objects properly classified, homogeneity improvement ratio 1.5) were obtained using 8 features in a BNN.

  12. Probabilistic precipitation and temperature downscaling of the Twentieth Century Reanalysis over France

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Caillouet, Laurie; Vidal, Jean -Philippe; Sauquet, Eric

    This work proposes a daily high-resolution probabilistic reconstruction of precipitation and temperature fields in France over the 1871–2012 period built on the NOAA Twentieth Century global extended atmospheric reanalysis (20CR). The objective is to fill in the spatial and temporal data gaps in surface observations in order to improve our knowledge on the local-scale climate variability from the late nineteenth century onwards. The SANDHY (Stepwise ANalogue Downscaling method for HYdrology) statistical downscaling method, initially developed for quantitative precipitation forecast, is used here to bridge the scale gap between large-scale 20CR predictors and local-scale predictands from the Safran high-resolution near-surface reanalysis,more » available from 1958 onwards only. SANDHY provides a daily ensemble of 125 analogue dates over the 1871–2012 period for 608 climatically homogeneous zones paving France. Large precipitation biases in intermediary seasons are shown to occur in regions with high seasonal asymmetry like the Mediterranean. Moreover, winter and summer temperatures are respectively over- and under-estimated over the whole of France. Two analogue subselection methods are therefore developed with the aim of keeping the structure of the SANDHY method unchanged while reducing those seasonal biases. The calendar selection keeps the analogues closest to the target calendar day. The stepwise selection applies two new analogy steps based on similarity of the sea surface temperature (SST) and the large-scale 2 m temperature ( T). Comparisons to the Safran reanalysis over 1959–2007 and to homogenized series over the whole twentieth century show that biases in the interannual cycle of precipitation and temperature are reduced with both methods. The stepwise subselection moreover leads to a large improvement of interannual correlation and reduction of errors in seasonal temperature time series. When the calendar subselection is an easily applicable method suitable in a quantitative precipitation forecast context, the stepwise subselection method allows for potential season shifts and SST trends and is therefore better suited for climate reconstructions and climate change studies. Furthermore, the probabilistic downscaling of 20CR over the period 1871–2012 with the SANDHY probabilistic downscaling method combined with the stepwise subselection thus constitutes a perfect framework for assessing the recent observed meteorological events but also future events projected by climate change impact studies and putting them in a historical perspective.« less

  13. Probabilistic precipitation and temperature downscaling of the Twentieth Century Reanalysis over France

    DOE PAGES

    Caillouet, Laurie; Vidal, Jean -Philippe; Sauquet, Eric; ...

    2016-03-16

    This work proposes a daily high-resolution probabilistic reconstruction of precipitation and temperature fields in France over the 1871–2012 period built on the NOAA Twentieth Century global extended atmospheric reanalysis (20CR). The objective is to fill in the spatial and temporal data gaps in surface observations in order to improve our knowledge on the local-scale climate variability from the late nineteenth century onwards. The SANDHY (Stepwise ANalogue Downscaling method for HYdrology) statistical downscaling method, initially developed for quantitative precipitation forecast, is used here to bridge the scale gap between large-scale 20CR predictors and local-scale predictands from the Safran high-resolution near-surface reanalysis,more » available from 1958 onwards only. SANDHY provides a daily ensemble of 125 analogue dates over the 1871–2012 period for 608 climatically homogeneous zones paving France. Large precipitation biases in intermediary seasons are shown to occur in regions with high seasonal asymmetry like the Mediterranean. Moreover, winter and summer temperatures are respectively over- and under-estimated over the whole of France. Two analogue subselection methods are therefore developed with the aim of keeping the structure of the SANDHY method unchanged while reducing those seasonal biases. The calendar selection keeps the analogues closest to the target calendar day. The stepwise selection applies two new analogy steps based on similarity of the sea surface temperature (SST) and the large-scale 2 m temperature ( T). Comparisons to the Safran reanalysis over 1959–2007 and to homogenized series over the whole twentieth century show that biases in the interannual cycle of precipitation and temperature are reduced with both methods. The stepwise subselection moreover leads to a large improvement of interannual correlation and reduction of errors in seasonal temperature time series. When the calendar subselection is an easily applicable method suitable in a quantitative precipitation forecast context, the stepwise subselection method allows for potential season shifts and SST trends and is therefore better suited for climate reconstructions and climate change studies. Furthermore, the probabilistic downscaling of 20CR over the period 1871–2012 with the SANDHY probabilistic downscaling method combined with the stepwise subselection thus constitutes a perfect framework for assessing the recent observed meteorological events but also future events projected by climate change impact studies and putting them in a historical perspective.« less

  14. Synthesis and high proton conductive performance of a quaternary vanadomolybdotungstosilicic heteropoly acid.

    PubMed

    Cai, Huaxue; Wu, Xuefei; Wu, Qingyin; Yan, Wenfu

    2016-09-28

    A new vanadium and molybdenum-substituted quaternary silicon-containing heteropoly acid H6SiW9MoV2O40·15H2O has been synthesized in this paper by the stepwise acidification and the stepwise addition of elemental solutions. The structural feature and hydration of this product were characterized by IR, UV, XRD and TG-DTA, and its proton conductivity was measured by electrochemical impedance spectroscopy (EIS). The result of EIS shows that H6SiW9MoV2O40·15H2O is a solid high-proton conductor with a conductivity of 6.01 × 10(-3) S cm(-1) at 22 °C and 80% relative humidity, which increases at higher temperatures. Its conductive activation energy is 27.5 kJ mol(-1), which suggests that the mechanism of proton conduction is dominated by the vehicle mechanism.

  15. Cation Selectivity in Biological Cation Channels Using Experimental Structural Information and Statistical Mechanical Simulation

    PubMed Central

    Finnerty, Justin John

    2015-01-01

    Cation selective channels constitute the gate for ion currents through the cell membrane. Here we present an improved statistical mechanical model based on atomistic structural information, cation hydration state and without tuned parameters that reproduces the selectivity of biological Na+ and Ca2+ ion channels. The importance of the inclusion of step-wise cation hydration in these results confirms the essential role partial dehydration plays in the bacterial Na+ channels. The model, proven reliable against experimental data, could be straightforwardly used for designing Na+ and Ca2+ selective nanopores. PMID:26460827

  16. Juvenile Offender Recidivism: An Examination of Risk Factors

    ERIC Educational Resources Information Center

    Calley, Nancy G.

    2012-01-01

    One hundred and seventy three male juvenile offenders were followed two years postrelease from a residential treatment facility to assess recidivism and factors related to recidivism. The overall recidivism rate was 23.9%. Logistic regression with stepwise and backward variable selection methods was used to examine the relationship between…

  17. Systematic feature selection improves accuracy of methylation-based forensic age estimation in Han Chinese males.

    PubMed

    Feng, Lei; Peng, Fuduan; Li, Shanfei; Jiang, Li; Sun, Hui; Ji, Anquan; Zeng, Changqing; Li, Caixia; Liu, Fan

    2018-03-23

    Estimating individual age from biomarkers may provide key information facilitating forensic investigations. Recent progress has shown DNA methylation at age-associated CpG sites as the most informative biomarkers for estimating the individual age of an unknown donor. Optimal feature selection plays a critical role in determining the performance of the final prediction model. In this study we investigate methylation levels at 153 age-associated CpG sites from 21 previously reported genomic regions using the EpiTYPER system for their predictive power on individual age in 390 Han Chinese males ranging from 15 to 75 years of age. We conducted a systematic feature selection using a stepwise backward multiple linear regression analysis as well as an exhaustive searching algorithm. Both approaches identified the same subset of 9 CpG sites, which in linear combination provided the optimal model fitting with mean absolute deviation (MAD) of 2.89 years of age and explainable variance (R 2 ) of 0.92. The final model was validated in two independent Han Chinese male samples (validation set 1, N = 65, MAD = 2.49, R 2  = 0.95, and validation set 2, N = 62, MAD = 3.36, R 2  = 0.89). Other competing models such as support vector machine and artificial neural network did not outperform the linear model to any noticeable degree. The validation set 1 was additionally analyzed using Pyrosequencing technology for cross-platform validation and was termed as validation set 3. Directly applying our model, in which the methylation levels were detected by the EpiTYPER system, to the data from pyrosequencing technology showed, however, less accurate results in terms of MAD (validation set 3, N = 65 Han Chinese males, MAD = 4.20, R 2  = 0.93), suggesting the presence of a batch effect between different data generation platforms. This batch effect could be partially overcome by a z-score transformation (MAD = 2.76, R 2  = 0.93). Overall, our systematic feature selection identified 9 CpG sites as the optimal subset for forensic age estimation and the prediction model consisting of these 9 markers demonstrated high potential in forensic practice. An age estimator implementing our prediction model allowing missing markers is freely available at http://liufan.big.ac.cn/AgePrediction. Copyright © 2018 Elsevier B.V. All rights reserved.

  18. The controversial role of food allergy in infantile colic: evidence and clinical management.

    PubMed

    Nocerino, Rita; Pezzella, Vincenza; Cosenza, Linda; Amoroso, Antonio; Di Scala, Carmen; Amato, Francesco; Iacono, Giuseppe; Canani, Roberto Berni

    2015-03-19

    Food allergies (FAs) are an increasing problem in Western countries, affecting up to 10% of young children. FAs are frequently associated with gastrointestinal manifestations. The role of FAs as a potential causative factor for infantile colic (IC) is still controversial. We report the most recent evidence on the pathogenesis, clinical and diagnostic aspects of FA-induced infantile colic (IC) and suggest a stepwise diagnostic approach. We selected articles on clinical and immunologic features, pathogenesis and management of FAs and IC from of 1981 to 2015. Original and review articles were identified through selective searches performed on PubMed, using the following terms: colic, infantile colic, food allergy and infantile colic, infantile colic treatment. The possible relationship between FAs and IC derives from the presence of dysmotility with visceral hypersensitivity and dysbiosis, demonstrated in both conditions, and the clinical response to dietary interventions. Unfortunately, the design of the studies, poor characterization of atopy and different dietary approaches limit the understanding of the importance of FAs in subjects with IC. The role of FAs in IC subjects without other symptoms of atopy remains controversial. However, where there is a suspicion of FAs, a short trial with an extensively hydrolyzed cow's proteins formula or, if breast fed, with maternal elimination diet may be considered a reasonable option.

  19. Classification and Progression Based on CFS-GA and C5.0 Boost Decision Tree of TCM Zheng in Chronic Hepatitis B.

    PubMed

    Chen, Xiao Yu; Ma, Li Zhuang; Chu, Na; Zhou, Min; Hu, Yiyang

    2013-01-01

    Chronic hepatitis B (CHB) is a serious public health problem, and Traditional Chinese Medicine (TCM) plays an important role in the control and treatment for CHB. In the treatment of TCM, zheng discrimination is the most important step. In this paper, an approach based on CFS-GA (Correlation based Feature Selection and Genetic Algorithm) and C5.0 boost decision tree is used for zheng classification and progression in the TCM treatment of CHB. The CFS-GA performs better than the typical method of CFS. By CFS-GA, the acquired attribute subset is classified by C5.0 boost decision tree for TCM zheng classification of CHB, and C5.0 decision tree outperforms two typical decision trees of NBTree and REPTree on CFS-GA, CFS, and nonselection in comparison. Based on the critical indicators from C5.0 decision tree, important lab indicators in zheng progression are obtained by the method of stepwise discriminant analysis for expressing TCM zhengs in CHB, and alterations of the important indicators are also analyzed in zheng progression. In conclusion, all the three decision trees perform better on CFS-GA than on CFS and nonselection, and C5.0 decision tree outperforms the two typical decision trees both on attribute selection and nonselection.

  20. Rapid initiation of fetal therapy services with a system of learner-centred training under proctorship: the National University Hospital (Singapore) experience

    PubMed Central

    Gosavi, Arundhati; Vijayakumar, Pradip D; Ng, Bryan SW; Loh, May-Han; Tan, Lay Geok; Johana, Nuryanti; Tan, Yi Wan; Sandikin, Dedy; Su, Lin Lin; Wataganara, Tuangsit; Biswas, Arijit; Choolani, Mahesh A; Mattar, Citra NZ

    2017-01-01

    INTRODUCTION Management of complicated monochorionic twins and certain intrauterine structural anomalies is a pressing challenge in communities that still lack advanced fetal therapy. We describe our efforts to rapidly initiate selective feticide using radiofrequency ablation (RFA) and selective fetoscopic laser photocoagulation (SFLP) for twin-to-twin transfusion syndrome (TTTS), and present the latter as a potential model for aspiring fetal therapy units. METHODS Five pregnancies with fetal complications were identified for RFA. Three pregnancies with Stage II TTTS were selected for SFLP. While RFA techniques utilising ultrasonography skills were quickly mastered, SFLP required stepwise technical learning with an overseas-based proctor, who provided real-time hands-off supervision. RESULTS All co-twins were live-born following selective feticide; one singleton pregnancy was lost. Fetoscopy techniques were learned in a stepwise manner and procedures were performed by a novice team of surgeons under proctorship. Dichorionisation was completed in only one patient. Five of six twins were live-born near term. One pregnancy developed twin anaemia-polycythaemia sequence, while another was complicated by co-twin demise. DISCUSSION Proctor-supervised directed learning facilitated the rapid provision of basic fetal therapy services by our unit. While traditional apprenticeship is important for building individual expertise, this system is complementary and may benefit other small units committed to providing these services. PMID:27439783

  1. Utility of Intermediate-Delay Washout CT Images for Differentiation of Malignant and Benign Adrenal Lesions: A Multivariate Analysis.

    PubMed

    Ng, Chaan S; Altinmakas, Emre; Wei, Wei; Ghosh, Payel; Li, Xiao; Grubbs, Elizabeth G; Perrier, Nancy D; Lee, Jeffrey E; Prieto, Victor G; Hobbs, Brian P

    2018-06-27

    The objective of this study was to identify features that impact the diagnostic performance of intermediate-delay washout CT for distinguishing malignant from benign adrenal lesions. This retrospective study evaluated 127 pathologically proven adrenal lesions (82 malignant, 45 benign) in 126 patients who had undergone portal venous phase and intermediate-delay washout CT (1-3 minutes after portal venous phase) with or without unenhanced images. Unenhanced images were available for 103 lesions. Quantitatively, lesion CT attenuation on unenhanced (UA) and delayed (DL) images, absolute and relative percentage of enhancement washout (APEW and RPEW, respectively), descriptive CT features (lesion size, margin characteristics, heterogeneity or homogeneity, fat, calcification), patient demographics, and medical history were evaluated for association with lesion status using multiple logistic regression with stepwise model selection. Area under the ROC curve (A z ) was calculated from both univariate and multivariate analyses. The predictive diagnostic performance of multivariate evaluations was ascertained through cross-validation. A z for DL, APEW, RPEW, and UA was 0.751, 0.795, 0.829, and 0.839, respectively. Multivariate analyses yielded the following significant CT quantitative features and associated A z when combined: RPEW and DL (A z = 0.861) when unenhanced images were not available and APEW and UA (A z = 0.889) when unenhanced images were available. Patient demographics and presence of a prior malignancy were additional significant factors, increasing A z to 0.903 and 0.927, respectively. The combined predictive classifier, without and with UA available, yielded 85.7% and 87.3% accuracies with cross-validation, respectively. When appropriately combined with other CT features, washout derived from intermediate-delay CT with or without additional clinical data has potential utility in differentiating malignant from benign adrenal lesions.

  2. Multivariate analysis of early and late nest sites of Abert's Towhees

    Treesearch

    Deborah M. Finch

    1985-01-01

    Seasonal variation in nest site selection by the Abert's towhee (Pipilo aberti) was studied in honey mesquite (Prosopis glandulosa) habitat along the lower Colorado River from March to July, 1981. Stepwise discriminant function analysis identified nest vegetation type, nest direction, and nest height as the three most important variables that characterized the...

  3. Shark Attack: high affinity binding proteins derived from shark vNAR domains by stepwise in vitro affinity maturation.

    PubMed

    Zielonka, Stefan; Weber, Niklas; Becker, Stefan; Doerner, Achim; Christmann, Andreas; Christmann, Christine; Uth, Christina; Fritz, Janine; Schäfer, Elena; Steinmann, Björn; Empting, Martin; Ockelmann, Pia; Lierz, Michael; Kolmar, Harald

    2014-12-10

    A novel method for stepwise in vitro affinity maturation of antigen-specific shark vNAR domains is described that exclusively relies on semi-synthetic repertoires derived from non-immunized sharks. Target-specific molecules were selected from a CDR3-randomized bamboo shark (Chiloscyllium plagiosum) vNAR library using yeast surface display as platform technology. Various antigen-binding vNAR domains were easily isolated by screening against several therapeutically relevant antigens, including the epithelial cell adhesion molecule (EpCAM), the Ephrin type-A receptor 2 (EphA2), and the human serine protease HTRA1. Affinity maturation was demonstrated for EpCAM and HTRA1 by diversifying CDR1 of target-enriched populations which allowed for the rapid selection of nanomolar binders. EpCAM-specific vNAR molecules were produced as soluble proteins and more extensively characterized via thermal shift assays and biolayer interferometry. Essentially, we demonstrate that high-affinity binders can be generated in vitro without largely compromising the desirable high thermostability of the vNAR scaffold. Copyright © 2014 Elsevier B.V. All rights reserved.

  4. Determination of benzo[a]pyrene in cigarette mainstream smoke by using mid-infrared spectroscopy associated with a novel chemometric algorithm.

    PubMed

    Zhang, Yan; Zou, Hong-Yan; Shi, Pei; Yang, Qin; Tang, Li-Juan; Jiang, Jian-Hui; Wu, Hai-Long; Yu, Ru-Qin

    2016-01-01

    Determination of benzo[a]pyrene (BaP) in cigarette smoke can be very important for the tobacco quality control and the assessment of its harm to human health. In this study, mid-infrared spectroscopy (MIR) coupled to chemometric algorithm (DPSO-WPT-PLS), which was based on the wavelet packet transform (WPT), discrete particle swarm optimization algorithm (DPSO) and partial least squares regression (PLS), was used to quantify harmful ingredient benzo[a]pyrene in the cigarette mainstream smoke with promising result. Furthermore, the proposed method provided better performance compared to several other chemometric models, i.e., PLS, radial basis function-based PLS (RBF-PLS), PLS with stepwise regression variable selection (Stepwise-PLS) as well as WPT-PLS with informative wavelet coefficients selected by correlation coefficient test (rtest-WPT-PLS). It can be expected that the proposed strategy could become a new effective, rapid quantitative analysis technique in analyzing the harmful ingredient BaP in cigarette mainstream smoke. Copyright © 2015 Elsevier B.V. All rights reserved.

  5. A novel simple QSAR model for the prediction of anti-HIV activity using multiple linear regression analysis.

    PubMed

    Afantitis, Antreas; Melagraki, Georgia; Sarimveis, Haralambos; Koutentis, Panayiotis A; Markopoulos, John; Igglessi-Markopoulou, Olga

    2006-08-01

    A quantitative-structure activity relationship was obtained by applying Multiple Linear Regression Analysis to a series of 80 1-[2-hydroxyethoxy-methyl]-6-(phenylthio) thymine (HEPT) derivatives with significant anti-HIV activity. For the selection of the best among 37 different descriptors, the Elimination Selection Stepwise Regression Method (ES-SWR) was utilized. The resulting QSAR model (R (2) (CV) = 0.8160; S (PRESS) = 0.5680) proved to be very accurate both in training and predictive stages.

  6. Selective laser ionisation of radionuclide 63Ni

    NASA Astrophysics Data System (ADS)

    Tsvetkov, G. O.; D'yachkov, A. B.; Gorkunov, A. A.; Labozin, A. V.; Mironov, S. M.; Firsov, V. A.; Panchenko, V. Ya.

    2017-02-01

    We report a search for a scheme of selective laser stepwise ionisation of radionuclide 63Ni by radiation of a dye laser pumped by a copper vapour laser. A three-stage scheme is found with ionisation through an autoionising state (AIS): 3d 84s2 3F4(E = 0) → 3d 94p 1Fo3(31030.99 cm-1) → 3d 94d 2[7/2]4(49322.56 cm-1) → AIS(67707.61 cm-1) which, by employing saturated radiation intensities provides the ionisation selectivity of above 1200 for 63Ni.

  7. Estimating annoyance to calculated wind turbine shadow flicker is improved when variables associated with wind turbine noise exposure are considered.

    PubMed

    Voicescu, Sonia A; Michaud, David S; Feder, Katya; Marro, Leonora; Than, John; Guay, Mireille; Denning, Allison; Bower, Tara; van den Berg, Frits; Broner, Norm; Lavigne, Eric

    2016-03-01

    The Community Noise and Health Study conducted by Health Canada included randomly selected participants aged 18-79 yrs (606 males, 632 females, response rate 78.9%), living between 0.25 and 11.22 km from operational wind turbines. Annoyance to wind turbine noise (WTN) and other features, including shadow flicker (SF) was assessed. The current analysis reports on the degree to which estimating high annoyance to wind turbine shadow flicker (HAWTSF) was improved when variables known to be related to WTN exposure were also considered. As SF exposure increased [calculated as maximum minutes per day (SFm)], HAWTSF increased from 3.8% at 0 ≤ SFm < 10 to 21.1% at SFm ≥ 30, p < 0.0001. For each unit increase in SFm the odds ratio was 2.02 [95% confidence interval: (1.68,2.43)]. Stepwise regression models for HAWTSF had a predictive strength of up to 53% with 10% attributed to SFm. Variables associated with HAWTSF included, but were not limited to, annoyance to other wind turbine-related features, concern for physical safety, and noise sensitivity. Reported dizziness was also retained in the final model at p = 0.0581. Study findings add to the growing science base in this area and may be helpful in identifying factors associated with community reactions to SF exposure from wind turbines.

  8. Assessing the accuracy and stability of variable selection methods for random forest modeling in ecology.

    PubMed

    Fox, Eric W; Hill, Ryan A; Leibowitz, Scott G; Olsen, Anthony R; Thornbrugh, Darren J; Weber, Marc H

    2017-07-01

    Random forest (RF) modeling has emerged as an important statistical learning method in ecology due to its exceptional predictive performance. However, for large and complex ecological data sets, there is limited guidance on variable selection methods for RF modeling. Typically, either a preselected set of predictor variables are used or stepwise procedures are employed which iteratively remove variables according to their importance measures. This paper investigates the application of variable selection methods to RF models for predicting probable biological stream condition. Our motivating data set consists of the good/poor condition of n = 1365 stream survey sites from the 2008/2009 National Rivers and Stream Assessment, and a large set (p = 212) of landscape features from the StreamCat data set as potential predictors. We compare two types of RF models: a full variable set model with all 212 predictors and a reduced variable set model selected using a backward elimination approach. We assess model accuracy using RF's internal out-of-bag estimate, and a cross-validation procedure with validation folds external to the variable selection process. We also assess the stability of the spatial predictions generated by the RF models to changes in the number of predictors and argue that model selection needs to consider both accuracy and stability. The results suggest that RF modeling is robust to the inclusion of many variables of moderate to low importance. We found no substantial improvement in cross-validated accuracy as a result of variable reduction. Moreover, the backward elimination procedure tended to select too few variables and exhibited numerous issues such as upwardly biased out-of-bag accuracy estimates and instabilities in the spatial predictions. We use simulations to further support and generalize results from the analysis of real data. A main purpose of this work is to elucidate issues of model selection bias and instability to ecologists interested in using RF to develop predictive models with large environmental data sets.

  9. Diagnosis of vulnerable atherosclerotic plaques by time-resolved fluorescence spectroscopy and ultrasound imaging.

    PubMed

    Jo, J A; Fang, Q; Papaioannou, T; Qiao, J H; Fishbein, M C; Beseth, B; Dorafshar, A H; Reil, T; Baker, D; Freischlag, J; Shung, K K; Sun, L; Marcu, L

    2006-01-01

    In this study, time-resolved laser-induced fluorescence spectroscopy (TR-LIFS) and ultrasonography were applied to detect vulnerable (high-risk) atherosclerotic plaque. A total of 813 TR-LIFS measurements were taken from carotid plaques of 65 patients, and subsequently analyzed using the Laguerre deconvolution technique. The investigated spots were classified by histopathology as thin, fibrotic, calcified, low-inflamed, inflamed and necrotic lesions. Spectral and time-resolved parameters (normalized intensity values and Laguerre expansion coefficients) were extracted from the TR-LIFS data. Feature selection for classification was performed by either analysis of variance (ANOVA) or principal component analysis (PCA). A stepwise linear discriminant analysis algorithm was developed for detecting inflamed and necrotic lesion, representing the most vulnerable plaques. These vulnerable plaques were detected with high sensitivity (>80%) and specificity (>90%). Ultrasound (US) imaging was obtained in 4 carotid plaques in addition to TR-LIFS examination. Preliminary results indicate that US provides important structural information of the plaques that could be combined with the compositional information obtained by TR-LIFS, to obtain a more accurate diagnosis of vulnerable atherosclerotic plaque.

  10. Theoretical Investigations of Transition Metal Surface Energies under Lattice Strain and CO Environment

    DOE PAGES

    Tang, Michael T.; Ulissi, Zachary W.; Chan, Karen

    2018-05-30

    Here, an understanding of the relative stability of surface facets is crucial to develop predictive models of catalyst activity and to fabricate catalysts with a controlled morphology. In this work, we present a systematic density functional theory study of the effect of lattice strain and CO environment on the surface formation energies of Cu, Pt, and Ni. First, we show that both compressive and tensile lattice strains favor the formation of stepped versus low-index terraces such as (111) and (100). Then, we investigate the effect of the CO environment using configurations of CO at various coverages, determined using a greedy,more » systematic approach, inspired by forward stepwise feature selection. We find that the CO environment favors stepped facets on Ni, Cu, and Pt. These trends are illustrated with the corresponding equilibrium Wulff shapes at various strains and CO pressures. In general, the surface energies of the studied transition metals are highly sensitive to strain and CO coverage, which should be considered when rationalizing trends in the catalytic activity.« less

  11. Theoretical Investigations of Transition Metal Surface Energies under Lattice Strain and CO Environment

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Tang, Michael T.; Ulissi, Zachary W.; Chan, Karen

    Here, an understanding of the relative stability of surface facets is crucial to develop predictive models of catalyst activity and to fabricate catalysts with a controlled morphology. In this work, we present a systematic density functional theory study of the effect of lattice strain and CO environment on the surface formation energies of Cu, Pt, and Ni. First, we show that both compressive and tensile lattice strains favor the formation of stepped versus low-index terraces such as (111) and (100). Then, we investigate the effect of the CO environment using configurations of CO at various coverages, determined using a greedy,more » systematic approach, inspired by forward stepwise feature selection. We find that the CO environment favors stepped facets on Ni, Cu, and Pt. These trends are illustrated with the corresponding equilibrium Wulff shapes at various strains and CO pressures. In general, the surface energies of the studied transition metals are highly sensitive to strain and CO coverage, which should be considered when rationalizing trends in the catalytic activity.« less

  12. Information Scent Determines Attention Allocation and Link Selection among Multiple Information Patches on a Webpage

    ERIC Educational Resources Information Center

    Blackmon, Marilyn Hughes

    2012-01-01

    This paper draws from cognitive psychology and cognitive neuroscience to develop a preliminary similarity-choice theory of how people allocate attention among information patches on webpages while completing search tasks in complex informational websites. Study 1 applied stepwise multiple regression to a large dataset and showed that success rate…

  13. Multiple linear regression analysis

    NASA Technical Reports Server (NTRS)

    Edwards, T. R.

    1980-01-01

    Program rapidly selects best-suited set of coefficients. User supplies only vectors of independent and dependent data and specifies confidence level required. Program uses stepwise statistical procedure for relating minimal set of variables to set of observations; final regression contains only most statistically significant coefficients. Program is written in FORTRAN IV for batch execution and has been implemented on NOVA 1200.

  14. The Relationship of Selected Supply- and Demand-Side Factors to Forms of Perceived Discrimination among Adults with Multiple Sclerosis

    ERIC Educational Resources Information Center

    Roessler, Richard T.; Neath, Jeanne; McMahon, Brian T.; Rumrill, Phillip D.

    2007-01-01

    Single-predictor and stepwise multinomial logistic regression analyses and an external validation were completed on 3,082 allegations of employment discrimination by adults with multiple sclerosis. Women filed two thirds of the allegations, and individuals between 31 and 50 made the vast majority of discrimination charges (73%). Allegations…

  15. Tracing the role of human civilization in the globalization of plant pathogens

    Treesearch

    Alberto Santini; Andrew Liebhold; Duccio Migliorini; Steve Woodward

    2018-01-01

    Co-evolution between plants and parasites, including herbivores and pathogens, has arguably generated much of Earth’s biological diversity. Within an ecosystem, coevolution of plants and pathogens is a stepwise reciprocal evolutionary interaction: epidemics result in intense selection pressures on both host and pathogen populations, ultimately allowing long-term...

  16. A stepwise recovery of metals from hybrid cathodes of spent Li-ion batteries with leaching-flotation-precipitation process

    NASA Astrophysics Data System (ADS)

    Huang, Yanfang; Han, Guihong; Liu, Jiongtian; Chai, Wencui; Wang, Wenjuan; Yang, Shuzhen; Su, Shengpeng

    2016-09-01

    The recovering of valuable metals in spent lithium-ion battery cathodes brings about economic and environmental benefits. A stepwise leaching-flotation-precipitation process is adopted to separate and recover Li/Fe/Mn from the mixed types of cathode materials (hybrid wastes of LiFePO4 and LiMn2O4). The optimal operating conditions for the stepwise recovery process are determined and analyzed by factorial design, thermodynamics calculation, XRD and SEM characterization in this study. First, Li/Fe/Mn ions are released from the cathode using HCl assisted with H2O2 in the acid leaching step. The leachability of metals follows the series Li > Fe > Mn in the acidic environment. Then Fe3+ ions are selectively floated and recovered as FeCl3 from the leachate in the flotation step. Finally, Mn2+/Mn3+ and Li+ ions are sequentially precipitated and separated as MnO2/Mn2O3 and Li3PO4 using saturated KMnO4 solution and hot saturated Na3PO4 solution, respectively. Under the optimized and advisable conditions, the total recovery of Li, Fe and Mn is respectively 80.93 ± 0.16%, 85.40 ± 0.12% and 81.02 ± 0.08%. The purity for lithium, ferrum and manganese compounds is respectively 99.32 ± 0.07%, 97.91 ± 0.05% and 98.73 ± 0.05%. This stepwise process could provide an alternative way for the effective separation and recovery of metal values from spent Li-ion battery cathodes in industry.

  17. Development of a Support Vector Machine - Based Image Analysis System for Focal Liver Lesions Classification in Magnetic Resonance Images

    NASA Astrophysics Data System (ADS)

    Gatos, I.; Tsantis, S.; Karamesini, M.; Skouroliakou, A.; Kagadis, G.

    2015-09-01

    Purpose: The design and implementation of a computer-based image analysis system employing the support vector machine (SVM) classifier system for the classification of Focal Liver Lesions (FLLs) on routine non-enhanced, T2-weighted Magnetic Resonance (MR) images. Materials and Methods: The study comprised 92 patients; each one of them has undergone MRI performed on a Magnetom Concerto (Siemens). Typical signs on dynamic contrast-enhanced MRI and biopsies were employed towards a three class categorization of the 92 cases: 40-benign FLLs, 25-Hepatocellular Carcinomas (HCC) within Cirrhotic liver parenchyma and 27-liver metastases from Non-Cirrhotic liver. Prior to FLLs classification an automated lesion segmentation algorithm based on Marcov Random Fields was employed in order to acquire each FLL Region of Interest. 42 texture features derived from the gray-level histogram, co-occurrence and run-length matrices and 12 morphological features were obtained from each lesion. Stepwise multi-linear regression analysis was utilized to avoid feature redundancy leading to a feature subset that fed the multiclass SVM classifier designed for lesion classification. SVM System evaluation was performed by means of leave-one-out method and ROC analysis. Results: Maximum accuracy for all three classes (90.0%) was obtained by means of the Radial Basis Kernel Function and three textural features (Inverse- Different-Moment, Sum-Variance and Long-Run-Emphasis) that describe lesion's contrast, variability and shape complexity. Sensitivity values for the three classes were 92.5%, 81.5% and 96.2% respectively, whereas specificity values were 94.2%, 95.3% and 95.5%. The AUC value achieved for the selected subset was 0.89 with 0.81 - 0.94 confidence interval. Conclusion: The proposed SVM system exhibit promising results that could be utilized as a second opinion tool to the radiologist in order to decrease the time/cost of diagnosis and the need for patients to undergo invasive examination.

  18. Assessment of global and local region-based bilateral mammographic feature asymmetry to predict short-term breast cancer risk

    NASA Astrophysics Data System (ADS)

    Li, Yane; Fan, Ming; Cheng, Hu; Zhang, Peng; Zheng, Bin; Li, Lihua

    2018-01-01

    This study aims to develop and test a new imaging marker-based short-term breast cancer risk prediction model. An age-matched dataset of 566 screening mammography cases was used. All ‘prior’ images acquired in the two screening series were negative, while in the ‘current’ screening images, 283 cases were positive for cancer and 283 cases remained negative. For each case, two bilateral cranio-caudal view mammograms acquired from the ‘prior’ negative screenings were selected and processed by a computer-aided image processing scheme, which segmented the entire breast area into nine strip-based local regions, extracted the element regions using difference of Gaussian filters, and computed both global- and local-based bilateral asymmetrical image features. An initial feature pool included 190 features related to the spatial distribution and structural similarity of grayscale values, as well as of the magnitude and phase responses of multidirectional Gabor filters. Next, a short-term breast cancer risk prediction model based on a generalized linear model was built using an embedded stepwise regression analysis method to select features and a leave-one-case-out cross-validation method to predict the likelihood of each woman having image-detectable cancer in the next sequential mammography screening. The area under the receiver operating characteristic curve (AUC) values significantly increased from 0.5863  ±  0.0237 to 0.6870  ±  0.0220 when the model trained by the image features extracted from the global regions and by the features extracted from both the global and the matched local regions (p  =  0.0001). The odds ratio values monotonically increased from 1.00-8.11 with a significantly increasing trend in slope (p  =  0.0028) as the model-generated risk score increased. In addition, the AUC values were 0.6555  ±  0.0437, 0.6958  ±  0.0290, and 0.7054  ±  0.0529 for the three age groups of 37-49, 50-65, and 66-87 years old, respectively. AUC values of 0.6529  ±  0.1100, 0.6820  ±  0.0353, 0.6836  ±  0.0302 and 0.8043  ±  0.1067 were yielded for the four mammography density sub-groups (BIRADS from 1-4), respectively. This study demonstrated that bilateral asymmetry features extracted from local regions combined with the global region in bilateral negative mammograms could be used as a new imaging marker to assist in the prediction of short-term breast cancer risk.

  19. Strategy for design NIR calibration sets based on process spectrum and model space: An innovative approach for process analytical technology.

    PubMed

    Cárdenas, V; Cordobés, M; Blanco, M; Alcalà, M

    2015-10-10

    The pharmaceutical industry is under stringent regulations on quality control of their products because is critical for both, productive process and consumer safety. According to the framework of "process analytical technology" (PAT), a complete understanding of the process and a stepwise monitoring of manufacturing are required. Near infrared spectroscopy (NIRS) combined with chemometrics have lately performed efficient, useful and robust for pharmaceutical analysis. One crucial step in developing effective NIRS-based methodologies is selecting an appropriate calibration set to construct models affording accurate predictions. In this work, we developed calibration models for a pharmaceutical formulation during its three manufacturing stages: blending, compaction and coating. A novel methodology is proposed for selecting the calibration set -"process spectrum"-, into which physical changes in the samples at each stage are algebraically incorporated. Also, we established a "model space" defined by Hotelling's T(2) and Q-residuals statistics for outlier identification - inside/outside the defined space - in order to select objectively the factors to be used in calibration set construction. The results obtained confirm the efficacy of the proposed methodology for stepwise pharmaceutical quality control, and the relevance of the study as a guideline for the implementation of this easy and fast methodology in the pharma industry. Copyright © 2015 Elsevier B.V. All rights reserved.

  20. The influence of attention deficits on functional recovery post stroke during the first 12 months after discharge from hospital.

    PubMed

    Hyndman, D; Pickering, R M; Ashburn, A

    2008-06-01

    Attention deficits have been linked to poor recovery after stroke and may predict outcome. We explored the influence of attention on functional recovery post stroke in the first 12 months after discharge from hospital. People with stroke completed measures of attention, balance, mobility and activities of daily living (ADL) ability at the point of discharge from hospital, and 6 and 12 months later. We used correlational analysis and stepwise linear regression to explore potential predictors of outcome. We recruited 122 men and women, mean age 70 years. At discharge, 56 (51%) had deficits of divided attention, 45 (37%) of sustained attention, 43 (36%) of auditory selective attention and 41 (37%) had visual selective attention deficits. Attention at discharge correlated with mobility, balance and ADL outcomes 12 months later. After controlling for the level of the outcome at discharge, correlations remained significant in only five of the 12 relationships. Stepwise linear regression revealed that the outcome measured at discharge, days until discharge and number of medications were better predictors of outcome: in no case was an attention variable at discharge selected as a predictor of outcome at 12 months. Although attention and function correlated significantly, this correlation was reduced after controlling for functional ability at discharge. Furthermore, side of lesion and the attention variables were not demonstrated as important predictors of outcome 12 months later.

  1. Students' Achievement, Skill and Confidence in Using Stepwise Problem-Solving Strategies

    ERIC Educational Resources Information Center

    Gok, Tolga

    2014-01-01

    The main purpose of this study was to examine the effects of Problem-Solving Strategy Steps (PSSS) on students' achievement, skill, and confidence. The study was conducted in a two-year college classroom with 70 students from two different groups enrolled in a physics course. One of them was randomly selected as an experimental group (EG) and the…

  2. Supramolecular self-assembly of heterobimetallic complexes: a new N,P-based, selective heteroditopic ligand.

    PubMed

    Hutchinson, Daniel John; Clauss, Reike; Sárosi, Menyhárt-Botond; Hey-Hawkins, Evamarie

    2018-01-23

    Pyrimidine-hydrazone and phosphole architectures have been combined to create a new heteroditopic ligand capable of forming heterobimetallic Zn II /Pd II , Pb II /Pd II and Cu II /Pd II complexes in high yielding stepwise or one pot reactions. The catalytic activity of these complexes in Heck coupling and Miyaura borylation reactions was investigated.

  3. Factors Associated with Level of Living in Washington County, Mississippi. Technical Bulletin No. 1501.

    ERIC Educational Resources Information Center

    McCoy, John L.

    Step-wise multiple regression and typological analysis were used to analyze the extent to which selected factors influence vertical mobility and achieved level of living. A sample of 418 male household heads who were 18 to 45 years old in Washington County, Mississippi were interviewed during 1971. A prescreening using census and local housing…

  4. Improved Variable Selection Algorithm Using a LASSO-Type Penalty, with an Application to Assessing Hepatitis B Infection Relevant Factors in Community Residents

    PubMed Central

    Guo, Pi; Zeng, Fangfang; Hu, Xiaomin; Zhang, Dingmei; Zhu, Shuming; Deng, Yu; Hao, Yuantao

    2015-01-01

    Objectives In epidemiological studies, it is important to identify independent associations between collective exposures and a health outcome. The current stepwise selection technique ignores stochastic errors and suffers from a lack of stability. The alternative LASSO-penalized regression model can be applied to detect significant predictors from a pool of candidate variables. However, this technique is prone to false positives and tends to create excessive biases. It remains challenging to develop robust variable selection methods and enhance predictability. Material and methods Two improved algorithms denoted the two-stage hybrid and bootstrap ranking procedures, both using a LASSO-type penalty, were developed for epidemiological association analysis. The performance of the proposed procedures and other methods including conventional LASSO, Bolasso, stepwise and stability selection models were evaluated using intensive simulation. In addition, methods were compared by using an empirical analysis based on large-scale survey data of hepatitis B infection-relevant factors among Guangdong residents. Results The proposed procedures produced comparable or less biased selection results when compared to conventional variable selection models. In total, the two newly proposed procedures were stable with respect to various scenarios of simulation, demonstrating a higher power and a lower false positive rate during variable selection than the compared methods. In empirical analysis, the proposed procedures yielding a sparse set of hepatitis B infection-relevant factors gave the best predictive performance and showed that the procedures were able to select a more stringent set of factors. The individual history of hepatitis B vaccination, family and individual history of hepatitis B infection were associated with hepatitis B infection in the studied residents according to the proposed procedures. Conclusions The newly proposed procedures improve the identification of significant variables and enable us to derive a new insight into epidemiological association analysis. PMID:26214802

  5. Demagnetization Analysis in Excel (DAIE) - An open source workbook in Excel for viewing and analyzing demagnetization data from paleomagnetic discrete samples and u-channels

    NASA Astrophysics Data System (ADS)

    Sagnotti, Leonardo

    2013-04-01

    Modern rock magnetometers and stepwise demagnetization procedures result in the production of large datasets, which need a versatile and fast software for their display and analysis. Various software packages for paleomagnetic analyses have been recently developed to overcome the problems linked to the limited capability and the loss of operability of early codes written in obsolete computer languages and/or platforms, not compatible with modern 64 bit processors. The Demagnetization Analysis in Excel (DAIE) workbook is a new software that has been designed to make the analysis of demagnetization data easy and accessible on an application (Microsoft Excel) widely diffused and available on both the Microsoft Windows and Mac OS X operating systems. The widespread diffusion of Excel should guarantee a long term working life, since compatibility and functionality of current Excel files should be most likely maintained during the development of new processors and operating systems. DAIE is designed for viewing and analyzing stepwise demagnetization data of both discrete and u-channel samples. DAIE consists of a single file and has an open modular structure organized in 10 distinct worksheets. The standard demagnetization diagrams and various parameters of common use are shown on the same worksheet including selectable parameters and user's choices. The remanence characteristic components may be computed by principal component analysis (PCA) on a selected interval of demagnetization steps. Saving of the PCA data can be done both sample by sample, or in automatic by applying the selected choices to all the samples included in the file. The DAIE open structure allows easy personalization, development and improvement. The workbook has the following features which may be valuable for various users: - Operability in nearly all the computers and platforms; - Easy inputs of demagnetization data by "copy and paste" from ASCII files; - Easy export of computed parameters and demagnetization plots; - Complete control of the whole workflow and possibility of implementation of the workbook by any user; - Modular structure in distinct worksheets for each type of analyses and plots, in order to make implementation and personalization easier; - Opportunity to use the workbook for educational purposes, since all the computations and analyses are easily traceable and accessible; - Automatic and fast analysis of a large batch of demagnetization data, such as those measured on u-channel samples. The DAIE workbook and the "User manual" are available for download on a dedicated web site (http://roma2.rm.ingv.it/en/facilities/software/49/daie).

  6. Step-wise loss of antidepressant effectiveness with repeated antidepressant trials in bipolar II depression.

    PubMed

    Amsterdam, Jay D; Lorenzo-Luaces, Lorenzo; DeRubeis, Robert J

    2016-11-01

    This study examined the relationship between the number of prior antidepressant treatment trials and step-wise increase in pharmacodynamic tolerance (or progressive loss of effectiveness) in subjects with bipolar II depression. Subjects ≥18 years old with bipolar II depression (n=129) were randomized to double-blind venlafaxine or lithium carbonate monotherapy for 12 weeks. Responders (n=59) received continuation monotherapy for six additional months. After controlling for baseline covariates of prior medications, there was a 25% reduction in the likelihood of response to treatment with each increase in the number of prior antidepressant trials (odds ratio [OR]=0.75, unstandardized coefficient [B]=-0.29, standard error (SE)=0.12; χ 2 =5.70, P<.02], as well as a 32% reduction in the likelihood of remission with each prior antidepressant trial (OR=0.68, B=-0.39, SE=0.13; χ 2 =9.71, P=.002). This step-wise increase in pharmacodynamic tolerance occurred in both treatment conditions. Prior selective serotonin reuptake inhibitor (SSRI) therapy was specifically associated with a step-wise increase in tolerance, whereas other prior antidepressants or mood stabilizers were not associated with pharmacodynamic tolerance. Neither the number of prior antidepressants, nor the number of prior SSRIs, or mood stabilizers, were associated with an increase in relapse during continuation therapy. The odds of responding or remitting during venlafaxine or lithium monotherapy were reduced by 25% and 32%, respectively, with each increase in the number of prior antidepressant treatment trials. There was no relationship between prior antidepressant exposure and depressive relapse during continuation therapy of bipolar II disorder. © 2016 John Wiley & Sons A/S. Published by John Wiley & Sons Ltd.

  7. Stepwise cycloaddition reaction of N-phenacylbenzothiazolium bromides and nitroalkenes for tetrahydro-, dihydro- and benzo[d]pyrrolo[2,1-b]thiazoles

    NASA Astrophysics Data System (ADS)

    Jin, Gong; Sun, Jing; Yang, Ren-Yin; Yan, Chao-Guo

    2017-04-01

    The triethylamine promoted stepwise 1,3-dipolar cycloaddition reaction of N-phenacylbenzothiazolium bromides with nitroalkenes in ethanol resulted in a mixture of two isomeric tetrahydrobenzo[d]pyrrolo[2,1-b]thiazoles with cis/trans/cis- and all-trans-configurations. More importantly, the corresponding dihydrobenzo[d]pyrrolo[2,1-b]thiazoles can be selectively prepared in refluxing ethanol and the benzo[d]pyrrolo[2,1-b]thiazoles can be obtained in satisfactory yields by sequential dehydrogenation with DDQ as oxidizer. On the other hand, the similar cycloaddition reaction of N-phenacylbenzothiazolium bromides with 1-methy-1-nitroalkenes in refluxing ethanol afforded benzo[d]pyrrolo[2,1-b]thiazoles with splitting out of nitro group. The stereochemistry of the spiro compounds was clearly elucidated on the basis of NMR spectra and sixteen single crystal structures.

  8. Controlled, Stepwise Reduction and Band Gap Manipulation of Graphene Oxide.

    PubMed

    Mathkar, Akshay; Tozier, Dylan; Cox, Paris; Ong, Peijie; Galande, Charudatta; Balakrishnan, Kaushik; Leela Mohana Reddy, Arava; Ajayan, Pulickel M

    2012-04-19

    Graphene oxide (GO) has drawn tremendous interest as a tunable precursor in numerous areas, due to its readily manipulable surface. However, its inhomogeneous and nonstoichiometric structure makes achieving chemical control a major challenge. Here, we present a room-temperature based, controlled method for the stepwise reduction of GO, with evidence of sequential removal of each organic moiety. By analyzing signature infrared absorption frequencies, we identify the carbonyl group as the first to be reduced, while the tertiary alcohol takes the longest to be completely removed from the GO surface. Controlled reduction allows for progressive tuning of the optical gap from 3.5 eV down to 1 eV, while XPS spectra show a concurrent increase in the C/O ratio. This study is the first step toward selectively enhancing the chemical homogeneity of GO, thus providing greater control over its structure, and elucidating the order of removal of functional groups and hydrazine-vapor reduction.

  9. High drug loading self-microemulsifying/micelle formulation: design by high-throughput formulation screening system and in vivo evaluation.

    PubMed

    Sakai, Kenichi; Obata, Kouki; Yoshikawa, Mayumi; Takano, Ryusuke; Shibata, Masaki; Maeda, Hiroyuki; Mizutani, Akihiko; Terada, Katsuhide

    2012-10-01

    To design a high drug loading formulation of self-microemulsifying/micelle system. A poorly-soluble model drug (CH5137291), 8 hydrophilic surfactants (HS), 10 lipophilic surfactants (LS), 5 oils, and PEG400 were used. A high loading formulation was designed by a following stepwise approach using a high-throughput formulation screening (HTFS) system: (1) an oil/solvent was selected by solubility of the drug; (2) a suitable HS for highly loading was selected by the screenings of emulsion/micelle size and phase stability in binary systems (HS, oil/solvent) with increasing loading levels; (3) a LS that formed a broad SMEDDS/micelle area on a phase diagram containing the HS and oil/solvent was selected by the same screenings; (4) an optimized formulation was selected by evaluating the loading capacity of the crystalline drug. Aqueous solubility behavior and oral absorption (Beagle dog) of the optimized formulation were compared with conventional formulations (jet-milled, PEG400). As an optimized formulation, d-α-tocopheryl polyoxyethylene 1000 succinic ester: PEG400 = 8:2 was selected, and achieved the target loading level (200 mg/mL). The formulation formed fine emulsion/micelle (49.1 nm), and generated and maintained a supersaturated state at a higher level compared with the conventional formulations. In the oral absorption test, the area under the plasma concentration-time curve of the optimized formulation was 16.5-fold higher than that of the jet-milled formulation. The high loading formulation designed by the stepwise approach using the HTFS system improved the oral absorption of the poorly-soluble model drug.

  10. [Fast optimization of stepwise gradient conditions for ternary mobile phase in reversed-phase high performance liquid chromatography].

    PubMed

    Shan, Yi-chu; Zhang, Yu-kui; Zhao, Rui-huan

    2002-07-01

    In high performance liquid chromatography, it is necessary to apply multi-composition gradient elution for the separation of complex samples such as environmental and biological samples. Multivariate stepwise gradient elution is one of the most efficient elution modes, because it combines the high selectivity of multi-composition mobile phase and shorter analysis time of gradient elution. In practical separations, the separation selectivity of samples can be effectively adjusted by using ternary mobile phase. For the optimization of these parameters, the retention equation of samples must be obtained at first. Traditionally, several isocratic experiments are used to get the retention equation of solute. However, it is time consuming especially for the separation of complex samples with a wide range of polarity. A new method for the fast optimization of ternary stepwise gradient elution was proposed based on the migration rule of solute in column. First, the coefficients of retention equation of solute are obtained by running several linear gradient experiments, then the optimal separation conditions are searched according to the hierarchical chromatography response function which acts as the optimization criterion. For each kind of organic modifier, two initial linear gradient experiments are used to obtain the primary coefficients of retention equation of each solute. For ternary mobile phase, only four linear gradient runs are needed to get the coefficients of retention equation. Then the retention times of solutes under arbitrary mobile phase composition can be predicted. The initial optimal mobile phase composition is obtained by resolution mapping for all of the solutes. A hierarchical chromatography response function is used to evaluate the separation efficiencies and search the optimal elution conditions. In subsequent optimization, the migrating distance of solute in the column is considered to decide the mobile phase composition and sustaining time of the latter steps until all the solutes are eluted out. Thus the first stepwise gradient elution conditions are predicted. If the resolution of samples under the predicted optimal separation conditions is satisfactory, the optimization procedure is stopped; otherwise, the coefficients of retention equation are adjusted according to the experimental results under the previously predicted elution conditions. Then the new stepwise gradient elution conditions are predicted repeatedly until satisfactory resolution is obtained. Normally, the satisfactory separation conditions can be found only after six experiments by using the proposed method. In comparison with the traditional optimization method, the time needed to finish the optimization procedure can be greatly reduced. The method has been validated by its application to the separation of several samples such as amino acid derivatives, aromatic amines, in which satisfactory separations were obtained with predicted resolution.

  11. Sequential ALK Inhibitors Can Select for Lorlatinib-Resistant Compound ALK Mutations in ALK-Positive Lung Cancer.

    PubMed

    Yoda, Satoshi; Lin, Jessica J; Lawrence, Michael S; Burke, Benjamin J; Friboulet, Luc; Langenbucher, Adam; Dardaei, Leila; Prutisto-Chang, Kylie; Dagogo-Jack, Ibiayi; Timofeevski, Sergei; Hubbeling, Harper; Gainor, Justin F; Ferris, Lorin A; Riley, Amanda K; Kattermann, Krystina E; Timonina, Daria; Heist, Rebecca S; Iafrate, A John; Benes, Cyril H; Lennerz, Jochen K; Mino-Kenudson, Mari; Engelman, Jeffrey A; Johnson, Ted W; Hata, Aaron N; Shaw, Alice T

    2018-06-01

    The cornerstone of treatment for advanced ALK-positive lung cancer is sequential therapy with increasingly potent and selective ALK inhibitors. The third-generation ALK inhibitor lorlatinib has demonstrated clinical activity in patients who failed previous ALK inhibitors. To define the spectrum of ALK mutations that confer lorlatinib resistance, we performed accelerated mutagenesis screening of Ba/F3 cells expressing EML4-ALK. Under comparable conditions, N -ethyl- N -nitrosourea (ENU) mutagenesis generated numerous crizotinib-resistant but no lorlatinib-resistant clones harboring single ALK mutations. In similar screens with EML4-ALK containing single ALK resistance mutations, numerous lorlatinib-resistant clones emerged harboring compound ALK mutations. To determine the clinical relevance of these mutations, we analyzed repeat biopsies from lorlatinib-resistant patients. Seven of 20 samples (35%) harbored compound ALK mutations, including two identified in the ENU screen. Whole-exome sequencing in three cases confirmed the stepwise accumulation of ALK mutations during sequential treatment. These results suggest that sequential ALK inhibitors can foster the emergence of compound ALK mutations, identification of which is critical to informing drug design and developing effective therapeutic strategies. Significance: Treatment with sequential first-, second-, and third-generation ALK inhibitors can select for compound ALK mutations that confer high-level resistance to ALK-targeted therapies. A more efficacious long-term strategy may be up-front treatment with a third-generation ALK inhibitor to prevent the emergence of on-target resistance. Cancer Discov; 8(6); 714-29. ©2018 AACR. This article is highlighted in the In This Issue feature, p. 663 . ©2018 American Association for Cancer Research.

  12. Stepwise pumping approach to improve free phase light hydrocarbon recovery from unconfined aquifers

    NASA Astrophysics Data System (ADS)

    Cooper, Grant S.; Peralta, Richard C.; Kaluarachchi, Jagath J.

    1995-04-01

    A stepwise, time-varying pumping approach is developed to improve free phase oil recovery of light non-aqueous phase liquids (LNAPL) from a homogeneous, unconfined aquifer. Stepwise pumping is used to contain the floating oil plume and obtain efficient free oil recovery. The graphical plots. The approach uses ARMOS ©, an areal two-dimensional multiphase flow, finite-element simulation model. Systematic simulations of free oil area changes to pumping rates are analyzed. Pumping rates are determined that achieve LNAPL plume containment at different times (i.e. 90, 180 and 360 days) for a planning period of 360 days. These pumping rates are used in reverse order as a stepwise (monotonically increasing) pumping strategy. This stepwise pumping strategy is analyzed further by performing additional simulations at different pumping rates for the last pumping period. The final stepwise pumping strategy is varied by factors of -25% and +30% to evaluate sensitivity in the free oil recovery process. Stepwise pumping is compared to steady pumping rates to determine the best free oil recovery strategy. Stepwise pumping is shown to improve oil recovery by increasing recoveredoil volume (11%) and decreasing residual oil (15%) when compared with traditional steady pumping strategies. The best stepwise pumping strategy recovers more free oil by reducing the amount of residual oil left in the system due to pumping drawdown. This stepwise pumping pproach can be used to enhance free oil recovery and provide for cost-effective design and management of LNAPL cleanup.

  13. Single neuropsychological test scores associated with rate of cognitive decline in early Alzheimer disease.

    PubMed

    Parikh, Mili; Hynan, Linda S; Weiner, Myron F; Lacritz, Laura; Ringe, Wendy; Cullum, C Munro

    2014-01-01

    Alzheimer disease (AD) characteristically begins with episodic memory impairment followed by other cognitive deficits; however, the course of illness varies, with substantial differences in the rate of cognitive decline. For research and clinical purposes it would be useful to distinguish between persons who will progress slowly from persons who will progress at an average or faster rate. Our objective was to use neurocognitive performance features and disease-specific and health information to determine a predictive model for the rate of cognitive decline in participants with mild AD. We reviewed the records of a series of 96 consecutive participants with mild AD from 1995 to 2011 who had been administered selected neurocognitive tests and clinical measures. Based on Clinical Dementia Rating (CDR) of functional and cognitive decline over 2 years, participants were classified as Faster (n = 45) or Slower (n = 51) Progressors. Stepwise logistic regression analyses using neurocognitive performance features, disease-specific, health, and demographic variables were performed. Neuropsychological scores that distinguished Faster from Slower Progressors included Trail Making Test - A, Digit Symbol, and California Verbal Learning Test (CVLT) Total Learned and Primacy Recall. No disease-specific, health, or demographic variable predicted rate of progression; however, history of heart disease showed a trend. Among the neuropsychological variables, Trail Making Test - A best distinguished Faster from Slower Progressors, with an overall accuracy of 68%. In an omnibus model including neuropsychological, disease-specific, health, and demographic variables, only Trail Making Test - A distinguished between groups. Several neuropsychological performance features were associated with the rate of cognitive decline in mild AD, with baseline Trail Making Test - A performance best separating those who declined at an average or faster rate from those who showed slower progression.

  14. Comparison of statistical methods for detection of serum lipid biomarkers for mesothelioma and asbestos exposure.

    PubMed

    Xu, Rengyi; Mesaros, Clementina; Weng, Liwei; Snyder, Nathaniel W; Vachani, Anil; Blair, Ian A; Hwang, Wei-Ting

    2017-07-01

    We compared three statistical methods in selecting a panel of serum lipid biomarkers for mesothelioma and asbestos exposure. Serum samples from mesothelioma, asbestos-exposed subjects and controls (40 per group) were analyzed. Three variable selection methods were considered: top-ranked predictors from univariate model, stepwise and least absolute shrinkage and selection operator. Crossed-validated area under the receiver operating characteristic curve was used to compare the prediction performance. Lipids with high crossed-validated area under the curve were identified. Lipid with mass-to-charge ratio of 372.31 was selected by all three methods comparing mesothelioma versus control. Lipids with mass-to-charge ratio of 1464.80 and 329.21 were selected by two models for asbestos exposure versus control. Different methods selected a similar set of serum lipids. Combining candidate biomarkers can improve prediction.

  15. In-Line Sorting of Harumanis Mango Based on External Quality Using Visible Imaging

    PubMed Central

    Ibrahim, Mohd Firdaus; Ahmad Sa’ad, Fathinul Syahir; Zakaria, Ammar; Md Shakaff, Ali Yeon

    2016-01-01

    The conventional method of grading Harumanis mango is time-consuming, costly and affected by human bias. In this research, an in-line system was developed to classify Harumanis mango using computer vision. The system was able to identify the irregularity of mango shape and its estimated mass. A group of images of mangoes of different size and shape was used as database set. Some important features such as length, height, centroid and parameter were extracted from each image. Fourier descriptor and size-shape parameters were used to describe the mango shape while the disk method was used to estimate the mass of the mango. Four features have been selected by stepwise discriminant analysis which was effective in sorting regular and misshapen mango. The volume from water displacement method was compared with the volume estimated by image processing using paired t-test and Bland-Altman method. The result between both measurements was not significantly different (P > 0.05). The average correct classification for shape classification was 98% for a training set composed of 180 mangoes. The data was validated with another testing set consist of 140 mangoes which have the success rate of 92%. The same set was used for evaluating the performance of mass estimation. The average success rate of the classification for grading based on its mass was 94%. The results indicate that the in-line sorting system using machine vision has a great potential in automatic fruit sorting according to its shape and mass. PMID:27801799

  16. In-Line Sorting of Harumanis Mango Based on External Quality Using Visible Imaging.

    PubMed

    Ibrahim, Mohd Firdaus; Ahmad Sa'ad, Fathinul Syahir; Zakaria, Ammar; Md Shakaff, Ali Yeon

    2016-10-27

    The conventional method of grading Harumanis mango is time-consuming, costly and affected by human bias. In this research, an in-line system was developed to classify Harumanis mango using computer vision. The system was able to identify the irregularity of mango shape and its estimated mass. A group of images of mangoes of different size and shape was used as database set. Some important features such as length, height, centroid and parameter were extracted from each image. Fourier descriptor and size-shape parameters were used to describe the mango shape while the disk method was used to estimate the mass of the mango. Four features have been selected by stepwise discriminant analysis which was effective in sorting regular and misshapen mango. The volume from water displacement method was compared with the volume estimated by image processing using paired t -test and Bland-Altman method. The result between both measurements was not significantly different (P > 0.05). The average correct classification for shape classification was 98% for a training set composed of 180 mangoes. The data was validated with another testing set consist of 140 mangoes which have the success rate of 92%. The same set was used for evaluating the performance of mass estimation. The average success rate of the classification for grading based on its mass was 94%. The results indicate that the in-line sorting system using machine vision has a great potential in automatic fruit sorting according to its shape and mass.

  17. Gynecomastia: a common indication for mammography in men of all age.

    PubMed

    Capasso, Raffaella; Sica, A; D'Amora, M; Mostardi, Maurizio; Martella, Ilenia; Totaro, Marilina; Della Casa, Giovanni; Vallara, Manuela; Pesce, Antonella; Gatta, G; Cappabianca, S

    2016-07-28

    Gynecomastia (GM) is the most frequent cause of male breast-related signs and symptoms and represents also the most common indication for mammography (MX) in men. In this article, our 7-year long experience with MX in men suffering from GM is reviewed, and the mammographic features of GM are presented. MXs performed in male patients at our institution from January 2009 to January 2016 were retrospectively reviewed and patients with mammographic features of GM were selected. Informed consent was waived by the local institutional review board given the retrospective nature of the study. Mammograms were performed in both cranio-caudal (CC) and medio-lateral-oblique (MLO) views according to diagnostic needs. Clinical and pathologic data were obtained by review of patient charts. 37 males (aged between 13-79 years, mean 59 years) referred for MX at our institution because of palpable lump (31/37; 83.8%), breast enlargement (33/37; 89.2%), tenderness or pain (25/37; 67.6%). Of the 37 patients evaluated, 32 (86.5%) had true GM while 5 (13.5%) had pseudoGM. The evaluation of GM can be complex but a stepwise approach that starts with careful history taking and physical examination may obviate the need for extensive work-up. In this context, MX has been shown to be an accurate diagnostic tool for detecting GM and should be the first imaging examination to be performed in all clinically suspicious lesions referred for imaging.

  18. Variation of facial features among three African populations: Body height match analyses.

    PubMed

    Taura, M G; Adamu, L H; Gudaji, A

    2017-01-01

    Body height is one of the variables that show a correlation with facial craniometry. Here we seek to discriminate the three populations (Nigerians, Ugandans and Kenyans) using facial craniometry based on different categories of body height of adult males. A total of 513 individuals comprising 234 Nigerians, 169 Ugandans and 110 Kenyans with mean age of 25.27, s=5.13 (18-40 years) participated. Paired and unpaired facial features were measured using direct craniometry. Multivariate and stepwise discriminate function analyses were used for differentiation of the three populations. The result showed significant overall facial differences among the three populations in all the body height categories. Skull height, total facial height, outer canthal distance, exophthalmometry, right ear width and nasal length were significantly different among the three different populations irrespective of body height categories. Other variables were sensitive to body height. Stepwise discriminant function analyses included maximum of six variables for better discrimination between the three populations. The single best discriminator of the groups was total facial height, however, for body height >1.70m the single best discriminator was nasal length. Most of the variables were better used with function 1, hence, better discrimination than function 2. In conclusion, adult body height in addition to other factors such as age, sex, and ethnicity should be considered in making decision on facial craniometry. However, not all the facial linear dimensions were sensitive to body height. Copyright © 2016 Elsevier GmbH. All rights reserved.

  19. Model selection bias and Freedman's paradox

    USGS Publications Warehouse

    Lukacs, P.M.; Burnham, K.P.; Anderson, D.R.

    2010-01-01

    In situations where limited knowledge of a system exists and the ratio of data points to variables is small, variable selection methods can often be misleading. Freedman (Am Stat 37:152-155, 1983) demonstrated how common it is to select completely unrelated variables as highly "significant" when the number of data points is similar in magnitude to the number of variables. A new type of model averaging estimator based on model selection with Akaike's AIC is used with linear regression to investigate the problems of likely inclusion of spurious effects and model selection bias, the bias introduced while using the data to select a single seemingly "best" model from a (often large) set of models employing many predictor variables. The new model averaging estimator helps reduce these problems and provides confidence interval coverage at the nominal level while traditional stepwise selection has poor inferential properties. ?? The Institute of Statistical Mathematics, Tokyo 2009.

  20. Personality features, dissociation, self-stigma, hope, and the complex treatment of depressive disorder

    PubMed Central

    Prasko, Jan; Ociskova, Marie; Grambal, Ales; Sigmundova, Zuzana; Kasalova, Petra; Marackova, Marketa; Holubova, Michaela; Vrbova, Kristyna; Latalova, Klara; Slepecky, Milos

    2016-01-01

    Objective Identifying the predictors of response to psychiatric and psychotherapeutic treatments may be useful for increasing treatment efficacy in pharmacoresistant depressive patients. The goal of this study was to examine the influence of dissociation, hope, personality trait, and selected demographic factors in treatment response of this group of patients. Methods Pharmacoresistant depressive inpatients were enrolled in the study. All patients completed Clinical Global Impression – both objective and subjective form (CGI), Beck Depression Inventory (BDI), and Beck Anxiety Inventory (BAI) at baseline and after 6 weeks of combined pharmacotherapy and psychotherapy (group cognitive-behavioral or group psychodynamic) treatment as an outcome measures. The Internalized Stigma of Mental Illness Scale (ISMI), Dissociative Experience Scale (DES), Adult Dispositional Hope Scale (ADHS), and Temperament and Character Inventory (TCI-R) were completed at the start of the treatment with the intention to find the predictors of treatment efficacy. Results The study included 72 patients who were hospitalized for the pharmacoresistant major depression; 63 of them completed the study. The mean scores of BDI-II, BAI, subjCGI, and objCGI significantly decreased during the treatment. BDI-II relative change statistically significantly correlated with the total ISMI score, Discrimination Experience (ISMI subscale), and Harm Avoidance (TCI-R personality trait). According to stepwise regression, the strongest factors connected to BDI-II relative change were the duration of the disorder and Discrimination Experience (domain of ISMI). ObjCGI relative change significantly correlated with the level of dissociation (DES), the total ISMI score, hope in ADHS total score, and Self-Directedness (TCI-R). According to stepwise regression, the strongest factor connected to objCGI relative change was Discrimination Experience (domain of ISMI). The existence of comorbid personality disorder did not influence the treatment response. Conclusion According to the results of the present study, patients with pharmacoresistant depressive disorders, who have had more experience with discrimination because of their mental struggles, showed a poorer response to treatment. PMID:27785031

  1. Modeling of sorption processes on solid-phase ion-exchangers

    NASA Astrophysics Data System (ADS)

    Dorofeeva, Ludmila; Kuan, Nguyen Anh

    2018-03-01

    Research of alkaline elements separation on solid-phase ion-exchangers is carried out to define the selectivity coefficients and height of an equivalent theoretical stage for both continuous and stepwise filling of column by ionite. On inorganic selective sorbents the increase in isotope enrichment factor up to 0.0127 is received. Also, parametrical models that are adequately describing dependence of the pressure difference and the magnitude expansion in the ion-exchange layer from the flow rate and temperature have been obtained. The concentration rate value under the optimum realization conditions of process and depending on type of a selective material changes in a range 1.021÷1.092. Calculated results show agreement with experimental data.

  2. Statistical learning and selective inference.

    PubMed

    Taylor, Jonathan; Tibshirani, Robert J

    2015-06-23

    We describe the problem of "selective inference." This addresses the following challenge: Having mined a set of data to find potential associations, how do we properly assess the strength of these associations? The fact that we have "cherry-picked"--searched for the strongest associations--means that we must set a higher bar for declaring significant the associations that we see. This challenge becomes more important in the era of big data and complex statistical modeling. The cherry tree (dataset) can be very large and the tools for cherry picking (statistical learning methods) are now very sophisticated. We describe some recent new developments in selective inference and illustrate their use in forward stepwise regression, the lasso, and principal components analysis.

  3. Selective cleavage of the C(α)-C(β) linkage in lignin model compounds via Baeyer-Villiger oxidation.

    PubMed

    Patil, Nikhil D; Yao, Soledad G; Meier, Mark S; Mobley, Justin K; Crocker, Mark

    2015-03-21

    Lignin is an amorphous aromatic polymer derived from plants and is a potential source of fuels and bulk chemicals. Herein, we present a survey of reagents for selective stepwise oxidation of lignin model compounds. Specifically, we have targeted the oxidative cleavage of Cα-Cβ bonds as a means to depolymerize lignin and obtain useful aromatic compounds. In this work, we prepared several lignin model compounds that possess structures, characteristic reactivity, and linkages closely related to the parent lignin polymer. We observed that selective oxidation of benzylic hydroxyl groups, followed by Baeyer-Villiger oxidation of the resulting ketones, successfully cleaves the Cα-Cβ linkage in these model compounds.

  4. Focal liver lesions segmentation and classification in nonenhanced T2-weighted MRI.

    PubMed

    Gatos, Ilias; Tsantis, Stavros; Karamesini, Maria; Spiliopoulos, Stavros; Karnabatidis, Dimitris; Hazle, John D; Kagadis, George C

    2017-07-01

    To automatically segment and classify focal liver lesions (FLLs) on nonenhanced T2-weighted magnetic resonance imaging (MRI) scans using a computer-aided diagnosis (CAD) algorithm. 71 FLLs (30 benign lesions, 19 hepatocellular carcinomas, and 22 metastases) on T2-weighted MRI scans were delineated by the proposed CAD scheme. The FLL segmentation procedure involved wavelet multiscale analysis to extract accurate edge information and mean intensity values for consecutive edges computed using horizontal and vertical analysis that were fed into the subsequent fuzzy C-means algorithm for final FLL border extraction. Texture information for each extracted lesion was derived using 42 first- and second-order textural features from grayscale value histogram, co-occurrence, and run-length matrices. Twelve morphological features were also extracted to capture any shape differentiation between classes. Feature selection was performed with stepwise multilinear regression analysis that led to a reduced feature subset. A multiclass Probabilistic Neural Network (PNN) classifier was then designed and used for lesion classification. PNN model evaluation was performed using the leave-one-out (LOO) method and receiver operating characteristic (ROC) curve analysis. The mean overlap between the automatically segmented FLLs and the manual segmentations performed by radiologists was 0.91 ± 0.12. The highest classification accuracies in the PNN model for the benign, hepatocellular carcinoma, and metastatic FLLs were 94.1%, 91.4%, and 94.1%, respectively, with sensitivity/specificity values of 90%/97.3%, 89.5%/92.2%, and 90.9%/95.6% respectively. The overall classification accuracy for the proposed system was 90.1%. Our diagnostic system using sophisticated FLL segmentation and classification algorithms is a powerful tool for routine clinical MRI-based liver evaluation and can be a supplement to contrast-enhanced MRI to prevent unnecessary invasive procedures. © 2017 American Association of Physicists in Medicine.

  5. Identification of environmental covariates of West Nile virus vector mosquito population abundance.

    PubMed

    Trawinski, Patricia R; Mackay, D Scott

    2010-06-01

    The rapid spread of West Nile virus (WNv) in North America is a major public health concern. Culex pipiens-restuans is the principle mosquito vector of WNv in the northeastern United States while Aedes vexans is an important bridge vector of the virus in this region. Vector mosquito abundance is directly dependent on physical environmental factors that provide mosquito habitats. The objective of this research is to determine landscape elements that explain the population abundance and distribution of WNv vector mosquitoes using stepwise linear regression. We developed a novel approach for examining a large set of landscape variables based on a land use and land cover classification by selecting variables in stages to minimize multicollinearity. We also investigated the distance at which landscape elements influence abundance of vector populations using buffer distances of 200, 400, and 1000 m. Results show landscape effects have a significant impact on Cx. pipiens-estuans population distribution while the effects of landscape features are less important for prediction of Ae. vexans population distributions. Cx. pipiens-restuans population abundance is positively correlated with human population density, housing unit density, and urban land use and land cover classes and negatively correlated with age of dwellings and amount of forested land.

  6. Modeling Governance KB with CATPCA to Overcome Multicollinearity in the Logistic Regression

    NASA Astrophysics Data System (ADS)

    Khikmah, L.; Wijayanto, H.; Syafitri, U. D.

    2017-04-01

    The problem often encounters in logistic regression modeling are multicollinearity problems. Data that have multicollinearity between explanatory variables with the result in the estimation of parameters to be bias. Besides, the multicollinearity will result in error in the classification. In general, to overcome multicollinearity in regression used stepwise regression. They are also another method to overcome multicollinearity which involves all variable for prediction. That is Principal Component Analysis (PCA). However, classical PCA in only for numeric data. Its data are categorical, one method to solve the problems is Categorical Principal Component Analysis (CATPCA). Data were used in this research were a part of data Demographic and Population Survey Indonesia (IDHS) 2012. This research focuses on the characteristic of women of using the contraceptive methods. Classification results evaluated using Area Under Curve (AUC) values. The higher the AUC value, the better. Based on AUC values, the classification of the contraceptive method using stepwise method (58.66%) is better than the logistic regression model (57.39%) and CATPCA (57.39%). Evaluation of the results of logistic regression using sensitivity, shows the opposite where CATPCA method (99.79%) is better than logistic regression method (92.43%) and stepwise (92.05%). Therefore in this study focuses on major class classification (using a contraceptive method), then the selected model is CATPCA because it can raise the level of the major class model accuracy.

  7. Stepwise evolution of resistance to toxic cardenolides via genetic substitutions in the Na+/K+ -ATPase of milkweed butterflies (lepidoptera: Danaini).

    PubMed

    Petschenka, Georg; Fandrich, Steffi; Sander, Nils; Wagschal, Vera; Boppré, Michael; Dobler, Susanne

    2013-09-01

    Despite the monarch butterfly (Danaus plexippus) being famous for its adaptations to the defensive traits of its milkweed host plants, little is known about the macroevolution of these traits. Unlike most other animal species, monarchs are largely insensitive to cardenolides, because their target site, the sodium pump (Na(+)/K(+) -ATPase), has evolved amino acid substitutions that reduce cardenolide binding (so-called target site insensitivity, TSI). Because many, but not all, species of milkweed butterflies (Danaini) are associated with cardenolide-containing host plants, we analyzed 16 species, representing all phylogenetic lineages of milkweed butterflies, for the occurrence of TSI by sequence analyses of the Na(+)/K(+) -ATPase gene and by enzymatic assays with extracted Na(+)/K(+) -ATPase. Here we report that sensitivity to cardenolides was reduced in a stepwise manner during the macroevolution of milkweed butterflies. Strikingly, not all Danaini typically consuming cardenolides showed TSI, but rather TSI was more strongly associated with sequestration of toxic cardenolides. Thus, the interplay between bottom-up selection by plant compounds and top-down selection by natural enemies can explain the evolutionary sequence of adaptations to these toxins. © 2013 The Author(s). Evolution © 2013 The Society for the Study of Evolution.

  8. Relation between trinucleotide GAA repeat length and sensory neuropathy in Friedreich's ataxia.

    PubMed

    Santoro, L; De Michele, G; Perretti, A; Crisci, C; Cocozza, S; Cavalcanti, F; Ragno, M; Monticelli, A; Filla, A; Caruso, G

    1999-01-01

    To verify if GAA expansion size in Friedreich's ataxia could account for the severity of sensory neuropathy. Retrospective study of 56 patients with Friedreich's ataxia selected according to homozygosity for GAA expansion and availability of electrophysiological findings. Orthodromic sensory conduction velocity in the median nerve was available in all patients and that of the tibial nerve in 46 of them. Data of sural nerve biopsy and of a morphometric analysis were available in 12 of the selected patients. The sensory action potential amplitude at the wrist (wSAP) and at the medial malleolus (m mal SAP) and the percentage of myelinated fibres with diameter larger than 7, 9, and 11 microm in the sural nerve were correlated with disease duration and GAA expansion size on the shorter (GAA1) and larger (GAA2) expanded allele in each pair. Pearson's correlation test and stepwise multiple regression were used for statistical analysis. A significant inverse correlation between GAA1 size and wSAP, m mal SAP, and percentage of myelinated fibres was found. Stepwise multiple regression showed that GAA1 size significantly affects electrophysiological and morphometric data, whereas duration of disease has no effect. The data suggest that the severity of the sensory neuropathy is probably genetically determined and that it is not progressive.

  9. Quantifying Parkinson's disease finger-tapping severity by extracting and synthesizing finger motion properties.

    PubMed

    Sano, Yuko; Kandori, Akihiko; Shima, Keisuke; Yamaguchi, Yuki; Tsuji, Toshio; Noda, Masafumi; Higashikawa, Fumiko; Yokoe, Masaru; Sakoda, Saburo

    2016-06-01

    We propose a novel index of Parkinson's disease (PD) finger-tapping severity, called "PDFTsi," for quantifying the severity of symptoms related to the finger tapping of PD patients with high accuracy. To validate the efficacy of PDFTsi, the finger-tapping movements of normal controls and PD patients were measured by using magnetic sensors, and 21 characteristics were extracted from the finger-tapping waveforms. To distinguish motor deterioration due to PD from that due to aging, the aging effect on finger tapping was removed from these characteristics. Principal component analysis (PCA) was applied to the age-normalized characteristics, and principal components that represented the motion properties of finger tapping were calculated. Multiple linear regression (MLR) with stepwise variable selection was applied to the principal components, and PDFTsi was calculated. The calculated PDFTsi indicates that PDFTsi has a high estimation ability, namely a mean square error of 0.45. The estimation ability of PDFTsi is higher than that of the alternative method, MLR with stepwise regression selection without PCA, namely a mean square error of 1.30. This result suggests that PDFTsi can quantify PD finger-tapping severity accurately. Furthermore, the result of interpreting a model for calculating PDFTsi indicated that motion wideness and rhythm disorder are important for estimating PD finger-tapping severity.

  10. Gold-catalyzed sequential annulations towards 3,4-fused bi/tri-cyclic furans involving a [3+2+2]-cycloaddition.

    PubMed

    Liu, Suna; Yang, Pu; Peng, Shiyong; Zhu, Chenghao; Cao, Shengyu; Li, Jian; Sun, Jiangtao

    2017-01-17

    A gold-catalyzed sequential annulation reaction to prepare 3,4-fused bicyclic furan compounds has been realized by employing 2-(1-alkynyl)-2-alken-1-ones and 1,3,5-triazines as the starting materials under mild reaction conditions. This protocol features multiple bond formation in a single operation with the incorporation of two nitrogen and two carbon atoms into the final products. A mechanistic investigation reveals that the sequential annulations involved an unprecedented stepwise [3+2+2]-cycloaddition.

  11. Variable Selection for Regression Models of Percentile Flows

    NASA Astrophysics Data System (ADS)

    Fouad, G.

    2017-12-01

    Percentile flows describe the flow magnitude equaled or exceeded for a given percent of time, and are widely used in water resource management. However, these statistics are normally unavailable since most basins are ungauged. Percentile flows of ungauged basins are often predicted using regression models based on readily observable basin characteristics, such as mean elevation. The number of these independent variables is too large to evaluate all possible models. A subset of models is typically evaluated using automatic procedures, like stepwise regression. This ignores a large variety of methods from the field of feature (variable) selection and physical understanding of percentile flows. A study of 918 basins in the United States was conducted to compare an automatic regression procedure to the following variable selection methods: (1) principal component analysis, (2) correlation analysis, (3) random forests, (4) genetic programming, (5) Bayesian networks, and (6) physical understanding. The automatic regression procedure only performed better than principal component analysis. Poor performance of the regression procedure was due to a commonly used filter for multicollinearity, which rejected the strongest models because they had cross-correlated independent variables. Multicollinearity did not decrease model performance in validation because of a representative set of calibration basins. Variable selection methods based strictly on predictive power (numbers 2-5 from above) performed similarly, likely indicating a limit to the predictive power of the variables. Similar performance was also reached using variables selected based on physical understanding, a finding that substantiates recent calls to emphasize physical understanding in modeling for predictions in ungauged basins. The strongest variables highlighted the importance of geology and land cover, whereas widely used topographic variables were the weakest predictors. Variables suffered from a high degree of multicollinearity, possibly illustrating the co-evolution of climatic and physiographic conditions. Given the ineffectiveness of many variables used here, future work should develop new variables that target specific processes associated with percentile flows.

  12. Step-wise pulling protocols for non-equilibrium dynamics

    NASA Astrophysics Data System (ADS)

    Ngo, Van Anh

    The fundamental laws of thermodynamics and statistical mechanics, and the deeper understandings of quantum mechanics have been rebuilt in recent years. It is partly because of the increasing power of computing resources nowadays, that allow shedding direct insights into the connections among the thermodynamics laws, statistical nature of our world, and the concepts of quantum mechanics, which have not yet been understood. But mostly, the most important reason, also the ultimate goal, is to understand the mechanisms, statistics and dynamics of biological systems, whose prevailing non-equilibrium processes violate the fundamental laws of thermodynamics, deviate from statistical mechanics, and finally complicate quantum effects. I believe that investigations of the fundamental laws of non-equilibrium dynamics will be a frontier research for at least several more decades. One of the fundamental laws was first discovered in 1997 by Jarzynski, so-called Jarzynski's Equality. Since then, different proofs, alternative descriptions of Jarzynski's Equality, and its further developments and applications have been quickly accumulated. My understandings, developments and applications of an alternative theory on Jarzynski's Equality form the bulk of this dissertation. The core of my theory is based on stepwise pulling protocols, which provide deeper insight into how fluctuations of reaction coordinates contribute to free-energy changes along a reaction pathway. We find that the most optimal pathways, having the largest contribution to free-energy changes, follow the principle of detailed balance. This is a glimpse of why the principle of detailed balance appears so powerful for sampling the most probable statistics of events. In a further development on Jarzynski's Equality, I have been trying to use it in the formalism of diagonal entropy to propose a way to extract useful thermodynamic quantities such temperature, work and free-energy profiles from far-from-equilibrium ensembles, which can be used to characterize non-equilibrium dynamics. Furthermore, we have applied the stepwise pulling protocols and Jarzynski's Equality to investigate the ion selectivity of potassium channels via molecular dynamics simulations. The mechanism of the potassium ion selectivity has remained poorly understood for over fifty years, although a Nobel Prize was awarded to the discovery of the molecular structure of a potassium-selective channel in 2003. In one year of performing simulations, we were able to reproduce the major results of ion selectivity accumulated in fifty years. We have been even boldly going further to propose a new model for ion selectivity based on the structural rearrangement of the selectivity filter of potassium-selective KcsA channels. This structural rearrangement has never been shown to play such a pivotal role in selecting and conducting potassium ions, but effectively rejecting sodium ions. Using the stepwise pulling protocols, we are also able to estimate conductance for ion channels, which remains elusive by using other methods. In the light of ion channels, we have also investigated how a synthetic channel of telemeric G-quadruplex conducts different types of ions. These two studies on ion selectivity not only constitute an interesting part of this dissertation, but also will enable us to further explore a new set of ion-selectivity principles. Beside the focus of my dissertation, I used million-atom molecular dynamics simulations to investigate the mechanical properties of body-centered-cubic (BCCS) and face-centered-cubic (FCCS) supercrystals of DNA-functionalized gold nanoparticles. These properties are valuable for examining whether these supercrystals can be used in gene delivery and gene therapy. The formation of such ordered supercrystals is useful to protect DNAs or RNAs from being attacked and destroyed by enzymes in cells. I also performed all-atom molecular dynamics simulations to study a pure oleic acid (OA) membrane in water that results into a triple-layer structure. The simulations show that the trans-membrane movement of water and OAs is cooperative and correlated, and agrees with experimentally measured absorption rates. The simulation results support the idea that OA flip-flop is more favorable than transport by means of functional proteins. This study might provide further insight into how primitive cell membranes work, and how the interplay and correlation between water and fatty acids may occur.

  13. Different clinical prognostic factors are associated with improved glycaemic control: findings from MARCH randomized trial.

    PubMed

    Han, J; Yu, H; Tu, Y; Pang, J; Liu, F; Bao, Y; Yang, W; Jia, W

    2017-04-01

    Metformin and acarbose have comparable efficacy as initial therapy for HbA 1c reduction in Chinese patients with newly diagnosed Type 2 diabetes. However, not all participants achieved glycaemic control. Our aim was to discover a monotherapy predictor for therapeutic response in Type 2 diabetes on the basis of baseline features. Data from the MARCH trial were collected, resulting in 698 individuals being available for longitudinal analyses. All participants were divided into subgroups based on successful and unsuccessful achievement of the glycaemic target according to primary endpoints at week 24 (HbA 1c < 53 mmol/mol; 7.0%). Logistic regression analysis with stepwise variable selection was performed to assess the independent risk factors for good glycaemic control of monotherapy with metformin or acarbose. Median HbA 1c was 66 ± 1 mmol/mol (8.2 ± 0.07%) in the metformin group at baseline, and 66 ± 1 mmol/mol (8.2 ± 0.07%) in the acarbose group. After 24 weeks of monotherapy, 79.8% of participants in the metformin group achieved glycaemic targets compared with 78.7% of those in the acarbose group. Multivariate regression analysis showed that BMI and fasting blood glucose were significant independent predictors for the maintenance of good glycaemic control in the metformin group, whereas phase I insulin secretion (Insulin/Glucose at 30 min, I30/G30) and duration of diabetes were associated with good glycaemic control in the acarbose group. For newly diagnosed Type 2 diabetes, some clinical features and laboratory parameters are important prognostic factors for predicting drug responsiveness. Participants with a higher BMI and lower fasting blood glucose achieved good glycaemic control when metformin was selected as the initial treatment. Acarbose was best for participants with higher phase I insulin secretion (I30/G30) and shorter duration of Type 2 diabetes. © 2016 Diabetes UK.

  14. Spectrometric Estimation of Total Nitrogen Concentration in Douglas-Fir Foliage

    NASA Technical Reports Server (NTRS)

    Johnson, Lee F.; Billow, Christine R.; Peterson, David L. (Technical Monitor)

    1995-01-01

    Spectral measurements of fresh and dehydrated Douglas-fir foliage, from trees cultivated under three fertilization treatments, were acquired with a laboratory spectrophotometer. The slope (first-derivative) of the fresh- and dry-leaf absorbance spectra at locations near known protein absorption features was strongly correlated with total nitrogen (TN) concentration of the foliage samples. Particularly strong correlation was observed between the first-derivative spectra in the 2150-2170 nm region and TN, reaching a local maximum in the fresh-leaf spectra of -0.84 at 2 160 nm. Stepwise regression was used to generate calibration equations relating first derivative spectra from fresh, dry/intact, and dry/ground samples to TN concentration. Standard errors of calibration were 1.52 mg g-1 (fresh), 1.33 (dry/intact), and 1.20 (dry/ground), with goodness-of-fit 0.94 and greater. Cross-validation was performed with the fresh-leaf dataset to examine the predictive capability of the regression method; standard errors of prediction ranged from 1.47 - 2.37 mg g(exp -1) across seven different validation sets, prediction goodness of fit ranged from .85-.94, and wavelength selection was fairly insensitive to the membership of the calibration set. All regressions in this study tended to select wavelengths in the 2100-2350 nm region, with the primary selection in the 2142-2172 nm region. The study provides positive evidence concerning the feasibility of assessing TN status of fresh-leaf samples by spectrometric means. We assert that the ability to extract biochemical information from fresh-leaf spectra is a necessary but insufficient condition regarding the use of remote sensing for canopy-level biochemical estimation.

  15. Canada's Deep Geological Repository For Used Nuclear Fuel -The Geoscientific Site Evaluation Process

    NASA Astrophysics Data System (ADS)

    Hirschorn, S.; Ben Belfadhel, M.; Blyth, A.; DesRoches, A. J.; McKelvie, J. R. M.; Parmenter, A.; Sanchez-Rico Castejon, M.; Urrutia-Bustos, A.; Vorauer, A.

    2014-12-01

    The Nuclear Waste Management Organization (NWMO) is responsible for implementing Adaptive Phased Management, the approach selected by the Government of Canada for long-term management of used nuclear fuel generated by Canadian nuclear reactors. In May 2010, the NWMO published and initiated a nine-step site selection process to find an informed and willing community to host a deep geological repository for Canada's used nuclear fuel. The site selection process is designed to address a broad range of technical and social, economic and cultural factors. The suitability of candidate areas will be assessed in a stepwise manner over a period of many years and include three main steps: Initial Screenings; Preliminary Assessments; and Detailed Site Characterizations. The Preliminary Assessment is conducted in two phases. NWMO has completed Phase 1 preliminary assessments for the first eight communities that entered into this step. While the Phase 1 desktop geoscientific assessments showed that each of the eight communities contains general areas that have the potential to satisfy the geoscientific safety requirements for hosting a deep geological repository, the assessment identified varying degrees of geoscientific complexity and uncertainty between communities, reflecting their different geological settings and structural histories. Phase 2 activities will include a sequence of high-resolution airborne geophysical surveys and focused geological field mapping to ground-truth lithology and structural features, followed by limited deep borehole drilling and testing. These activities will further evaluate the site's ability to meet the safety functions that a site would need to ultimately satisfy in order to be considered suitable. This paper provides an update on the site evaluation process and describes the approach, methods and criteria that are being used to conduct the geoscientific Preliminary Assessments.

  16. Prognostic stratification model for patients with stage I non-small cell lung cancer adenocarcinoma treated with surgical resection without adjuvant therapies using metabolic features measured on F-18 FDG PET and postoperative pathologic factors.

    PubMed

    Kang, Yeon-Koo; Song, Yoo Sung; Cho, Sukki; Jheon, Sanghoon; Lee, Won Woo; Kim, Kwhanmien; Kim, Sang Eun

    2018-05-01

    In the management of non-small cell lung cancer (NSCLC), the prognostic stratification of stage I tumors without indication of adjuvant therapy, remains to be elucidated in order to better select patients who can benefit from additional therapies. We aimed to stratify the prognosis of patients with stage I NSCLC adenocarcinoma using clinicopathologic factors and F-18 FDG PET. We retrospectively enrolled 128 patients with stage I NSCLC without any high-risk factors, who underwent curative surgical resection without adjuvant therapies. Preoperative clinical and postoperative pathologic factors were evaluated by medical record review. Standardized uptake value corrected with lean body mass (SUL max ) was measured on F-18 FDG PET. Among the factors, independent predictors for recurrence-free survival (RFS) were selected using univariate and stepwise multivariate survival analyses. A prognostic stratification model for RFS was designed using the selected factors. Tumors recurred in nineteen patients (14.8%). Among the investigated clinicopathologic and FDG PET factors, SUL max on PET and spread through air spaces (STAS) on pathologic review were determined to be independent prognostic factors for RFS. A prognostic model was designed using these two factors in the following manner: (1) Low-risk: SUL max  ≤ 1.9 and no STAS, (2) intermediate-risk: neither low-risk nor high-risk, (3) high-risk: SUL max> 1.9 and observed STAS. This model exhibited significant predictive power for RFS. We showed that FDG uptake and STAS are significant prognostic markers in stage I NSCLC adenocarcinoma treated with surgical resection without adjuvant therapies. Copyright © 2018 Elsevier B.V. All rights reserved.

  17. The World Health Organization STEPwise Approach to Noncommunicable Disease Risk-Factor Surveillance: Methods, Challenges, and Opportunities

    PubMed Central

    Guthold, Regina; Cowan, Melanie; Savin, Stefan; Bhatti, Lubna; Armstrong, Timothy; Bonita, Ruth

    2016-01-01

    Objectives. We sought to outline the framework and methods used by the World Health Organization (WHO) STEPwise approach to noncommunicable disease (NCD) surveillance (STEPS), describe the development and current status, and discuss strengths, limitations, and future directions of STEPS surveillance. Methods. STEPS is a WHO-developed, standardized but flexible framework for countries to monitor the main NCD risk factors through questionnaire assessment and physical and biochemical measurements. It is coordinated by national authorities of the implementing country. The STEPS surveys are generally household-based and interviewer-administered, with scientifically selected samples of around 5000 participants. Results. To date, 122 countries across all 6 WHO regions have completed data collection for STEPS or STEPS-aligned surveys. Conclusions. STEPS data are being used to inform NCD policies and track risk-factor trends. Future priorities include strengthening these linkages from data to action on NCDs at the country level, and continuing to develop STEPS’ capacities to enable a regular and continuous cycle of risk-factor surveillance worldwide. PMID:26696288

  18. The World Health Organization STEPwise Approach to Noncommunicable Disease Risk-Factor Surveillance: Methods, Challenges, and Opportunities.

    PubMed

    Riley, Leanne; Guthold, Regina; Cowan, Melanie; Savin, Stefan; Bhatti, Lubna; Armstrong, Timothy; Bonita, Ruth

    2016-01-01

    We sought to outline the framework and methods used by the World Health Organization (WHO) STEPwise approach to noncommunicable disease (NCD) surveillance (STEPS), describe the development and current status, and discuss strengths, limitations, and future directions of STEPS surveillance. STEPS is a WHO-developed, standardized but flexible framework for countries to monitor the main NCD risk factors through questionnaire assessment and physical and biochemical measurements. It is coordinated by national authorities of the implementing country. The STEPS surveys are generally household-based and interviewer-administered, with scientifically selected samples of around 5000 participants. To date, 122 countries across all 6 WHO regions have completed data collection for STEPS or STEPS-aligned surveys. STEPS data are being used to inform NCD policies and track risk-factor trends. Future priorities include strengthening these linkages from data to action on NCDs at the country level, and continuing to develop STEPS' capacities to enable a regular and continuous cycle of risk-factor surveillance worldwide.

  19. Association Between Borderline Personality Features and Temporal Summation of Second Pain: A Cross-Sectional Study.

    PubMed

    You, Dokyoung S; Meagher, Mary W

    2017-01-01

    Individuals with greater borderline personality features may be vulnerable to chronic pain. Because pain is an unpleasant sensory and emotional experience, affect dysregulation as the core personality feature may be linked to pain hypersensitivity. Studies have found that greater borderline features are associated with increased intensity in clinical and experimental pain, and that depression mediates this increase. The current study further examined the association between borderline features and heat pain sensitivity, the contribution of affect dysregulation and the other borderline personality factors (identity problems, negative relationships, self-harming/impulsivity) to the association, and depression as a mediator. Additionally, we examined whether blunted sympathetic responses mediate the association between borderline features and temporal summation of second pain (TSSP). Thermal pain threshold, thermal TSSP and aftersensations pain were assessed in 79 healthy individuals with varying degrees of borderline features. TSSP is a proxy measure for central sensitization and refers to the gradual increase in pain to repeated nociceptive stimuli. A regression analysis showed that greater borderline features predicted greater TSSP (β = .22, p = .050, R 2 = .05). Borderline features were unrelated to pain threshold and TSSP decay. A stepwise regression showed greater TSSP in individuals with greater borderline features was accounted for by the negative relationships factor rather than the affect dysregulation factor. The results of mediational analyses showed depression and blunted sympathetic skin conductance responses mediated the positive association between TSSP and borderline features.

  20. Automated detection of tuberculosis on sputum smeared slides using stepwise classification

    NASA Astrophysics Data System (ADS)

    Divekar, Ajay; Pangilinan, Corina; Coetzee, Gerrit; Sondh, Tarlochan; Lure, Fleming Y. M.; Kennedy, Sean

    2012-03-01

    Routine visual slide screening for identification of tuberculosis (TB) bacilli in stained sputum slides under microscope system is a tedious labor-intensive task and can miss up to 50% of TB. Based on the Shannon cofactor expansion on Boolean function for classification, a stepwise classification (SWC) algorithm is developed to remove different types of false positives, one type at a time, and to increase the detection of TB bacilli at different concentrations. Both bacilli and non-bacilli objects are first analyzed and classified into several different categories including scanty positive, high concentration positive, and several non-bacilli categories: small bright objects, beaded, dim elongated objects, etc. The morphological and contrast features are extracted based on aprior clinical knowledge. The SWC is composed of several individual classifiers. Individual classifier to increase the bacilli counts utilizes an adaptive algorithm based on a microbiologist's statistical heuristic decision process. Individual classifier to reduce false positive is developed through minimization from a binary decision tree to classify different types of true and false positive based on feature vectors. Finally, the detection algorithm is was tested on 102 independent confirmed negative and 74 positive cases. A multi-class task analysis shows high accordance rate for negative, scanty, and high-concentration as 88.24%, 56.00%, and 97.96%, respectively. A binary-class task analysis using a receiver operating characteristics method with the area under the curve (Az) is also utilized to analyze the performance of this detection algorithm, showing the superior detection performance on the high-concentration cases (Az=0.913) and cases mixed with high-concentration and scanty cases (Az=0.878).

  1. Development and Application of a Three-Dimensional Finite Element Vapor Intrusion Model

    PubMed Central

    Pennell, Kelly G.; Bozkurt, Ozgur; Suuberg, Eric M.

    2010-01-01

    Details of a three-dimensional finite element model of soil vapor intrusion, including the overall modeling process and the stepwise approach, are provided. The model is a quantitative modeling tool that can help guide vapor intrusion characterization efforts. It solves the soil gas continuity equation coupled with the chemical transport equation, allowing for both advective and diffusive transport. Three-dimensional pressure, velocity, and chemical concentration fields are produced from the model. Results from simulations involving common site features, such as impervious surfaces, porous foundation sub-base material, and adjacent structures are summarized herein. The results suggest that site-specific features are important to consider when characterizing vapor intrusion risks. More importantly, the results suggest that soil gas or subslab gas samples taken without proper regard for particular site features may not be suitable for evaluating vapor intrusion risks; rather, careful attention needs to be given to the many factors that affect chemical transport into and around buildings. PMID:19418819

  2. Investigating a physical basis for spectroscopic estimates of leaf nitrogen concentration

    USGS Publications Warehouse

    Kokaly, R.F.

    2001-01-01

    The reflectance spectra of dried and ground plant foliage are examined for changes directly due to increasing nitrogen concentration. A broadening of the 2.1-??m absorption feature is observed as nitrogen concentration increases. The broadening is shown to arise from two absorptions at 2.054 ??m and 2.172 ??m. The wavelength positions of these absorptions coincide with the absorption characteristics of the nitrogen-containing amide bonds in proteins. The observed presence of these absorption features in the reflectance spectra of dried foliage is suggested to form a physical basis for high correlations established by stepwise multiple linear regression techniques between the reflectance of dry plant samples and their nitrogen concentration. The consistent change in the 2.1-??m absorption feature as nitrogen increases and the offset position of protein absorptions compared to those of other plant components together indicate that a generally applicable algorithm may be developed for spectroscopic estimates of nitrogen concentration from the reflectance spectra of dried plant foliage samples. ?? 2001 Published by Elsevier Science Ireland Ltd.

  3. Quantum-enhanced feature selection with forward selection and backward elimination

    NASA Astrophysics Data System (ADS)

    He, Zhimin; Li, Lvzhou; Huang, Zhiming; Situ, Haozhen

    2018-07-01

    Feature selection is a well-known preprocessing technique in machine learning, which can remove irrelevant features to improve the generalization capability of a classifier and reduce training and inference time. However, feature selection is time-consuming, particularly for the applications those have thousands of features, such as image retrieval, text mining and microarray data analysis. It is crucial to accelerate the feature selection process. We propose a quantum version of wrapper-based feature selection, which converts a classical feature selection to its quantum counterpart. It is valuable for machine learning on quantum computer. In this paper, we focus on two popular kinds of feature selection methods, i.e., wrapper-based forward selection and backward elimination. The proposed feature selection algorithm can quadratically accelerate the classical one.

  4. Designing of a fluoride selective receptor through molecular orbital engineering

    NASA Astrophysics Data System (ADS)

    Mishra, Rakesh K.; Kumar, Virendra; Diwan, Uzra; Upadhyay, K. K.; Roy Chowdhury, P. K.

    2012-11-01

    The stepwise substitution of appropriate groups over the 3-[(2,4-dinitro-phenyl)-hydrazono]-butyric acid ethyl ester (R3) lead receptor R1 which showed selectivity towards fluoride in DMSO. The UV-vis and 1H NMR titration studies revealed the details of the binding between receptor R1 and fluoride. The receptor R1 also recognized fluoride in a toothpaste solution to as low as 50 ppm. The theoretical simulations of recognition event at Density Functional Theory (DFT) level using B3LYP/6-31G** basis set and polarizable continuum model (PCM) approach lead a semi-quantitative match with the experimental results.

  5. Natural Resources Inventory and Land Evaluation in Switzerland

    NASA Technical Reports Server (NTRS)

    Haefner, H. (Principal Investigator)

    1975-01-01

    The author has identified the following significant results. A system was developed to operationally map and measure the areal extent of various land use categories for updating existing and producing new and actual thematic maps showing the latest state of rural and urban landscapes and its changes. The processing system includes: (1) preprocessing steps for radiometric and geometric corrections; (2) classification of the data by a multivariate procedure, using a stepwise linear discriminant analysis based on carefully selected training cells; and (3) output in form of color maps by printing black and white theme overlays of a selected scale with photomation system and its coloring and combination into a color composite.

  6. Study of the thermal properties of selected PCMs for latent heat storage in buildings

    NASA Astrophysics Data System (ADS)

    Valentova, Katerina; Pechackova, Katerina; Prikryl, Radek; Ostry, Milan; Zmeskal, Oldrich

    2017-07-01

    The paper is focused on measurements of thermal properties of selected phase change materials (PCMs) which can be used for latent heat storage in building structures. The thermal properties were measured by the transient step-wise method and analyzed by the thermal spectroscopy. The results of three different materials (RT18HC, RT28HC, and RT35HC) and their thermal properties in solid, liquid, and phase change region were determined. They were correlated with the differential scanning calorimetry (DSC) measurement. The results will be used to determine the optimum ratio of components for the construction of drywall and plasters containing listed ingredients, respectively.

  7. Factors related to HIV-associated neurocognitive impairment differ with age.

    PubMed

    Fogel, Gary B; Lamers, Susanna L; Levine, Andrew J; Valdes-Sueiras, Miguel; McGrath, Michael S; Shapshak, Paul; Singer, Elyse J

    2015-02-01

    Over 50% of HIV-infected (HIV+) persons are expected to be over age 50 by 2015. The pathogenic effects of HIV, particularly in cases of long-term infection, may intersect with those of age-related illnesses and prolonged exposure to combined antiretroviral therapy (cART). One potential outcome is an increased prevalence of neurocognitive impairment in older HIV+ individuals, as well as an altered presentation of HIV-associated neurocognitive disorders (HANDs). In this study, we employed stepwise regression to examine 24 features sometimes associated with HAND in 40 older (55-73 years of age) and 30 younger (32-50 years of age) HIV+, cART-treated participants without significant central nervous system confounds. The features most effective in generating a true assessment of the likelihood of HAND diagnosis differed between older and younger cohorts, with the younger cohort containing features associated with drug abuse that were correlated to HAND and the older cohort containing features that were associated with lipid disorders mildly associated with HAND. As the HIV-infected population grows and the demographics of the epidemic change, it is increasingly important to re-evaluate features associated with neurocognitive impairment. Here, we have identified features, routinely collected in primary care settings, that provide more accurate diagnostic value than a neurocognitive screening measure among younger and older HIV individuals.

  8. Comparative interactomics provides evidence for functional specialization of the nuclear pore complex.

    PubMed

    Obado, Samson O; Field, Mark C; Rout, Michael P

    2017-07-04

    The core architecture of the eukaryotic cell was established well over one billion years ago, and is largely retained in all extant lineages. However, eukaryotic cells also possess lineage-specific features, frequently keyed to specific functional requirements. One quintessential core eukaryotic structure is the nuclear pore complex (NPC), responsible for regulating exchange of macromolecules between the nucleus and cytoplasm as well as acting as a nuclear organizational hub. NPC architecture has been best documented in one eukaryotic supergroup, the Opisthokonts (e.g. Saccharomyces cerevisiae and Homo sapiens), which although compositionally similar, have significant variations in certain NPC subcomplex structures. The variation of NPC structure across other taxa in the eukaryotic kingdom however, remains poorly understood. We explored trypanosomes, highly divergent organisms, and mapped and assigned their NPC proteins to specific substructures to reveal their NPC architecture. We showed that the NPC central structural scaffold is conserved, likely across all eukaryotes, but more peripheral elements can exhibit very significant lineage-specific losses, duplications or other alterations in their components. Amazingly, trypanosomes lack the major components of the mRNA export platform that are asymmetrically localized within yeast and vertebrate NPCs. Concomitant with this, the trypanosome NPC is ALMOST completely symmetric with the nuclear basket being the only major source of asymmetry. We suggest these features point toward a stepwise evolution of the NPC in which a coating scaffold first stabilized the pore after which selective gating emerged and expanded, leading to the addition of peripheral remodeling machineries on the nucleoplasmic and cytoplasmic sides of the pore.

  9. Probing cluster surface morphology by cryo spectroscopy of N2 on cationic nickel clusters

    NASA Astrophysics Data System (ADS)

    Dillinger, Sebastian; Mohrbach, Jennifer; Niedner-Schatteburg, Gereon

    2017-11-01

    We present the cryogenic (26 K) IR spectra of selected [Nin(N2)m]+ (n = 5-20, m = 1 - mmax), which strongly reveal n- and m-dependent features in the N2 stretching region, in conjunction with density functional theory modeling of some of these findings. The observed spectral features allow us to refine the kinetic classification [cf. J. Mohrbach, S. Dillinger, and G. Niedner-Schatteburg, J. Chem. Phys. 147, 184304 (2017)] and to define four classes of structure related surface adsorption behavior: Class (1) of Ni6+, Ni13+, and Ni19+ are highly symmetrical clusters with all smooth surfaces of equally coordinated Ni atoms that entertain stepwise N2 adsorption up to stoichiometric N2:Nisurface saturation. Class (2) of Ni12+ and Ni18+ are highly symmetrical clusters minus one. Their relaxed smooth surfaces reorganize by enhanced N2 uptake toward some low coordinated Ni surface atoms with double N2 occupation. Class (3) of Ni5+ and Ni7+ through Ni11+ are small clusters of rough surfaces with low coordinated Ni surface atoms, and some reveal semi-internal Ni atoms of high next-neighbor coordination. Surface reorganization upon N2 uptake turns rough into rough surface by Ni atom migration and turns octahedral based structures into pentagonal bipyramidal structures. Class (4) of Ni14+ through Ni17+ and Ni20+ are large clusters with rough and smooth surface areas. They possess smooth icosahedral surfaces with some proximate capping atom(s) on one hemisphere of the icosahedron with the other one largely unaffected.

  10. Feature Selection for Chemical Sensor Arrays Using Mutual Information

    PubMed Central

    Wang, X. Rosalind; Lizier, Joseph T.; Nowotny, Thomas; Berna, Amalia Z.; Prokopenko, Mikhail; Trowell, Stephen C.

    2014-01-01

    We address the problem of feature selection for classifying a diverse set of chemicals using an array of metal oxide sensors. Our aim is to evaluate a filter approach to feature selection with reference to previous work, which used a wrapper approach on the same data set, and established best features and upper bounds on classification performance. We selected feature sets that exhibit the maximal mutual information with the identity of the chemicals. The selected features closely match those found to perform well in the previous study using a wrapper approach to conduct an exhaustive search of all permitted feature combinations. By comparing the classification performance of support vector machines (using features selected by mutual information) with the performance observed in the previous study, we found that while our approach does not always give the maximum possible classification performance, it always selects features that achieve classification performance approaching the optimum obtained by exhaustive search. We performed further classification using the selected feature set with some common classifiers and found that, for the selected features, Bayesian Networks gave the best performance. Finally, we compared the observed classification performances with the performance of classifiers using randomly selected features. We found that the selected features consistently outperformed randomly selected features for all tested classifiers. The mutual information filter approach is therefore a computationally efficient method for selecting near optimal features for chemical sensor arrays. PMID:24595058

  11. Rough sets and Laplacian score based cost-sensitive feature selection

    PubMed Central

    Yu, Shenglong

    2018-01-01

    Cost-sensitive feature selection learning is an important preprocessing step in machine learning and data mining. Recently, most existing cost-sensitive feature selection algorithms are heuristic algorithms, which evaluate the importance of each feature individually and select features one by one. Obviously, these algorithms do not consider the relationship among features. In this paper, we propose a new algorithm for minimal cost feature selection called the rough sets and Laplacian score based cost-sensitive feature selection. The importance of each feature is evaluated by both rough sets and Laplacian score. Compared with heuristic algorithms, the proposed algorithm takes into consideration the relationship among features with locality preservation of Laplacian score. We select a feature subset with maximal feature importance and minimal cost when cost is undertaken in parallel, where the cost is given by three different distributions to simulate different applications. Different from existing cost-sensitive feature selection algorithms, our algorithm simultaneously selects out a predetermined number of “good” features. Extensive experimental results show that the approach is efficient and able to effectively obtain the minimum cost subset. In addition, the results of our method are more promising than the results of other cost-sensitive feature selection algorithms. PMID:29912884

  12. Rough sets and Laplacian score based cost-sensitive feature selection.

    PubMed

    Yu, Shenglong; Zhao, Hong

    2018-01-01

    Cost-sensitive feature selection learning is an important preprocessing step in machine learning and data mining. Recently, most existing cost-sensitive feature selection algorithms are heuristic algorithms, which evaluate the importance of each feature individually and select features one by one. Obviously, these algorithms do not consider the relationship among features. In this paper, we propose a new algorithm for minimal cost feature selection called the rough sets and Laplacian score based cost-sensitive feature selection. The importance of each feature is evaluated by both rough sets and Laplacian score. Compared with heuristic algorithms, the proposed algorithm takes into consideration the relationship among features with locality preservation of Laplacian score. We select a feature subset with maximal feature importance and minimal cost when cost is undertaken in parallel, where the cost is given by three different distributions to simulate different applications. Different from existing cost-sensitive feature selection algorithms, our algorithm simultaneously selects out a predetermined number of "good" features. Extensive experimental results show that the approach is efficient and able to effectively obtain the minimum cost subset. In addition, the results of our method are more promising than the results of other cost-sensitive feature selection algorithms.

  13. Orientation-free and differentially pumped addition of a low-flux reactive gas beam to a surface analysis system.

    PubMed

    Harthcock, Colin; Jahanbekam, Abdolreza; Eskelsen, Jeremy R; Lee, David Y

    2016-11-01

    We describe an example of a piecewise gas chamber that can be customized to incorporate a low flux of gas-phase radicals with an existing surface analysis chamber for in situ and stepwise gas-surface interaction experiments without any constraint in orientation. The piecewise nature of this gas chamber provides complete angular freedom and easy alignment and does not require any modification of the existing surface analysis chamber. In addition, the entire gas-surface system is readily differentially pumped with the surface chamber kept under ultra-high-vacuum during the gas-surface measurements. This new design also allows not only straightforward reconstruction to accommodate the orientation of different surface chambers but also for the addition of other desired features, such as an additional pump to the current configuration. Stepwise interaction between atomic oxygen and a highly ordered pyrolytic graphite surface was chosen to test the effectiveness of this design, and the site-dependent O-atom chemisorption and clustering on the graphite surface were resolved by a scanning tunneling microscope in the nm-scale. X-ray photoelectron spectroscopy was used to further confirm the identity of the chemisorbed species on the graphite surface as oxygen.

  14. Super-resolution fluorescence microscopy by stepwise optical saturation

    PubMed Central

    Zhang, Yide; Nallathamby, Prakash D.; Vigil, Genevieve D.; Khan, Aamir A.; Mason, Devon E.; Boerckel, Joel D.; Roeder, Ryan K.; Howard, Scott S.

    2018-01-01

    Super-resolution fluorescence microscopy is an important tool in biomedical research for its ability to discern features smaller than the diffraction limit. However, due to its difficult implementation and high cost, the super-resolution microscopy is not feasible in many applications. In this paper, we propose and demonstrate a saturation-based super-resolution fluorescence microscopy technique that can be easily implemented and requires neither additional hardware nor complex post-processing. The method is based on the principle of stepwise optical saturation (SOS), where M steps of raw fluorescence images are linearly combined to generate an image with a M-fold increase in resolution compared with conventional diffraction-limited images. For example, linearly combining (scaling and subtracting) two images obtained at regular powers extends the resolution by a factor of 1.4 beyond the diffraction limit. The resolution improvement in SOS microscopy is theoretically infinite but practically is limited by the signal-to-noise ratio. We perform simulations and experimentally demonstrate super-resolution microscopy with both one-photon (confocal) and multiphoton excitation fluorescence. We show that with the multiphoton modality, the SOS microscopy can provide super-resolution imaging deep in scattering samples. PMID:29675306

  15. Step-wise refolding of recombinant proteins.

    PubMed

    Tsumoto, Kouhei; Arakawa, Tsutomu; Chen, Linda

    2010-04-01

    Protein refolding is still on trial-and-error basis. Here we describe step-wise dialysis refolding, in which denaturant concentration is altered in step-wise fashion. This technology controls the folding pathway by adjusting the concentrations of the denaturant and other solvent additives to induce sequential folding or disulfide formation.

  16. Site-selective oxidation, amination and epimerization reactions of complex polyols enabled by transfer hydrogenation

    NASA Astrophysics Data System (ADS)

    Hill, Christopher K.; Hartwig, John F.

    2017-12-01

    Polyoxygenated hydrocarbons that bear one or more hydroxyl groups comprise a large set of natural and synthetic compounds, often with potent biological activity. In synthetic chemistry, alcohols are important precursors to carbonyl groups, which then can be converted into a wide range of oxygen- or nitrogen-based functionality. Therefore, the selective conversion of a single hydroxyl group in natural products into a ketone would enable the selective introduction of unnatural functionality. However, the methods known to convert a simple alcohol, or even an alcohol in a molecule that contains multiple protected functional groups, are not suitable for selective reactions of complex polyol structures. We present a new ruthenium catalyst with a unique efficacy for the selective oxidation of a single hydroxyl group among many in unprotected polyol natural products. This oxidation enables the introduction of nitrogen-based functional groups into such structures that lack nitrogen atoms and enables a selective alcohol epimerization by stepwise or reversible oxidation and reduction.

  17. Diels–Alder Reactions of Allene with Benzene and Butadiene: Concerted, Stepwise, and Ambimodal Transition States

    PubMed Central

    2015-01-01

    Multiconfigurational complete active space methods (CASSCF and CASPT2) have been used to investigate the (4 + 2) cycloadditions of allene with butadiene and with benzene. Both concerted and stepwise radical pathways were examined to determine the mechanism of the Diels–Alder reactions with an allene dienophile. Reaction with butadiene occurs via a single ambimodal transition state that can lead to either the concerted or stepwise trajectories along the potential energy surface, while reaction with benzene involves two separate transition states and favors the concerted mechanism relative to the stepwise mechanism via a diradical intermediate. PMID:25216056

  18. Shaped Ceria Nanocrystals Catalyze Efficient and Selective Para-Hydrogen-Enhanced Polarization.

    PubMed

    Zhao, Evan W; Zheng, Haibin; Zhou, Ronghui; Hagelin-Weaver, Helena E; Bowers, Clifford R

    2015-11-23

    Intense para-hydrogen-enhanced NMR signals are observed in the hydrogenation of propene and propyne over ceria nanocubes, nano-octahedra, and nanorods. The well-defined ceria shapes, synthesized by a hydrothermal method, expose different crystalline facets with various oxygen vacancy densities, which are known to play a role in hydrogenation and oxidation catalysis. While the catalytic activity of the hydrogenation of propene over ceria is strongly facet-dependent, the pairwise selectivity is low (2.4% at 375 °C), which is consistent with stepwise H atom transfer, and it is the same for all three nanocrystal shapes. Selective semi-hydrogenation of propyne over ceria nanocubes yields hyperpolarized propene with a similar pairwise selectivity of (2.7% at 300 °C), indicating product formation predominantly by a non-pairwise addition. Ceria is also shown to be an efficient pairwise replacement catalyst for propene. © 2015 WILEY-VCH Verlag GmbH & Co. KGaA, Weinheim.

  19. Stepwise Analysis of Differential Item Functioning Based on Multiple-Group Partial Credit Model.

    ERIC Educational Resources Information Center

    Muraki, Eiji

    1999-01-01

    Extended an Item Response Theory (IRT) method for detection of differential item functioning to the partial credit model and applied the method to simulated data using a stepwise procedure. Then applied the stepwise DIF analysis based on the multiple-group partial credit model to writing trend data from the National Assessment of Educational…

  20. Online feature selection with streaming features.

    PubMed

    Wu, Xindong; Yu, Kui; Ding, Wei; Wang, Hao; Zhu, Xingquan

    2013-05-01

    We propose a new online feature selection framework for applications with streaming features where the knowledge of the full feature space is unknown in advance. We define streaming features as features that flow in one by one over time whereas the number of training examples remains fixed. This is in contrast with traditional online learning methods that only deal with sequentially added observations, with little attention being paid to streaming features. The critical challenges for Online Streaming Feature Selection (OSFS) include 1) the continuous growth of feature volumes over time, 2) a large feature space, possibly of unknown or infinite size, and 3) the unavailability of the entire feature set before learning starts. In the paper, we present a novel Online Streaming Feature Selection method to select strongly relevant and nonredundant features on the fly. An efficient Fast-OSFS algorithm is proposed to improve feature selection performance. The proposed algorithms are evaluated extensively on high-dimensional datasets and also with a real-world case study on impact crater detection. Experimental results demonstrate that the algorithms achieve better compactness and higher prediction accuracy than existing streaming feature selection algorithms.

  1. Discrimination of Active and Weakly Active Human BACE1 Inhibitors Using Self-Organizing Map and Support Vector Machine.

    PubMed

    Li, Hang; Wang, Maolin; Gong, Ya-Nan; Yan, Aixia

    2016-01-01

    β-secretase (BACE1) is an aspartyl protease, which is considered as a novel vital target in Alzheimer`s disease therapy. We collected a data set of 294 BACE1 inhibitors, and built six classification models to discriminate active and weakly active inhibitors using Kohonen's Self-Organizing Map (SOM) method and Support Vector Machine (SVM) method. Each molecular descriptor was calculated using the program ADRIANA.Code. We adopted two different methods: random method and Self-Organizing Map method, for training/test set split. The descriptors were selected by F-score and stepwise linear regression analysis. The best SVM model Model2C has a good prediction performance on test set with prediction accuracy, sensitivity (SE), specificity (SP) and Matthews correlation coefficient (MCC) of 89.02%, 90%, 88%, 0.78, respectively. Model 1A is the best SOM model, whose accuracy and MCC of the test set were 94.57% and 0.98, respectively. The lone pair electronegativity and polarizability related descriptors importantly contributed to bioactivity of BACE1 inhibitor. The Extended-Connectivity Finger-Prints_4 (ECFP_4) analysis found some vitally key substructural features, which could be helpful for further drug design research. The SOM and SVM models built in this study can be obtained from the authors by email or other contacts.

  2. Encapsulation of CH3NH3PbBr3 Perovskite Quantum Dots in MOF-5 Microcrystals as a Stable Platform for Temperature and Aqueous Heavy Metal Ion Detection.

    PubMed

    Zhang, Diwei; Xu, Yan; Liu, Quanlin; Xia, Zhiguo

    2018-04-16

    The stability issue of organometallic halide perovskites remains a great challenge for future research as to their applicability in different functional material fields. Herein, a novel and facile two-step synthesis procedure is reported for encapsulation of CH 3 NH 3 PbBr 3 perovskite quantum dots (QDs) in MOF-5 microcrystals, where PbBr 2 and CH 3 NH 3 Br precursors are added stepwise to fabricate stable CH 3 NH 3 PbBr 3 @MOF-5 composites. In comparison to CH 3 NH 3 PbBr 3 QDs, CH 3 NH 3 PbBr 3 @MOF-5 composites exhibited highly improved water resistance and thermal stability, as well as better pH adaptability over a wide range. Luminescent investigations demonstrate that CH 3 NH 3 PbBr 3 @MOF-5 composites not only featured excellent sensing properties with respect to temperature changes from 30 to 230 °C but also exhibited significant selective luminescent response to several different metal ions in aqueous solution. These outstanding characteristics indicate that the stable CH 3 NH 3 PbBr 3 @MOF-5 composites are potentially interesting for application in fluorescence sensors or detectors.

  3. Detecting obstructive sleep apnea in children by self-affine visualization of oximetry.

    PubMed

    Garde, Ainara; Dekhordi, Parastoo; Petersen, Christian L; Ansermino, J Mark; Dumont, Guy A

    2017-07-01

    Obstructive sleep apnea (OSA), characterized by cessations of breathing during sleep due to upper airway collapse, can affect the healthy growth and development of children. The gold standard for OSA diagnosis, polysomnography(PSG), is expensive and resource intensive, resulting in long waiting lists to perform a PSG. Previously, we investigated the time-frequency analysis of blood oxygen saturation (SpO 2 ) to screen for OSA. We used overnight pulse oximetry from 146 children, collected using a smartphone-based pulse oximeter (Phone Oximeter), simultaneously with standard PSG. Sleep technicians manually scored PSG and provided the average of apnea/hypoapnea events per hour (AHI). In this study, we proposed an alternative method for analyzing SpO 2 , in which a set of contracting transformations form a self-affine set with a 2D attractor, previously developed for qualitative visualization of the photoplethysmogram and electroencephalogram. We applied this technique to the overnight SpO 2 signal from individual patients and extracted features based on the distribution of points (radius and angle) in the visualization. The cloud of points in children without OSA (NonOSA) was more confined than in children with OSA, which was reflected by more empty pixels (radius and angles). The maximum value, skewness and standard deviation of the distribution of points located at different radius and angles were significantly (Bonferroni corrected) higher in NonOSA compared to OSA children. To detect OSA defined at different levels (AHI≥5, AHI≥10 and AHI≥15), three multivariate logistic regression models were implemented using a stepwise feature selection and internally validated through bootstrapping. The models (AHI≥5, AHI≥10, AHI≥15), consisting of 3, 4 and 1 features respectively, provided a bootstrap-corrected AUC of 73%, 81%, 73%. Thus, applying this visualization to nocturnal SpO 2 could yield both visual and quantitative information that might be useful for screening children for OSA.

  4. Enhancing the Discrimination Ability of a Gas Sensor Array Based on a Novel Feature Selection and Fusion Framework.

    PubMed

    Deng, Changjian; Lv, Kun; Shi, Debo; Yang, Bo; Yu, Song; He, Zhiyi; Yan, Jia

    2018-06-12

    In this paper, a novel feature selection and fusion framework is proposed to enhance the discrimination ability of gas sensor arrays for odor identification. Firstly, we put forward an efficient feature selection method based on the separability and the dissimilarity to determine the feature selection order for each type of feature when increasing the dimension of selected feature subsets. Secondly, the K-nearest neighbor (KNN) classifier is applied to determine the dimensions of the optimal feature subsets for different types of features. Finally, in the process of establishing features fusion, we come up with a classification dominance feature fusion strategy which conducts an effective basic feature. Experimental results on two datasets show that the recognition rates of Database I and Database II achieve 97.5% and 80.11%, respectively, when k = 1 for KNN classifier and the distance metric is correlation distance (COR), which demonstrates the superiority of the proposed feature selection and fusion framework in representing signal features. The novel feature selection method proposed in this paper can effectively select feature subsets that are conducive to the classification, while the feature fusion framework can fuse various features which describe the different characteristics of sensor signals, for enhancing the discrimination ability of gas sensors and, to a certain extent, suppressing drift effect.

  5. 4D-LQTA-QSAR and docking study on potent Gram-negative specific LpxC inhibitors: a comparison to CoMFA modeling.

    PubMed

    Ghasemi, Jahan B; Safavi-Sohi, Reihaneh; Barbosa, Euzébio G

    2012-02-01

    A quasi 4D-QSAR has been carried out on a series of potent Gram-negative LpxC inhibitors. This approach makes use of the molecular dynamics (MD) trajectories and topology information retrieved from the GROMACS package. This new methodology is based on the generation of a conformational ensemble profile, CEP, for each compound instead of only one conformation, followed by the calculation intermolecular interaction energies at each grid point considering probes and all aligned conformations resulting from MD simulations. These interaction energies are independent variables employed in a QSAR analysis. The comparison of the proposed methodology to comparative molecular field analysis (CoMFA) formalism was performed. This methodology explores jointly the main features of CoMFA and 4D-QSAR models. Step-wise multiple linear regression was used for the selection of the most informative variables. After variable selection, multiple linear regression (MLR) and partial least squares (PLS) methods used for building the regression models. Leave-N-out cross-validation (LNO), and Y-randomization were performed in order to confirm the robustness of the model in addition to analysis of the independent test set. Best models provided the following statistics: [Formula in text] (PLS) and [Formula in text] (MLR). Docking study was applied to investigate the major interactions in protein-ligand complex with CDOCKER algorithm. Visualization of the descriptors of the best model helps us to interpret the model from the chemical point of view, supporting the applicability of this new approach in rational drug design.

  6. Accurate discrimination of Alzheimer's disease from other dementia and/or normal subjects using SPECT specific volume analysis

    NASA Astrophysics Data System (ADS)

    Iyatomi, Hitoshi; Hashimoto, Jun; Yoshii, Fumuhito; Kazama, Toshiki; Kawada, Shuichi; Imai, Yutaka

    2014-03-01

    Discrimination between Alzheimer's disease and other dementia is clinically significant, however it is often difficult. In this study, we developed classification models among Alzheimer's disease (AD), other dementia (OD) and/or normal subjects (NC) using patient factors and indices obtained by brain perfusion SPECT. SPECT is commonly used to assess cerebral blood flow (CBF) and allows the evaluation of the severity of hypoperfusion by introducing statistical parametric mapping (SPM). We investigated a total of 150 cases (50 cases each for AD, OD, and NC) from Tokai University Hospital, Japan. In each case, we obtained a total of 127 candidate parameters from: (A) 2 patient factors (age and sex), (B) 12 CBF parameters and 113 SPM parameters including (C) 3 from specific volume analysis (SVA), and (D) 110 from voxel-based analysis stereotactic extraction estimation (vbSEE). We built linear classifiers with a statistical stepwise feature selection and evaluated the performance with the leave-one-out cross validation strategy. Our classifiers achieved very high classification performances with reasonable number of selected parameters. In the most significant discrimination in clinical, namely those of AD from OD, our classifier achieved both sensitivity (SE) and specificity (SP) of 96%. In a similar way, our classifiers achieved a SE of 90% and a SP of 98% in AD from NC, as well as a SE of 88% and a SP of 86% in AD from OD and NC cases. Introducing SPM indices such as SVA and vbSEE, classification performances improved around 7-15%. We confirmed that these SPM factors are quite important for diagnosing Alzheimer's disease.

  7. Brucine diol-copper-catalyzed asymmetric synthesis of endo-pyrrolidines: the mechanistic dichotomy of imino esters.

    PubMed

    Li, Jian-Yuan; Kim, Hun Young; Oh, Kyungsoo

    2015-03-06

    Enantio- and diastereodivergent approaches to pyrrolidines are described by using catalyst- and substrate-controlled reaction pathways. A concerted endo-selective [3 + 2]-cycloaddition pathway is developed for the reaction of methyl imino ester, whereas endo-pyrrolidines with an opposite absolute stereochemical outcome are prepared by using the stepwise reaction pathway of tert-butyl imino ester. The development of catalyst- and substrate-controlled stereodivergent approaches highlights the inherent substrate-catalyst interactions in the [3 + 2]-cycloaddition reactions of metalated azomethine ylides.

  8. A simple randomisation procedure for validating discriminant analysis: a methodological note.

    PubMed

    Wastell, D G

    1987-04-01

    Because the goal of discriminant analysis (DA) is to optimise classification, it designedly exaggerates between-group differences. This bias complicates validation of DA. Jack-knifing has been used for validation but is inappropriate when stepwise selection (SWDA) is employed. A simple randomisation test is presented which is shown to give correct decisions for SWDA. The general superiority of randomisation tests over orthodox significance tests is discussed. Current work on non-parametric methods of estimating the error rates of prediction rules is briefly reviewed.

  9. The ground state of the Frenkel-Kontorova model

    NASA Astrophysics Data System (ADS)

    Babushkin, A. Yu.; Abkaryan, A. K.; Dobronets, B. S.; Krasikov, V. S.; Filonov, A. N.

    2016-09-01

    The continual approximation of the ground state of the discrete Frenkel-Kontorova model is tested using a symmetric algorithm of numerical simulation. A "kaleidoscope effect" is found, which means that the curves representing the dependences of the relative extension of an N-atom chain vary periodically with increasing N. Stairs of structural transitions for N ≫ 1 are analyzed by the channel selection method with the approximation N = ∞. Images of commensurable and incommensurable structures are constructed. The commensurable-incommensurable phase transitions are stepwise.

  10. A Multi-Faceted Approach to Inquiry-Based Learning

    NASA Astrophysics Data System (ADS)

    Brudzinski, M. R.; Sikorski, J.

    2009-12-01

    In order to fully attain the benefits of inquiry-based learning, instructors who typically employ the traditional lecture format need to make several adjustments to their approach. This change in styles can be intimidating and logistically difficult to overcome. A stepwise approach to this transformation is likely to be more manageable for individual faculty or departments. In this session, we will describe several features that we are implementing in our introductory geology course with the ultimate goal of converting to an entirely inquiry-based approach. Our project is part of the Miami University initiative in the top 25 enrolled courses to move towards the “student as scholar” model for engaged learning. Some of the features we developed for our course include: student learning outcomes, student development outcomes, out-of-class content quizzes, in-class conceptests, pre-/post-course assessment, reflective knowledge surveys, and daily group activities.

  11. Selective Audiovisual Semantic Integration Enabled by Feature-Selective Attention.

    PubMed

    Li, Yuanqing; Long, Jinyi; Huang, Biao; Yu, Tianyou; Wu, Wei; Li, Peijun; Fang, Fang; Sun, Pei

    2016-01-13

    An audiovisual object may contain multiple semantic features, such as the gender and emotional features of the speaker. Feature-selective attention and audiovisual semantic integration are two brain functions involved in the recognition of audiovisual objects. Humans often selectively attend to one or several features while ignoring the other features of an audiovisual object. Meanwhile, the human brain integrates semantic information from the visual and auditory modalities. However, how these two brain functions correlate with each other remains to be elucidated. In this functional magnetic resonance imaging (fMRI) study, we explored the neural mechanism by which feature-selective attention modulates audiovisual semantic integration. During the fMRI experiment, the subjects were presented with visual-only, auditory-only, or audiovisual dynamical facial stimuli and performed several feature-selective attention tasks. Our results revealed that a distribution of areas, including heteromodal areas and brain areas encoding attended features, may be involved in audiovisual semantic integration. Through feature-selective attention, the human brain may selectively integrate audiovisual semantic information from attended features by enhancing functional connectivity and thus regulating information flows from heteromodal areas to brain areas encoding the attended features.

  12. Development of a food frequency questionnaire for Sri Lankan adults

    PubMed Central

    2012-01-01

    Background Food Frequency Questionnaires (FFQs) are commonly used in epidemiologic studies to assess long-term nutritional exposure. Because of wide variations in dietary habits in different countries, a FFQ must be developed to suit the specific population. Sri Lanka is undergoing nutritional transition and diet-related chronic diseases are emerging as an important health problem. Currently, no FFQ has been developed for Sri Lankan adults. In this study, we developed a FFQ to assess the regular dietary intake of Sri Lankan adults. Methods A nationally representative sample of 600 adults was selected by a multi-stage random cluster sampling technique and dietary intake was assessed by random 24-h dietary recall. Nutrient analysis of the FFQ required the selection of foods, development of recipes and application of these to cooked foods to develop a nutrient database. We constructed a comprehensive food list with the units of measurement. A stepwise regression method was used to identify foods contributing to a cumulative 90% of variance to total energy and macronutrients. In addition, a series of photographs were included. Results We obtained dietary data from 482 participants and 312 different food items were recorded. Nutritionists grouped similar food items which resulted in a total of 178 items. After performing step-wise multiple regression, 93 foods explained 90% of the variance for total energy intake, carbohydrates, protein, total fat and dietary fibre. Finally, 90 food items and 12 photographs were selected. Conclusion We developed a FFQ and the related nutrient composition database for Sri Lankan adults. Culturally specific dietary tools are central to capturing the role of diet in risk for chronic disease in Sri Lanka. The next step will involve the verification of FFQ reproducibility and validity. PMID:22937734

  13. Is there a step-wise migration in Nigeria? A case study of the migrational histories of migrants in Lagos.

    PubMed

    Afolayan, A A

    1985-09-01

    "The paper sets out to test whether or not the movement pattern of people in Nigeria is step-wise. It examines the spatial order in the country and the movement pattern of people. It then analyzes the survey data and tests for the validity of step-wise migration in the country. The findings show that step-wise migration cannot adequately describe all the patterns observed." The presence of large-scale circulatory migration between rural and urban areas is noted. Ways to decrease the pressure on Lagos by developing intermediate urban areas are considered. excerpt

  14. EEG feature selection method based on decision tree.

    PubMed

    Duan, Lijuan; Ge, Hui; Ma, Wei; Miao, Jun

    2015-01-01

    This paper aims to solve automated feature selection problem in brain computer interface (BCI). In order to automate feature selection process, we proposed a novel EEG feature selection method based on decision tree (DT). During the electroencephalogram (EEG) signal processing, a feature extraction method based on principle component analysis (PCA) was used, and the selection process based on decision tree was performed by searching the feature space and automatically selecting optimal features. Considering that EEG signals are a series of non-linear signals, a generalized linear classifier named support vector machine (SVM) was chosen. In order to test the validity of the proposed method, we applied the EEG feature selection method based on decision tree to BCI Competition II datasets Ia, and the experiment showed encouraging results.

  15. Compensatory selection for roads over natural linear features by wolves in northern Ontario: Implications for caribou conservation

    PubMed Central

    Patterson, Brent R.; Anderson, Morgan L.; Rodgers, Arthur R.; Vander Vennen, Lucas M.; Fryxell, John M.

    2017-01-01

    Woodland caribou (Rangifer tarandus caribou) in Ontario are a threatened species that have experienced a substantial retraction of their historic range. Part of their decline has been attributed to increasing densities of anthropogenic linear features such as trails, roads, railways, and hydro lines. These features have been shown to increase the search efficiency and kill rate of wolves. However, it is unclear whether selection for anthropogenic linear features is additive or compensatory to selection for natural (water) linear features which may also be used for travel. We studied the selection of water and anthropogenic linear features by 52 resident wolves (Canis lupus x lycaon) over four years across three study areas in northern Ontario that varied in degrees of forestry activity and human disturbance. We used Euclidean distance-based resource selection functions (mixed-effects logistic regression) at the seasonal range scale with random coefficients for distance to water linear features, primary/secondary roads/railways, and hydro lines, and tertiary roads to estimate the strength of selection for each linear feature and for several habitat types, while accounting for availability of each feature. Next, we investigated the trade-off between selection for anthropogenic and water linear features. Wolves selected both anthropogenic and water linear features; selection for anthropogenic features was stronger than for water during the rendezvous season. Selection for anthropogenic linear features increased with increasing density of these features on the landscape, while selection for natural linear features declined, indicating compensatory selection of anthropogenic linear features. These results have implications for woodland caribou conservation. Prey encounter rates between wolves and caribou seem to be strongly influenced by increasing linear feature densities. This behavioral mechanism–a compensatory functional response to anthropogenic linear feature density resulting in decreased use of natural travel corridors–has negative consequences for the viability of woodland caribou. PMID:29117234

  16. Compensatory selection for roads over natural linear features by wolves in northern Ontario: Implications for caribou conservation.

    PubMed

    Newton, Erica J; Patterson, Brent R; Anderson, Morgan L; Rodgers, Arthur R; Vander Vennen, Lucas M; Fryxell, John M

    2017-01-01

    Woodland caribou (Rangifer tarandus caribou) in Ontario are a threatened species that have experienced a substantial retraction of their historic range. Part of their decline has been attributed to increasing densities of anthropogenic linear features such as trails, roads, railways, and hydro lines. These features have been shown to increase the search efficiency and kill rate of wolves. However, it is unclear whether selection for anthropogenic linear features is additive or compensatory to selection for natural (water) linear features which may also be used for travel. We studied the selection of water and anthropogenic linear features by 52 resident wolves (Canis lupus x lycaon) over four years across three study areas in northern Ontario that varied in degrees of forestry activity and human disturbance. We used Euclidean distance-based resource selection functions (mixed-effects logistic regression) at the seasonal range scale with random coefficients for distance to water linear features, primary/secondary roads/railways, and hydro lines, and tertiary roads to estimate the strength of selection for each linear feature and for several habitat types, while accounting for availability of each feature. Next, we investigated the trade-off between selection for anthropogenic and water linear features. Wolves selected both anthropogenic and water linear features; selection for anthropogenic features was stronger than for water during the rendezvous season. Selection for anthropogenic linear features increased with increasing density of these features on the landscape, while selection for natural linear features declined, indicating compensatory selection of anthropogenic linear features. These results have implications for woodland caribou conservation. Prey encounter rates between wolves and caribou seem to be strongly influenced by increasing linear feature densities. This behavioral mechanism-a compensatory functional response to anthropogenic linear feature density resulting in decreased use of natural travel corridors-has negative consequences for the viability of woodland caribou.

  17. Blind prediction of noncanonical RNA structure at atomic accuracy.

    PubMed

    Watkins, Andrew M; Geniesse, Caleb; Kladwang, Wipapat; Zakrevsky, Paul; Jaeger, Luc; Das, Rhiju

    2018-05-01

    Prediction of RNA structure from nucleotide sequence remains an unsolved grand challenge of biochemistry and requires distinct concepts from protein structure prediction. Despite extensive algorithmic development in recent years, modeling of noncanonical base pairs of new RNA structural motifs has not been achieved in blind challenges. We report a stepwise Monte Carlo (SWM) method with a unique add-and-delete move set that enables predictions of noncanonical base pairs of complex RNA structures. A benchmark of 82 diverse motifs establishes the method's general ability to recover noncanonical pairs ab initio, including multistrand motifs that have been refractory to prior approaches. In a blind challenge, SWM models predicted nucleotide-resolution chemical mapping and compensatory mutagenesis experiments for three in vitro selected tetraloop/receptors with previously unsolved structures (C7.2, C7.10, and R1). As a final test, SWM blindly and correctly predicted all noncanonical pairs of a Zika virus double pseudoknot during a recent community-wide RNA-Puzzle. Stepwise structure formation, as encoded in the SWM method, enables modeling of noncanonical RNA structure in a variety of previously intractable problems.

  18. McTwo: a two-step feature selection algorithm based on maximal information coefficient.

    PubMed

    Ge, Ruiquan; Zhou, Manli; Luo, Youxi; Meng, Qinghan; Mai, Guoqin; Ma, Dongli; Wang, Guoqing; Zhou, Fengfeng

    2016-03-23

    High-throughput bio-OMIC technologies are producing high-dimension data from bio-samples at an ever increasing rate, whereas the training sample number in a traditional experiment remains small due to various difficulties. This "large p, small n" paradigm in the area of biomedical "big data" may be at least partly solved by feature selection algorithms, which select only features significantly associated with phenotypes. Feature selection is an NP-hard problem. Due to the exponentially increased time requirement for finding the globally optimal solution, all the existing feature selection algorithms employ heuristic rules to find locally optimal solutions, and their solutions achieve different performances on different datasets. This work describes a feature selection algorithm based on a recently published correlation measurement, Maximal Information Coefficient (MIC). The proposed algorithm, McTwo, aims to select features associated with phenotypes, independently of each other, and achieving high classification performance of the nearest neighbor algorithm. Based on the comparative study of 17 datasets, McTwo performs about as well as or better than existing algorithms, with significantly reduced numbers of selected features. The features selected by McTwo also appear to have particular biomedical relevance to the phenotypes from the literature. McTwo selects a feature subset with very good classification performance, as well as a small feature number. So McTwo may represent a complementary feature selection algorithm for the high-dimensional biomedical datasets.

  19. Evaluation of interpolation techniques for the creation of gridded daily precipitation (1 × 1 km2); Cyprus, 1980-2010

    NASA Astrophysics Data System (ADS)

    Camera, Corrado; Bruggeman, Adriana; Hadjinicolaou, Panos; Pashiardis, Stelios; Lange, Manfred A.

    2014-01-01

    High-resolution gridded daily data sets are essential for natural resource management and the analyses of climate changes and their effects. This study aims to evaluate the performance of 15 simple or complex interpolation techniques in reproducing daily precipitation at a resolution of 1 km2 over topographically complex areas. Methods are tested considering two different sets of observation densities and different rainfall amounts. We used rainfall data that were recorded at 74 and 145 observational stations, respectively, spread over the 5760 km2 of the Republic of Cyprus, in the Eastern Mediterranean. Regression analyses utilizing geographical copredictors and neighboring interpolation techniques were evaluated both in isolation and combined. Linear multiple regression (LMR) and geographically weighted regression methods (GWR) were tested. These included a step-wise selection of covariables, as well as inverse distance weighting (IDW), kriging, and 3D-thin plate splines (TPS). The relative rank of the different techniques changes with different station density and rainfall amounts. Our results indicate that TPS performs well for low station density and large-scale events and also when coupled with regression models. It performs poorly for high station density. The opposite is observed when using IDW. Simple IDW performs best for local events, while a combination of step-wise GWR and IDW proves to be the best method for large-scale events and high station density. This study indicates that the use of step-wise regression with a variable set of geographic parameters can improve the interpolation of large-scale events because it facilitates the representation of local climate dynamics.

  20. A stepwise approach to the evaluation and treatment of subclinical hyperthyroidism.

    PubMed

    Mai, Vinh Q; Burch, Henry B

    2012-01-01

    To review a stepwise approach to the evaluation and treatment of subclinical hyperthyroidism. English-language articles regarding clinical management of subclinical hyperthyroidism published between 2007 and 2012 were reviewed. Subclinical hyperthyroidism is encountered on a daily basis in clinical practice. When evaluating patients with a suppressed serum thyrotropin value, it is important to exclude other potential etiologies such as overt triiodothyronine toxicosis, drug effect, nonthyroidal illness, and central hypothyroidism. In younger patients with mild thyrotropin suppression, it is acceptable to perform testing again in 3 to 6 months to assess for persistence before performing further diagnostic testing. In older patients or patients with thyrotropin values less than 0.1 mIU/L, diagnostic testing should proceed without delay. Persistence of thyrotropin suppression is more typical of nodular thyroid autonomy, whereas thyroiditis and mild Graves disease frequently resolve spontaneously. The clinical consequences of subclinical hyperthyroidism, such as atrial dysrhythmia, accelerated bone loss, increased fracture rate, and higher rates of cardiovascular mortality, are dependent on age and severity. The decision to treat subclinical hyperthyroidism is directly tied to an assessment of the potential for clinical consequences in untreated disease. Definitive therapy is generally selected for patients with nodular autonomous function, whereas antithyroid drug therapy is more appropriate for mild, persistent Graves disease. The presented stepwise approach to the care of patients presenting with an isolated suppression of serum thyrotropin focuses on the differential diagnosis, a prediction of the likelihood of persistence, an assessment of potential risks posed to the patient, and, finally, a personalized choice of therapy.

  1. Attentional Selection Can Be Predicted by Reinforcement Learning of Task-relevant Stimulus Features Weighted by Value-independent Stickiness.

    PubMed

    Balcarras, Matthew; Ardid, Salva; Kaping, Daniel; Everling, Stefan; Womelsdorf, Thilo

    2016-02-01

    Attention includes processes that evaluate stimuli relevance, select the most relevant stimulus against less relevant stimuli, and bias choice behavior toward the selected information. It is not clear how these processes interact. Here, we captured these processes in a reinforcement learning framework applied to a feature-based attention task that required macaques to learn and update the value of stimulus features while ignoring nonrelevant sensory features, locations, and action plans. We found that value-based reinforcement learning mechanisms could account for feature-based attentional selection and choice behavior but required a value-independent stickiness selection process to explain selection errors while at asymptotic behavior. By comparing different reinforcement learning schemes, we found that trial-by-trial selections were best predicted by a model that only represents expected values for the task-relevant feature dimension, with nonrelevant stimulus features and action plans having only a marginal influence on covert selections. These findings show that attentional control subprocesses can be described by (1) the reinforcement learning of feature values within a restricted feature space that excludes irrelevant feature dimensions, (2) a stochastic selection process on feature-specific value representations, and (3) value-independent stickiness toward previous feature selections akin to perseveration in the motor domain. We speculate that these three mechanisms are implemented by distinct but interacting brain circuits and that the proposed formal account of feature-based stimulus selection will be important to understand how attentional subprocesses are implemented in primate brain networks.

  2. Stepwise group sparse regression (SGSR): gene-set-based pharmacogenomic predictive models with stepwise selection of functional priors.

    PubMed

    Jang, In Sock; Dienstmann, Rodrigo; Margolin, Adam A; Guinney, Justin

    2015-01-01

    Complex mechanisms involving genomic aberrations in numerous proteins and pathways are believed to be a key cause of many diseases such as cancer. With recent advances in genomics, elucidating the molecular basis of cancer at a patient level is now feasible, and has led to personalized treatment strategies whereby a patient is treated according to his or her genomic profile. However, there is growing recognition that existing treatment modalities are overly simplistic, and do not fully account for the deep genomic complexity associated with sensitivity or resistance to cancer therapies. To overcome these limitations, large-scale pharmacogenomic screens of cancer cell lines--in conjunction with modern statistical learning approaches--have been used to explore the genetic underpinnings of drug response. While these analyses have demonstrated the ability to infer genetic predictors of compound sensitivity, to date most modeling approaches have been data-driven, i.e. they do not explicitly incorporate domain-specific knowledge (priors) in the process of learning a model. While a purely data-driven approach offers an unbiased perspective of the data--and may yield unexpected or novel insights--this strategy introduces challenges for both model interpretability and accuracy. In this study, we propose a novel prior-incorporated sparse regression model in which the choice of informative predictor sets is carried out by knowledge-driven priors (gene sets) in a stepwise fashion. Under regularization in a linear regression model, our algorithm is able to incorporate prior biological knowledge across the predictive variables thereby improving the interpretability of the final model with no loss--and often an improvement--in predictive performance. We evaluate the performance of our algorithm compared to well-known regularization methods such as LASSO, Ridge and Elastic net regression in the Cancer Cell Line Encyclopedia (CCLE) and Genomics of Drug Sensitivity in Cancer (Sanger) pharmacogenomics datasets, demonstrating that incorporation of the biological priors selected by our model confers improved predictability and interpretability, despite much fewer predictors, over existing state-of-the-art methods.

  3. Marker-Based Hierarchical Segmentation and Classification Approach for Hyperspectral Imagery

    NASA Technical Reports Server (NTRS)

    Tarabalka, Yuliya; Tilton, James C.; Benediktsson, Jon Atli; Chanussot, Jocelyn

    2011-01-01

    The Hierarchical SEGmentation (HSEG) algorithm, which is a combination of hierarchical step-wise optimization and spectral clustering, has given good performances for hyperspectral image analysis. This technique produces at its output a hierarchical set of image segmentations. The automated selection of a single segmentation level is often necessary. We propose and investigate the use of automatically selected markers for this purpose. In this paper, a novel Marker-based HSEG (M-HSEG) method for spectral-spatial classification of hyperspectral images is proposed. First, pixelwise classification is performed and the most reliably classified pixels are selected as markers, with the corresponding class labels. Then, a novel constrained marker-based HSEG algorithm is applied, resulting in a spectral-spatial classification map. The experimental results show that the proposed approach yields accurate segmentation and classification maps, and thus is attractive for hyperspectral image analysis.

  4. The effect of feature selection methods on computer-aided detection of masses in mammograms

    NASA Astrophysics Data System (ADS)

    Hupse, Rianne; Karssemeijer, Nico

    2010-05-01

    In computer-aided diagnosis (CAD) research, feature selection methods are often used to improve generalization performance of classifiers and shorten computation times. In an application that detects malignant masses in mammograms, we investigated the effect of using a selection criterion that is similar to the final performance measure we are optimizing, namely the mean sensitivity of the system in a predefined range of the free-response receiver operating characteristics (FROC). To obtain the generalization performance of the selected feature subsets, a cross validation procedure was performed on a dataset containing 351 abnormal and 7879 normal regions, each region providing a set of 71 mass features. The same number of noise features, not containing any information, were added to investigate the ability of the feature selection algorithms to distinguish between useful and non-useful features. It was found that significantly higher performances were obtained using feature sets selected by the general test statistic Wilks' lambda than using feature sets selected by the more specific FROC measure. Feature selection leads to better performance when compared to a system in which all features were used.

  5. Covariate Selection for Multilevel Models with Missing Data

    PubMed Central

    Marino, Miguel; Buxton, Orfeu M.; Li, Yi

    2017-01-01

    Missing covariate data hampers variable selection in multilevel regression settings. Current variable selection techniques for multiply-imputed data commonly address missingness in the predictors through list-wise deletion and stepwise-selection methods which are problematic. Moreover, most variable selection methods are developed for independent linear regression models and do not accommodate multilevel mixed effects regression models with incomplete covariate data. We develop a novel methodology that is able to perform covariate selection across multiply-imputed data for multilevel random effects models when missing data is present. Specifically, we propose to stack the multiply-imputed data sets from a multiple imputation procedure and to apply a group variable selection procedure through group lasso regularization to assess the overall impact of each predictor on the outcome across the imputed data sets. Simulations confirm the advantageous performance of the proposed method compared with the competing methods. We applied the method to reanalyze the Healthy Directions-Small Business cancer prevention study, which evaluated a behavioral intervention program targeting multiple risk-related behaviors in a working-class, multi-ethnic population. PMID:28239457

  6. A Java-based tool for the design of classification microarrays.

    PubMed

    Meng, Da; Broschat, Shira L; Call, Douglas R

    2008-08-04

    Classification microarrays are used for purposes such as identifying strains of bacteria and determining genetic relationships to understand the epidemiology of an infectious disease. For these cases, mixed microarrays, which are composed of DNA from more than one organism, are more effective than conventional microarrays composed of DNA from a single organism. Selection of probes is a key factor in designing successful mixed microarrays because redundant sequences are inefficient and limited representation of diversity can restrict application of the microarray. We have developed a Java-based software tool, called PLASMID, for use in selecting the minimum set of probe sequences needed to classify different groups of plasmids or bacteria. The software program was successfully applied to several different sets of data. The utility of PLASMID was illustrated using existing mixed-plasmid microarray data as well as data from a virtual mixed-genome microarray constructed from different strains of Streptococcus. Moreover, use of data from expression microarray experiments demonstrated the generality of PLASMID. In this paper we describe a new software tool for selecting a set of probes for a classification microarray. While the tool was developed for the design of mixed microarrays-and mixed-plasmid microarrays in particular-it can also be used to design expression arrays. The user can choose from several clustering methods (including hierarchical, non-hierarchical, and a model-based genetic algorithm), several probe ranking methods, and several different display methods. A novel approach is used for probe redundancy reduction, and probe selection is accomplished via stepwise discriminant analysis. Data can be entered in different formats (including Excel and comma-delimited text), and dendrogram, heat map, and scatter plot images can be saved in several different formats (including jpeg and tiff). Weights generated using stepwise discriminant analysis can be stored for analysis of subsequent experimental data. Additionally, PLASMID can be used to construct virtual microarrays with genomes from public databases, which can then be used to identify an optimal set of probes.

  7. A statistical model for Windstorm Variability over the British Isles based on Large-scale Atmospheric and Oceanic Mechanisms

    NASA Astrophysics Data System (ADS)

    Kirchner-Bossi, Nicolas; Befort, Daniel J.; Wild, Simon B.; Ulbrich, Uwe; Leckebusch, Gregor C.

    2016-04-01

    Time-clustered winter storms are responsible for a majority of the wind-induced losses in Europe. Over last years, different atmospheric and oceanic large-scale mechanisms as the North Atlantic Oscillation (NAO) or the Meridional Overturning Circulation (MOC) have been proven to drive some significant portion of the windstorm variability over Europe. In this work we systematically investigate the influence of different large-scale natural variability modes: more than 20 indices related to those mechanisms with proven or potential influence on the windstorm frequency variability over Europe - mostly SST- or pressure-based - are derived by means of ECMWF ERA-20C reanalysis during the last century (1902-2009), and compared to the windstorm variability for the European winter (DJF). Windstorms are defined and tracked as in Leckebusch et al. (2008). The derived indices are then employed to develop a statistical procedure including a stepwise Multiple Linear Regression (MLR) and an Artificial Neural Network (ANN), aiming to hindcast the inter-annual (DJF) regional windstorm frequency variability in a case study for the British Isles. This case study reveals 13 indices with a statistically significant coupling with seasonal windstorm counts. The Scandinavian Pattern (SCA) showed the strongest correlation (0.61), followed by the NAO (0.48) and the Polar/Eurasia Pattern (0.46). The obtained indices (standard-normalised) are selected as predictors for a windstorm variability hindcast model applied for the British Isles. First, a stepwise linear regression is performed, to identify which mechanisms can explain windstorm variability best. Finally, the indices retained by the stepwise regression are used to develop a multlayer perceptron-based ANN that hindcasted seasonal windstorm frequency and clustering. Eight indices (SCA, NAO, EA, PDO, W.NAtl.SST, AMO (unsmoothed), EA/WR and Trop.N.Atl SST) are retained by the stepwise regression. Among them, SCA showed the highest linear coefficient, followed by SST in western Atlantic, AMO and NAO. The explanatory regression model (considering all time steps) provided a Coefficient of Determination (R^2) of 0.75. A predictive version of the linear model applying a leave-one-out cross-validation (LOOCV) shows an R2 of 0.56 and a relative RMSE of 4.67 counts/season. An ANN-based nonlinear hindcast model for the seasonal windstorm frequency is developed with the aim to improve the stepwise hindcast ability and thus better predict a time-clustered season over the case study. A 7 node-hidden layer perceptron is set, and the LOOCV procedure reveals a R2 of 0.71. In comparison to the stepwise MLR the RMSE is reduced a 20%. This work shows that for the British Isles case study, most of the interannual variability can be explained by certain large-scale mechanisms, considering also nonlinear effects (ANN). This allows to discern a time-clustered season from a non-clustered one - a key issue for applications e.g., in the (re)insurance industry.

  8. Multiple-Primitives Hierarchical Classification of Airborne Laser Scanning Data in Urban Areas

    NASA Astrophysics Data System (ADS)

    Ni, H.; Lin, X. G.; Zhang, J. X.

    2017-09-01

    A hierarchical classification method for Airborne Laser Scanning (ALS) data of urban areas is proposed in this paper. This method is composed of three stages among which three types of primitives are utilized, i.e., smooth surface, rough surface, and individual point. In the first stage, the input ALS data is divided into smooth surfaces and rough surfaces by employing a step-wise point cloud segmentation method. In the second stage, classification based on smooth surfaces and rough surfaces is performed. Points in the smooth surfaces are first classified into ground and buildings based on semantic rules. Next, features of rough surfaces are extracted. Then, points in rough surfaces are classified into vegetation and vehicles based on the derived features and Random Forests (RF). In the third stage, point-based features are extracted for the ground points, and then, an individual point classification procedure is performed to classify the ground points into bare land, artificial ground and greenbelt. Moreover, the shortages of the existing studies are analyzed, and experiments show that the proposed method overcomes these shortages and handles more types of objects.

  9. Speech Emotion Feature Selection Method Based on Contribution Analysis Algorithm of Neural Network

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Wang Xiaojia; Mao Qirong; Zhan Yongzhao

    There are many emotion features. If all these features are employed to recognize emotions, redundant features may be existed. Furthermore, recognition result is unsatisfying and the cost of feature extraction is high. In this paper, a method to select speech emotion features based on contribution analysis algorithm of NN is presented. The emotion features are selected by using contribution analysis algorithm of NN from the 95 extracted features. Cluster analysis is applied to analyze the effectiveness for the features selected, and the time of feature extraction is evaluated. Finally, 24 emotion features selected are used to recognize six speech emotions.more » The experiments show that this method can improve the recognition rate and the time of feature extraction.« less

  10. Feature Selection Method Based on Neighborhood Relationships: Applications in EEG Signal Identification and Chinese Character Recognition

    PubMed Central

    Zhao, Yu-Xiang; Chou, Chien-Hsing

    2016-01-01

    In this study, a new feature selection algorithm, the neighborhood-relationship feature selection (NRFS) algorithm, is proposed for identifying rat electroencephalogram signals and recognizing Chinese characters. In these two applications, dependent relationships exist among the feature vectors and their neighboring feature vectors. Therefore, the proposed NRFS algorithm was designed for solving this problem. By applying the NRFS algorithm, unselected feature vectors have a high priority of being added into the feature subset if the neighboring feature vectors have been selected. In addition, selected feature vectors have a high priority of being eliminated if the neighboring feature vectors are not selected. In the experiments conducted in this study, the NRFS algorithm was compared with two feature algorithms. The experimental results indicated that the NRFS algorithm can extract the crucial frequency bands for identifying rat vigilance states and identifying crucial character regions for recognizing Chinese characters. PMID:27314346

  11. Aboveground Biomass Estimation Using Reconstructed Feature of Airborne Discrete-Return LIDAR by Auto-Encoder Neural Network

    NASA Astrophysics Data System (ADS)

    Li, T.; Wang, Z.; Peng, J.

    2018-04-01

    Aboveground biomass (AGB) estimation is critical for quantifying carbon stocks and essential for evaluating carbon cycle. In recent years, airborne LiDAR shows its great ability for highly-precision AGB estimation. Most of the researches estimate AGB by the feature metrics extracted from the canopy height distribution of the point cloud which calculated based on precise digital terrain model (DTM). However, if forest canopy density is high, the probability of the LiDAR signal penetrating the canopy is lower, resulting in ground points is not enough to establish DTM. Then the distribution of forest canopy height is imprecise and some critical feature metrics which have a strong correlation with biomass such as percentiles, maximums, means and standard deviations of canopy point cloud can hardly be extracted correctly. In order to address this issue, we propose a strategy of first reconstructing LiDAR feature metrics through Auto-Encoder neural network and then using the reconstructed feature metrics to estimate AGB. To assess the prediction ability of the reconstructed feature metrics, both original and reconstructed feature metrics were regressed against field-observed AGB using the multiple stepwise regression (MS) and the partial least squares regression (PLS) respectively. The results showed that the estimation model using reconstructed feature metrics improved R2 by 5.44 %, 18.09 %, decreased RMSE value by 10.06 %, 22.13 % and reduced RMSEcv by 10.00 %, 21.70 % for AGB, respectively. Therefore, reconstructing LiDAR point feature metrics has potential for addressing AGB estimation challenge in dense canopy area.

  12. Selective Audiovisual Semantic Integration Enabled by Feature-Selective Attention

    PubMed Central

    Li, Yuanqing; Long, Jinyi; Huang, Biao; Yu, Tianyou; Wu, Wei; Li, Peijun; Fang, Fang; Sun, Pei

    2016-01-01

    An audiovisual object may contain multiple semantic features, such as the gender and emotional features of the speaker. Feature-selective attention and audiovisual semantic integration are two brain functions involved in the recognition of audiovisual objects. Humans often selectively attend to one or several features while ignoring the other features of an audiovisual object. Meanwhile, the human brain integrates semantic information from the visual and auditory modalities. However, how these two brain functions correlate with each other remains to be elucidated. In this functional magnetic resonance imaging (fMRI) study, we explored the neural mechanism by which feature-selective attention modulates audiovisual semantic integration. During the fMRI experiment, the subjects were presented with visual-only, auditory-only, or audiovisual dynamical facial stimuli and performed several feature-selective attention tasks. Our results revealed that a distribution of areas, including heteromodal areas and brain areas encoding attended features, may be involved in audiovisual semantic integration. Through feature-selective attention, the human brain may selectively integrate audiovisual semantic information from attended features by enhancing functional connectivity and thus regulating information flows from heteromodal areas to brain areas encoding the attended features. PMID:26759193

  13. Highly efficient spin polarizer based on individual heterometallic cubane single-molecule magnets

    NASA Astrophysics Data System (ADS)

    Dong, Damin

    2015-09-01

    The spin-polarized transport across a single-molecule magnet [Mn3Zn(hmp)3O(N3)3(C3H5O2)3].2CHCl3 has been investigated using a density functional theory combined with Keldysh non-equilibrium Green's function formalism. It is shown that this single-molecule magnet has perfect spin filter behaviour. By adsorbing Ni3 cluster onto non-magnetic Au electrode, a large magnetoresistance exceeding 172% is found displaying molecular spin valve feature. Due to the tunneling via discrete quantum-mechanical states, the I-V curve has a stepwise character and negative differential resistance behaviour.

  14. [Population of Lytechinus variegatus (Echinoidea: Toxopneustidae) and structural characteristics of seagrass of Thalassia testudinum in Mochima Bay, Venezuela)].

    PubMed

    Noriega, Nicida; Cróquer, Aldo; Pauls, Sheila M

    2002-03-01

    To compare the general features of Thalassia testudinum seagrass at Mochima Bay with sea urchin (Lxtechinus variegatus) abundance and distribution, three T. testudinum seagrass beds were selected, from the mouth (strong wave exposure) to the inner bay (calm waters). Each site was surveyed by using 5 line transects (20 m long) parallel to the coast and 1 m2 quadrats. In situ measurements of T. testudinum cover, shoot and leaf density were taken. Estimation of dry biomass for each seagrass fraction (leaves, rhizomes and roots) and leaf length were obtained from 25 vegetation samples extracted per site using cores (15 cm diameter). A multivariate analysis of variance (Manova) and a less significative difference test (LSD) were performed to examine differences between sites and within sites at different depths. A stepwise multiple regression analysis was done, dependent variable was sea urchin density; independent variables: vegetation values at each site. The only seagrass species found in the three sites was T. testudinum, and cover was 56-100%, leaf density 100-1000 leaf/m2, lengths 6-18.8 cm and shoot density 20-475 shoots/m2. The highest sea urchin densities were found at Isla Redonda and Ensenada Toporo (1-3.6 ind/m2), the lowest at Playa Colorada (0.6-0.8 ind/m2). Significant differences in seagrass features between sites were obtained (Manova p < 0.001), but not between depths (Manova p < 0.320). The regression coefficient between sea urchin density and seagrass parameters was statistically significant (r2 = 0.154, p < 0.007), however, total biomass was the only variable with a significant effect on sea urchin distribution (beta = 0.308, p < 0.032). The other variables did not explain satisfactorily L. variegatus abundance and distribution.

  15. Ultrasonographic markers and preoperative CA-125 to distinguish between borderline ovarian tumors and stage I ovarian cancer.

    PubMed

    Zacharakis, Dimitrios; Thomakos, Nikolaos; Biliatis, Ioannis; Rodolakis, Alexandros; Simou, Maria; Daskalakis, Georgios; Bamias, Aris; Antsaklis, Aris

    2013-03-01

    Preoperative evaluation of ovarian masses has become increasingly important for optimal planning of treatment. The aim of this study was to assess the role of preoperative serum cancer antigen 125 (CA-125) levels in correlation with ultrasonographic features in order to distinguish between borderline ovarian tumors (BOTs) and stage I epithelial ovarian carcinoma (EOC). Retrospective study. Tertiary University Hospital. We reviewed all women with BOTs and stage I EOC from January 2000 to December 2010. Data from 165 women (66 BOTs and 99 stage I EOC) were analyzed. Multivariable logistic regression with stepwise selection of variables was used to determine which clinical variables, ultrasound features and CA-125 level were independently associated with invasiveness. Utility of ultrasonographic markers and CA-125 in the preoperative differential diagnosis between BOTs and stage I EOC. Women with CA-125 > 100 IU mL(-1) had almost three times greater likelihood of belonging in the EOC group [odds ratio (OR) 3.02; confidence interval (CI) 95%: 1.13-8.12]. Furthermore, the presence of large solid component (≥20% of the tumor comprised of solid components) was associated with 4.25 times greater odds of it to representing ovarian cancer rather than a BOT (OR 4.25; 95% CI: 2.05-8.82). In contrast, the presence of papillary projections was associated with a 73% lower likelihood of EOC (OR 0.27; 95% CI: 0.13-0.58). Preoperative CA-125 > 100 IU mL(-1) combined with the presence of a large solid component and the absence of papillary projections seems to improve the discriminative ability in favor of stage I EOC. © 2012 The Authors Acta Obstetricia et Gynecologica Scandinavica © 2012 Nordic Federation of Societies of Obstetrics and Gynecology.

  16. Natural image classification driven by human brain activity

    NASA Astrophysics Data System (ADS)

    Zhang, Dai; Peng, Hanyang; Wang, Jinqiao; Tang, Ming; Xue, Rong; Zuo, Zhentao

    2016-03-01

    Natural image classification has been a hot topic in computer vision and pattern recognition research field. Since the performance of an image classification system can be improved by feature selection, many image feature selection methods have been developed. However, the existing supervised feature selection methods are typically driven by the class label information that are identical for different samples from the same class, ignoring with-in class image variability and therefore degrading the feature selection performance. In this study, we propose a novel feature selection method, driven by human brain activity signals collected using fMRI technique when human subjects were viewing natural images of different categories. The fMRI signals associated with subjects viewing different images encode the human perception of natural images, and therefore may capture image variability within- and cross- categories. We then select image features with the guidance of fMRI signals from brain regions with active response to image viewing. Particularly, bag of words features based on GIST descriptor are extracted from natural images for classification, and a sparse regression base feature selection method is adapted to select image features that can best predict fMRI signals. Finally, a classification model is built on the select image features to classify images without fMRI signals. The validation experiments for classifying images from 4 categories of two subjects have demonstrated that our method could achieve much better classification performance than the classifiers built on image feature selected by traditional feature selection methods.

  17. EFS: an ensemble feature selection tool implemented as R-package and web-application.

    PubMed

    Neumann, Ursula; Genze, Nikita; Heider, Dominik

    2017-01-01

    Feature selection methods aim at identifying a subset of features that improve the prediction performance of subsequent classification models and thereby also simplify their interpretability. Preceding studies demonstrated that single feature selection methods can have specific biases, whereas an ensemble feature selection has the advantage to alleviate and compensate for these biases. The software EFS (Ensemble Feature Selection) makes use of multiple feature selection methods and combines their normalized outputs to a quantitative ensemble importance. Currently, eight different feature selection methods have been integrated in EFS, which can be used separately or combined in an ensemble. EFS identifies relevant features while compensating specific biases of single methods due to an ensemble approach. Thereby, EFS can improve the prediction accuracy and interpretability in subsequent binary classification models. EFS can be downloaded as an R-package from CRAN or used via a web application at http://EFS.heiderlab.de.

  18. RRegrs: an R package for computer-aided model selection with multiple regression models.

    PubMed

    Tsiliki, Georgia; Munteanu, Cristian R; Seoane, Jose A; Fernandez-Lozano, Carlos; Sarimveis, Haralambos; Willighagen, Egon L

    2015-01-01

    Predictive regression models can be created with many different modelling approaches. Choices need to be made for data set splitting, cross-validation methods, specific regression parameters and best model criteria, as they all affect the accuracy and efficiency of the produced predictive models, and therefore, raising model reproducibility and comparison issues. Cheminformatics and bioinformatics are extensively using predictive modelling and exhibit a need for standardization of these methodologies in order to assist model selection and speed up the process of predictive model development. A tool accessible to all users, irrespectively of their statistical knowledge, would be valuable if it tests several simple and complex regression models and validation schemes, produce unified reports, and offer the option to be integrated into more extensive studies. Additionally, such methodology should be implemented as a free programming package, in order to be continuously adapted and redistributed by others. We propose an integrated framework for creating multiple regression models, called RRegrs. The tool offers the option of ten simple and complex regression methods combined with repeated 10-fold and leave-one-out cross-validation. Methods include Multiple Linear regression, Generalized Linear Model with Stepwise Feature Selection, Partial Least Squares regression, Lasso regression, and Support Vector Machines Recursive Feature Elimination. The new framework is an automated fully validated procedure which produces standardized reports to quickly oversee the impact of choices in modelling algorithms and assess the model and cross-validation results. The methodology was implemented as an open source R package, available at https://www.github.com/enanomapper/RRegrs, by reusing and extending on the caret package. The universality of the new methodology is demonstrated using five standard data sets from different scientific fields. Its efficiency in cheminformatics and QSAR modelling is shown with three use cases: proteomics data for surface-modified gold nanoparticles, nano-metal oxides descriptor data, and molecular descriptors for acute aquatic toxicity data. The results show that for all data sets RRegrs reports models with equal or better performance for both training and test sets than those reported in the original publications. Its good performance as well as its adaptability in terms of parameter optimization could make RRegrs a popular framework to assist the initial exploration of predictive models, and with that, the design of more comprehensive in silico screening applications.Graphical abstractRRegrs is a computer-aided model selection framework for R multiple regression models; this is a fully validated procedure with application to QSAR modelling.

  19. Selection and outcome of the potential live liver donor.

    PubMed

    Pamecha, Viniyendra; Mahansaria, Shyam Sunder; Bharathy, Kishore G S; Kumar, Senthil; Sasturkar, Shridhar Vasantrao; Sinha, Piyush Kumar; Sarin, Shiv Kumar

    2016-07-01

    A thorough donor evaluation in the living donation process is mandatory to ensure a safe outcome in an otherwise healthy individual. The aim of the current study was to evaluate the reasons for not proceeding to donation and the outcome of live liver donors. A prospective study of potential donors who underwent evaluation and proceeded to surgery from 1 April 2012 to 31 January 2015 was conducted. The process of donor selection, its outcome and peri-operative complications were recorded. A total of 460 donors were evaluated in a stepwise manner for 367 potential recipients. Of the 321 (69.7 %) donors not proceeding to donation, the reasons were donor-related in 63.6 % and recipient-related in the rest. Common donor-related reasons were: donor reluctance (23.5 %), negative liver attenuation index (16.2 %), anatomic variations (10.3 %), inadequate remnant liver volume (9.8 %), unacceptable liver biopsy (8.8 %), and inadequate graft volume (5.4 %). A majority of donors (82.8 %) were turned down early in the (steps 1 and 2) evaluation process. Recipient death was the most common recipient-related reason [n = 51 (43.6 %)] for not proceeding to donation. There was no donor mortality. The overall complication rate was 19.8 % and major complication rate (grade 3 or higher) was 4.4 %. A stringent stepwise donor evaluation process leads to early recognition of unsuitable donors and a low complication rate.

  20. Facies-dependent variations in sediment physical properties on the Mississippi River Delta Front, USA: evidence for depositional and post-depositional processes

    NASA Astrophysics Data System (ADS)

    Smith, J. E., IV; Bentley, S. J.; Courtois, A. J.; Obelcz, J.; Chaytor, J. D.; Maloney, J. M.; Georgiou, I. Y.; Xu, K.; Miner, M. D.

    2017-12-01

    Recent studies on Mississippi River Delta have documented sub-aerial land loss, driven in part by declining sediment load over the past century. Impacts of changing sediment load on the subaqueous delta are less well known. The subaqueous Mississippi River Delta Front is known to be shaped by extensive submarine mudflows operating at a range of temporal and spatial scales, however impacts of changing sediment delivery on mudflow deposits have not been investigated. To better understand seabed morphology and stratigraphy as impacted by plume sedimentation and mudflows, an integrated geological/geophysical study was undertaken in delta front regions offshore the three main passes of the Mississippi River Delta. This study focuses on stratigraphy and physical properties of 30 piston cores (5-9 m length) collected in June 2017. Coring locations were selected in gully, lobe and prodelta settings based on multibeam bathymetry and seismic profiles collected in mid-May 2017. Cores were analyzed for density, magnetic susceptibility, P-wave speed, and resistivity using a Geotek multi sensor core logger; here, we focus on density data. Core density profiles generally vary systematically across facies. Density profiles of gully cores are nearly invariant with some downward stepwise increases delineating units meters thick, and abundant gaps likely caused by gas expansion. Lobe cores generally have subtle downward increases in density, some stepwise density increases, and fewer gaps. Prodelta cores show more pronounced downward density increases, decimeter-scale peaks and valleys in density profiles, but stepwise increases are less evident. We hypothesize that density profiles in gully and lobe settings (uniform profiles except for stepwise increases) reflect remolding by mudflows, whereas density variations in prodelta settings instead reflect grain size variations (decimeter-scale) and more advanced consolidation (overall downward density increase) consistent with slower sediment deposition. These hypotheses will be evaluated by a more detailed study of seismic stratigraphy and core properties, including geochronology, grain size distribution and X-radiographic imaging, to further relate important sedimentary processes with resulting deposits.

  1. Proton radius from electron scattering data

    NASA Astrophysics Data System (ADS)

    Higinbotham, Douglas W.; Kabir, Al Amin; Lin, Vincent; Meekins, David; Norum, Blaine; Sawatzky, Brad

    2016-05-01

    Background: The proton charge radius extracted from recent muonic hydrogen Lamb shift measurements is significantly smaller than that extracted from atomic hydrogen and electron scattering measurements. The discrepancy has become known as the proton radius puzzle. Purpose: In an attempt to understand the discrepancy, we review high-precision electron scattering results from Mainz, Jefferson Lab, Saskatoon, and Stanford. Methods: We make use of stepwise regression techniques using the F test as well as the Akaike information criterion to systematically determine the predictive variables to use for a given set and range of electron scattering data as well as to provide multivariate error estimates. Results: Starting with the precision, low four-momentum transfer (Q2) data from Mainz (1980) and Saskatoon (1974), we find that a stepwise regression of the Maclaurin series using the F test as well as the Akaike information criterion justify using a linear extrapolation which yields a value for the proton radius that is consistent with the result obtained from muonic hydrogen measurements. Applying the same Maclaurin series and statistical criteria to the 2014 Rosenbluth results on GE from Mainz, we again find that the stepwise regression tends to favor a radius consistent with the muonic hydrogen radius but produces results that are extremely sensitive to the range of data included in the fit. Making use of the high-Q2 data on GE to select functions which extrapolate to high Q2, we find that a Padé (N =M =1 ) statistical model works remarkably well, as does a dipole function with a 0.84 fm radius, GE(Q2) =(1+Q2/0.66 GeV2) -2 . Conclusions: Rigorous applications of stepwise regression techniques and multivariate error estimates result in the extraction of a proton charge radius that is consistent with the muonic hydrogen result of 0.84 fm; either from linear extrapolation of the extremely-low-Q2 data or by use of the Padé approximant for extrapolation using a larger range of data. Thus, based on a purely statistical analysis of electron scattering data, we conclude that the electron scattering results and the muonic hydrogen results are consistent. It is the atomic hydrogen results that are the outliers.

  2. Feature selection methods for big data bioinformatics: A survey from the search perspective.

    PubMed

    Wang, Lipo; Wang, Yaoli; Chang, Qing

    2016-12-01

    This paper surveys main principles of feature selection and their recent applications in big data bioinformatics. Instead of the commonly used categorization into filter, wrapper, and embedded approaches to feature selection, we formulate feature selection as a combinatorial optimization or search problem and categorize feature selection methods into exhaustive search, heuristic search, and hybrid methods, where heuristic search methods may further be categorized into those with or without data-distilled feature ranking measures. Copyright © 2016 Elsevier Inc. All rights reserved.

  3. The effect of thermal history on crystalline structure and mechanical properties of β-nucleated isotactic polypropylene

    NASA Astrophysics Data System (ADS)

    Tian, Yefei; Zhou, Jian; Feng, Jiachun

    2018-04-01

    The effect of thermal history on β-nucleated iPP was systematically investigated by comparing the variance of crystalline microstructures and mechanical property of stepwise crystallized sample and annealed sample, which experienced different thermal history. The mechanical property tests exhibit that that the toughness of stepwise crystallized sample and annealed sample were both decreased compared to control sample, while the tensile strength of the stepwise crystallized sample increased slightly. Structure investigation showed that the α-relaxation peak, which is related to the assignment of chains in rigid amorphous phase, moved to the high temperature region for stepwise crystallized sample, while it moved to the low temperature region for annealed sample. The results indicated the weakening in rigid amorphous fraction (RAF) and the increase in lamellar thickness of β-iPP after stepwise crystallization treatment. For annealed sample, the RAF strengthened and lamellar thickness decreased slightly after thermal treatment. A mechanism of crystalline microstructures evolution of restricted area between the main lamellar under different treatments was proposed.

  4. Improving meat quality of organic pork through post mortem handling of carcasses: an innovative approach.

    PubMed

    Therkildsen, Margrethe; Kristensen, Lars; Kyed, Sybille; Oksbjerg, Niels

    2012-06-01

    This study was conducted to examine the best combination of post mortem chilling, suspension and ageing in order to optimize tenderness of organic pork at slaughter, which may be tougher than conventionally produced pork, because of lower daily gain. Combinations of stepwise chilling with a holding period of 6h at 10°C or traditional blast tunnel chilling, suspension in the pelvic bone or Achilles Tendon and ageing 2 or 4 days post mortem were tested. Stepwise chilling and ageing improved tenderness of the loin, and the effects were additive, whereas pelvic suspension was less effective in texture improvements, and non-additive to stepwise chilling. Stepwise chilling improved tenderness to a similar degree as can be obtained within 2-4 days of extended ageing, however, the minimum temperature during the holding period seems to be crucial in order to obtain a positive effect of stepwise chilling, and it should be above 7.5°C. Copyright © 2011 Elsevier Ltd. All rights reserved.

  5. Feature selection method based on multi-fractal dimension and harmony search algorithm and its application

    NASA Astrophysics Data System (ADS)

    Zhang, Chen; Ni, Zhiwei; Ni, Liping; Tang, Na

    2016-10-01

    Feature selection is an important method of data preprocessing in data mining. In this paper, a novel feature selection method based on multi-fractal dimension and harmony search algorithm is proposed. Multi-fractal dimension is adopted as the evaluation criterion of feature subset, which can determine the number of selected features. An improved harmony search algorithm is used as the search strategy to improve the efficiency of feature selection. The performance of the proposed method is compared with that of other feature selection algorithms on UCI data-sets. Besides, the proposed method is also used to predict the daily average concentration of PM2.5 in China. Experimental results show that the proposed method can obtain competitive results in terms of both prediction accuracy and the number of selected features.

  6. Feet deformities are correlated with impaired balance and postural stability in seniors over 75

    PubMed Central

    Puszczalowska-Lizis, Ewa; Bujas, Przemyslaw; Omorczyk, Jaroslaw; Jandzis, Slawomir

    2017-01-01

    Objective Understanding the factors and mechanisms that determine balance in seniors appears vital in terms of their self-reliance and overall safety. The study aimed to determine the relationship between the features of feet structure and the indicators of postural stability in the elderly. Methods The study group comprised 80 seniors (41F, 39M; aged 75–85 years). CQ-ST podoscope and the CQ-Stab 2P two-platform posturograph were used as primary research tools. The data were analyzed based on Spearman’s rank correlation and forward stepwise regression. Results Analysis of forward stepwise regression identified the left foot length in females and Clarke’s angle of the left foot in men as significant and independent predictors of postural up to 30% of the variance of dependent variables. Conclusions Longer feet provide older women with better stability, whereas in men, the lowering of the longitudinal arch results in postural deterioration. In the elderly, the left lower limb shows greater activity in the stabilizing processes in the standing position than the right one. In gerontological rehabilitation special attention should be paid to the individually tailored, gender-specific treatment, with a view to enhancing overall safety and quality of seniors’ lives. PMID:28877185

  7. Assessment of the integrity of concrete bridge structures by acoustic emission technique

    NASA Astrophysics Data System (ADS)

    Yoon, Dong-Jin; Park, Philip; Jung, Juong-Chae; Lee, Seung-Seok

    2002-06-01

    This study was aimed at developing a new method for assessing the integrity of concrete structures. Especially acoustic emission technique was used in carrying out both laboratory experiment and field application. From the previous laboratory study, we confirmed that AE analysis provided a promising approach for estimating the level of damage and distress in concrete structures. The Felicity ratio, one of the key parameter for assessing damage, exhibits a favorable correlation with the overall damage level. The total number of AE events under stepwise cyclic loading also showed a good agreement with the damage level. In this study, a new suggested technique was applied to several concrete bridges in Korea in order to verify the applicability in field. The AE response was analyzed to obtain key parameters such as the total number and rate of AE events, AE parameter analysis for each event, and the characteristic features of the waveform as well as Felicity ratio analysis. Stepwise loading-unloading procedure for AE generation was introduced in field test by using each different weight of vehicle. According to the condition of bridge, for instance new or old bridge, AE event rate and AE generation behavior indicated many different aspects. The results showed that the suggested analyzing method would be a promising approach for assessing the integrity of concrete structures.

  8. Stepwise introduction of laparoscopic liver surgery: validation of guideline recommendations.

    PubMed

    van der Poel, Marcel J; Huisman, Floor; Busch, Olivier R; Abu Hilal, Mohammad; van Gulik, Thomas M; Tanis, Pieter J; Besselink, Marc G

    2017-10-01

    Uncontrolled introduction of laparoscopic liver surgery (LLS) could compromise postoperative outcomes. A stepwise introduction of LLS combined with structured training is advised. This study aimed to evaluate the impact of such a stepwise introduction. A retrospective, single-center case series assessing short term outcomes of all consecutive LLS in the period November 2006-January 2017. The technique was implemented in a stepwise fashion. To evaluate the impact of this stepwise approach combined with structured training, outcomes of LLS before and after a laparoscopic HPB fellowship were compared. A total of 135 laparoscopic resections were performed. Overall conversion rate was 4% (n = 5), clinically relevant complication rate 13% (n = 18) and mortality 0.7% (n = 1). A significant increase in patients with major LLS, multiple liver resections, previous abdominal surgery, malignancies and lesions located in posterior segments was observed after the fellowship as well as a decrease in the use of hand-assistance. Increasing complexity in the post fellowship period was reflected by an increase in operating times, but without comprising other surgical outcomes. A stepwise introduction of LLS combined with structured training reduced the clinical impact of the learning curve, thereby confirming guideline recommendations. Copyright © 2017 International Hepato-Pancreato-Biliary Association Inc. Published by Elsevier Ltd. All rights reserved.

  9. Kernel-based Joint Feature Selection and Max-Margin Classification for Early Diagnosis of Parkinson’s Disease

    NASA Astrophysics Data System (ADS)

    Adeli, Ehsan; Wu, Guorong; Saghafi, Behrouz; An, Le; Shi, Feng; Shen, Dinggang

    2017-01-01

    Feature selection methods usually select the most compact and relevant set of features based on their contribution to a linear regression model. Thus, these features might not be the best for a non-linear classifier. This is especially crucial for the tasks, in which the performance is heavily dependent on the feature selection techniques, like the diagnosis of neurodegenerative diseases. Parkinson’s disease (PD) is one of the most common neurodegenerative disorders, which progresses slowly while affects the quality of life dramatically. In this paper, we use the data acquired from multi-modal neuroimaging data to diagnose PD by investigating the brain regions, known to be affected at the early stages. We propose a joint kernel-based feature selection and classification framework. Unlike conventional feature selection techniques that select features based on their performance in the original input feature space, we select features that best benefit the classification scheme in the kernel space. We further propose kernel functions, specifically designed for our non-negative feature types. We use MRI and SPECT data of 538 subjects from the PPMI database, and obtain a diagnosis accuracy of 97.5%, which outperforms all baseline and state-of-the-art methods.

  10. Kernel-based Joint Feature Selection and Max-Margin Classification for Early Diagnosis of Parkinson’s Disease

    PubMed Central

    Adeli, Ehsan; Wu, Guorong; Saghafi, Behrouz; An, Le; Shi, Feng; Shen, Dinggang

    2017-01-01

    Feature selection methods usually select the most compact and relevant set of features based on their contribution to a linear regression model. Thus, these features might not be the best for a non-linear classifier. This is especially crucial for the tasks, in which the performance is heavily dependent on the feature selection techniques, like the diagnosis of neurodegenerative diseases. Parkinson’s disease (PD) is one of the most common neurodegenerative disorders, which progresses slowly while affects the quality of life dramatically. In this paper, we use the data acquired from multi-modal neuroimaging data to diagnose PD by investigating the brain regions, known to be affected at the early stages. We propose a joint kernel-based feature selection and classification framework. Unlike conventional feature selection techniques that select features based on their performance in the original input feature space, we select features that best benefit the classification scheme in the kernel space. We further propose kernel functions, specifically designed for our non-negative feature types. We use MRI and SPECT data of 538 subjects from the PPMI database, and obtain a diagnosis accuracy of 97.5%, which outperforms all baseline and state-of-the-art methods. PMID:28120883

  11. Talent identification and selection in elite youth football: An Australian context.

    PubMed

    O'Connor, Donna; Larkin, Paul; Mark Williams, A

    2016-10-01

    We identified the perceptual-cognitive skills and player history variables that differentiate players selected or not selected into an elite youth football (i.e. soccer) programme in Australia. A sample of elite youth male football players (n = 127) completed an adapted participation history questionnaire and video-based assessments of perceptual-cognitive skills. Following data collection, 22 of these players were offered a full-time scholarship for enrolment at an elite player residential programme. Participants selected for the scholarship programme recorded superior performance on the combined perceptual-cognitive skills tests compared to the non-selected group. There were no significant between group differences on the player history variables. Stepwise discriminant function analysis identified four predictor variables that resulted in the best categorization of selected and non-selected players (i.e. recent match-play performance, region, number of other sports participated, combined perceptual-cognitive performance). The effectiveness of the discriminant function is reflected by 93.7% of players being correctly classified, with the four variables accounting for 57.6% of the variance. Our discriminating model for selection may provide a greater understanding of the factors that influence elite youth talent selection and identification.

  12. Integrated feature extraction and selection for neuroimage classification

    NASA Astrophysics Data System (ADS)

    Fan, Yong; Shen, Dinggang

    2009-02-01

    Feature extraction and selection are of great importance in neuroimage classification for identifying informative features and reducing feature dimensionality, which are generally implemented as two separate steps. This paper presents an integrated feature extraction and selection algorithm with two iterative steps: constrained subspace learning based feature extraction and support vector machine (SVM) based feature selection. The subspace learning based feature extraction focuses on the brain regions with higher possibility of being affected by the disease under study, while the possibility of brain regions being affected by disease is estimated by the SVM based feature selection, in conjunction with SVM classification. This algorithm can not only take into account the inter-correlation among different brain regions, but also overcome the limitation of traditional subspace learning based feature extraction methods. To achieve robust performance and optimal selection of parameters involved in feature extraction, selection, and classification, a bootstrapping strategy is used to generate multiple versions of training and testing sets for parameter optimization, according to the classification performance measured by the area under the ROC (receiver operating characteristic) curve. The integrated feature extraction and selection method is applied to a structural MR image based Alzheimer's disease (AD) study with 98 non-demented and 100 demented subjects. Cross-validation results indicate that the proposed algorithm can improve performance of the traditional subspace learning based classification.

  13. Compact cancer biomarkers discovery using a swarm intelligence feature selection algorithm.

    PubMed

    Martinez, Emmanuel; Alvarez, Mario Moises; Trevino, Victor

    2010-08-01

    Biomarker discovery is a typical application from functional genomics. Due to the large number of genes studied simultaneously in microarray data, feature selection is a key step. Swarm intelligence has emerged as a solution for the feature selection problem. However, swarm intelligence settings for feature selection fail to select small features subsets. We have proposed a swarm intelligence feature selection algorithm based on the initialization and update of only a subset of particles in the swarm. In this study, we tested our algorithm in 11 microarray datasets for brain, leukemia, lung, prostate, and others. We show that the proposed swarm intelligence algorithm successfully increase the classification accuracy and decrease the number of selected features compared to other swarm intelligence methods. Copyright © 2010 Elsevier Ltd. All rights reserved.

  14. A stepwise procedure for science communication in the field

    NASA Astrophysics Data System (ADS)

    Nisancioglu, Kerim; Paasche, Øyvind

    2017-04-01

    Communicating and disseminating earth science to laypersons, high-school students and their teachers are becoming increasingly important considering the overwhelming impact human civilization have on the planet. One of the main challenges with this type of dissemination arises from the cross-disciplinary nature of the Earth system as it encompasses anything from cloud physics to the geological evidence of ice ages being played out on millennial time scales. During the last four years we have tested and developed an approach referred to as «Turspor» which can be translated to 'Trail Tracks'. The ambition with "Turspor" is to inspire participants to seek in-depth knowledge relating to observations of features made in the field (glacial moraines, active permafrost, clouds, winds and so forth) as we have come to learn that observations made in the field enhances students capability to grasp the bare essentials related to the phenomena in question. By engaging master and PhD students in the process we create a platform where students can improve their teaching and communicative skills through a stepwise procedure. The initial concept was tested on 35 high school students during the summer of 2012 in the mountainous area of Snøheim on Dovre, Southern Norway. Before the arrival of the high school students, the university students prepared one page written summaries describing relevant geological or meteorological features and trained on how to best disseminate a basic scientific understanding of these. Specific examples were patterned ground caused by permafrost, glacier flour, katabatic winds, and equilibrium line altitude of glaciers. Based on the success of the program over the past 4 years with field trips together with local schools, we are in the process of developing the concept to be offered as a course at the master and PhD level, including a week of training in didactics applied to topics in the geosciences as well as practical training in the field. The university students who complete the course will be in charge of guiding local high school students in the field (mostly at university field stations in the mountains of Norway), as well as follow up on the topics discussed in collaboration with the high-school teachers during a subsequent visit to the classroom. By participating in the program, high-school students will experience inquiry-based learning, adding to their understanding of the scientific process. Sites developed through this program so far have been documented with a GPS tag, detailed description, background and pictures and hosted by the Norwegian Trekking Association website: ut.no as well as an associated app. The stepwise procedure of this concept has been developed through continous interaction with teachers, students and the general public. As a consequence, our capability to address and explain many earth science features of large societal relevance such as glacier retreat, thawing permafrost, flooding, snow avalanches, vegetation dynamics and climate change in general has been greatly improved. For examples see https://www.ut.no/gruppe/8.1819/

  15. Selective CO2 Sequestration with Monolithic Bimodal Micro/Macroporous Carbon Aerogels Derived from Stepwise Pyrolytic Decomposition of Polyamide-Polyimide-Polyurea Random Copolymers.

    PubMed

    Saeed, Adnan M; Rewatkar, Parwani M; Majedi Far, Hojat; Taghvaee, Tahereh; Donthula, Suraj; Mandal, Chandana; Sotiriou-Leventis, Chariklia; Leventis, Nicholas

    2017-04-19

    Polymeric aerogels (PA-xx) were synthesized via room-temperature reaction of an aromatic triisocyanate (tris(4-isocyanatophenyl) methane) with pyromellitic acid. Using solid-state CPMAS 13 C and 15 N NMR, it was found that the skeletal framework of PA-xx was a statistical copolymer of polyamide, polyurea, polyimide, and of the primary condensation product of the two reactants, a carbamic-anhydride adduct. Stepwise pyrolytic decomposition of those components yielded carbon aerogels with both open and closed microporosity. The open micropore surface area increased from <15 m 2 g -1 in PA-xx to 340 m 2 g -1 in the carbons. Next, reactive etching at 1,000 °C with CO 2 opened access to the closed pores and the micropore area increased by almost 4× to 1150 m 2 g -1 (out of 1750 m 2 g -1 of a total BET surface area). At 0 °C, etched carbon aerogels demonstrated a good balance of adsorption capacity for CO 2 (up to 4.9 mmol g -1 ), and selectivity toward other gases (via Henry's law). The selectivity for CO 2 versus H 2 (up to 928:1) is suitable for precombustion fuel purification. Relevant to postcombustion CO 2 capture and sequestration (CCS), the selectivity for CO 2 versus N 2 was in the 17:1 to 31:1 range. In addition to typical factors involved in gas sorption (kinetic diameters, quadrupole moments and polarizabilities of the adsorbates), it is also suggested that CO 2 is preferentially engaged by surface pyridinic and pyridonic N on carbon (identified with XPS) in an energy-neutral surface reaction. Relatively high uptake of CH 4 (2.16 mmol g -1 at 0 °C/1 bar) was attributed to its low polarizability, and that finding paves the way for further studies on adsorption of higher (i.e., more polarizable) hydrocarbons. Overall, high CO 2 selectivities, in combination with attractive CO 2 adsorption capacities, low monomer cost, and the innate physicochemical stability of carbon render the materials of this study reasonable candidates for further practical consideration.

  16. Multi-task feature selection in microarray data by binary integer programming.

    PubMed

    Lan, Liang; Vucetic, Slobodan

    2013-12-20

    A major challenge in microarray classification is that the number of features is typically orders of magnitude larger than the number of examples. In this paper, we propose a novel feature filter algorithm to select the feature subset with maximal discriminative power and minimal redundancy by solving a quadratic objective function with binary integer constraints. To improve the computational efficiency, the binary integer constraints are relaxed and a low-rank approximation to the quadratic term is applied. The proposed feature selection algorithm was extended to solve multi-task microarray classification problems. We compared the single-task version of the proposed feature selection algorithm with 9 existing feature selection methods on 4 benchmark microarray data sets. The empirical results show that the proposed method achieved the most accurate predictions overall. We also evaluated the multi-task version of the proposed algorithm on 8 multi-task microarray datasets. The multi-task feature selection algorithm resulted in significantly higher accuracy than when using the single-task feature selection methods.

  17. Melanin fluorescence spectra by step-wise three photon excitation

    NASA Astrophysics Data System (ADS)

    Lai, Zhenhua; Kerimo, Josef; DiMarzio, Charles A.

    2012-03-01

    Melanin is the characteristic chromophore of human skin with various potential biological functions. Kerimo discovered enhanced melanin fluorescence by stepwise three-photon excitation in 2011. In this article, step-wise three-photon excited fluorescence (STPEF) spectrum between 450 nm -700 nm of melanin is reported. The melanin STPEF spectrum exhibited an exponential increase with wavelength. However, there was a probability of about 33% that another kind of step-wise multi-photon excited fluorescence (SMPEF) that peaks at 525 nm, shown by previous research, could also be generated using the same process. Using an excitation source at 920 nm as opposed to 830 nm increased the potential for generating SMPEF peaks at 525 nm. The SMPEF spectrum peaks at 525 nm photo-bleached faster than STPEF spectrum.

  18. Attentional Selection of Feature Conjunctions Is Accomplished by Parallel and Independent Selection of Single Features.

    PubMed

    Andersen, Søren K; Müller, Matthias M; Hillyard, Steven A

    2015-07-08

    Experiments that study feature-based attention have often examined situations in which selection is based on a single feature (e.g., the color red). However, in more complex situations relevant stimuli may not be set apart from other stimuli by a single defining property but by a specific combination of features. Here, we examined sustained attentional selection of stimuli defined by conjunctions of color and orientation. Human observers attended to one out of four concurrently presented superimposed fields of randomly moving horizontal or vertical bars of red or blue color to detect brief intervals of coherent motion. Selective stimulus processing in early visual cortex was assessed by recordings of steady-state visual evoked potentials (SSVEPs) elicited by each of the flickering fields of stimuli. We directly contrasted attentional selection of single features and feature conjunctions and found that SSVEP amplitudes on conditions in which selection was based on a single feature only (color or orientation) exactly predicted the magnitude of attentional enhancement of SSVEPs when attending to a conjunction of both features. Furthermore, enhanced SSVEP amplitudes elicited by attended stimuli were accompanied by equivalent reductions of SSVEP amplitudes elicited by unattended stimuli in all cases. We conclude that attentional selection of a feature-conjunction stimulus is accomplished by the parallel and independent facilitation of its constituent feature dimensions in early visual cortex. The ability to perceive the world is limited by the brain's processing capacity. Attention affords adaptive behavior by selectively prioritizing processing of relevant stimuli based on their features (location, color, orientation, etc.). We found that attentional mechanisms for selection of different features belonging to the same object operate independently and in parallel: concurrent attentional selection of two stimulus features is simply the sum of attending to each of those features separately. This result is key to understanding attentional selection in complex (natural) scenes, where relevant stimuli are likely to be defined by a combination of stimulus features. Copyright © 2015 the authors 0270-6474/15/359912-08$15.00/0.

  19. Collective feature selection to identify crucial epistatic variants.

    PubMed

    Verma, Shefali S; Lucas, Anastasia; Zhang, Xinyuan; Veturi, Yogasudha; Dudek, Scott; Li, Binglan; Li, Ruowang; Urbanowicz, Ryan; Moore, Jason H; Kim, Dokyoon; Ritchie, Marylyn D

    2018-01-01

    Machine learning methods have gained popularity and practicality in identifying linear and non-linear effects of variants associated with complex disease/traits. Detection of epistatic interactions still remains a challenge due to the large number of features and relatively small sample size as input, thus leading to the so-called "short fat data" problem. The efficiency of machine learning methods can be increased by limiting the number of input features. Thus, it is very important to perform variable selection before searching for epistasis. Many methods have been evaluated and proposed to perform feature selection, but no single method works best in all scenarios. We demonstrate this by conducting two separate simulation analyses to evaluate the proposed collective feature selection approach. Through our simulation study we propose a collective feature selection approach to select features that are in the "union" of the best performing methods. We explored various parametric, non-parametric, and data mining approaches to perform feature selection. We choose our top performing methods to select the union of the resulting variables based on a user-defined percentage of variants selected from each method to take to downstream analysis. Our simulation analysis shows that non-parametric data mining approaches, such as MDR, may work best under one simulation criteria for the high effect size (penetrance) datasets, while non-parametric methods designed for feature selection, such as Ranger and Gradient boosting, work best under other simulation criteria. Thus, using a collective approach proves to be more beneficial for selecting variables with epistatic effects also in low effect size datasets and different genetic architectures. Following this, we applied our proposed collective feature selection approach to select the top 1% of variables to identify potential interacting variables associated with Body Mass Index (BMI) in ~ 44,000 samples obtained from Geisinger's MyCode Community Health Initiative (on behalf of DiscovEHR collaboration). In this study, we were able to show that selecting variables using a collective feature selection approach could help in selecting true positive epistatic variables more frequently than applying any single method for feature selection via simulation studies. We were able to demonstrate the effectiveness of collective feature selection along with a comparison of many methods in our simulation analysis. We also applied our method to identify non-linear networks associated with obesity.

  20. AVC: Selecting discriminative features on basis of AUC by maximizing variable complementarity.

    PubMed

    Sun, Lei; Wang, Jun; Wei, Jinmao

    2017-03-14

    The Receiver Operator Characteristic (ROC) curve is well-known in evaluating classification performance in biomedical field. Owing to its superiority in dealing with imbalanced and cost-sensitive data, the ROC curve has been exploited as a popular metric to evaluate and find out disease-related genes (features). The existing ROC-based feature selection approaches are simple and effective in evaluating individual features. However, these approaches may fail to find real target feature subset due to their lack of effective means to reduce the redundancy between features, which is essential in machine learning. In this paper, we propose to assess feature complementarity by a trick of measuring the distances between the misclassified instances and their nearest misses on the dimensions of pairwise features. If a misclassified instance and its nearest miss on one feature dimension are far apart on another feature dimension, the two features are regarded as complementary to each other. Subsequently, we propose a novel filter feature selection approach on the basis of the ROC analysis. The new approach employs an efficient heuristic search strategy to select optimal features with highest complementarities. The experimental results on a broad range of microarray data sets validate that the classifiers built on the feature subset selected by our approach can get the minimal balanced error rate with a small amount of significant features. Compared with other ROC-based feature selection approaches, our new approach can select fewer features and effectively improve the classification performance.

  1. A Feature and Algorithm Selection Method for Improving the Prediction of Protein Structural Class.

    PubMed

    Ni, Qianwu; Chen, Lei

    2017-01-01

    Correct prediction of protein structural class is beneficial to investigation on protein functions, regulations and interactions. In recent years, several computational methods have been proposed in this regard. However, based on various features, it is still a great challenge to select proper classification algorithm and extract essential features to participate in classification. In this study, a feature and algorithm selection method was presented for improving the accuracy of protein structural class prediction. The amino acid compositions and physiochemical features were adopted to represent features and thirty-eight machine learning algorithms collected in Weka were employed. All features were first analyzed by a feature selection method, minimum redundancy maximum relevance (mRMR), producing a feature list. Then, several feature sets were constructed by adding features in the list one by one. For each feature set, thirtyeight algorithms were executed on a dataset, in which proteins were represented by features in the set. The predicted classes yielded by these algorithms and true class of each protein were collected to construct a dataset, which were analyzed by mRMR method, yielding an algorithm list. From the algorithm list, the algorithm was taken one by one to build an ensemble prediction model. Finally, we selected the ensemble prediction model with the best performance as the optimal ensemble prediction model. Experimental results indicate that the constructed model is much superior to models using single algorithm and other models that only adopt feature selection procedure or algorithm selection procedure. The feature selection procedure or algorithm selection procedure are really helpful for building an ensemble prediction model that can yield a better performance. Copyright© Bentham Science Publishers; For any queries, please email at epub@benthamscience.org.

  2. Non-negative matrix factorization in texture feature for classification of dementia with MRI data

    NASA Astrophysics Data System (ADS)

    Sarwinda, D.; Bustamam, A.; Ardaneswari, G.

    2017-07-01

    This paper investigates applications of non-negative matrix factorization as feature selection method to select the features from gray level co-occurrence matrix. The proposed approach is used to classify dementia using MRI data. In this study, texture analysis using gray level co-occurrence matrix is done to feature extraction. In the feature extraction process of MRI data, we found seven features from gray level co-occurrence matrix. Non-negative matrix factorization selected three features that influence of all features produced by feature extractions. A Naïve Bayes classifier is adapted to classify dementia, i.e. Alzheimer's disease, Mild Cognitive Impairment (MCI) and normal control. The experimental results show that non-negative factorization as feature selection method able to achieve an accuracy of 96.4% for classification of Alzheimer's and normal control. The proposed method also compared with other features selection methods i.e. Principal Component Analysis (PCA).

  3. Laboratory quality improvement in Tanzania.

    PubMed

    Andiric, Linda R; Massambu, Charles G

    2015-04-01

    The article describes the implementation and improvement in the first groups of medical laboratories in Tanzania selected to participate in the training program on Strengthening Laboratory Management Toward Accreditation (SLMTA). As in many other African nations, the selected improvement plan consisted of formalized hands-on training (SLMTA) that teaches the tasks and skills of laboratory management and provides the tools for implementation of best laboratory practice. Implementation of the improvements learned during training was verified before and after SLMTA with the World Health Organization African Region Stepwise Laboratory Improvement Process Towards Accreditation checklist. During a 4-year period, the selected laboratories described in this article demonstrated improvement with a range of 2% to 203% (cohort I) and 12% to 243% (cohort II) over baseline scores. The article describes the progress made in Tanzania's first cohorts, the obstacles encountered, and the lessons learned during the pilot and subsequent implementations. Copyright© by the American Society for Clinical Pathology.

  4. Comparison of Genetic Algorithm, Particle Swarm Optimization and Biogeography-based Optimization for Feature Selection to Classify Clusters of Microcalcifications

    NASA Astrophysics Data System (ADS)

    Khehra, Baljit Singh; Pharwaha, Amar Partap Singh

    2017-04-01

    Ductal carcinoma in situ (DCIS) is one type of breast cancer. Clusters of microcalcifications (MCCs) are symptoms of DCIS that are recognized by mammography. Selection of robust features vector is the process of selecting an optimal subset of features from a large number of available features in a given problem domain after the feature extraction and before any classification scheme. Feature selection reduces the feature space that improves the performance of classifier and decreases the computational burden imposed by using many features on classifier. Selection of an optimal subset of features from a large number of available features in a given problem domain is a difficult search problem. For n features, the total numbers of possible subsets of features are 2n. Thus, selection of an optimal subset of features problem belongs to the category of NP-hard problems. In this paper, an attempt is made to find the optimal subset of MCCs features from all possible subsets of features using genetic algorithm (GA), particle swarm optimization (PSO) and biogeography-based optimization (BBO). For simulation, a total of 380 benign and malignant MCCs samples have been selected from mammogram images of DDSM database. A total of 50 features extracted from benign and malignant MCCs samples are used in this study. In these algorithms, fitness function is correct classification rate of classifier. Support vector machine is used as a classifier. From experimental results, it is also observed that the performance of PSO-based and BBO-based algorithms to select an optimal subset of features for classifying MCCs as benign or malignant is better as compared to GA-based algorithm.

  5. Feature Selection for Classification of Polar Regions Using a Fuzzy Expert System

    NASA Technical Reports Server (NTRS)

    Penaloza, Mauel A.; Welch, Ronald M.

    1996-01-01

    Labeling, feature selection, and the choice of classifier are critical elements for classification of scenes and for image understanding. This study examines several methods for feature selection in polar regions, including the list, of a fuzzy logic-based expert system for further refinement of a set of selected features. Six Advanced Very High Resolution Radiometer (AVHRR) Local Area Coverage (LAC) arctic scenes are classified into nine classes: water, snow / ice, ice cloud, land, thin stratus, stratus over water, cumulus over water, textured snow over water, and snow-covered mountains. Sixty-seven spectral and textural features are computed and analyzed by the feature selection algorithms. The divergence, histogram analysis, and discriminant analysis approaches are intercompared for their effectiveness in feature selection. The fuzzy expert system method is used not only to determine the effectiveness of each approach in classifying polar scenes, but also to further reduce the features into a more optimal set. For each selection method,features are ranked from best to worst, and the best half of the features are selected. Then, rules using these selected features are defined. The results of running the fuzzy expert system with these rules show that the divergence method produces the best set features, not only does it produce the highest classification accuracy, but also it has the lowest computation requirements. A reduction of the set of features produced by the divergence method using the fuzzy expert system results in an overall classification accuracy of over 95 %. However, this increase of accuracy has a high computation cost.

  6. Selective catalytic two-step process for ethylene glycol from carbon monoxide

    PubMed Central

    Dong, Kaiwu; Elangovan, Saravanakumar; Sang, Rui; Spannenberg, Anke; Jackstell, Ralf; Junge, Kathrin; Li, Yuehui; Beller, Matthias

    2016-01-01

    Upgrading C1 chemicals (for example, CO, CO/H2, MeOH and CO2) with C–C bond formation is essential for the synthesis of bulk chemicals. In general, these industrially important processes (for example, Fischer Tropsch) proceed at drastic reaction conditions (>250 °C; high pressure) and suffer from low selectivity, which makes high capital investment necessary and requires additional purifications. Here, a different strategy for the preparation of ethylene glycol (EG) via initial oxidative coupling and subsequent reduction is presented. Separating coupling and reduction steps allows for a completely selective formation of EG (99%) from CO. This two-step catalytic procedure makes use of a Pd-catalysed oxycarbonylation of amines to oxamides at room temperature (RT) and subsequent Ru- or Fe-catalysed hydrogenation to EG. Notably, in the first step the required amines can be efficiently reused. The presented stepwise oxamide-mediated coupling provides the basis for a new strategy for selective upgrading of C1 chemicals. PMID:27377550

  7. Stepwise hydrolysis to improve carbon releasing efficiency from sludge.

    PubMed

    Liu, Hongbo; Wang, Yuanyuan; Wang, Ling; Yu, Tiantian; Fu, Bo; Liu, He

    2017-08-01

    Based on thermal alkaline hydrolysis (TAH), a novel strategy of stepwise hydrolysis was developed to improve carbon releasing efficiency from waste activated sludge (WAS). By stepwise increasing hydrolysis intensity, conventional sludge hydrolysis (the control) was divided into four stages for separately recovering sludge carbon sources with different bonding strengths, namely stage 1 (60 °C, pH 6.0-8.0), stage 2 (80 °C, pH 6.0-8.0), stage 3 (80 °C, pH 10.0) and stage 4 (90 °C, pH 12.0). Results indicate stepwise hydrolysis could enhance the amount of released soluble chemical oxygen demand (SCOD) for almost 2 times, from 7200 to 14,693 mg/L, and the released carbon presented better biodegradability, with BOD/COD of 0.47 and volatile fatty acids (VFAs) yield of 0.37 g VFAs/g SCOD via anaerobic fermentation. Moreover, stepwise hydrolysis also improved the dewaterability of hydrolyzed sludge, capillary suction time (CST) reducing from 2500 to 1600 s. Economic assessment indicates stepwise hydrolysis shows less alkali demand and lower thermal energy consumption than those of the control. Furthermore, results of this study help support the concepts of improving carbon recovery in wastewater by manipulating WAS composition and the idea of classifiably recovering the nutrients in WAS. Copyright © 2017 Elsevier Ltd. All rights reserved.

  8. Discrimination of Geographical Origin of Asian Garlic Using Isotopic and Chemical Datasets under Stepwise Principal Component Analysis.

    PubMed

    Liu, Tsang-Sen; Lin, Jhen-Nan; Peng, Tsung-Ren

    2018-01-16

    Isotopic compositions of δ 2 H, δ 18 O, δ 13 C, and δ 15 N and concentrations of 22 trace elements from garlic samples were analyzed and processed with stepwise principal component analysis (PCA) to discriminate garlic's country of origin among Asian regions including South Korea, Vietnam, Taiwan, and China. Results indicate that there is no single trace-element concentration or isotopic composition that can accomplish the study's purpose and the stepwise PCA approach proposed does allow for discrimination between countries on a regional basis. Sequentially, Step-1 PCA distinguishes garlic's country of origin among Taiwanese, South Korean, and Vietnamese samples; Step-2 PCA discriminates Chinese garlic from South Korean garlic; and Step-3 and Step-4 PCA, Chinese garlic from Vietnamese garlic. In model tests, countries of origin of all audit samples were correctly discriminated by stepwise PCA. Consequently, this study demonstrates that stepwise PCA as applied is a simple and effective approach to discriminating country of origin among Asian garlics. © 2018 American Academy of Forensic Sciences.

  9. Unbiased feature selection in learning random forests for high-dimensional data.

    PubMed

    Nguyen, Thanh-Tung; Huang, Joshua Zhexue; Nguyen, Thuy Thi

    2015-01-01

    Random forests (RFs) have been widely used as a powerful classification method. However, with the randomization in both bagging samples and feature selection, the trees in the forest tend to select uninformative features for node splitting. This makes RFs have poor accuracy when working with high-dimensional data. Besides that, RFs have bias in the feature selection process where multivalued features are favored. Aiming at debiasing feature selection in RFs, we propose a new RF algorithm, called xRF, to select good features in learning RFs for high-dimensional data. We first remove the uninformative features using p-value assessment, and the subset of unbiased features is then selected based on some statistical measures. This feature subset is then partitioned into two subsets. A feature weighting sampling technique is used to sample features from these two subsets for building trees. This approach enables one to generate more accurate trees, while allowing one to reduce dimensionality and the amount of data needed for learning RFs. An extensive set of experiments has been conducted on 47 high-dimensional real-world datasets including image datasets. The experimental results have shown that RFs with the proposed approach outperformed the existing random forests in increasing the accuracy and the AUC measures.

  10. Texture analysis of tissues in Gleason grading of prostate cancer

    NASA Astrophysics Data System (ADS)

    Alexandratou, Eleni; Yova, Dido; Gorpas, Dimitris; Maragos, Petros; Agrogiannis, George; Kavantzas, Nikolaos

    2008-02-01

    Prostate cancer is a common malignancy among maturing men and the second leading cause of cancer death in USA. Histopathological grading of prostate cancer is based on tissue structural abnormalities. Gleason grading system is the gold standard and is based on the organization features of prostatic glands. Although Gleason score has contributed on cancer prognosis and on treatment planning, its accuracy is about 58%, with this percentage to be lower in GG2, GG3 and GG5 grading. On the other hand it is strongly affected by "inter- and intra observer variations", making the whole process very subjective. Therefore, there is need for the development of grading tools based on imaging and computer vision techniques for a more accurate prostate cancer prognosis. The aim of this paper is the development of a novel method for objective grading of biopsy specimen in order to support histopathological prognosis of the tumor. This new method is based on texture analysis techniques, and particularly on Gray Level Co-occurrence Matrix (GLCM) that estimates image properties related to second order statistics. Histopathological images of prostate cancer, from Gleason grade2 to Gleason grade 5, were acquired and subjected to image texture analysis. Thirteen texture characteristics were calculated from this matrix as they were proposed by Haralick. Using stepwise variable selection, a subset of four characteristics were selected and used for the description and classification of each image field. The selected characteristics profile was used for grading the specimen with the multiparameter statistical method of multiple logistic discrimination analysis. The subset of these characteristics provided 87% correct grading of the specimens. The addition of any of the remaining characteristics did not improve significantly the diagnostic ability of the method. This study demonstrated that texture analysis techniques could provide valuable grading decision support to the pathologists, concerning prostate cancer prognosis.

  11. Quantitative structure activity relationship studies of sulfamide derivatives as carbonic anhydrase inhibitor: as antiglaucoma agents.

    PubMed

    Kumar, Surendra; Singh, Vineet; Tiwari, Meena

    2007-07-01

    Selective inhibition of ciliary process enzyme i.e. Carbonic Anhydrase-II is an excellent approach in reducing elevated intraocular pressure, thus treating glaucoma. Due to characteristic physicochemical properties of sulphonamide (Inhibition of Carbonic Anhydrase), they are clinically effective against glaucoma. But the non-specificity of sulphonamide derivatives to isozyme, leads to a range of side effects. Presently, the absence of comparative studies related to the binding of the sulphonamides as inhibitors to CA isozymes limits their use. In this paper we have represented "Three Dimensional Quantitative Structure Activity Relationship" study to characterize structural features of Sulfamide derivative [RR'NSO(2)NH(2)] as inhibitors, that are required for selective binding of carbonic anhydrase isozymes (CAI and CAII). In the analysis, stepwise multiple linear regression was performed using physiochemical parameters as independent variable and CA-I and CA-II inhibitory activity as dependent variable, respectively. The best multiparametric QSAR model obtained for CA-I inhibitory activity shows good statistical significance (r= 0.9714) and predictability (Q(2)=0.8921), involving the Electronic descriptors viz. Highest Occupied Molecular Orbital, Lowest Unoccupied Molecular Orbital and Steric descriptors viz. Principal moment of Inertia at X axis. Similarly, CA-II inhibitory activity also shows good statistical significance (r=0.9644) and predictability (Q(2)=0.8699) involving aforementioned descriptors. The predictive power of the model was successfully tested externally using a set of six compounds as test set for CA-I inhibitory activity and a set of seven compounds in case of CA-II inhibitory activity with good predictive squared correlation coefficient, r(2)(pred)=0.6016 and 0.7662, respectively. Overview of analysis favours substituents with high electronegativity and less bulk at R and R' positions of the parent nucleus, provides a basis to design new Sulfamide derivatives possessing potent and selective carbonic anhydrase-II inhibitory activity.

  12. Train axle bearing fault detection using a feature selection scheme based multi-scale morphological filter

    NASA Astrophysics Data System (ADS)

    Li, Yifan; Liang, Xihui; Lin, Jianhui; Chen, Yuejian; Liu, Jianxin

    2018-02-01

    This paper presents a novel signal processing scheme, feature selection based multi-scale morphological filter (MMF), for train axle bearing fault detection. In this scheme, more than 30 feature indicators of vibration signals are calculated for axle bearings with different conditions and the features which can reflect fault characteristics more effectively and representatively are selected using the max-relevance and min-redundancy principle. Then, a filtering scale selection approach for MMF based on feature selection and grey relational analysis is proposed. The feature selection based MMF method is tested on diagnosis of artificially created damages of rolling bearings of railway trains. Experimental results show that the proposed method has a superior performance in extracting fault features of defective train axle bearings. In addition, comparisons are performed with the kurtosis criterion based MMF and the spectral kurtosis criterion based MMF. The proposed feature selection based MMF method outperforms these two methods in detection of train axle bearing faults.

  13. Surgery-Independent Language Function Decline in Patients Undergoing Awake Craniotomy.

    PubMed

    Gonen, Tal; Sela, Gal; Yanakee, Ranin; Ram, Zvi; Grossman, Rachel

    2017-03-01

    Despite selection process before awake-craniotomy, some patients experience an unexpected decline in language functions in the operating room (OR), compared with their baseline evaluation, which may impair their functional monitoring. To investigate this phenomenon we prospectively compared language function the day before surgery and on entrance to the OR. Data were collected prospectively from consecutive patients undergoing awake-craniotomy with intraoperative cortical mapping for resection of gliomas affecting language areas. Language functions of 79 patients were evaluated and compared 1-2 days before surgery and after entering the OR. Changes in functional linguistic performance were analyzed with respect to demographic, clinical, and pathologic characteristics. There was a significant decline in language function, beyond sedation effect, after entering the OR, (from median/interquartile range: 0.94/0.72-0.98 to median/interquartile range: 0.86/0.51-0.94; Z = -7.19, P < 0.001). Univariate analyses revealed that this decline was related to age, preoperative Karnofsky Performance Scale, tumor location, tumor pathology, and preexisting language deficits. Multivariate stepwise regression identified tumor pathology and the presence of preoperative language deficit as significant independent predictors for this functional decline. Patients undergoing awake-craniotomy may experience a substantial decline in language functioning after entering the OR. Tumor grade and the presence of preoperative language deficits were significant risk factors for this phenomenon, suggesting a possible relation between cognitive reserve, psychobehavioral coping abilities and histologic features of a tumor involving language areas. Capturing and identifying this unique population of patients who are prone to experience such language decline may improve our ability in the future to select patients eligible for awake-craniotomy. Copyright © 2016 Elsevier Inc. All rights reserved.

  14. Feature extraction and selection from volatile compounds for analytical classification of Chinese red wines from different varieties.

    PubMed

    Zhang, Jian; Li, Li; Gao, Nianfa; Wang, Depei; Gao, Qiang; Jiang, Shengping

    2010-03-10

    This work was undertaken to evaluate whether it is possible to determine the variety of a Chinese wine on the basis of its volatile compounds, and to investigate if discrimination models could be developed with the experimental wines that could be used for the commercial ones. A headspace solid-phase microextraction gas chromatographic (HS-SPME-GC) procedure was used to determine the volatile compounds and a blind analysis based on Ac/Ais (peak area of volatile compound/peak area of internal standard) was carried out for statistical purposes. One way analysis of variance (ANOVA), principal component analysis (PCA) and stepwise linear discriminant analysis (SLDA) were used to process data and to develop discriminant models. Only 11 peaks enabled to differentiate and classify the experimental wines. SLDA allowed 100% recognition ability for three grape varieties, 100% prediction ability for Cabernet Sauvignon and Cabernet Gernischt wines, but only 92.31% for Merlot wines. A more valid and robust way was to use the PCA scores to do the discriminant analysis. When we performed SLDA this way, 100% recognition ability and 100% prediction ability were obtained. At last, 11 peaks which selected by SLDA from raw analysis set had been identified. When we demonstrated the models using commercial wines, the models showed 100% recognition ability for the wines collected directly from winery and without ageing, but only 65% for the others. Therefore, the varietal factor was currently discredited as a differentiating parameter for commercial wines in China. Nevertheless, this method could be applied as a screening tool and as a complement to other methods for grape base liquors which do not need ageing and blending procedures. 2010 Elsevier B.V. All rights reserved.

  15. Genetic Particle Swarm Optimization-Based Feature Selection for Very-High-Resolution Remotely Sensed Imagery Object Change Detection.

    PubMed

    Chen, Qiang; Chen, Yunhao; Jiang, Weiguo

    2016-07-30

    In the field of multiple features Object-Based Change Detection (OBCD) for very-high-resolution remotely sensed images, image objects have abundant features and feature selection affects the precision and efficiency of OBCD. Through object-based image analysis, this paper proposes a Genetic Particle Swarm Optimization (GPSO)-based feature selection algorithm to solve the optimization problem of feature selection in multiple features OBCD. We select the Ratio of Mean to Variance (RMV) as the fitness function of GPSO, and apply the proposed algorithm to the object-based hybrid multivariate alternative detection model. Two experiment cases on Worldview-2/3 images confirm that GPSO can significantly improve the speed of convergence, and effectively avoid the problem of premature convergence, relative to other feature selection algorithms. According to the accuracy evaluation of OBCD, GPSO is superior at overall accuracy (84.17% and 83.59%) and Kappa coefficient (0.6771 and 0.6314) than other algorithms. Moreover, the sensitivity analysis results show that the proposed algorithm is not easily influenced by the initial parameters, but the number of features to be selected and the size of the particle swarm would affect the algorithm. The comparison experiment results reveal that RMV is more suitable than other functions as the fitness function of GPSO-based feature selection algorithm.

  16. Genetic Programming and Frequent Itemset Mining to Identify Feature Selection Patterns of iEEG and fMRI Epilepsy Data

    PubMed Central

    Smart, Otis; Burrell, Lauren

    2014-01-01

    Pattern classification for intracranial electroencephalogram (iEEG) and functional magnetic resonance imaging (fMRI) signals has furthered epilepsy research toward understanding the origin of epileptic seizures and localizing dysfunctional brain tissue for treatment. Prior research has demonstrated that implicitly selecting features with a genetic programming (GP) algorithm more effectively determined the proper features to discern biomarker and non-biomarker interictal iEEG and fMRI activity than conventional feature selection approaches. However for each the iEEG and fMRI modalities, it is still uncertain whether the stochastic properties of indirect feature selection with a GP yield (a) consistent results within a patient data set and (b) features that are specific or universal across multiple patient data sets. We examined the reproducibility of implicitly selecting features to classify interictal activity using a GP algorithm by performing several selection trials and subsequent frequent itemset mining (FIM) for separate iEEG and fMRI epilepsy patient data. We observed within-subject consistency and across-subject variability with some small similarity for selected features, indicating a clear need for patient-specific features and possible need for patient-specific feature selection or/and classification. For the fMRI, using nearest-neighbor classification and 30 GP generations, we obtained over 60% median sensitivity and over 60% median selectivity. For the iEEG, using nearest-neighbor classification and 30 GP generations, we obtained over 65% median sensitivity and over 65% median selectivity except one patient. PMID:25580059

  17. Method of generating features optimal to a dataset and classifier

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Bruillard, Paul J.; Gosink, Luke J.; Jarman, Kenneth D.

    A method of generating features optimal to a particular dataset and classifier is disclosed. A dataset of messages is inputted and a classifier is selected. An algebra of features is encoded. Computable features that are capable of describing the dataset from the algebra of features are selected. Irredundant features that are optimal for the classifier and the dataset are selected.

  18. Temporal Correlation Mechanisms and Their Role in Feature Selection: A Single-Unit Study in Primate Somatosensory Cortex

    PubMed Central

    Gomez-Ramirez, Manuel; Trzcinski, Natalie K.; Mihalas, Stefan; Niebur, Ernst

    2014-01-01

    Studies in vision show that attention enhances the firing rates of cells when it is directed towards their preferred stimulus feature. However, it is unknown whether other sensory systems employ this mechanism to mediate feature selection within their modalities. Moreover, whether feature-based attention modulates the correlated activity of a population is unclear. Indeed, temporal correlation codes such as spike-synchrony and spike-count correlations (rsc) are believed to play a role in stimulus selection by increasing the signal and reducing the noise in a population, respectively. Here, we investigate (1) whether feature-based attention biases the correlated activity between neurons when attention is directed towards their common preferred feature, (2) the interplay between spike-synchrony and rsc during feature selection, and (3) whether feature attention effects are common across the visual and tactile systems. Single-unit recordings were made in secondary somatosensory cortex of three non-human primates while animals engaged in tactile feature (orientation and frequency) and visual discrimination tasks. We found that both firing rate and spike-synchrony between neurons with similar feature selectivity were enhanced when attention was directed towards their preferred feature. However, attention effects on spike-synchrony were twice as large as those on firing rate, and had a tighter relationship with behavioral performance. Further, we observed increased rsc when attention was directed towards the visual modality (i.e., away from touch). These data suggest that similar feature selection mechanisms are employed in vision and touch, and that temporal correlation codes such as spike-synchrony play a role in mediating feature selection. We posit that feature-based selection operates by implementing multiple mechanisms that reduce the overall noise levels in the neural population and synchronize activity across subpopulations that encode the relevant features of sensory stimuli. PMID:25423284

  19. Feature Selection Methods for Zero-Shot Learning of Neural Activity.

    PubMed

    Caceres, Carlos A; Roos, Matthew J; Rupp, Kyle M; Milsap, Griffin; Crone, Nathan E; Wolmetz, Michael E; Ratto, Christopher R

    2017-01-01

    Dimensionality poses a serious challenge when making predictions from human neuroimaging data. Across imaging modalities, large pools of potential neural features (e.g., responses from particular voxels, electrodes, and temporal windows) have to be related to typically limited sets of stimuli and samples. In recent years, zero-shot prediction models have been introduced for mapping between neural signals and semantic attributes, which allows for classification of stimulus classes not explicitly included in the training set. While choices about feature selection can have a substantial impact when closed-set accuracy, open-set robustness, and runtime are competing design objectives, no systematic study of feature selection for these models has been reported. Instead, a relatively straightforward feature stability approach has been adopted and successfully applied across models and imaging modalities. To characterize the tradeoffs in feature selection for zero-shot learning, we compared correlation-based stability to several other feature selection techniques on comparable data sets from two distinct imaging modalities: functional Magnetic Resonance Imaging and Electrocorticography. While most of the feature selection methods resulted in similar zero-shot prediction accuracies and spatial/spectral patterns of selected features, there was one exception; A novel feature/attribute correlation approach was able to achieve those accuracies with far fewer features, suggesting the potential for simpler prediction models that yield high zero-shot classification accuracy.

  20. Modelling the evolution and diversity of cumulative culture

    PubMed Central

    Enquist, Magnus; Ghirlanda, Stefano; Eriksson, Kimmo

    2011-01-01

    Previous work on mathematical models of cultural evolution has mainly focused on the diffusion of simple cultural elements. However, a characteristic feature of human cultural evolution is the seemingly limitless appearance of new and increasingly complex cultural elements. Here, we develop a general modelling framework to study such cumulative processes, in which we assume that the appearance and disappearance of cultural elements are stochastic events that depend on the current state of culture. Five scenarios are explored: evolution of independent cultural elements, stepwise modification of elements, differentiation or combination of elements and systems of cultural elements. As one application of our framework, we study the evolution of cultural diversity (in time as well as between groups). PMID:21199845

  1. Size effects influence on conducting properties of Cu-Nb alloy microcomposites at cryogenic temperature

    NASA Astrophysics Data System (ADS)

    Guryev, Valentin V.; Polikarpova, Maria V.; Lukyanov, Pavel A.; Khlebova, Natalya E.; Pantsyrny, Viktor I.

    2018-03-01

    A comprehensive study has been carried out in relation to the conductivity of heavily deformed Cu-16wt%Nb nanostructured wires at room and cryogenic temperatures. When the true strain exceeds 5, the growth rates of the resistivity qualitatively change the behavior at all temperatures. It is shown that such behavior is defined mostly by interface scattering. At 10 K the stepwise increasing of resistivity has been found, what is speculated as a feature of amorphous regions formation at the interface of Cu/Nb. Simultaneously the superconducting transition temperature (Tcs) falls down due to proximity effect. The deviation of experimental Tcs values from predicted by classical model is discussed.

  2. In-line metrology for roll-to-roll UV assisted nanoimprint lithography using diffractometry

    NASA Astrophysics Data System (ADS)

    Kreuzer, Martin; Whitworth, Guy L.; Francone, Achille; Gomis-Bresco, Jordi; Kehagias, Nikolaos; Sotomayor-Torres, Clivia M.

    2018-05-01

    We describe and discuss the optical design of a diffractometer to carry out in-line quality control during roll-to-roll nanoimprinting. The tool measures diffractograms in reflection geometry, through an aspheric lens to gain fast, non-invasive information of any changes to the critical dimensions of target grating structures. A stepwise tapered linear grating with constant period was fabricated in order to detect the variation in grating linewidth through diffractometry. The minimum feature change detected was ˜40 nm to a precision of 10 nm. The diffractometer was then integrated with a roll-to-roll UV assisted nanoimprint lithography machine to gain dynamic measurements in situ.

  3. A novel structural tree for wrap-proteins, a subclass of (α+β)-proteins.

    PubMed

    Boshkova, Eugenia A; Gordeev, Alexey B; Efimov, Alexander V

    2014-01-01

    In this paper, a novel structural subclass of (α+β)-proteins is presented. A characteristic feature of these proteins and domains is that they consist of strongly twisted and coiled β-sheets wrapped around one or two α-helices, so they are referred to here as wrap-proteins. It is shown that overall folds of the wrap-proteins can be obtained by stepwise addition of α-helices and/or β-strands to the strongly twisted and coiled β-hairpin taken as the starting structure in modeling. As a result of modeling, a structural tree for the wrap-proteins was constructed that includes 201 folds of which 49 occur in known nonhomologous proteins.

  4. Insulin degludec/insulin aspart once daily in Type 2 diabetes: a comparison of simple or stepwise titration algorithms (BOOST® : SIMPLE USE).

    PubMed

    Park, S W; Bebakar, W M W; Hernandez, P G; Macura, S; Hersløv, M L; de la Rosa, R

    2017-02-01

    To compare the efficacy and safety of two titration algorithms for insulin degludec/insulin aspart (IDegAsp) administered once daily with metformin in participants with insulin-naïve Type 2 diabetes mellitus. This open-label, parallel-group, 26-week, multicentre, treat-to-target trial, randomly allocated participants (1:1) to two titration arms. The Simple algorithm titrated IDegAsp twice weekly based on a single pre-breakfast self-monitored plasma glucose (SMPG) measurement. The Stepwise algorithm titrated IDegAsp once weekly based on the lowest of three consecutive pre-breakfast SMPG measurements. In both groups, IDegAsp once daily was titrated to pre-breakfast plasma glucose values of 4.0-5.0 mmol/l. Primary endpoint was change from baseline in HbA 1c (%) after 26 weeks. Change in HbA 1c at Week 26 was IDegAsp Simple -14.6 mmol/mol (-1.3%) (to 52.4 mmol/mol; 6.9%) and IDegAsp Stepwise -11.9 mmol/mol (-1.1%) (to 54.7 mmol/mol; 7.2%). The estimated between-group treatment difference was -1.97 mmol/mol [95% confidence interval (CI) -4.1, 0.2] (-0.2%, 95% CI -0.4, 0.02), confirming the non-inferiority of IDegAsp Simple to IDegAsp Stepwise (non-inferiority limit of ≤ 0.4%). Mean reduction in fasting plasma glucose and 8-point SMPG profiles were similar between groups. Rates of confirmed hypoglycaemia were lower for IDegAsp Stepwise [2.1 per patient years of exposure (PYE)] vs. IDegAsp Simple (3.3 PYE) (estimated rate ratio IDegAsp Simple /IDegAsp Stepwise 1.8; 95% CI 1.1, 2.9). Nocturnal hypoglycaemia rates were similar between groups. No severe hypoglycaemic events were reported. In participants with insulin-naïve Type 2 diabetes mellitus, the IDegAsp Simple titration algorithm improved HbA 1c levels as effectively as a Stepwise titration algorithm. Hypoglycaemia rates were lower in the Stepwise arm. © 2016 The Authors. Diabetic Medicine published by John Wiley & Sons Ltd on behalf of Diabetes UK.

  5. Specificity in endoplasmic reticulum-stress signaling in yeast entails a step-wise engagement of HAC1 mRNA to clusters of the stress sensor Ire1.

    PubMed

    van Anken, Eelco; Pincus, David; Coyle, Scott; Aragón, Tomás; Osman, Christof; Lari, Federica; Gómez Puerta, Silvia; Korennykh, Alexei V; Walter, Peter

    2014-12-30

    Insufficient protein-folding capacity in the endoplasmic reticulum (ER) induces the unfolded protein response (UPR). In the ER lumen, accumulation of unfolded proteins activates the transmembrane ER-stress sensor Ire1 and drives its oligomerization. In the cytosol, Ire1 recruits HAC1 mRNA, mediating its non-conventional splicing. The spliced mRNA is translated into Hac1, the key transcription activator of UPR target genes that mitigate ER-stress. In this study, we report that oligomeric assembly of the ER-lumenal domain is sufficient to drive Ire1 clustering. Clustering facilitates Ire1's cytosolic oligomeric assembly and HAC1 mRNA docking onto a positively charged motif in Ire1's cytosolic linker domain that tethers the kinase/RNase to the transmembrane domain. By the use of a synthetic bypass, we demonstrate that mRNA docking per se is a pre-requisite for initiating Ire1's RNase activity and, hence, splicing. We posit that such step-wise engagement between Ire1 and its mRNA substrate contributes to selectivity and efficiency in UPR signaling.

  6. Enhancement of In Situ Enzymatic Saccharification of Corn Stover by a Stepwise Sodium Hydroxide and Organic Acid Pretreatment.

    PubMed

    Qing, Qing; Guo, Qi; Zhou, Linlin; He, Yucai; Wang, Liqun; Zhang, Yue

    2017-01-01

    A stepwise pretreatment method that combines sodium hydroxide and organic acid pretreatments was proposed and investigated to maximize the recovery of main constituents of lignocellulose. The sodium hydroxide pretreatment was firstly optimized by a designed orthogonal experiment with the optimum pretreatment conditions determined as 1 wt% NaOH at 70 °C for 1 h, and 60.42 % of lignin was successfully removed during this stage. In the second stage, 0.5 % acetic acid was selected to pretreat the first-stage solid residue at 80 °C for 40 min in order to decompose hemicelluloses to soluble oligomers or monomers. Then, the whole slurry was subjected to in situ enzymatic saccharification by cellullase with a supplementation of xylanase to further degrade the xylooligosaccharides generated during the acetic acid pretreatment. The maximum reducing sugar and glucose yields achieved were 20.74 and 12.03 g/L, respectively. Furthermore, rapid ethanol fermentation and a yield of 80.3 % also testified this pretreatment method, and the in situ saccharification did not bring any negative impact on ethanol fermentation and has a broad application prospect.

  7. Computational Studies on the Synthesis of β-Lactams via [2+2] Thermal Cycloadditions

    NASA Astrophysics Data System (ADS)

    Arrieta, Ana; Lecea, Begoña; Cossío, Fernando P.

    The main computational studies on the formation of β-lactams through [2+2] cycloadditions published during 1992-2008 are reported with special emphasis on the mechanistic and selectivity aspects of these reactions. Disconnection of the N1-C2 and C3-C4 bonds of the azetidin-2-one ring leads to the reaction between ketenes and imines. Computational and experimental results point to a stepwise mechanism for this reaction. The first step consists of a nucleophilic attack of the iminic nitrogen on the sp-hybridized carbon atom of the ketene. The zwitterionic intermediate thus formed yields the corresponding β-lactam by means of a four-electron conrotatoty electrocyclization. The steroecontrol and the periselectivity of the reaction support this two-step mechanism. The [2+2] cycloaddition between isocyanates and alkenes takes place via a concerted (but asynchronous) mechanism that can be interpreted in terms of a [π2s + (π2s + π2s)] interaction between both reactants. Both the regio and the stereochemistry observed are compatible with this computational model. However, the combination of solvent and substituent effects can result in a stepwise mechanism.

  8. Retro-regression--another important multivariate regression improvement.

    PubMed

    Randić, M

    2001-01-01

    We review the serious problem associated with instabilities of the coefficients of regression equations, referred to as the MRA (multivariate regression analysis) "nightmare of the first kind". This is manifested when in a stepwise regression a descriptor is included or excluded from a regression. The consequence is an unpredictable change of the coefficients of the descriptors that remain in the regression equation. We follow with consideration of an even more serious problem, referred to as the MRA "nightmare of the second kind", arising when optimal descriptors are selected from a large pool of descriptors. This process typically causes at different steps of the stepwise regression a replacement of several previously used descriptors by new ones. We describe a procedure that resolves these difficulties. The approach is illustrated on boiling points of nonanes which are considered (1) by using an ordered connectivity basis; (2) by using an ordering resulting from application of greedy algorithm; and (3) by using an ordering derived from an exhaustive search for optimal descriptors. A novel variant of multiple regression analysis, called retro-regression (RR), is outlined showing how it resolves the ambiguities associated with both "nightmares" of the first and the second kind of MRA.

  9. Preparative isolation of alkaloids from Corydalis bungeana Turcz. by high-speed counter-current chromatography using stepwise elution

    PubMed Central

    Niu, Lili; Xie, Zhensheng; Cai, Tanxi; Wu, Peng; Xue, Peng; Chen, Xiulan; Wu, Zhiyong; Ito, Yoichiro; Li, Famei; Yang, Fuquan

    2011-01-01

    High-speed counter-current chromatography (HSCCC) was successfully applied for the preparative separation and purification of alkaloids from Corydalis bungeana Turcz. (Kudiding in Chinese) for the first time. After the measurement of partition coefficient of seven target alkaloids in the nine two-phase solvent systems composed of CHCl3–MeOH–(0.1 M; 0.2 M; 0.3 M) HCl (4:1.5:2; 4:2:2; 4:3:2, v/v), CHCl3–MeOH–0.2 M HCl (4:2:2, v/v) and CHCl3–MeOH–0.3 M HCl (4:3:2, v/v) were finally selected for the HSCCC separation using the first upper phase as the stationary phase and the stepwise elution of the two lower mobile phases. Consequently, sanguinarine (10 mg), corynoline (25 mg), protopine (20 mg), corynoloxine (18 mg), and 12-hydroxycorynoline (8 mg) were obtained from 200 mg of crude alkaloid extracts with purities of 94–99% as determined by HPLC. Their chemical structures were characterized on the basis of 1H-NMR, 13C-NMR, and LC-ESI-Q-TOF-MS/MS analyses. PMID:21387560

  10. Classification of lung cancer patients and controls by chromatography of modified nucleosides in serum

    USGS Publications Warehouse

    McEntire, John E.; Kuo, Kenneth C.; Smith, Mark E.; Stalling, David L.; Richens, Jack W.; Zumwalt, Robert W.; Gehrke, Charles W.; Papermaster, Ben W.

    1989-01-01

    A wide spectrum of modified nucleosides has been quantified by high-performance liquid chromatography in serum of 49 male lung cancer patients, 35 patients with other cancers, and 48 patients hospitalized for nonneoplastic diseases. Data for 29 modified nucleoside peaks were normalized to an internal standard and analyzed by discriminant analysis and stepwise discriminant analysis. A model based on peaks selected by a stepwise discriminant procedure correctly classified 79% of the cancer and 75% of the noncancer subjects. It also demonstrated 84% sensitivity and 79% specificity when comparing lung cancer to noncancer subjects, and 80% sensitivity and 55% specificity in comparing lung cancer to other cancers. The nucleoside peaks having the greatest influence on the models varied dependent on the subgroups compared, confirming the importance of quantifying a wide array of nucleosides. These data support and expand previous studies which reported the utility of measuring modified nucleoside levels in serum and show that precise measurement of an array of 29 modified nucleosides in serum by high-performance liquid chromatography with UV scanning with subsequent data modeling may provide a clinically useful approach to patient classification in diagnosis and subsequent therapeutic monitoring.

  11. A tunable algorithm for collective decision-making.

    PubMed

    Pratt, Stephen C; Sumpter, David J T

    2006-10-24

    Complex biological systems are increasingly understood in terms of the algorithms that guide the behavior of system components and the information pathways that link them. Much attention has been given to robust algorithms, or those that allow a system to maintain its functions in the face of internal or external perturbations. At the same time, environmental variation imposes a complementary need for algorithm versatility, or the ability to alter system function adaptively as external circumstances change. An important goal of systems biology is thus the identification of biological algorithms that can meet multiple challenges rather than being narrowly specified to particular problems. Here we show that emigrating colonies of the ant Temnothorax curvispinosus tune the parameters of a single decision algorithm to respond adaptively to two distinct problems: rapid abandonment of their old nest in a crisis and deliberative selection of the best available new home when their old nest is still intact. The algorithm uses a stepwise commitment scheme and a quorum rule to integrate information gathered by numerous individual ants visiting several candidate homes. By varying the rates at which they search for and accept these candidates, the ants yield a colony-level response that adaptively emphasizes either speed or accuracy. We propose such general but tunable algorithms as a design feature of complex systems, each algorithm providing elegant solutions to a wide range of problems.

  12. Utilization of matrix-assisted laser desorption/ionization time-of-flight mass spectrometry for structural studies related to biology and disease

    NASA Astrophysics Data System (ADS)

    Costello, Catherine E.; Helin, Jari; Ngoka, Lambert C. M.

    1996-04-01

    Matrix-assisted laser desorption/ionization time-of-flight mass spectrometry (MALDI-TOF MS), because of its high sensitivity and relatively straightforward requirements for sample preparation, is contributing to the solution of structural problems in biology and to the development of therapeutic approaches through increased understanding of pharmacology and enhanced capabilities for quality control of pharmaceuticals. We are using a reflectron TOF- MS for the determination of molecular weights of individual compounds and the components of mixtures that are naturally occurring or are generated through enzymic digests, and employing the post-source decay mode to elucidate structural details. To maximize the sensitivity and information content of the spectra, varied matrices, derivative, and stepwise degradation procedures are being explored. Present studies include investigations of oligosaccharides, neutral glycolipids, gangliosides, glycoproteins, neuropeptides and proteins. Rules for fragmentation are being developed with model compounds and used for the structural elucidation of unknowns. When adequate sample amounts are available, the results are compared with low- and high-energy collision-induced decomposition spectra obtained with tandem MS in order to provide a data base for the correlation of spectral features and guidance in selection of approaches for scarce biological samples. Current projects include biophysical studies of glycoplipids, glycoproteins and oligosaccharides and investigations of the substance P receptor, transthyretin genetic variants and cisplatin-DNA interactions.

  13. Leaf Photosynthetic Parameters Related to Biomass Accumulation in a Global Rice Diversity Survey1[OPEN

    PubMed Central

    Zheng, Guangyong; Hamdani, Saber; Essemine, Jemaa; Song, Qingfeng; Wang, Hongru

    2017-01-01

    Mining natural variations is a major approach to identify new options to improve crop light use efficiency. So far, successes in identifying photosynthetic parameters positively related to crop biomass accumulation through this approach are scarce, possibly due to the earlier emphasis on properties related to leaf instead of canopy photosynthetic efficiency. This study aims to uncover rice (Oryza sativa) natural variations to identify leaf physiological parameters that are highly correlated with biomass accumulation, a surrogate of canopy photosynthesis. To do this, we systematically investigated 14 photosynthetic parameters and four morphological traits in a rice population, which consists of 204 U.S. Department of Agriculture-curated minicore accessions collected globally and 11 elite Chinese rice cultivars in both Beijing and Shanghai. To identify key components responsible for the variance of biomass accumulation, we applied a stepwise feature-selection approach based on linear regression models. Although there are large variations in photosynthetic parameters measured in different environments, we observed that photosynthetic rate under low light (Alow) was highly related to biomass accumulation and also exhibited high genomic inheritability in both environments, suggesting its great potential to be used as a target for future rice breeding programs. Large variations in Alow among modern rice cultivars further suggest the great potential of using this parameter in contemporary rice breeding for the improvement of biomass and, hence, yield potential. PMID:28739819

  14. A conceptual framework for the collection of food products in a Total Diet Study.

    PubMed

    Turrini, Aida; Lombardi-Boccia, Ginevra; Aureli, Federica; Cubadda, Francesco; D'Addezio, Laura; D'Amato, Marilena; D'Evoli, Laura; Darnerud, PerOla; Devlin, Niamh; Dias, Maria Graça; Jurković, Marina; Kelleher, Cecily; Le Donne, Cinzia; López Esteban, Maite; Lucarini, Massimo; Martinez Burgos, Maria Alba; Martínez-Victoria, Emilio; McNulty, Breige; Mistura, Lorenza; Nugent, Anne; Oktay Basegmez, Hatice Imge; Oliveira, Luisa; Ozer, Hayrettin; Perelló, Gemma; Pite, Marina; Presser, Karl; Sokolić, Darja; Vasco, Elsa; Volatier, Jean-Luc

    2018-02-01

    A total diet study (TDS) provides representative and realistic data for assessing the dietary intake of chemicals, such as contaminants and residues, and nutrients, at a population level. Reproducing the diet through collection of customarily consumed foods and their preparation as habitually eaten is crucial to ensure representativeness, i.e., all relevant foods are included and all potential dietary sources of the substances investigated are captured. Having this in mind, a conceptual framework for building a relevant food-shopping list was developed as a research task in the European Union's 7th Framework Program project, 'Total Diet Study Exposure' (TDS-Exposure), aimed at standardising methods for food sampling, analyses, exposure assessment calculations and modelling, priority foods, and selection of chemical contaminants. A stepwise approach following the knowledge translation (KT) model for concept analysis is proposed to set up a general protocol for the collection of food products in a TDS in terms of steps (characterisation of the food list, development of the food-shopping list, food products collection) and pillars (background documentation, procedures, and tools). A simple model for structuring the information in a way to support the implementation of the process, by presenting relevant datasets, forms to store inherent information, and folders to record the results is also proposed. Reproducibility of the process and possibility to exploit the gathered information are two main features of such a system for future applications.

  15. A new computational strategy for predicting essential genes.

    PubMed

    Cheng, Jian; Wu, Wenwu; Zhang, Yinwen; Li, Xiangchen; Jiang, Xiaoqian; Wei, Gehong; Tao, Shiheng

    2013-12-21

    Determination of the minimum gene set for cellular life is one of the central goals in biology. Genome-wide essential gene identification has progressed rapidly in certain bacterial species; however, it remains difficult to achieve in most eukaryotic species. Several computational models have recently been developed to integrate gene features and used as alternatives to transfer gene essentiality annotations between organisms. We first collected features that were widely used by previous predictive models and assessed the relationships between gene features and gene essentiality using a stepwise regression model. We found two issues that could significantly reduce model accuracy: (i) the effect of multicollinearity among gene features and (ii) the diverse and even contrasting correlations between gene features and gene essentiality existing within and among different species. To address these issues, we developed a novel model called feature-based weighted Naïve Bayes model (FWM), which is based on Naïve Bayes classifiers, logistic regression, and genetic algorithm. The proposed model assesses features and filters out the effects of multicollinearity and diversity. The performance of FWM was compared with other popular models, such as support vector machine, Naïve Bayes model, and logistic regression model, by applying FWM to reciprocally predict essential genes among and within 21 species. Our results showed that FWM significantly improves the accuracy and robustness of essential gene prediction. FWM can remarkably improve the accuracy of essential gene prediction and may be used as an alternative method for other classification work. This method can contribute substantially to the knowledge of the minimum gene sets required for living organisms and the discovery of new drug targets.

  16. Variable selection for zero-inflated and overdispersed data with application to health care demand in Germany

    PubMed Central

    Wang, Zhu; Shuangge, Ma; Wang, Ching-Yun

    2017-01-01

    In health services and outcome research, count outcomes are frequently encountered and often have a large proportion of zeros. The zero-inflated negative binomial (ZINB) regression model has important applications for this type of data. With many possible candidate risk factors, this paper proposes new variable selection methods for the ZINB model. We consider maximum likelihood function plus a penalty including the least absolute shrinkage and selection operator (LASSO), smoothly clipped absolute deviation (SCAD) and minimax concave penalty (MCP). An EM (expectation-maximization) algorithm is proposed for estimating the model parameters and conducting variable selection simultaneously. This algorithm consists of estimating penalized weighted negative binomial models and penalized logistic models via the coordinated descent algorithm. Furthermore, statistical properties including the standard error formulae are provided. A simulation study shows that the new algorithm not only has more accurate or at least comparable estimation, also is more robust than the traditional stepwise variable selection. The proposed methods are applied to analyze the health care demand in Germany using an open-source R package mpath. PMID:26059498

  17. Visualizing Nanoscopic Topography and Patterns in Freely Standing Thin Films

    NASA Astrophysics Data System (ADS)

    Sharma, Vivek; Zhang, Yiran; Yilixiati, Subinuer

    Thin liquid films containing micelles, nanoparticles, polyelectrolyte-surfactant complexes and smectic liquid crystals undergo thinning in a discontinuous, step-wise fashion. The discontinuous jumps in thickness are often characterized by quantifying changes in the intensity of reflected monochromatic light, modulated by thin film interference from a region of interest. Stratifying thin films exhibit a mosaic pattern in reflected white light microscopy, attributed to the coexistence of domains with various thicknesses, separated by steps. Using Interferometry Digital Imaging Optical Microscopy (IDIOM) protocols developed in the course of this study, we spatially resolve for the first time, the landscape of stratifying freely standing thin films. We distinguish nanoscopic rims, mesas and craters, and follow their emergence and growth. In particular, for thin films containing micelles of sodium dodecyl sulfate (SDS), these topological features involve discontinuous, thickness transitions with concentration-dependent steps of 5-25 nm. These non-flat features result from oscillatory, periodic, supramolecular structural forces that arise in confined fluids, and arise due to complex coupling of hydrodynamic and thermodynamic effects at the nanoscale.

  18. Feature engineering for drug name recognition in biomedical texts: feature conjunction and feature selection.

    PubMed

    Liu, Shengyu; Tang, Buzhou; Chen, Qingcai; Wang, Xiaolong; Fan, Xiaoming

    2015-01-01

    Drug name recognition (DNR) is a critical step for drug information extraction. Machine learning-based methods have been widely used for DNR with various types of features such as part-of-speech, word shape, and dictionary feature. Features used in current machine learning-based methods are usually singleton features which may be due to explosive features and a large number of noisy features when singleton features are combined into conjunction features. However, singleton features that can only capture one linguistic characteristic of a word are not sufficient to describe the information for DNR when multiple characteristics should be considered. In this study, we explore feature conjunction and feature selection for DNR, which have never been reported. We intuitively select 8 types of singleton features and combine them into conjunction features in two ways. Then, Chi-square, mutual information, and information gain are used to mine effective features. Experimental results show that feature conjunction and feature selection can improve the performance of the DNR system with a moderate number of features and our DNR system significantly outperforms the best system in the DDIExtraction 2013 challenge.

  19. Effect of feature-selective attention on neuronal responses in macaque area MT

    PubMed Central

    Chen, X.; Hoffmann, K.-P.; Albright, T. D.

    2012-01-01

    Attention influences visual processing in striate and extrastriate cortex, which has been extensively studied for spatial-, object-, and feature-based attention. Most studies exploring neural signatures of feature-based attention have trained animals to attend to an object identified by a certain feature and ignore objects/displays identified by a different feature. Little is known about the effects of feature-selective attention, where subjects attend to one stimulus feature domain (e.g., color) of an object while features from different domains (e.g., direction of motion) of the same object are ignored. To study this type of feature-selective attention in area MT in the middle temporal sulcus, we trained macaque monkeys to either attend to and report the direction of motion of a moving sine wave grating (a feature for which MT neurons display strong selectivity) or attend to and report its color (a feature for which MT neurons have very limited selectivity). We hypothesized that neurons would upregulate their firing rate during attend-direction conditions compared with attend-color conditions. We found that feature-selective attention significantly affected 22% of MT neurons. Contrary to our hypothesis, these neurons did not necessarily increase firing rate when animals attended to direction of motion but fell into one of two classes. In one class, attention to color increased the gain of stimulus-induced responses compared with attend-direction conditions. The other class displayed the opposite effects. Feature-selective activity modulations occurred earlier in neurons modulated by attention to color compared with neurons modulated by attention to motion direction. Thus feature-selective attention influences neuronal processing in macaque area MT but often exhibited a mismatch between the preferred stimulus dimension (direction of motion) and the preferred attention dimension (attention to color). PMID:22170961

  20. Effect of feature-selective attention on neuronal responses in macaque area MT.

    PubMed

    Chen, X; Hoffmann, K-P; Albright, T D; Thiele, A

    2012-03-01

    Attention influences visual processing in striate and extrastriate cortex, which has been extensively studied for spatial-, object-, and feature-based attention. Most studies exploring neural signatures of feature-based attention have trained animals to attend to an object identified by a certain feature and ignore objects/displays identified by a different feature. Little is known about the effects of feature-selective attention, where subjects attend to one stimulus feature domain (e.g., color) of an object while features from different domains (e.g., direction of motion) of the same object are ignored. To study this type of feature-selective attention in area MT in the middle temporal sulcus, we trained macaque monkeys to either attend to and report the direction of motion of a moving sine wave grating (a feature for which MT neurons display strong selectivity) or attend to and report its color (a feature for which MT neurons have very limited selectivity). We hypothesized that neurons would upregulate their firing rate during attend-direction conditions compared with attend-color conditions. We found that feature-selective attention significantly affected 22% of MT neurons. Contrary to our hypothesis, these neurons did not necessarily increase firing rate when animals attended to direction of motion but fell into one of two classes. In one class, attention to color increased the gain of stimulus-induced responses compared with attend-direction conditions. The other class displayed the opposite effects. Feature-selective activity modulations occurred earlier in neurons modulated by attention to color compared with neurons modulated by attention to motion direction. Thus feature-selective attention influences neuronal processing in macaque area MT but often exhibited a mismatch between the preferred stimulus dimension (direction of motion) and the preferred attention dimension (attention to color).

  1. Comparison of Feature Selection Techniques in Machine Learning for Anatomical Brain MRI in Dementia.

    PubMed

    Tohka, Jussi; Moradi, Elaheh; Huttunen, Heikki

    2016-07-01

    We present a comparative split-half resampling analysis of various data driven feature selection and classification methods for the whole brain voxel-based classification analysis of anatomical magnetic resonance images. We compared support vector machines (SVMs), with or without filter based feature selection, several embedded feature selection methods and stability selection. While comparisons of the accuracy of various classification methods have been reported previously, the variability of the out-of-training sample classification accuracy and the set of selected features due to independent training and test sets have not been previously addressed in a brain imaging context. We studied two classification problems: 1) Alzheimer's disease (AD) vs. normal control (NC) and 2) mild cognitive impairment (MCI) vs. NC classification. In AD vs. NC classification, the variability in the test accuracy due to the subject sample did not vary between different methods and exceeded the variability due to different classifiers. In MCI vs. NC classification, particularly with a large training set, embedded feature selection methods outperformed SVM-based ones with the difference in the test accuracy exceeding the test accuracy variability due to the subject sample. The filter and embedded methods produced divergent feature patterns for MCI vs. NC classification that suggests the utility of the embedded feature selection for this problem when linked with the good generalization performance. The stability of the feature sets was strongly correlated with the number of features selected, weakly correlated with the stability of classification accuracy, and uncorrelated with the average classification accuracy.

  2. Modified Bat Algorithm for Feature Selection with the Wisconsin Diagnosis Breast Cancer (WDBC) Dataset

    PubMed

    Jeyasingh, Suganthi; Veluchamy, Malathi

    2017-05-01

    Early diagnosis of breast cancer is essential to save lives of patients. Usually, medical datasets include a large variety of data that can lead to confusion during diagnosis. The Knowledge Discovery on Database (KDD) process helps to improve efficiency. It requires elimination of inappropriate and repeated data from the dataset before final diagnosis. This can be done using any of the feature selection algorithms available in data mining. Feature selection is considered as a vital step to increase the classification accuracy. This paper proposes a Modified Bat Algorithm (MBA) for feature selection to eliminate irrelevant features from an original dataset. The Bat algorithm was modified using simple random sampling to select the random instances from the dataset. Ranking was with the global best features to recognize the predominant features available in the dataset. The selected features are used to train a Random Forest (RF) classification algorithm. The MBA feature selection algorithm enhanced the classification accuracy of RF in identifying the occurrence of breast cancer. The Wisconsin Diagnosis Breast Cancer Dataset (WDBC) was used for estimating the performance analysis of the proposed MBA feature selection algorithm. The proposed algorithm achieved better performance in terms of Kappa statistic, Mathew’s Correlation Coefficient, Precision, F-measure, Recall, Mean Absolute Error (MAE), Root Mean Square Error (RMSE), Relative Absolute Error (RAE) and Root Relative Squared Error (RRSE). Creative Commons Attribution License

  3. Genetic Particle Swarm Optimization–Based Feature Selection for Very-High-Resolution Remotely Sensed Imagery Object Change Detection

    PubMed Central

    Chen, Qiang; Chen, Yunhao; Jiang, Weiguo

    2016-01-01

    In the field of multiple features Object-Based Change Detection (OBCD) for very-high-resolution remotely sensed images, image objects have abundant features and feature selection affects the precision and efficiency of OBCD. Through object-based image analysis, this paper proposes a Genetic Particle Swarm Optimization (GPSO)-based feature selection algorithm to solve the optimization problem of feature selection in multiple features OBCD. We select the Ratio of Mean to Variance (RMV) as the fitness function of GPSO, and apply the proposed algorithm to the object-based hybrid multivariate alternative detection model. Two experiment cases on Worldview-2/3 images confirm that GPSO can significantly improve the speed of convergence, and effectively avoid the problem of premature convergence, relative to other feature selection algorithms. According to the accuracy evaluation of OBCD, GPSO is superior at overall accuracy (84.17% and 83.59%) and Kappa coefficient (0.6771 and 0.6314) than other algorithms. Moreover, the sensitivity analysis results show that the proposed algorithm is not easily influenced by the initial parameters, but the number of features to be selected and the size of the particle swarm would affect the algorithm. The comparison experiment results reveal that RMV is more suitable than other functions as the fitness function of GPSO-based feature selection algorithm. PMID:27483285

  4. Toward optimal feature and time segment selection by divergence method for EEG signals classification.

    PubMed

    Wang, Jie; Feng, Zuren; Lu, Na; Luo, Jing

    2018-06-01

    Feature selection plays an important role in the field of EEG signals based motor imagery pattern classification. It is a process that aims to select an optimal feature subset from the original set. Two significant advantages involved are: lowering the computational burden so as to speed up the learning procedure and removing redundant and irrelevant features so as to improve the classification performance. Therefore, feature selection is widely employed in the classification of EEG signals in practical brain-computer interface systems. In this paper, we present a novel statistical model to select the optimal feature subset based on the Kullback-Leibler divergence measure, and automatically select the optimal subject-specific time segment. The proposed method comprises four successive stages: a broad frequency band filtering and common spatial pattern enhancement as preprocessing, features extraction by autoregressive model and log-variance, the Kullback-Leibler divergence based optimal feature and time segment selection and linear discriminate analysis classification. More importantly, this paper provides a potential framework for combining other feature extraction models and classification algorithms with the proposed method for EEG signals classification. Experiments on single-trial EEG signals from two public competition datasets not only demonstrate that the proposed method is effective in selecting discriminative features and time segment, but also show that the proposed method yields relatively better classification results in comparison with other competitive methods. Copyright © 2018 Elsevier Ltd. All rights reserved.

  5. Sentiment analysis of feature ranking methods for classification accuracy

    NASA Astrophysics Data System (ADS)

    Joseph, Shashank; Mugauri, Calvin; Sumathy, S.

    2017-11-01

    Text pre-processing and feature selection are important and critical steps in text mining. Text pre-processing of large volumes of datasets is a difficult task as unstructured raw data is converted into structured format. Traditional methods of processing and weighing took much time and were less accurate. To overcome this challenge, feature ranking techniques have been devised. A feature set from text preprocessing is fed as input for feature selection. Feature selection helps improve text classification accuracy. Of the three feature selection categories available, the filter category will be the focus. Five feature ranking methods namely: document frequency, standard deviation information gain, CHI-SQUARE, and weighted-log likelihood -ratio is analyzed.

  6. Mutual information criterion for feature selection with application to classification of breast microcalcifications

    NASA Astrophysics Data System (ADS)

    Diamant, Idit; Shalhon, Moran; Goldberger, Jacob; Greenspan, Hayit

    2016-03-01

    Classification of clustered breast microcalcifications into benign and malignant categories is an extremely challenging task for computerized algorithms and expert radiologists alike. In this paper we present a novel method for feature selection based on mutual information (MI) criterion for automatic classification of microcalcifications. We explored the MI based feature selection for various texture features. The proposed method was evaluated on a standardized digital database for screening mammography (DDSM). Experimental results demonstrate the effectiveness and the advantage of using the MI-based feature selection to obtain the most relevant features for the task and thus to provide for improved performance as compared to using all features.

  7. Multivariate Profiles of Selected versus Non-Selected Elite Youth Brazilian Soccer Players

    PubMed Central

    Alves, Isabella S.; Padilha, Maickel B.; Casanova, Filipe; Puggina, Enrico F.; Maia, José

    2017-01-01

    Abstract This study determined whether a multivariate profile more effectively discriminated selected than non-selected elite youth Brazilian soccer players. This examination was carried out on 66 youth soccer players (selected, n = 28, mean age 16.3 ± 0.1; non-selected, n = 38, mean age 16.7 ± 0.4) using objective instruments. Multivariate profiles were assessed through anthropometric characteristics, biological maturation, tactical-technical skills, and motor performance. The Student’s t-test identified that selected players exhibited significantly higher values for height (t = 2.331, p = 0.02), lean body mass (t = 2.441, p = 0.01), and maturity offset (t = 4.559, p < 0.001), as well as performed better in declarative tactical knowledge (t = 10.484, p < 0.001), shooting (t = 2.188, p = 0.03), dribbling (t = 5.914, p < 0.001), speed – 30 m (t = 8.304, p < 0.001), countermovement jump (t = 2.718, p = 0.008), and peak power tests (t = 2.454, p = 0.01). Forward stepwise discriminant function analysis showed that declarative tactical knowledge, running speed –30 m, maturity offset, dribbling, height, and peak power correctly classified 97% of the selected players. These findings may have implications for a highly efficient selection process with objective measures of youth players in soccer clubs. PMID:29339991

  8. Effects of stepwise dry/wet-aging and freezing on meat quality of beef loins.

    PubMed

    Kim, Yuan H Brad; Meyers, Brandon; Kim, Hyun-Wook; Liceaga, Andrea M; Lemenager, Ronald P

    2017-01-01

    The objective of this study was to evaluate the effects of stepwise dry/wet-aging and freezing method on quality attributes of beef loins. Paired loins (M. Longissimus lumborum) from eight carcasses were assigned to either stepwise dry/wet-aging (carcass dry-aging for 10days then further wet-aging for 7days in vacuum bags) or carcass dry-aging only for 17days. Then, each loin was divided into three sections for freezing (never-frozen, blast or cryogenic freezing). Stepwise dry/wet-aged loin had lower purge/drip loss and shear force than conventionally dry-aged loin (P<0.05), but similar color and sensory characteristics (P>0.05). The cryogenic freezing resulted in a significant decrease in shear force values and a significant improvement in water-holding capacity (WHC). These findings indicate that the stepwise dry/wet-aging coupled with cryogenic freezing could provide beneficial impacts to the local meat industry by providing equivalent quality attributes as conventional dry-aging and improving WHC of frozen/thawed meat, while reducing the time needed for dry-aging. Copyright © 2016 Elsevier Ltd. All rights reserved.

  9. Feature Selection Methods for Zero-Shot Learning of Neural Activity

    PubMed Central

    Caceres, Carlos A.; Roos, Matthew J.; Rupp, Kyle M.; Milsap, Griffin; Crone, Nathan E.; Wolmetz, Michael E.; Ratto, Christopher R.

    2017-01-01

    Dimensionality poses a serious challenge when making predictions from human neuroimaging data. Across imaging modalities, large pools of potential neural features (e.g., responses from particular voxels, electrodes, and temporal windows) have to be related to typically limited sets of stimuli and samples. In recent years, zero-shot prediction models have been introduced for mapping between neural signals and semantic attributes, which allows for classification of stimulus classes not explicitly included in the training set. While choices about feature selection can have a substantial impact when closed-set accuracy, open-set robustness, and runtime are competing design objectives, no systematic study of feature selection for these models has been reported. Instead, a relatively straightforward feature stability approach has been adopted and successfully applied across models and imaging modalities. To characterize the tradeoffs in feature selection for zero-shot learning, we compared correlation-based stability to several other feature selection techniques on comparable data sets from two distinct imaging modalities: functional Magnetic Resonance Imaging and Electrocorticography. While most of the feature selection methods resulted in similar zero-shot prediction accuracies and spatial/spectral patterns of selected features, there was one exception; A novel feature/attribute correlation approach was able to achieve those accuracies with far fewer features, suggesting the potential for simpler prediction models that yield high zero-shot classification accuracy. PMID:28690513

  10. Ovarian tissue cryopreservation by stepped vitrification and monitored by X-ray computed tomography.

    PubMed

    Corral, Ariadna; Clavero, Macarena; Gallardo, Miguel; Balcerzyk, Marcin; Amorim, Christiani A; Parrado-Gallego, Ángel; Dolmans, Marie-Madeleine; Paulini, Fernanda; Morris, John; Risco, Ramón

    2018-04-01

    Ovarian tissue cryopreservation is, in most cases, the only fertility preservation option available for female patients soon to undergo gonadotoxic treatment. To date, cryopreservation of ovarian tissue has been carried out by both traditional slow freezing method and vitrification, but even with the best techniques, there is still a considerable loss of follicle viability. In this report, we investigated a stepped cryopreservation procedure which combines features of slow cooling and vitrification (hereafter called stepped vitrification). Bovine ovarian tissue was used as a tissue model. Stepwise increments of the Me 2 SO concentration coupled with stepwise drops-in temperature in a device specifically designed for this purpose and X-ray computed tomography were combined to investigate loading times at each step, by monitoring the attenuation of the radiation proportional to Me 2 SO permeation. Viability analysis was performed in warmed tissues by immunohistochemistry. Although further viability tests should be conducted after transplantation, preliminary results are very promising. Four protocols were explored. Two of them showed a poor permeation of the vitrification solution (P1 and P2). The other two (P3 and P4), with higher permeation, were studied in deeper detail. Out of these two protocols, P4, with a longer permeation time at -40 °C, showed the same histological integrity after warming as fresh controls. Copyright © 2018 Elsevier Inc. All rights reserved.

  11. Mps1 kinase-dependent Sgo2 centromere localisation mediates cohesin protection in mouse oocyte meiosis I.

    PubMed

    El Yakoubi, Warif; Buffin, Eulalie; Cladière, Damien; Gryaznova, Yulia; Berenguer, Inés; Touati, Sandra A; Gómez, Rocío; Suja, José A; van Deursen, Jan M; Wassmann, Katja

    2017-09-25

    A key feature of meiosis is the step-wise removal of cohesin, the protein complex holding sister chromatids together, first from arms in meiosis I and then from the centromere region in meiosis II. Centromeric cohesin is protected by Sgo2 from Separase-mediated cleavage, in order to maintain sister chromatids together until their separation in meiosis II. Failures in step-wise cohesin removal result in aneuploid gametes, preventing the generation of healthy embryos. Here, we report that kinase activities of Bub1 and Mps1 are required for Sgo2 localisation to the centromere region. Mps1 inhibitor-treated oocytes are defective in centromeric cohesin protection, whereas oocytes devoid of Bub1 kinase activity, which cannot phosphorylate H2A at T121, are not perturbed in cohesin protection as long as Mps1 is functional. Mps1 and Bub1 kinase activities localise Sgo2 in meiosis I preferentially to the centromere and pericentromere respectively, indicating that Sgo2 at the centromere is required for protection.In meiosis I centromeric cohesin is protected by Sgo2 from Separase-mediated cleavage ensuring that sister chromatids are kept together until their separation in meiosis II. Here the authors demonstrate that Bub1 and Mps1 kinase activities are required for Sgo2 localisation to the centromere region.

  12. Factors associated with fall-related fractures in Parkinson's disease.

    PubMed

    Cheng, Kuei-Yueh; Lin, Wei-Che; Chang, Wen-Neng; Lin, Tzu-Kong; Tsai, Nai-Wen; Huang, Chih-Cheng; Wang, Hung-Chen; Huang, Yung-Cheng; Chang, Hsueh-Wen; Lin, Yu-Jun; Lee, Lian-Hui; Cheng, Ben-Chung; Kung, Chia-Te; Chang, Ya-Ting; Su, Chih-Min; Chiang, Yi-Fang; Su, Yu-Jih; Lu, Cheng-Hsien

    2014-01-01

    Fall-related fracture is one of the most disabling features of idiopathic Parkinson's disease (PD). A better understanding of the associated factors is needed to predict PD patients who will require treatment. This prospective study enrolled 100 adult idiopathic PD patients. Stepwise logistic regressions were used to evaluate the relationships between clinical factors and fall-related fracture. Falls occurred in 56 PD patients, including 32 with fall-related fractures. The rate of falls in the study period was 2.2 ± 1.4 per 18 months. The percentage of osteoporosis was 34% (19/56) and 11% in PD patients with and without falls, respectively. Risk factors associated with fall-related fracture were sex, underlying knee osteoarthritis, mean Unified Parkinson's Disease Rating Scale score, mean Morse fall scale, mean Hoehn and Yahr stage, and exercise habit. By stepwise logistic regression, sex and mean Morse fall scale were independently associated with fall-related fracture. Females had an odds ratio of 3.8 compared to males and the cut-off value of the Morse fall scale for predicting fall-related fracture was 72.5 (sensitivity 72% and specificity 70%). Higher mean Morse fall scales (>72.5) and female sex are associated with higher risk of fall-related fractures. Preventing falls in the high-risk PD group is an important safety issue and highly relevant for their quality of life. Copyright © 2013 Elsevier Ltd. All rights reserved.

  13. Relationship between COMLEX-USA scores and performance on the American Osteopathic Board of Emergency Medicine Part I certifying examination.

    PubMed

    Li, Feiming; Gimpel, John R; Arenson, Ethan; Song, Hao; Bates, Bruce P; Ludwin, Fredric

    2014-04-01

    Few studies have investigated how well scores from the Comprehensive Osteopathic Medical Licensing Examination-USA (COMLEX-USA) series predict resident outcomes, such as performance on board certification examinations. To determine how well COMLEX-USA predicts performance on the American Osteopathic Board of Emergency Medicine (AOBEM) Part I certification examination. The target study population was first-time examinees who took AOBEM Part I in 2011 and 2012 with matched performances on COMLEX-USA Level 1, Level 2-Cognitive Evaluation (CE), and Level 3. Pearson correlations were computed between AOBEM Part I first-attempt scores and COMLEX-USA performances to measure the association between these examinations. Stepwise linear regression analysis was conducted to predict AOBEM Part I scores by the 3 COMLEX-USA scores. An independent t test was conducted to compare mean COMLEX-USA performances between candidates who passed and who failed AOBEM Part I, and a stepwise logistic regression analysis was used to predict the log-odds of passing AOBEM Part I on the basis of COMLEX-USA scores. Scores from AOBEM Part I had the highest correlation with COMLEX-USA Level 3 scores (.57) and slightly lower correlation with COMLEX-USA Level 2-CE scores (.53). The lowest correlation was between AOBEM Part I and COMLEX-USA Level 1 scores (.47). According to the stepwise regression model, COMLEX-USA Level 1 and Level 2-CE scores, which residency programs often use as selection criteria, together explained 30% of variance in AOBEM Part I scores. Adding Level 3 scores explained 37% of variance. The independent t test indicated that the 397 examinees passing AOBEM Part I performed significantly better than the 54 examinees failing AOBEM Part I in all 3 COMLEX-USA levels (P<.001 for all 3 levels). The logistic regression model showed that COMLEX-USA Level 1 and Level 3 scores predicted the log-odds of passing AOBEM Part I (P=.03 and P<.001, respectively). The present study empirically supported the predictive and discriminant validities of the COMLEX-USA series in relation to the AOBEM Part I certification examination. Although residency programs may use COMLEX-USA Level 1 and Level 2-CE scores as partial criteria in selecting residents, Level 3 scores, though typically not available at the time of application, are actually the most statistically related to performances on AOBEM Part I.

  14. Enhancing the Performance of LibSVM Classifier by Kernel F-Score Feature Selection

    NASA Astrophysics Data System (ADS)

    Sarojini, Balakrishnan; Ramaraj, Narayanasamy; Nickolas, Savarimuthu

    Medical Data mining is the search for relationships and patterns within the medical datasets that could provide useful knowledge for effective clinical decisions. The inclusion of irrelevant, redundant and noisy features in the process model results in poor predictive accuracy. Much research work in data mining has gone into improving the predictive accuracy of the classifiers by applying the techniques of feature selection. Feature selection in medical data mining is appreciable as the diagnosis of the disease could be done in this patient-care activity with minimum number of significant features. The objective of this work is to show that selecting the more significant features would improve the performance of the classifier. We empirically evaluate the classification effectiveness of LibSVM classifier on the reduced feature subset of diabetes dataset. The evaluations suggest that the feature subset selected improves the predictive accuracy of the classifier and reduce false negatives and false positives.

  15. The fate of task-irrelevant visual motion: perceptual load versus feature-based attention.

    PubMed

    Taya, Shuichiro; Adams, Wendy J; Graf, Erich W; Lavie, Nilli

    2009-11-18

    We tested contrasting predictions derived from perceptual load theory and from recent feature-based selection accounts. Observers viewed moving, colored stimuli and performed low or high load tasks associated with one stimulus feature, either color or motion. The resultant motion aftereffect (MAE) was used to evaluate attentional allocation. We found that task-irrelevant visual features received less attention than co-localized task-relevant features of the same objects. Moreover, when color and motion features were co-localized yet perceived to belong to two distinct surfaces, feature-based selection was further increased at the expense of object-based co-selection. Load theory predicts that the MAE for task-irrelevant motion would be reduced with a higher load color task. However, this was not seen for co-localized features; perceptual load only modulated the MAE for task-irrelevant motion when this was spatially separated from the attended color location. Our results suggest that perceptual load effects are mediated by spatial selection and do not generalize to the feature domain. Feature-based selection operates to suppress processing of task-irrelevant, co-localized features, irrespective of perceptual load.

  16. Classification Influence of Features on Given Emotions and Its Application in Feature Selection

    NASA Astrophysics Data System (ADS)

    Xing, Yin; Chen, Chuang; Liu, Li-Long

    2018-04-01

    In order to solve the problem that there is a large amount of redundant data in high-dimensional speech emotion features, we analyze deeply the extracted speech emotion features and select better features. Firstly, a given emotion is classified by each feature. Secondly, the recognition rate is ranked in descending order. Then, the optimal threshold of features is determined by rate criterion. Finally, the better features are obtained. When applied in Berlin and Chinese emotional data set, the experimental results show that the feature selection method outperforms the other traditional methods.

  17. Feature Selection in Classification of Eye Movements Using Electrooculography for Activity Recognition

    PubMed Central

    Mala, S.; Latha, K.

    2014-01-01

    Activity recognition is needed in different requisition, for example, reconnaissance system, patient monitoring, and human-computer interfaces. Feature selection plays an important role in activity recognition, data mining, and machine learning. In selecting subset of features, an efficient evolutionary algorithm Differential Evolution (DE), a very efficient optimizer, is used for finding informative features from eye movements using electrooculography (EOG). Many researchers use EOG signals in human-computer interactions with various computational intelligence methods to analyze eye movements. The proposed system involves analysis of EOG signals using clearness based features, minimum redundancy maximum relevance features, and Differential Evolution based features. This work concentrates more on the feature selection algorithm based on DE in order to improve the classification for faultless activity recognition. PMID:25574185

  18. Feature selection in classification of eye movements using electrooculography for activity recognition.

    PubMed

    Mala, S; Latha, K

    2014-01-01

    Activity recognition is needed in different requisition, for example, reconnaissance system, patient monitoring, and human-computer interfaces. Feature selection plays an important role in activity recognition, data mining, and machine learning. In selecting subset of features, an efficient evolutionary algorithm Differential Evolution (DE), a very efficient optimizer, is used for finding informative features from eye movements using electrooculography (EOG). Many researchers use EOG signals in human-computer interactions with various computational intelligence methods to analyze eye movements. The proposed system involves analysis of EOG signals using clearness based features, minimum redundancy maximum relevance features, and Differential Evolution based features. This work concentrates more on the feature selection algorithm based on DE in order to improve the classification for faultless activity recognition.

  19. Feldspar 40Ar/39Ar dating of ICDP PALEOVAN cores

    NASA Astrophysics Data System (ADS)

    Engelhardt, Jonathan Franz; Sudo, Masafumi; Stockhecke, Mona; Oberhänsli, Roland

    2017-11-01

    Volcaniclastic fall deposits in ICDP drilling cores from Lake Van, Turkey, contain sodium-rich sanidine and calcium-rich anorthoclase, which both comprise a variety of textural zoning and inclusions. An age model records the lake's history and is based on climate-stratigraphic correlations, tephrostratigraphy, paleomagnetics, and earlier 40Ar/39Ar analyses (Stockhecke et al., 2014b). Results from total fusion and stepwise heating 40Ar/39Ar analyses presented in this study allow for the comparison of radiometric constraints from texturally diversified feldspar and the multi-proxy lacustrine age model and vice versa. This study has investigated several grain-size fractions of feldspar from 13 volcaniclastic units. The feldspars show textural features that are visible in cathodoluminescence (CL) or back-scattered electron (BSE) images and can be subdivided into three dominant zoning-types: (1) compositional zoning, (2) round pseudo-oscillatory zoning and (3) resorbed and patchy zoning (Ginibre et al., 2004). Round pseudo-oscillatory zoning records a sensitive alternation of Fe and Ca that also reflects resorption processes. This is only visible in CL images. Compositional zoning reflects anticorrelated anorthite and orthoclase contents and is visible in BSE. Eleven inverse isochron ages from total fusion and three from stepwise heating analyses fit the age model. Four experiments resulted in older inverse isochron ages that do not concur with the model within 2σ uncertainties and that deviate from 1 ka to 17 ka minimum. C- and R-type zoning are interpreted as representing growth in magma chamber cupolas, as wall mushes, or in narrow conduits. Persistent compositions of PO-type crystals and abundant surfaces recording dissolution features correspond to formation within a magma chamber. C-type zoning and R-type zoning have revealed an irregular incorporation of melt and fluid inclusions. These two types of zoning in feldspar are interpreted as preferentially contributing either heterogeneously distributed excess 40Ar or inherited 40Ar to the deviating 40Ar/39Ar ages that are discussed in this study.

  20. Perceptual quality estimation of H.264/AVC videos using reduced-reference and no-reference models

    NASA Astrophysics Data System (ADS)

    Shahid, Muhammad; Pandremmenou, Katerina; Kondi, Lisimachos P.; Rossholm, Andreas; Lövström, Benny

    2016-09-01

    Reduced-reference (RR) and no-reference (NR) models for video quality estimation, using features that account for the impact of coding artifacts, spatio-temporal complexity, and packet losses, are proposed. The purpose of this study is to analyze a number of potentially quality-relevant features in order to select the most suitable set of features for building the desired models. The proposed sets of features have not been used in the literature and some of the features are used for the first time in this study. The features are employed by the least absolute shrinkage and selection operator (LASSO), which selects only the most influential of them toward perceptual quality. For comparison, we apply feature selection in the complete feature sets and ridge regression on the reduced sets. The models are validated using a database of H.264/AVC encoded videos that were subjectively assessed for quality in an ITU-T compliant laboratory. We infer that just two features selected by RR LASSO and two bitstream-based features selected by NR LASSO are able to estimate perceptual quality with high accuracy, higher than that of ridge, which uses more features. The comparisons with competing works and two full-reference metrics also verify the superiority of our models.

  1. Feature Grouping and Selection Over an Undirected Graph.

    PubMed

    Yang, Sen; Yuan, Lei; Lai, Ying-Cheng; Shen, Xiaotong; Wonka, Peter; Ye, Jieping

    2012-01-01

    High-dimensional regression/classification continues to be an important and challenging problem, especially when features are highly correlated. Feature selection, combined with additional structure information on the features has been considered to be promising in promoting regression/classification performance. Graph-guided fused lasso (GFlasso) has recently been proposed to facilitate feature selection and graph structure exploitation, when features exhibit certain graph structures. However, the formulation in GFlasso relies on pairwise sample correlations to perform feature grouping, which could introduce additional estimation bias. In this paper, we propose three new feature grouping and selection methods to resolve this issue. The first method employs a convex function to penalize the pairwise l ∞ norm of connected regression/classification coefficients, achieving simultaneous feature grouping and selection. The second method improves the first one by utilizing a non-convex function to reduce the estimation bias. The third one is the extension of the second method using a truncated l 1 regularization to further reduce the estimation bias. The proposed methods combine feature grouping and feature selection to enhance estimation accuracy. We employ the alternating direction method of multipliers (ADMM) and difference of convex functions (DC) programming to solve the proposed formulations. Our experimental results on synthetic data and two real datasets demonstrate the effectiveness of the proposed methods.

  2. Multiobjective Model of Time-of-Use and Stepwise Power Tariff for Residential Consumers in Regulated Power Markets

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Zhou, Bin; Yang, Rui; Li, Canbing

    Here, time-of-use (TOU) rates and stepwise power tariff (SPT) are important economic levers to motivate residents to shift their electricity usage in response to electricity price. In this paper, a new multiobjective optimal tariff-making model of time-of-use and stepwise power tariff (TOUSPT) is proposed, which combines the complementary characteristics of two power tariffs, for residential energy conservation and peak load shaving. In the proposed approach, the residential demand response with price elasticity in regulated power market is considered to determine the optimum peak-valley TOU tariffs for each stepwise electricity partition. Furthermore, a practical case study is implemented to test themore » effectiveness of the proposed TOUSPT, and the results demonstrate that TOUSPT can achieve efficient end-use energy saving and also shift load from peak to off-peak periods.« less

  3. Multiobjective Model of Time-of-Use and Stepwise Power Tariff for Residential Consumers in Regulated Power Markets

    DOE PAGES

    Zhou, Bin; Yang, Rui; Li, Canbing; ...

    2017-07-04

    Here, time-of-use (TOU) rates and stepwise power tariff (SPT) are important economic levers to motivate residents to shift their electricity usage in response to electricity price. In this paper, a new multiobjective optimal tariff-making model of time-of-use and stepwise power tariff (TOUSPT) is proposed, which combines the complementary characteristics of two power tariffs, for residential energy conservation and peak load shaving. In the proposed approach, the residential demand response with price elasticity in regulated power market is considered to determine the optimum peak-valley TOU tariffs for each stepwise electricity partition. Furthermore, a practical case study is implemented to test themore » effectiveness of the proposed TOUSPT, and the results demonstrate that TOUSPT can achieve efficient end-use energy saving and also shift load from peak to off-peak periods.« less

  4. Proton radius from electron scattering data

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Higinbotham, Douglas W.; Kabir, Al Amin; Lin, Vincent

    Background: The proton charge radius extracted from recent muonic hydrogen Lamb shift measurements is significantly smaller than that extracted from atomic hydrogen and electron scattering measurements. The discrepancy has become known as the proton radius puzzle. Purpose: In an attempt to understand the discrepancy, we review high-precision electron scattering results from Mainz, Jefferson Lab, Saskatoon and Stanford. Methods: We make use of stepwise regression techniques using the F-test as well as the Akaike information criterion to systematically determine the predictive variables to use for a given set and range of electron scattering data as well as to provide multivariate errormore » estimates. Results: Starting with the precision, low four-momentum transfer (Q 2) data from Mainz (1980) and Saskatoon (1974), we find that a stepwise regression of the Maclaurin series using the F-test as well as the Akaike information criterion justify using a linear extrapolation which yields a value for the proton radius that is consistent with the result obtained from muonic hydrogen measurements. Applying the same Maclaurin series and statistical criteria to the 2014 Rosenbluth results on GE from Mainz, we again find that the stepwise regression tends to favor a radius consistent with the muonic hydrogen radius but produces results that are extremely sensitive to the range of data included in the fit. Making use of the high-Q 2 data on G E to select functions which extrapolate to high Q 2, we find that a Pad´e (N = M = 1) statistical model works remarkably well, as does a dipole function with a 0.84 fm radius, G E(Q 2) = (1 + Q 2/0.66 GeV 2) -2. Conclusions: Rigorous applications of stepwise regression techniques and multivariate error estimates result in the extraction of a proton charge radius that is consistent with the muonic hydrogen result of 0.84 fm; either from linear extrapolation of the extreme low-Q 2 data or by use of the Pad´e approximant for extrapolation using a larger range of data. Thus, based on a purely statistical analysis of electron scattering data, we conclude that the electron scattering result and the muonic hydrogen result are consistent. Lastly, it is the atomic hydrogen results that are the outliers.« less

  5. Proton radius from electron scattering data

    DOE PAGES

    Higinbotham, Douglas W.; Kabir, Al Amin; Lin, Vincent; ...

    2016-05-31

    Background: The proton charge radius extracted from recent muonic hydrogen Lamb shift measurements is significantly smaller than that extracted from atomic hydrogen and electron scattering measurements. The discrepancy has become known as the proton radius puzzle. Purpose: In an attempt to understand the discrepancy, we review high-precision electron scattering results from Mainz, Jefferson Lab, Saskatoon and Stanford. Methods: We make use of stepwise regression techniques using the F-test as well as the Akaike information criterion to systematically determine the predictive variables to use for a given set and range of electron scattering data as well as to provide multivariate errormore » estimates. Results: Starting with the precision, low four-momentum transfer (Q 2) data from Mainz (1980) and Saskatoon (1974), we find that a stepwise regression of the Maclaurin series using the F-test as well as the Akaike information criterion justify using a linear extrapolation which yields a value for the proton radius that is consistent with the result obtained from muonic hydrogen measurements. Applying the same Maclaurin series and statistical criteria to the 2014 Rosenbluth results on GE from Mainz, we again find that the stepwise regression tends to favor a radius consistent with the muonic hydrogen radius but produces results that are extremely sensitive to the range of data included in the fit. Making use of the high-Q 2 data on G E to select functions which extrapolate to high Q 2, we find that a Pad´e (N = M = 1) statistical model works remarkably well, as does a dipole function with a 0.84 fm radius, G E(Q 2) = (1 + Q 2/0.66 GeV 2) -2. Conclusions: Rigorous applications of stepwise regression techniques and multivariate error estimates result in the extraction of a proton charge radius that is consistent with the muonic hydrogen result of 0.84 fm; either from linear extrapolation of the extreme low-Q 2 data or by use of the Pad´e approximant for extrapolation using a larger range of data. Thus, based on a purely statistical analysis of electron scattering data, we conclude that the electron scattering result and the muonic hydrogen result are consistent. Lastly, it is the atomic hydrogen results that are the outliers.« less

  6. Analysis of model development strategies: predicting ventral hernia recurrence.

    PubMed

    Holihan, Julie L; Li, Linda T; Askenasy, Erik P; Greenberg, Jacob A; Keith, Jerrod N; Martindale, Robert G; Roth, J Scott; Liang, Mike K

    2016-11-01

    There have been many attempts to identify variables associated with ventral hernia recurrence; however, it is unclear which statistical modeling approach results in models with greatest internal and external validity. We aim to assess the predictive accuracy of models developed using five common variable selection strategies to determine variables associated with hernia recurrence. Two multicenter ventral hernia databases were used. Database 1 was randomly split into "development" and "internal validation" cohorts. Database 2 was designated "external validation". The dependent variable for model development was hernia recurrence. Five variable selection strategies were used: (1) "clinical"-variables considered clinically relevant, (2) "selective stepwise"-all variables with a P value <0.20 were assessed in a step-backward model, (3) "liberal stepwise"-all variables were included and step-backward regression was performed, (4) "restrictive internal resampling," and (5) "liberal internal resampling." Variables were included with P < 0.05 for the Restrictive model and P < 0.10 for the Liberal model. A time-to-event analysis using Cox regression was performed using these strategies. The predictive accuracy of the developed models was tested on the internal and external validation cohorts using Harrell's C-statistic where C > 0.70 was considered "reasonable". The recurrence rate was 32.9% (n = 173/526; median/range follow-up, 20/1-58 mo) for the development cohort, 36.0% (n = 95/264, median/range follow-up 20/1-61 mo) for the internal validation cohort, and 12.7% (n = 155/1224, median/range follow-up 9/1-50 mo) for the external validation cohort. Internal validation demonstrated reasonable predictive accuracy (C-statistics = 0.772, 0.760, 0.767, 0.757, 0.763), while on external validation, predictive accuracy dipped precipitously (C-statistic = 0.561, 0.557, 0.562, 0.553, 0.560). Predictive accuracy was equally adequate on internal validation among models; however, on external validation, all five models failed to demonstrate utility. Future studies should report multiple variable selection techniques and demonstrate predictive accuracy on external data sets for model validation. Copyright © 2016 Elsevier Inc. All rights reserved.

  7. Natural image statistics and low-complexity feature selection.

    PubMed

    Vasconcelos, Manuela; Vasconcelos, Nuno

    2009-02-01

    Low-complexity feature selection is analyzed in the context of visual recognition. It is hypothesized that high-order dependences of bandpass features contain little information for discrimination of natural images. This hypothesis is characterized formally by the introduction of the concepts of conjunctive interference and decomposability order of a feature set. Necessary and sufficient conditions for the feasibility of low-complexity feature selection are then derived in terms of these concepts. It is shown that the intrinsic complexity of feature selection is determined by the decomposability order of the feature set and not its dimension. Feature selection algorithms are then derived for all levels of complexity and are shown to be approximated by existing information-theoretic methods, which they consistently outperform. The new algorithms are also used to objectively test the hypothesis of low decomposability order through comparison of classification performance. It is shown that, for image classification, the gain of modeling feature dependencies has strongly diminishing returns: best results are obtained under the assumption of decomposability order 1. This suggests a generic law for bandpass features extracted from natural images: that the effect, on the dependence of any two features, of observing any other feature is constant across image classes.

  8. Socio-economic variables influencing mean age at marriage in Karnataka and Kerala.

    PubMed

    Prakasam, C P; Upadhyay, R B

    1985-01-01

    "In this paper an attempt was made to study the influence of certain socio-economic variables on the male and the female age at marriage in Karnataka and Kerala [India] for the year 1971. Step-wise regression method has been used to select the predictor variables influencing mean age at marriage. The results reveal that percent female literate...and percent female in labour force...are found to influence female mean age at marriage in Kerala, while the variables for Karnataka were percent female literate..., percent male literate..., and percent urban male population...." excerpt

  9. Sex determination of the Acadian Flycatcher using discriminant analysis

    USGS Publications Warehouse

    Wilson, R.R.

    1999-01-01

    I used five morphometric variables from 114 individuals captured in Arkansas to develop a discriminant model to predict the sex of Acadian Flycatchers (Empidonax virescens). Stepwise discriminant function analyses selected wing chord and tail length as the most parsimonious subset of variables for discriminating sex. This two-variable model correctly classified 80% of females and 97% of males used to develop the model. Validation of the model using 19 individuals from Louisiana and Virginia resulted in 100% correct classification of males and females. This model provides criteria for sexing monomorphic Acadian Flycatchers during the breeding season and possibly during the winter.

  10. Impact of chemical oxidation on indigenous bacteria and mobilization of nutrients and subsequent bioremediation of crude oil-contaminated soil.

    PubMed

    Xu, Jinlan; Deng, Xin; Cui, Yiwei; Kong, Fanxing

    2016-12-15

    Fenton pre-oxidation provides nutrients to promote bioremediation. However, the effects of the indigenous bacteria that remain following Fenton oxidation on nutrient mobilization and subsequent bioremediation remain unclear. Experiments were performed with inoculation with native bacteria and foreign bacteria or without inoculation after four regimens of stepwise pre-oxidations. The effects of the indigenous bacteria remaining after stepwise oxidation on nutrient mobilization and subsequent bioremediation over 80 days were investigated. After stepwise Fenton pre-oxidation at a low H 2 O 2 concentration (225×4), the remaining indigenous bacterial populations reached their peak (4.8±0.17×10 6 CFU/g), the nutrients were mobilized rapidly, and the subsequent bioremediation of crude oil was improved (biodegradation efficiency of 35%). However, after stepwise Fenton pre-oxidation at a high H 2 O 2 concentration (450×4), only 3.6±0.16×10 3 CFU/g of indigenous bacteria remained, and the indigenous bacteria that degrade C 15 -C 30 alkanes were inhibited. The nutrient mobilization was then highly limited, and only 19% of total petroleum hydrocarbon was degraded. Furthermore, the recovery period after the low H 2 O 2 concentration stepwise Fenton pre-oxidation (225×4) was less than 20 days, which was 20-30 days shorter than with the other pre-oxidation treatments. Therefore, stepwise Fenton pre-oxidation at a low H 2 O 2 concentration protects indigenous bacterial populations and improves the nutrient mobilization and subsequent bioremediation. Copyright © 2016 Elsevier B.V. All rights reserved.

  11. Effective traffic features selection algorithm for cyber-attacks samples

    NASA Astrophysics Data System (ADS)

    Li, Yihong; Liu, Fangzheng; Du, Zhenyu

    2018-05-01

    By studying the defense scheme of Network attacks, this paper propose an effective traffic features selection algorithm based on k-means++ clustering to deal with the problem of high dimensionality of traffic features which extracted from cyber-attacks samples. Firstly, this algorithm divide the original feature set into attack traffic feature set and background traffic feature set by the clustering. Then, we calculates the variation of clustering performance after removing a certain feature. Finally, evaluating the degree of distinctiveness of the feature vector according to the result. Among them, the effective feature vector is whose degree of distinctiveness exceeds the set threshold. The purpose of this paper is to select out the effective features from the extracted original feature set. In this way, it can reduce the dimensionality of the features so as to reduce the space-time overhead of subsequent detection. The experimental results show that the proposed algorithm is feasible and it has some advantages over other selection algorithms.

  12. Relevance popularity: A term event model based feature selection scheme for text classification.

    PubMed

    Feng, Guozhong; An, Baiguo; Yang, Fengqin; Wang, Han; Zhang, Libiao

    2017-01-01

    Feature selection is a practical approach for improving the performance of text classification methods by optimizing the feature subsets input to classifiers. In traditional feature selection methods such as information gain and chi-square, the number of documents that contain a particular term (i.e. the document frequency) is often used. However, the frequency of a given term appearing in each document has not been fully investigated, even though it is a promising feature to produce accurate classifications. In this paper, we propose a new feature selection scheme based on a term event Multinomial naive Bayes probabilistic model. According to the model assumptions, the matching score function, which is based on the prediction probability ratio, can be factorized. Finally, we derive a feature selection measurement for each term after replacing inner parameters by their estimators. On a benchmark English text datasets (20 Newsgroups) and a Chinese text dataset (MPH-20), our numerical experiment results obtained from using two widely used text classifiers (naive Bayes and support vector machine) demonstrate that our method outperformed the representative feature selection methods.

  13. Hybrid feature selection for supporting lightweight intrusion detection systems

    NASA Astrophysics Data System (ADS)

    Song, Jianglong; Zhao, Wentao; Liu, Qiang; Wang, Xin

    2017-08-01

    Redundant and irrelevant features not only cause high resource consumption but also degrade the performance of Intrusion Detection Systems (IDS), especially when coping with big data. These features slow down the process of training and testing in network traffic classification. Therefore, a hybrid feature selection approach in combination with wrapper and filter selection is designed in this paper to build a lightweight intrusion detection system. Two main phases are involved in this method. The first phase conducts a preliminary search for an optimal subset of features, in which the chi-square feature selection is utilized. The selected set of features from the previous phase is further refined in the second phase in a wrapper manner, in which the Random Forest(RF) is used to guide the selection process and retain an optimized set of features. After that, we build an RF-based detection model and make a fair comparison with other approaches. The experimental results on NSL-KDD datasets show that our approach results are in higher detection accuracy as well as faster training and testing processes.

  14. Classification of epileptic EEG signals based on simple random sampling and sequential feature selection.

    PubMed

    Ghayab, Hadi Ratham Al; Li, Yan; Abdulla, Shahab; Diykh, Mohammed; Wan, Xiangkui

    2016-06-01

    Electroencephalogram (EEG) signals are used broadly in the medical fields. The main applications of EEG signals are the diagnosis and treatment of diseases such as epilepsy, Alzheimer, sleep problems and so on. This paper presents a new method which extracts and selects features from multi-channel EEG signals. This research focuses on three main points. Firstly, simple random sampling (SRS) technique is used to extract features from the time domain of EEG signals. Secondly, the sequential feature selection (SFS) algorithm is applied to select the key features and to reduce the dimensionality of the data. Finally, the selected features are forwarded to a least square support vector machine (LS_SVM) classifier to classify the EEG signals. The LS_SVM classifier classified the features which are extracted and selected from the SRS and the SFS. The experimental results show that the method achieves 99.90, 99.80 and 100 % for classification accuracy, sensitivity and specificity, respectively.

  15. Joint L2,1 Norm and Fisher Discrimination Constrained Feature Selection for Rational Synthesis of Microporous Aluminophosphates.

    PubMed

    Qi, Miao; Wang, Ting; Yi, Yugen; Gao, Na; Kong, Jun; Wang, Jianzhong

    2017-04-01

    Feature selection has been regarded as an effective tool to help researchers understand the generating process of data. For mining the synthesis mechanism of microporous AlPOs, this paper proposes a novel feature selection method by joint l 2,1 norm and Fisher discrimination constraints (JNFDC). In order to obtain more effective feature subset, the proposed method can be achieved in two steps. The first step is to rank the features according to sparse and discriminative constraints. The second step is to establish predictive model with the ranked features, and select the most significant features in the light of the contribution of improving the predictive accuracy. To the best of our knowledge, JNFDC is the first work which employs the sparse representation theory to explore the synthesis mechanism of six kinds of pore rings. Numerical simulations demonstrate that our proposed method can select significant features affecting the specified structural property and improve the predictive accuracy. Moreover, comparison results show that JNFDC can obtain better predictive performances than some other state-of-the-art feature selection methods. © 2017 Wiley-VCH Verlag GmbH & Co. KGaA, Weinheim.

  16. Investigating a memory-based account of negative priming: support for selection-feature mismatch.

    PubMed

    MacDonald, P A; Joordens, S

    2000-08-01

    Using typical and modified negative priming tasks, the selection-feature mismatch account of negative priming was tested. In the modified task, participants performed selections on the basis of a semantic feature (e.g., referent size). This procedure has been shown to enhance negative priming (P. A. MacDonald, S. Joordens, & K. N. Seergobin, 1999). Across 3 experiments, negative priming occurred only when the repeated item mismatched in terms of the feature used as the basis for selections. When the repeated item was congruent on the selection feature across the prime and probe displays, positive priming arose. This pattern of results appeared in both the ignored- and the attended-repetition conditions. Negative priming does not result from previously ignoring an item. These findings strongly support the selection-feature mismatch account of negative priming and refute both the distractor inhibition and the episodic-retrieval explanations.

  17. Ordinal feature selection for iris and palmprint recognition.

    PubMed

    Sun, Zhenan; Wang, Libin; Tan, Tieniu

    2014-09-01

    Ordinal measures have been demonstrated as an effective feature representation model for iris and palmprint recognition. However, ordinal measures are a general concept of image analysis and numerous variants with different parameter settings, such as location, scale, orientation, and so on, can be derived to construct a huge feature space. This paper proposes a novel optimization formulation for ordinal feature selection with successful applications to both iris and palmprint recognition. The objective function of the proposed feature selection method has two parts, i.e., misclassification error of intra and interclass matching samples and weighted sparsity of ordinal feature descriptors. Therefore, the feature selection aims to achieve an accurate and sparse representation of ordinal measures. And, the optimization subjects to a number of linear inequality constraints, which require that all intra and interclass matching pairs are well separated with a large margin. Ordinal feature selection is formulated as a linear programming (LP) problem so that a solution can be efficiently obtained even on a large-scale feature pool and training database. Extensive experimental results demonstrate that the proposed LP formulation is advantageous over existing feature selection methods, such as mRMR, ReliefF, Boosting, and Lasso for biometric recognition, reporting state-of-the-art accuracy on CASIA and PolyU databases.

  18. Economic indicators selection for crime rates forecasting using cooperative feature selection

    NASA Astrophysics Data System (ADS)

    Alwee, Razana; Shamsuddin, Siti Mariyam Hj; Salleh Sallehuddin, Roselina

    2013-04-01

    Features selection in multivariate forecasting model is very important to ensure that the model is accurate. The purpose of this study is to apply the Cooperative Feature Selection method for features selection. The features are economic indicators that will be used in crime rate forecasting model. The Cooperative Feature Selection combines grey relational analysis and artificial neural network to establish a cooperative model that can rank and select the significant economic indicators. Grey relational analysis is used to select the best data series to represent each economic indicator and is also used to rank the economic indicators according to its importance to the crime rate. After that, the artificial neural network is used to select the significant economic indicators for forecasting the crime rates. In this study, we used economic indicators of unemployment rate, consumer price index, gross domestic product and consumer sentiment index, as well as data rates of property crime and violent crime for the United States. Levenberg-Marquardt neural network is used in this study. From our experiments, we found that consumer price index is an important economic indicator that has a significant influence on the violent crime rate. While for property crime rate, the gross domestic product, unemployment rate and consumer price index are the influential economic indicators. The Cooperative Feature Selection is also found to produce smaller errors as compared to Multiple Linear Regression in forecasting property and violent crime rates.

  19. Feature Selection for Ridge Regression with Provable Guarantees.

    PubMed

    Paul, Saurabh; Drineas, Petros

    2016-04-01

    We introduce single-set spectral sparsification as a deterministic sampling-based feature selection technique for regularized least-squares classification, which is the classification analog to ridge regression. The method is unsupervised and gives worst-case guarantees of the generalization power of the classification function after feature selection with respect to the classification function obtained using all features. We also introduce leverage-score sampling as an unsupervised randomized feature selection method for ridge regression. We provide risk bounds for both single-set spectral sparsification and leverage-score sampling on ridge regression in the fixed design setting and show that the risk in the sampled space is comparable to the risk in the full-feature space. We perform experiments on synthetic and real-world data sets; a subset of TechTC-300 data sets, to support our theory. Experimental results indicate that the proposed methods perform better than the existing feature selection methods.

  20. Improving Classification of Protein Interaction Articles Using Context Similarity-Based Feature Selection.

    PubMed

    Chen, Yifei; Sun, Yuxing; Han, Bing-Qing

    2015-01-01

    Protein interaction article classification is a text classification task in the biological domain to determine which articles describe protein-protein interactions. Since the feature space in text classification is high-dimensional, feature selection is widely used for reducing the dimensionality of features to speed up computation without sacrificing classification performance. Many existing feature selection methods are based on the statistical measure of document frequency and term frequency. One potential drawback of these methods is that they treat features separately. Hence, first we design a similarity measure between the context information to take word cooccurrences and phrase chunks around the features into account. Then we introduce the similarity of context information to the importance measure of the features to substitute the document and term frequency. Hence we propose new context similarity-based feature selection methods. Their performance is evaluated on two protein interaction article collections and compared against the frequency-based methods. The experimental results reveal that the context similarity-based methods perform better in terms of the F1 measure and the dimension reduction rate. Benefiting from the context information surrounding the features, the proposed methods can select distinctive features effectively for protein interaction article classification.

  1. Feature Selection Using Information Gain for Improved Structural-Based Alert Correlation

    PubMed Central

    Siraj, Maheyzah Md; Zainal, Anazida; Elshoush, Huwaida Tagelsir; Elhaj, Fatin

    2016-01-01

    Grouping and clustering alerts for intrusion detection based on the similarity of features is referred to as structurally base alert correlation and can discover a list of attack steps. Previous researchers selected different features and data sources manually based on their knowledge and experience, which lead to the less accurate identification of attack steps and inconsistent performance of clustering accuracy. Furthermore, the existing alert correlation systems deal with a huge amount of data that contains null values, incomplete information, and irrelevant features causing the analysis of the alerts to be tedious, time-consuming and error-prone. Therefore, this paper focuses on selecting accurate and significant features of alerts that are appropriate to represent the attack steps, thus, enhancing the structural-based alert correlation model. A two-tier feature selection method is proposed to obtain the significant features. The first tier aims at ranking the subset of features based on high information gain entropy in decreasing order. The‏ second tier extends additional features with a better discriminative ability than the initially ranked features. Performance analysis results show the significance of the selected features in terms of the clustering accuracy using 2000 DARPA intrusion detection scenario-specific dataset. PMID:27893821

  2. A New Direction of Cancer Classification: Positive Effect of Low-Ranking MicroRNAs.

    PubMed

    Li, Feifei; Piao, Minghao; Piao, Yongjun; Li, Meijing; Ryu, Keun Ho

    2014-10-01

    Many studies based on microRNA (miRNA) expression profiles showed a new aspect of cancer classification. Because one characteristic of miRNA expression data is the high dimensionality, feature selection methods have been used to facilitate dimensionality reduction. The feature selection methods have one shortcoming thus far: they just consider the problem of where feature to class is 1:1 or n:1. However, because one miRNA may influence more than one type of cancer, human miRNA is considered to be ranked low in traditional feature selection methods and are removed most of the time. In view of the limitation of the miRNA number, low-ranking miRNAs are also important to cancer classification. We considered both high- and low-ranking features to cover all problems (1:1, n:1, 1:n, and m:n) in cancer classification. First, we used the correlation-based feature selection method to select the high-ranking miRNAs, and chose the support vector machine, Bayes network, decision tree, k-nearest-neighbor, and logistic classifier to construct cancer classification. Then, we chose Chi-square test, information gain, gain ratio, and Pearson's correlation feature selection methods to build the m:n feature subset, and used the selected miRNAs to determine cancer classification. The low-ranking miRNA expression profiles achieved higher classification accuracy compared with just using high-ranking miRNAs in traditional feature selection methods. Our results demonstrate that the m:n feature subset made a positive impression of low-ranking miRNAs in cancer classification.

  3. Cushing's syndrome: Stepwise approach to diagnosis

    PubMed Central

    Lila, Anurag R.; Sarathi, Vijaya; Jagtap, Varsha S.; Bandgar, Tushar; Menon, Padmavathy; Shah, Nalini S.

    2011-01-01

    The projected prevalence of Cushing's syndrome (CS) inclusive of subclinical cases in the adult population ranges from 0.2–2% and it may no longer be considered as an orphan disease (2–3 cases/million/year). The recognition of CS by physicians is important for early diagnosis and treatment. Late-night salivary cortisol, dexamethasone suppressiontesti, or 24-h urine free cortisol are good screening tests. Positively screened cases need stepwise evaluation by an endocrinologist. This paper discusses the importance of screening for CS and suggests a stepwise diagnostic approach to a case of suspected hypercortisolism. PMID:22145134

  4. Stepwise pH-responsive nanoparticles for enhanced cellular uptake and on-demand intracellular release of doxorubicin.

    PubMed

    Chen, Wei-Liang; Li, Fang; Tang, Yan; Yang, Shu-di; Li, Ji-Zhao; Yuan, Zhi-Qiang; Liu, Yang; Zhou, Xiao-Feng; Liu, Chun; Zhang, Xue-Nong

    2017-01-01

    Physicochemical properties, including particle size, zeta potential, and drug release behavior, affect targeting efficiency, cellular uptake, and antitumor effect of nanocarriers in a formulated drug-delivery system. In this study, a novel stepwise pH-responsive nanodrug delivery system was developed to efficiently deliver and significantly promote the therapeutic effect of doxorubicin (DOX). The system comprised dimethylmaleic acid-chitosan-urocanic acid and elicited stepwise responses to extracellular and intracellular pH. The nanoparticles (NPs), which possessed negative surface charge under physiological conditions and an appropriate nanosize, exhibited advantageous stability during blood circulation and enhanced accumulation in tumor sites via enhanced permeability and retention effect. The tumor cellular uptake of DOX-loaded NPs was significantly promoted by the first-step pH response, wherein surface charge reversion of NPs from negative to positive was triggered by the slightly acidic tumor extracellular environment. After internalization into tumor cells, the second-step pH response in endo/lysosome acidic environment elicited the on-demand intracellular release of DOX from NPs, thereby increasing cytotoxicity against tumor cells. Furthermore, stepwise pH-responsive NPs showed enhanced antiproliferation effect and reduced systemic side effect in vivo. Hence, the stepwise pH-responsive NPs provide a promising strategy for efficient delivery of antitumor agents.

  5. Stepwise pH-responsive nanoparticles for enhanced cellular uptake and on-demand intracellular release of doxorubicin

    PubMed Central

    Chen, Wei-liang; Li, Fang; Tang, Yan; Yang, Shu-di; Li, Ji-zhao; Yuan, Zhi-qiang; Liu, Yang; Zhou, Xiao-feng; Liu, Chun; Zhang, Xue-nong

    2017-01-01

    Physicochemical properties, including particle size, zeta potential, and drug release behavior, affect targeting efficiency, cellular uptake, and antitumor effect of nanocarriers in a formulated drug-delivery system. In this study, a novel stepwise pH-responsive nanodrug delivery system was developed to efficiently deliver and significantly promote the therapeutic effect of doxorubicin (DOX). The system comprised dimethylmaleic acid-chitosan-urocanic acid and elicited stepwise responses to extracellular and intracellular pH. The nanoparticles (NPs), which possessed negative surface charge under physiological conditions and an appropriate nanosize, exhibited advantageous stability during blood circulation and enhanced accumulation in tumor sites via enhanced permeability and retention effect. The tumor cellular uptake of DOX-loaded NPs was significantly promoted by the first-step pH response, wherein surface charge reversion of NPs from negative to positive was triggered by the slightly acidic tumor extracellular environment. After internalization into tumor cells, the second-step pH response in endo/lysosome acidic environment elicited the on-demand intracellular release of DOX from NPs, thereby increasing cytotoxicity against tumor cells. Furthermore, stepwise pH-responsive NPs showed enhanced antiproliferation effect and reduced systemic side effect in vivo. Hence, the stepwise pH-responsive NPs provide a promising strategy for efficient delivery of antitumor agents. PMID:28652730

  6. Artificial bee colony algorithm for single-trial electroencephalogram analysis.

    PubMed

    Hsu, Wei-Yen; Hu, Ya-Ping

    2015-04-01

    In this study, we propose an analysis system combined with feature selection to further improve the classification accuracy of single-trial electroencephalogram (EEG) data. Acquiring event-related brain potential data from the sensorimotor cortices, the system comprises artifact and background noise removal, feature extraction, feature selection, and feature classification. First, the artifacts and background noise are removed automatically by means of independent component analysis and surface Laplacian filter, respectively. Several potential features, such as band power, autoregressive model, and coherence and phase-locking value, are then extracted for subsequent classification. Next, artificial bee colony (ABC) algorithm is used to select features from the aforementioned feature combination. Finally, selected subfeatures are classified by support vector machine. Comparing with and without artifact removal and feature selection, using a genetic algorithm on single-trial EEG data for 6 subjects, the results indicate that the proposed system is promising and suitable for brain-computer interface applications. © EEG and Clinical Neuroscience Society (ECNS) 2014.

  7. Application of quantum-behaved particle swarm optimization to motor imagery EEG classification.

    PubMed

    Hsu, Wei-Yen

    2013-12-01

    In this study, we propose a recognition system for single-trial analysis of motor imagery (MI) electroencephalogram (EEG) data. Applying event-related brain potential (ERP) data acquired from the sensorimotor cortices, the system chiefly consists of automatic artifact elimination, feature extraction, feature selection and classification. In addition to the use of independent component analysis, a similarity measure is proposed to further remove the electrooculographic (EOG) artifacts automatically. Several potential features, such as wavelet-fractal features, are then extracted for subsequent classification. Next, quantum-behaved particle swarm optimization (QPSO) is used to select features from the feature combination. Finally, selected sub-features are classified by support vector machine (SVM). Compared with without artifact elimination, feature selection using a genetic algorithm (GA) and feature classification with Fisher's linear discriminant (FLD) on MI data from two data sets for eight subjects, the results indicate that the proposed method is promising in brain-computer interface (BCI) applications.

  8. K/T age for the popigai impact event

    NASA Technical Reports Server (NTRS)

    Deino, A. L.; Garvin, J. B.; Montanari, S.

    1991-01-01

    The multi-ringed POPIGAI structure, with an outer ring diameter of over 100 km, is the largest impact feature currently recognized on Earth with an Phanerozoic age. The target rocks in this relatively unglaciated region consist of upper Proterozoic through Mesozoic platform sediments and igneous rocks overlying Precambrian crystalline basement. The reported absolute age of the Popigai impact event ranges from 30.5 to 39 Ma. With the intent of refining this age estimate, a melt-breccia (suevite) sample from the inner regions of the Popigai structure was prepared for total fusion and step-wise heating Ar-40/Ar-39 analysis. Although the total fusion and step-heating experiments suggest some degree of age heterogeneity, the recurring theme is an age of around 64 to 66 Ma.

  9. Improved Hierarchical Optimization-Based Classification of Hyperspectral Images Using Shape Analysis

    NASA Technical Reports Server (NTRS)

    Tarabalka, Yuliya; Tilton, James C.

    2012-01-01

    A new spectral-spatial method for classification of hyperspectral images is proposed. The HSegClas method is based on the integration of probabilistic classification and shape analysis within the hierarchical step-wise optimization algorithm. First, probabilistic support vector machines classification is applied. Then, at each iteration two neighboring regions with the smallest Dissimilarity Criterion (DC) are merged, and classification probabilities are recomputed. The important contribution of this work consists in estimating a DC between regions as a function of statistical, classification and geometrical (area and rectangularity) features. Experimental results are presented on a 102-band ROSIS image of the Center of Pavia, Italy. The developed approach yields more accurate classification results when compared to previously proposed methods.

  10. Programmable DNA-Mediated Multitasking Processor.

    PubMed

    Shu, Jian-Jun; Wang, Qi-Wen; Yong, Kian-Yan; Shao, Fangwei; Lee, Kee Jin

    2015-04-30

    Because of DNA appealing features as perfect material, including minuscule size, defined structural repeat and rigidity, programmable DNA-mediated processing is a promising computing paradigm, which employs DNAs as information storing and processing substrates to tackle the computational problems. The massive parallelism of DNA hybridization exhibits transcendent potential to improve multitasking capabilities and yield a tremendous speed-up over the conventional electronic processors with stepwise signal cascade. As an example of multitasking capability, we present an in vitro programmable DNA-mediated optimal route planning processor as a functional unit embedded in contemporary navigation systems. The novel programmable DNA-mediated processor has several advantages over the existing silicon-mediated methods, such as conducting massive data storage and simultaneous processing via much fewer materials than conventional silicon devices.

  11. Combining living anionic polymerization with branching reactions in an iterative fashion to design branched polymers.

    PubMed

    Higashihara, Tomoya; Sugiyama, Kenji; Yoo, Hee-Soo; Hayashi, Mayumi; Hirao, Akira

    2010-06-16

    This paper reviews the precise synthesis of many-armed and multi-compositional star-branched polymers, exact graft (co)polymers, and structurally well-defined dendrimer-like star-branched polymers, which are synthetically difficult, by a commonly-featured iterative methodology combining living anionic polymerization with branched reactions to design branched polymers. The methodology basically involves only two synthetic steps; (a) preparation of a polymeric building block corresponding to each branched polymer and (b) connection of the resulting building unit to another unit. The synthetic steps were repeated in a stepwise fashion several times to successively synthesize a series of well-defined target branched polymers. Copyright © 2010 WILEY-VCH Verlag GmbH & Co. KGaA, Weinheim.

  12. Bootstrap investigation of the stability of a Cox regression model.

    PubMed

    Altman, D G; Andersen, P K

    1989-07-01

    We describe a bootstrap investigation of the stability of a Cox proportional hazards regression model resulting from the analysis of a clinical trial of azathioprine versus placebo in patients with primary biliary cirrhosis. We have considered stability to refer both to the choice of variables included in the model and, more importantly, to the predictive ability of the model. In stepwise Cox regression analyses of 100 bootstrap samples using 17 candidate variables, the most frequently selected variables were those selected in the original analysis, and no other important variable was identified. Thus there was no reason to doubt the model obtained in the original analysis. For each patient in the trial, bootstrap confidence intervals were constructed for the estimated probability of surviving two years. It is shown graphically that these intervals are markedly wider than those obtained from the original model.

  13. Multivariate calibration on NIR data: development of a model for the rapid evaluation of ethanol content in bakery products.

    PubMed

    Bello, Alessandra; Bianchi, Federica; Careri, Maria; Giannetto, Marco; Mori, Giovanni; Musci, Marilena

    2007-11-05

    A new NIR method based on multivariate calibration for determination of ethanol in industrially packed wholemeal bread was developed and validated. GC-FID was used as reference method for the determination of actual ethanol concentration of different samples of wholemeal bread with proper content of added ethanol, ranging from 0 to 3.5% (w/w). Stepwise discriminant analysis was carried out on the NIR dataset, in order to reduce the number of original variables by selecting those that were able to discriminate between the samples of different ethanol concentrations. With the so selected variables a multivariate calibration model was then obtained by multiple linear regression. The prediction power of the linear model was optimized by a new "leave one out" method, so that the number of original variables resulted further reduced.

  14. Optimum location of external markers using feature selection algorithms for real‐time tumor tracking in external‐beam radiotherapy: a virtual phantom study

    PubMed Central

    Nankali, Saber; Miandoab, Payam Samadi; Baghizadeh, Amin

    2016-01-01

    In external‐beam radiotherapy, using external markers is one of the most reliable tools to predict tumor position, in clinical applications. The main challenge in this approach is tumor motion tracking with highest accuracy that depends heavily on external markers location, and this issue is the objective of this study. Four commercially available feature selection algorithms entitled 1) Correlation‐based Feature Selection, 2) Classifier, 3) Principal Components, and 4) Relief were proposed to find optimum location of external markers in combination with two “Genetic” and “Ranker” searching procedures. The performance of these algorithms has been evaluated using four‐dimensional extended cardiac‐torso anthropomorphic phantom. Six tumors in lung, three tumors in liver, and 49 points on the thorax surface were taken into account to simulate internal and external motions, respectively. The root mean square error of an adaptive neuro‐fuzzy inference system (ANFIS) as prediction model was considered as metric for quantitatively evaluating the performance of proposed feature selection algorithms. To do this, the thorax surface region was divided into nine smaller segments and predefined tumors motion was predicted by ANFIS using external motion data of given markers at each small segment, separately. Our comparative results showed that all feature selection algorithms can reasonably select specific external markers from those segments where the root mean square error of the ANFIS model is minimum. Moreover, the performance accuracy of proposed feature selection algorithms was compared, separately. For this, each tumor motion was predicted using motion data of those external markers selected by each feature selection algorithm. Duncan statistical test, followed by F‐test, on final results reflected that all proposed feature selection algorithms have the same performance accuracy for lung tumors. But for liver tumors, a correlation‐based feature selection algorithm, in combination with a genetic search algorithm, proved to yield best performance accuracy for selecting optimum markers. PACS numbers: 87.55.km, 87.56.Fc PMID:26894358

  15. Optimum location of external markers using feature selection algorithms for real-time tumor tracking in external-beam radiotherapy: a virtual phantom study.

    PubMed

    Nankali, Saber; Torshabi, Ahmad Esmaili; Miandoab, Payam Samadi; Baghizadeh, Amin

    2016-01-08

    In external-beam radiotherapy, using external markers is one of the most reliable tools to predict tumor position, in clinical applications. The main challenge in this approach is tumor motion tracking with highest accuracy that depends heavily on external markers location, and this issue is the objective of this study. Four commercially available feature selection algorithms entitled 1) Correlation-based Feature Selection, 2) Classifier, 3) Principal Components, and 4) Relief were proposed to find optimum location of external markers in combination with two "Genetic" and "Ranker" searching procedures. The performance of these algorithms has been evaluated using four-dimensional extended cardiac-torso anthropomorphic phantom. Six tumors in lung, three tumors in liver, and 49 points on the thorax surface were taken into account to simulate internal and external motions, respectively. The root mean square error of an adaptive neuro-fuzzy inference system (ANFIS) as prediction model was considered as metric for quantitatively evaluating the performance of proposed feature selection algorithms. To do this, the thorax surface region was divided into nine smaller segments and predefined tumors motion was predicted by ANFIS using external motion data of given markers at each small segment, separately. Our comparative results showed that all feature selection algorithms can reasonably select specific external markers from those segments where the root mean square error of the ANFIS model is minimum. Moreover, the performance accuracy of proposed feature selection algorithms was compared, separately. For this, each tumor motion was predicted using motion data of those external markers selected by each feature selection algorithm. Duncan statistical test, followed by F-test, on final results reflected that all proposed feature selection algorithms have the same performance accuracy for lung tumors. But for liver tumors, a correlation-based feature selection algorithm, in combination with a genetic search algorithm, proved to yield best performance accuracy for selecting optimum markers.

  16. Planning for robust reserve networks using uncertainty analysis

    USGS Publications Warehouse

    Moilanen, A.; Runge, M.C.; Elith, Jane; Tyre, A.; Carmel, Y.; Fegraus, E.; Wintle, B.A.; Burgman, M.; Ben-Haim, Y.

    2006-01-01

    Planning land-use for biodiversity conservation frequently involves computer-assisted reserve selection algorithms. Typically such algorithms operate on matrices of species presence?absence in sites, or on species-specific distributions of model predicted probabilities of occurrence in grid cells. There are practically always errors in input data?erroneous species presence?absence data, structural and parametric uncertainty in predictive habitat models, and lack of correspondence between temporal presence and long-run persistence. Despite these uncertainties, typical reserve selection methods proceed as if there is no uncertainty in the data or models. Having two conservation options of apparently equal biological value, one would prefer the option whose value is relatively insensitive to errors in planning inputs. In this work we show how uncertainty analysis for reserve planning can be implemented within a framework of information-gap decision theory, generating reserve designs that are robust to uncertainty. Consideration of uncertainty involves modifications to the typical objective functions used in reserve selection. Search for robust-optimal reserve structures can still be implemented via typical reserve selection optimization techniques, including stepwise heuristics, integer-programming and stochastic global search.

  17. Automatic migraine classification via feature selection committee and machine learning techniques over imaging and questionnaire data.

    PubMed

    Garcia-Chimeno, Yolanda; Garcia-Zapirain, Begonya; Gomez-Beldarrain, Marian; Fernandez-Ruanova, Begonya; Garcia-Monco, Juan Carlos

    2017-04-13

    Feature selection methods are commonly used to identify subsets of relevant features to facilitate the construction of models for classification, yet little is known about how feature selection methods perform in diffusion tensor images (DTIs). In this study, feature selection and machine learning classification methods were tested for the purpose of automating diagnosis of migraines using both DTIs and questionnaire answers related to emotion and cognition - factors that influence of pain perceptions. We select 52 adult subjects for the study divided into three groups: control group (15), subjects with sporadic migraine (19) and subjects with chronic migraine and medication overuse (18). These subjects underwent magnetic resonance with diffusion tensor to see white matter pathway integrity of the regions of interest involved in pain and emotion. The tests also gather data about pathology. The DTI images and test results were then introduced into feature selection algorithms (Gradient Tree Boosting, L1-based, Random Forest and Univariate) to reduce features of the first dataset and classification algorithms (SVM (Support Vector Machine), Boosting (Adaboost) and Naive Bayes) to perform a classification of migraine group. Moreover we implement a committee method to improve the classification accuracy based on feature selection algorithms. When classifying the migraine group, the greatest improvements in accuracy were made using the proposed committee-based feature selection method. Using this approach, the accuracy of classification into three types improved from 67 to 93% when using the Naive Bayes classifier, from 90 to 95% with the support vector machine classifier, 93 to 94% in boosting. The features that were determined to be most useful for classification included are related with the pain, analgesics and left uncinate brain (connected with the pain and emotions). The proposed feature selection committee method improved the performance of migraine diagnosis classifiers compared to individual feature selection methods, producing a robust system that achieved over 90% accuracy in all classifiers. The results suggest that the proposed methods can be used to support specialists in the classification of migraines in patients undergoing magnetic resonance imaging.

  18. Joint Feature Selection and Classification for Multilabel Learning.

    PubMed

    Huang, Jun; Li, Guorong; Huang, Qingming; Wu, Xindong

    2018-03-01

    Multilabel learning deals with examples having multiple class labels simultaneously. It has been applied to a variety of applications, such as text categorization and image annotation. A large number of algorithms have been proposed for multilabel learning, most of which concentrate on multilabel classification problems and only a few of them are feature selection algorithms. Current multilabel classification models are mainly built on a single data representation composed of all the features which are shared by all the class labels. Since each class label might be decided by some specific features of its own, and the problems of classification and feature selection are often addressed independently, in this paper, we propose a novel method which can perform joint feature selection and classification for multilabel learning, named JFSC. Different from many existing methods, JFSC learns both shared features and label-specific features by considering pairwise label correlations, and builds the multilabel classifier on the learned low-dimensional data representations simultaneously. A comparative study with state-of-the-art approaches manifests a competitive performance of our proposed method both in classification and feature selection for multilabel learning.

  19. A hybrid feature selection method using multiclass SVM for diagnosis of erythemato-squamous disease

    NASA Astrophysics Data System (ADS)

    Maryam, Setiawan, Noor Akhmad; Wahyunggoro, Oyas

    2017-08-01

    The diagnosis of erythemato-squamous disease is a complex problem and difficult to detect in dermatology. Besides that, it is a major cause of skin cancer. Data mining implementation in the medical field helps expert to diagnose precisely, accurately, and inexpensively. In this research, we use data mining technique to developed a diagnosis model based on multiclass SVM with a novel hybrid feature selection method to diagnose erythemato-squamous disease. Our hybrid feature selection method, named ChiGA (Chi Square and Genetic Algorithm), uses the advantages from filter and wrapper methods to select the optimal feature subset from original feature. Chi square used as filter method to remove redundant features and GA as wrapper method to select the ideal feature subset with SVM used as classifier. Experiment performed with 10 fold cross validation on erythemato-squamous diseases dataset taken from University of California Irvine (UCI) machine learning database. The experimental result shows that the proposed model based multiclass SVM with Chi Square and GA can give an optimum feature subset. There are 18 optimum features with 99.18% accuracy.

  20. Selective attention to temporal features on nested time scales.

    PubMed

    Henry, Molly J; Herrmann, Björn; Obleser, Jonas

    2015-02-01

    Meaningful auditory stimuli such as speech and music often vary simultaneously along multiple time scales. Thus, listeners must selectively attend to, and selectively ignore, separate but intertwined temporal features. The current study aimed to identify and characterize the neural network specifically involved in this feature-selective attention to time. We used a novel paradigm where listeners judged either the duration or modulation rate of auditory stimuli, and in which the stimulation, working memory demands, response requirements, and task difficulty were held constant. A first analysis identified all brain regions where individual brain activation patterns were correlated with individual behavioral performance patterns, which thus supported temporal judgments generically. A second analysis then isolated those brain regions that specifically regulated selective attention to temporal features: Neural responses in a bilateral fronto-parietal network including insular cortex and basal ganglia decreased with degree of change of the attended temporal feature. Critically, response patterns in these regions were inverted when the task required selectively ignoring this feature. The results demonstrate how the neural analysis of complex acoustic stimuli with multiple temporal features depends on a fronto-parietal network that simultaneously regulates the selective gain for attended and ignored temporal features. © The Author 2013. Published by Oxford University Press. All rights reserved. For Permissions, please e-mail: journals.permissions@oup.com.

  1. A combinatorial feature selection approach to describe the QSAR of dual site inhibitors of acetylcholinesterase.

    PubMed

    Asadabadi, Ebrahim Barzegari; Abdolmaleki, Parviz; Barkooie, Seyyed Mohsen Hosseini; Jahandideh, Samad; Rezaei, Mohammad Ali

    2009-12-01

    Regarding the great potential of dual binding site inhibitors of acetylcholinesterase as the future potent drugs of Alzheimer's disease, this study was devoted to extraction of the most effective structural features of these inhibitors from among a large number of quantitative descriptors. To do this, we adopted a unique approach in quantitative structure-activity relationships. An efficient feature selection method was emphasized in such an approach, using the confirmative results of different routine and novel feature selection methods. The proposed methods generated quite consistent results ensuring the effectiveness of the selected structural features.

  2. A combined Fisher and Laplacian score for feature selection in QSAR based drug design using compounds with known and unknown activities.

    PubMed

    Valizade Hasanloei, Mohammad Amin; Sheikhpour, Razieh; Sarram, Mehdi Agha; Sheikhpour, Elnaz; Sharifi, Hamdollah

    2018-02-01

    Quantitative structure-activity relationship (QSAR) is an effective computational technique for drug design that relates the chemical structures of compounds to their biological activities. Feature selection is an important step in QSAR based drug design to select the most relevant descriptors. One of the most popular feature selection methods for classification problems is Fisher score which aim is to minimize the within-class distance and maximize the between-class distance. In this study, the properties of Fisher criterion were extended for QSAR models to define the new distance metrics based on the continuous activity values of compounds with known activities. Then, a semi-supervised feature selection method was proposed based on the combination of Fisher and Laplacian criteria which exploits both compounds with known and unknown activities to select the relevant descriptors. To demonstrate the efficiency of the proposed semi-supervised feature selection method in selecting the relevant descriptors, we applied the method and other feature selection methods on three QSAR data sets such as serine/threonine-protein kinase PLK3 inhibitors, ROCK inhibitors and phenol compounds. The results demonstrated that the QSAR models built on the selected descriptors by the proposed semi-supervised method have better performance than other models. This indicates the efficiency of the proposed method in selecting the relevant descriptors using the compounds with known and unknown activities. The results of this study showed that the compounds with known and unknown activities can be helpful to improve the performance of the combined Fisher and Laplacian based feature selection methods.

  3. A combined Fisher and Laplacian score for feature selection in QSAR based drug design using compounds with known and unknown activities

    NASA Astrophysics Data System (ADS)

    Valizade Hasanloei, Mohammad Amin; Sheikhpour, Razieh; Sarram, Mehdi Agha; Sheikhpour, Elnaz; Sharifi, Hamdollah

    2018-02-01

    Quantitative structure-activity relationship (QSAR) is an effective computational technique for drug design that relates the chemical structures of compounds to their biological activities. Feature selection is an important step in QSAR based drug design to select the most relevant descriptors. One of the most popular feature selection methods for classification problems is Fisher score which aim is to minimize the within-class distance and maximize the between-class distance. In this study, the properties of Fisher criterion were extended for QSAR models to define the new distance metrics based on the continuous activity values of compounds with known activities. Then, a semi-supervised feature selection method was proposed based on the combination of Fisher and Laplacian criteria which exploits both compounds with known and unknown activities to select the relevant descriptors. To demonstrate the efficiency of the proposed semi-supervised feature selection method in selecting the relevant descriptors, we applied the method and other feature selection methods on three QSAR data sets such as serine/threonine-protein kinase PLK3 inhibitors, ROCK inhibitors and phenol compounds. The results demonstrated that the QSAR models built on the selected descriptors by the proposed semi-supervised method have better performance than other models. This indicates the efficiency of the proposed method in selecting the relevant descriptors using the compounds with known and unknown activities. The results of this study showed that the compounds with known and unknown activities can be helpful to improve the performance of the combined Fisher and Laplacian based feature selection methods.

  4. Irrelevant reward and selection histories have different influences on task-relevant attentional selection.

    PubMed

    MacLean, Mary H; Giesbrecht, Barry

    2015-07-01

    Task-relevant and physically salient features influence visual selective attention. In the present study, we investigated the influence of task-irrelevant and physically nonsalient reward-associated features on visual selective attention. Two hypotheses were tested: One predicts that the effects of target-defining task-relevant and task-irrelevant features interact to modulate visual selection; the other predicts that visual selection is determined by the independent combination of relevant and irrelevant feature effects. These alternatives were tested using a visual search task that contained multiple targets, placing a high demand on the need for selectivity, and that was data-limited and required unspeeded responses, emphasizing early perceptual selection processes. One week prior to the visual search task, participants completed a training task in which they learned to associate particular colors with a specific reward value. In the search task, the reward-associated colors were presented surrounding targets and distractors, but were neither physically salient nor task-relevant. In two experiments, the irrelevant reward-associated features influenced performance, but only when they were presented in a task-relevant location. The costs induced by the irrelevant reward-associated features were greater when they oriented attention to a target than to a distractor. In a third experiment, we examined the effects of selection history in the absence of reward history and found that the interaction between task relevance and selection history differed, relative to when the features had previously been associated with reward. The results indicate that under conditions that demand highly efficient perceptual selection, physically nonsalient task-irrelevant and task-relevant factors interact to influence visual selective attention.

  5. Multi-level gene/MiRNA feature selection using deep belief nets and active learning.

    PubMed

    Ibrahim, Rania; Yousri, Noha A; Ismail, Mohamed A; El-Makky, Nagwa M

    2014-01-01

    Selecting the most discriminative genes/miRNAs has been raised as an important task in bioinformatics to enhance disease classifiers and to mitigate the dimensionality curse problem. Original feature selection methods choose genes/miRNAs based on their individual features regardless of how they perform together. Considering group features instead of individual ones provides a better view for selecting the most informative genes/miRNAs. Recently, deep learning has proven its ability in representing the data in multiple levels of abstraction, allowing for better discrimination between different classes. However, the idea of using deep learning for feature selection is not widely used in the bioinformatics field yet. In this paper, a novel multi-level feature selection approach named MLFS is proposed for selecting genes/miRNAs based on expression profiles. The approach is based on both deep and active learning. Moreover, an extension to use the technique for miRNAs is presented by considering the biological relation between miRNAs and genes. Experimental results show that the approach was able to outperform classical feature selection methods in hepatocellular carcinoma (HCC) by 9%, lung cancer by 6% and breast cancer by around 10% in F1-measure. Results also show the enhancement in F1-measure of our approach over recently related work in [1] and [2].

  6. A Hierarchical Feature and Sample Selection Framework and Its Application for Alzheimer’s Disease Diagnosis

    NASA Astrophysics Data System (ADS)

    An, Le; Adeli, Ehsan; Liu, Mingxia; Zhang, Jun; Lee, Seong-Whan; Shen, Dinggang

    2017-03-01

    Classification is one of the most important tasks in machine learning. Due to feature redundancy or outliers in samples, using all available data for training a classifier may be suboptimal. For example, the Alzheimer’s disease (AD) is correlated with certain brain regions or single nucleotide polymorphisms (SNPs), and identification of relevant features is critical for computer-aided diagnosis. Many existing methods first select features from structural magnetic resonance imaging (MRI) or SNPs and then use those features to build the classifier. However, with the presence of many redundant features, the most discriminative features are difficult to be identified in a single step. Thus, we formulate a hierarchical feature and sample selection framework to gradually select informative features and discard ambiguous samples in multiple steps for improved classifier learning. To positively guide the data manifold preservation process, we utilize both labeled and unlabeled data during training, making our method semi-supervised. For validation, we conduct experiments on AD diagnosis by selecting mutually informative features from both MRI and SNP, and using the most discriminative samples for training. The superior classification results demonstrate the effectiveness of our approach, as compared with the rivals.

  7. Inclusion of Endogenous Hormone Levels in Risk Prediction Models of Postmenopausal Breast Cancer

    PubMed Central

    Tworoger, Shelley S.; Zhang, Xuehong; Eliassen, A. Heather; Qian, Jing; Colditz, Graham A.; Willett, Walter C.; Rosner, Bernard A.; Kraft, Peter; Hankinson, Susan E.

    2014-01-01

    Purpose Endogenous hormones are risk factors for postmenopausal breast cancer, and their measurement may improve our ability to identify high-risk women. Therefore, we evaluated whether inclusion of plasma estradiol, estrone, estrone sulfate, testosterone, dehydroepiandrosterone sulfate, prolactin, and sex hormone–binding globulin (SHBG) improved risk prediction for postmenopausal invasive breast cancer (n = 437 patient cases and n = 775 controls not using postmenopausal hormones) in the Nurses' Health Study. Methods We evaluated improvement in the area under the curve (AUC) for 5-year risk of invasive breast cancer by adding each hormone to the Gail and Rosner-Colditz risk scores. We used stepwise regression to identify the subset of hormones most associated with risk and assessed AUC improvement; we used 10-fold cross validation to assess model overfitting. Results Each hormone was associated with breast cancer risk (odds ratio doubling, 0.82 [SHBG] to 1.37 [estrone sulfate]). Individual hormones improved the AUC by 1.3 to 5.2 units relative to the Gail score and 0.3 to 2.9 for the Rosner-Colditz score. Estrone sulfate, testosterone, and prolactin were selected by stepwise regression and increased the AUC by 5.9 units (P = .003) for the Gail score and 3.4 (P = .04) for the Rosner-Colditz score. In cross validation, the average AUC change across the validation data sets was 6.0 (P = .002) and 3.0 units (P = .03), respectively. Similar results were observed for estrogen receptor–positive disease (selected hormones: estrone sulfate, testosterone, prolactin, and SHBG; change in AUC, 8.8 [P < .001] for Gail score and 5.8 [P = .004] for Rosner-Colditz score). Conclusion Our results support that endogenous hormones improve risk prediction for invasive breast cancer and could help identify women who may benefit from chemoprevention or more screening. PMID:25135988

  8. What to do When Scalar Invariance Fails: The Extended Alignment Method for Multi-Group Factor Analysis Comparison of Latent Means Across Many Groups.

    PubMed

    Marsh, Herbert W; Guo, Jiesi; Parker, Philip D; Nagengast, Benjamin; Asparouhov, Tihomir; Muthén, Bengt; Dicke, Theresa

    2017-01-12

    Scalar invariance is an unachievable ideal that in practice can only be approximated; often using potentially questionable approaches such as partial invariance based on a stepwise selection of parameter estimates with large modification indices. Study 1 demonstrates an extension of the power and flexibility of the alignment approach for comparing latent factor means in large-scale studies (30 OECD countries, 8 factors, 44 items, N = 249,840), for which scalar invariance is typically not supported in the traditional confirmatory factor analysis approach to measurement invariance (CFA-MI). Importantly, we introduce an alignment-within-CFA (AwC) approach, transforming alignment from a largely exploratory tool into a confirmatory tool, and enabling analyses that previously have not been possible with alignment (testing the invariance of uniquenesses and factor variances/covariances; multiple-group MIMIC models; contrasts on latent means) and structural equation models more generally. Specifically, it also allowed a comparison of gender differences in a 30-country MIMIC AwC (i.e., a SEM with gender as a covariate) and a 60-group AwC CFA (i.e., 30 countries × 2 genders) analysis. Study 2, a simulation study following up issues raised in Study 1, showed that latent means were more accurately estimated with alignment than with the scalar CFA-MI, and particularly with partial invariance scalar models based on the heavily criticized stepwise selection strategy. In summary, alignment augmented by AwC provides applied researchers from diverse disciplines considerable flexibility to address substantively important issues when the traditional CFA-MI scalar model does not fit the data. (PsycINFO Database Record (c) 2017 APA, all rights reserved).

  9. Updated Value of Service Reliability Estimates for Electric Utility Customers in the United States

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Sullivan, Michael; Schellenberg, Josh; Blundell, Marshall

    2015-01-01

    This report updates the 2009 meta-analysis that provides estimates of the value of service reliability for electricity customers in the United States (U.S.). The meta-dataset now includes 34 different datasets from surveys fielded by 10 different utility companies between 1989 and 2012. Because these studies used nearly identical interruption cost estimation or willingness-to-pay/accept methods, it was possible to integrate their results into a single meta-dataset describing the value of electric service reliability observed in all of them. Once the datasets from the various studies were combined, a two-part regression model was used to estimate customer damage functions that can bemore » generally applied to calculate customer interruption costs per event by season, time of day, day of week, and geographical regions within the U.S. for industrial, commercial, and residential customers. This report focuses on the backwards stepwise selection process that was used to develop the final revised model for all customer classes. Across customer classes, the revised customer interruption cost model has improved significantly because it incorporates more data and does not include the many extraneous variables that were in the original specification from the 2009 meta-analysis. The backwards stepwise selection process led to a more parsimonious model that only included key variables, while still achieving comparable out-of-sample predictive performance. In turn, users of interruption cost estimation tools such as the Interruption Cost Estimate (ICE) Calculator will have less customer characteristics information to provide and the associated inputs page will be far less cumbersome. The upcoming new version of the ICE Calculator is anticipated to be released in 2015.« less

  10. An improved wrapper-based feature selection method for machinery fault diagnosis

    PubMed Central

    2017-01-01

    A major issue of machinery fault diagnosis using vibration signals is that it is over-reliant on personnel knowledge and experience in interpreting the signal. Thus, machine learning has been adapted for machinery fault diagnosis. The quantity and quality of the input features, however, influence the fault classification performance. Feature selection plays a vital role in selecting the most representative feature subset for the machine learning algorithm. In contrast, the trade-off relationship between capability when selecting the best feature subset and computational effort is inevitable in the wrapper-based feature selection (WFS) method. This paper proposes an improved WFS technique before integration with a support vector machine (SVM) model classifier as a complete fault diagnosis system for a rolling element bearing case study. The bearing vibration dataset made available by the Case Western Reserve University Bearing Data Centre was executed using the proposed WFS and its performance has been analysed and discussed. The results reveal that the proposed WFS secures the best feature subset with a lower computational effort by eliminating the redundancy of re-evaluation. The proposed WFS has therefore been found to be capable and efficient to carry out feature selection tasks. PMID:29261689

  11. Asymmetric bagging and feature selection for activities prediction of drug molecules.

    PubMed

    Li, Guo-Zheng; Meng, Hao-Hua; Lu, Wen-Cong; Yang, Jack Y; Yang, Mary Qu

    2008-05-28

    Activities of drug molecules can be predicted by QSAR (quantitative structure activity relationship) models, which overcomes the disadvantages of high cost and long cycle by employing the traditional experimental method. With the fact that the number of drug molecules with positive activity is rather fewer than that of negatives, it is important to predict molecular activities considering such an unbalanced situation. Here, asymmetric bagging and feature selection are introduced into the problem and asymmetric bagging of support vector machines (asBagging) is proposed on predicting drug activities to treat the unbalanced problem. At the same time, the features extracted from the structures of drug molecules affect prediction accuracy of QSAR models. Therefore, a novel algorithm named PRIFEAB is proposed, which applies an embedded feature selection method to remove redundant and irrelevant features for asBagging. Numerical experimental results on a data set of molecular activities show that asBagging improve the AUC and sensitivity values of molecular activities and PRIFEAB with feature selection further helps to improve the prediction ability. Asymmetric bagging can help to improve prediction accuracy of activities of drug molecules, which can be furthermore improved by performing feature selection to select relevant features from the drug molecules data sets.

  12. System Complexity Reduction via Feature Selection

    ERIC Educational Resources Information Center

    Deng, Houtao

    2011-01-01

    This dissertation transforms a set of system complexity reduction problems to feature selection problems. Three systems are considered: classification based on association rules, network structure learning, and time series classification. Furthermore, two variable importance measures are proposed to reduce the feature selection bias in tree…

  13. Classification of motor imagery tasks for BCI with multiresolution analysis and multiobjective feature selection.

    PubMed

    Ortega, Julio; Asensio-Cubero, Javier; Gan, John Q; Ortiz, Andrés

    2016-07-15

    Brain-computer interfacing (BCI) applications based on the classification of electroencephalographic (EEG) signals require solving high-dimensional pattern classification problems with such a relatively small number of training patterns that curse of dimensionality problems usually arise. Multiresolution analysis (MRA) has useful properties for signal analysis in both temporal and spectral analysis, and has been broadly used in the BCI field. However, MRA usually increases the dimensionality of the input data. Therefore, some approaches to feature selection or feature dimensionality reduction should be considered for improving the performance of the MRA based BCI. This paper investigates feature selection in the MRA-based frameworks for BCI. Several wrapper approaches to evolutionary multiobjective feature selection are proposed with different structures of classifiers. They are evaluated by comparing with baseline methods using sparse representation of features or without feature selection. The statistical analysis, by applying the Kolmogorov-Smirnoff and Kruskal-Wallis tests to the means of the Kappa values evaluated by using the test patterns in each approach, has demonstrated some advantages of the proposed approaches. In comparison with the baseline MRA approach used in previous studies, the proposed evolutionary multiobjective feature selection approaches provide similar or even better classification performances, with significant reduction in the number of features that need to be computed.

  14. Feature Selection for Speech Emotion Recognition in Spanish and Basque: On the Use of Machine Learning to Improve Human-Computer Interaction

    PubMed Central

    Arruti, Andoni; Cearreta, Idoia; Álvarez, Aitor; Lazkano, Elena; Sierra, Basilio

    2014-01-01

    Study of emotions in human–computer interaction is a growing research area. This paper shows an attempt to select the most significant features for emotion recognition in spoken Basque and Spanish Languages using different methods for feature selection. RekEmozio database was used as the experimental data set. Several Machine Learning paradigms were used for the emotion classification task. Experiments were executed in three phases, using different sets of features as classification variables in each phase. Moreover, feature subset selection was applied at each phase in order to seek for the most relevant feature subset. The three phases approach was selected to check the validity of the proposed approach. Achieved results show that an instance-based learning algorithm using feature subset selection techniques based on evolutionary algorithms is the best Machine Learning paradigm in automatic emotion recognition, with all different feature sets, obtaining a mean of 80,05% emotion recognition rate in Basque and a 74,82% in Spanish. In order to check the goodness of the proposed process, a greedy searching approach (FSS-Forward) has been applied and a comparison between them is provided. Based on achieved results, a set of most relevant non-speaker dependent features is proposed for both languages and new perspectives are suggested. PMID:25279686

  15. STEP and STEPSPL: Computer programs for aerodynamic model structure determination and parameter estimation

    NASA Technical Reports Server (NTRS)

    Batterson, J. G.

    1986-01-01

    The successful parametric modeling of the aerodynamics for an airplane operating at high angles of attack or sideslip is performed in two phases. First the aerodynamic model structure must be determined and second the associated aerodynamic parameters (stability and control derivatives) must be estimated for that model. The purpose of this paper is to document two versions of a stepwise regression computer program which were developed for the determination of airplane aerodynamic model structure and to provide two examples of their use on computer generated data. References are provided for the application of the programs to real flight data. The two computer programs that are the subject of this report, STEP and STEPSPL, are written in FORTRAN IV (ANSI l966) compatible with a CDC FTN4 compiler. Both programs are adaptations of a standard forward stepwise regression algorithm. The purpose of the adaptation is to facilitate the selection of a adequate mathematical model of the aerodynamic force and moment coefficients of an airplane from flight test data. The major difference between STEP and STEPSPL is in the basis for the model. The basis for the model in STEP is the standard polynomial Taylor's series expansion of the aerodynamic function about some steady-state trim condition. Program STEPSPL utilizes a set of spline basis functions.

  16. Sulfide response analysis for sulfide control using a pS electrode in sulfate reducing bioreactors.

    PubMed

    Villa-Gomez, D K; Cassidy, J; Keesman, K J; Sampaio, R; Lens, P N L

    2014-03-01

    Step changes in the organic loading rate (OLR) through variations in the influent chemical oxygen demand (CODin) concentration or in the hydraulic retention time (HRT) at constant COD/SO4(2-) ratio (0.67) were applied to create sulfide responses for the design of a sulfide control in sulfate reducing bioreactors. The sulfide was measured using a sulfide ion selective electrode (pS) and the values obtained were used to calculate proportional-integral-derivative (PID) controller parameters. The experiments were performed in an inverse fluidized bed bioreactor with automated operation using the LabVIEW software version 2009(®). A rapid response and high sulfide increment was obtained through a stepwise increase in the CODin concentration, while a stepwise decrease to the HRT exhibited a slower response with smaller sulfide increment. Irrespective of the way the OLR was decreased, the pS response showed a time-varying behavior due to sulfide accumulation (HRT change) or utilization of substrate sources that were not accounted for (CODin change). The pS electrode response, however, showed to be informative for applications in sulfate reducing bioreactors. Nevertheless, the recorded pS values need to be corrected for pH variations and high sulfide concentrations (>200 mg/L). Copyright © 2013 Elsevier Ltd. All rights reserved.

  17. Enzymatically enhanced collisions on ultramicroelectrodes for specific and rapid detection of individual viruses

    PubMed Central

    Dick, Jeffrey E.; Hilterbrand, Adam T.; Strawsine, Lauren M.; Upton, Jason W.; Bard, Allen J.

    2016-01-01

    We report the specific collision of a single murine cytomegalovirus (MCMV) on a platinum ultramicroelectrode (UME, radius of 1 μm). Antibody directed against the viral surface protein glycoprotein B functionalized with glucose oxidase (GOx) allowed for specific detection of the virus in solution and a biological sample (urine). The oxidation of ferrocene methanol to ferrocenium methanol was carried out at the electrode surface, and the ferrocenium methanol acted as the cosubstrate to GOx to catalyze the oxidation of glucose to gluconolactone. In the presence of glucose, the incident collision of a GOx-covered virus onto the UME while ferrocene methanol was being oxidized produced stepwise increases in current as observed by amperometry. These current increases were observed due to the feedback loop of ferrocene methanol to the surface of the electrode after GOx reduces ferrocenium methanol back to ferrocene. Negative controls (i) without glucose, (ii) with an irrelevant virus (murine gammaherpesvirus 68), and (iii) without either virus do not display these current increases. Stepwise current decreases were observed for the prior two negative controls and no discrete events were observed for the latter. We further apply this method to the detection of MCMV in urine of infected mice. The method provides for a selective, rapid, and sensitive detection technique based on electrochemical collisions. PMID:27217569

  18. The impact of feature selection on one and two-class classification performance for plant microRNAs.

    PubMed

    Khalifa, Waleed; Yousef, Malik; Saçar Demirci, Müşerref Duygu; Allmer, Jens

    2016-01-01

    MicroRNAs (miRNAs) are short nucleotide sequences that form a typical hairpin structure which is recognized by a complex enzyme machinery. It ultimately leads to the incorporation of 18-24 nt long mature miRNAs into RISC where they act as recognition keys to aid in regulation of target mRNAs. It is involved to determine miRNAs experimentally and, therefore, machine learning is used to complement such endeavors. The success of machine learning mostly depends on proper input data and appropriate features for parameterization of the data. Although, in general, two-class classification (TCC) is used in the field; because negative examples are hard to come by, one-class classification (OCC) has been tried for pre-miRNA detection. Since both positive and negative examples are currently somewhat limited, feature selection can prove to be vital for furthering the field of pre-miRNA detection. In this study, we compare the performance of OCC and TCC using eight feature selection methods and seven different plant species providing positive pre-miRNA examples. Feature selection was very successful for OCC where the best feature selection method achieved an average accuracy of 95.6%, thereby being ∼29% better than the worst method which achieved 66.9% accuracy. While the performance is comparable to TCC, which performs up to 3% better than OCC, TCC is much less affected by feature selection and its largest performance gap is ∼13% which only occurs for two of the feature selection methodologies. We conclude that feature selection is crucially important for OCC and that it can perform on par with TCC given the proper set of features.

  19. Variable selection for zero-inflated and overdispersed data with application to health care demand in Germany.

    PubMed

    Wang, Zhu; Ma, Shuangge; Wang, Ching-Yun

    2015-09-01

    In health services and outcome research, count outcomes are frequently encountered and often have a large proportion of zeros. The zero-inflated negative binomial (ZINB) regression model has important applications for this type of data. With many possible candidate risk factors, this paper proposes new variable selection methods for the ZINB model. We consider maximum likelihood function plus a penalty including the least absolute shrinkage and selection operator (LASSO), smoothly clipped absolute deviation (SCAD), and minimax concave penalty (MCP). An EM (expectation-maximization) algorithm is proposed for estimating the model parameters and conducting variable selection simultaneously. This algorithm consists of estimating penalized weighted negative binomial models and penalized logistic models via the coordinated descent algorithm. Furthermore, statistical properties including the standard error formulae are provided. A simulation study shows that the new algorithm not only has more accurate or at least comparable estimation, but also is more robust than the traditional stepwise variable selection. The proposed methods are applied to analyze the health care demand in Germany using the open-source R package mpath. © 2015 WILEY-VCH Verlag GmbH & Co. KGaA, Weinheim.

  20. FSR: feature set reduction for scalable and accurate multi-class cancer subtype classification based on copy number.

    PubMed

    Wong, Gerard; Leckie, Christopher; Kowalczyk, Adam

    2012-01-15

    Feature selection is a key concept in machine learning for microarray datasets, where features represented by probesets are typically several orders of magnitude larger than the available sample size. Computational tractability is a key challenge for feature selection algorithms in handling very high-dimensional datasets beyond a hundred thousand features, such as in datasets produced on single nucleotide polymorphism microarrays. In this article, we present a novel feature set reduction approach that enables scalable feature selection on datasets with hundreds of thousands of features and beyond. Our approach enables more efficient handling of higher resolution datasets to achieve better disease subtype classification of samples for potentially more accurate diagnosis and prognosis, which allows clinicians to make more informed decisions in regards to patient treatment options. We applied our feature set reduction approach to several publicly available cancer single nucleotide polymorphism (SNP) array datasets and evaluated its performance in terms of its multiclass predictive classification accuracy over different cancer subtypes, its speedup in execution as well as its scalability with respect to sample size and array resolution. Feature Set Reduction (FSR) was able to reduce the dimensions of an SNP array dataset by more than two orders of magnitude while achieving at least equal, and in most cases superior predictive classification performance over that achieved on features selected by existing feature selection methods alone. An examination of the biological relevance of frequently selected features from FSR-reduced feature sets revealed strong enrichment in association with cancer. FSR was implemented in MATLAB R2010b and is available at http://ww2.cs.mu.oz.au/~gwong/FSR.

  1. Selective processing of multiple features in the human brain: effects of feature type and salience.

    PubMed

    McGinnis, E Menton; Keil, Andreas

    2011-02-09

    Identifying targets in a stream of items at a given constant spatial location relies on selection of aspects such as color, shape, or texture. Such attended (target) features of a stimulus elicit a negative-going event-related brain potential (ERP), termed Selection Negativity (SN), which has been used as an index of selective feature processing. In two experiments, participants viewed a series of Gabor patches in which targets were defined as a specific combination of color, orientation, and shape. Distracters were composed of different combinations of color, orientation, and shape of the target stimulus. This design allows comparisons of items with and without specific target features. Consistent with previous ERP research, SN deflections extended between 160-300 ms. Data from the subsequent P3 component (300-450 ms post-stimulus) were also examined, and were regarded as an index of target processing. In Experiment A, predominant effects of target color on SN and P3 amplitudes were found, along with smaller ERP differences in response to variations of orientation and shape. Manipulating color to be less salient while enhancing the saliency of the orientation of the Gabor patch (Experiment B) led to delayed color selection and enhanced orientation selection. Topographical analyses suggested that the location of SN on the scalp reliably varies with the nature of the to-be-attended feature. No interference of non-target features on the SN was observed. These results suggest that target feature selection operates by means of electrocortical facilitation of feature-specific sensory processes, and that selective electrocortical facilitation is more effective when stimulus saliency is heightened.

  2. Feature selection for the classification of traced neurons.

    PubMed

    López-Cabrera, José D; Lorenzo-Ginori, Juan V

    2018-06-01

    The great availability of computational tools to calculate the properties of traced neurons leads to the existence of many descriptors which allow the automated classification of neurons from these reconstructions. This situation determines the necessity to eliminate irrelevant features as well as making a selection of the most appropriate among them, in order to improve the quality of the classification obtained. The dataset used contains a total of 318 traced neurons, classified by human experts in 192 GABAergic interneurons and 126 pyramidal cells. The features were extracted by means of the L-measure software, which is one of the most used computational tools in neuroinformatics to quantify traced neurons. We review some current feature selection techniques as filter, wrapper, embedded and ensemble methods. The stability of the feature selection methods was measured. For the ensemble methods, several aggregation methods based on different metrics were applied to combine the subsets obtained during the feature selection process. The subsets obtained applying feature selection methods were evaluated using supervised classifiers, among which Random Forest, C4.5, SVM, Naïve Bayes, Knn, Decision Table and the Logistic classifier were used as classification algorithms. Feature selection methods of types filter, embedded, wrappers and ensembles were compared and the subsets returned were tested in classification tasks for different classification algorithms. L-measure features EucDistanceSD, PathDistanceSD, Branch_pathlengthAve, Branch_pathlengthSD and EucDistanceAve were present in more than 60% of the selected subsets which provides evidence about their importance in the classification of this neurons. Copyright © 2018 Elsevier B.V. All rights reserved.

  3. Color-selective attention need not be mediated by spatial attention.

    PubMed

    Andersen, Søren K; Müller, Matthias M; Hillyard, Steven A

    2009-06-08

    It is well-established that attention can select stimuli for preferential processing on the basis of non-spatial features such as color, orientation, or direction of motion. Evidence is mixed, however, as to whether feature-selective attention acts by increasing the signal strength of to-be-attended features irrespective of their spatial locations or whether it acts by guiding the spotlight of spatial attention to locations containing the relevant feature. To address this question, we designed a task in which feature-selective attention could not be mediated by spatial selection. Participants observed a display of intermingled dots of two colors, which rapidly and unpredictably changed positions, with the task of detecting brief intervals of reduced luminance of 20% of the dots of one or the other color. Both behavioral indices and electrophysiological measures of steady-state visual evoked potentials showed selectively enhanced processing of the attended-color items. The results demonstrate that feature-selective attention produces a sensory gain enhancement at early levels of the visual cortex that occurs without mediation by spatial attention.

  4. RESIDENTIAL RADON RESISTANT CONSTRUCTION FEATURE SELECTION SYSTEM

    EPA Science Inventory

    The report describes a proposed residential radon resistant construction feature selection system. The features consist of engineered barriers to reduce radon entry and accumulation indoors. The proposed Florida standards require radon resistant features in proportion to regional...

  5. Object-based selection from spatially-invariant representations: evidence from a feature-report task.

    PubMed

    Matsukura, Michi; Vecera, Shaun P

    2011-02-01

    Attention selects objects as well as locations. When attention selects an object's features, observers identify two features from a single object more accurately than two features from two different objects (object-based effect of attention; e.g., Duncan, Journal of Experimental Psychology: General, 113, 501-517, 1984). Several studies have demonstrated that object-based attention can operate at a late visual processing stage that is independent of objects' spatial information (Awh, Dhaliwal, Christensen, & Matsukura, Psychological Science, 12, 329-334, 2001; Matsukura & Vecera, Psychonomic Bulletin & Review, 16, 529-536, 2009; Vecera, Journal of Experimental Psychology: General, 126, 14-18, 1997; Vecera & Farah, Journal of Experimental Psychology: General, 123, 146-160, 1994). In the present study, we asked two questions regarding this late object-based selection mechanism. In Part I, we investigated how observers' foreknowledge of to-be-reported features allows attention to select objects, as opposed to individual features. Using a feature-report task, a significant object-based effect was observed when to-be-reported features were known in advance but not when this advance knowledge was absent. In Part II, we examined what drives attention to select objects rather than individual features in the absence of observers' foreknowledge of to-be-reported features. Results suggested that, when there was no opportunity for observers to direct their attention to objects that possess to-be-reported features at the time of stimulus presentation, these stimuli must retain strong perceptual cues to establish themselves as separate objects.

  6. Constraint programming based biomarker optimization.

    PubMed

    Zhou, Manli; Luo, Youxi; Sun, Guoquan; Mai, Guoqin; Zhou, Fengfeng

    2015-01-01

    Efficient and intuitive characterization of biological big data is becoming a major challenge for modern bio-OMIC based scientists. Interactive visualization and exploration of big data is proven to be one of the successful solutions. Most of the existing feature selection algorithms do not allow the interactive inputs from users in the optimizing process of feature selection. This study investigates this question as fixing a few user-input features in the finally selected feature subset and formulates these user-input features as constraints for a programming model. The proposed algorithm, fsCoP (feature selection based on constrained programming), performs well similar to or much better than the existing feature selection algorithms, even with the constraints from both literature and the existing algorithms. An fsCoP biomarker may be intriguing for further wet lab validation, since it satisfies both the classification optimization function and the biomedical knowledge. fsCoP may also be used for the interactive exploration of bio-OMIC big data by interactively adding user-defined constraints for modeling.

  7. Predication of different stages of Alzheimer's disease using neighborhood component analysis and ensemble decision tree.

    PubMed

    Jin, Mingwu; Deng, Weishu

    2018-05-15

    There is a spectrum of the progression from healthy control (HC) to mild cognitive impairment (MCI) without conversion to Alzheimer's disease (AD), to MCI with conversion to AD (cMCI), and to AD. This study aims to predict the different disease stages using brain structural information provided by magnetic resonance imaging (MRI) data. The neighborhood component analysis (NCA) is applied to select most powerful features for prediction. The ensemble decision tree classifier is built to predict which group the subject belongs to. The best features and model parameters are determined by cross validation of the training data. Our results show that 16 out of a total of 429 features were selected by NCA using 240 training subjects, including MMSE score and structural measures in memory-related regions. The boosting tree model with NCA features can achieve prediction accuracy of 56.25% on 160 test subjects. Principal component analysis (PCA) and sequential feature selection (SFS) are used for feature selection, while support vector machine (SVM) is used for classification. The boosting tree model with NCA features outperforms all other combinations of feature selection and classification methods. The results suggest that NCA be a better feature selection strategy than PCA and SFS for the data used in this study. Ensemble tree classifier with boosting is more powerful than SVM to predict the subject group. However, more advanced feature selection and classification methods or additional measures besides structural MRI may be needed to improve the prediction performance. Copyright © 2018 Elsevier B.V. All rights reserved.

  8. [Feature extraction for breast cancer data based on geometric algebra theory and feature selection using differential evolution].

    PubMed

    Li, Jing; Hong, Wenxue

    2014-12-01

    The feature extraction and feature selection are the important issues in pattern recognition. Based on the geometric algebra representation of vector, a new feature extraction method using blade coefficient of geometric algebra was proposed in this study. At the same time, an improved differential evolution (DE) feature selection method was proposed to solve the elevated high dimension issue. The simple linear discriminant analysis was used as the classifier. The result of the 10-fold cross-validation (10 CV) classification of public breast cancer biomedical dataset was more than 96% and proved superior to that of the original features and traditional feature extraction method.

  9. Application of machine learning on brain cancer multiclass classification

    NASA Astrophysics Data System (ADS)

    Panca, V.; Rustam, Z.

    2017-07-01

    Classification of brain cancer is a problem of multiclass classification. One approach to solve this problem is by first transforming it into several binary problems. The microarray gene expression dataset has the two main characteristics of medical data: extremely many features (genes) and only a few number of samples. The application of machine learning on microarray gene expression dataset mainly consists of two steps: feature selection and classification. In this paper, the features are selected using a method based on support vector machine recursive feature elimination (SVM-RFE) principle which is improved to solve multiclass classification, called multiple multiclass SVM-RFE. Instead of using only the selected features on a single classifier, this method combines the result of multiple classifiers. The features are divided into subsets and SVM-RFE is used on each subset. Then, the selected features on each subset are put on separate classifiers. This method enhances the feature selection ability of each single SVM-RFE. Twin support vector machine (TWSVM) is used as the method of the classifier to reduce computational complexity. While ordinary SVM finds single optimum hyperplane, the main objective Twin SVM is to find two non-parallel optimum hyperplanes. The experiment on the brain cancer microarray gene expression dataset shows this method could classify 71,4% of the overall test data correctly, using 100 and 1000 genes selected from multiple multiclass SVM-RFE feature selection method. Furthermore, the per class results show that this method could classify data of normal and MD class with 100% accuracy.

  10. Local Feature Selection for Data Classification.

    PubMed

    Armanfard, Narges; Reilly, James P; Komeili, Majid

    2016-06-01

    Typical feature selection methods choose an optimal global feature subset that is applied over all regions of the sample space. In contrast, in this paper we propose a novel localized feature selection (LFS) approach whereby each region of the sample space is associated with its own distinct optimized feature set, which may vary both in membership and size across the sample space. This allows the feature set to optimally adapt to local variations in the sample space. An associated method for measuring the similarities of a query datum to each of the respective classes is also proposed. The proposed method makes no assumptions about the underlying structure of the samples; hence the method is insensitive to the distribution of the data over the sample space. The method is efficiently formulated as a linear programming optimization problem. Furthermore, we demonstrate the method is robust against the over-fitting problem. Experimental results on eleven synthetic and real-world data sets demonstrate the viability of the formulation and the effectiveness of the proposed algorithm. In addition we show several examples where localized feature selection produces better results than a global feature selection method.

  11. Decision Variants for the Automatic Determination of Optimal Feature Subset in RF-RFE.

    PubMed

    Chen, Qi; Meng, Zhaopeng; Liu, Xinyi; Jin, Qianguo; Su, Ran

    2018-06-15

    Feature selection, which identifies a set of most informative features from the original feature space, has been widely used to simplify the predictor. Recursive feature elimination (RFE), as one of the most popular feature selection approaches, is effective in data dimension reduction and efficiency increase. A ranking of features, as well as candidate subsets with the corresponding accuracy, is produced through RFE. The subset with highest accuracy (HA) or a preset number of features (PreNum) are often used as the final subset. However, this may lead to a large number of features being selected, or if there is no prior knowledge about this preset number, it is often ambiguous and subjective regarding final subset selection. A proper decision variant is in high demand to automatically determine the optimal subset. In this study, we conduct pioneering work to explore the decision variant after obtaining a list of candidate subsets from RFE. We provide a detailed analysis and comparison of several decision variants to automatically select the optimal feature subset. Random forest (RF)-recursive feature elimination (RF-RFE) algorithm and a voting strategy are introduced. We validated the variants on two totally different molecular biology datasets, one for a toxicogenomic study and the other one for protein sequence analysis. The study provides an automated way to determine the optimal feature subset when using RF-RFE.

  12. New Magnetic and 10Be/9Be results from ODP site 851 (East Equatorial Pacific)

    NASA Astrophysics Data System (ADS)

    Valet, J. P.; Savranskaia, T.; Anojh, T.; Meynadier, L.; Thouveny, N.; Gacem, L.; L2NC, A. T.; Bassinot, F. C.; Simon, Q.

    2017-12-01

    The paleomagnetic record from ODP site 851 was the first long data of relative paleointensity that attempted to describe 4 Ma of geomagnetic variations. Among other features, it was characterized by an asymmetrical saw-tooth pattern of the intensity changes across reversals. The upper part of the record (0 to 1.1 Ma) was documented by stepwise alternating field (af) demagnetization of U-channels, while the deeper part could not be sampled by U-channels and instead combined shipboard measurements and stepwise demagnetized single samples within specific intervals. Thermal demagnetization was also conducted within specific intervals to assess the absence of viscous component. We performed a new detailed study using U-channels and single samples that were taken along a continuous splice section that covers the upper 80 meters of sediment. Stepwise demagnetization of the natural magnetization and of the anhysteretic magnetization were carried out for all samples and U-channels in order to improve the resolution and the reliability of relative paleointensity for the older part of the record. The new results improve the detailed magnetostratigraphy that was formerly established and provide additional details to the paleointensity results. In parallel, 10Be/9Be measurements were carried out at the same levels as the magnetic measurements to test further the controversial asymmetrical pattern of relative paleointensity. Unfortunately, the 10Be/9Be results did not provide any consistent signal. This failure most likely results from high carbonate concentration (about 85%) that yields poor adsorption of beryllium by the sediment particles and therefore generates large fluctuations. The reliability of the paleointensity record is linked to downcore homogeneity of the sediment that is characterized by little variability of carbonate content and therefore little changes in the magnetization response to the field. Summarizing poor clay content appears to be a favorable situation for relative paleointensity, while 10Be studies would benefit from the presence of clay-rich sediments. More generally, this lithological dependence should help understand differences between amplitudes and patterns of relative paleointensity and beryllium records.

  13. Outpost in Jovian system - a stepwise long-term undertaking

    NASA Astrophysics Data System (ADS)

    Yasaka, Tetsuo

    2003-11-01

    Space has been attracting human attention since the dawn of our history, and clues thus given have triggered scientific and cultural evolutions. Now the space is in our hands. Near earth space has been developed, providing benefits to daily life. Moon and Mars will become the stage of human activity in a few decades. What will be the next logical step? The next step should be an undertaking that promises substantial influence to human history, both in knowledge and productive activities. Looking into the future directions of technology development combined with their outcome, Kyushu University selected a stepwise long-term undertaking toward establishment of an outpost in the Jovian system. Jupiter is our closest gas planet, which is a replica of the Sun. Its true understanding is essential to our knowledge of the universe. Its satellites abounds versatility providing not only the crucial knowledge of science but energy and materials vital to space activities. Jovian outpost consists of the central station on or around Callisto, controlling several laboratories on other Galilean satellites and dispatching probes to the main planet including Jovian-Crafts to cruise within its atmosphere and Deep Probes to explore the depth of the hydrogen ocean. Utilization of materials especially water on Europa will enable energy management of the stations and probes, and will further provide sound base toward exploration of the outskirts of the solar system and beyond. This understanding needs a long term endeavor that should be handed over many generations. This is a technology development program but education is an essential part of the process. The task is based on a series of short (5 year) targets. Each target provides stepwise solution to the objective, yet provides substantial outputs to the society and industries in a timely manner. The paper describes the overall program and details of the first 5 year targets.

  14. Outpost in Jovian system—a stepwise long-term undertaking

    NASA Astrophysics Data System (ADS)

    Yasaka, Tetsuo

    2006-10-01

    Space has been attracting human attention since the dawn of our history, and clues thus given have triggered scientific and cultural evolutions. Now the space is in our hands. Near earth space has been developed, providing benefits to daily life. Moon and Mars will become the stage of human activity in a few decades. What will be the next logical step? The next step should be an undertaking that promises substantial influence to human history, both in knowledge and productive activities. Looking into the future directions of technology development combined with their outcome, Kyushu University selected a stepwise long-term undertaking toward establishment of an outpost in the Jovian system. Jupiter is our closest gas planet, which is a replica of the Sun. Its true understanding is essential to our knowledge of the universe. Its satellites abounds versatility providing not only the crucial knowledge of science but energy and materials vital to space activities. Jovian outpost consists of the central station on or around Callisto, controlling several laboratories on other Galilean satellites and dispatching probes to the main planet including Jovian-Crafts to cruise within its atmosphere and Deep Probes to explore the depth of the hydrogen ocean. Utilization of materials especially water on Europa will enable energy management of the stations and probes, and will further provide sound base toward exploration of the outskirts of the solar system and beyond. This understanding needs a long-term endeavor that should be handed over many generations. This is a technology development program but education is an essential part of the process. The task is based on a series of short (5 year) targets. Each target provides stepwise solution to the objective, yet provides substantial outputs to the society and industries in a timely manner. The paper describes the overall program and details of the first 5 year targets.

  15. Determination of Failure Point of Asphalt-Mixture Fatigue-Test Results Using the Flow Number Method

    NASA Astrophysics Data System (ADS)

    Wulan, C. E. P.; Setyawan, A.; Pramesti, F. P.

    2018-03-01

    The failure point of the results of fatigue tests of asphalt mixtures performed in controlled stress mode is difficult to determine. However, several methods from empirical studies are available to solve this problem. The objectives of this study are to determine the fatigue failure point of the results of indirect tensile fatigue tests using the Flow Number Method and to determine the best Flow Number model for the asphalt mixtures tested. In order to achieve these goals, firstly the best asphalt mixture of three was selected based on their Marshall properties. Next, the Indirect Tensile Fatigue Test was performed on the chosen asphalt mixture. The stress-controlled fatigue tests were conducted at a temperature of 20°C and frequency of 10 Hz, with the application of three loads: 500, 600, and 700 kPa. The last step was the application of the Flow Number methods, namely the Three-Stages Model, FNest Model, Francken Model, and Stepwise Method, to the results of the fatigue tests to determine the failure point of the specimen. The chosen asphalt mixture is EVA (Ethyl Vinyl Acetate) polymer -modified asphalt mixture with 6.5% OBC (Optimum Bitumen Content). Furthermore, the result of this study shows that the failure points of the EVA-modified asphalt mixture under loads of 500, 600, and 700 kPa are 6621, 4841, and 611 for the Three-Stages Model; 4271, 3266, and 537 for the FNest Model; 3401, 2431, and 421 for the Francken Model, and 6901, 6841, and 1291 for the Stepwise Method, respectively. These different results show that the bigger the loading, the smaller the number of cycles to failure. However, the best FN results are shown by the Three-Stages Model and the Stepwise Method, which exhibit extreme increases after the constant development of accumulated strain.

  16. Exploring non-stationarity patterns in schizophrenia: neural reorganization abnormalities in the alpha band

    NASA Astrophysics Data System (ADS)

    Núñez, Pablo; Poza, Jesús; Bachiller, Alejandro; Gomez-Pilar, Javier; Lubeiro, Alba; Molina, Vicente; Hornero, Roberto

    2017-08-01

    Objective. The aim of this paper was to characterize brain non-stationarity during an auditory oddball task in schizophrenia (SCH). The level of non-stationarity was measured in the baseline and response windows of relevant tones in SCH patients and healthy controls. Approach. Event-related potentials were recorded from 28 SCH patients and 51 controls. Non-stationarity was estimated in the conventional electroencephalography frequency bands by means of Kullback-Leibler divergence (KLD). Relative power (RP) was also computed to assess a possible complementarity with KLD. Main results. Results showed a widespread statistically significant increase in the level of non-stationarity from baseline to response in all frequency bands for both groups. Statistically significant differences in non-stationarity were found between SCH patients and controls in beta-2 and in the alpha band. SCH patients showed more non-stationarity in the left parieto-occipital region during the baseline window in the beta-2 band. A leave-one-out cross validation classification study with feature selection based on binary stepwise logistic regression to discriminate between SCH patients and controls provided a positive predictive value of 72.73% and negative predictive value of 78.95%. Significance. KLD can characterize transient neural reorganization during an attentional task in response to novelty and relevance. Our findings suggest anomalous reorganization of neural dynamics in SCH during an oddball task. The abnormal frequency-dependent modulation found in SCH patients during relevant tones is in agreement with the hypothesis of aberrant salience detection in SCH. The increase in non-stationarity in the alpha band during the active task supports the notion that this band is involved in top-down processing. The baseline differences in the beta-2 band suggest that hyperactivation of the default mode network during attention tasks may be related to SCH symptoms. Furthermore, the classification improved when features from both KLD and RP were used, supporting the idea that these measures can be complementary.

  17. Finite cohesion due to chain entanglement in polymer melts.

    PubMed

    Cheng, Shiwang; Lu, Yuyuan; Liu, Gengxin; Wang, Shi-Qing

    2016-04-14

    Three different types of experiments, quiescent stress relaxation, delayed rate-switching during stress relaxation, and elastic recovery after step strain, are carried out in this work to elucidate the existence of a finite cohesion barrier against free chain retraction in entangled polymers. Our experiments show that there is little hastened stress relaxation from step-wise shear up to γ = 0.7 and step-wise extension up to the stretching ratio λ = 1.5 at any time before or after the Rouse time. In contrast, a noticeable stress drop stemming from the built-in barrier-free chain retraction is predicted using the GLaMM model. In other words, the experiment reveals a threshold magnitude of step-wise deformation below which the stress relaxation follows identical dynamics whereas the GLaMM or Doi-Edwards model indicates a monotonic acceleration of the stress relaxation dynamics as a function of the magnitude of the step-wise deformation. Furthermore, a sudden application of startup extension during different stages of stress relaxation after a step-wise extension, i.e. the delayed rate-switching experiment, shows that the geometric condensation of entanglement strands in the cross-sectional area survives beyond the reptation time τd that is over 100 times the Rouse time τR. Our results point to the existence of a cohesion barrier that can prevent free chain retraction upon moderate deformation in well-entangled polymer melts.

  18. Novel methods of time-resolved fluorescence data analysis for in-vivo tissue characterization: application to atherosclerosis.

    PubMed

    Jo, J A; Fang, Q; Papaioannou, T; Qiao, J H; Fishbein, M C; Dorafshar, A; Reil, T; Baker, D; Freischlag, J; Marcu, L

    2004-01-01

    This study investigates the ability of new analytical methods of time-resolved laser-induced fluorescence spectroscopy (TR-LIFS) data to characterize tissue in-vivo, such as the composition of atherosclerotic vulnerable plaques. A total of 73 TR-LIFS measurements were taken in-vivo from the aorta of 8 rabbits, and subsequently analyzed using the Laguerre deconvolution technique. The investigated spots were classified as normal aorta, thin or thick lesions, and lesions rich in either collagen or macrophages/foam-cells. Different linear and nonlinear classification algorithms (linear discriminant analysis, stepwise linear discriminant analysis, principal component analysis, and feedforward neural networks) were developed using spectral and TR features (ratios of intensity values and Laguerre expansion coefficients, respectively). Normal intima and thin lesions were discriminated from thick lesions (sensitivity >90%, specificity 100%) using only spectral features. However, both spectral and time-resolved features were necessary to discriminate thick lesions rich in collagen from thick lesions rich in foam cells (sensitivity >85%, specificity >93%), and thin lesions rich in foam cells from normal aorta and thin lesions rich in collagen (sensitivity >85%, specificity >94%). Based on these findings, we believe that TR-LIFS information derived from the Laguerre expansion coefficients can provide a valuable additional dimension for in-vivo tissue characterization.

  19. A case definition and photographic screening tool for the facial phenotype of fetal alcohol syndrome.

    PubMed

    Astley, S J; Clarren, S K

    1996-07-01

    The purpose of this study was to demonstrate that a quantitative, multivariate case definition of the fetal alcohol syndrome (FAS) facial phenotype could be derived from photographs of individuals with FAS and to demonstrate how this case definition and photographic approach could be used to develop efficient, accurate, and precise screening tools, diagnostic aids, and possibly surveillance tools. Frontal facial photographs of 42 subjects (from birth to 27 years of age) with FAS were matched to 84 subjects without FAS. The study population was randomly divided in half. Group 1 was used to identify the facial features that best differentiated individuals with and without FAS. Group 2 was used for cross validation. In group 1, stepwise discriminant analysis identified three facial features (reduced palpebral fissure length/inner canthal distance ratio, smooth philtrum, and thin upper lip) as the cluster of features that differentiated individuals with and without FAS in groups 1 and 2 with 100% accuracy. Sensitivity and specificity were unaffected by race, gender, and age. The phenotypic case definition derived from photographs accurately distinguished between individuals with and without FAS, demonstrating the potential of this approach for developing screening, diagnostic, and surveillance tools. Further evaluation of the validity and generalizability of this method will be needed.

  20. Computer-assisted sperm morphometry fluorescence-based analysis has potential to determine progeny sex.

    PubMed

    Santolaria, Pilar; Pauciullo, Alfredo; Silvestre, Miguel A; Vicente-Fiel, Sandra; Villanova, Leyre; Pinton, Alain; Viruel, Juan; Sales, Ester; Yániz, Jesús L

    2016-01-01

    This study was designed to determine the ability of computer-assisted sperm morphometry analysis (CASA-Morph) with fluorescence to discriminate between spermatozoa carrying different sex chromosomes from the nuclear morphometrics generated and different statistical procedures in the bovine species. The study was divided into two experiments. The first was to study the morphometric differences between X- and Y-chromosome-bearing spermatozoa (SX and SY, respectively). Spermatozoa from eight bulls were processed to assess simultaneously the sex chromosome by FISH and sperm morphometry by fluorescence-based CASA-Morph. SX cells were larger than SY cells on average (P < 0.001) although with important differences between bulls. A simultaneous evaluation of all the measured features by discriminant analysis revealed that nuclear area and average fluorescence intensity were the variables selected by stepwise discriminant function analysis as the best discriminators between SX and SY. In the second experiment, the sperm nuclear morphometric results from CASA-Morph in nonsexed (mixed SX and SY) and sexed (SX) semen samples from four bulls were compared. FISH allowed a successful classification of spermatozoa according to their sex chromosome content. X-sexed spermatozoa displayed a larger size and fluorescence intensity than nonsexed spermatozoa (P < 0.05). We conclude that the CASA-Morph fluorescence-based method has the potential to find differences between X- and Y-chromosome-bearing spermatozoa in bovine species although more studies are needed to increase the precision of sex determination by this technique.

  1. A genomic copy number signature predicts radiation exposure in post-Chernobyl breast cancer.

    PubMed

    Wilke, Christina M; Braselmann, Herbert; Hess, Julia; Klymenko, Sergiy V; Chumak, Vadim V; Zakhartseva, Liubov M; Bakhanova, Elena V; Walch, Axel K; Selmansberger, Martin; Samaga, Daniel; Weber, Peter; Schneider, Ludmila; Fend, Falko; Bösmüller, Hans C; Zitzelsberger, Horst; Unger, Kristian

    2018-04-16

    Breast cancer is the second leading cause of cancer death among women worldwide and besides life style, age and genetic risk factors, exposure to ionizing radiation is known to increase the risk for breast cancer. Further, DNA copy number alterations (CNAs), which can result from radiation-induced double-strand breaks, are frequently occurring in breast cancer cells. We set out to identify a signature of CNAs discriminating breast cancers from radiation-exposed and non-exposed female patients. We analyzed resected breast cancer tissues from 68 exposed female Chernobyl clean-up workers and evacuees and 68 matched non-exposed control patients for CNAs by array comparative genomic hybridization analysis (aCGH). Using a stepwise forward-backward selection approach a non-complex CNA signature, that is, less than ten features, was identified in the training data set, which could be subsequently validated in the validation data set (p value < 0.05). The signature consisted of nine copy number regions located on chromosomal bands 7q11.22-11.23, 7q21.3, 16q24.3, 17q21.31, 20p11.23-11.21, 1p21.1, 2q35, 2q35, 6p22.2. The signature was independent of any clinical characteristics of the patients. In all, we identified a CNA signature that has the potential to allow identification of radiation-associated breast cancer at the individual level. © 2018 UICC.

  2. Successful Stepwise Development of Patient Research Partnership: 14 Years' Experience of Actions and Consequences in Outcome Measures in Rheumatology (OMERACT).

    PubMed

    de Wit, Maarten; Kirwan, John R; Tugwell, Peter; Beaton, Dorcas; Boers, Maarten; Brooks, Peter; Collins, Sarah; Conaghan, Philip G; D'Agostino, Maria-Antonietta; Hofstetter, Cathie; Hughes, Rod; Leong, Amye; Lyddiatt, Ann; March, Lyn; May, James; Montie, Pamela; Richards, Pamela; Simon, Lee S; Singh, Jasvinder A; Strand, Vibeke; Voshaar, Marieke; Bingham, Clifton O; Gossec, Laure

    2017-04-01

    There is increasing interest in making patient participation an integral component of medical research. However, practical guidance on optimizing this engagement in healthcare is scarce. Since 2002, patient involvement has been one of the key features of the Outcome Measures in Rheumatology (OMERACT) international consensus effort. Based on a review of cumulative data from qualitative studies and internal surveys among OMERACT participants, we explored the potential benefits and challenges of involving patient research partners in conferences and working group activities. We supplemented our review with personal experiences and reflections regarding patient participation in the OMERACT process. We found that between 2002 and 2016, 67 patients have attended OMERACT conferences, of whom 28 had sustained involvement; many other patients contributed to OMERACT working groups. Their participation provided face validity to the OMERACT process and expanded the research agenda. Essential facilitators have been the financial commitment to guarantee sustainable involvement of patients at these conferences, procedures for recruitment, selection and support, and dedicated time allocated in the program for patient issues. Current challenges include the representativeness of the patient panel, risk of pseudo-professionalization, and disparity in patients' and researchers' perception of involvement. In conclusion, OMERACT has embedded long-term patient involvement in the consensus-building process on the measurement of core health outcomes. This integrative process continues to evolve iteratively. We believe that the practical points raised here can improve participatory research implementation.

  3. A robust and efficient stepwise regression method for building sparse polynomial chaos expansions

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Abraham, Simon, E-mail: Simon.Abraham@ulb.ac.be; Raisee, Mehrdad; Ghorbaniasl, Ghader

    2017-03-01

    Polynomial Chaos (PC) expansions are widely used in various engineering fields for quantifying uncertainties arising from uncertain parameters. The computational cost of classical PC solution schemes is unaffordable as the number of deterministic simulations to be calculated grows dramatically with the number of stochastic dimension. This considerably restricts the practical use of PC at the industrial level. A common approach to address such problems is to make use of sparse PC expansions. This paper presents a non-intrusive regression-based method for building sparse PC expansions. The most important PC contributions are detected sequentially through an automatic search procedure. The variable selectionmore » criterion is based on efficient tools relevant to probabilistic method. Two benchmark analytical functions are used to validate the proposed algorithm. The computational efficiency of the method is then illustrated by a more realistic CFD application, consisting of the non-deterministic flow around a transonic airfoil subject to geometrical uncertainties. To assess the performance of the developed methodology, a detailed comparison is made with the well established LAR-based selection technique. The results show that the developed sparse regression technique is able to identify the most significant PC contributions describing the problem. Moreover, the most important stochastic features are captured at a reduced computational cost compared to the LAR method. The results also demonstrate the superior robustness of the method by repeating the analyses using random experimental designs.« less

  4. Celiac disease and Helicobacter pylori infection in children: Is there any Association?

    PubMed

    Narang, Manish; Puri, Amarender Singh; Sachdeva, Sanjeev; Singh, Jatinderpal; Kumar, Ajay; Saran, Ravindra K

    2017-06-01

    Helicobacter pylori (HP) infection can influence the inflammatory and immune responses in the gut and may therefore play a role in the development of gluten-related enteropathy in genetically susceptible individuals. Our objective was to assess the relationship between celiac disease and HP infection in children. Children (1-18 years) diagnosed as celiac disease (CD) (n = 324) with submission of gastric and duodenal biopsies and duodenal histology having Marsh grade III features were eligible for the study. Non-celiac patients referred for endoscopy were selected as controls. We studied proportion of HP prevalence in children with confirmed CD as compared with HP prevalence in reference group comprising non-celiac children referred for endoscopy. We also evaluated predictors of HP infection in children with celiac disease. Of the 324 participants with CD, gastric HP was seen in 37 (11.4%) patients. The prevalence of HP in patients without CD (50%, P < 0.001) was significantly higher. Among patients with CD, HP infection was most frequent in patients with Marsh IIIa. In the stepwise regression analysis for risk factors of HP infection in CD patients: presence of gastritis, hemoglobin, and absence of scalloping were found to be independent predictors in a multivariate setup. Celiac disease and gastric HP infection have inverse relationship that raises the question whether development of HP infection confers protection against CD. © 2016 Journal of Gastroenterology and Hepatology Foundation and John Wiley & Sons Australia, Ltd.

  5. The ATLAS high level trigger steering

    NASA Astrophysics Data System (ADS)

    Berger, N.; Bold, T.; Eifert, T.; Fischer, G.; George, S.; Haller, J.; Hoecker, A.; Masik, J.; Nedden, M. Z.; Reale, V. P.; Risler, C.; Schiavi, C.; Stelzer, J.; Wu, X.

    2008-07-01

    The High Level Trigger (HLT) of the ATLAS experiment at the Large Hadron Collider receives events which pass the LVL1 trigger at ~75 kHz and has to reduce the rate to ~200 Hz while retaining the most interesting physics. It is a software trigger and performs the reduction in two stages: the LVL2 trigger and the Event Filter (EF). At the heart of the HLT is the Steering software. To minimise processing time and data transfers it implements the novel event selection strategies of seeded, step-wise reconstruction and early rejection. The HLT is seeded by regions of interest identified at LVL1. These and the static configuration determine which algorithms are run to reconstruct event data and test the validity of trigger signatures. The decision to reject the event or continue is based on the valid signatures, taking into account pre-scale and pass-through. After the EF, event classification tags are assigned for streaming purposes. Several new features for commissioning and operation have been added: comprehensive monitoring is now built in to the framework; for validation and debugging, reconstructed data can be written out; the steering is integrated with the new configuration (presented separately), and topological and global triggers have been added. This paper will present details of the final design and its implementation, the principles behind it, and the requirements and constraints it is subject to. The experience gained from technical runs with realistic trigger menus will be described.

  6. Quantifying Ciliary Dynamics during Assembly Reveals Step-wise Waveform Maturation in Airway Cells.

    PubMed

    Oltean, Alina; Schaffer, Andrew J; Bayly, Philip V; Brody, Steven L

    2018-05-31

    Motile cilia are essential for clearance of particulates and pathogens from airways. For effective transport, ciliary motor proteins and axonemal structures interact to generate the rhythmic, propulsive bending, but the mechanisms that produce a dynamic waveform remain incompletely understood. Biomechanical measures of human cilia motion and their relationships to cilia assembly are needed to illuminate the biophysics of normal cilia function, and to quantify dysfunction in ciliopathies. To these ends, we analyzed cilia motion from high-speed video microscopy of ciliated cells sampled from human lung airways compared to primary-culture cells that undergo ciliogenesis in vitro. Quantitative assessment of waveform parameters showed variations in waveform shape between individual cilia; however, general trends in waveform parameters emerged, associated with progression of cilia length and stage of differentiation. When cilia emerged from cultured cells, beat frequency was initially elevated, then fell and remained stable as cilia lengthened. In contrast, the average bending amplitude and the ability to generate force gradually increased and eventually approached values observed in ex vivo samples. Dynein arm motor proteins DNAH5, DNAH9, DNAH11, and DNAH6 were localized within specific regions of the axoneme in the ex vivo cells; however distinct stages of in vitro waveform development identified by biomechanical features were associated with the progressive movement of dyneins to the appropriate proximal or distal sections of the cilium. These observations suggest that the step-wise variation in waveform development during ciliogenesis is dependent on cilia length and potentially outer dynein arm assembly.

  7. Analyzing urban ecosystem variation in the City of Dongguan: A stepwise cluster modeling approach.

    PubMed

    Sun, J; Li, Y P; Gao, P P; Suo, C; Xia, B C

    2018-06-13

    In this study, a stepwise cluster modeling approach (SCMA) is developed for analyzing urban ecosystem variation via Normalized Difference Vegetation Index (NDVI). NDVI is an indicator of vegetation growth and coverage and useful in reflecting urban ecosystem. SCMA is established on a cluster tree that can characterize the complex relationship between independent and dependent variables. SCMA is applied to the City of Dongguan for simulating the urban NDVI and identifying associated drivers of human activity, topography and meteorology without specific functions. Results show that SCMA performances better than conventional statistical methods, illustrating the ability of SCMA in capturing the complex and nonlinear features of urban ecosystem. Results disclose that human activities play negative effects on NDVI due to the destruction of green space for pursuing more space for buildings. NDVI reduces gradually from the south part to the north part of Dongguan due to increased gross domestic product and population density, indicating that the ecosystem in Dongguan is better in the south part. NDVI in the northeast part (dominated by agriculture) is sensitive to the growth of economy and population. More attention should be paid to this part for sustainable development, such as increasing afforestation, planting grass and constructing parks. Precipitation has a positive effect on NDVI due to the promotion of soil moisture that is beneficial to plants' growth. Awareness of these complexities is helpful for sustainable development of urban ecosystem. Copyright © 2018 Elsevier Inc. All rights reserved.

  8. Influence of time and length size feature selections for human activity sequences recognition.

    PubMed

    Fang, Hongqing; Chen, Long; Srinivasan, Raghavendiran

    2014-01-01

    In this paper, Viterbi algorithm based on a hidden Markov model is applied to recognize activity sequences from observed sensors events. Alternative features selections of time feature values of sensors events and activity length size feature values are tested, respectively, and then the results of activity sequences recognition performances of Viterbi algorithm are evaluated. The results show that the selection of larger time feature values of sensor events and/or smaller activity length size feature values will generate relatively better results on the activity sequences recognition performances. © 2013 ISA Published by ISA All rights reserved.

  9. Adaptive runtime for a multiprocessing API

    DOEpatents

    Antao, Samuel F.; Bertolli, Carlo; Eichenberger, Alexandre E.; O'Brien, John K.

    2016-11-15

    A computer-implemented method includes selecting a runtime for executing a program. The runtime includes a first combination of feature implementations, where each feature implementation implements a feature of an application programming interface (API). Execution of the program is monitored, and the execution uses the runtime. Monitor data is generated based on the monitoring. A second combination of feature implementations are selected, by a computer processor, where the selection is based at least in part on the monitor data. The runtime is modified by activating the second combination of feature implementations to replace the first combination of feature implementations.

  10. Adaptive runtime for a multiprocessing API

    DOEpatents

    Antao, Samuel F.; Bertolli, Carlo; Eichenberger, Alexandre E.; O'Brien, John K.

    2016-10-11

    A computer-implemented method includes selecting a runtime for executing a program. The runtime includes a first combination of feature implementations, where each feature implementation implements a feature of an application programming interface (API). Execution of the program is monitored, and the execution uses the runtime. Monitor data is generated based on the monitoring. A second combination of feature implementations are selected, by a computer processor, where the selection is based at least in part on the monitor data. The runtime is modified by activating the second combination of feature implementations to replace the first combination of feature implementations.

  11. Informative Feature Selection for Object Recognition via Sparse PCA

    DTIC Science & Technology

    2011-04-07

    constraint on images collected from low-power camera net- works instead of high-end photography is that establishing wide-baseline feature correspondence of...variable selection tool for selecting informative features in the object images captured from low-resolution cam- era sensor networks. Firstly, we...More examples can be found in Figure 4 later. 3. Identifying Informative Features Classical PCA is a well established tool for the analysis of high

  12. Diagnostic Features of Common Oral Ulcerative Lesions: An Updated Decision Tree

    PubMed Central

    Safi, Yaser

    2016-01-01

    Diagnosis of oral ulcerative lesions might be quite challenging. This narrative review article aims to introduce an updated decision tree for diagnosing oral ulcerative lesions on the basis of their diagnostic features. Various general search engines and specialized databases including PubMed, PubMed Central, Medline Plus, EBSCO, Science Direct, Scopus, Embase, and authenticated textbooks were used to find relevant topics by means of MeSH keywords such as “oral ulcer,” “stomatitis,” and “mouth diseases.” Thereafter, English-language articles published since 1983 to 2015 in both medical and dental journals including reviews, meta-analyses, original papers, and case reports were appraised. Upon compilation of the relevant data, oral ulcerative lesions were categorized into three major groups: acute, chronic, and recurrent ulcers and into five subgroups: solitary acute, multiple acute, solitary chronic, multiple chronic, and solitary/multiple recurrent, based on the number and duration of lesions. In total, 29 entities were organized in the form of a decision tree in order to help clinicians establish a logical diagnosis by stepwise progression. PMID:27781066

  13. Template-based preparation of free-standing semiconducting polymeric nanorod arrays on conductive substrates.

    PubMed

    Haberkorn, Niko; Weber, Stefan A L; Berger, Rüdiger; Theato, Patrick

    2010-06-01

    We describe the synthesis and characterization of a cross-linkable siloxane-derivatized tetraphenylbenzidine (DTMS-TPD), which was used for the fabrication of semiconducting highly ordered nanorod arrays on conductive indium tin oxide or Pt-coated substrates. The stepwise process allow fabricating of macroscopic areas of well-ordered free-standing nanorod arrays, which feature a high resistance against organic solvents, semiconducting properties and a good adhesion to the substrate. Thin films of the TPD derivate with good hole-conducting properties could be prepared by cross-linking and covalently attaching to hydroxylated substrates utilizing an initiator-free thermal curing at 160 degrees C. The nanorod arrays composed of cross-linked DTMS-TPD were fabricated by an anodic aluminum oxide (AAO) template approach. Furthermore, the nanorod arrays were investigated by a recently introduced method allowing to probe local conductivity on fragile structures. It revealed that more than 98% of the nanorods exhibit electrical conductance and consequently feature a good electrical contact to the substrate. The prepared nanorod arrays have the potential to find application in the fabrication of multilayered device architectures for building well-ordered bulk-heterojunction solar cells.

  14. Endoscopic features of gastro-intestinal lymphomas: From diagnosis to follow-up

    PubMed Central

    Vetro, Calogero; Romano, Alessandra; Amico, Irene; Conticello, Concetta; Motta, Giovanna; Figuera, Amalia; Chiarenza, Annalisa; Di Raimondo, Cosimo; Giulietti, Giorgio; Bonanno, Giacomo; Palumbo, Giuseppe Alberto; Di Raimondo, Francesco

    2014-01-01

    Many progresses have been done in the management of gastrointestinal (GI) lymphomas during last decades, especially after the discovery of Helicobacter pylori-dependent lymphoma development. The stepwise implementation of new endoscopic techniques, by means of echoendoscopy or double-balloon enteroscopy, enabled us to more precisely describe the endoscopic features of GI lymphomas with substantial contribution in patient management and in tailoring the treatment strategy with organ preserving approaches. In this review, we describe the recent progresses in GI lymphoma management from disease diagnosis to follow-up with a specific focus on the endoscopic presentation according to the involved site and the lymphoma subtype. Additionally, new or emerging endoscopic technologies that have an impact on the management of gastrointestinal lymphomas are reported. We here discuss the two most common subtypes of GI lymphomas: the mucosa-associated lymphoid tissue and the diffuse large B cell lymphoma. A general outline on the state-of-the-art of the disease and on the role of endoscopy in both diagnosis and follow-up will be performed. PMID:25278693

  15. Fallopian tube cytology: a histocorrelative study of 150 washings.

    PubMed

    Mulvany, N J; Arnstein, M; Ostör, A G

    1997-06-01

    The aim of the study was to assess the relationship between fallopian tube lavage cytology and recognized microscopic prognostic features in cancer of the uterine corpus. Tubal (TW) and peritoneal washing cytology (PW), endometrial tumor grade, and tumor involvement of the cervix, myometrium, myometrial vessels, and peritoneum were assessed in 150 patients. Endometrioid adenocarcinoma grade I was considered a low-grade tumor, while endometrioid carcinoma grades 2/3, serous/clear cell carcinoma, carcinosarcoma, and high-grade stromal sarcoma were considered high grade. The overall concordance rate for paired TWs and PWs was 72% (108/150). Forward stepwise logistic regression analysis of the 150 tumors revealed that only PWs and cervical involvement were independently predictive of TWs. No relationship was evident between TWs and depth of myometrial invasion, myometrial vascular involvement, or peritoneal metastases. It is concluded that retrograde transtubal spread by malignant endometrial cells occurs independently of myometrial histoprognostic features. TWs provide supporting evidence for diagnostically difficult PWs, and malignant TWs may be detected in the presence of minimally invasive serous/clear cell carcinoma and carcinosarcoma of the endometrium.

  16. Subsolid pulmonary nodules: imaging evaluation and strategic management.

    PubMed

    Godoy, Myrna C B; Sabloff, Bradley; Naidich, David P

    2012-07-01

    Given the higher rate of malignancy of subsolid pulmonary nodules and the considerably lower growth rate of ground-glass nodules (GGNs), dedicated standardized guidelines for management of these nodules have been proposed, including long-term low-dose computed tomography (CT) follow-up (≥3 years). Physicians must be familiar with the strategic management of subsolid pulmonary nodules, and should be able to identify imaging features that suggest invasive adenocarcinoma requiring a more aggressive management. Low-dose CT screening studies for early detection of lung cancer have increased our knowledge of pulmonary nodules, and in particular our understanding of the strong although imperfect correlation of the subsolid pulmonary nodules, including pure GGNs and part-solid nodules, with the spectrum of preinvasive to invasive lung adenocarcinoma. Serial CT imaging has shown stepwise progression in a subset of these nodules, characterized by increase in size and density of pure GGNs and development of a solid component, the latter usually indicating invasive adenocarcinoma. There is close correlation between the CT features of subsolid nodules (SSNs) and the spectrum of lung adenocarcinoma. Standardized guidelines are suggested for management of SSNs.

  17. Emotional textile image classification based on cross-domain convolutional sparse autoencoders with feature selection

    NASA Astrophysics Data System (ADS)

    Li, Zuhe; Fan, Yangyu; Liu, Weihua; Yu, Zeqi; Wang, Fengqin

    2017-01-01

    We aim to apply sparse autoencoder-based unsupervised feature learning to emotional semantic analysis for textile images. To tackle the problem of limited training data, we present a cross-domain feature learning scheme for emotional textile image classification using convolutional autoencoders. We further propose a correlation-analysis-based feature selection method for the weights learned by sparse autoencoders to reduce the number of features extracted from large size images. First, we randomly collect image patches on an unlabeled image dataset in the source domain and learn local features with a sparse autoencoder. We then conduct feature selection according to the correlation between different weight vectors corresponding to the autoencoder's hidden units. We finally adopt a convolutional neural network including a pooling layer to obtain global feature activations of textile images in the target domain and send these global feature vectors into logistic regression models for emotional image classification. The cross-domain unsupervised feature learning method achieves 65% to 78% average accuracy in the cross-validation experiments corresponding to eight emotional categories and performs better than conventional methods. Feature selection can reduce the computational cost of global feature extraction by about 50% while improving classification performance.

  18. Experience, Reflect, Critique: The End of the "Learning Cycles" Era

    ERIC Educational Resources Information Center

    Seaman, Jayson

    2008-01-01

    According to prevailing models, experiential learning is by definition a stepwise process beginning with direct experience, followed by reflection, followed by learning. It has been argued, however, that stepwise models inadequately explain the holistic learning processes that are central to learning from experience, and that they lack scientific…

  19. A Latent-Variable Causal Model of Faculty Reputational Ratings.

    ERIC Educational Resources Information Center

    King, Suzanne; Wolfle, Lee M.

    A reanalysis was conducted of Saunier's research (1985) on sources of variation in the National Research Council (NRC) reputational ratings of university faculty. Saunier conducted a stepwise regression analysis using 12 predictor variables. Due to problems with multicollinearity and because of the atheoretical nature of stepwise regression,…

  20. SEPARATION PROCESS FOR THORIUM SALTS

    DOEpatents

    Bridger, G.L.; Whatley, M.E.; Shaw, K.G.

    1957-12-01

    A process is described for the separation of uranium, thorium, and rare earths extracted from monazite by digesting with sulfuric acid. By carefully increasing the pH of the solution, stepwise, over the range 0.8 to 5.5, a series of selective precipitations will be achieved, with the thorium values coming out at lower pH, the rare earths at intermediate pH and the uranium last. Some mixed precipitates will be obtained, and these may be treated by dissolving in HNO/sub 3/ and contacting with dibutyl phosphate, whereby thorium or uranium are taken up by the organic phase while the rare earths preferentially remain in the aqueous solution.

  1. Neuropsychology 3.0: Evidence-Based Science and Practice

    PubMed Central

    Bilder, Robert M.

    2011-01-01

    Neuropsychology is poised for transformations of its concepts and methods, leveraging advances in neuroimaging, the human genome project, psychometric theory, and information technologies. It is argued that a paradigm shift towards evidence-based science and practice can be enabled by innovations, including: (1) formal definition of neuropsychological concepts and tasks in cognitive ontologies; (2) creation of collaborative neuropsychological knowledgebases; and (3) design of web-based assessment methods that permit free development, large-sample implementation, and dynamic refinement of neuropsychological tests and the constructs these aim to assess. This article considers these opportunities, highlights selected obstacles, and offers suggestions for stepwise progress towards these goals. PMID:21092355

  2. Hydraulic geometry and streamflow of channels in the Piceance Basin, Rio Blanco and Garfield counties, Colorado

    USGS Publications Warehouse

    Elliott, J.G.; Cartier, K.D.

    1986-01-01

    The influence of streamflow and basin characteristics on channel geometry was investigated at 18 perennial and ephemeral stream reaches in the Piceance basin of northwestern Colorado. Results of stepwise multiple regression analyses indicated that the variabilities of mean bankfull depth (D) and bankfull cross-sectional flow area (Af) were predominantly a function of bankfull discharge (QB), and that most of the variability in channel slopes (S) could be explained by drainage area (DA). None of the independent variables selected for the study could account for a large part of the variability in bankfull channel width (W). (USGS)

  3. Semantic point cloud interpretation based on optimal neighborhoods, relevant features and efficient classifiers

    NASA Astrophysics Data System (ADS)

    Weinmann, Martin; Jutzi, Boris; Hinz, Stefan; Mallet, Clément

    2015-07-01

    3D scene analysis in terms of automatically assigning 3D points a respective semantic label has become a topic of great importance in photogrammetry, remote sensing, computer vision and robotics. In this paper, we address the issue of how to increase the distinctiveness of geometric features and select the most relevant ones among these for 3D scene analysis. We present a new, fully automated and versatile framework composed of four components: (i) neighborhood selection, (ii) feature extraction, (iii) feature selection and (iv) classification. For each component, we consider a variety of approaches which allow applicability in terms of simplicity, efficiency and reproducibility, so that end-users can easily apply the different components and do not require expert knowledge in the respective domains. In a detailed evaluation involving 7 neighborhood definitions, 21 geometric features, 7 approaches for feature selection, 10 classifiers and 2 benchmark datasets, we demonstrate that the selection of optimal neighborhoods for individual 3D points significantly improves the results of 3D scene analysis. Additionally, we show that the selection of adequate feature subsets may even further increase the quality of the derived results while significantly reducing both processing time and memory consumption.

  4. Patient feature based dosimetric Pareto front prediction in esophageal cancer radiotherapy.

    PubMed

    Wang, Jiazhou; Jin, Xiance; Zhao, Kuaike; Peng, Jiayuan; Xie, Jiang; Chen, Junchao; Zhang, Zhen; Studenski, Matthew; Hu, Weigang

    2015-02-01

    To investigate the feasibility of the dosimetric Pareto front (PF) prediction based on patient's anatomic and dosimetric parameters for esophageal cancer patients. Eighty esophagus patients in the authors' institution were enrolled in this study. A total of 2928 intensity-modulated radiotherapy plans were obtained and used to generate PF for each patient. On average, each patient had 36.6 plans. The anatomic and dosimetric features were extracted from these plans. The mean lung dose (MLD), mean heart dose (MHD), spinal cord max dose, and PTV homogeneity index were recorded for each plan. Principal component analysis was used to extract overlap volume histogram (OVH) features between PTV and other organs at risk. The full dataset was separated into two parts; a training dataset and a validation dataset. The prediction outcomes were the MHD and MLD. The spearman's rank correlation coefficient was used to evaluate the correlation between the anatomical features and dosimetric features. The stepwise multiple regression method was used to fit the PF. The cross validation method was used to evaluate the model. With 1000 repetitions, the mean prediction error of the MHD was 469 cGy. The most correlated factor was the first principal components of the OVH between heart and PTV and the overlap between heart and PTV in Z-axis. The mean prediction error of the MLD was 284 cGy. The most correlated factors were the first principal components of the OVH between heart and PTV and the overlap between lung and PTV in Z-axis. It is feasible to use patients' anatomic and dosimetric features to generate a predicted Pareto front. Additional samples and further studies are required improve the prediction model.

  5. TU-C-17A-10: Patient Features Based Dosimetric Pareto Front Prediction In Esophagus Cancer Radiotherapy

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Wang, J; Zhao, K; Peng, J

    2014-06-15

    Purpose: The purpose of this study is to study the feasibility of the dosimetric pareto front (PF) prediction based on patient anatomic and dosimetric parameters for esophagus cancer patients. Methods: Sixty esophagus patients in our institution were enrolled in this study. A total 2920 IMRT plans were created to generated PF for each patient. On average, each patient had 48 plans. The anatomic and dosimetric features were extracted from those plans. The mean lung dose (MLD), mean heart dose (MHD), spinal cord max dose and PTV homogeneous index (PTVHI) were recorded for each plan. The principal component analysis (PCA) wasmore » used to extract overlap volume histogram (OVH) features between PTV and other critical organs. The full dataset was separated into two parts include the training dataset and the validation dataset. The prediction outcomes were the MHD and MLD for the current study. The spearman rank correlation coefficient was used to evaluate the correlation between the anatomical features and dosimetric features. The PF was fit by the the stepwise multiple regression method. The cross-validation method was used to evaluation the model. Results: The mean prediction error of the MHD was 465 cGy with 100 repetitions. The most correlated factors were the first principal components of the OVH between heart and PTV, and the overlap between heart and PTV in Z-axis. The mean prediction error of the MLD was 195 cGy. The most correlated factors were the first principal components of the OVH between lung and PTV, and the overlap between lung and PTV in Z-axis. Conclusion: It is feasible to use patients anatomic and dosimetric features to generate a predicted PF. Additional samples and further studies were required to get a better prediction model.« less

  6. Patient feature based dosimetric Pareto front prediction in esophageal cancer radiotherapy

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Wang, Jiazhou; Zhao, Kuaike; Peng, Jiayuan

    2015-02-15

    Purpose: To investigate the feasibility of the dosimetric Pareto front (PF) prediction based on patient’s anatomic and dosimetric parameters for esophageal cancer patients. Methods: Eighty esophagus patients in the authors’ institution were enrolled in this study. A total of 2928 intensity-modulated radiotherapy plans were obtained and used to generate PF for each patient. On average, each patient had 36.6 plans. The anatomic and dosimetric features were extracted from these plans. The mean lung dose (MLD), mean heart dose (MHD), spinal cord max dose, and PTV homogeneity index were recorded for each plan. Principal component analysis was used to extract overlapmore » volume histogram (OVH) features between PTV and other organs at risk. The full dataset was separated into two parts; a training dataset and a validation dataset. The prediction outcomes were the MHD and MLD. The spearman’s rank correlation coefficient was used to evaluate the correlation between the anatomical features and dosimetric features. The stepwise multiple regression method was used to fit the PF. The cross validation method was used to evaluate the model. Results: With 1000 repetitions, the mean prediction error of the MHD was 469 cGy. The most correlated factor was the first principal components of the OVH between heart and PTV and the overlap between heart and PTV in Z-axis. The mean prediction error of the MLD was 284 cGy. The most correlated factors were the first principal components of the OVH between heart and PTV and the overlap between lung and PTV in Z-axis. Conclusions: It is feasible to use patients’ anatomic and dosimetric features to generate a predicted Pareto front. Additional samples and further studies are required improve the prediction model.« less

  7. Efficient feature selection using a hybrid algorithm for the task of epileptic seizure detection

    NASA Astrophysics Data System (ADS)

    Lai, Kee Huong; Zainuddin, Zarita; Ong, Pauline

    2014-07-01

    Feature selection is a very important aspect in the field of machine learning. It entails the search of an optimal subset from a very large data set with high dimensional feature space. Apart from eliminating redundant features and reducing computational cost, a good selection of feature also leads to higher prediction and classification accuracy. In this paper, an efficient feature selection technique is introduced in the task of epileptic seizure detection. The raw data are electroencephalography (EEG) signals. Using discrete wavelet transform, the biomedical signals were decomposed into several sets of wavelet coefficients. To reduce the dimension of these wavelet coefficients, a feature selection method that combines the strength of both filter and wrapper methods is proposed. Principal component analysis (PCA) is used as part of the filter method. As for wrapper method, the evolutionary harmony search (HS) algorithm is employed. This metaheuristic method aims at finding the best discriminating set of features from the original data. The obtained features were then used as input for an automated classifier, namely wavelet neural networks (WNNs). The WNNs model was trained to perform a binary classification task, that is, to determine whether a given EEG signal was normal or epileptic. For comparison purposes, different sets of features were also used as input. Simulation results showed that the WNNs that used the features chosen by the hybrid algorithm achieved the highest overall classification accuracy.

  8. A study of metaheuristic algorithms for high dimensional feature selection on microarray data

    NASA Astrophysics Data System (ADS)

    Dankolo, Muhammad Nasiru; Radzi, Nor Haizan Mohamed; Sallehuddin, Roselina; Mustaffa, Noorfa Haszlinna

    2017-11-01

    Microarray systems enable experts to examine gene profile at molecular level using machine learning algorithms. It increases the potentials of classification and diagnosis of many diseases at gene expression level. Though, numerous difficulties may affect the efficiency of machine learning algorithms which includes vast number of genes features comprised in the original data. Many of these features may be unrelated to the intended analysis. Therefore, feature selection is necessary to be performed in the data pre-processing. Many feature selection algorithms are developed and applied on microarray which including the metaheuristic optimization algorithms. This paper discusses the application of the metaheuristics algorithms for feature selection in microarray dataset. This study reveals that, the algorithms have yield an interesting result with limited resources thereby saving computational expenses of machine learning algorithms.

  9. A step towards considering the spatial heterogeneity of urban key features in urban hydrology flood modelling

    NASA Astrophysics Data System (ADS)

    Leandro, J.; Schumann, A.; Pfister, A.

    2016-04-01

    Some of the major challenges in modelling rainfall-runoff in urbanised areas are the complex interaction between the sewer system and the overland surface, and the spatial heterogeneity of the urban key features. The former requires the sewer network and the system of surface flow paths to be solved simultaneously. The latter is still an unresolved issue because the heterogeneity of runoff formation requires high detailed information and includes a large variety of feature specific rainfall-runoff dynamics. This paper discloses a methodology for considering the variability of building types and the spatial heterogeneity of land surfaces. The former is achieved by developing a specific conceptual rainfall-runoff model and the latter by defining a fully distributed approach for infiltration processes in urban areas with limited storage capacity dependent on OpenStreetMaps (OSM). The model complexity is increased stepwise by adding components to an existing 2D overland flow model. The different steps are defined as modelling levels. The methodology is applied in a German case study. Results highlight that: (a) spatial heterogeneity of urban features has a medium to high impact on the estimated overland flood-depths, (b) the addition of multiple urban features have a higher cumulative effect due to the dynamic effects simulated by the model, (c) connecting the runoff from buildings to the sewer contributes to the non-linear effects observed on the overland flood-depths, and (d) OSM data is useful in identifying pounding areas (for which infiltration plays a decisive role) and permeable natural surface flow paths (which delay the flood propagation).

  10. Improved sparse decomposition based on a smoothed L0 norm using a Laplacian kernel to select features from fMRI data.

    PubMed

    Zhang, Chuncheng; Song, Sutao; Wen, Xiaotong; Yao, Li; Long, Zhiying

    2015-04-30

    Feature selection plays an important role in improving the classification accuracy of multivariate classification techniques in the context of fMRI-based decoding due to the "few samples and large features" nature of functional magnetic resonance imaging (fMRI) data. Recently, several sparse representation methods have been applied to the voxel selection of fMRI data. Despite the low computational efficiency of the sparse representation methods, they still displayed promise for applications that select features from fMRI data. In this study, we proposed the Laplacian smoothed L0 norm (LSL0) approach for feature selection of fMRI data. Based on the fast sparse decomposition using smoothed L0 norm (SL0) (Mohimani, 2007), the LSL0 method used the Laplacian function to approximate the L0 norm of sources. Results of the simulated and real fMRI data demonstrated the feasibility and robustness of LSL0 for the sparse source estimation and feature selection. Simulated results indicated that LSL0 produced more accurate source estimation than SL0 at high noise levels. The classification accuracy using voxels that were selected by LSL0 was higher than that by SL0 in both simulated and real fMRI experiment. Moreover, both LSL0 and SL0 showed higher classification accuracy and required less time than ICA and t-test for the fMRI decoding. LSL0 outperformed SL0 in sparse source estimation at high noise level and in feature selection. Moreover, LSL0 and SL0 showed better performance than ICA and t-test for feature selection. Copyright © 2015 Elsevier B.V. All rights reserved.

  11. Feature-Selective Attention Adaptively Shifts Noise Correlations in Primary Auditory Cortex.

    PubMed

    Downer, Joshua D; Rapone, Brittany; Verhein, Jessica; O'Connor, Kevin N; Sutter, Mitchell L

    2017-05-24

    Sensory environments often contain an overwhelming amount of information, with both relevant and irrelevant information competing for neural resources. Feature attention mediates this competition by selecting the sensory features needed to form a coherent percept. How attention affects the activity of populations of neurons to support this process is poorly understood because population coding is typically studied through simulations in which one sensory feature is encoded without competition. Therefore, to study the effects of feature attention on population-based neural coding, investigations must be extended to include stimuli with both relevant and irrelevant features. We measured noise correlations ( r noise ) within small neural populations in primary auditory cortex while rhesus macaques performed a novel feature-selective attention task. We found that the effect of feature-selective attention on r noise depended not only on the population tuning to the attended feature, but also on the tuning to the distractor feature. To attempt to explain how these observed effects might support enhanced perceptual performance, we propose an extension of a simple and influential model in which shifts in r noise can simultaneously enhance the representation of the attended feature while suppressing the distractor. These findings present a novel mechanism by which attention modulates neural populations to support sensory processing in cluttered environments. SIGNIFICANCE STATEMENT Although feature-selective attention constitutes one of the building blocks of listening in natural environments, its neural bases remain obscure. To address this, we developed a novel auditory feature-selective attention task and measured noise correlations ( r noise ) in rhesus macaque A1 during task performance. Unlike previous studies showing that the effect of attention on r noise depends on population tuning to the attended feature, we show that the effect of attention depends on the tuning to the distractor feature as well. We suggest that these effects represent an efficient process by which sensory cortex simultaneously enhances relevant information and suppresses irrelevant information. Copyright © 2017 the authors 0270-6474/17/375378-15$15.00/0.

  12. Feature-Selective Attention Adaptively Shifts Noise Correlations in Primary Auditory Cortex

    PubMed Central

    2017-01-01

    Sensory environments often contain an overwhelming amount of information, with both relevant and irrelevant information competing for neural resources. Feature attention mediates this competition by selecting the sensory features needed to form a coherent percept. How attention affects the activity of populations of neurons to support this process is poorly understood because population coding is typically studied through simulations in which one sensory feature is encoded without competition. Therefore, to study the effects of feature attention on population-based neural coding, investigations must be extended to include stimuli with both relevant and irrelevant features. We measured noise correlations (rnoise) within small neural populations in primary auditory cortex while rhesus macaques performed a novel feature-selective attention task. We found that the effect of feature-selective attention on rnoise depended not only on the population tuning to the attended feature, but also on the tuning to the distractor feature. To attempt to explain how these observed effects might support enhanced perceptual performance, we propose an extension of a simple and influential model in which shifts in rnoise can simultaneously enhance the representation of the attended feature while suppressing the distractor. These findings present a novel mechanism by which attention modulates neural populations to support sensory processing in cluttered environments. SIGNIFICANCE STATEMENT Although feature-selective attention constitutes one of the building blocks of listening in natural environments, its neural bases remain obscure. To address this, we developed a novel auditory feature-selective attention task and measured noise correlations (rnoise) in rhesus macaque A1 during task performance. Unlike previous studies showing that the effect of attention on rnoise depends on population tuning to the attended feature, we show that the effect of attention depends on the tuning to the distractor feature as well. We suggest that these effects represent an efficient process by which sensory cortex simultaneously enhances relevant information and suppresses irrelevant information. PMID:28432139

  13. Higher criticism thresholding: Optimal feature selection when useful features are rare and weak.

    PubMed

    Donoho, David; Jin, Jiashun

    2008-09-30

    In important application fields today-genomics and proteomics are examples-selecting a small subset of useful features is crucial for success of Linear Classification Analysis. We study feature selection by thresholding of feature Z-scores and introduce a principle of threshold selection, based on the notion of higher criticism (HC). For i = 1, 2, ..., p, let pi(i) denote the two-sided P-value associated with the ith feature Z-score and pi((i)) denote the ith order statistic of the collection of P-values. The HC threshold is the absolute Z-score corresponding to the P-value maximizing the HC objective (i/p - pi((i)))/sqrt{i/p(1-i/p)}. We consider a rare/weak (RW) feature model, where the fraction of useful features is small and the useful features are each too weak to be of much use on their own. HC thresholding (HCT) has interesting behavior in this setting, with an intimate link between maximizing the HC objective and minimizing the error rate of the designed classifier, and very different behavior from popular threshold selection procedures such as false discovery rate thresholding (FDRT). In the most challenging RW settings, HCT uses an unconventionally low threshold; this keeps the missed-feature detection rate under better control than FDRT and yields a classifier with improved misclassification performance. Replacing cross-validated threshold selection in the popular Shrunken Centroid classifier with the computationally less expensive and simpler HCT reduces the variance of the selected threshold and the error rate of the constructed classifier. Results on standard real datasets and in asymptotic theory confirm the advantages of HCT.

  14. Higher criticism thresholding: Optimal feature selection when useful features are rare and weak

    PubMed Central

    Donoho, David; Jin, Jiashun

    2008-01-01

    In important application fields today—genomics and proteomics are examples—selecting a small subset of useful features is crucial for success of Linear Classification Analysis. We study feature selection by thresholding of feature Z-scores and introduce a principle of threshold selection, based on the notion of higher criticism (HC). For i = 1, 2, …, p, let πi denote the two-sided P-value associated with the ith feature Z-score and π(i) denote the ith order statistic of the collection of P-values. The HC threshold is the absolute Z-score corresponding to the P-value maximizing the HC objective (i/p − π(i))/i/p(1−i/p). We consider a rare/weak (RW) feature model, where the fraction of useful features is small and the useful features are each too weak to be of much use on their own. HC thresholding (HCT) has interesting behavior in this setting, with an intimate link between maximizing the HC objective and minimizing the error rate of the designed classifier, and very different behavior from popular threshold selection procedures such as false discovery rate thresholding (FDRT). In the most challenging RW settings, HCT uses an unconventionally low threshold; this keeps the missed-feature detection rate under better control than FDRT and yields a classifier with improved misclassification performance. Replacing cross-validated threshold selection in the popular Shrunken Centroid classifier with the computationally less expensive and simpler HCT reduces the variance of the selected threshold and the error rate of the constructed classifier. Results on standard real datasets and in asymptotic theory confirm the advantages of HCT. PMID:18815365

  15. Application-Dedicated Selection of Filters (ADSF) using covariance maximization and orthogonal projection.

    PubMed

    Hadoux, Xavier; Kumar, Dinesh Kant; Sarossy, Marc G; Roger, Jean-Michel; Gorretta, Nathalie

    2016-05-19

    Visible and near-infrared (Vis-NIR) spectra are generated by the combination of numerous low resolution features. Spectral variables are thus highly correlated, which can cause problems for selecting the most appropriate ones for a given application. Some decomposition bases such as Fourier or wavelet generally help highlighting spectral features that are important, but are by nature constraint to have both positive and negative components. Thus, in addition to complicating the selected features interpretability, it impedes their use for application-dedicated sensors. In this paper we have proposed a new method for feature selection: Application-Dedicated Selection of Filters (ADSF). This method relaxes the shape constraint by enabling the selection of any type of user defined custom features. By considering only relevant features, based on the underlying nature of the data, high regularization of the final model can be obtained, even in the small sample size context often encountered in spectroscopic applications. For larger scale deployment of application-dedicated sensors, these predefined feature constraints can lead to application specific optical filters, e.g., lowpass, highpass, bandpass or bandstop filters with positive only coefficients. In a similar fashion to Partial Least Squares, ADSF successively selects features using covariance maximization and deflates their influences using orthogonal projection in order to optimally tune the selection to the data with limited redundancy. ADSF is well suited for spectroscopic data as it can deal with large numbers of highly correlated variables in supervised learning, even with many correlated responses. Copyright © 2016 Elsevier B.V. All rights reserved.

  16. Selective anaerobic oxidation of methane enables direct synthesis of methanol.

    PubMed

    Sushkevich, Vitaly L; Palagin, Dennis; Ranocchiari, Marco; van Bokhoven, Jeroen A

    2017-05-05

    Direct functionalization of methane in natural gas remains a key challenge. We present a direct stepwise method for converting methane into methanol with high selectivity (~97%) over a copper-containing zeolite, based on partial oxidation with water. The activation in helium at 673 kelvin (K), followed by consecutive catalyst exposures to 7 bars of methane and then water at 473 K, consistently produced 0.204 mole of CH 3 OH per mole of copper in zeolite. Isotopic labeling confirmed water as the source of oxygen to regenerate the zeolite active centers and renders methanol desorption energetically favorable. On the basis of in situ x-ray absorption spectroscopy, infrared spectroscopy, and density functional theory calculations, we propose a mechanism involving methane oxidation at Cu II oxide active centers, followed by Cu I reoxidation by water with concurrent formation of hydrogen. Copyright © 2017, American Association for the Advancement of Science.

  17. Dispersion-free radial transmission lines

    DOEpatents

    Caporaso, George J [Livermore, CA; Nelson, Scott D [Patterson, CA

    2011-04-12

    A dispersion-free radial transmission line ("DFRTL") preferably for linear accelerators, having two plane conductors each with a central hole, and an electromagnetically permeable material ("EPM") between the two conductors and surrounding a channel connecting the two holes. At least one of the material parameters of relative magnetic permeability, relative dielectric permittivity, and axial width of the EPM is varied as a function of radius, so that the characteristic impedance of the DFRTL is held substantially constant, and pulse transmission therethrough is substantially dispersion-free. Preferably, the EPM is divided into concentric radial sections, with the varied material parameters held constant in each respective section but stepwise varied between sections as a step function of the radius. The radial widths of the concentric sections are selected so that pulse traversal time across each section is the same, and the varied material parameters of the concentric sections are selected to minimize traversal error.

  18. The hydrodeoxygenation of bioderived furans into alkanes.

    PubMed

    Sutton, Andrew D; Waldie, Fraser D; Wu, Ruilian; Schlaf, Marcel; Silks, Louis A Pete; Gordon, John C

    2013-05-01

    The conversion of biomass into fuels and chemical feedstocks is one part of a drive to reduce the world's dependence on crude oil. For transportation fuels in particular, wholesale replacement of a fuel is logistically problematic, not least because of the infrastructure that is already in place. Here, we describe the catalytic defunctionalization of a series of biomass-derived molecules to provide linear alkanes suitable for use as transportation fuels. These biomass-derived molecules contain a variety of functional groups, including olefins, furan rings and carbonyl groups. We describe the removal of these in either a stepwise process or a one-pot process using common reagents and catalysts under mild reaction conditions to provide n-alkanes in good yields and with high selectivities. Our general synthetic approach is applicable to a range of precursors with different carbon content (chain length). This allows the selective generation of linear alkanes with carbon chain lengths between eight and sixteen carbons.

  19. The hydrodeoxygenation of bioderived furans into alkanes

    NASA Astrophysics Data System (ADS)

    Sutton, Andrew D.; Waldie, Fraser D.; Wu, Ruilian; Schlaf, Marcel; ‘Pete' Silks, Louis A.; Gordon, John C.

    2013-05-01

    The conversion of biomass into fuels and chemical feedstocks is one part of a drive to reduce the world's dependence on crude oil. For transportation fuels in particular, wholesale replacement of a fuel is logistically problematic, not least because of the infrastructure that is already in place. Here, we describe the catalytic defunctionalization of a series of biomass-derived molecules to provide linear alkanes suitable for use as transportation fuels. These biomass-derived molecules contain a variety of functional groups, including olefins, furan rings and carbonyl groups. We describe the removal of these in either a stepwise process or a one-pot process using common reagents and catalysts under mild reaction conditions to provide n-alkanes in good yields and with high selectivities. Our general synthetic approach is applicable to a range of precursors with different carbon content (chain length). This allows the selective generation of linear alkanes with carbon chain lengths between eight and sixteen carbons.

  20. Multisensor-based real-time quality monitoring by means of feature extraction, selection and modeling for Al alloy in arc welding

    NASA Astrophysics Data System (ADS)

    Zhang, Zhifen; Chen, Huabin; Xu, Yanling; Zhong, Jiyong; Lv, Na; Chen, Shanben

    2015-08-01

    Multisensory data fusion-based online welding quality monitoring has gained increasing attention in intelligent welding process. This paper mainly focuses on the automatic detection of typical welding defect for Al alloy in gas tungsten arc welding (GTAW) by means of analzing arc spectrum, sound and voltage signal. Based on the developed algorithms in time and frequency domain, 41 feature parameters were successively extracted from these signals to characterize the welding process and seam quality. Then, the proposed feature selection approach, i.e., hybrid fisher-based filter and wrapper was successfully utilized to evaluate the sensitivity of each feature and reduce the feature dimensions. Finally, the optimal feature subset with 19 features was selected to obtain the highest accuracy, i.e., 94.72% using established classification model. This study provides a guideline for feature extraction, selection and dynamic modeling based on heterogeneous multisensory data to achieve a reliable online defect detection system in arc welding.

  1. SVM-RFE based feature selection and Taguchi parameters optimization for multiclass SVM classifier.

    PubMed

    Huang, Mei-Ling; Hung, Yung-Hsiang; Lee, W M; Li, R K; Jiang, Bo-Ru

    2014-01-01

    Recently, support vector machine (SVM) has excellent performance on classification and prediction and is widely used on disease diagnosis or medical assistance. However, SVM only functions well on two-group classification problems. This study combines feature selection and SVM recursive feature elimination (SVM-RFE) to investigate the classification accuracy of multiclass problems for Dermatology and Zoo databases. Dermatology dataset contains 33 feature variables, 1 class variable, and 366 testing instances; and the Zoo dataset contains 16 feature variables, 1 class variable, and 101 testing instances. The feature variables in the two datasets were sorted in descending order by explanatory power, and different feature sets were selected by SVM-RFE to explore classification accuracy. Meanwhile, Taguchi method was jointly combined with SVM classifier in order to optimize parameters C and γ to increase classification accuracy for multiclass classification. The experimental results show that the classification accuracy can be more than 95% after SVM-RFE feature selection and Taguchi parameter optimization for Dermatology and Zoo databases.

  2. SVM-RFE Based Feature Selection and Taguchi Parameters Optimization for Multiclass SVM Classifier

    PubMed Central

    Huang, Mei-Ling; Hung, Yung-Hsiang; Lee, W. M.; Li, R. K.; Jiang, Bo-Ru

    2014-01-01

    Recently, support vector machine (SVM) has excellent performance on classification and prediction and is widely used on disease diagnosis or medical assistance. However, SVM only functions well on two-group classification problems. This study combines feature selection and SVM recursive feature elimination (SVM-RFE) to investigate the classification accuracy of multiclass problems for Dermatology and Zoo databases. Dermatology dataset contains 33 feature variables, 1 class variable, and 366 testing instances; and the Zoo dataset contains 16 feature variables, 1 class variable, and 101 testing instances. The feature variables in the two datasets were sorted in descending order by explanatory power, and different feature sets were selected by SVM-RFE to explore classification accuracy. Meanwhile, Taguchi method was jointly combined with SVM classifier in order to optimize parameters C and γ to increase classification accuracy for multiclass classification. The experimental results show that the classification accuracy can be more than 95% after SVM-RFE feature selection and Taguchi parameter optimization for Dermatology and Zoo databases. PMID:25295306

  3. Feature selection in feature network models: finding predictive subsets of features with the Positive Lasso.

    PubMed

    Frank, Laurence E; Heiser, Willem J

    2008-05-01

    A set of features is the basis for the network representation of proximity data achieved by feature network models (FNMs). Features are binary variables that characterize the objects in an experiment, with some measure of proximity as response variable. Sometimes features are provided by theory and play an important role in the construction of the experimental conditions. In some research settings, the features are not known a priori. This paper shows how to generate features in this situation and how to select an adequate subset of features that takes into account a good compromise between model fit and model complexity, using a new version of least angle regression that restricts coefficients to be non-negative, called the Positive Lasso. It will be shown that features can be generated efficiently with Gray codes that are naturally linked to the FNMs. The model selection strategy makes use of the fact that FNM can be considered as univariate multiple regression model. A simulation study shows that the proposed strategy leads to satisfactory results if the number of objects is less than or equal to 22. If the number of objects is larger than 22, the number of features selected by our method exceeds the true number of features in some conditions.

  4. Infrared laser driven double proton transfer. An optimal control theory study

    NASA Astrophysics Data System (ADS)

    Abdel-Latif, Mahmoud K.; Kühn, Oliver

    2010-02-01

    Laser control of ultrafast double proton transfer is investigated for a two-dimensional model system describing stepwise and concerted transfer pathways. The pulse design has been done by employing optimal control theory in combination with the multiconfiguration time-dependent Hartree wave packet propagation. The obtained laser fields correspond to multiple pump-dump pulse sequences. Special emphasis is paid to the relative importance of stepwise and concerted transfer pathways for the driven wave packet and its dependence on the parameters of the model Hamiltonian as well as on the propagation time. While stepwise transfer is dominating in all cases considered, for high barrier systems concerted transfer proceeding via tunneling can make a contribution.

  5. Response time of mitochondrial oxygen consumption following stepwise changes in cardiac energy demand.

    PubMed

    van Beek, J H; Westerhof, N

    1990-01-01

    We determined the speed with which mitochondrial oxygen consumption and therefore the mitochondrial ATP-synthesis adapted to changes in metabolic demand in the rabbit heart. This was done by measuring the oxygen uptake of the whole heart during a stepwise change in heart rate and correcting for the time taken by diffusion and by convective transport in the blood vessels. Data for the correction for transport time were obtained from the response of venous oxygen concentration to a stepwise change of arterial oxygen concentration. The time constant of the response of mitochondrial oxygen consumption to a step change in heart rate was found to be 4-8 s.

  6. Application of stepwise multiple regression techniques to inversion of Nimbus 'IRIS' observations.

    NASA Technical Reports Server (NTRS)

    Ohring, G.

    1972-01-01

    Exploratory studies with Nimbus-3 infrared interferometer-spectrometer (IRIS) data indicate that, in addition to temperature, such meteorological parameters as geopotential heights of pressure surfaces, tropopause pressure, and tropopause temperature can be inferred from the observed spectra with the use of simple regression equations. The technique of screening the IRIS spectral data by means of stepwise regression to obtain the best radiation predictors of meteorological parameters is validated. The simplicity of application of the technique and the simplicity of the derived linear regression equations - which contain only a few terms - suggest usefulness for this approach. Based upon the results obtained, suggestions are made for further development and exploitation of the stepwise regression analysis technique.

  7. A Filter Feature Selection Method Based on MFA Score and Redundancy Excluding and It's Application to Tumor Gene Expression Data Analysis.

    PubMed

    Li, Jiangeng; Su, Lei; Pang, Zenan

    2015-12-01

    Feature selection techniques have been widely applied to tumor gene expression data analysis in recent years. A filter feature selection method named marginal Fisher analysis score (MFA score) which is based on graph embedding has been proposed, and it has been widely used mainly because it is superior to Fisher score. Considering the heavy redundancy in gene expression data, we proposed a new filter feature selection technique in this paper. It is named MFA score+ and is based on MFA score and redundancy excluding. We applied it to an artificial dataset and eight tumor gene expression datasets to select important features and then used support vector machine as the classifier to classify the samples. Compared with MFA score, t test and Fisher score, it achieved higher classification accuracy.

  8. An ant colony optimization based feature selection for web page classification.

    PubMed

    Saraç, Esra; Özel, Selma Ayşe

    2014-01-01

    The increased popularity of the web has caused the inclusion of huge amount of information to the web, and as a result of this explosive information growth, automated web page classification systems are needed to improve search engines' performance. Web pages have a large number of features such as HTML/XML tags, URLs, hyperlinks, and text contents that should be considered during an automated classification process. The aim of this study is to reduce the number of features to be used to improve runtime and accuracy of the classification of web pages. In this study, we used an ant colony optimization (ACO) algorithm to select the best features, and then we applied the well-known C4.5, naive Bayes, and k nearest neighbor classifiers to assign class labels to web pages. We used the WebKB and Conference datasets in our experiments, and we showed that using the ACO for feature selection improves both accuracy and runtime performance of classification. We also showed that the proposed ACO based algorithm can select better features with respect to the well-known information gain and chi square feature selection methods.

  9. Best Merge Region Growing with Integrated Probabilistic Classification for Hyperspectral Imagery

    NASA Technical Reports Server (NTRS)

    Tarabalka, Yuliya; Tilton, James C.

    2011-01-01

    A new method for spectral-spatial classification of hyperspectral images is proposed. The method is based on the integration of probabilistic classification within the hierarchical best merge region growing algorithm. For this purpose, preliminary probabilistic support vector machines classification is performed. Then, hierarchical step-wise optimization algorithm is applied, by iteratively merging regions with the smallest Dissimilarity Criterion (DC). The main novelty of this method consists in defining a DC between regions as a function of region statistical and geometrical features along with classification probabilities. Experimental results are presented on a 200-band AVIRIS image of the Northwestern Indiana s vegetation area and compared with those obtained by recently proposed spectral-spatial classification techniques. The proposed method improves classification accuracies when compared to other classification approaches.

  10. Reflectance spectroscopy of fresh whole leaves for the estimation of chemical concentration

    NASA Technical Reports Server (NTRS)

    Curran, Paul J.; Dungan, Jennifer L.; Macler, Bruce A.; Plummer, Stephen E.; Peterson, David L.

    1992-01-01

    Remotely sensed plant-canopy data in the visible and near-IR ranges are used to establish relations between the canopy reflectance and the chemical content of the leaves. The mathematical relation is generated by means of stepwise regression based on the derivative reflectance at certain wavelengths. Fourier filtering and sample control are used to minimize instrument noise and spectral overlap respectively, and absorption features are noted that correspond to sugar and protein. The coefficients of determination between estimated and measured concentrations are at least 0.82 for such substances as starch and chlorophyll. It is recommended in the analysis of remotly sensed canopy data that the chemicals with strong spectral overlaps with the chemical of interest be accounted for in order to estimate foliar chemical concentrations accurately.

  11. Vessel Classification in Cosmo-Skymed SAR Data Using Hierarchical Feature Selection

    NASA Astrophysics Data System (ADS)

    Makedonas, A.; Theoharatos, C.; Tsagaris, V.; Anastasopoulos, V.; Costicoglou, S.

    2015-04-01

    SAR based ship detection and classification are important elements of maritime monitoring applications. Recently, high-resolution SAR data have opened new possibilities to researchers for achieving improved classification results. In this work, a hierarchical vessel classification procedure is presented based on a robust feature extraction and selection scheme that utilizes scale, shape and texture features in a hierarchical way. Initially, different types of feature extraction algorithms are implemented in order to form the utilized feature pool, able to represent the structure, material, orientation and other vessel type characteristics. A two-stage hierarchical feature selection algorithm is utilized next in order to be able to discriminate effectively civilian vessels into three distinct types, in COSMO-SkyMed SAR images: cargos, small ships and tankers. In our analysis, scale and shape features are utilized in order to discriminate smaller types of vessels present in the available SAR data, or shape specific vessels. Then, the most informative texture and intensity features are incorporated in order to be able to better distinguish the civilian types with high accuracy. A feature selection procedure that utilizes heuristic measures based on features' statistical characteristics, followed by an exhaustive research with feature sets formed by the most qualified features is carried out, in order to discriminate the most appropriate combination of features for the final classification. In our analysis, five COSMO-SkyMed SAR data with 2.2m x 2.2m resolution were used to analyse the detailed characteristics of these types of ships. A total of 111 ships with available AIS data were used in the classification process. The experimental results show that this method has good performance in ship classification, with an overall accuracy reaching 83%. Further investigation of additional features and proper feature selection is currently in progress.

  12. Features selection and classification to estimate elbow movements

    NASA Astrophysics Data System (ADS)

    Rubiano, A.; Ramírez, J. L.; El Korso, M. N.; Jouandeau, N.; Gallimard, L.; Polit, O.

    2015-11-01

    In this paper, we propose a novel method to estimate the elbow motion, through the features extracted from electromyography (EMG) signals. The features values are normalized and then compared to identify potential relationships between the EMG signal and the kinematic information as angle and angular velocity. We propose and implement a method to select the best set of features, maximizing the distance between the features that correspond to flexion and extension movements. Finally, we test the selected features as inputs to a non-linear support vector machine in the presence of non-idealistic conditions, obtaining an accuracy of 99.79% in the motion estimation results.

  13. Efficient feature subset selection with probabilistic distance criteria. [pattern recognition

    NASA Technical Reports Server (NTRS)

    Chittineni, C. B.

    1979-01-01

    Recursive expressions are derived for efficiently computing the commonly used probabilistic distance measures as a change in the criteria both when a feature is added to and when a feature is deleted from the current feature subset. A combinatorial algorithm for generating all possible r feature combinations from a given set of s features in (s/r) steps with a change of a single feature at each step is presented. These expressions can also be used for both forward and backward sequential feature selection.

  14. Reactivity of Household Oxygen Bleaches: A Stepwise Laboratory Exercise in High School Chemistry Course

    ERIC Educational Resources Information Center

    Nakano, Masayoshi; Ogasawara, Haruka; Wada, Takeshi; Koga, Nobuyoshi

    2016-01-01

    This paper reports on a learning program designed for high school chemistry classes that involves laboratory exercises using household oxygen bleaches. In this program, students are taught the chemistry of oxygen bleaches through a stepwise inquiry using laboratory exercises organized with different pedagogical intents. Through comparative…

  15. FSMRank: feature selection algorithm for learning to rank.

    PubMed

    Lai, Han-Jiang; Pan, Yan; Tang, Yong; Yu, Rong

    2013-06-01

    In recent years, there has been growing interest in learning to rank. The introduction of feature selection into different learning problems has been proven effective. These facts motivate us to investigate the problem of feature selection for learning to rank. We propose a joint convex optimization formulation which minimizes ranking errors while simultaneously conducting feature selection. This optimization formulation provides a flexible framework in which we can easily incorporate various importance measures and similarity measures of the features. To solve this optimization problem, we use the Nesterov's approach to derive an accelerated gradient algorithm with a fast convergence rate O(1/T(2)). We further develop a generalization bound for the proposed optimization problem using the Rademacher complexities. Extensive experimental evaluations are conducted on the public LETOR benchmark datasets. The results demonstrate that the proposed method shows: 1) significant ranking performance gain compared to several feature selection baselines for ranking, and 2) very competitive performance compared to several state-of-the-art learning-to-rank algorithms.

  16. The concern of emergence of multi-station reaction pathways that might make stepwise the mechanism of the 1,3-dipolar cycloadditions of azides and alkynes

    NASA Astrophysics Data System (ADS)

    Mohtat, Bita; Siadati, Seyyed Amir; Khalilzadeh, Mohammad Ali; Zareyee, Daryoush

    2018-03-01

    After hot debates on the concerted or stepwise nature of the mechanism of the catalyst-free 1,3-dipolar cycloadditions (DC)s, nowadays, it is being believed that for the reaction of each dipole and dipolarophile, there is a possibility that the reaction mechanism becomes stepwise, intermediates emerge, and the reaction becomes non-stereospecific. Yield of even minimal amounts of unwanted side products or stereoisomers as impurities could bring many troubles like difficult purification steps. In this project, we have made attempts to study all probable reaction channels of the azide cycloadditions with two functionalized alkynes, in order to answer this question: "is there any possibility that intermediates evolve in the catalyst-free click 1,3-DC reaction of azide-alkynes?". During the calculations, several multi-station reaction pathways supporting the stepwise and concerted mechanisms were detected. Also, the born-oppenheimer molecular dynamic (BOMD) simulation was used to find trustable geometries which could be emerged during the reaction coordinate.

  17. Cloud field classification based on textural features

    NASA Technical Reports Server (NTRS)

    Sengupta, Sailes Kumar

    1989-01-01

    An essential component in global climate research is accurate cloud cover and type determination. Of the two approaches to texture-based classification (statistical and textural), only the former is effective in the classification of natural scenes such as land, ocean, and atmosphere. In the statistical approach that was adopted, parameters characterizing the stochastic properties of the spatial distribution of grey levels in an image are estimated and then used as features for cloud classification. Two types of textural measures were used. One is based on the distribution of the grey level difference vector (GLDV), and the other on a set of textural features derived from the MaxMin cooccurrence matrix (MMCM). The GLDV method looks at the difference D of grey levels at pixels separated by a horizontal distance d and computes several statistics based on this distribution. These are then used as features in subsequent classification. The MaxMin tectural features on the other hand are based on the MMCM, a matrix whose (I,J)th entry give the relative frequency of occurrences of the grey level pair (I,J) that are consecutive and thresholded local extremes separated by a given pixel distance d. Textural measures are then computed based on this matrix in much the same manner as is done in texture computation using the grey level cooccurrence matrix. The database consists of 37 cloud field scenes from LANDSAT imagery using a near IR visible channel. The classification algorithm used is the well known Stepwise Discriminant Analysis. The overall accuracy was estimated by the percentage or correct classifications in each case. It turns out that both types of classifiers, at their best combination of features, and at any given spatial resolution give approximately the same classification accuracy. A neural network based classifier with a feed forward architecture and a back propagation training algorithm is used to increase the classification accuracy, using these two classes of features. Preliminary results based on the GLDV textural features alone look promising.

  18. Computational Prediction of Protein Epsilon Lysine Acetylation Sites Based on a Feature Selection Method.

    PubMed

    Gao, JianZhao; Tao, Xue-Wen; Zhao, Jia; Feng, Yuan-Ming; Cai, Yu-Dong; Zhang, Ning

    2017-01-01

    Lysine acetylation, as one type of post-translational modifications (PTM), plays key roles in cellular regulations and can be involved in a variety of human diseases. However, it is often high-cost and time-consuming to use traditional experimental approaches to identify the lysine acetylation sites. Therefore, effective computational methods should be developed to predict the acetylation sites. In this study, we developed a position-specific method for epsilon lysine acetylation site prediction. Sequences of acetylated proteins were retrieved from the UniProt database. Various kinds of features such as position specific scoring matrix (PSSM), amino acid factors (AAF), and disorders were incorporated. A feature selection method based on mRMR (Maximum Relevance Minimum Redundancy) and IFS (Incremental Feature Selection) was employed. Finally, 319 optimal features were selected from total 541 features. Using the 319 optimal features to encode peptides, a predictor was constructed based on dagging. As a result, an accuracy of 69.56% with MCC of 0.2792 was achieved. We analyzed the optimal features, which suggested some important factors determining the lysine acetylation sites. We developed a position-specific method for epsilon lysine acetylation site prediction. A set of optimal features was selected. Analysis of the optimal features provided insights into the mechanism of lysine acetylation sites, providing guidance of experimental validation. Copyright© Bentham Science Publishers; For any queries, please email at epub@benthamscience.org.

  19. Skin lesion computational diagnosis of dermoscopic images: Ensemble models based on input feature manipulation.

    PubMed

    Oliveira, Roberta B; Pereira, Aledir S; Tavares, João Manuel R S

    2017-10-01

    The number of deaths worldwide due to melanoma has risen in recent times, in part because melanoma is the most aggressive type of skin cancer. Computational systems have been developed to assist dermatologists in early diagnosis of skin cancer, or even to monitor skin lesions. However, there still remains a challenge to improve classifiers for the diagnosis of such skin lesions. The main objective of this article is to evaluate different ensemble classification models based on input feature manipulation to diagnose skin lesions. Input feature manipulation processes are based on feature subset selections from shape properties, colour variation and texture analysis to generate diversity for the ensemble models. Three subset selection models are presented here: (1) a subset selection model based on specific feature groups, (2) a correlation-based subset selection model, and (3) a subset selection model based on feature selection algorithms. Each ensemble classification model is generated using an optimum-path forest classifier and integrated with a majority voting strategy. The proposed models were applied on a set of 1104 dermoscopic images using a cross-validation procedure. The best results were obtained by the first ensemble classification model that generates a feature subset ensemble based on specific feature groups. The skin lesion diagnosis computational system achieved 94.3% accuracy, 91.8% sensitivity and 96.7% specificity. The input feature manipulation process based on specific feature subsets generated the greatest diversity for the ensemble classification model with very promising results. Copyright © 2017 Elsevier B.V. All rights reserved.

  20. Theoretical study on the reaction mechanism of hydrogenation of furfural to furfuryl alcohol on Lewis acidic BEA zeolites: effects of defect structure and tetravalent metals substitution.

    PubMed

    Injongkol, Yuwanda; Maihom, Thana; Treesukul, Piti; Sirijaraensre, Jakkapan; Boekfa, Bundet; Limtrakul, Jumras

    2017-09-13

    Furfural acquired from agricultural sources is receiving extensive attention in the petrochemical industry as it offers an alternative route to generate more valuable hydrocarbon compounds. Herein, we investigate the furfural hydrogenation to furfuryl alcohol catalyzed by Lewis acidic BEA zeolites at the molecular level by means of the M06-L density functional theory. The mechanistic pictures in the catalytic procedure are revealed. The possible reaction pathways are considered to proceed via either concerted or stepwise mechanisms. With the contribution of zeolite oxygen bridging for the H-H splitting, the rate determining step activation barrier for the stepwise mechanism is 14.7 kcal mol -1 lower than that for the concerted mechanism. The stepwise reaction therefore seems to be favored compared to the concerted one. The catalytic effect of the defect zeolite framework on the stepwise mechanism is also investigated. The activation energy for the stepwise rate-determining step over this site is significantly lower than the corresponding step over the perfect one by 14.1 kcal mol -1 . Finally, the catalytic activity of tetravalent metal centers (Sn, Ge, Zr and Hf) substituted in BEA is also preliminarily compared and it is found to follow the order of Hf > Zr > Sn > Ge based on activation energies and the reaction rate. The difference in the activation energy can be traced back to the difference in the charge transfer from the catalytic site to the adsorbed molecules.

  1. Controlled breathing protocols probe human autonomic cardiovascular rhythms

    NASA Technical Reports Server (NTRS)

    Cooke, W. H.; Cox, J. F.; Diedrich, A. M.; Taylor, J. A.; Beightol, L. A.; Ames, J. E. 4th; Hoag, J. B.; Seidel, H.; Eckberg, D. L.

    1998-01-01

    The purpose of this study was to determine how breathing protocols requiring varying degrees of control affect cardiovascular dynamics. We measured inspiratory volume, end-tidal CO2, R-R interval, and arterial pressure spectral power in 10 volunteers who followed the following 5 breathing protocols: 1) uncontrolled breathing for 5 min; 2) stepwise frequency breathing (at 0.3, 0.25, 0.2, 0.15, 0.1, and 0.05 Hz for 2 min each); 3) stepwise frequency breathing as above, but with prescribed tidal volumes; 4) random-frequency breathing (approximately 0.5-0.05 Hz) for 6 min; and 5) fixed-frequency breathing (0.25 Hz) for 5 min. During stepwise breathing, R-R interval and arterial pressure spectral power increased as breathing frequency decreased. Control of inspired volume reduced R-R interval spectral power during 0.1 Hz breathing (P < 0.05). Stepwise and random-breathing protocols yielded comparable coherence and transfer functions between respiration and R-R intervals and systolic pressure and R-R intervals. Random- and fixed-frequency breathing reduced end-tidal CO2 modestly (P < 0.05). Our data suggest that stringent tidal volume control attenuates low-frequency R-R interval oscillations and that fixed- and random-rate breathing may decrease CO2 chemoreceptor stimulation. We conclude that autonomic rhythms measured during different breathing protocols have much in common but that a stepwise protocol without stringent control of inspired volume may allow for the most efficient assessment of short-term respiratory-mediated autonomic oscillations.

  2. Toward high-throughput phenotyping: unbiased automated feature extraction and selection from knowledge sources.

    PubMed

    Yu, Sheng; Liao, Katherine P; Shaw, Stanley Y; Gainer, Vivian S; Churchill, Susanne E; Szolovits, Peter; Murphy, Shawn N; Kohane, Isaac S; Cai, Tianxi

    2015-09-01

    Analysis of narrative (text) data from electronic health records (EHRs) can improve population-scale phenotyping for clinical and genetic research. Currently, selection of text features for phenotyping algorithms is slow and laborious, requiring extensive and iterative involvement by domain experts. This paper introduces a method to develop phenotyping algorithms in an unbiased manner by automatically extracting and selecting informative features, which can be comparable to expert-curated ones in classification accuracy. Comprehensive medical concepts were collected from publicly available knowledge sources in an automated, unbiased fashion. Natural language processing (NLP) revealed the occurrence patterns of these concepts in EHR narrative notes, which enabled selection of informative features for phenotype classification. When combined with additional codified features, a penalized logistic regression model was trained to classify the target phenotype. The authors applied our method to develop algorithms to identify patients with rheumatoid arthritis and coronary artery disease cases among those with rheumatoid arthritis from a large multi-institutional EHR. The area under the receiver operating characteristic curves (AUC) for classifying RA and CAD using models trained with automated features were 0.951 and 0.929, respectively, compared to the AUCs of 0.938 and 0.929 by models trained with expert-curated features. Models trained with NLP text features selected through an unbiased, automated procedure achieved comparable or slightly higher accuracy than those trained with expert-curated features. The majority of the selected model features were interpretable. The proposed automated feature extraction method, generating highly accurate phenotyping algorithms with improved efficiency, is a significant step toward high-throughput phenotyping. © The Author 2015. Published by Oxford University Press on behalf of the American Medical Informatics Association. All rights reserved. For Permissions, please email: journals.permissions@oup.com.

  3. Compact Representation of High-Dimensional Feature Vectors for Large-Scale Image Recognition and Retrieval.

    PubMed

    Zhang, Yu; Wu, Jianxin; Cai, Jianfei

    2016-05-01

    In large-scale visual recognition and image retrieval tasks, feature vectors, such as Fisher vector (FV) or the vector of locally aggregated descriptors (VLAD), have achieved state-of-the-art results. However, the combination of the large numbers of examples and high-dimensional vectors necessitates dimensionality reduction, in order to reduce its storage and CPU costs to a reasonable range. In spite of the popularity of various feature compression methods, this paper shows that the feature (dimension) selection is a better choice for high-dimensional FV/VLAD than the feature (dimension) compression methods, e.g., product quantization. We show that strong correlation among the feature dimensions in the FV and the VLAD may not exist, which renders feature selection a natural choice. We also show that, many dimensions in FV/VLAD are noise. Throwing them away using feature selection is better than compressing them and useful dimensions altogether using feature compression methods. To choose features, we propose an efficient importance sorting algorithm considering both the supervised and unsupervised cases, for visual recognition and image retrieval, respectively. Combining with the 1-bit quantization, feature selection has achieved both higher accuracy and less computational cost than feature compression methods, such as product quantization, on the FV and the VLAD image representations.

  4. Automatic MeSH term assignment and quality assessment.

    PubMed Central

    Kim, W.; Aronson, A. R.; Wilbur, W. J.

    2001-01-01

    For computational purposes documents or other objects are most often represented by a collection of individual attributes that may be strings or numbers. Such attributes are often called features and success in solving a given problem can depend critically on the nature of the features selected to represent documents. Feature selection has received considerable attention in the machine learning literature. In the area of document retrieval we refer to feature selection as indexing. Indexing has not traditionally been evaluated by the same methods used in machine learning feature selection. Here we show how indexing quality may be evaluated in a machine learning setting and apply this methodology to results of the Indexing Initiative at the National Library of Medicine. PMID:11825203

  5. Optimization of breast mass classification using sequential forward floating selection (SFFS) and a support vector machine (SVM) model

    PubMed Central

    Tan, Maxine; Pu, Jiantao; Zheng, Bin

    2014-01-01

    Purpose: Improving radiologists’ performance in classification between malignant and benign breast lesions is important to increase cancer detection sensitivity and reduce false-positive recalls. For this purpose, developing computer-aided diagnosis (CAD) schemes has been attracting research interest in recent years. In this study, we investigated a new feature selection method for the task of breast mass classification. Methods: We initially computed 181 image features based on mass shape, spiculation, contrast, presence of fat or calcifications, texture, isodensity, and other morphological features. From this large image feature pool, we used a sequential forward floating selection (SFFS)-based feature selection method to select relevant features, and analyzed their performance using a support vector machine (SVM) model trained for the classification task. On a database of 600 benign and 600 malignant mass regions of interest (ROIs), we performed the study using a ten-fold cross-validation method. Feature selection and optimization of the SVM parameters were conducted on the training subsets only. Results: The area under the receiver operating characteristic curve (AUC) = 0.805±0.012 was obtained for the classification task. The results also showed that the most frequently-selected features by the SFFS-based algorithm in 10-fold iterations were those related to mass shape, isodensity and presence of fat, which are consistent with the image features frequently used by radiologists in the clinical environment for mass classification. The study also indicated that accurately computing mass spiculation features from the projection mammograms was difficult, and failed to perform well for the mass classification task due to tissue overlap within the benign mass regions. Conclusions: In conclusion, this comprehensive feature analysis study provided new and valuable information for optimizing computerized mass classification schemes that may have potential to be useful as a “second reader” in future clinical practice. PMID:24664267

  6. A stepwise protocol for the treatment of refractory gastroesophageal reflux-induced chronic cough

    PubMed Central

    Xu, Xianghuai; Lv, Hanjing; Yu, Li; Chen, Qiang; Liang, Siwei

    2016-01-01

    Background Refractory gastroesophageal reflux-induced chronic cough (GERC) is difficult to manage. The purpose of the study is to evaluate the efficacy of a novel stepwise protocol for treating this condition. Methods A total of 103 consecutive patients with suspected refractory reflux-induced chronic cough failing to a standard anti-reflux therapy were treated with a stepwise therapy. Treatment commences with high-dose omeprazole and, if necessary, is escalated to subsequent sequential treatment with ranitidine and finally baclofen. The primary end-point was overall cough resolution, and the secondary end-point was cough resolution after each treatment step. Results High-dose omeprazole eliminated or improved cough in 28.1% of patients (n=29). Further stepwise of treatment with the addition of ranitide yielded a favorable response in an additional 12.6% (n=13) of patients, and subsequent escalation to baclofen provoked response in another 36.9% (n=38) of patients. Overall, this stepwise protocol was successful in 77.6% (n=80) of patients. The diurnal cough symptom score fell from 3 [1] to 1 [0] (Z=6.316, P=0.000), and the nocturnal cough symptom score decreased from 1 [1] to 0 [1] (Z=–4.511, P=0.000), with a corresponding reduction in the Gastroesophageal Reflux Diagnostic Questionnaire score from 8.6±1.7 to 6.8±0.7 (t=3.612, P=0.000). Conversely, the cough threshold C2 to capsaicin was increased from 0.49 (0.49) µmol/L to 1.95 (2.92) µmol/L (Z=–5.892, P=0.000), and the cough threshold C5 was increased from 1.95 (2.92) µmol/L to 7.8 (5.85) µmol/L (Z=–5.171, P=0.000). Conclusions Sequential stepwise anti-reflux therapy is a useful therapeutic strategy for refractory reflux-induced chronic cough. PMID:26904227

  7. Checklist/Guide to Selecting a Small Computer.

    ERIC Educational Resources Information Center

    Bennett, Wilma E.

    This 322-point checklist was designed to help executives make an intelligent choice when selecting a small computer for a business. For ease of use the questions have been divided into ten categories: Display Features, Keyboard Features, Printer Features, Controller Features, Software, Word Processing, Service, Training, Miscellaneous, and Costs.…

  8. Feature selection methods for object-based classification of sub-decimeter resolution digital aerial imagery

    USDA-ARS?s Scientific Manuscript database

    Due to the availability of numerous spectral, spatial, and contextual features, the determination of optimal features and class separabilities can be a time consuming process in object-based image analysis (OBIA). While several feature selection methods have been developed to assist OBIA, a robust c...

  9. News video story segmentation method using fusion of audio-visual features

    NASA Astrophysics Data System (ADS)

    Wen, Jun; Wu, Ling-da; Zeng, Pu; Luan, Xi-dao; Xie, Yu-xiang

    2007-11-01

    News story segmentation is an important aspect for news video analysis. This paper presents a method for news video story segmentation. Different form prior works, which base on visual features transform, the proposed technique uses audio features as baseline and fuses visual features with it to refine the results. At first, it selects silence clips as audio features candidate points, and selects shot boundaries and anchor shots as two kinds of visual features candidate points. Then this paper selects audio feature candidates as cues and develops different fusion method, which effectively using diverse type visual candidates to refine audio candidates, to get story boundaries. Experiment results show that this method has high efficiency and adaptability to different kinds of news video.

  10. A comparative analysis of swarm intelligence techniques for feature selection in cancer classification.

    PubMed

    Gunavathi, Chellamuthu; Premalatha, Kandasamy

    2014-01-01

    Feature selection in cancer classification is a central area of research in the field of bioinformatics and used to select the informative genes from thousands of genes of the microarray. The genes are ranked based on T-statistics, signal-to-noise ratio (SNR), and F-test values. The swarm intelligence (SI) technique finds the informative genes from the top-m ranked genes. These selected genes are used for classification. In this paper the shuffled frog leaping with Lévy flight (SFLLF) is proposed for feature selection. In SFLLF, the Lévy flight is included to avoid premature convergence of shuffled frog leaping (SFL) algorithm. The SI techniques such as particle swarm optimization (PSO), cuckoo search (CS), SFL, and SFLLF are used for feature selection which identifies informative genes for classification. The k-nearest neighbour (k-NN) technique is used to classify the samples. The proposed work is applied on 10 different benchmark datasets and examined with SI techniques. The experimental results show that the results obtained from k-NN classifier through SFLLF feature selection method outperform PSO, CS, and SFL.

  11. Feature selection for elderly faller classification based on wearable sensors.

    PubMed

    Howcroft, Jennifer; Kofman, Jonathan; Lemaire, Edward D

    2017-05-30

    Wearable sensors can be used to derive numerous gait pattern features for elderly fall risk and faller classification; however, an appropriate feature set is required to avoid high computational costs and the inclusion of irrelevant features. The objectives of this study were to identify and evaluate smaller feature sets for faller classification from large feature sets derived from wearable accelerometer and pressure-sensing insole gait data. A convenience sample of 100 older adults (75.5 ± 6.7 years; 76 non-fallers, 24 fallers based on 6 month retrospective fall occurrence) walked 7.62 m while wearing pressure-sensing insoles and tri-axial accelerometers at the head, pelvis, left and right shanks. Feature selection was performed using correlation-based feature selection (CFS), fast correlation based filter (FCBF), and Relief-F algorithms. Faller classification was performed using multi-layer perceptron neural network, naïve Bayesian, and support vector machine classifiers, with 75:25 single stratified holdout and repeated random sampling. The best performing model was a support vector machine with 78% accuracy, 26% sensitivity, 95% specificity, 0.36 F1 score, and 0.31 MCC and one posterior pelvis accelerometer input feature (left acceleration standard deviation). The second best model achieved better sensitivity (44%) and used a support vector machine with 74% accuracy, 83% specificity, 0.44 F1 score, and 0.29 MCC. This model had ten input features: maximum, mean and standard deviation posterior acceleration; maximum, mean and standard deviation anterior acceleration; mean superior acceleration; and three impulse features. The best multi-sensor model sensitivity (56%) was achieved using posterior pelvis and both shank accelerometers and a naïve Bayesian classifier. The best single-sensor model sensitivity (41%) was achieved using the posterior pelvis accelerometer and a naïve Bayesian classifier. Feature selection provided models with smaller feature sets and improved faller classification compared to faller classification without feature selection. CFS and FCBF provided the best feature subset (one posterior pelvis accelerometer feature) for faller classification. However, better sensitivity was achieved by the second best model based on a Relief-F feature subset with three pressure-sensing insole features and seven head accelerometer features. Feature selection should be considered as an important step in faller classification using wearable sensors.

  12. Select Features in "Finale 2011" for Music Educators

    ERIC Educational Resources Information Center

    Thompson, Douglas Earl

    2011-01-01

    A feature-laden software program such as "Finale" is an overwhelming tool to master--if one hopes to master many features in a short amount of time. Believing that working with a fewer number of features can be a helpful approach, this article looks at a select number of features in "Finale 2011" of obvious use to music educators. These features…

  13. Effects of Spatial and Feature Attention on Disparity-Rendered Structure-From-Motion Stimuli in the Human Visual Cortex

    PubMed Central

    Ip, Ifan Betina; Bridge, Holly; Parker, Andrew J.

    2014-01-01

    An important advance in the study of visual attention has been the identification of a non-spatial component of attention that enhances the response to similar features or objects across the visual field. Here we test whether this non-spatial component can co-select individual features that are perceptually bound into a coherent object. We combined human psychophysics and functional magnetic resonance imaging (fMRI) to demonstrate the ability to co-select individual features from perceptually coherent objects. Our study used binocular disparity and visual motion to define disparity structure-from-motion (dSFM) stimuli. Although the spatial attention system induced strong modulations of the fMRI response in visual regions, the non-spatial system’s ability to co-select features of the dSFM stimulus was less pronounced and variable across subjects. Our results demonstrate that feature and global feature attention effects are variable across participants, suggesting that the feature attention system may be limited in its ability to automatically select features within the attended object. Careful comparison of the task design suggests that even minor differences in the perceptual task may be critical in revealing the presence of global feature attention. PMID:24936974

  14. IMMAN: free software for information theory-based chemometric analysis.

    PubMed

    Urias, Ricardo W Pino; Barigye, Stephen J; Marrero-Ponce, Yovani; García-Jacas, César R; Valdes-Martiní, José R; Perez-Gimenez, Facundo

    2015-05-01

    The features and theoretical background of a new and free computational program for chemometric analysis denominated IMMAN (acronym for Information theory-based CheMoMetrics ANalysis) are presented. This is multi-platform software developed in the Java programming language, designed with a remarkably user-friendly graphical interface for the computation of a collection of information-theoretic functions adapted for rank-based unsupervised and supervised feature selection tasks. A total of 20 feature selection parameters are presented, with the unsupervised and supervised frameworks represented by 10 approaches in each case. Several information-theoretic parameters traditionally used as molecular descriptors (MDs) are adapted for use as unsupervised rank-based feature selection methods. On the other hand, a generalization scheme for the previously defined differential Shannon's entropy is discussed, as well as the introduction of Jeffreys information measure for supervised feature selection. Moreover, well-known information-theoretic feature selection parameters, such as information gain, gain ratio, and symmetrical uncertainty are incorporated to the IMMAN software ( http://mobiosd-hub.com/imman-soft/ ), following an equal-interval discretization approach. IMMAN offers data pre-processing functionalities, such as missing values processing, dataset partitioning, and browsing. Moreover, single parameter or ensemble (multi-criteria) ranking options are provided. Consequently, this software is suitable for tasks like dimensionality reduction, feature ranking, as well as comparative diversity analysis of data matrices. Simple examples of applications performed with this program are presented. A comparative study between IMMAN and WEKA feature selection tools using the Arcene dataset was performed, demonstrating similar behavior. In addition, it is revealed that the use of IMMAN unsupervised feature selection methods improves the performance of both IMMAN and WEKA supervised algorithms. Graphic representation for Shannon's distribution of MD calculating software.

  15. How are patients influenced when counseled for minimally invasive lumbar spine surgeries? A stepwise model assessing pivotal information for decision-making.

    PubMed

    Hey, Hwee Weng Dennis; Kumar, Nishant; Teo, Alex Quok An; Tan, Kimberly-Anne; Kumar, Naresh; Liu, Ka-Po Gabriel; Wong, Hee-Kit

    2017-08-01

    Although minimally invasive surgery (MIS)-transforaminal lumbar interbody fusion (TLIF) has many evidence-based short-term benefits over open TLIF, both procedures have similar long-term outcomes. Patients' preference for MIS over open TLIF may be confounded by a lack of understanding of what each approach entails. The study aimed to identify the various factors influencing patients' choice between MIS and open TLIF. This is a cross-sectional study conducted at a tertiary health-care institution. Patients, for whom TLIF procedures were indicated, were recruited over a 3-month period from specialist outpatient clinics. The outcome measure was patients' choice of surgical approach (MIS or open). All patients were subjected to a stepwise interviewing process and were asked to select between open and MIS approaches at each step. Further subgroup analysis stratifying subjects based on stages of decision-making was performed to identify key predictors of selection changes. No sources of funding were required for this study and there are no conflicts of interests. Fifty-four patients with a mean age of 55.8 years participated in the study. Thirteen (24.1%) consistently selected a single approach, whereas 31 (57.4%) changed their selection more than once during the interviewing process. Overall, 12 patients (22.2%) had a final decision different from their initial choice, and 15 patients (27.8%) were unable to decide. A large proportion of patients (65.0%) initially favored the open approach's midline incision. This proportion dropped to 16.7% (p<.001) upon mention of the term MIS. The proportion of patients favoring MIS dropped significantly following discussion on the pros and cons (p=.002) of each approach, as well as conversion or revision surgery (p=.017). Radiation and cosmesis were identified as the two most important factors influencing patients' final decisions. The longer midline incision of the open approach is cosmetically more appealing to patients than the paramedian stab wounds of MIS. The advantages of the MIS approach may not be as valued by patients as they are by surgeons. Given the equivalent long-term outcomes of both approaches, it is crucial that patients are adequately informed during preoperative counseling to achieve the best consensus decision. Copyright © 2017 Elsevier Inc. All rights reserved.

  16. Efficient least angle regression for identification of linear-in-the-parameters models

    PubMed Central

    Beach, Thomas H.; Rezgui, Yacine

    2017-01-01

    Least angle regression, as a promising model selection method, differentiates itself from conventional stepwise and stagewise methods, in that it is neither too greedy nor too slow. It is closely related to L1 norm optimization, which has the advantage of low prediction variance through sacrificing part of model bias property in order to enhance model generalization capability. In this paper, we propose an efficient least angle regression algorithm for model selection for a large class of linear-in-the-parameters models with the purpose of accelerating the model selection process. The entire algorithm works completely in a recursive manner, where the correlations between model terms and residuals, the evolving directions and other pertinent variables are derived explicitly and updated successively at every subset selection step. The model coefficients are only computed when the algorithm finishes. The direct involvement of matrix inversions is thereby relieved. A detailed computational complexity analysis indicates that the proposed algorithm possesses significant computational efficiency, compared with the original approach where the well-known efficient Cholesky decomposition is involved in solving least angle regression. Three artificial and real-world examples are employed to demonstrate the effectiveness, efficiency and numerical stability of the proposed algorithm. PMID:28293140

  17. Biomarker selection for medical diagnosis using the partial area under the ROC curve

    PubMed Central

    2014-01-01

    Background A biomarker is usually used as a diagnostic or assessment tool in medical research. Finding an ideal biomarker is not easy and combining multiple biomarkers provides a promising alternative. Moreover, some biomarkers based on the optimal linear combination do not have enough discriminatory power. As a result, the aim of this study was to find the significant biomarkers based on the optimal linear combination maximizing the pAUC for assessment of the biomarkers. Methods Under the binormality assumption we obtain the optimal linear combination of biomarkers maximizing the partial area under the receiver operating characteristic curve (pAUC). Related statistical tests are developed for assessment of a biomarker set and of an individual biomarker. Stepwise biomarker selections are introduced to identify those biomarkers of statistical significance. Results The results of simulation study and three real examples, Duchenne Muscular Dystrophy disease, heart disease, and breast tissue example are used to show that our methods are most suitable biomarker selection for the data sets of a moderate number of biomarkers. Conclusions Our proposed biomarker selection approaches can be used to find the significant biomarkers based on hypothesis testing. PMID:24410929

  18. Oculomotor selection underlies feature retention in visual working memory.

    PubMed

    Hanning, Nina M; Jonikaitis, Donatas; Deubel, Heiner; Szinte, Martin

    2016-02-01

    Oculomotor selection, spatial task relevance, and visual working memory (WM) are described as three processes highly intertwined and sustained by similar cortical structures. However, because task-relevant locations always constitute potential saccade targets, no study so far has been able to distinguish between oculomotor selection and spatial task relevance. We designed an experiment that allowed us to dissociate in humans the contribution of task relevance, oculomotor selection, and oculomotor execution to the retention of feature representations in WM. We report that task relevance and oculomotor selection lead to dissociable effects on feature WM maintenance. In a first task, in which an object's location was encoded as a saccade target, its feature representations were successfully maintained in WM, whereas they declined at nonsaccade target locations. Likewise, we observed a similar WM benefit at the target of saccades that were prepared but never executed. In a second task, when an object's location was marked as task relevant but constituted a nonsaccade target (a location to avoid), feature representations maintained at that location did not benefit. Combined, our results demonstrate that oculomotor selection is consistently associated with WM, whereas task relevance is not. This provides evidence for an overlapping circuitry serving saccade target selection and feature-based WM that can be dissociated from processes encoding task-relevant locations. Copyright © 2016 the American Physiological Society.

  19. JCDSA: a joint covariate detection tool for survival analysis on tumor expression profiles.

    PubMed

    Wu, Yiming; Liu, Yanan; Wang, Yueming; Shi, Yan; Zhao, Xudong

    2018-05-29

    Survival analysis on tumor expression profiles has always been a key issue for subsequent biological experimental validation. It is crucial how to select features which closely correspond to survival time. Furthermore, it is important how to select features which best discriminate between low-risk and high-risk group of patients. Common features derived from the two aspects may provide variable candidates for prognosis of cancer. Based on the provided two-step feature selection strategy, we develop a joint covariate detection tool for survival analysis on tumor expression profiles. Significant features, which are not only consistent with survival time but also associated with the categories of patients with different survival risks, are chosen. Using the miRNA expression data (Level 3) of 548 patients with glioblastoma multiforme (GBM) as an example, miRNA candidates for prognosis of cancer are selected. The reliability of selected miRNAs using this tool is demonstrated by 100 simulations. Furthermore, It is discovered that significant covariates are not directly composed of individually significant variables. Joint covariate detection provides a viewpoint for selecting variables which are not individually but jointly significant. Besides, it helps to select features which are not only consistent with survival time but also associated with prognosis risk. The software is available at http://bio-nefu.com/resource/jcdsa .

  20. Adaptive feature selection using v-shaped binary particle swarm optimization.

    PubMed

    Teng, Xuyang; Dong, Hongbin; Zhou, Xiurong

    2017-01-01

    Feature selection is an important preprocessing method in machine learning and data mining. This process can be used not only to reduce the amount of data to be analyzed but also to build models with stronger interpretability based on fewer features. Traditional feature selection methods evaluate the dependency and redundancy of features separately, which leads to a lack of measurement of their combined effect. Moreover, a greedy search considers only the optimization of the current round and thus cannot be a global search. To evaluate the combined effect of different subsets in the entire feature space, an adaptive feature selection method based on V-shaped binary particle swarm optimization is proposed. In this method, the fitness function is constructed using the correlation information entropy. Feature subsets are regarded as individuals in a population, and the feature space is searched using V-shaped binary particle swarm optimization. The above procedure overcomes the hard constraint on the number of features, enables the combined evaluation of each subset as a whole, and improves the search ability of conventional binary particle swarm optimization. The proposed algorithm is an adaptive method with respect to the number of feature subsets. The experimental results show the advantages of optimizing the feature subsets using the V-shaped transfer function and confirm the effectiveness and efficiency of the feature subsets obtained under different classifiers.

Top