Liu, Zitao; Hauskrecht, Milos
2017-11-01
Building of an accurate predictive model of clinical time series for a patient is critical for understanding of the patient condition, its dynamics, and optimal patient management. Unfortunately, this process is not straightforward. First, patient-specific variations are typically large and population-based models derived or learned from many different patients are often unable to support accurate predictions for each individual patient. Moreover, time series observed for one patient at any point in time may be too short and insufficient to learn a high-quality patient-specific model just from the patient's own data. To address these problems we propose, develop and experiment with a new adaptive forecasting framework for building multivariate clinical time series models for a patient and for supporting patient-specific predictions. The framework relies on the adaptive model switching approach that at any point in time selects the most promising time series model out of the pool of many possible models, and consequently, combines advantages of the population, patient-specific and short-term individualized predictive models. We demonstrate that the adaptive model switching framework is very promising approach to support personalized time series prediction, and that it is able to outperform predictions based on pure population and patient-specific models, as well as, other patient-specific model adaptation strategies.
On the effects of alternative optima in context-specific metabolic model predictions
Nikoloski, Zoran
2017-01-01
The integration of experimental data into genome-scale metabolic models can greatly improve flux predictions. This is achieved by restricting predictions to a more realistic context-specific domain, like a particular cell or tissue type. Several computational approaches to integrate data have been proposed—generally obtaining context-specific (sub)models or flux distributions. However, these approaches may lead to a multitude of equally valid but potentially different models or flux distributions, due to possible alternative optima in the underlying optimization problems. Although this issue introduces ambiguity in context-specific predictions, it has not been generally recognized, especially in the case of model reconstructions. In this study, we analyze the impact of alternative optima in four state-of-the-art context-specific data integration approaches, providing both flux distributions and/or metabolic models. To this end, we present three computational methods and apply them to two particular case studies: leaf-specific predictions from the integration of gene expression data in a metabolic model of Arabidopsis thaliana, and liver-specific reconstructions derived from a human model with various experimental data sources. The application of these methods allows us to obtain the following results: (i) we sample the space of alternative flux distributions in the leaf- and the liver-specific case and quantify the ambiguity of the predictions. In addition, we show how the inclusion of ℓ1-regularization during data integration reduces the ambiguity in both cases. (ii) We generate sets of alternative leaf- and liver-specific models that are optimal to each one of the evaluated model reconstruction approaches. We demonstrate that alternative models of the same context contain a marked fraction of disparate reactions. Further, we show that a careful balance between model sparsity and metabolic functionality helps in reducing the discrepancies between alternative models. Finally, our findings indicate that alternative optima must be taken into account for rendering the context-specific metabolic model predictions less ambiguous. PMID:28557990
On the effects of alternative optima in context-specific metabolic model predictions.
Robaina-Estévez, Semidán; Nikoloski, Zoran
2017-05-01
The integration of experimental data into genome-scale metabolic models can greatly improve flux predictions. This is achieved by restricting predictions to a more realistic context-specific domain, like a particular cell or tissue type. Several computational approaches to integrate data have been proposed-generally obtaining context-specific (sub)models or flux distributions. However, these approaches may lead to a multitude of equally valid but potentially different models or flux distributions, due to possible alternative optima in the underlying optimization problems. Although this issue introduces ambiguity in context-specific predictions, it has not been generally recognized, especially in the case of model reconstructions. In this study, we analyze the impact of alternative optima in four state-of-the-art context-specific data integration approaches, providing both flux distributions and/or metabolic models. To this end, we present three computational methods and apply them to two particular case studies: leaf-specific predictions from the integration of gene expression data in a metabolic model of Arabidopsis thaliana, and liver-specific reconstructions derived from a human model with various experimental data sources. The application of these methods allows us to obtain the following results: (i) we sample the space of alternative flux distributions in the leaf- and the liver-specific case and quantify the ambiguity of the predictions. In addition, we show how the inclusion of ℓ1-regularization during data integration reduces the ambiguity in both cases. (ii) We generate sets of alternative leaf- and liver-specific models that are optimal to each one of the evaluated model reconstruction approaches. We demonstrate that alternative models of the same context contain a marked fraction of disparate reactions. Further, we show that a careful balance between model sparsity and metabolic functionality helps in reducing the discrepancies between alternative models. Finally, our findings indicate that alternative optima must be taken into account for rendering the context-specific metabolic model predictions less ambiguous.
Musite, a tool for global prediction of general and kinase-specific phosphorylation sites.
Gao, Jianjiong; Thelen, Jay J; Dunker, A Keith; Xu, Dong
2010-12-01
Reversible protein phosphorylation is one of the most pervasive post-translational modifications, regulating diverse cellular processes in various organisms. High throughput experimental studies using mass spectrometry have identified many phosphorylation sites, primarily from eukaryotes. However, the vast majority of phosphorylation sites remain undiscovered, even in well studied systems. Because mass spectrometry-based experimental approaches for identifying phosphorylation events are costly, time-consuming, and biased toward abundant proteins and proteotypic peptides, in silico prediction of phosphorylation sites is potentially a useful alternative strategy for whole proteome annotation. Because of various limitations, current phosphorylation site prediction tools were not well designed for comprehensive assessment of proteomes. Here, we present a novel software tool, Musite, specifically designed for large scale predictions of both general and kinase-specific phosphorylation sites. We collected phosphoproteomics data in multiple organisms from several reliable sources and used them to train prediction models by a comprehensive machine-learning approach that integrates local sequence similarities to known phosphorylation sites, protein disorder scores, and amino acid frequencies. Application of Musite on several proteomes yielded tens of thousands of phosphorylation site predictions at a high stringency level. Cross-validation tests show that Musite achieves some improvement over existing tools in predicting general phosphorylation sites, and it is at least comparable with those for predicting kinase-specific phosphorylation sites. In Musite V1.0, we have trained general prediction models for six organisms and kinase-specific prediction models for 13 kinases or kinase families. Although the current pretrained models were not correlated with any particular cellular conditions, Musite provides a unique functionality for training customized prediction models (including condition-specific models) from users' own data. In addition, with its easily extensible open source application programming interface, Musite is aimed at being an open platform for community-based development of machine learning-based phosphorylation site prediction applications. Musite is available at http://musite.sourceforge.net/.
Learning Instance-Specific Predictive Models
Visweswaran, Shyam; Cooper, Gregory F.
2013-01-01
This paper introduces a Bayesian algorithm for constructing predictive models from data that are optimized to predict a target variable well for a particular instance. This algorithm learns Markov blanket models, carries out Bayesian model averaging over a set of models to predict a target variable of the instance at hand, and employs an instance-specific heuristic to locate a set of suitable models to average over. We call this method the instance-specific Markov blanket (ISMB) algorithm. The ISMB algorithm was evaluated on 21 UCI data sets using five different performance measures and its performance was compared to that of several commonly used predictive algorithms, including nave Bayes, C4.5 decision tree, logistic regression, neural networks, k-Nearest Neighbor, Lazy Bayesian Rules, and AdaBoost. Over all the data sets, the ISMB algorithm performed better on average on all performance measures against all the comparison algorithms. PMID:25045325
Practical approach to subject-specific estimation of knee joint contact force.
Knarr, Brian A; Higginson, Jill S
2015-08-20
Compressive forces experienced at the knee can significantly contribute to cartilage degeneration. Musculoskeletal models enable predictions of the internal forces experienced at the knee, but validation is often not possible, as experimental data detailing loading at the knee joint is limited. Recently available data reporting compressive knee force through direct measurement using instrumented total knee replacements offer a unique opportunity to evaluate the accuracy of models. Previous studies have highlighted the importance of subject-specificity in increasing the accuracy of model predictions; however, these techniques may be unrealistic outside of a research setting. Therefore, the goal of our work was to identify a practical approach for accurate prediction of tibiofemoral knee contact force (KCF). Four methods for prediction of knee contact force were compared: (1) standard static optimization, (2) uniform muscle coordination weighting, (3) subject-specific muscle coordination weighting and (4) subject-specific strength adjustments. Walking trials for three subjects with instrumented knee replacements were used to evaluate the accuracy of model predictions. Predictions utilizing subject-specific muscle coordination weighting yielded the best agreement with experimental data; however this method required in vivo data for weighting factor calibration. Including subject-specific strength adjustments improved models' predictions compared to standard static optimization, with errors in peak KCF less than 0.5 body weight for all subjects. Overall, combining clinical assessments of muscle strength with standard tools available in the OpenSim software package, such as inverse kinematics and static optimization, appears to be a practical method for predicting joint contact force that can be implemented for many applications. Copyright © 2015 Elsevier Ltd. All rights reserved.
Discriminative value of FRAX for fracture prediction in a cohort of Chinese postmenopausal women.
Cheung, E Y N; Bow, C H; Cheung, C L; Soong, C; Yeung, S; Loong, C; Kung, A
2012-03-01
We followed 2,266 postmenopausal Chinese women for 4.5 years to determine which model best predicts osteoporotic fracture. A model that contains ethnic-specific risk factors, some of which reflect frailty, performed as well as or better than the well-established FRAX model. Clinical risk assessment, with or without T-score, can predict fractures in Chinese postmenopausal women although it is unknown which combination of clinical risk factors is most effective. This prospective study sought to compare the accuracy for fracture prediction using various models including FRAX, our ethnic-specific clinical risk factors (CRF) and other simple models. This study is part of the Hong Kong Osteoporosis Study. A total of 2,266 treatment naïve postmenopausal women underwent clinical risk factor and bone mineral density assessment. Subjects were followed up for outcome of major osteoporotic fracture and receiver operating characteristic (ROC) curves for different models were compared. The percentage of subjects in different quartiles of risk according to various models who actually fractured was also compared. The mean age at baseline was 62.1 ± 8.5 years and mean follow-up time was 4.5 ± 2.8 years. A total of 106 new major osteoporotic fractures were reported, of which 21 were hip fractures. Ethnic-specific CRF with T-score performed better than FRAX with T-score (based on both Chinese normative and National Health and Nutrition Examination Survey (NHANES) databases) in terms of AUC comparison for prediction of major osteoporotic fracture. The two models were similar in hip fracture prediction. The ethnic-specific CRF model had a 10% higher sensitivity than FRAX at a specificity of 0.8 or above. CRF related to frailty and differences in lifestyle between populations are likely to be important in fracture prediction. Further work is required to determine which and how CRF can be applied to develop a fracture prediction model in our population.
Dantigny, Philippe; Guilmart, Audrey; Bensoussan, Maurice
2005-04-15
For over 20 years, predictive microbiology focused on food-pathogenic bacteria. Few studies concerned modelling fungal development. On one hand, most of food mycologists are not familiar with modelling techniques; on the other hand, people involved in modelling are developing tools dedicated to bacteria. Therefore, there is a tendency to extend the use of models that were developed for bacteria to moulds. However, some mould specificities should be taken into account. The use of specific models for predicting germination and growth of fungi was advocated previously []. This paper provides a short review of fungal modelling studies.
Predicting survival across chronic interstitial lung disease: the ILD-GAP model.
Ryerson, Christopher J; Vittinghoff, Eric; Ley, Brett; Lee, Joyce S; Mooney, Joshua J; Jones, Kirk D; Elicker, Brett M; Wolters, Paul J; Koth, Laura L; King, Talmadge E; Collard, Harold R
2014-04-01
Risk prediction is challenging in chronic interstitial lung disease (ILD) because of heterogeneity in disease-specific and patient-specific variables. Our objective was to determine whether mortality is accurately predicted in patients with chronic ILD using the GAP model, a clinical prediction model based on sex, age, and lung physiology, that was previously validated in patients with idiopathic pulmonary fibrosis. Patients with idiopathic pulmonary fibrosis (n=307), chronic hypersensitivity pneumonitis (n=206), connective tissue disease-associated ILD (n=281), idiopathic nonspecific interstitial pneumonia (n=45), or unclassifiable ILD (n=173) were selected from an ongoing database (N=1,012). Performance of the previously validated GAP model was compared with novel prediction models in each ILD subtype and the combined cohort. Patients with follow-up pulmonary function data were used for longitudinal model validation. The GAP model had good performance in all ILD subtypes (c-index, 74.6 in the combined cohort), which was maintained at all stages of disease severity and during follow-up evaluation. The GAP model had similar performance compared with alternative prediction models. A modified ILD-GAP Index was developed for application across all ILD subtypes to provide disease-specific survival estimates using a single risk prediction model. This was done by adding a disease subtype variable that accounted for better adjusted survival in connective tissue disease-associated ILD, chronic hypersensitivity pneumonitis, and idiopathic nonspecific interstitial pneumonia. The GAP model accurately predicts risk of death in chronic ILD. The ILD-GAP model accurately predicts mortality in major chronic ILD subtypes and at all stages of disease.
Maciejewski, Matthew L; Liu, Chuan-Fen; Fihn, Stephan D
2009-01-01
To compare the ability of generic comorbidity and risk adjustment measures, a diabetes-specific measure, and a self-reported functional status measure to explain variation in health care expenditures for individuals with diabetes. This study included a retrospective cohort of 3,092 diabetic veterans participating in a multisite trial. Two comorbidity measures, four risk adjusters, a functional status measure, a diabetes complication count, and baseline expenditures were constructed from administrative and survey data. Outpatient, inpatient, and total expenditure models were estimated using ordinary least squares regression. Adjusted R(2) statistics and predictive ratios were compared across measures to assess overall explanatory power and explanatory power of low- and high-cost subgroups. Administrative data-based risk adjusters performed better than the comorbidity, functional status, and diabetes-specific measures in all expenditure models. The diagnostic cost groups (DCGs) measure had the greatest predictive power overall and for the low- and high-cost subgroups, while the diabetes-specific measure had the lowest predictive power. A model with DCGs and the diabetes-specific measure modestly improved predictive power. Existing generic measures can be useful for diabetes-specific research and policy applications, but more predictive diabetes-specific measures are needed.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Valencia, Antoni; Prous, Josep; Mora, Oscar
As indicated in ICH M7 draft guidance, in silico predictive tools including statistically-based QSARs and expert analysis may be used as a computational assessment for bacterial mutagenicity for the qualification of impurities in pharmaceuticals. To address this need, we developed and validated a QSAR model to predict Salmonella t. mutagenicity (Ames assay outcome) of pharmaceutical impurities using Prous Institute's Symmetry℠, a new in silico solution for drug discovery and toxicity screening, and the Mold2 molecular descriptor package (FDA/NCTR). Data was sourced from public benchmark databases with known Ames assay mutagenicity outcomes for 7300 chemicals (57% mutagens). Of these data, 90%more » was used to train the model and the remaining 10% was set aside as a holdout set for validation. The model's applicability to drug impurities was tested using a FDA/CDER database of 951 structures, of which 94% were found within the model's applicability domain. The predictive performance of the model is acceptable for supporting regulatory decision-making with 84 ± 1% sensitivity, 81 ± 1% specificity, 83 ± 1% concordance and 79 ± 1% negative predictivity based on internal cross-validation, while the holdout dataset yielded 83% sensitivity, 77% specificity, 80% concordance and 78% negative predictivity. Given the importance of having confidence in negative predictions, an additional external validation of the model was also carried out, using marketed drugs known to be Ames-negative, and obtained 98% coverage and 81% specificity. Additionally, Ames mutagenicity data from FDA/CFSAN was used to create another data set of 1535 chemicals for external validation of the model, yielding 98% coverage, 73% sensitivity, 86% specificity, 81% concordance and 84% negative predictivity. - Highlights: • A new in silico QSAR model to predict Ames mutagenicity is described. • The model is extensively validated with chemicals from the FDA and the public domain. • Validation tests show desirable high sensitivity and high negative predictivity. • The model predicted 14 reportedly difficult to predict drug impurities with accuracy. • The model is suitable to support risk evaluation of potentially mutagenic compounds.« less
Development of estrogen receptor beta binding prediction model using large sets of chemicals.
Sakkiah, Sugunadevi; Selvaraj, Chandrabose; Gong, Ping; Zhang, Chaoyang; Tong, Weida; Hong, Huixiao
2017-11-03
We developed an ER β binding prediction model to facilitate identification of chemicals specifically bind ER β or ER α together with our previously developed ER α binding model. Decision Forest was used to train ER β binding prediction model based on a large set of compounds obtained from EADB. Model performance was estimated through 1000 iterations of 5-fold cross validations. Prediction confidence was analyzed using predictions from the cross validations. Informative chemical features for ER β binding were identified through analysis of the frequency data of chemical descriptors used in the models in the 5-fold cross validations. 1000 permutations were conducted to assess the chance correlation. The average accuracy of 5-fold cross validations was 93.14% with a standard deviation of 0.64%. Prediction confidence analysis indicated that the higher the prediction confidence the more accurate the predictions. Permutation testing results revealed that the prediction model is unlikely generated by chance. Eighteen informative descriptors were identified to be important to ER β binding prediction. Application of the prediction model to the data from ToxCast project yielded very high sensitivity of 90-92%. Our results demonstrated ER β binding of chemicals could be accurately predicted using the developed model. Coupling with our previously developed ER α prediction model, this model could be expected to facilitate drug development through identification of chemicals that specifically bind ER β or ER α .
Practical approach to subject-specific estimation of knee joint contact force
Knarr, Brian A.; Higginson, Jill S.
2015-01-01
Compressive forces experienced at the knee can significantly contribute to cartilage degeneration. Musculoskeletal models enable predictions of the internal forces experienced at the knee, but validation is often not possible, as experimental data detailing loading at the knee joint is limited. Recently available data reporting compressive knee force through direct measurement using instrumented total knee replacements offer a unique opportunity to evaluate the accuracy of models. Previous studies have highlighted the importance of subject-specificity in increasing the accuracy of model predictions; however, these techniques may be unrealistic outside of a research setting. Therefore, the goal of our work was to identify a practical approach for accurate prediction of tibiofemoral knee contact force (KCF). Four methods for prediction of knee contact force were compared: (1) standard static optimization, (2) uniform muscle coordination weighting, (3) subject-specific muscle coordination weighting and (4) subject-specific strength adjustments. Walking trials for three subjects with instrumented knee replacements were used to evaluate the accuracy of model predictions. Predictions utilizing subject-specific muscle coordination weighting yielded the best agreement with experimental data, however this method required in vivo data for weighting factor calibration. Including subject-specific strength adjustments improved models’ predictions compared to standard static optimization, with errors in peak KCF less than 0.5 body weight for all subjects. Overall, combining clinical assessments of muscle strength with standard tools available in the OpenSim software package, such as inverse kinematics and static optimization, appears to be a practical method for predicting joint contact force that can be implemented for many applications. PMID:25952546
Roth, Christian J; Becher, Tobias; Frerichs, Inéz; Weiler, Norbert; Wall, Wolfgang A
2017-04-01
Providing optimal personalized mechanical ventilation for patients with acute or chronic respiratory failure is still a challenge within a clinical setting for each case anew. In this article, we integrate electrical impedance tomography (EIT) monitoring into a powerful patient-specific computational lung model to create an approach for personalizing protective ventilatory treatment. The underlying computational lung model is based on a single computed tomography scan and able to predict global airflow quantities, as well as local tissue aeration and strains for any ventilation maneuver. For validation, a novel "virtual EIT" module is added to our computational lung model, allowing to simulate EIT images based on the patient's thorax geometry and the results of our numerically predicted tissue aeration. Clinically measured EIT images are not used to calibrate the computational model. Thus they provide an independent method to validate the computational predictions at high temporal resolution. The performance of this coupling approach has been tested in an example patient with acute respiratory distress syndrome. The method shows good agreement between computationally predicted and clinically measured airflow data and EIT images. These results imply that the proposed framework can be used for numerical prediction of patient-specific responses to certain therapeutic measures before applying them to an actual patient. In the long run, definition of patient-specific optimal ventilation protocols might be assisted by computational modeling. NEW & NOTEWORTHY In this work, we present a patient-specific computational lung model that is able to predict global and local ventilatory quantities for a given patient and any selected ventilation protocol. For the first time, such a predictive lung model is equipped with a virtual electrical impedance tomography module allowing real-time validation of the computed results with the patient measurements. First promising results obtained in an acute respiratory distress syndrome patient show the potential of this approach for personalized computationally guided optimization of mechanical ventilation in future. Copyright © 2017 the American Physiological Society.
Tomás, Inmaculada; Regueira-Iglesias, Alba; López, Maria; Arias-Bujanda, Nora; Novoa, Lourdes; Balsa-Castro, Carlos; Tomás, Maria
2017-01-01
Currently, there is little evidence available on the development of predictive models for the diagnosis or prognosis of chronic periodontitis based on the qPCR quantification of subgingival pathobionts. Our objectives were to: (1) analyze and internally validate pathobiont-based models that could be used to distinguish different periodontal conditions at site-specific level within the same patient with chronic periodontitis; (2) develop nomograms derived from predictive models. Subgingival plaque samples were obtained from control and periodontal sites (probing pocket depth and clinical attachment loss <4 mm and >4 mm, respectively) from 40 patients with moderate-severe generalized chronic periodontitis. The samples were analyzed by qPCR using TaqMan probes and specific primers to determine the concentrations of Actinobacillus actinomycetemcomitans (Aa) , Fusobacterium nucleatum (Fn) , Parvimonas micra (Pm) , Porphyromonas gingivalis (Pg) , Prevotella intermedia (Pi) , Tannerella forsythia (Tf) , and Treponema denticola (Td) . The pathobiont-based models were obtained using multivariate binary logistic regression. The best models were selected according to specified criteria. The discrimination was assessed using receiver operating characteristic curves and numerous classification measures were thus obtained. The nomograms were built based on the best predictive models. Eight bacterial cluster-based models showed an area under the curve (AUC) ≥0.760 and a sensitivity and specificity ≥75.0%. The PiTfFn cluster showed an AUC of 0.773 (sensitivity and specificity = 75.0%). When Pm and AaPm were incorporated in the TdPiTfFn cluster, we detected the two best predictive models with an AUC of 0.788 and 0.789, respectively (sensitivity and specificity = 77.5%). The TdPiTfAa cluster had an AUC of 0.785 (sensitivity and specificity = 75.0%). When Pm was incorporated in this cluster, a new predictive model appeared with better AUC and specificity values (0.787 and 80.0%, respectively). Distinct clusters formed by species with different etiopathogenic role (belonging to different Socransky's complexes) had a good predictive accuracy for distinguishing a site with periodontal destruction in a periodontal patient. The predictive clusters with the lowest number of bacteria were PiTfFn and TdPiTfAa , while TdPiTfAaFnPm had the highest number. In all the developed nomograms, high concentrations of these clusters were associated with an increased probability of having a periodontal site in a patient with chronic periodontitis.
Tomás, Inmaculada; Regueira-Iglesias, Alba; López, Maria; Arias-Bujanda, Nora; Novoa, Lourdes; Balsa-Castro, Carlos; Tomás, Maria
2017-01-01
Currently, there is little evidence available on the development of predictive models for the diagnosis or prognosis of chronic periodontitis based on the qPCR quantification of subgingival pathobionts. Our objectives were to: (1) analyze and internally validate pathobiont-based models that could be used to distinguish different periodontal conditions at site-specific level within the same patient with chronic periodontitis; (2) develop nomograms derived from predictive models. Subgingival plaque samples were obtained from control and periodontal sites (probing pocket depth and clinical attachment loss <4 mm and >4 mm, respectively) from 40 patients with moderate-severe generalized chronic periodontitis. The samples were analyzed by qPCR using TaqMan probes and specific primers to determine the concentrations of Actinobacillus actinomycetemcomitans (Aa), Fusobacterium nucleatum (Fn), Parvimonas micra (Pm), Porphyromonas gingivalis (Pg), Prevotella intermedia (Pi), Tannerella forsythia (Tf), and Treponema denticola (Td). The pathobiont-based models were obtained using multivariate binary logistic regression. The best models were selected according to specified criteria. The discrimination was assessed using receiver operating characteristic curves and numerous classification measures were thus obtained. The nomograms were built based on the best predictive models. Eight bacterial cluster-based models showed an area under the curve (AUC) ≥0.760 and a sensitivity and specificity ≥75.0%. The PiTfFn cluster showed an AUC of 0.773 (sensitivity and specificity = 75.0%). When Pm and AaPm were incorporated in the TdPiTfFn cluster, we detected the two best predictive models with an AUC of 0.788 and 0.789, respectively (sensitivity and specificity = 77.5%). The TdPiTfAa cluster had an AUC of 0.785 (sensitivity and specificity = 75.0%). When Pm was incorporated in this cluster, a new predictive model appeared with better AUC and specificity values (0.787 and 80.0%, respectively). Distinct clusters formed by species with different etiopathogenic role (belonging to different Socransky’s complexes) had a good predictive accuracy for distinguishing a site with periodontal destruction in a periodontal patient. The predictive clusters with the lowest number of bacteria were PiTfFn and TdPiTfAa, while TdPiTfAaFnPm had the highest number. In all the developed nomograms, high concentrations of these clusters were associated with an increased probability of having a periodontal site in a patient with chronic periodontitis. PMID:28848499
Category-Specific Neural Oscillations Predict Recall Organization During Memory Search
Morton, Neal W.; Kahana, Michael J.; Rosenberg, Emily A.; Baltuch, Gordon H.; Litt, Brian; Sharan, Ashwini D.; Sperling, Michael R.; Polyn, Sean M.
2013-01-01
Retrieved-context models of human memory propose that as material is studied, retrieval cues are constructed that allow one to target particular aspects of past experience. We examined the neural predictions of these models by using electrocorticographic/depth recordings and scalp electroencephalography (EEG) to characterize category-specific oscillatory activity, while participants studied and recalled items from distinct, neurally discriminable categories. During study, these category-specific patterns predict whether a studied item will be recalled. In the scalp EEG experiment, category-specific activity during study also predicts whether a given item will be recalled adjacent to other same-category items, consistent with the proposal that a category-specific retrieval cue is used to guide memory search. Retrieved-context models suggest that integrative neural circuitry is involved in the construction and maintenance of the retrieval cue. Consistent with this hypothesis, we observe category-specific patterns that rise in strength as multiple same-category items are studied sequentially, and find that individual differences in this category-specific neural integration during study predict the degree to which a participant will use category information to organize memory search. Finally, we track the deployment of this retrieval cue during memory search: Category-specific patterns are stronger when participants organize their responses according to the category of the studied material. PMID:22875859
McKenna, James E.; Carlson, Douglas M.; Payne-Wynne, Molly L.
2013-01-01
Aim: Rare aquatic species are a substantial component of biodiversity, and their conservation is a major objective of many management plans. However, they are difficult to assess, and their optimal habitats are often poorly known. Methods to effectively predict the likely locations of suitable rare aquatic species habitats are needed. We combine two modelling approaches to predict occurrence and general abundance of several rare fish species. Location: Allegheny watershed of western New York State (USA) Methods: Our method used two empirical neural network modelling approaches (species specific and assemblage based) to predict stream-by-stream occurrence and general abundance of rare darters, based on broad-scale habitat conditions. Species-specific models were developed for longhead darter (Percina macrocephala), spotted darter (Etheostoma maculatum) and variegate darter (Etheostoma variatum) in the Allegheny drainage. An additional model predicted the type of rare darter-containing assemblage expected in each stream reach. Predictions from both models were then combined inclusively and exclusively and compared with additional independent data. Results Example rare darter predictions demonstrate the method's effectiveness. Models performed well (R2 ≥ 0.79), identified where suitable darter habitat was most likely to occur, and predictions matched well to those of collection sites. Additional independent data showed that the most conservative (exclusive) model slightly underestimated the distributions of these rare darters or predictions were displaced by one stream reach, suggesting that new darter habitat types were detected in the later collections. Main conclusions Broad-scale habitat variables can be used to effectively identify rare species' habitats. Combining species-specific and assemblage-based models enhances our ability to make use of the sparse data on rare species and to identify habitat units most likely and least likely to support those species. This hybrid approach may assist managers with the prioritization of habitats to be examined or conserved for rare species.
Individualized Prediction of Reading Comprehension Ability Using Gray Matter Volume.
Cui, Zaixu; Su, Mengmeng; Li, Liangjie; Shu, Hua; Gong, Gaolang
2018-05-01
Reading comprehension is a crucial reading skill for learning and putatively contains 2 key components: reading decoding and linguistic comprehension. Current understanding of the neural mechanism underlying these reading comprehension components is lacking, and whether and how neuroanatomical features can be used to predict these 2 skills remain largely unexplored. In the present study, we analyzed a large sample from the Human Connectome Project (HCP) dataset and successfully built multivariate predictive models for these 2 skills using whole-brain gray matter volume features. The results showed that these models effectively captured individual differences in these 2 skills and were able to significantly predict these components of reading comprehension for unseen individuals. The strict cross-validation using the HCP cohort and another independent cohort of children demonstrated the model generalizability. The identified gray matter regions contributing to the skill prediction consisted of a wide range of regions covering the putative reading, cerebellum, and subcortical systems. Interestingly, there were gender differences in the predictive models, with the female-specific model overestimating the males' abilities. Moreover, the identified contributing gray matter regions for the female-specific and male-specific models exhibited considerable differences, supporting a gender-dependent neuroanatomical substrate for reading comprehension.
Muscle Synergies Facilitate Computational Prediction of Subject-Specific Walking Motions
Meyer, Andrew J.; Eskinazi, Ilan; Jackson, Jennifer N.; Rao, Anil V.; Patten, Carolynn; Fregly, Benjamin J.
2016-01-01
Researchers have explored a variety of neurorehabilitation approaches to restore normal walking function following a stroke. However, there is currently no objective means for prescribing and implementing treatments that are likely to maximize recovery of walking function for any particular patient. As a first step toward optimizing neurorehabilitation effectiveness, this study develops and evaluates a patient-specific synergy-controlled neuromusculoskeletal simulation framework that can predict walking motions for an individual post-stroke. The main question we addressed was whether driving a subject-specific neuromusculoskeletal model with muscle synergy controls (5 per leg) facilitates generation of accurate walking predictions compared to a model driven by muscle activation controls (35 per leg) or joint torque controls (5 per leg). To explore this question, we developed a subject-specific neuromusculoskeletal model of a single high-functioning hemiparetic subject using instrumented treadmill walking data collected at the subject’s self-selected speed of 0.5 m/s. The model included subject-specific representations of lower-body kinematic structure, foot–ground contact behavior, electromyography-driven muscle force generation, and neural control limitations and remaining capabilities. Using direct collocation optimal control and the subject-specific model, we evaluated the ability of the three control approaches to predict the subject’s walking kinematics and kinetics at two speeds (0.5 and 0.8 m/s) for which experimental data were available from the subject. We also evaluated whether synergy controls could predict a physically realistic gait period at one speed (1.1 m/s) for which no experimental data were available. All three control approaches predicted the subject’s walking kinematics and kinetics (including ground reaction forces) well for the model calibration speed of 0.5 m/s. However, only activation and synergy controls could predict the subject’s walking kinematics and kinetics well for the faster non-calibration speed of 0.8 m/s, with synergy controls predicting the new gait period the most accurately. When used to predict how the subject would walk at 1.1 m/s, synergy controls predicted a gait period close to that estimated from the linear relationship between gait speed and stride length. These findings suggest that our neuromusculoskeletal simulation framework may be able to bridge the gap between patient-specific muscle synergy information and resulting functional capabilities and limitations. PMID:27790612
Maciejewski, Matthew L.; Liu, Chuan-Fen; Fihn, Stephan D.
2009-01-01
OBJECTIVE—To compare the ability of generic comorbidity and risk adjustment measures, a diabetes-specific measure, and a self-reported functional status measure to explain variation in health care expenditures for individuals with diabetes. RESEARCH DESIGN AND METHODS—This study included a retrospective cohort of 3,092 diabetic veterans participating in a multisite trial. Two comorbidity measures, four risk adjusters, a functional status measure, a diabetes complication count, and baseline expenditures were constructed from administrative and survey data. Outpatient, inpatient, and total expenditure models were estimated using ordinary least squares regression. Adjusted R2 statistics and predictive ratios were compared across measures to assess overall explanatory power and explanatory power of low- and high-cost subgroups. RESULTS—Administrative data–based risk adjusters performed better than the comorbidity, functional status, and diabetes-specific measures in all expenditure models. The diagnostic cost groups (DCGs) measure had the greatest predictive power overall and for the low- and high-cost subgroups, while the diabetes-specific measure had the lowest predictive power. A model with DCGs and the diabetes-specific measure modestly improved predictive power. CONCLUSIONS—Existing generic measures can be useful for diabetes-specific research and policy applications, but more predictive diabetes-specific measures are needed. PMID:18945927
Yamamoto, Yumi; Välitalo, Pyry A.; Huntjens, Dymphy R.; Proost, Johannes H.; Vermeulen, An; Krauwinkel, Walter; Beukers, Margot W.; van den Berg, Dirk‐Jan; Hartman, Robin; Wong, Yin Cheong; Danhof, Meindert; van Hasselt, John G. C.
2017-01-01
Drug development targeting the central nervous system (CNS) is challenging due to poor predictability of drug concentrations in various CNS compartments. We developed a generic physiologically based pharmacokinetic (PBPK) model for prediction of drug concentrations in physiologically relevant CNS compartments. System‐specific and drug‐specific model parameters were derived from literature and in silico predictions. The model was validated using detailed concentration‐time profiles from 10 drugs in rat plasma, brain extracellular fluid, 2 cerebrospinal fluid sites, and total brain tissue. These drugs, all small molecules, were selected to cover a wide range of physicochemical properties. The concentration‐time profiles for these drugs were adequately predicted across the CNS compartments (symmetric mean absolute percentage error for the model prediction was <91%). In conclusion, the developed PBPK model can be used to predict temporal concentration profiles of drugs in multiple relevant CNS compartments, which we consider valuable information for efficient CNS drug development. PMID:28891201
A Subject-Specific Kinematic Model to Predict Human Motion in Exoskeleton-Assisted Gait.
Torricelli, Diego; Cortés, Camilo; Lete, Nerea; Bertelsen, Álvaro; Gonzalez-Vargas, Jose E; Del-Ama, Antonio J; Dimbwadyo, Iris; Moreno, Juan C; Florez, Julian; Pons, Jose L
2018-01-01
The relative motion between human and exoskeleton is a crucial factor that has remarkable consequences on the efficiency, reliability and safety of human-robot interaction. Unfortunately, its quantitative assessment has been largely overlooked in the literature. Here, we present a methodology that allows predicting the motion of the human joints from the knowledge of the angular motion of the exoskeleton frame. Our method combines a subject-specific skeletal model with a kinematic model of a lower limb exoskeleton (H2, Technaid), imposing specific kinematic constraints between them. To calibrate the model and validate its ability to predict the relative motion in a subject-specific way, we performed experiments on seven healthy subjects during treadmill walking tasks. We demonstrate a prediction accuracy lower than 3.5° globally, and around 1.5° at the hip level, which represent an improvement up to 66% compared to the traditional approach assuming no relative motion between the user and the exoskeleton.
A Subject-Specific Kinematic Model to Predict Human Motion in Exoskeleton-Assisted Gait
Torricelli, Diego; Cortés, Camilo; Lete, Nerea; Bertelsen, Álvaro; Gonzalez-Vargas, Jose E.; del-Ama, Antonio J.; Dimbwadyo, Iris; Moreno, Juan C.; Florez, Julian; Pons, Jose L.
2018-01-01
The relative motion between human and exoskeleton is a crucial factor that has remarkable consequences on the efficiency, reliability and safety of human-robot interaction. Unfortunately, its quantitative assessment has been largely overlooked in the literature. Here, we present a methodology that allows predicting the motion of the human joints from the knowledge of the angular motion of the exoskeleton frame. Our method combines a subject-specific skeletal model with a kinematic model of a lower limb exoskeleton (H2, Technaid), imposing specific kinematic constraints between them. To calibrate the model and validate its ability to predict the relative motion in a subject-specific way, we performed experiments on seven healthy subjects during treadmill walking tasks. We demonstrate a prediction accuracy lower than 3.5° globally, and around 1.5° at the hip level, which represent an improvement up to 66% compared to the traditional approach assuming no relative motion between the user and the exoskeleton. PMID:29755336
Development of machine learning models for diagnosis of glaucoma.
Kim, Seong Jae; Cho, Kyong Jin; Oh, Sejong
2017-01-01
The study aimed to develop machine learning models that have strong prediction power and interpretability for diagnosis of glaucoma based on retinal nerve fiber layer (RNFL) thickness and visual field (VF). We collected various candidate features from the examination of retinal nerve fiber layer (RNFL) thickness and visual field (VF). We also developed synthesized features from original features. We then selected the best features proper for classification (diagnosis) through feature evaluation. We used 100 cases of data as a test dataset and 399 cases of data as a training and validation dataset. To develop the glaucoma prediction model, we considered four machine learning algorithms: C5.0, random forest (RF), support vector machine (SVM), and k-nearest neighbor (KNN). We repeatedly composed a learning model using the training dataset and evaluated it by using the validation dataset. Finally, we got the best learning model that produces the highest validation accuracy. We analyzed quality of the models using several measures. The random forest model shows best performance and C5.0, SVM, and KNN models show similar accuracy. In the random forest model, the classification accuracy is 0.98, sensitivity is 0.983, specificity is 0.975, and AUC is 0.979. The developed prediction models show high accuracy, sensitivity, specificity, and AUC in classifying among glaucoma and healthy eyes. It will be used for predicting glaucoma against unknown examination records. Clinicians may reference the prediction results and be able to make better decisions. We may combine multiple learning models to increase prediction accuracy. The C5.0 model includes decision rules for prediction. It can be used to explain the reasons for specific predictions.
Improving stability of prediction models based on correlated omics data by using network approaches.
Tissier, Renaud; Houwing-Duistermaat, Jeanine; Rodríguez-Girondo, Mar
2018-01-01
Building prediction models based on complex omics datasets such as transcriptomics, proteomics, metabolomics remains a challenge in bioinformatics and biostatistics. Regularized regression techniques are typically used to deal with the high dimensionality of these datasets. However, due to the presence of correlation in the datasets, it is difficult to select the best model and application of these methods yields unstable results. We propose a novel strategy for model selection where the obtained models also perform well in terms of overall predictability. Several three step approaches are considered, where the steps are 1) network construction, 2) clustering to empirically derive modules or pathways, and 3) building a prediction model incorporating the information on the modules. For the first step, we use weighted correlation networks and Gaussian graphical modelling. Identification of groups of features is performed by hierarchical clustering. The grouping information is included in the prediction model by using group-based variable selection or group-specific penalization. We compare the performance of our new approaches with standard regularized regression via simulations. Based on these results we provide recommendations for selecting a strategy for building a prediction model given the specific goal of the analysis and the sizes of the datasets. Finally we illustrate the advantages of our approach by application of the methodology to two problems, namely prediction of body mass index in the DIetary, Lifestyle, and Genetic determinants of Obesity and Metabolic syndrome study (DILGOM) and prediction of response of each breast cancer cell line to treatment with specific drugs using a breast cancer cell lines pharmacogenomics dataset.
Glass Transition Temperature- and Specific Volume- Composition Models for Tellurite Glasses
DOE Office of Scientific and Technical Information (OSTI.GOV)
Riley, Brian J.; Vienna, John D.
This report provides models for predicting composition-properties for tellurite glasses, namely specific gravity and glass transition temperature. Included are the partial specific coefficients for each model, the component validity ranges, and model fit parameters.
Youth Sport Readiness: A Predictive Model for Success.
ERIC Educational Resources Information Center
Aicinena, Steven
1992-01-01
A model for predicting organized youth sport participation readiness has four predictive components: sport-related fundamental motor skill development; sport-specific knowledge; motivation; and socialization. Physical maturation is also important. The model emphasizes the importance of preparing children for successful participation through…
Magheli, Ahmed; Hinz, Stefan; Hege, Claudia; Stephan, Carsten; Jung, Klaus; Miller, Kurt; Lein, Michael
2010-01-01
We investigated the value of pretreatment prostate specific antigen density to predict Gleason score upgrading in light of significant changes in grading routine in the last 2 decades. Of 1,061 consecutive men who underwent radical prostatectomy between 1999 and 2004, 843 were eligible for study. Prostate specific antigen density was calculated and a cutoff for highest accuracy to predict Gleason upgrading was determined using ROC curve analysis. The predictive accuracy of prostate specific antigen and prostate specific antigen density to predict Gleason upgrading was evaluated using ROC curve analysis based on predicted probabilities from logistic regression models. Prostate specific antigen and prostate specific antigen density predicted Gleason upgrading on univariate analysis (as continuous variables OR 1.07 and 7.21, each p <0.001) and on multivariate analysis (as continuous variables with prostate specific antigen density adjusted for prostate specific antigen OR 1.07, p <0.001 and OR 4.89, p = 0.037, respectively). When prostate specific antigen density was added to the model including prostate specific antigen and other Gleason upgrading predictors, prostate specific antigen lost its predictive value (OR 1.02, p = 0.423), while prostate specific antigen density remained an independent predictor (OR 4.89, p = 0.037). Prostate specific antigen density was more accurate than prostate specific antigen to predict Gleason upgrading (AUC 0.61 vs 0.57, p = 0.030). Prostate specific antigen density is a significant independent predictor of Gleason upgrading even when accounting for prostate specific antigen. This could be especially important in patients with low risk prostate cancer who seek less invasive therapy such as active surveillance since potentially life threatening disease may be underestimated. Further studies are warranted to help evaluate the role of prostate specific antigen density in Gleason upgrading and its significance for biochemical outcome.
An age-specific biokinetic model for iodine
DOE Office of Scientific and Technical Information (OSTI.GOV)
Leggett, Richard Wayne
This study reviews age-specific biokinetic data for iodine in humans and extends to pre-adult ages the baseline parameter values of the author’s previously published model for systemic iodine in adult humans. Compared with the ICRP’s current age-specific model for iodine introduced in Publication 56 (1989), the present model provides a more detailed description of the behavior of iodine in the human body; predicts greater cumulative (integrated) activity in the thyroid for short-lived isotopes of iodine; predicts similar cumulative activity in the thyroid for isotopes with half-time greater than a few hours; and, for most iodine isotopes, predicts much greater cumulativemore » activity in salivary glands, stomach wall, liver, and kidneys.« less
An age-specific biokinetic model for iodine
Leggett, Richard Wayne
2017-10-26
This study reviews age-specific biokinetic data for iodine in humans and extends to pre-adult ages the baseline parameter values of the author’s previously published model for systemic iodine in adult humans. Compared with the ICRP’s current age-specific model for iodine introduced in Publication 56 (1989), the present model provides a more detailed description of the behavior of iodine in the human body; predicts greater cumulative (integrated) activity in the thyroid for short-lived isotopes of iodine; predicts similar cumulative activity in the thyroid for isotopes with half-time greater than a few hours; and, for most iodine isotopes, predicts much greater cumulativemore » activity in salivary glands, stomach wall, liver, and kidneys.« less
Predicting coronary artery disease using different artificial neural network models.
Colak, M Cengiz; Colak, Cemil; Kocatürk, Hasan; Sağiroğlu, Seref; Barutçu, Irfan
2008-08-01
Eight different learning algorithms used for creating artificial neural network (ANN) models and the different ANN models in the prediction of coronary artery disease (CAD) are introduced. This work was carried out as a retrospective case-control study. Overall, 124 consecutive patients who had been diagnosed with CAD by coronary angiography (at least 1 coronary stenosis > 50% in major epicardial arteries) were enrolled in the work. Angiographically, the 113 people (group 2) with normal coronary arteries were taken as control subjects. Multi-layered perceptrons ANN architecture were applied. The ANN models trained with different learning algorithms were performed in 237 records, divided into training (n=171) and testing (n=66) data sets. The performance of prediction was evaluated by sensitivity, specificity and accuracy values based on standard definitions. The results have demonstrated that ANN models trained with eight different learning algorithms are promising because of high (greater than 71%) sensitivity, specificity and accuracy values in the prediction of CAD. Accuracy, sensitivity and specificity values varied between 83.63%-100%, 86.46%-100% and 74.67%-100% for training, respectively. For testing, the values were more than 71% for sensitivity, 76% for specificity and 81% for accuracy. It may be proposed that the use of different learning algorithms other than backpropagation and larger sample sizes can improve the performance of prediction. The proposed ANN models trained with these learning algorithms could be used a promising approach for predicting CAD without the need for invasive diagnostic methods and could help in the prognostic clinical decision.
Martini, Alberto; Gupta, Akriti; Lewis, Sara C; Cumarasamy, Shivaram; Haines, Kenneth G; Briganti, Alberto; Montorsi, Francesco; Tewari, Ashutosh K
2018-04-19
To develop a nomogram for predicting side-specific extracapsular extension (ECE) for planning nerve-sparing radical prostatectomy. We retrospectively analysed data from 561 patients who underwent robot-assisted radical prostatectomy between February 2014 and October 2015. To develop a side-specific predictive model, we considered the prostatic lobes separately. Four variables were included: prostate-specific antigen; highest ipsilateral biopsy Gleason grade; highest ipsilateral percentage core involvement; and ECE on multiparametric magnetic resonance imaging (mpMRI). A multivariable logistic regression analysis was fitted to predict side-specific ECE. A nomogram was built based on the coefficients of the logit function. Internal validation was performed using 'leave-one-out' cross-validation. Calibration was graphically investigated. The decision curve analysis was used to evaluate the net clinical benefit. The study population consisted of 829 side-specific cases, after excluding negative biopsy observations (n = 293). ECE was reported on mpMRI and final pathology in 115 (14%) and 142 (17.1%) cases, respectively. Among these, mpMRI was able to predict ECE correctly in 57 (40.1%) cases. All variables in the model except highest percentage core involvement were predictors of ECE (all P ≤ 0.006). All variables were considered for inclusion in the nomogram. After internal validation, the area under the curve was 82.11%. The model demonstrated excellent calibration and improved clinical risk prediction, especially when compared with relying on mpMRI prediction of ECE alone. When retrospectively applying the nomogram-derived probability, using a 20% threshold for performing nerve-sparing, nine out of 14 positive surgical margins (PSMs) at the site of ECE resulted above the threshold. We developed an easy-to-use model for the prediction of side-specific ECE, and hope it serves as a tool for planning nerve-sparing radical prostatectomy and in the reduction of PSM in future series. © 2018 The Authors BJU International © 2018 BJU International Published by John Wiley & Sons Ltd.
Luboz, Vincent; Chabanas, Matthieu; Swider, Pascal; Payan, Yohan
2005-08-01
This paper addresses an important issue raised for the clinical relevance of Computer-Assisted Surgical applications, namely the methodology used to automatically build patient-specific finite element (FE) models of anatomical structures. From this perspective, a method is proposed, based on a technique called the mesh-matching method, followed by a process that corrects mesh irregularities. The mesh-matching algorithm generates patient-specific volume meshes from an existing generic model. The mesh regularization process is based on the Jacobian matrix transform related to the FE reference element and the current element. This method for generating patient-specific FE models is first applied to computer-assisted maxillofacial surgery, and more precisely, to the FE elastic modelling of patient facial soft tissues. For each patient, the planned bone osteotomies (mandible, maxilla, chin) are used as boundary conditions to deform the FE face model, in order to predict the aesthetic outcome of the surgery. Seven FE patient-specific models were successfully generated by our method. For one patient, the prediction of the FE model is qualitatively compared with the patient's post-operative appearance, measured from a computer tomography scan. Then, our methodology is applied to computer-assisted orbital surgery. It is, therefore, evaluated for the generation of 11 patient-specific FE poroelastic models of the orbital soft tissues. These models are used to predict the consequences of the surgical decompression of the orbit. More precisely, an average law is extrapolated from the simulations carried out for each patient model. This law links the size of the osteotomy (i.e. the surgical gesture) and the backward displacement of the eyeball (the consequence of the surgical gesture).
New Space Weather Systems Under Development and Their Contribution to Space Weather Management
NASA Astrophysics Data System (ADS)
Tobiska, W.; Bouwer, D.; Schunk, R.; Garrett, H.; Mertens, C.; Bowman, B.
2008-12-01
There have been notable successes during the past decade in the development of operational space environment systems. Examples include the Magnetospheric Specification Model (MSM) of the Earth's magnetosphere, 2000; SOLAR2000 (S2K) solar spectral irradiances, 2001; High Accuracy Satellite Drag Model (HASDM) neutral atmosphere densities, 2004; Global Assimilation of Ionospheric Measurements (GAIM) ionosphere specification, 2006; Hakamada-Akasofu-Fry (HAF) solar wind parameters, 2007; Communication Alert and Prediction System (CAPS) ionosphere, high frequency radio, and scintillation S4 index prediction, 2008; and GEO Alert and Prediction System (GAPS) geosynchronous environment satellite charging specification and forecast, 2008. Operational systems that are in active operational implementation include the Jacchia-Bowman 2006/2008 (JB2006/2008) neutral atmosphere, 2009, and the Nowcast of Atmospheric Ionizing Radiation for Aviation Safety (NAIRAS) aviation radiation model using the Radiation Alert and Prediction System (RAPS), 2010. U.S. national agency and commercial assets will soon reach a state where specification and prediction will become ubiquitous and where coordinated management of the space environment and space weather will become a necessity. We describe the status of the CAPS, GAPS, RAPS, and JB2008 operational development. We additionally discuss the conditions that are laying the groundwork for space weather management and estimate the unfilled needs as we move beyond specification and prediction efforts.
Chande, Ruchi D; Wayne, Jennifer S
2017-09-01
Computational models of diarthrodial joints serve to inform the biomechanical function of these structures, and as such, must be supplied appropriate inputs for performance that is representative of actual joint function. Inputs for these models are sourced from both imaging modalities as well as literature. The latter is often the source of mechanical properties for soft tissues, like ligament stiffnesses; however, such data are not always available for all the soft tissues nor is it known for patient-specific work. In the current research, a method to improve the ligament stiffness definition for a computational foot/ankle model was sought with the greater goal of improving the predictive ability of the computational model. Specifically, the stiffness values were optimized using artificial neural networks (ANNs); both feedforward and radial basis function networks (RBFNs) were considered. Optimal networks of each type were determined and subsequently used to predict stiffnesses for the foot/ankle model. Ultimately, the predicted stiffnesses were considered reasonable and resulted in enhanced performance of the computational model, suggesting that artificial neural networks can be used to optimize stiffness inputs.
Specimen-specific modeling of hip fracture pattern and repair.
Ali, Azhar A; Cristofolini, Luca; Schileo, Enrico; Hu, Haixiang; Taddei, Fulvia; Kim, Raymond H; Rullkoetter, Paul J; Laz, Peter J
2014-01-22
Hip fracture remains a major health problem for the elderly. Clinical studies have assessed fracture risk based on bone quality in the aging population and cadaveric testing has quantified bone strength and fracture loads. Prior modeling has primarily focused on quantifying the strain distribution in bone as an indicator of fracture risk. Recent advances in the extended finite element method (XFEM) enable prediction of the initiation and propagation of cracks without requiring a priori knowledge of the crack path. Accordingly, the objectives of this study were to predict femoral fracture in specimen-specific models using the XFEM approach, to perform one-to-one comparisons of predicted and in vitro fracture patterns, and to develop a framework to assess the mechanics and load transfer in the fractured femur when it is repaired with an osteosynthesis implant. Five specimen-specific femur models were developed from in vitro experiments under a simulated stance loading condition. Predicted fracture patterns closely matched the in vitro patterns; however, predictions of fracture load differed by approximately 50% due to sensitivity to local material properties. Specimen-specific intertrochanteric fractures were induced by subjecting the femur models to a sideways fall and repaired with a contemporary implant. Under a post-surgical stance loading, model-predicted load sharing between the implant and bone across the fracture surface varied from 59%:41% to 89%:11%, underscoring the importance of considering anatomic and fracture variability in the evaluation of implants. XFEM modeling shows potential as a macro-level analysis enabling fracture investigations of clinical cohorts, including at-risk groups, and the design of robust implants. © 2013 Published by Elsevier Ltd.
Automatically updating predictive modeling workflows support decision-making in drug design.
Muegge, Ingo; Bentzien, Jörg; Mukherjee, Prasenjit; Hughes, Robert O
2016-09-01
Using predictive models for early decision-making in drug discovery has become standard practice. We suggest that model building needs to be automated with minimum input and low technical maintenance requirements. Models perform best when tailored to answering specific compound optimization related questions. If qualitative answers are required, 2-bin classification models are preferred. Integrating predictive modeling results with structural information stimulates better decision making. For in silico models supporting rapid structure-activity relationship cycles the performance deteriorates within weeks. Frequent automated updates of predictive models ensure best predictions. Consensus between multiple modeling approaches increases the prediction confidence. Combining qualified and nonqualified data optimally uses all available information. Dose predictions provide a holistic alternative to multiple individual property predictions for reaching complex decisions.
Wu, Zheyang; Yang, Chun; Tang, Dalin
2011-06-01
It has been hypothesized that mechanical risk factors may be used to predict future atherosclerotic plaque rupture. Truly predictive methods for plaque rupture and methods to identify the best predictor(s) from all the candidates are lacking in the literature. A novel combination of computational and statistical models based on serial magnetic resonance imaging (MRI) was introduced to quantify sensitivity and specificity of mechanical predictors to identify the best candidate for plaque rupture site prediction. Serial in vivo MRI data of carotid plaque from one patient was acquired with follow-up scan showing ulceration. 3D computational fluid-structure interaction (FSI) models using both baseline and follow-up data were constructed and plaque wall stress (PWS) and strain (PWSn) and flow maximum shear stress (FSS) were extracted from all 600 matched nodal points (100 points per matched slice, baseline matching follow-up) on the lumen surface for analysis. Each of the 600 points was marked "ulcer" or "nonulcer" using follow-up scan. Predictive statistical models for each of the seven combinations of PWS, PWSn, and FSS were trained using the follow-up data and applied to the baseline data to assess their sensitivity and specificity using the 600 data points for ulcer predictions. Sensitivity of prediction is defined as the proportion of the true positive outcomes that are predicted to be positive. Specificity of prediction is defined as the proportion of the true negative outcomes that are correctly predicted to be negative. Using probability 0.3 as a threshold to infer ulcer occurrence at the prediction stage, the combination of PWS and PWSn provided the best predictive accuracy with (sensitivity, specificity) = (0.97, 0.958). Sensitivity and specificity given by PWS, PWSn, and FSS individually were (0.788, 0.968), (0.515, 0.968), and (0.758, 0.928), respectively. The proposed computational-statistical process provides a novel method and a framework to assess the sensitivity and specificity of various risk indicators and offers the potential to identify the optimized predictor for plaque rupture using serial MRI with follow-up scan showing ulceration as the gold standard for method validation. While serial MRI data with actual rupture are hard to acquire, this single-case study suggests that combination of multiple predictors may provide potential improvement to existing plaque assessment schemes. With large-scale patient studies, this predictive modeling process may provide more solid ground for rupture predictor selection strategies and methods for image-based plaque vulnerability assessment.
Heil, Lieke; Kwisthout, Johan; van Pelt, Stan; van Rooij, Iris; Bekkering, Harold
2018-01-01
Evidence is accumulating that our brains process incoming information using top-down predictions. If lower level representations are correctly predicted by higher level representations, this enhances processing. However, if they are incorrectly predicted, additional processing is required at higher levels to "explain away" prediction errors. Here, we explored the potential nature of the models generating such predictions. More specifically, we investigated whether a predictive processing model with a hierarchical structure and causal relations between its levels is able to account for the processing of agent-caused events. In Experiment 1, participants watched animated movies of "experienced" and "novice" bowlers. The results are in line with the idea that prediction errors at a lower level of the hierarchy (i.e., the outcome of how many pins fell down) slow down reporting of information at a higher level (i.e., which agent was throwing the ball). Experiments 2 and 3 suggest that this effect is specific to situations in which the predictor is causally related to the outcome. Overall, the study supports the idea that a hierarchical predictive processing model can account for the processing of observed action outcomes and that the predictions involved are specific to cases where action outcomes can be predicted based on causal knowledge.
A predictive model for biomimetic plate type broadband frequency sensor
NASA Astrophysics Data System (ADS)
Ahmed, Riaz U.; Banerjee, Sourav
2016-04-01
In this work, predictive model for a bio-inspired broadband frequency sensor is developed. Broadband frequency sensing is essential in many domains of science and technology. One great example of such sensor is human cochlea, where it senses a frequency band of 20 Hz to 20 KHz. Developing broadband sensor adopting the physics of human cochlea has found tremendous interest in recent years. Although few experimental studies have been reported, a true predictive model to design such sensors is missing. A predictive model is utmost necessary for accurate design of selective broadband sensors that are capable of sensing very selective band of frequencies. Hence, in this study, we proposed a novel predictive model for the cochlea-inspired broadband sensor, aiming to select the frequency band and model parameters predictively. Tapered plate geometry is considered mimicking the real shape of the basilar membrane in the human cochlea. The predictive model is intended to develop flexible enough that can be employed in a wide variety of scientific domains. To do that, the predictive model is developed in such a way that, it can not only handle homogeneous but also any functionally graded model parameters. Additionally, the predictive model is capable of managing various types of boundary conditions. It has been found that, using the homogeneous model parameters, it is possible to sense a specific frequency band from a specific portion (B) of the model length (L). It is also possible to alter the attributes of `B' using functionally graded model parameters, which confirms the predictive frequency selection ability of the developed model.
NASA Astrophysics Data System (ADS)
Ahdika, Atina; Lusiyana, Novyan
2017-02-01
World Health Organization (WHO) noted Indonesia as the country with the highest dengue (DHF) cases in Southeast Asia. There are no vaccine and specific treatment for DHF. One of the efforts which can be done by both government and resident is doing a prevention action. In statistics, there are some methods to predict the number of DHF cases to be used as the reference to prevent the DHF cases. In this paper, a discrete time series model, INAR(1)-Poisson model in specific, and Markov prediction model are used to predict the number of DHF patients in West Java Indonesia. The result shows that MPM is the best model since it has the smallest value of MAE (mean absolute error) and MAPE (mean absolute percentage error).
Miles, Brad; Kolos, Elizabeth; Walter, William L; Appleyard, Richard; Shi, Angela; Li, Qing; Ruys, Andrew J
2015-06-01
Subject-specific finite element (FE) modeling methodology could predict peri-prosthetic femoral fracture (PFF) for cementless hip arthoplasty in the early postoperative period. This study develops methodology for subject-specific finite element modeling by using the element deactivation technique to simulate bone failure and validate with experimental testing, thereby predicting peri-prosthetic femoral fracture in the early postoperative period. Material assignments for biphasic and triphasic models were undertaken. Failure modeling with the element deactivation feature available in ABAQUS 6.9 was used to simulate a crack initiation and propagation in the bony tissue based upon a threshold of fracture strain. The crack mode for the biphasic models was very similar to the experimental testing crack mode, with a similar shape and path of the crack. The fracture load is sensitive to the friction coefficient at the implant-bony interface. The development of a novel technique to simulate bone failure by element deactivation of subject-specific finite element models could aid prediction of fracture load in addition to fracture risk characterization for PFF. Copyright © 2015 IPEM. Published by Elsevier Ltd. All rights reserved.
Paul, Keryn I; Roxburgh, Stephen H; Chave, Jerome; England, Jacqueline R; Zerihun, Ayalsew; Specht, Alison; Lewis, Tom; Bennett, Lauren T; Baker, Thomas G; Adams, Mark A; Huxtable, Dan; Montagu, Kelvin D; Falster, Daniel S; Feller, Mike; Sochacki, Stan; Ritson, Peter; Bastin, Gary; Bartle, John; Wildy, Dan; Hobbs, Trevor; Larmour, John; Waterworth, Rob; Stewart, Hugh T L; Jonson, Justin; Forrester, David I; Applegate, Grahame; Mendham, Daniel; Bradford, Matt; O'Grady, Anthony; Green, Daryl; Sudmeyer, Rob; Rance, Stan J; Turner, John; Barton, Craig; Wenk, Elizabeth H; Grove, Tim; Attiwill, Peter M; Pinkard, Elizabeth; Butler, Don; Brooksbank, Kim; Spencer, Beren; Snowdon, Peter; O'Brien, Nick; Battaglia, Michael; Cameron, David M; Hamilton, Steve; McAuthur, Geoff; Sinclair, Jenny
2016-06-01
Accurate ground-based estimation of the carbon stored in terrestrial ecosystems is critical to quantifying the global carbon budget. Allometric models provide cost-effective methods for biomass prediction. But do such models vary with ecoregion or plant functional type? We compiled 15 054 measurements of individual tree or shrub biomass from across Australia to examine the generality of allometric models for above-ground biomass prediction. This provided a robust case study because Australia includes ecoregions ranging from arid shrublands to tropical rainforests, and has a rich history of biomass research, particularly in planted forests. Regardless of ecoregion, for five broad categories of plant functional type (shrubs; multistemmed trees; trees of the genus Eucalyptus and closely related genera; other trees of high wood density; and other trees of low wood density), relationships between biomass and stem diameter were generic. Simple power-law models explained 84-95% of the variation in biomass, with little improvement in model performance when other plant variables (height, bole wood density), or site characteristics (climate, age, management) were included. Predictions of stand-based biomass from allometric models of varying levels of generalization (species-specific, plant functional type) were validated using whole-plot harvest data from 17 contrasting stands (range: 9-356 Mg ha(-1) ). Losses in efficiency of prediction were <1% if generalized models were used in place of species-specific models. Furthermore, application of generalized multispecies models did not introduce significant bias in biomass prediction in 92% of the 53 species tested. Further, overall efficiency of stand-level biomass prediction was 99%, with a mean absolute prediction error of only 13%. Hence, for cost-effective prediction of biomass across a wide range of stands, we recommend use of generic allometric models based on plant functional types. Development of new species-specific models is only warranted when gains in accuracy of stand-based predictions are relatively high (e.g. high-value monocultures). © 2015 John Wiley & Sons Ltd.
Bosboom, E. Marielle H.; Kroon, Wilco; van der Linden, Wim P. M.; Planken, R. Nils; van de Vosse, Frans N.; Tordoir, Jan H. M.
2012-01-01
Introduction Inadequate flow enhancement on the one hand, and excessive flow enhancement on the other hand, remain frequent complications of arteriovenous fistula (AVF) creation, and hamper hemodialysis therapy in patients with end-stage renal disease. In an effort to reduce these, a patient-specific computational model, capable of predicting postoperative flow, has been developed. The purpose of this study was to determine the accuracy of the patient-specific model and to investigate its feasibility to support decision-making in AVF surgery. Methods Patient-specific pulse wave propagation models were created for 25 patients awaiting AVF creation. Model input parameters were obtained from clinical measurements and literature. For every patient, a radiocephalic AVF, a brachiocephalic AVF, and a brachiobasilic AVF configuration were simulated and analyzed for their postoperative flow. The most distal configuration with a predicted flow between 400 and 1500 ml/min was considered the preferred location for AVF surgery. The suggestion of the model was compared to the choice of an experienced vascular surgeon. Furthermore, predicted flows were compared to measured postoperative flows. Results Taken into account the confidence interval (25th and 75th percentile interval), overlap between predicted and measured postoperative flows was observed in 70% of the patients. Differentiation between upper and lower arm configuration was similar in 76% of the patients, whereas discrimination between two upper arm AVF configurations was more difficult. In 3 patients the surgeon created an upper arm AVF, while model based predictions allowed for lower arm AVF creation, thereby preserving proximal vessels. In one patient early thrombosis in a radiocephalic AVF was observed which might have been indicated by the low predicted postoperative flow. Conclusions Postoperative flow can be predicted relatively accurately for multiple AVF configurations by using computational modeling. This model may therefore be considered a valuable additional tool in the preoperative work-up of patients awaiting AVF creation. PMID:22496816
SVM-Based System for Prediction of Epileptic Seizures from iEEG Signal
Cherkassky, Vladimir; Lee, Jieun; Veber, Brandon; Patterson, Edward E.; Brinkmann, Benjamin H.; Worrell, Gregory A.
2017-01-01
Objective This paper describes a data-analytic modeling approach for prediction of epileptic seizures from intracranial electroencephalogram (iEEG) recording of brain activity. Even though it is widely accepted that statistical characteristics of iEEG signal change prior to seizures, robust seizure prediction remains a challenging problem due to subject-specific nature of data-analytic modeling. Methods Our work emphasizes understanding of clinical considerations important for iEEG-based seizure prediction, and proper translation of these clinical considerations into data-analytic modeling assumptions. Several design choices during pre-processing and post-processing are considered and investigated for their effect on seizure prediction accuracy. Results Our empirical results show that the proposed SVM-based seizure prediction system can achieve robust prediction of preictal and interictal iEEG segments from dogs with epilepsy. The sensitivity is about 90–100%, and the false-positive rate is about 0–0.3 times per day. The results also suggest good prediction is subject-specific (dog or human), in agreement with earlier studies. Conclusion Good prediction performance is possible only if the training data contain sufficiently many seizure episodes, i.e., at least 5–7 seizures. Significance The proposed system uses subject-specific modeling and unbalanced training data. This system also utilizes three different time scales during training and testing stages. PMID:27362758
NASA Astrophysics Data System (ADS)
Hughes, J. D.; White, J.; Doherty, J.
2011-12-01
Linear prediction uncertainty analysis in a Bayesian framework was applied to guide the conditioning of an integrated surface water/groundwater model that will be used to predict the effects of groundwater withdrawals on surface-water and groundwater flows. Linear prediction uncertainty analysis is an effective approach for identifying (1) raw and processed data most effective for model conditioning prior to inversion, (2) specific observations and periods of time critically sensitive to specific predictions, and (3) additional observation data that would reduce model uncertainty relative to specific predictions. We present results for a two-dimensional groundwater model of a 2,186 km2 area of the Biscayne aquifer in south Florida implicitly coupled to a surface-water routing model of the actively managed canal system. The model domain includes 5 municipal well fields withdrawing more than 1 Mm3/day and 17 operable surface-water control structures that control freshwater releases from the Everglades and freshwater discharges to Biscayne Bay. More than 10 years of daily observation data from 35 groundwater wells and 24 surface water gages are available to condition model parameters. A dense parameterization was used to fully characterize the contribution of the inversion null space to predictive uncertainty and included bias-correction parameters. This approach allows better resolution of the boundary between the inversion null space and solution space. Bias-correction parameters (e.g., rainfall, potential evapotranspiration, and structure flow multipliers) absorb information that is present in structural noise that may otherwise contaminate the estimation of more physically-based model parameters. This allows greater precision in predictions that are entirely solution-space dependent, and reduces the propensity for bias in predictions that are not. Results show that application of this analysis is an effective means of identifying those surface-water and groundwater data, both raw and processed, that minimize predictive uncertainty, while simultaneously identifying the maximum solution-space dimensionality of the inverse problem supported by the data.
Refining metabolic models and accounting for regulatory effects.
Kim, Joonhoon; Reed, Jennifer L
2014-10-01
Advances in genome-scale metabolic modeling allow us to investigate and engineer metabolism at a systems level. Metabolic network reconstructions have been made for many organisms and computational approaches have been developed to convert these reconstructions into predictive models. However, due to incomplete knowledge these reconstructions often have missing or extraneous components and interactions, which can be identified by reconciling model predictions with experimental data. Recent studies have provided methods to further improve metabolic model predictions by incorporating transcriptional regulatory interactions and high-throughput omics data to yield context-specific metabolic models. Here we discuss recent approaches for resolving model-data discrepancies and building context-specific metabolic models. Once developed highly accurate metabolic models can be used in a variety of biotechnology applications. Copyright © 2014 Elsevier Ltd. All rights reserved.
Subject-specific knee joint geometry improves predictions of medial tibiofemoral contact forces.
Gerus, Pauline; Sartori, Massimo; Besier, Thor F; Fregly, Benjamin J; Delp, Scott L; Banks, Scott A; Pandy, Marcus G; D'Lima, Darryl D; Lloyd, David G
2013-11-15
Estimating tibiofemoral joint contact forces is important for understanding the initiation and progression of knee osteoarthritis. However, tibiofemoral contact force predictions are influenced by many factors including muscle forces and anatomical representations of the knee joint. This study aimed to investigate the influence of subject-specific geometry and knee joint kinematics on the prediction of tibiofemoral contact forces using a calibrated EMG-driven neuromusculoskeletal model of the knee. One participant fitted with an instrumented total knee replacement walked at a self-selected speed while medial and lateral tibiofemoral contact forces, ground reaction forces, whole-body kinematics, and lower-limb muscle activity were simultaneously measured. The combination of generic and subject-specific knee joint geometry and kinematics resulted in four different OpenSim models used to estimate muscle-tendon lengths and moment arms. The subject-specific geometric model was created from CT scans and the subject-specific knee joint kinematics representing the translation of the tibia relative to the femur was obtained from fluoroscopy. The EMG-driven model was calibrated using one walking trial, but with three different cost functions that tracked the knee flexion/extension moments with and without constraint over the estimated joint contact forces. The calibrated models then predicted the medial and lateral tibiofemoral contact forces for five other different walking trials. The use of subject-specific models with minimization of the peak tibiofemoral contact forces improved the accuracy of medial contact forces by 47% and lateral contact forces by 7%, respectively compared with the use of generic musculoskeletal model. © 2013 Published by Elsevier Ltd.
Subject-specific knee joint geometry improves predictions of medial tibiofemoral contact forces
Gerus, Pauline; Sartori, Massimo; Besier, Thor F.; Fregly, Benjamin J.; Delp, Scott L.; Banks, Scott A.; Pandy, Marcus G.; D’Lima, Darryl D.; Lloyd, David G.
2013-01-01
Estimating tibiofemoral joint contact forces is important for understanding the initiation and progression of knee osteoarthritis. However, tibiofemoral contact force predictions are influenced by many factors including muscle forces and anatomical representations of the knee joint. This study aimed to investigate the influence of subject-specific geometry and knee joint kinematics on the prediction of tibiofemoral contact forces using a calibrated EMG-driven neuromusculoskeletal model of the knee. One participant fitted with an instrumented total knee replacement walked at a self-selected speed while medial and lateral tibiofemoral contact forces, ground reaction forces, whole-body kinematics, and lower-limb muscle activity were simultaneously measured. The combination of generic and subject-specific knee joint geometry and kinematics resulted in four different OpenSim models used to estimate muscle-tendon lengths and moment arms. The subject-specific geometric model was created from CT scans and the subject-specific knee joint kinematics representing the translation of the tibia relative to the femur was obtained from fluoroscopy. The EMG-driven model was calibrated using one walking trial, but with three different cost functions that tracked the knee flexion/extension moments with and without constraint over the estimated joint contact forces. The calibrated models then predicted the medial and lateral tibiofemoral contact forces for five other different walking trials. The use of subject-specific models with minimization of the peak tibiofemoral contact forces improved the accuracy of medial contact forces by 47% and lateral contact forces by 7%, respectively compared with the use of generic musculoskeletal model. PMID:24074941
Quantifying the predictive consequences of model error with linear subspace analysis
White, Jeremy T.; Doherty, John E.; Hughes, Joseph D.
2014-01-01
All computer models are simplified and imperfect simulators of complex natural systems. The discrepancy arising from simplification induces bias in model predictions, which may be amplified by the process of model calibration. This paper presents a new method to identify and quantify the predictive consequences of calibrating a simplified computer model. The method is based on linear theory, and it scales efficiently to the large numbers of parameters and observations characteristic of groundwater and petroleum reservoir models. The method is applied to a range of predictions made with a synthetic integrated surface-water/groundwater model with thousands of parameters. Several different observation processing strategies and parameterization/regularization approaches are examined in detail, including use of the Karhunen-Loève parameter transformation. Predictive bias arising from model error is shown to be prediction specific and often invisible to the modeler. The amount of calibration-induced bias is influenced by several factors, including how expert knowledge is applied in the design of parameterization schemes, the number of parameters adjusted during calibration, how observations and model-generated counterparts are processed, and the level of fit with observations achieved through calibration. Failure to properly implement any of these factors in a prediction-specific manner may increase the potential for predictive bias in ways that are not visible to the calibration and uncertainty analysis process.
Dynamic prediction in functional concurrent regression with an application to child growth.
Leroux, Andrew; Xiao, Luo; Crainiceanu, Ciprian; Checkley, William
2018-04-15
In many studies, it is of interest to predict the future trajectory of subjects based on their historical data, referred to as dynamic prediction. Mixed effects models have traditionally been used for dynamic prediction. However, the commonly used random intercept and slope model is often not sufficiently flexible for modeling subject-specific trajectories. In addition, there may be useful exposures/predictors of interest that are measured concurrently with the outcome, complicating dynamic prediction. To address these problems, we propose a dynamic functional concurrent regression model to handle the case where both the functional response and the functional predictors are irregularly measured. Currently, such a model cannot be fit by existing software. We apply the model to dynamically predict children's length conditional on prior length, weight, and baseline covariates. Inference on model parameters and subject-specific trajectories is conducted using the mixed effects representation of the proposed model. An extensive simulation study shows that the dynamic functional regression model provides more accurate estimation and inference than existing methods. Methods are supported by fast, flexible, open source software that uses heavily tested smoothing techniques. © 2017 The Authors. Statistics in Medicine Published by John Wiley & Sons Ltd.
Thermophysical properties of liquid UO2, ZrO2 and corium by molecular dynamics and predictive models
NASA Astrophysics Data System (ADS)
Kim, Woong Kee; Shim, Ji Hoon; Kaviany, Massoud
2017-08-01
Predicting the fate of accident-melted nuclear fuel-cladding requires the understanding of the thermophysical properties which are lacking or have large scatter due to high-temperature experimental challenges. Using equilibrium classical molecular dynamics (MD), we predict the properties of melted UO2 and ZrO2 and compare them with the available experimental data and the predictive models. The existing interatomic potential models have been developed mainly for the polymorphic solid phases of these oxides, so they cannot be used to predict all the properties accurately. We compare and decipher the distinctions of those MD predictions using the specific property-related autocorrelation decays. The predicted properties are density, specific heat, heat of fusion, compressibility, viscosity, surface tension, and the molecular and electronic thermal conductivities. After the comparisons, we provide readily usable temperature-dependent correlations (including UO2-ZrO2 compounds, i.e. corium melt).
Simultaneous prediction of binding free energy and specificity for PDZ domain-peptide interactions
NASA Astrophysics Data System (ADS)
Crivelli, Joseph J.; Lemmon, Gordon; Kaufmann, Kristian W.; Meiler, Jens
2013-12-01
Interactions between protein domains and linear peptides underlie many biological processes. Among these interactions, the recognition of C-terminal peptides by PDZ domains is one of the most ubiquitous. In this work, we present a mathematical model for PDZ domain-peptide interactions capable of predicting both affinity and specificity of binding based on X-ray crystal structures and comparative modeling with R osetta. We developed our mathematical model using a large phage display dataset describing binding specificity for a wild type PDZ domain and 91 single mutants, as well as binding affinity data for a wild type PDZ domain binding to 28 different peptides. Structural refinement was carried out through several R osetta protocols, the most accurate of which included flexible peptide docking and several iterations of side chain repacking and backbone minimization. Our findings emphasize the importance of backbone flexibility and the energetic contributions of side chain-side chain hydrogen bonds in accurately predicting interactions. We also determined that predicting PDZ domain-peptide interactions became increasingly challenging as the length of the peptide increased in the N-terminal direction. In the training dataset, predicted binding energies correlated with those derived through calorimetry and specificity switches introduced through single mutations at interface positions were recapitulated. In independent tests, our best performing protocol was capable of predicting dissociation constants well within one order of magnitude of the experimental values and specificity profiles at the level of accuracy of previous studies. To our knowledge, this approach represents the first integrated protocol for predicting both affinity and specificity for PDZ domain-peptide interactions.
Modeled changes of cerebellar activity in mutant mice are predictive of their learning impairments
NASA Astrophysics Data System (ADS)
Badura, Aleksandra; Clopath, Claudia; Schonewille, Martijn; de Zeeuw, Chris I.
2016-11-01
Translating neuronal activity to measurable behavioral changes has been a long-standing goal of systems neuroscience. Recently, we have developed a model of phase-reversal learning of the vestibulo-ocular reflex, a well-established, cerebellar-dependent task. The model, comprising both the cerebellar cortex and vestibular nuclei, reproduces behavioral data and accounts for the changes in neural activity during learning in wild type mice. Here, we used our model to predict Purkinje cell spiking as well as behavior before and after learning of five different lines of mutant mice with distinct cell-specific alterations of the cerebellar cortical circuitry. We tested these predictions by obtaining electrophysiological data depicting changes in neuronal spiking. We show that our data is largely consistent with the model predictions for simple spike modulation of Purkinje cells and concomitant behavioral learning in four of the mutants. In addition, our model accurately predicts a shift in simple spike activity in a mutant mouse with a brainstem specific mutation. This combination of electrophysiological and computational techniques opens a possibility of predicting behavioral impairments from neural activity.
Modeled changes of cerebellar activity in mutant mice are predictive of their learning impairments
Badura, Aleksandra; Clopath, Claudia; Schonewille, Martijn; De Zeeuw, Chris I.
2016-01-01
Translating neuronal activity to measurable behavioral changes has been a long-standing goal of systems neuroscience. Recently, we have developed a model of phase-reversal learning of the vestibulo-ocular reflex, a well-established, cerebellar-dependent task. The model, comprising both the cerebellar cortex and vestibular nuclei, reproduces behavioral data and accounts for the changes in neural activity during learning in wild type mice. Here, we used our model to predict Purkinje cell spiking as well as behavior before and after learning of five different lines of mutant mice with distinct cell-specific alterations of the cerebellar cortical circuitry. We tested these predictions by obtaining electrophysiological data depicting changes in neuronal spiking. We show that our data is largely consistent with the model predictions for simple spike modulation of Purkinje cells and concomitant behavioral learning in four of the mutants. In addition, our model accurately predicts a shift in simple spike activity in a mutant mouse with a brainstem specific mutation. This combination of electrophysiological and computational techniques opens a possibility of predicting behavioral impairments from neural activity. PMID:27805050
Drought Prediction Site Specific and Regional up to Three Years in Advance
NASA Astrophysics Data System (ADS)
Suhler, G.; O'Brien, D. P.
2002-12-01
Dynamic Predictables has developed proprietary software that analyzes and predicts future climatic behavior based on past data. The programs employ both a regional thermodynamic model together with a unique predictive algorithm to achieve a high degree of prediction accuracy up to 36 months. The thermodynamic model was developed initially to explain the results of a study on global circulation models done at SUNY-Stony Brook by S. Hameed, R.G. Currie, and H. LaGrone (Int. Jour. Climatology, 15, pp.852-871, 1995). The authors pointed out that on a time scale of 2-70 months the spectrum of sea level pressure is dominated by the harmonics and subharmonics of the seasonal cycle and their combination tones. These oscillations are fundamental to an understanding of climatic variations on a sub-regional to continental basis. The oscillatory nature of these variations allows them to be used as broad based climate predictors. In addition, they can be subtracted from the data to yield residuals. The residuals are then analyzed to determine components that are predictable. The program then combines both the thermodynamic model results (the primary predictive model) with those from the residual data (the secondary model) to yield an estimate of the future behavior of the climatic variable. Spatial resolution is site specific or aggregated regional based upon appropriate length (45 years or more monthly data) and reasonable quality weather observation records. Most climate analysis has been based on monthly time-step data, but time scales on the order of days can be used. Oregon Climate Division 1 (Coastal) precipitation provides an example relating DynaPred's method to nature's observed elements in the early 2000s. The prediction's leading dynamic factors are the strong seasonal in the primary model combined with high secondary model contributions from planet Earth's Chandler Wobble (near 15 months) and what has been called the Quasi-Triennial Oscillation (QTO, near 36 months) in equatorial regions. Examples of regional aggregate and site-specific predictions previously made blind forward and publicly available (AASC Annual Meetings 1998-2002) will be shown. Certain climate dynamics features relevant to extrema prediction and specifically drought prediction will then be discussed. Time steps presented will be monthly. Climate variables examined are mean temperature and accumulated precipitation. NINO3 SST, interior continental and marine/continental transition area examples will be shown. http://www.dynamicpredictables.com
Using a knowledge-based planning solution to select patients for proton therapy.
Delaney, Alexander R; Dahele, Max; Tol, Jim P; Kuijper, Ingrid T; Slotman, Ben J; Verbakel, Wilko F A R
2017-08-01
Patient selection for proton therapy by comparing proton/photon treatment plans is time-consuming and prone to bias. RapidPlan™, a knowledge-based-planning solution, uses plan-libraries to model and predict organ-at-risk (OAR) dose-volume-histograms (DVHs). We investigated whether RapidPlan, utilizing an algorithm based only on photon beam characteristics, could generate proton DVH-predictions and whether these could correctly identify patients for proton therapy. Model PROT and Model PHOT comprised 30 head-and-neck cancer proton and photon plans, respectively. Proton and photon knowledge-based-plans (KBPs) were made for ten evaluation-patients. DVH-prediction accuracy was analyzed by comparing predicted-vs-achieved mean OAR doses. KBPs and manual plans were compared using salivary gland and swallowing muscle mean doses. For illustration, patients were selected for protons if predicted Model PHOT mean dose minus predicted Model PROT mean dose (ΔPrediction) for combined OARs was ≥6Gy, and benchmarked using achieved KBP doses. Achieved and predicted Model PROT /Model PHOT mean dose R 2 was 0.95/0.98. Generally, achieved mean dose for Model PHOT /Model PROT KBPs was respectively lower/higher than predicted. Comparing Model PROT /Model PHOT KBPs with manual plans, salivary and swallowing mean doses increased/decreased by <2Gy, on average. ΔPrediction≥6Gy correctly selected 4 of 5 patients for protons. Knowledge-based DVH-predictions can provide efficient, patient-specific selection for protons. A proton-specific RapidPlan-solution could improve results. Copyright © 2017 Elsevier B.V. All rights reserved.
Modeling ready biodegradability of fragrance materials.
Ceriani, Lidia; Papa, Ester; Kovarich, Simona; Boethling, Robert; Gramatica, Paola
2015-06-01
In the present study, quantitative structure activity relationships were developed for predicting ready biodegradability of approximately 200 heterogeneous fragrance materials. Two classification methods, classification and regression tree (CART) and k-nearest neighbors (kNN), were applied to perform the modeling. The models were validated with multiple external prediction sets, and the structural applicability domain was verified by the leverage approach. The best models had good sensitivity (internal ≥80%; external ≥68%), specificity (internal ≥80%; external 73%), and overall accuracy (≥75%). Results from the comparison with BIOWIN global models, based on group contribution method, show that specific models developed in the present study perform better in prediction than BIOWIN6, in particular for the correct classification of not readily biodegradable fragrance materials. © 2015 SETAC.
Ding, Ziyun; Nolte, Daniel; Kit Tsang, Chui; Cleather, Daniel J; Kedgley, Angela E; Bull, Anthony M J
2016-02-01
Segment-based musculoskeletal models allow the prediction of muscle, ligament, and joint forces without making assumptions regarding joint degrees-of-freedom (DOF). The dataset published for the "Grand Challenge Competition to Predict in vivo Knee Loads" provides directly measured tibiofemoral contact forces for activities of daily living (ADL). For the Sixth Grand Challenge Competition to Predict in vivo Knee Loads, blinded results for "smooth" and "bouncy" gait trials were predicted using a customized patient-specific musculoskeletal model. For an unblinded comparison, the following modifications were made to improve the predictions: further customizations, including modifications to the knee center of rotation; reductions to the maximum allowable muscle forces to represent known loss of strength in knee arthroplasty patients; and a kinematic constraint to the hip joint to address the sensitivity of the segment-based approach to motion tracking artifact. For validation, the improved model was applied to normal gait, squat, and sit-to-stand for three subjects. Comparisons of the predictions with measured contact forces showed that segment-based musculoskeletal models using patient-specific input data can estimate tibiofemoral contact forces with root mean square errors (RMSEs) of 0.48-0.65 times body weight (BW) for normal gait trials. Comparisons between measured and predicted tibiofemoral contact forces yielded an average coefficient of determination of 0.81 and RMSEs of 0.46-1.01 times BW for squatting and 0.70-0.99 times BW for sit-to-stand tasks. This is comparable to the best validations in the literature using alternative models.
Rios, Anthony; Kavuluru, Ramakanth
2017-11-01
The CEGS N-GRID 2016 Shared Task in Clinical Natural Language Processing (NLP) provided a set of 1000 neuropsychiatric notes to participants as part of a competition to predict psychiatric symptom severity scores. This paper summarizes our methods, results, and experiences based on our participation in the second track of the shared task. Classical methods of text classification usually fall into one of three problem types: binary, multi-class, and multi-label classification. In this effort, we study ordinal regression problems with text data where misclassifications are penalized differently based on how far apart the ground truth and model predictions are on the ordinal scale. Specifically, we present our entries (methods and results) in the N-GRID shared task in predicting research domain criteria (RDoC) positive valence ordinal symptom severity scores (absent, mild, moderate, and severe) from psychiatric notes. We propose a novel convolutional neural network (CNN) model designed to handle ordinal regression tasks on psychiatric notes. Broadly speaking, our model combines an ordinal loss function, a CNN, and conventional feature engineering (wide features) into a single model which is learned end-to-end. Given interpretability is an important concern with nonlinear models, we apply a recent approach called locally interpretable model-agnostic explanation (LIME) to identify important words that lead to instance specific predictions. Our best model entered into the shared task placed third among 24 teams and scored a macro mean absolute error (MMAE) based normalized score (100·(1-MMAE)) of 83.86. Since the competition, we improved our score (using basic ensembling) to 85.55, comparable with the winning shared task entry. Applying LIME to model predictions, we demonstrate the feasibility of instance specific prediction interpretation by identifying words that led to a particular decision. In this paper, we present a method that successfully uses wide features and an ordinal loss function applied to convolutional neural networks for ordinal text classification specifically in predicting psychiatric symptom severity scores. Our approach leads to excellent performance on the N-GRID shared task and is also amenable to interpretability using existing model-agnostic approaches. Copyright © 2017 Elsevier Inc. All rights reserved.
Dusenberry, Michael W; Brown, Charles K; Brewer, Kori L
2017-02-01
To construct an artificial neural network (ANN) model that can predict the presence of acute CT findings with both high sensitivity and high specificity when applied to the population of patients≥age 65years who have incurred minor head injury after a fall. An ANN was created in the Python programming language using a population of 514 patients ≥ age 65 years presenting to the ED with minor head injury after a fall. The patient dataset was divided into three parts: 60% for "training", 20% for "cross validation", and 20% for "testing". Sensitivity, specificity, positive and negative predictive values, and accuracy were determined by comparing the model's predictions to the actual correct answers for each patient. On the "cross validation" data, the model attained a sensitivity ("recall") of 100.00%, specificity of 78.95%, PPV ("precision") of 78.95%, NPV of 100.00%, and accuracy of 88.24% in detecting the presence of positive head CTs. On the "test" data, the model attained a sensitivity of 97.78%, specificity of 89.47%, PPV of 88.00%, NPV of 98.08%, and accuracy of 93.14% in detecting the presence of positive head CTs. ANNs show great potential for predicting CT findings in the population of patients ≥ 65 years of age presenting with minor head injury after a fall. As a good first step, the ANN showed comparable sensitivity, predictive values, and accuracy, with a much higher specificity than the existing decision rules in clinical usage for predicting head CTs with acute intracranial findings. Copyright © 2016 Elsevier Inc. All rights reserved.
Meguid, Robert A; Bronsert, Michael R; Juarez-Colunga, Elizabeth; Hammermeister, Karl E; Henderson, William G
2016-07-01
To develop parsimonious prediction models for postoperative mortality, overall morbidity, and 6 complication clusters applicable to a broad range of surgical operations in adult patients. Quantitative risk assessment tools are not routinely used for preoperative patient assessment, shared decision making, informed consent, and preoperative patient optimization, likely due in part to the burden of data collection and the complexity of incorporation into routine surgical practice. Multivariable forward selection stepwise logistic regression analyses were used to develop predictive models for 30-day mortality, overall morbidity, and 6 postoperative complication clusters, using 40 preoperative variables from 2,275,240 surgical cases in the American College of Surgeons National Surgical Quality Improvement Program data set, 2005 to 2012. For the mortality and overall morbidity outcomes, prediction models were compared with and without preoperative laboratory variables, and generic models (based on all of the data from 9 surgical specialties) were compared with specialty-specific models. In each model, the cumulative c-index was used to examine the contribution of each added predictor variable. C-indexes, Hosmer-Lemeshow analyses, and Brier scores were used to compare discrimination and calibration between models. For the mortality and overall morbidity outcomes, the prediction models without the preoperative laboratory variables performed as well as the models with the laboratory variables, and the generic models performed as well as the specialty-specific models. The c-indexes were 0.938 for mortality, 0.810 for overall morbidity, and for the 6 complication clusters ranged from 0.757 for infectious to 0.897 for pulmonary complications. Across the 8 prediction models, the first 7 to 11 variables entered accounted for at least 99% of the c-index of the full model (using up to 28 nonlaboratory predictor variables). Our results suggest that it will be possible to develop parsimonious models to predict 8 important postoperative outcomes for a broad surgical population, without the need for surgeon specialty-specific models or inclusion of laboratory variables.
Predicting readmission risk with institution-specific prediction models.
Yu, Shipeng; Farooq, Faisal; van Esbroeck, Alexander; Fung, Glenn; Anand, Vikram; Krishnapuram, Balaji
2015-10-01
The ability to predict patient readmission risk is extremely valuable for hospitals, especially under the Hospital Readmission Reduction Program of the Center for Medicare and Medicaid Services which went into effect starting October 1, 2012. There is a plethora of work in the literature that deals with developing readmission risk prediction models, but most of them do not have sufficient prediction accuracy to be deployed in a clinical setting, partly because different hospitals may have different characteristics in their patient populations. We propose a generic framework for institution-specific readmission risk prediction, which takes patient data from a single institution and produces a statistical risk prediction model optimized for that particular institution and, optionally, for a specific condition. This provides great flexibility in model building, and is also able to provide institution-specific insights in its readmitted patient population. We have experimented with classification methods such as support vector machines, and prognosis methods such as the Cox regression. We compared our methods with industry-standard methods such as the LACE model, and showed the proposed framework is not only more flexible but also more effective. We applied our framework to patient data from three hospitals, and obtained some initial results for heart failure (HF), acute myocardial infarction (AMI), pneumonia (PN) patients as well as patients with all conditions. On Hospital 2, the LACE model yielded AUC 0.57, 0.56, 0.53 and 0.55 for AMI, HF, PN and All Cause readmission prediction, respectively, while the proposed model yielded 0.66, 0.65, 0.63, 0.74 for the corresponding conditions, all significantly better than the LACE counterpart. The proposed models that leverage all features at discharge time is more accurate than the models that only leverage features at admission time (0.66 vs. 0.61 for AMI, 0.65 vs. 0.61 for HF, 0.63 vs. 0.56 for PN, 0.74 vs. 0.60 for All Cause). Furthermore, the proposed admission-time models already outperform the performance of LACE, which is a discharge-time model (0.61 vs. 0.57 for AMI, 0.61 vs. 0.56 for HF, 0.56 vs. 0.53 for PN, 0.60 vs. 0.55 for All Cause). Similar conclusions can be drawn from other hospitals as well. The same performance comparison also holds for precision and recall at top-decile predictions. Most of the performance improvements are statistically significant. The institution-specific readmission risk prediction framework is more flexible and more effective than the one-size-fit-all models like the LACE, sometimes twice and three-time more effective. The admission-time models are able to give early warning signs compared to the discharge-time models, and may be able to help hospital staff intervene early while the patient is still in the hospital. Copyright © 2015 Elsevier B.V. All rights reserved.
Latent Patient Cluster Discovery for Robust Future Forecasting and New-Patient Generalization
Masino, Aaron J.
2016-01-01
Commonly referred to as predictive modeling, the use of machine learning and statistical methods to improve healthcare outcomes has recently gained traction in biomedical informatics research. Given the vast opportunities enabled by large Electronic Health Records (EHR) data and powerful resources for conducting predictive modeling, we argue that it is yet crucial to first carefully examine the prediction task and then choose predictive methods accordingly. Specifically, we argue that there are at least three distinct prediction tasks that are often conflated in biomedical research: 1) data imputation, where a model fills in the missing values in a dataset, 2) future forecasting, where a model projects the development of a medical condition for a known patient based on existing observations, and 3) new-patient generalization, where a model transfers the knowledge learned from previously observed patients to newly encountered ones. Importantly, the latter two tasks—future forecasting and new-patient generalizations—tend to be more difficult than data imputation as they require predictions to be made on potentially out-of-sample data (i.e., data following a different predictable pattern from what has been learned by the model). Using hearing loss progression as an example, we investigate three regression models and show that the modeling of latent clusters is a robust method for addressing the more challenging prediction scenarios. Overall, our findings suggest that there exist significant differences between various kinds of prediction tasks and that it is important to evaluate the merits of a predictive model relative to the specific purpose of a prediction task. PMID:27636203
Latent Patient Cluster Discovery for Robust Future Forecasting and New-Patient Generalization.
Qian, Ting; Masino, Aaron J
2016-01-01
Commonly referred to as predictive modeling, the use of machine learning and statistical methods to improve healthcare outcomes has recently gained traction in biomedical informatics research. Given the vast opportunities enabled by large Electronic Health Records (EHR) data and powerful resources for conducting predictive modeling, we argue that it is yet crucial to first carefully examine the prediction task and then choose predictive methods accordingly. Specifically, we argue that there are at least three distinct prediction tasks that are often conflated in biomedical research: 1) data imputation, where a model fills in the missing values in a dataset, 2) future forecasting, where a model projects the development of a medical condition for a known patient based on existing observations, and 3) new-patient generalization, where a model transfers the knowledge learned from previously observed patients to newly encountered ones. Importantly, the latter two tasks-future forecasting and new-patient generalizations-tend to be more difficult than data imputation as they require predictions to be made on potentially out-of-sample data (i.e., data following a different predictable pattern from what has been learned by the model). Using hearing loss progression as an example, we investigate three regression models and show that the modeling of latent clusters is a robust method for addressing the more challenging prediction scenarios. Overall, our findings suggest that there exist significant differences between various kinds of prediction tasks and that it is important to evaluate the merits of a predictive model relative to the specific purpose of a prediction task.
Qian, Weiguo; Tang, Yunjia; Yan, Wenhua; Sun, Ling; Lv, Haitao
2018-03-09
Kawasaki disease (KD) is the most common pediatric vasculitis. Several models have been established to predict intravenous immunoglobulin (IVIG) resistance. The present study was aimed to evaluate the efficacy of prediction models using the medical data of KD patients. We collected the medical records of patients hospitalized in the Department of Cardiology in Children's Hospital of Soochow University with a diagnosis of KD from Jan 2015 to Dec 2016. IVIG resistance was defined as recrudescent or persistent fever ≥36 h after the end of their IVIG infusion. Patients with IVIG resistance tended to be younger, have higher occurrence of rash and changes of extremities. They had higher levels of c-reactive protein, aspartate aminotransferase, neutrophils proportion (N%), total bilirubin and lower level of albumin. Our prediction model had a sensitivity of 0.72 and a specificity of 0.75. Sensitivity of Kobayashi, Egami, Kawamura, Sano and Formosa were 0.72, 0.44, 0.48, 0.20, and 0.68, respectively. Specificity of these models were 0.62, 0.82, 0.66, 0.91, and 0.48, respectively. Our prediction model had a powerful predictive value in this area, followed by Kobayashi model while all the other prediction models had less excellent performances than ours.
Modelling proteins' hidden conformations to predict antibiotic resistance
NASA Astrophysics Data System (ADS)
Hart, Kathryn M.; Ho, Chris M. W.; Dutta, Supratik; Gross, Michael L.; Bowman, Gregory R.
2016-10-01
TEM β-lactamase confers bacteria with resistance to many antibiotics and rapidly evolves activity against new drugs. However, functional changes are not easily explained by differences in crystal structures. We employ Markov state models to identify hidden conformations and explore their role in determining TEM's specificity. We integrate these models with existing drug-design tools to create a new technique, called Boltzmann docking, which better predicts TEM specificity by accounting for conformational heterogeneity. Using our MSMs, we identify hidden states whose populations correlate with activity against cefotaxime. To experimentally detect our predicted hidden states, we use rapid mass spectrometric footprinting and confirm our models' prediction that increased cefotaxime activity correlates with reduced Ω-loop flexibility. Finally, we design novel variants to stabilize the hidden cefotaximase states, and find their populations predict activity against cefotaxime in vitro and in vivo. Therefore, we expect this framework to have numerous applications in drug and protein design.
A First Step towards a Clinical Decision Support System for Post-traumatic Stress Disorders.
Ma, Sisi; Galatzer-Levy, Isaac R; Wang, Xuya; Fenyö, David; Shalev, Arieh Y
2016-01-01
PTSD is distressful and debilitating, following a non-remitting course in about 10% to 20% of trauma survivors. Numerous risk indicators of PTSD have been identified, but individual level prediction remains elusive. As an effort to bridge the gap between scientific discovery and practical application, we designed and implemented a clinical decision support pipeline to provide clinically relevant recommendation for trauma survivors. To meet the specific challenge of early prediction, this work uses data obtained within ten days of a traumatic event. The pipeline creates personalized predictive model for each individual, and computes quality metrics for each predictive model. Clinical recommendations are made based on both the prediction of the model and its quality, thus avoiding making potentially detrimental recommendations based on insufficient information or suboptimal model. The current pipeline outperforms the acute stress disorder, a commonly used clinical risk factor for PTSD development, both in terms of sensitivity and specificity.
Francy, Donna S.; Brady, Amie M.G.; Carvin, Rebecca B.; Corsi, Steven R.; Fuller, Lori M.; Harrison, John H.; Hayhurst, Brett A.; Lant, Jeremiah; Nevers, Meredith B.; Terrio, Paul J.; Zimmerman, Tammy M.
2013-01-01
Predictive models have been used at beaches to improve the timeliness and accuracy of recreational water-quality assessments over the most common current approach to water-quality monitoring, which relies on culturing fecal-indicator bacteria such as Escherichia coli (E. coli.). Beach-specific predictive models use environmental and water-quality variables that are easily and quickly measured as surrogates to estimate concentrations of fecal-indicator bacteria or to provide the probability that a State recreational water-quality standard will be exceeded. When predictive models are used for beach closure or advisory decisions, they are referred to as “nowcasts.” During the recreational seasons of 2010-12, the U.S. Geological Survey (USGS), in cooperation with 23 local and State agencies, worked to improve existing nowcasts at 4 beaches, validate predictive models at another 38 beaches, and collect data for predictive-model development at 7 beaches throughout the Great Lakes. This report summarizes efforts to collect data and develop predictive models by multiple agencies and to compile existing information on the beaches and beach-monitoring programs into one comprehensive report. Local agencies measured E. coli concentrations and variables expected to affect E. coli concentrations such as wave height, turbidity, water temperature, and numbers of birds at the time of sampling. In addition to these field measurements, equipment was installed by the USGS or local agencies at or near several beaches to collect water-quality and metrological measurements in near real time, including nearshore buoys, weather stations, and tributary staff gages and monitors. The USGS worked with local agencies to retrieve data from existing sources either manually or by use of tools designed specifically to compile and process data for predictive-model development. Predictive models were developed by use of linear regression and (or) partial least squares techniques for 42 beaches that had at least 2 years of data (2010-11 and sometimes earlier) and for 1 beach that had 1 year of data. For most models, software designed for model development by the U.S. Environmental Protection Agency (Virtual Beach) was used. The selected model for each beach was based on a combination of explanatory variables including, most commonly, turbidity, day of the year, change in lake level over 24 hours, wave height, wind direction and speed, and antecedent rainfall for various time periods. Forty-two predictive models were validated against data collected during an independent year (2012) and compared to the current method for assessing recreational water quality-using the previous day’s E. coli concentration (persistence model). Goals for good predictive-model performance were responses that were at least 5 percent greater than the persistence model and overall correct responses greater than or equal to 80 percent, sensitivities (percentage of exceedances of the bathing-water standard that were correctly predicted by the model) greater than or equal to 50 percent, and specificities (percentage of nonexceedances correctly predicted by the model) greater than or equal to 85 percent. Out of 42 predictive models, 24 models yielded over-all correct responses that were at least 5 percent greater than the use of the persistence model. Predictive-model responses met the performance goals more often than the persistence-model responses in terms of overall correctness (28 versus 17 models, respectively), sensitivity (17 versus 4 models), and specificity (34 versus 25 models). Gaining knowledge of each beach and the factors that affect E. coli concentrations is important for developing good predictive models. Collection of additional years of data with a wide range of environmental conditions may also help to improve future model performance. The USGS will continue to work with local agencies in 2013 and beyond to develop and validate predictive models at beaches and improve existing nowcasts, restructuring monitoring activities to accommodate future uncertainties in funding and resources.
Neale, Peta A; Leusch, Frederic D L; Escher, Beate I
2017-04-01
Pharmaceuticals and antibiotics co-occur in the aquatic environment but mixture studies to date have mainly focused on pharmaceuticals alone or antibiotics alone, although differences in mode of action may lead to different effects in mixtures. In this study we used the Bacterial Luminescence Toxicity Screen (BLT-Screen) after acute (0.5 h) and chronic (16 h) exposure to evaluate how non-specifically acting pharmaceuticals and specifically acting antibiotics act together in mixtures. Three models were applied to predict mixture toxicity including concentration addition, independent action and the two-step prediction (TSP) model, which groups similarly acting chemicals together using concentration addition, followed by independent action to combine the two groups. All non-antibiotic pharmaceuticals had similar EC 50 values at both 0.5 and 16 h, indicating together with a QSAR (Quantitative Structure-Activity Relationship) analysis that they act as baseline toxicants. In contrast, the antibiotics' EC 50 values decreased by up to three orders of magnitude after 16 h, which can be explained by their specific effect on bacteria. Equipotent mixtures of non-antibiotic pharmaceuticals only, antibiotics only and both non-antibiotic pharmaceuticals and antibiotics were prepared based on the single chemical results. The mixture toxicity models were all in close agreement with the experimental results, with predicted EC 50 values within a factor of two of the experimental results. This suggests that concentration addition can be applied to bacterial assays to model the mixture effects of environmental samples containing both specifically and non-specifically acting chemicals. Copyright © 2017 Elsevier Ltd. All rights reserved.
Stuart P. Cottrell; Alan R. Graefe
1995-01-01
This paper examines predictors of boater behavior in a specific behavior situation, namely the percentage of raw sewage discharged from recreational vessels in a sanitation pumpout facility on the Chesapeake Bay. Results of a multiple regression analysis show knowledge predicts behavior in specific issue situations. In addition, the more specific the...
Collins, G S; Reitsma, J B; Altman, D G; Moons, K G M
2015-01-20
Prediction models are developed to aid health-care providers in estimating the probability or risk that a specific disease or condition is present (diagnostic models) or that a specific event will occur in the future (prognostic models), to inform their decision making. However, the overwhelming evidence shows that the quality of reporting of prediction model studies is poor. Only with full and clear reporting of information on all aspects of a prediction model can risk of bias and potential usefulness of prediction models be adequately assessed. The Transparent Reporting of a multivariable prediction model for Individual Prognosis Or Diagnosis (TRIPOD) Initiative developed a set of recommendations for the reporting of studies developing, validating, or updating a prediction model, whether for diagnostic or prognostic purposes. This article describes how the TRIPOD Statement was developed. An extensive list of items based on a review of the literature was created, which was reduced after a Web-based survey and revised during a 3-day meeting in June 2011 with methodologists, health-care professionals, and journal editors. The list was refined during several meetings of the steering group and in e-mail discussions with the wider group of TRIPOD contributors. The resulting TRIPOD Statement is a checklist of 22 items, deemed essential for transparent reporting of a prediction model study. The TRIPOD Statement aims to improve the transparency of the reporting of a prediction model study regardless of the study methods used. The TRIPOD Statement is best used in conjunction with the TRIPOD explanation and elaboration document. To aid the editorial process and readers of prediction model studies, it is recommended that authors include a completed checklist in their submission (also available at www.tripod-statement.org).
Collins, Gary S; Reitsma, Johannes B; Altman, Douglas G; Moons, Karel G M
2015-02-01
Prediction models are developed to aid healthcare providers in estimating the probability or risk that a specific disease or condition is present (diagnostic models) or that a specific event will occur in the future (prognostic models), to inform their decision-making. However, the overwhelming evidence shows that the quality of reporting of prediction model studies is poor. Only with full and clear reporting of information on all aspects of a prediction model can risk of bias and potential usefulness of prediction models be adequately assessed. The Transparent Reporting of a multivariable prediction model for Individual Prognosis Or Diagnosis (TRIPOD) initiative developed a set of recommendations for the reporting of studies developing, validating or updating a prediction model, whether for diagnostic or prognostic purposes. This article describes how the TRIPOD Statement was developed. An extensive list of items based on a review of the literature was created, which was reduced after a Web-based survey and revised during a 3-day meeting in June 2011 with methodologists, healthcare professionals and journal editors. The list was refined during several meetings of the steering group and in e-mail discussions with the wider group of TRIPOD contributors. The resulting TRIPOD Statement is a checklist of 22 items, deemed essential for transparent reporting of a prediction model study. The TRIPOD Statement aims to improve the transparency of the reporting of a prediction model study regardless of the study methods used. The TRIPOD Statement is best used in conjunction with the TRIPOD explanation and elaboration document. To aid the editorial process and readers of prediction model studies, it is recommended that authors include a completed checklist in their submission (also available at www.tripod-statement.org). © 2015 Stichting European Society for Clinical Investigation Journal Foundation.
Collins, Gary S; Reitsma, Johannes B; Altman, Douglas G; Moons, Karel G M
2015-01-06
Prediction models are developed to aid health care providers in estimating the probability or risk that a specific disease or condition is present (diagnostic models) or that a specific event will occur in the future (prognostic models), to inform their decision making. However, the overwhelming evidence shows that the quality of reporting of prediction model studies is poor. Only with full and clear reporting of information on all aspects of a prediction model can risk of bias and potential usefulness of prediction models be adequately assessed. The Transparent Reporting of a multivariable prediction model for Individual Prognosis Or Diagnosis (TRIPOD) Initiative developed a set of recommendations for the reporting of studies developing, validating, or updating a prediction model, whether for diagnostic or prognostic purposes. This article describes how the TRIPOD Statement was developed. An extensive list of items based on a review of the literature was created, which was reduced after a Web-based survey and revised during a 3-day meeting in June 2011 with methodologists, health care professionals, and journal editors. The list was refined during several meetings of the steering group and in e-mail discussions with the wider group of TRIPOD contributors. The resulting TRIPOD Statement is a checklist of 22 items, deemed essential for transparent reporting of a prediction model study. The TRIPOD Statement aims to improve the transparency of the reporting of a prediction model study regardless of the study methods used. The TRIPOD Statement is best used in conjunction with the TRIPOD explanation and elaboration document. To aid the editorial process and readers of prediction model studies, it is recommended that authors include a completed checklist in their submission (also available at www.tripod-statement.org).
Reitsma, Johannes B.; Altman, Douglas G.; Moons, Karel G.M.
2015-01-01
Background— Prediction models are developed to aid health care providers in estimating the probability or risk that a specific disease or condition is present (diagnostic models) or that a specific event will occur in the future (prognostic models), to inform their decision making. However, the overwhelming evidence shows that the quality of reporting of prediction model studies is poor. Only with full and clear reporting of information on all aspects of a prediction model can risk of bias and potential usefulness of prediction models be adequately assessed. Methods— The Transparent Reporting of a multivariable prediction model for Individual Prognosis Or Diagnosis (TRIPOD) Initiative developed a set of recommendations for the reporting of studies developing, validating, or updating a prediction model, whether for diagnostic or prognostic purposes. This article describes how the TRIPOD Statement was developed. An extensive list of items based on a review of the literature was created, which was reduced after a Web-based survey and revised during a 3-day meeting in June 2011 with methodologists, health care professionals, and journal editors. The list was refined during several meetings of the steering group and in e-mail discussions with the wider group of TRIPOD contributors. Results— The resulting TRIPOD Statement is a checklist of 22 items, deemed essential for transparent reporting of a prediction model study. The TRIPOD Statement aims to improve the transparency of the reporting of a prediction model study regardless of the study methods used. The TRIPOD Statement is best used in conjunction with the TRIPOD explanation and elaboration document. Conclusions— To aid the editorial process and readers of prediction model studies, it is recommended that authors include a completed checklist in their submission (also available at www.tripod-statement.org). PMID:25561516
Collins, G S; Reitsma, J B; Altman, D G; Moons, K G M
2015-01-01
Prediction models are developed to aid health-care providers in estimating the probability or risk that a specific disease or condition is present (diagnostic models) or that a specific event will occur in the future (prognostic models), to inform their decision making. However, the overwhelming evidence shows that the quality of reporting of prediction model studies is poor. Only with full and clear reporting of information on all aspects of a prediction model can risk of bias and potential usefulness of prediction models be adequately assessed. The Transparent Reporting of a multivariable prediction model for Individual Prognosis Or Diagnosis (TRIPOD) Initiative developed a set of recommendations for the reporting of studies developing, validating, or updating a prediction model, whether for diagnostic or prognostic purposes. This article describes how the TRIPOD Statement was developed. An extensive list of items based on a review of the literature was created, which was reduced after a Web-based survey and revised during a 3-day meeting in June 2011 with methodologists, health-care professionals, and journal editors. The list was refined during several meetings of the steering group and in e-mail discussions with the wider group of TRIPOD contributors. The resulting TRIPOD Statement is a checklist of 22 items, deemed essential for transparent reporting of a prediction model study. The TRIPOD Statement aims to improve the transparency of the reporting of a prediction model study regardless of the study methods used. The TRIPOD Statement is best used in conjunction with the TRIPOD explanation and elaboration document. To aid the editorial process and readers of prediction model studies, it is recommended that authors include a completed checklist in their submission (also available at www.tripod-statement.org). PMID:25562432
Collins, G S; Reitsma, J B; Altman, D G; Moons, K G M
2015-02-01
Prediction models are developed to aid health care providers in estimating the probability or risk that a specific disease or condition is present (diagnostic models) or that a specific event will occur in the future (prognostic models), to inform their decision making. However, the overwhelming evidence shows that the quality of reporting of prediction model studies is poor. Only with full and clear reporting of information on all aspects of a prediction model can risk of bias and potential usefulness of prediction models be adequately assessed. The Transparent Reporting of a multivariable prediction model for Individual Prognosis or Diagnosis (TRIPOD) Initiative developed a set of recommendations for the reporting of studies developing, validating, or updating a prediction model, whether for diagnostic or prognostic purposes. This article describes how the TRIPOD Statement was developed. An extensive list of items based on a review of the literature was created, which was reduced after a Web-based survey and revised during a 3-day meeting in June 2011 with methodologists, health care professionals, and journal editors. The list was refined during several meetings of the steering group and in e-mail discussions with the wider group of TRIPOD contributors. The resulting TRIPOD Statement is a checklist of 22 items, deemed essential for transparent reporting of a prediction model study. The TRIPOD Statement aims to improve the transparency of the reporting of a prediction model study regardless of the study methods used. The TRIPOD Statement is best used in conjunction with the TRIPOD explanation and elaboration document. To aid the editorial process and readers of prediction model studies, it is recommended that authors include a completed checklist in their submission (also available at www.tripod-statement.org). © 2015 Royal College of Obstetricians and Gynaecologists.
Collins, Gary S; Reitsma, Johannes B; Altman, Douglas G; Moons, Karel G M
2015-01-13
Prediction models are developed to aid health care providers in estimating the probability or risk that a specific disease or condition is present (diagnostic models) or that a specific event will occur in the future (prognostic models), to inform their decision making. However, the overwhelming evidence shows that the quality of reporting of prediction model studies is poor. Only with full and clear reporting of information on all aspects of a prediction model can risk of bias and potential usefulness of prediction models be adequately assessed. The Transparent Reporting of a multivariable prediction model for Individual Prognosis Or Diagnosis (TRIPOD) Initiative developed a set of recommendations for the reporting of studies developing, validating, or updating a prediction model, whether for diagnostic or prognostic purposes. This article describes how the TRIPOD Statement was developed. An extensive list of items based on a review of the literature was created, which was reduced after a Web-based survey and revised during a 3-day meeting in June 2011 with methodologists, health care professionals, and journal editors. The list was refined during several meetings of the steering group and in e-mail discussions with the wider group of TRIPOD contributors. The resulting TRIPOD Statement is a checklist of 22 items, deemed essential for transparent reporting of a prediction model study. The TRIPOD Statement aims to improve the transparency of the reporting of a prediction model study regardless of the study methods used. The TRIPOD Statement is best used in conjunction with the TRIPOD explanation and elaboration document. To aid the editorial process and readers of prediction model studies, it is recommended that authors include a completed checklist in their submission (also available at www.tripod-statement.org). © 2015 The Authors.
Collins, Gary S; Reitsma, Johannes B; Altman, Douglas G; Moons, Karel G M
2015-01-06
Prediction models are developed to aid health care providers in estimating the probability or risk that a specific disease or condition is present (diagnostic models) or that a specific event will occur in the future (prognostic models), to inform their decision making. However, the overwhelming evidence shows that the quality of reporting of prediction model studies is poor. Only with full and clear reporting of information on all aspects of a prediction model can risk of bias and potential usefulness of prediction models be adequately assessed. The Transparent Reporting of a multivariable prediction model for Individual Prognosis Or Diagnosis (TRIPOD) Initiative developed a set of recommendations for the reporting of studies developing, validating, or updating a prediction model, whether for diagnostic or prognostic purposes. This article describes how the TRIPOD Statement was developed. An extensive list of items based on a review of the literature was created, which was reduced after a Web-based survey and revised during a 3-day meeting in June 2011 with methodologists, health care professionals, and journal editors. The list was refined during several meetings of the steering group and in e-mail discussions with the wider group of TRIPOD contributors. The resulting TRIPOD Statement is a checklist of 22 items, deemed essential for transparent reporting of a prediction model study. The TRIPOD Statement aims to improve the transparency of the reporting of a prediction model study regardless of the study methods used. The TRIPOD Statement is best used in conjunction with the TRIPOD explanation and elaboration document. To aid the editorial process and readers of prediction model studies, it is recommended that authors include a completed checklist in their submission (also available at www.tripod-statement.org).
Collins, Gary S; Reitsma, Johannes B; Altman, Douglas G; Moons, Karel G M
2015-02-01
Prediction models are developed to aid health care providers in estimating the probability or risk that a specific disease or condition is present (diagnostic models) or that a specific event will occur in the future (prognostic models), to inform their decision making. However, the overwhelming evidence shows that the quality of reporting of prediction model studies is poor. Only with full and clear reporting of information on all aspects of a prediction model can risk of bias and potential usefulness of prediction models be adequately assessed. The Transparent Reporting of a multivariable prediction model for Individual Prognosis Or Diagnosis (TRIPOD) Initiative developed a set of recommendations for the reporting of studies developing, validating, or updating a prediction model, whether for diagnostic or prognostic purposes. This article describes how the TRIPOD Statement was developed. An extensive list of items based on a review of the literature was created, which was reduced after a Web-based survey and revised during a 3-day meeting in June 2011 with methodologists, health care professionals, and journal editors. The list was refined during several meetings of the steering group and in e-mail discussions with the wider group of TRIPOD contributors. The resulting TRIPOD Statement is a checklist of 22 items, deemed essential for transparent reporting of a prediction model study. The TRIPOD Statement aims to improve the transparency of the reporting of a prediction model study regardless of the study methods used. The TRIPOD Statement is best used in conjunction with the TRIPOD explanation and elaboration document. To aid the editorial process and readers of prediction model studies, it is recommended that authors include a completed checklist in their submission (also available at www.tripod-statement.org). Copyright © 2015 Elsevier Inc. All rights reserved.
ERIC Educational Resources Information Center
Miller, Daniel C.; Maricle, Denise E.; Jones, Alicia M.
2016-01-01
Processing Strengths and Weaknesses (PSW) models have been proposed as a method for identifying specific learning disabilities. Three PSW models were examined for their ability to predict expert identified specific learning disabilities cases. The Dual Discrepancy/Consistency Model (DD/C; Flanagan, Ortiz, & Alfonso, 2013) as operationalized by…
Reeves, Mari Kathryn; Perdue, Margaret; Munk, Lee Ann; Hagedorn, Birgit
2018-07-15
Studies of environmental processes exhibit spatial variation within data sets. The ability to derive predictions of risk from field data is a critical path forward in understanding the data and applying the information to land and resource management. Thanks to recent advances in predictive modeling, open source software, and computing, the power to do this is within grasp. This article provides an example of how we predicted relative trace element pollution risk from roads across a region by combining site specific trace element data in soils with regional land cover and planning information in a predictive model framework. In the Kenai Peninsula of Alaska, we sampled 36 sites (191 soil samples) adjacent to roads for trace elements. We then combined this site specific data with freely-available land cover and urban planning data to derive a predictive model of landscape scale environmental risk. We used six different model algorithms to analyze the dataset, comparing these in terms of their predictive abilities and the variables identified as important. Based on comparable predictive abilities (mean R 2 from 30 to 35% and mean root mean square error from 65 to 68%), we averaged all six model outputs to predict relative levels of trace element deposition in soils-given the road surface, traffic volume, sample distance from the road, land cover category, and impervious surface percentage. Mapped predictions of environmental risk from toxic trace element pollution can show land managers and transportation planners where to prioritize road renewal or maintenance by each road segment's relative environmental and human health risk. Published by Elsevier B.V.
A Predictive Model for Readmissions Among Medicare Patients in a California Hospital.
Duncan, Ian; Huynh, Nhan
2017-11-17
Predictive models for hospital readmission rates are in high demand because of the Centers for Medicare & Medicaid Services (CMS) Hospital Readmission Reduction Program (HRRP). The LACE index is one of the most popular predictive tools among hospitals in the United States. The LACE index is a simple tool with 4 parameters: Length of stay, Acuity of admission, Comorbidity, and Emergency visits in the previous 6 months. The authors applied logistic regression to develop a predictive model for a medium-sized not-for-profit community hospital in California using patient-level data with more specific patient information (including 13 explanatory variables). Specifically, the logistic regression is applied to 2 populations: a general population including all patients and the specific group of patients targeted by the CMS penalty (characterized as ages 65 or older with select conditions). The 2 resulting logistic regression models have a higher sensitivity rate compared to the sensitivity of the LACE index. The C statistic values of the model applied to both populations demonstrate moderate levels of predictive power. The authors also build an economic model to demonstrate the potential financial impact of the use of the model for targeting high-risk patients in a sample hospital and demonstrate that, on balance, whether the hospital gains or loses from reducing readmissions depends on its margin and the extent of its readmission penalties.
Li, Mao; Miller, Karol; Joldes, Grand Roman; Kikinis, Ron; Wittek, Adam
2016-01-01
Patient-specific biomechanical models have been advocated as a tool for predicting deformations of soft body organs/tissue for medical image registration (aligning two sets of images) when differences between the images are large. However, complex and irregular geometry of the body organs makes generation of patient-specific biomechanical models very time consuming. Meshless discretisation has been proposed to solve this challenge. However, applications so far have been limited to 2-D models and computing single organ deformations. In this study, 3-D comprehensive patient-specific non-linear biomechanical models implemented using Meshless Total Lagrangian Explicit Dynamics (MTLED) algorithms are applied to predict a 3-D deformation field for whole-body image registration. Unlike a conventional approach which requires dividing (segmenting) the image into non-overlapping constituents representing different organs/tissues, the mechanical properties are assigned using the Fuzzy C-Means (FCM) algorithm without the image segmentation. Verification indicates that the deformations predicted using the proposed meshless approach are for practical purposes the same as those obtained using the previously validated finite element models. To quantitatively evaluate the accuracy of the predicted deformations, we determined the spatial misalignment between the registered (i.e. source images warped using the predicted deformations) and target images by computing the edge-based Hausdorff distance. The Hausdorff distance-based evaluation determines that our meshless models led to successful registration of the vast majority of the image features. PMID:26791945
Pneumococcal vaccine targeting strategy for older adults: customized risk profiling.
Balicer, Ran D; Cohen, Chandra J; Leibowitz, Morton; Feldman, Becca S; Brufman, Ilan; Roberts, Craig; Hoshen, Moshe
2014-02-12
Current pneumococcal vaccine campaigns take a broad, primarily age-based approach to immunization targeting, overlooking many clinical and administrative considerations necessary in disease prevention and resource planning for specific patient populations. We aim to demonstrate the utility of a population-specific predictive model for hospital-treated pneumonia to direct effective vaccine targeting. Data was extracted for 1,053,435 members of an Israeli HMO, age 50 and older, during the study period 2008-2010. We developed and validated a logistic regression model to predict hospital-treated pneumonia using training and test samples, including a set of standard and population-specific risk factors. The model's predictive value was tested for prospectively identifying cases of pneumonia and invasive pneumococcal disease (IPD), and was compared to the existing international paradigm for patient immunization targeting. In a multivariate regression, age, co-morbidity burden and previous pneumonia events were most strongly positively associated with hospital-treated pneumonia. The model predicting hospital-treated pneumonia yielded a c-statistic of 0.80. Utilizing the predictive model, the top 17% highest-risk within the study validation population were targeted to detect 54% of those members who were subsequently treated for hospitalized pneumonia in the follow up period. The high-risk population identified through this model included 46% of the follow-up year's IPD cases, and 27% of community-treated pneumonia cases. These outcomes were compared with international guidelines for risk for pneumococcal diseases that accurately identified only 35% of hospitalized pneumonia, 41% of IPD cases and 21% of community-treated pneumonia. We demonstrate that a customized model for vaccine targeting performs better than international guidelines, and therefore, risk modeling may allow for more precise vaccine targeting and resource allocation than current national and international guidelines. Health care managers and policy-makers may consider the strategic potential of utilizing clinical and administrative databases for creating population-specific risk prediction models to inform vaccination campaigns. Copyright © 2013 Elsevier Ltd. All rights reserved.
Predicting What Will Happen When You Intervene.
Cartwright, Nancy; Hardie, Jeremy
2017-01-01
This paper offers some rules of thumb that practicing social workers can use for case studies that aim to construct, albeit not fully and never entirely reliably, models designed to help predict what will happen if they intervene in specific ways to help this particular client, here and now. We call these 'ex ante case-specific causal models'. 'Ex ante' because they are for before-the-fact prediction of what the likely effects of proposed actions are. 'Case-specific' because we are not concerned with studies that provide evidence for some general conclusion but rather with using what general and local knowledge one can get to predict what will happen to a specific client in the real settings in which they live. 'Causal' because this kind of case study aims to trace out as best possible the web of causal processes that will be responsible for what happens. In this sense our case studies resemble post facto realist evaluations.
Ramezani Tehrani, Fahimeh; Mansournia, Mohammad Ali; Solaymani-Dodaran, Masoud; Steyerberg, Ewout; Azizi, Fereidoun
2016-06-01
This study aimed to improve existing prediction models for age at menopause. We identified all reproductive aged women with regular menstrual cycles who met our eligibility criteria (n = 1,015) in the Tehran Lipid and Glucose Study-an ongoing population-based cohort study initiated in 1998. Participants were examined every 3 years and their reproductive histories were recorded. Blood levels of antimüllerian hormone (AMH) were measured at the time of recruitment. Age at menopause was estimated based on serum concentrations of AMH using flexible parametric survival models. The optimum model was selected according to Akaike Information Criteria and the realness of the range of predicted median menopause age. We followed study participants for a median of 9.8 years during which 277 women reached menopause and found that a spline-based proportional odds model including age-specific AMH percentiles as the covariate performed well in terms of statistical criteria and provided the most clinically relevant and realistic predictions. The range of predicted median age at menopause for this model was 47.1 to 55.9 years. For those who reached menopause, the median of the absolute mean difference between actual and predicted age at menopause was 1.9 years (interquartile range 2.9). The model including the age-specific AMH percentiles as the covariate and using proportional odds as its covariate metrics meets all the statistical criteria for the best model and provides the most clinically relevant and realistic predictions for age at menopause for reproductive-aged women.
Xiao, Li-Hong; Chen, Pei-Ran; Gou, Zhong-Ping; Li, Yong-Zhong; Li, Mei; Xiang, Liang-Cheng; Feng, Ping
2017-01-01
The aim of this study is to evaluate the ability of the random forest algorithm that combines data on transrectal ultrasound findings, age, and serum levels of prostate-specific antigen to predict prostate carcinoma. Clinico-demographic data were analyzed for 941 patients with prostate diseases treated at our hospital, including age, serum prostate-specific antigen levels, transrectal ultrasound findings, and pathology diagnosis based on ultrasound-guided needle biopsy of the prostate. These data were compared between patients with and without prostate cancer using the Chi-square test, and then entered into the random forest model to predict diagnosis. Patients with and without prostate cancer differed significantly in age and serum prostate-specific antigen levels (P < 0.001), as well as in all transrectal ultrasound characteristics (P < 0.05) except uneven echo (P = 0.609). The random forest model based on age, prostate-specific antigen and ultrasound predicted prostate cancer with an accuracy of 83.10%, sensitivity of 65.64%, and specificity of 93.83%. Positive predictive value was 86.72%, and negative predictive value was 81.64%. By integrating age, prostate-specific antigen levels and transrectal ultrasound findings, the random forest algorithm shows better diagnostic performance for prostate cancer than either diagnostic indicator on its own. This algorithm may help improve diagnosis of the disease by identifying patients at high risk for biopsy.
Knowledge-guided fuzzy logic modeling to infer cellular signaling networks from proteomic data
Liu, Hui; Zhang, Fan; Mishra, Shital Kumar; Zhou, Shuigeng; Zheng, Jie
2016-01-01
Modeling of signaling pathways is crucial for understanding and predicting cellular responses to drug treatments. However, canonical signaling pathways curated from literature are seldom context-specific and thus can hardly predict cell type-specific response to external perturbations; purely data-driven methods also have drawbacks such as limited biological interpretability. Therefore, hybrid methods that can integrate prior knowledge and real data for network inference are highly desirable. In this paper, we propose a knowledge-guided fuzzy logic network model to infer signaling pathways by exploiting both prior knowledge and time-series data. In particular, the dynamic time warping algorithm is employed to measure the goodness of fit between experimental and predicted data, so that our method can model temporally-ordered experimental observations. We evaluated the proposed method on a synthetic dataset and two real phosphoproteomic datasets. The experimental results demonstrate that our model can uncover drug-induced alterations in signaling pathways in cancer cells. Compared with existing hybrid models, our method can model feedback loops so that the dynamical mechanisms of signaling networks can be uncovered from time-series data. By calibrating generic models of signaling pathways against real data, our method supports precise predictions of context-specific anticancer drug effects, which is an important step towards precision medicine. PMID:27774993
Aguiar, Fabio S; Almeida, Luciana L; Ruffino-Netto, Antonio; Kritski, Afranio Lineu; Mello, Fernanda Cq; Werneck, Guilherme L
2012-08-07
Tuberculosis (TB) remains a public health issue worldwide. The lack of specific clinical symptoms to diagnose TB makes the correct decision to admit patients to respiratory isolation a difficult task for the clinician. Isolation of patients without the disease is common and increases health costs. Decision models for the diagnosis of TB in patients attending hospitals can increase the quality of care and decrease costs, without the risk of hospital transmission. We present a predictive model for predicting pulmonary TB in hospitalized patients in a high prevalence area in order to contribute to a more rational use of isolation rooms without increasing the risk of transmission. Cross sectional study of patients admitted to CFFH from March 2003 to December 2004. A classification and regression tree (CART) model was generated and validated. The area under the ROC curve (AUC), sensitivity, specificity, positive and negative predictive values were used to evaluate the performance of model. Validation of the model was performed with a different sample of patients admitted to the same hospital from January to December 2005. We studied 290 patients admitted with clinical suspicion of TB. Diagnosis was confirmed in 26.5% of them. Pulmonary TB was present in 83.7% of the patients with TB (62.3% with positive sputum smear) and HIV/AIDS was present in 56.9% of patients. The validated CART model showed sensitivity, specificity, positive predictive value and negative predictive value of 60.00%, 76.16%, 33.33%, and 90.55%, respectively. The AUC was 79.70%. The CART model developed for these hospitalized patients with clinical suspicion of TB had fair to good predictive performance for pulmonary TB. The most important variable for prediction of TB diagnosis was chest radiograph results. Prospective validation is still necessary, but our model offer an alternative for decision making in whether to isolate patients with clinical suspicion of TB in tertiary health facilities in countries with limited resources.
Helbling, Damian E; Johnson, David R; Lee, Tae Kwon; Scheidegger, Andreas; Fenner, Kathrin
2015-03-01
The rates at which wastewater treatment plant (WWTP) microbial communities biotransform specific substrates can differ by orders of magnitude among WWTP communities. Differences in taxonomic compositions among WWTP communities may predict differences in the rates of some types of biotransformations. In this work, we present a novel framework for establishing predictive relationships between specific bacterial 16S rRNA sequence abundances and biotransformation rates. We selected ten WWTPs with substantial variation in their environmental and operational metrics and measured the in situ ammonia biotransformation rate constants in nine of them. We isolated total RNA from samples from each WWTP and analyzed 16S rRNA sequence reads. We then developed multivariate models between the measured abundances of specific bacterial 16S rRNA sequence reads and the ammonia biotransformation rate constants. We constructed model scenarios that systematically explored the effects of model regularization, model linearity and non-linearity, and aggregation of 16S rRNA sequences into operational taxonomic units (OTUs) as a function of sequence dissimilarity threshold (SDT). A large percentage (greater than 80%) of model scenarios resulted in well-performing and significant models at intermediate SDTs of 0.13-0.14 and 0.26. The 16S rRNA sequences consistently selected into the well-performing and significant models at those SDTs were classified as Nitrosomonas and Nitrospira groups. We then extend the framework by applying it to the biotransformation rate constants of ten micropollutants measured in batch reactors seeded with the ten WWTP communities. We identified phylogenetic groups that were robustly selected into all well-performing and significant models constructed with biotransformation rates of isoproturon, propachlor, ranitidine, and venlafaxine. These phylogenetic groups can be used as predictive biomarkers of WWTP microbial community activity towards these specific micropollutants. This work is an important step towards developing tools to predict biotransformation rates in WWTPs based on taxonomic composition. Copyright © 2014 Elsevier Ltd. All rights reserved.
Fukuda, Haruhisa; Kuroki, Manabu
2016-03-01
To develop and internally validate a surgical site infection (SSI) prediction model for Japan. Retrospective observational cohort study. We analyzed surveillance data submitted to the Japan Nosocomial Infections Surveillance system for patients who had undergone target surgical procedures from January 1, 2010, through December 31, 2012. Logistic regression analyses were used to develop statistical models for predicting SSIs. An SSI prediction model was constructed for each of the procedure categories by statistically selecting the appropriate risk factors from among the collected surveillance data and determining their optimal categorization. Standard bootstrapping techniques were applied to assess potential overfitting. The C-index was used to compare the predictive performances of the new statistical models with those of models based on conventional risk index variables. The study sample comprised 349,987 cases from 428 participant hospitals throughout Japan, and the overall SSI incidence was 7.0%. The C-indices of the new statistical models were significantly higher than those of the conventional risk index models in 21 (67.7%) of the 31 procedure categories (P<.05). No significant overfitting was detected. Japan-specific SSI prediction models were shown to generally have higher accuracy than conventional risk index models. These new models may have applications in assessing hospital performance and identifying high-risk patients in specific procedure categories.
Polak, Marta E; Ung, Chuin Ying; Masapust, Joanna; Freeman, Tom C; Ardern-Jones, Michael R
2017-04-06
Langerhans cells (LCs) are able to orchestrate adaptive immune responses in the skin by interpreting the microenvironmental context in which they encounter foreign substances, but the regulatory basis for this has not been established. Utilising systems immunology approaches combining in silico modelling of a reconstructed gene regulatory network (GRN) with in vitro validation of the predictions, we sought to determine the mechanisms of regulation of immune responses in human primary LCs. The key role of Interferon regulatory factors (IRFs) as controllers of the human Langerhans cell response to epidermal cytokines was revealed by whole transcriptome analysis. Applying Boolean logic we assembled a Petri net-based model of the IRF-GRN which provides molecular pathway predictions for the induction of different transcriptional programmes in LCs. In silico simulations performed after model parameterisation with transcription factor expression values predicted that human LC activation of antigen-specific CD8 T cells would be differentially regulated by epidermal cytokine induction of specific IRF-controlled pathways. This was confirmed by in vitro measurement of IFN-γ production by activated T cells. As a proof of concept, this approach shows that stochastic modelling of a specific immune networks renders transcriptome data valuable for the prediction of functional outcomes of immune responses.
Panagou, Efstathios Z; Nychas, George-John E
2008-09-01
A product-specific model was developed and validated under dynamic temperature conditions for predicting the growth of Listeria monocytogenes in pasteurized vanilla cream, a traditional milk-based product. Model performance was also compared with Growth Predictor and Sym'Previus predictive microbiology software packages. Commercially prepared vanilla cream samples were artificially inoculated with a five-strain cocktail of L. monocytogenes, with an initial concentration of 102 CFU g(-1), and stored at 3, 5, 10, and 15 degrees C for 36 days. The growth kinetic parameters at each temperature were determined by the primary model of Baranyi and Roberts. The maximum specific growth rate (mu(max)) was further modeled as a function of temperature by means of a square root-type model. The performance of the model in predicting the growth of the pathogen under dynamic temperature conditions was based on two different temperature scenarios with periodic changes from 4 to 15 degrees C. Growth prediction for dynamic temperature profiles was based on the square root model and the differential equations of the Baranyi and Roberts model, which were numerically integrated with respect to time. Model performance was based on the bias factor (B(f)), the accuracy factor (A(f)), the goodness-of-fit index (GoF), and the percent relative errors between observed and predicted growth. The product-specific model developed in the present study accurately predicted the growth of L. monocytogenes under dynamic temperature conditions. The average values for the performance indices were 1.038, 1.068, and 0.397 for B(f), A(f), and GoF, respectively for both temperature scenarios assayed. Predictions from Growth Predictor and Sym'Previus overestimated pathogen growth. The average values of B(f), A(f), and GoF were 1.173, 1.174, 1.162, and 0.956, 1.115, 0.713 for [corrected] Growth Predictor and Sym'Previus, respectively.
The Rangeland Hydrology and Erosion Model: A dynamic approach for predicting soil loss on rangelands
USDA-ARS?s Scientific Manuscript database
In this study we present the improved Rangeland Hydrology and Erosion Model (RHEM V2.3), a process-based erosion prediction tool specific for rangeland application. The article provides the mathematical formulation of the model and parameter estimation equations. Model performance is assessed agains...
ERIC Educational Resources Information Center
Simpkins, Sandra D.; Fredricks, Jennifer A.; Eccles, Jacquelynne S.
2012-01-01
The Eccles' expectancy-value model posits that a cascade of mechanisms explain associations between parents' beliefs and youths' achievement-related behaviors. Specifically, parents' beliefs predict parents' behaviors; in turn, parents' behaviors predict youths' motivational beliefs, and youths' motivational beliefs predict their behaviors. This…
Accurate and dynamic predictive model for better prediction in medicine and healthcare.
Alanazi, H O; Abdullah, A H; Qureshi, K N; Ismail, A S
2018-05-01
Information and communication technologies (ICTs) have changed the trend into new integrated operations and methods in all fields of life. The health sector has also adopted new technologies to improve the systems and provide better services to customers. Predictive models in health care are also influenced from new technologies to predict the different disease outcomes. However, still, existing predictive models have suffered from some limitations in terms of predictive outcomes performance. In order to improve predictive model performance, this paper proposed a predictive model by classifying the disease predictions into different categories. To achieve this model performance, this paper uses traumatic brain injury (TBI) datasets. TBI is one of the serious diseases worldwide and needs more attention due to its seriousness and serious impacts on human life. The proposed predictive model improves the predictive performance of TBI. The TBI data set is developed and approved by neurologists to set its features. The experiment results show that the proposed model has achieved significant results including accuracy, sensitivity, and specificity.
ERIC Educational Resources Information Center
Bhagat, Rabi S.; And Others
The role of attitudes in the conduct of buyer behavior is examined in the context of two competitive models of attitude structure and attitude-behavior relationship. Specifically, the objectives of the study were to compare the Fishbein and Sheth models on the criteria of predictive as well as cross validities. Data on both the models were…
Nøst, Therese Haugdahl; Breivik, Knut; Wania, Frank; Rylander, Charlotta; Odland, Jon Øyvind; Sandanger, Torkjel Manning
2015-01-01
Background Studies on the health effects of polychlorinated biphenyls (PCBs) call for an understanding of past and present human exposure. Time-resolved mechanistic models may supplement information on concentrations in individuals obtained from measurements and/or statistical approaches if they can be shown to reproduce empirical data. Objectives Here, we evaluated the capability of one such mechanistic model to reproduce measured PCB concentrations in individual Norwegian women. We also assessed individual life-course concentrations. Methods Concentrations of four PCB congeners in pregnant (n = 310, sampled in 2007–2009) and postmenopausal (n = 244, 2005) women were compared with person-specific predictions obtained using CoZMoMAN, an emission-based environmental fate and human food-chain bioaccumulation model. Person-specific predictions were also made using statistical regression models including dietary and lifestyle variables and concentrations. Results CoZMoMAN accurately reproduced medians and ranges of measured concentrations in the two study groups. Furthermore, rank correlations between measurements and predictions from both CoZMoMAN and regression analyses were strong (Spearman’s r > 0.67). Precision in quartile assignments from predictions was strong overall as evaluated by weighted Cohen’s kappa (> 0.6). Simulations indicated large inter-individual differences in concentrations experienced in the past. Conclusions The mechanistic model reproduced all measurements of PCB concentrations within a factor of 10, and subject ranking and quartile assignments were overall largely consistent, although they were weak within each study group. Contamination histories for individuals predicted by CoZMoMAN revealed variation between study subjects, particularly in the timing of peak concentrations. Mechanistic models can provide individual PCB exposure metrics that could serve as valuable supplements to measurements. Citation Nøst TH, Breivik K, Wania F, Rylander C, Odland JØ, Sandanger TM. 2016. Estimating time-varying PCB exposures using person-specific predictions to supplement measured values: a comparison of observed and predicted values in two cohorts of Norwegian women. Environ Health Perspect 124:299–305; http://dx.doi.org/10.1289/ehp.1409191 PMID:26186800
Pre-operative prediction of surgical morbidity in children: comparison of five statistical models.
Cooper, Jennifer N; Wei, Lai; Fernandez, Soledad A; Minneci, Peter C; Deans, Katherine J
2015-02-01
The accurate prediction of surgical risk is important to patients and physicians. Logistic regression (LR) models are typically used to estimate these risks. However, in the fields of data mining and machine-learning, many alternative classification and prediction algorithms have been developed. This study aimed to compare the performance of LR to several data mining algorithms for predicting 30-day surgical morbidity in children. We used the 2012 National Surgical Quality Improvement Program-Pediatric dataset to compare the performance of (1) a LR model that assumed linearity and additivity (simple LR model) (2) a LR model incorporating restricted cubic splines and interactions (flexible LR model) (3) a support vector machine, (4) a random forest and (5) boosted classification trees for predicting surgical morbidity. The ensemble-based methods showed significantly higher accuracy, sensitivity, specificity, PPV, and NPV than the simple LR model. However, none of the models performed better than the flexible LR model in terms of the aforementioned measures or in model calibration or discrimination. Support vector machines, random forests, and boosted classification trees do not show better performance than LR for predicting pediatric surgical morbidity. After further validation, the flexible LR model derived in this study could be used to assist with clinical decision-making based on patient-specific surgical risks. Copyright © 2014 Elsevier Ltd. All rights reserved.
Predictive modeling of complications.
Osorio, Joseph A; Scheer, Justin K; Ames, Christopher P
2016-09-01
Predictive analytic algorithms are designed to identify patterns in the data that allow for accurate predictions without the need for a hypothesis. Therefore, predictive modeling can provide detailed and patient-specific information that can be readily applied when discussing the risks of surgery with a patient. There are few studies using predictive modeling techniques in the adult spine surgery literature. These types of studies represent the beginning of the use of predictive analytics in spine surgery outcomes. We will discuss the advancements in the field of spine surgery with respect to predictive analytics, the controversies surrounding the technique, and the future directions.
Subseasonal-to-Seasonal Science and Prediction Initiatives of the NOAA MAPP Program
NASA Astrophysics Data System (ADS)
Archambault, H. M.; Barrie, D.; Mariotti, A.
2016-12-01
There is great practical interest in developing predictions beyond the 2-week weather timescale. Scientific communities have historically organized themselves around the weather and climate problems, but the subseasonal-to-seasonal (S2S) timescale range overall is recognized as new territory for which a concerted shared effort is needed. For instance, the climate community, as part of programs like CLIVAR, has historically tackled coupled phenomena and modeling, keys to harnessing predictability on longer timescales. In contrast, the weather community has focused on synoptic dynamics, higher-resolution modeling, and enhanced model initialization, of importance at the shorter timescales and especially for the prediction of extremes. The processes and phenomena specific to timescales between weather and climate require a unified approach to science, modeling, and predictions. Internationally, the WWRP/WCRP S2S Prediction Project is a promising catalyzer for these types of activities. Among the various contributing U.S. research programs, the Modeling, Analysis, Predictions and Projections (MAPP) program, as part of the NOAA Climate Program Office, has launched coordinated research and transition activities that help to meet the agency's goals to fill the weather-to-climate prediction gap and will contribute to advance international goals. This presentation will describe ongoing MAPP program S2S science and prediction initiatives, specifically the MAPP S2S Task Force and the SubX prediction experiment.
Madaniyazi, Lina; Guo, Yuming; Chen, Renjie; Kan, Haidong; Tong, Shilu
2016-01-01
Estimating the burden of mortality associated with particulates requires knowledge of exposure-response associations. However, the evidence on exposure-response associations is limited in many cities, especially in developing countries. In this study, we predicted associations of particulates smaller than 10 μm in aerodynamic diameter (PM10) with mortality in 73 Chinese cities. The meta-regression model was used to test and quantify which city-specific characteristics contributed significantly to the heterogeneity of PM10-mortality associations for 16 Chinese cities. Then, those city-specific characteristics with statistically significant regression coefficients were treated as independent variables to build multivariate meta-regression models. The model with the best fitness was used to predict PM10-mortality associations in 73 Chinese cities in 2010. Mean temperature, PM10 concentration and green space per capita could best explain the heterogeneity in PM10-mortality associations. Based on city-specific characteristics, we were able to develop multivariate meta-regression models to predict associations between air pollutants and health outcomes reasonably well. Copyright © 2015 Elsevier Ltd. All rights reserved.
ERIC Educational Resources Information Center
Litchfield, Bradley C.
2013-01-01
This study examined the use of an institutionally-specific risk prediction model in the university's College of Education. Set in a large, urban, public university, the risk model predicted incoming students' first-semester GPAs, which, in turn, predicted the students' risk of attrition. Additionally, the study investigated advising practices…
Lee, Chu-Hee; Landham, Priyan R; Eastell, Richard; Adams, Michael A; Dolan, Patricia; Yang, Lang
2017-09-01
Finite element models of an isolated vertebral body cannot accurately predict compressive strength of the spinal column because, in life, compressive load is variably distributed across the vertebral body and neural arch. The purpose of this study was to develop and validate a patient-specific finite element model of a functional spinal unit, and then use the model to predict vertebral strength from medical images. A total of 16 cadaveric functional spinal units were scanned and then tested mechanically in bending and compression to generate a vertebral wedge fracture. Before testing, an image processing and finite element analysis framework (SpineVox-Pro), developed previously in MATLAB using ANSYS APDL, was used to generate a subject-specific finite element model with eight-node hexahedral elements. Transversely isotropic linear-elastic material properties were assigned to vertebrae, and simple homogeneous linear-elastic properties were assigned to the intervertebral disc. Forward bending loading conditions were applied to simulate manual handling. Results showed that vertebral strengths measured by experiment were positively correlated with strengths predicted by the functional spinal unit finite element model with von Mises or Drucker-Prager failure criteria ( R 2 = 0.80-0.87), with areal bone mineral density measured by dual-energy X-ray absorptiometry ( R 2 = 0.54) and with volumetric bone mineral density from quantitative computed tomography ( R 2 = 0.79). Large-displacement non-linear analyses on all specimens did not improve predictions. We conclude that subject-specific finite element models of a functional spinal unit have potential to estimate the vertebral strength better than bone mineral density alone.
Collins, G S; Reitsma, J B; Altman, D G; Moons, K G M
2015-02-01
Prediction models are developed to aid healthcare providers in estimating the probability or risk that a specific disease or condition is present (diagnostic models) or that a specific event will occur in the future (prognostic models), to inform their decision-making. However, the overwhelming evidence shows that the quality of reporting of prediction model studies is poor. Only with full and clear reporting of information on all aspects of a prediction model can risk of bias and potential usefulness of prediction models be adequately assessed. The Transparent Reporting of a multivariable prediction model for Individual Prognosis Or Diagnosis (TRIPOD) Initiative developed a set of recommendations for the reporting of studies developing, validating or updating a prediction model, whether for diagnostic or prognostic purposes. This article describes how the TRIPOD Statement was developed. An extensive list of items based on a review of the literature was created, which was reduced after a web-based survey and revised during a 3-day meeting in June 2011 with methodologists, healthcare professionals and journal editors. The list was refined during several meetings of the steering group and in e-mail discussions with the wider group of TRIPOD contributors. The resulting TRIPOD Statement is a checklist of 22 items, deemed essential for transparent reporting of a prediction model study. The TRIPOD Statement aims to improve the transparency of the reporting of a prediction model study regardless of the study methods used. The TRIPOD Statement is best used in conjunction with the TRIPOD explanation and elaboration document. A complete checklist is available at http://www.tripod-statement.org. © 2015 American College of Physicians.
Carbon and energy fluxes in cropland ecosystems: a model-data comparison
Lokupitiya, E.; Denning, A. Scott; Schaefer, K.; Ricciuto, D.; Anderson, R.; Arain, M. A.; Baker, I.; Barr, A. G.; Chen, G.; Chen, J.M.; Ciais, P.; Cook, D.R.; Dietze, M.C.; El Maayar, M.; Fischer, M.; Grant, R.; Hollinger, D.; Izaurralde, C.; Jain, A.; Kucharik, C.J.; Li, Z.; Liu, S.; Li, L.; Matamala, R.; Peylin, P.; Price, D.; Running, S. W.; Sahoo, A.; Sprintsin, M.; Suyker, A.E.; Tian, H.; Tonitto, Christina; Torn, M.S.; Verbeeck, Hans; Verma, S.B.; Xue, Y.
2016-01-01
Croplands are highly productive ecosystems that contribute to land–atmosphere exchange of carbon, energy, and water during their short growing seasons. We evaluated and compared net ecosystem exchange (NEE), latent heat flux (LE), and sensible heat flux (H) simulated by a suite of ecosystem models at five agricultural eddy covariance flux tower sites in the central United States as part of the North American Carbon Program Site Synthesis project. Most of the models overestimated H and underestimated LE during the growing season, leading to overall higher Bowen ratios compared to the observations. Most models systematically under predicted NEE, especially at rain-fed sites. Certain crop-specific models that were developed considering the high productivity and associated physiological changes in specific crops better predicted the NEE and LE at both rain-fed and irrigated sites. Models with specific parameterization for different crops better simulated the inter-annual variability of NEE for maize-soybean rotation compared to those models with a single generic crop type. Stratification according to basic model formulation and phenological methodology did not explain significant variation in model performance across these sites and crops. The under prediction of NEE and LE and over prediction of H by most of the models suggests that models developed and parameterized for natural ecosystems cannot accurately predict the more robust physiology of highly bred and intensively managed crop ecosystems. When coupled in Earth System Models, it is likely that the excessive physiological stress simulated in many land surface component models leads to overestimation of temperature and atmospheric boundary layer depth, and underestimation of humidity and CO2 seasonal uptake over agricultural regions.
Carbon and energy fluxes in cropland ecosystems: a model-data comparison
DOE Office of Scientific and Technical Information (OSTI.GOV)
Lokupitiya, E.; Denning, A. S.; Schaefer, K.
2016-06-03
Croplands are highly productive ecosystems that contribute to land–atmosphere exchange of carbon, energy, and water during their short growing seasons. We evaluated and compared net ecosystem exchange (NEE), latent heat flux (LE), and sensible heat flux (H) simulated by a suite of ecosystem models at five agricultural eddy covariance flux tower sites in the central United States as part of the North American Carbon Program Site Synthesis project. Most of the models overestimated H and underestimated LE during the growing season, leading to overall higher Bowen ratios compared to the observations. Most models systematically under predicted NEE, especially at rain-fedmore » sites. Certain crop-specific models that were developed considering the high productivity and associated physiological changes in specific crops better predicted the NEE and LE at both rain-fed and irrigated sites. Models with specific parameterization for different crops better simulated the inter-annual variability of NEE for maize-soybean rotation compared to those models with a single generic crop type. Stratification according to basic model formulation and phenological methodology did not explain significant variation in model performance across these sites and crops. The under prediction of NEE and LE and over prediction of H by most of the models suggests that models developed and parameterized for natural ecosystems cannot accurately predict the more robust physiology of highly bred and intensively managed crop ecosystems. When coupled in Earth System Models, it is likely that the excessive physiological stress simulated in many land surface component models leads to overestimation of temperature and atmospheric boundary layer depth, and underestimation of humidity and CO 2 seasonal uptake over agricultural regions.« less
Chu, Haitao; Nie, Lei; Cole, Stephen R; Poole, Charles
2009-08-15
In a meta-analysis of diagnostic accuracy studies, the sensitivities and specificities of a diagnostic test may depend on the disease prevalence since the severity and definition of disease may differ from study to study due to the design and the population considered. In this paper, we extend the bivariate nonlinear random effects model on sensitivities and specificities to jointly model the disease prevalence, sensitivities and specificities using trivariate nonlinear random-effects models. Furthermore, as an alternative parameterization, we also propose jointly modeling the test prevalence and the predictive values, which reflect the clinical utility of a diagnostic test. These models allow investigators to study the complex relationship among the disease prevalence, sensitivities and specificities; or among test prevalence and the predictive values, which can reveal hidden information about test performance. We illustrate the proposed two approaches by reanalyzing the data from a meta-analysis of radiological evaluation of lymph node metastases in patients with cervical cancer and a simulation study. The latter illustrates the importance of carefully choosing an appropriate normality assumption for the disease prevalence, sensitivities and specificities, or the test prevalence and the predictive values. In practice, it is recommended to use model selection techniques to identify a best-fitting model for making statistical inference. In summary, the proposed trivariate random effects models are novel and can be very useful in practice for meta-analysis of diagnostic accuracy studies. Copyright 2009 John Wiley & Sons, Ltd.
Vlachopoulos, Lazaros; Lüthi, Marcel; Carrillo, Fabio; Gerber, Christian; Székely, Gábor; Fürnstahl, Philipp
2018-04-18
In computer-assisted reconstructive surgeries, the contralateral anatomy is established as the best available reconstruction template. However, existing intra-individual bilateral differences or a pathological, contralateral humerus may limit the applicability of the method. The aim of the study was to evaluate whether a statistical shape model (SSM) has the potential to predict accurately the pretraumatic anatomy of the humerus from the posttraumatic condition. Three-dimensional (3D) triangular surface models were extracted from the computed tomographic data of 100 paired cadaveric humeri without a pathological condition. An SSM was constructed, encoding the characteristic shape variations among the individuals. To predict the patient-specific anatomy of the proximal (or distal) part of the humerus with the SSM, we generated segments of the humerus of predefined length excluding the part to predict. The proximal and distal humeral prediction (p-HP and d-HP) errors, defined as the deviation of the predicted (bone) model from the original (bone) model, were evaluated. For comparison with the state-of-the-art technique, i.e., the contralateral registration method, we used the same segments of the humerus to evaluate whether the SSM or the contralateral anatomy yields a more accurate reconstruction template. The p-HP error (mean and standard deviation, 3.8° ± 1.9°) using 85% of the distal end of the humerus to predict the proximal humeral anatomy was significantly smaller (p = 0.001) compared with the contralateral registration method. The difference between the d-HP error (mean, 5.5° ± 2.9°), using 85% of the proximal part of the humerus to predict the distal humeral anatomy, and the contralateral registration method was not significant (p = 0.61). The restoration of the humeral length was not significantly different between the SSM and the contralateral registration method. SSMs accurately predict the patient-specific anatomy of the proximal and distal aspects of the humerus. The prediction errors of the SSM depend on the size of the healthy part of the humerus. The prediction of the patient-specific anatomy of the humerus is of fundamental importance for computer-assisted reconstructive surgeries.
Rothe, Jessica; Watkins, Norman E.; Nagy, Marion
2012-01-01
Allele-specific extension reactions (ASERs) use 3′ terminus-specific primers for the selective extension of completely annealed matches by polymerase. The ability of the polymerase to extend non-specific 3′ terminal mismatches leads to a failure of the reaction, a process that is only partly understood and predictable, and often requires time-consuming assay design. In our studies we investigated haplotype-specific extraction (HSE) for the separation of male DNA mixtures. HSE is an ASER and provides the ability to distinguish between diploid chromosomes from one or more individuals. Here, we show that the success of HSE and allele-specific extension depend strongly on the concentration difference between complete match and 3′ terminal mismatch. Using the oligonucleotide-modeling platform Visual Omp, we demonstrated the dependency of the discrimination power of the polymerase on match- and mismatch-target hybridization between different probe lengths. Therefore, the probe specificity in HSE could be predicted by performing a relative comparison of different probe designs with their simulated differences between the duplex concentration of target-probe match and mismatches. We tested this new model for probe design in more than 300 HSE reactions with 137 different probes and obtained an accordance of 88%. PMID:23049901
Rothe, Jessica; Watkins, Norman E; Nagy, Marion
2012-01-01
Allele-specific extension reactions (ASERs) use 3' terminus-specific primers for the selective extension of completely annealed matches by polymerase. The ability of the polymerase to extend non-specific 3' terminal mismatches leads to a failure of the reaction, a process that is only partly understood and predictable, and often requires time-consuming assay design. In our studies we investigated haplotype-specific extraction (HSE) for the separation of male DNA mixtures. HSE is an ASER and provides the ability to distinguish between diploid chromosomes from one or more individuals. Here, we show that the success of HSE and allele-specific extension depend strongly on the concentration difference between complete match and 3' terminal mismatch. Using the oligonucleotide-modeling platform Visual Omp, we demonstrated the dependency of the discrimination power of the polymerase on match- and mismatch-target hybridization between different probe lengths. Therefore, the probe specificity in HSE could be predicted by performing a relative comparison of different probe designs with their simulated differences between the duplex concentration of target-probe match and mismatches. We tested this new model for probe design in more than 300 HSE reactions with 137 different probes and obtained an accordance of 88%.
A Prospective Test of Cognitive Vulnerability Models of Depression With Adolescent Girls
Bohon, Cara; Stice, Eric; Burton, Emily; Fudell, Molly; Nolen-Hoeksema, Susan
2009-01-01
This study sought to provide a more rigorous prospective test of two cognitive vulnerability models of depression with longitudinal data from 496 adolescent girls. Results supported the cognitive vulnerability model in that stressors predicted future increases in depressive symptoms and onset of clinically significant major depression for individuals with a negative attributional style, but not for those with a positive attributional style, although these effects were small. This model appeared to be specific to depression, in that it did not predict future increases in bulimia nervosa or substance abuse symptoms. In contrast, results did not support the integrated cognitive vulnerability self-esteem model that asserts stressors should only predict increased depression for individuals with a confluence of negative attributional style and low self-esteem, and this model did not appear to be specific to depression. PMID:18328873
DOE Office of Scientific and Technical Information (OSTI.GOV)
Hall, David C.; Trofimov, Alexei V.; Winey, Brian A.
Purpose: To predict the organ at risk (OAR) dose levels achievable with proton beam therapy (PBT), solely based on the geometric arrangement of the target volume in relation to the OARs. A comparison with an alternative therapy yields a prediction of the patient-specific benefits offered by PBT. This could enable physicians at hospitals without proton capabilities to make a better-informed referral decision or aid patient selection in model-based clinical trials. Methods and Materials: Skull-base tumors were chosen to test the method, owing to their geometric complexity and multitude of nearby OARs. By exploiting the correlations between the dose and distance-to-targetmore » in existing PBT plans, the models were independently trained for 6 types of OARs: brainstem, cochlea, optic chiasm, optic nerve, parotid gland, and spinal cord. Once trained, the models could estimate the feasible dose–volume histogram and generalized equivalent uniform dose (gEUD) for OAR structures of new patients. The models were trained using 20 patients and validated using an additional 21 patients. Validation was achieved by comparing the predicted gEUD to that of the actual PBT plan. Results: The predicted and planned gEUD were in good agreement. Considering all OARs, the prediction error was +1.4 ± 5.1 Gy (mean ± standard deviation), and Pearson's correlation coefficient was 93%. By comparing with an intensity modulated photon treatment plan, the model could classify whether an OAR structure would experience a gain, with a sensitivity of 93% (95% confidence interval: 87%-97%) and specificity of 63% (95% confidence interval: 38%-84%). Conclusions: We trained and validated models that could quickly and accurately predict the patient-specific benefits of PBT for skull-base tumors. Similar models could be developed for other tumor sites. Such models will be useful when an estimation of the feasible benefits of PBT is desired but the experience and/or resources required for treatment planning are unavailable.« less
Predicting Homework Effort: Support for a Domain-Specific, Multilevel Homework Model
ERIC Educational Resources Information Center
Trautwein, Ulrich; Ludtke, Oliver; Schnyder, Inge; Niggli, Alois
2006-01-01
According to the domain-specific, multilevel homework model proposed in the present study, students' homework effort is influenced by expectancy and value beliefs, homework characteristics, parental homework behavior, and conscientiousness. The authors used structural equation modeling and hierarchical linear modeling analyses to test the model in…
Logistic Mixed Models to Investigate Implicit and Explicit Belief Tracking.
Lages, Martin; Scheel, Anne
2016-01-01
We investigated the proposition of a two-systems Theory of Mind in adults' belief tracking. A sample of N = 45 participants predicted the choice of one of two opponent players after observing several rounds in an animated card game. Three matches of this card game were played and initial gaze direction on target and subsequent choice predictions were recorded for each belief task and participant. We conducted logistic regressions with mixed effects on the binary data and developed Bayesian logistic mixed models to infer implicit and explicit mentalizing in true belief and false belief tasks. Although logistic regressions with mixed effects predicted the data well a Bayesian logistic mixed model with latent task- and subject-specific parameters gave a better account of the data. As expected explicit choice predictions suggested a clear understanding of true and false beliefs (TB/FB). Surprisingly, however, model parameters for initial gaze direction also indicated belief tracking. We discuss why task-specific parameters for initial gaze directions are different from choice predictions yet reflect second-order perspective taking.
Prediction of frozen food properties during freezing using product composition.
Boonsupthip, W; Heldman, D R
2007-06-01
Frozen water fraction (FWF), as a function of temperature, is an important parameter for use in the design of food freezing processes. An FWF-prediction model, based on concentrations and molecular weights of specific product components, has been developed. Published food composition data were used to determine the identity and composition of key components. The model proposed in this investigation had been verified using published experimental FWF data and initial freezing temperature data, and by comparison to outputs from previously published models. It was found that specific food components with significant influence on freezing temperature depression of food products included low molecular weight water-soluble compounds with molality of 50 micromol per 100 g food or higher. Based on an analysis of 200 high-moisture food products, nearly 45% of the experimental initial freezing temperature data were within an absolute difference (AD) of +/- 0.15 degrees C and standard error (SE) of +/- 0.65 degrees C when compared to values predicted by the proposed model. The predicted relationship between temperature and FWF for all analyzed food products provided close agreements with experimental data (+/- 0.06 SE). The proposed model provided similar prediction capability for high- and intermediate-moisture food products. In addition, the proposed model provided statistically better prediction of initial freezing temperature and FWF than previous published models.
Modelling proteins’ hidden conformations to predict antibiotic resistance
Hart, Kathryn M.; Ho, Chris M. W.; Dutta, Supratik; Gross, Michael L.; Bowman, Gregory R.
2016-01-01
TEM β-lactamase confers bacteria with resistance to many antibiotics and rapidly evolves activity against new drugs. However, functional changes are not easily explained by differences in crystal structures. We employ Markov state models to identify hidden conformations and explore their role in determining TEM’s specificity. We integrate these models with existing drug-design tools to create a new technique, called Boltzmann docking, which better predicts TEM specificity by accounting for conformational heterogeneity. Using our MSMs, we identify hidden states whose populations correlate with activity against cefotaxime. To experimentally detect our predicted hidden states, we use rapid mass spectrometric footprinting and confirm our models’ prediction that increased cefotaxime activity correlates with reduced Ω-loop flexibility. Finally, we design novel variants to stabilize the hidden cefotaximase states, and find their populations predict activity against cefotaxime in vitro and in vivo. Therefore, we expect this framework to have numerous applications in drug and protein design. PMID:27708258
Li, Mao; Miller, Karol; Joldes, Grand Roman; Kikinis, Ron; Wittek, Adam
2016-12-01
Patient-specific biomechanical models have been advocated as a tool for predicting deformations of soft body organs/tissue for medical image registration (aligning two sets of images) when differences between the images are large. However, complex and irregular geometry of the body organs makes generation of patient-specific biomechanical models very time-consuming. Meshless discretisation has been proposed to solve this challenge. However, applications so far have been limited to 2D models and computing single organ deformations. In this study, 3D comprehensive patient-specific nonlinear biomechanical models implemented using meshless Total Lagrangian explicit dynamics algorithms are applied to predict a 3D deformation field for whole-body image registration. Unlike a conventional approach that requires dividing (segmenting) the image into non-overlapping constituents representing different organs/tissues, the mechanical properties are assigned using the fuzzy c-means algorithm without the image segmentation. Verification indicates that the deformations predicted using the proposed meshless approach are for practical purposes the same as those obtained using the previously validated finite element models. To quantitatively evaluate the accuracy of the predicted deformations, we determined the spatial misalignment between the registered (i.e. source images warped using the predicted deformations) and target images by computing the edge-based Hausdorff distance. The Hausdorff distance-based evaluation determines that our meshless models led to successful registration of the vast majority of the image features. Copyright © 2016 John Wiley & Sons, Ltd. Copyright © 2016 John Wiley & Sons, Ltd.
Muscle Synergies May Improve Optimization Prediction of Knee Contact Forces During Walking
Walter, Jonathan P.; Kinney, Allison L.; Banks, Scott A.; D'Lima, Darryl D.; Besier, Thor F.; Lloyd, David G.; Fregly, Benjamin J.
2014-01-01
The ability to predict patient-specific joint contact and muscle forces accurately could improve the treatment of walking-related disorders. Muscle synergy analysis, which decomposes a large number of muscle electromyographic (EMG) signals into a small number of synergy control signals, could reduce the dimensionality and thus redundancy of the muscle and contact force prediction process. This study investigated whether use of subject-specific synergy controls can improve optimization prediction of knee contact forces during walking. To generate the predictions, we performed mixed dynamic muscle force optimizations (i.e., inverse skeletal dynamics with forward muscle activation and contraction dynamics) using data collected from a subject implanted with a force-measuring knee replacement. Twelve optimization problems (three cases with four subcases each) that minimized the sum of squares of muscle excitations were formulated to investigate how synergy controls affect knee contact force predictions. The three cases were: (1) Calibrate+Match where muscle model parameter values were calibrated and experimental knee contact forces were simultaneously matched, (2) Precalibrate+Predict where experimental knee contact forces were predicted using precalibrated muscle model parameters values from the first case, and (3) Calibrate+Predict where muscle model parameter values were calibrated and experimental knee contact forces were simultaneously predicted, all while matching inverse dynamic loads at the hip, knee, and ankle. The four subcases used either 44 independent controls or five synergy controls with and without EMG shape tracking. For the Calibrate+Match case, all four subcases closely reproduced the measured medial and lateral knee contact forces (R2 ≥ 0.94, root-mean-square (RMS) error < 66 N), indicating sufficient model fidelity for contact force prediction. For the Precalibrate+Predict and Calibrate+Predict cases, synergy controls yielded better contact force predictions (0.61 < R2 < 0.90, 83 N < RMS error < 161 N) than did independent controls (-0.15 < R2 < 0.79, 124 N < RMS error < 343 N) for corresponding subcases. For independent controls, contact force predictions improved when precalibrated model parameter values or EMG shape tracking was used. For synergy controls, contact force predictions were relatively insensitive to how model parameter values were calibrated, while EMG shape tracking made lateral (but not medial) contact force predictions worse. For the subject and optimization cost function analyzed in this study, use of subject-specific synergy controls improved the accuracy of knee contact force predictions, especially for lateral contact force when EMG shape tracking was omitted, and reduced prediction sensitivity to uncertainties in muscle model parameter values. PMID:24402438
Muscle synergies may improve optimization prediction of knee contact forces during walking.
Walter, Jonathan P; Kinney, Allison L; Banks, Scott A; D'Lima, Darryl D; Besier, Thor F; Lloyd, David G; Fregly, Benjamin J
2014-02-01
The ability to predict patient-specific joint contact and muscle forces accurately could improve the treatment of walking-related disorders. Muscle synergy analysis, which decomposes a large number of muscle electromyographic (EMG) signals into a small number of synergy control signals, could reduce the dimensionality and thus redundancy of the muscle and contact force prediction process. This study investigated whether use of subject-specific synergy controls can improve optimization prediction of knee contact forces during walking. To generate the predictions, we performed mixed dynamic muscle force optimizations (i.e., inverse skeletal dynamics with forward muscle activation and contraction dynamics) using data collected from a subject implanted with a force-measuring knee replacement. Twelve optimization problems (three cases with four subcases each) that minimized the sum of squares of muscle excitations were formulated to investigate how synergy controls affect knee contact force predictions. The three cases were: (1) Calibrate+Match where muscle model parameter values were calibrated and experimental knee contact forces were simultaneously matched, (2) Precalibrate+Predict where experimental knee contact forces were predicted using precalibrated muscle model parameters values from the first case, and (3) Calibrate+Predict where muscle model parameter values were calibrated and experimental knee contact forces were simultaneously predicted, all while matching inverse dynamic loads at the hip, knee, and ankle. The four subcases used either 44 independent controls or five synergy controls with and without EMG shape tracking. For the Calibrate+Match case, all four subcases closely reproduced the measured medial and lateral knee contact forces (R2 ≥ 0.94, root-mean-square (RMS) error < 66 N), indicating sufficient model fidelity for contact force prediction. For the Precalibrate+Predict and Calibrate+Predict cases, synergy controls yielded better contact force predictions (0.61 < R2 < 0.90, 83 N < RMS error < 161 N) than did independent controls (-0.15 < R2 < 0.79, 124 N < RMS error < 343 N) for corresponding subcases. For independent controls, contact force predictions improved when precalibrated model parameter values or EMG shape tracking was used. For synergy controls, contact force predictions were relatively insensitive to how model parameter values were calibrated, while EMG shape tracking made lateral (but not medial) contact force predictions worse. For the subject and optimization cost function analyzed in this study, use of subject-specific synergy controls improved the accuracy of knee contact force predictions, especially for lateral contact force when EMG shape tracking was omitted, and reduced prediction sensitivity to uncertainties in muscle model parameter values.
Prediction of TF target sites based on atomistic models of protein-DNA complexes
Angarica, Vladimir Espinosa; Pérez, Abel González; Vasconcelos, Ana T; Collado-Vides, Julio; Contreras-Moreira, Bruno
2008-01-01
Background The specific recognition of genomic cis-regulatory elements by transcription factors (TFs) plays an essential role in the regulation of coordinated gene expression. Studying the mechanisms determining binding specificity in protein-DNA interactions is thus an important goal. Most current approaches for modeling TF specific recognition rely on the knowledge of large sets of cognate target sites and consider only the information contained in their primary sequence. Results Here we describe a structure-based methodology for predicting sequence motifs starting from the coordinates of a TF-DNA complex. Our algorithm combines information regarding the direct and indirect readout of DNA into an atomistic statistical model, which is used to estimate the interaction potential. We first measure the ability of our method to correctly estimate the binding specificities of eight prokaryotic and eukaryotic TFs that belong to different structural superfamilies. Secondly, the method is applied to two homology models, finding that sampling of interface side-chain rotamers remarkably improves the results. Thirdly, the algorithm is compared with a reference structural method based on contact counts, obtaining comparable predictions for the experimental complexes and more accurate sequence motifs for the homology models. Conclusion Our results demonstrate that atomic-detail structural information can be feasibly used to predict TF binding sites. The computational method presented here is universal and might be applied to other systems involving protein-DNA recognition. PMID:18922190
New Paradigm for Translational Modeling to Predict Long‐term Tuberculosis Treatment Response
Bartelink, IH; Zhang, N; Keizer, RJ; Strydom, N; Converse, PJ; Dooley, KE; Nuermberger, EL
2017-01-01
Abstract Disappointing results of recent tuberculosis chemotherapy trials suggest that knowledge gained from preclinical investigations was not utilized to maximal effect. A mouse‐to‐human translational pharmacokinetics (PKs) – pharmacodynamics (PDs) model built on a rich mouse database may improve clinical trial outcome predictions. The model included Mycobacterium tuberculosis growth function in mice, adaptive immune response effect on bacterial growth, relationships among moxifloxacin, rifapentine, and rifampin concentrations accelerating bacterial death, clinical PK data, species‐specific protein binding, drug‐drug interactions, and patient‐specific pathology. Simulations of recent trials testing 4‐month regimens predicted 65% (95% confidence interval [CI], 55–74) relapse‐free patients vs. 80% observed in the REMox‐TB trial, and 79% (95% CI, 72–87) vs. 82% observed in the Rifaquin trial. Simulation of 6‐month regimens predicted 97% (95% CI, 93–99) vs. 92% and 95% observed in 2RHZE/4RH control arms, and 100% predicted and observed in the 35 mg/kg rifampin arm of PanACEA MAMS. These results suggest that the model can inform regimen optimization and predict outcomes of ongoing trials. PMID:28561946
A PROBABILISTIC POPULATION EXPOSURE MODEL FOR PM10 AND PM 2.5
A first generation probabilistic population exposure model for Particulate Matter (PM), specifically for predicting PM10, and PM2.5, exposures of an urban, population has been developed. This model is intended to be used to predict exposure (magnitude, frequency, and duration) ...
Volkova, Svitlana; Ayton, Ellyn; Porterfield, Katherine; ...
2017-12-15
This work is the first to take advantage of recurrent neural networks to predict influenza-like-illness (ILI) dynamics from various linguistic signals extracted from social media data. Unlike other approaches that rely on timeseries analysis of historical ILI data [1, 2] and the state-of-the-art machine learning models [3, 4], we build and evaluate the predictive power of Long Short Term Memory (LSTMs) architectures capable of nowcasting (predicting in \\real-time") and forecasting (predicting the future) ILI dynamics in the 2011 { 2014 influenza seasons. To build our models we integrate information people post in social media e.g., topics, stylistic and syntactic patterns,more » emotions and opinions, and communication behavior. We then quantitatively evaluate the predictive power of different social media signals and contrast the performance of the-state-of-the-art regression models with neural networks. Finally, we combine ILI and social media signals to build joint neural network models for ILI dynamics prediction. Unlike the majority of the existing work, we specifically focus on developing models for local rather than national ILI surveillance [1], specifically for military rather than general populations [3] in 26 U.S. and six international locations. Our approach demonstrates several advantages: (a) Neural network models learned from social media data yield the best performance compared to previously used regression models. (b) Previously under-explored language and communication behavior features are more predictive of ILI dynamics than syntactic and stylistic signals expressed in social media. (c) Neural network models learned exclusively from social media signals yield comparable or better performance to the models learned from ILI historical data, thus, signals from social media can be potentially used to accurately forecast ILI dynamics for the regions where ILI historical data is not available. (d) Neural network models learned from combined ILI and social media signals significantly outperform models that rely solely on ILI historical data, which adds to a great potential of alternative public sources for ILI dynamics prediction. (e) Location-specific models outperform previously used location-independent models e.g., U.S. only. (f) Prediction results significantly vary across geolocations depending on the amount of social media data available and ILI activity patterns.« less
DOE Office of Scientific and Technical Information (OSTI.GOV)
Volkova, Svitlana; Ayton, Ellyn; Porterfield, Katherine
This work is the first to take advantage of recurrent neural networks to predict influenza-like-illness (ILI) dynamics from various linguistic signals extracted from social media data. Unlike other approaches that rely on timeseries analysis of historical ILI data [1, 2] and the state-of-the-art machine learning models [3, 4], we build and evaluate the predictive power of Long Short Term Memory (LSTMs) architectures capable of nowcasting (predicting in \\real-time") and forecasting (predicting the future) ILI dynamics in the 2011 { 2014 influenza seasons. To build our models we integrate information people post in social media e.g., topics, stylistic and syntactic patterns,more » emotions and opinions, and communication behavior. We then quantitatively evaluate the predictive power of different social media signals and contrast the performance of the-state-of-the-art regression models with neural networks. Finally, we combine ILI and social media signals to build joint neural network models for ILI dynamics prediction. Unlike the majority of the existing work, we specifically focus on developing models for local rather than national ILI surveillance [1], specifically for military rather than general populations [3] in 26 U.S. and six international locations. Our approach demonstrates several advantages: (a) Neural network models learned from social media data yield the best performance compared to previously used regression models. (b) Previously under-explored language and communication behavior features are more predictive of ILI dynamics than syntactic and stylistic signals expressed in social media. (c) Neural network models learned exclusively from social media signals yield comparable or better performance to the models learned from ILI historical data, thus, signals from social media can be potentially used to accurately forecast ILI dynamics for the regions where ILI historical data is not available. (d) Neural network models learned from combined ILI and social media signals significantly outperform models that rely solely on ILI historical data, which adds to a great potential of alternative public sources for ILI dynamics prediction. (e) Location-specific models outperform previously used location-independent models e.g., U.S. only. (f) Prediction results significantly vary across geolocations depending on the amount of social media data available and ILI activity patterns.« less
A predictive framework for evaluating models of semantic organization in free recall
Morton, Neal W; Polyn, Sean M.
2016-01-01
Research in free recall has demonstrated that semantic associations reliably influence the organization of search through episodic memory. However, the specific structure of these associations and the mechanisms by which they influence memory search remain unclear. We introduce a likelihood-based model-comparison technique, which embeds a model of semantic structure within the context maintenance and retrieval (CMR) model of human memory search. Within this framework, model variants are evaluated in terms of their ability to predict the specific sequence in which items are recalled. We compare three models of semantic structure, latent semantic analysis (LSA), global vectors (GloVe), and word association spaces (WAS), and find that models using WAS have the greatest predictive power. Furthermore, we find evidence that semantic and temporal organization is driven by distinct item and context cues, rather than a single context cue. This finding provides important constraint for theories of memory search. PMID:28331243
Nøst, Therese Haugdahl; Breivik, Knut; Wania, Frank; Rylander, Charlotta; Odland, Jon Øyvind; Sandanger, Torkjel Manning
2016-03-01
Studies on the health effects of polychlorinated biphenyls (PCBs) call for an understanding of past and present human exposure. Time-resolved mechanistic models may supplement information on concentrations in individuals obtained from measurements and/or statistical approaches if they can be shown to reproduce empirical data. Here, we evaluated the capability of one such mechanistic model to reproduce measured PCB concentrations in individual Norwegian women. We also assessed individual life-course concentrations. Concentrations of four PCB congeners in pregnant (n = 310, sampled in 2007-2009) and postmenopausal (n = 244, 2005) women were compared with person-specific predictions obtained using CoZMoMAN, an emission-based environmental fate and human food-chain bioaccumulation model. Person-specific predictions were also made using statistical regression models including dietary and lifestyle variables and concentrations. CoZMoMAN accurately reproduced medians and ranges of measured concentrations in the two study groups. Furthermore, rank correlations between measurements and predictions from both CoZMoMAN and regression analyses were strong (Spearman's r > 0.67). Precision in quartile assignments from predictions was strong overall as evaluated by weighted Cohen's kappa (> 0.6). Simulations indicated large inter-individual differences in concentrations experienced in the past. The mechanistic model reproduced all measurements of PCB concentrations within a factor of 10, and subject ranking and quartile assignments were overall largely consistent, although they were weak within each study group. Contamination histories for individuals predicted by CoZMoMAN revealed variation between study subjects, particularly in the timing of peak concentrations. Mechanistic models can provide individual PCB exposure metrics that could serve as valuable supplements to measurements.
Prediction of high incidence of dengue in the Philippines.
Buczak, Anna L; Baugher, Benjamin; Babin, Steven M; Ramac-Thomas, Liane C; Guven, Erhan; Elbert, Yevgeniy; Koshute, Phillip T; Velasco, John Mark S; Roque, Vito G; Tayag, Enrique A; Yoon, In-Kyu; Lewis, Sheri H
2014-04-01
Accurate prediction of dengue incidence levels weeks in advance of an outbreak may reduce the morbidity and mortality associated with this neglected disease. Therefore, models were developed to predict high and low dengue incidence in order to provide timely forewarnings in the Philippines. Model inputs were chosen based on studies indicating variables that may impact dengue incidence. The method first uses Fuzzy Association Rule Mining techniques to extract association rules from these historical epidemiological, environmental, and socio-economic data, as well as climate data indicating future weather patterns. Selection criteria were used to choose a subset of these rules for a classifier, thereby generating a Prediction Model. The models predicted high or low incidence of dengue in a Philippines province four weeks in advance. The threshold between high and low was determined relative to historical incidence data. Model accuracy is described by Positive Predictive Value (PPV), Negative Predictive Value (NPV), Sensitivity, and Specificity computed on test data not previously used to develop the model. Selecting a model using the F0.5 measure, which gives PPV more importance than Sensitivity, gave these results: PPV = 0.780, NPV = 0.938, Sensitivity = 0.547, Specificity = 0.978. Using the F3 measure, which gives Sensitivity more importance than PPV, the selected model had PPV = 0.778, NPV = 0.948, Sensitivity = 0.627, Specificity = 0.974. The decision as to which model has greater utility depends on how the predictions will be used in a particular situation. This method builds prediction models for future dengue incidence in the Philippines and is capable of being modified for use in different situations; for diseases other than dengue; and for regions beyond the Philippines. The Philippines dengue prediction models predicted high or low incidence of dengue four weeks in advance of an outbreak with high accuracy, as measured by PPV, NPV, Sensitivity, and Specificity.
Prediction of High Incidence of Dengue in the Philippines
Buczak, Anna L.; Baugher, Benjamin; Babin, Steven M.; Ramac-Thomas, Liane C.; Guven, Erhan; Elbert, Yevgeniy; Koshute, Phillip T.; Velasco, John Mark S.; Roque, Vito G.; Tayag, Enrique A.; Yoon, In-Kyu; Lewis, Sheri H.
2014-01-01
Background Accurate prediction of dengue incidence levels weeks in advance of an outbreak may reduce the morbidity and mortality associated with this neglected disease. Therefore, models were developed to predict high and low dengue incidence in order to provide timely forewarnings in the Philippines. Methods Model inputs were chosen based on studies indicating variables that may impact dengue incidence. The method first uses Fuzzy Association Rule Mining techniques to extract association rules from these historical epidemiological, environmental, and socio-economic data, as well as climate data indicating future weather patterns. Selection criteria were used to choose a subset of these rules for a classifier, thereby generating a Prediction Model. The models predicted high or low incidence of dengue in a Philippines province four weeks in advance. The threshold between high and low was determined relative to historical incidence data. Principal Findings Model accuracy is described by Positive Predictive Value (PPV), Negative Predictive Value (NPV), Sensitivity, and Specificity computed on test data not previously used to develop the model. Selecting a model using the F0.5 measure, which gives PPV more importance than Sensitivity, gave these results: PPV = 0.780, NPV = 0.938, Sensitivity = 0.547, Specificity = 0.978. Using the F3 measure, which gives Sensitivity more importance than PPV, the selected model had PPV = 0.778, NPV = 0.948, Sensitivity = 0.627, Specificity = 0.974. The decision as to which model has greater utility depends on how the predictions will be used in a particular situation. Conclusions This method builds prediction models for future dengue incidence in the Philippines and is capable of being modified for use in different situations; for diseases other than dengue; and for regions beyond the Philippines. The Philippines dengue prediction models predicted high or low incidence of dengue four weeks in advance of an outbreak with high accuracy, as measured by PPV, NPV, Sensitivity, and Specificity. PMID:24722434
NASA Astrophysics Data System (ADS)
Jarrett, Angela M.; Hormuth, David A.; Barnes, Stephanie L.; Feng, Xinzeng; Huang, Wei; Yankeelov, Thomas E.
2018-05-01
Clinical methods for assessing tumor response to therapy are largely rudimentary, monitoring only temporal changes in tumor size. Our goal is to predict the response of breast tumors to therapy using a mathematical model that utilizes magnetic resonance imaging (MRI) data obtained non-invasively from individual patients. We extended a previously established, mechanically coupled, reaction-diffusion model for predicting tumor response initialized with patient-specific diffusion weighted MRI (DW-MRI) data by including the effects of chemotherapy drug delivery, which is estimated using dynamic contrast-enhanced (DCE-) MRI data. The extended, drug incorporated, model is initialized using patient-specific DW-MRI and DCE-MRI data. Data sets from five breast cancer patients were used—obtained before, after one cycle, and at mid-point of neoadjuvant chemotherapy. The DCE-MRI data was used to estimate spatiotemporal variations in tumor perfusion with the extended Kety–Tofts model. The physiological parameters derived from DCE-MRI were used to model changes in delivery of therapy drugs within the tumor for incorporation in the extended model. We simulated the original model and the extended model in both 2D and 3D and compare the results for this five-patient cohort. Preliminary results show reductions in the error of model predicted tumor cellularity and size compared to the experimentally-measured results for the third MRI scan when therapy was incorporated. Comparing the two models for agreement between the predicted total cellularity and the calculated total cellularity (from the DW-MRI data) reveals an increased concordance correlation coefficient from 0.81 to 0.98 for the 2D analysis and 0.85 to 0.99 for the 3D analysis (p < 0.01 for each) when the extended model was used in place of the original model. This study demonstrates the plausibility of using DCE-MRI data as a means to estimate drug delivery on a patient-specific basis in predictive models and represents a step toward the goal of achieving individualized prediction of tumor response to therapy.
A site specific model and analysis of the neutral somatic mutation rate in whole-genome cancer data.
Bertl, Johanna; Guo, Qianyun; Juul, Malene; Besenbacher, Søren; Nielsen, Morten Muhlig; Hornshøj, Henrik; Pedersen, Jakob Skou; Hobolth, Asger
2018-04-19
Detailed modelling of the neutral mutational process in cancer cells is crucial for identifying driver mutations and understanding the mutational mechanisms that act during cancer development. The neutral mutational process is very complex: whole-genome analyses have revealed that the mutation rate differs between cancer types, between patients and along the genome depending on the genetic and epigenetic context. Therefore, methods that predict the number of different types of mutations in regions or specific genomic elements must consider local genomic explanatory variables. A major drawback of most methods is the need to average the explanatory variables across the entire region or genomic element. This procedure is particularly problematic if the explanatory variable varies dramatically in the element under consideration. To take into account the fine scale of the explanatory variables, we model the probabilities of different types of mutations for each position in the genome by multinomial logistic regression. We analyse 505 cancer genomes from 14 different cancer types and compare the performance in predicting mutation rate for both regional based models and site-specific models. We show that for 1000 randomly selected genomic positions, the site-specific model predicts the mutation rate much better than regional based models. We use a forward selection procedure to identify the most important explanatory variables. The procedure identifies site-specific conservation (phyloP), replication timing, and expression level as the best predictors for the mutation rate. Finally, our model confirms and quantifies certain well-known mutational signatures. We find that our site-specific multinomial regression model outperforms the regional based models. The possibility of including genomic variables on different scales and patient specific variables makes it a versatile framework for studying different mutational mechanisms. Our model can serve as the neutral null model for the mutational process; regions that deviate from the null model are candidates for elements that drive cancer development.
Klement, William; Wilk, Szymon; Michalowski, Wojtek; Farion, Ken J; Osmond, Martin H; Verter, Vedat
2012-03-01
Using an automatic data-driven approach, this paper develops a prediction model that achieves more balanced performance (in terms of sensitivity and specificity) than the Canadian Assessment of Tomography for Childhood Head Injury (CATCH) rule, when predicting the need for computed tomography (CT) imaging of children after a minor head injury. CT is widely considered an effective tool for evaluating patients with minor head trauma who have potentially suffered serious intracranial injury. However, its use poses possible harmful effects, particularly for children, due to exposure to radiation. Safety concerns, along with issues of cost and practice variability, have led to calls for the development of effective methods to decide when CT imaging is needed. Clinical decision rules represent such methods and are normally derived from the analysis of large prospectively collected patient data sets. The CATCH rule was created by a group of Canadian pediatric emergency physicians to support the decision of referring children with minor head injury to CT imaging. The goal of the CATCH rule was to maximize the sensitivity of predictions of potential intracranial lesion while keeping specificity at a reasonable level. After extensive analysis of the CATCH data set, characterized by severe class imbalance, and after a thorough evaluation of several data mining methods, we derived an ensemble of multiple Naive Bayes classifiers as the prediction model for CT imaging decisions. In the first phase of the experiment we compared the proposed ensemble model to other ensemble models employing rule-, tree- and instance-based member classifiers. Our prediction model demonstrated the best performance in terms of AUC, G-mean and sensitivity measures. In the second phase, using a bootstrapping experiment similar to that reported by the CATCH investigators, we showed that the proposed ensemble model achieved a more balanced predictive performance than the CATCH rule with an average sensitivity of 82.8% and an average specificity of 74.4% (vs. 98.1% and 50.0% for the CATCH rule respectively). Automatically derived prediction models cannot replace a physician's acumen. However, they help establish reference performance indicators for the purpose of developing clinical decision rules so the trade-off between prediction sensitivity and specificity is better understood. Copyright © 2011 Elsevier B.V. All rights reserved.
Hara, Tomohiko; Nakanishi, Hiroyuki; Nakagawa, Tohru; Komiyama, Motokiyo; Kawahara, Takashi; Manabe, Tomoko; Miyake, Mototaka; Arai, Eri; Kanai, Yae; Fujimoto, Hiroyuki
2013-10-01
Recent studies have shown an improvement in prostate cancer diagnosis with the use of 3.0-Tesla magnetic resonance imaging. We retrospectively assessed the ability of this imaging technique to predict side-specific extracapsular extension of prostate cancer. From October 2007 to August 2011, prostatectomy was carried out in 396 patients after preoperative 3.0-Tesla magnetic resonance imaging. Among these, 132 (primary sample) and 134 patients (validation sample) underwent 12-core prostate biopsy at the National Cancer Center Hospital of Tokyo, Japan, and at other institutions, respectively. In the primary dataset, univariate and multivariate analyses were carried out to predict side-specific extracapsular extension using variables determined preoperatively, including 3.0-Tesla magnetic resonance imaging findings (T2-weighted and diffusion-weighted imaging). A prediction model was then constructed and applied to the validation study sample. Multivariate analysis identified four significant independent predictors (P < 0.05), including a biopsy Gleason score of ≥8, positive 3.0-Tesla diffusion-weighted magnetic resonance imaging findings, ≥2 positive biopsy cores on each side and a maximum percentage of positive cores ≥31% on each side. The negative predictive value was 93.9% in the combination model with these four predictors, meanwhile the positive predictive value was 33.8%. Good reproducibility of these four significant predictors and the combination model was observed in the validation study sample. The side-specific extracapsular extension prediction by the biopsy Gleason score and factors associated with tumor location, including a positive 3.0-Tesla diffusion-weighted magnetic resonance imaging finding, have a high negative predictive value, but a low positive predictive value. © 2013 The Japanese Urological Association.
Wen, Ping-Ping; Shi, Shao-Ping; Xu, Hao-Dong; Wang, Li-Na; Qiu, Jian-Ding
2016-10-15
As one of the most important reversible types of post-translational modification, protein methylation catalyzed by methyltransferases carries many pivotal biological functions as well as many essential biological processes. Identification of methylation sites is prerequisite for decoding methylation regulatory networks in living cells and understanding their physiological roles. Experimental methods are limitations of labor-intensive and time-consuming. While in silicon approaches are cost-effective and high-throughput manner to predict potential methylation sites, but those previous predictors only have a mixed model and their prediction performances are not fully satisfactory now. Recently, with increasing availability of quantitative methylation datasets in diverse species (especially in eukaryotes), there is a growing need to develop a species-specific predictor. Here, we designed a tool named PSSMe based on information gain (IG) feature optimization method for species-specific methylation site prediction. The IG method was adopted to analyze the importance and contribution of each feature, then select the valuable dimension feature vectors to reconstitute a new orderly feature, which was applied to build the finally prediction model. Finally, our method improves prediction performance of accuracy about 15% comparing with single features. Furthermore, our species-specific model significantly improves the predictive performance compare with other general methylation prediction tools. Hence, our prediction results serve as useful resources to elucidate the mechanism of arginine or lysine methylation and facilitate hypothesis-driven experimental design and validation. The tool online service is implemented by C# language and freely available at http://bioinfo.ncu.edu.cn/PSSMe.aspx CONTACT: jdqiu@ncu.edu.cnSupplementary information: Supplementary data are available at Bioinformatics online. © The Author 2016. Published by Oxford University Press. All rights reserved. For Permissions, please e-mail: journals.permissions@oup.com.
Jin, H; Wu, S; Vidyanti, I; Di Capua, P; Wu, B
2015-01-01
This article is part of the Focus Theme of Methods of Information in Medicine on "Big Data and Analytics in Healthcare". Depression is a common and often undiagnosed condition for patients with diabetes. It is also a condition that significantly impacts healthcare outcomes, use, and cost as well as elevating suicide risk. Therefore, a model to predict depression among diabetes patients is a promising and valuable tool for providers to proactively assess depressive symptoms and identify those with depression. This study seeks to develop a generalized multilevel regression model, using a longitudinal data set from a recent large-scale clinical trial, to predict depression severity and presence of major depression among patients with diabetes. Severity of depression was measured by the Patient Health Questionnaire PHQ-9 score. Predictors were selected from 29 candidate factors to develop a 2-level Poisson regression model that can make population-average predictions for all patients and subject-specific predictions for individual patients with historical records. Newly obtained patient records can be incorporated with historical records to update the prediction model. Root-mean-square errors (RMSE) were used to evaluate predictive accuracy of PHQ-9 scores. The study also evaluated the classification ability of using the predicted PHQ-9 scores to classify patients as having major depression. Two time-invariant and 10 time-varying predictors were selected for the model. Incorporating historical records and using them to update the model may improve both predictive accuracy of PHQ-9 scores and classification ability of the predicted scores. Subject-specific predictions (for individual patients with historical records) achieved RMSE about 4 and areas under the receiver operating characteristic (ROC) curve about 0.9 and are better than population-average predictions. The study developed a generalized multilevel regression model to predict depression and demonstrated that using generalized multilevel regression based on longitudinal patient records can achieve high predictive ability.
Mamoshina, Polina; Kochetov, Kirill; Putin, Evgeny; Cortese, Franco; Aliper, Alexander; Lee, Won-Suk; Ahn, Sung-Min; Uhn, Lee; Skjodt, Neil; Kovalchuk, Olga; Scheibye-Knudsen, Morten; Zhavoronkov, Alex
2018-01-11
Accurate and physiologically meaningful biomarkers for human aging are key to assessing anti-aging therapies. Given ethnic differences in health, diet, lifestyle, behaviour, environmental exposures and even average rate of biological aging, it stands to reason that aging clocks trained on datasets obtained from specific ethnic populations are more likely to account for these potential confounding factors, resulting in an enhanced capacity to predict chronological age and quantify biological age. Here we present a deep learning-based hematological aging clock modeled using the large combined dataset of Canadian, South Korean and Eastern European population blood samples that show increased predictive accuracy in individual populations compared to population-specific hematologic aging clocks. The performance of models was also evaluated on publicly-available samples of the American population from the National Health and Nutrition Examination Survey (NHANES). In addition, we explored the association between age predicted by both population-specific and combined hematological clocks and all-cause mortality. Overall, this study suggests a) the population-specificity of aging patterns and b) hematologic clocks predicts all-cause mortality. Proposed models added to the freely available Aging.AI system allowing improved ability to assess human aging. © The Author(s) 2018. Published by Oxford University Press on behalf of The Gerontological Society of America.
Diehm, Nicolas; Sin, Sangmun; Hoppe, Hanno; Baumgartner, Iris; Büchler, Philippe
2011-06-01
To assess if finite element (FE) models can be used to predict deformation of the femoropopliteal segment during knee flexion. Magnetic resonance angiography (MRA) images were acquired on the lower limbs of 8 healthy volunteers (5 men; mean age 28 ± 4 years). Images were taken in 2 natural positions, with the lower limb fully extended and with the knee bent at ~ 40°. Patient-specific FE models were developed and used to simulate the experimental situation. The displacements of the artery during knee bending as predicted by the numerical model were compared to the corresponding positions measured on the MRA images. The numerical predictions showed a good overall agreement between the calculated displacements of the motion measures from MRA images. The average position error comparing the calculated vs. actual displacements of the femoropopliteal intersection measured on the MRA was 8 ± 4 mm. Two of the 8 subjects showed large prediction errors (average 13 ± 5 mm); these 2 volunteers were the tallest subjects involved in the study and had a low body mass index (20.5 kg/m²). The present computational model is able to capture the gross mechanical environment of the femoropopliteal intersection during knee bending and provide a better understanding of the complex biomechanical behavior. However, results suggest that patient-specific mechanical properties and detailed muscle modeling are required to provide accurate patient-specific numerical predictions of arterial displacement. Further adaptation of this model is expected to provide an improved ability to predict the multiaxial deformation of this arterial segment during leg movements and to optimize future stent designs.
Nguyen, Van-Nui; Huang, Kai-Yao; Huang, Chien-Hsun; Chang, Tzu-Hao; Bretaña, Neil; Lai, K; Weng, Julia; Lee, Tzong-Yi
2015-01-01
In eukaryotes, ubiquitin-conjugation is an important mechanism underlying proteasome-mediated degradation of proteins, and as such, plays an essential role in the regulation of many cellular processes. In the ubiquitin-proteasome pathway, E3 ligases play important roles by recognizing a specific protein substrate and catalyzing the attachment of ubiquitin to a lysine (K) residue. As more and more experimental data on ubiquitin conjugation sites become available, it becomes possible to develop prediction models that can be scaled to big data. However, no development that focuses on the investigation of ubiquitinated substrate specificities has existed. Herein, we present an approach that exploits an iteratively statistical method to identify ubiquitin conjugation sites with substrate site specificities. In this investigation, totally 6259 experimentally validated ubiquitinated proteins were obtained from dbPTM. After having filtered out homologous fragments with 40% sequence identity, the training data set contained 2658 ubiquitination sites (positive data) and 5532 non-ubiquitinated sites (negative data). Due to the difficulty in characterizing the substrate site specificities of E3 ligases by conventional sequence logo analysis, a recursively statistical method has been applied to obtain significant conserved motifs. The profile hidden Markov model (profile HMM) was adopted to construct the predictive models learned from the identified substrate motifs. A five-fold cross validation was then used to evaluate the predictive model, achieving sensitivity, specificity, and accuracy of 73.07%, 65.46%, and 67.93%, respectively. Additionally, an independent testing set, completely blind to the training data of the predictive model, was used to demonstrate that the proposed method could provide a promising accuracy (76.13%) and outperform other ubiquitination site prediction tool. A case study demonstrated the effectiveness of the characterized substrate motifs for identifying ubiquitination sites. The proposed method presents a practical means of preliminary analysis and greatly diminishes the total number of potential targets required for further experimental confirmation. This method may help unravel their mechanisms and roles in E3 recognition and ubiquitin-mediated protein degradation.
ProTSAV: A protein tertiary structure analysis and validation server.
Singh, Ankita; Kaushik, Rahul; Mishra, Avinash; Shanker, Asheesh; Jayaram, B
2016-01-01
Quality assessment of predicted model structures of proteins is as important as the protein tertiary structure prediction. A highly efficient quality assessment of predicted model structures directs further research on function. Here we present a new server ProTSAV, capable of evaluating predicted model structures based on some popular online servers and standalone tools. ProTSAV furnishes the user with a single quality score in case of individual protein structure along with a graphical representation and ranking in case of multiple protein structure assessment. The server is validated on ~64,446 protein structures including experimental structures from RCSB and predicted model structures for CASP targets and from public decoy sets. ProTSAV succeeds in predicting quality of protein structures with a specificity of 100% and a sensitivity of 98% on experimentally solved structures and achieves a specificity of 88%and a sensitivity of 91% on predicted protein structures of CASP11 targets under 2Å.The server overcomes the limitations of any single server/method and is seen to be robust in helping in quality assessment. ProTSAV is freely available at http://www.scfbio-iitd.res.in/software/proteomics/protsav.jsp. Copyright © 2015 Elsevier B.V. All rights reserved.
Candido Dos Reis, Francisco J; Wishart, Gordon C; Dicks, Ed M; Greenberg, David; Rashbass, Jem; Schmidt, Marjanka K; van den Broek, Alexandra J; Ellis, Ian O; Green, Andrew; Rakha, Emad; Maishman, Tom; Eccles, Diana M; Pharoah, Paul D P
2017-05-22
PREDICT is a breast cancer prognostic and treatment benefit model implemented online. The overall fit of the model has been good in multiple independent case series, but PREDICT has been shown to underestimate breast cancer specific mortality in women diagnosed under the age of 40. Another limitation is the use of discrete categories for tumour size and node status resulting in 'step' changes in risk estimates on moving between categories. We have refitted the PREDICT prognostic model using the original cohort of cases from East Anglia with updated survival time in order to take into account age at diagnosis and to smooth out the survival function for tumour size and node status. Multivariable Cox regression models were used to fit separate models for ER negative and ER positive disease. Continuous variables were fitted using fractional polynomials and a smoothed baseline hazard was obtained by regressing the baseline cumulative hazard for each patients against time using fractional polynomials. The fit of the prognostic models were then tested in three independent data sets that had also been used to validate the original version of PREDICT. In the model fitting data, after adjusting for other prognostic variables, there is an increase in risk of breast cancer specific mortality in younger and older patients with ER positive disease, with a substantial increase in risk for women diagnosed before the age of 35. In ER negative disease the risk increases slightly with age. The association between breast cancer specific mortality and both tumour size and number of positive nodes was non-linear with a more marked increase in risk with increasing size and increasing number of nodes in ER positive disease. The overall calibration and discrimination of the new version of PREDICT (v2) was good and comparable to that of the previous version in both model development and validation data sets. However, the calibration of v2 improved over v1 in patients diagnosed under the age of 40. The PREDICT v2 is an improved prognostication and treatment benefit model compared with v1. The online version should continue to aid clinical decision making in women with early breast cancer.
Collins, Gary S; Reitsma, Johannes B; Altman, Douglas G; Moons, Karel G M
2015-06-01
Prediction models are developed to aid health care providers in estimating the probability or risk that a specific disease or condition is present (diagnostic models) or that a specific event will occur in the future (prognostic models), to inform their decision making. However, the overwhelming evidence shows that the quality of reporting of prediction model studies is poor. Only with full and clear reporting of information on all aspects of a prediction model can risk of bias and potential usefulness of prediction models be adequately assessed. The Transparent Reporting of a multivariable prediction model for Individual Prognosis Or Diagnosis (TRIPOD) Initiative developed a set of recommendations for the reporting of studies developing, validating, or updating a prediction model, whether for diagnostic or prognostic purposes. This article describes how the TRIPOD Statement was developed. An extensive list of items based on a review of the literature was created, which was reduced after a Web-based survey and revised during a 3-day meeting in June 2011 with methodologists, health care professionals, and journal editors. The list was refined during several meetings of the steering group and in e-mail discussions with the wider group of TRIPOD contributors. The resulting TRIPOD Statement is a checklist of 22 items, deemed essential for transparent reporting of a prediction model study. The TRIPOD Statement aims to improve the transparency of the reporting of a prediction model study regardless of the study methods used. The TRIPOD Statement is best used in conjunction with the TRIPOD explanation and elaboration document. To aid the editorial process and readers of prediction model studies, it is recommended that authors include a completed checklist in their submission (also available at www.tripod-statement.org). The Transparent Reporting of a multivariable prediction model for Individual Prognosis Or Diagnosis (TRIPOD) Initiative developed a set of recommendations for the reporting of studies developing, validating, or updating a prediction model, whether for diagnostic or prognostic purposes. Copyright © 2014 The Authors. Published by Elsevier B.V. All rights reserved.
A microscale emission factor model (MicroFacPM) for predicting real-time site-specific motor vehicle particulate matter emissions was presented in the companion paper entitled "Development of a Microscale Emission Factor Model for Particulate Matter (MicroFacPM) for Predicting Re...
Aircraft measurements made downwind from specific coal fired power plants during the 2013 Southeast Nexus field campaign provide a unique opportunity to evaluate single source photochemical model predictions of both O3 and secondary PM2.5 species. The model did well at predicting...
Collins, G S; Reitsma, J B; Altman, D G; Moons, K G M
2015-02-01
Prediction models are developed to aid healthcare providers in estimating the probability or risk that a specific disease or condition is present (diagnostic models) or that a specific event will occur in the future (prognostic models), to inform their decision-making. However, the overwhelming evidence shows that the quality of reporting of prediction model studies is poor. Only with full and clear reporting of information on all aspects of a prediction model can risk of bias and potential usefulness of prediction models be adequately assessed. The Transparent Reporting of a multivariable prediction model for Individual Prognosis or Diagnosis (TRIPOD) initiative developed a set of recommendations for the reporting of studies developing, validating or updating a prediction model, whether for diagnostic or prognostic purposes. This article describes how the TRIPOD Statement was developed. An extensive list of items based on a review of the literature was created, which was reduced after a web-based survey and revised during a 3-day meeting in June 2011 with methodologists, healthcare professionals and journal editors. The list was refined during several meetings of the steering group and in e-mail discussions with the wider group of TRIPOD contributors. The resulting TRIPOD Statement is a checklist of 22 items, deemed essential for transparent reporting of a prediction model study. The TRIPOD Statement aims to improve the transparency of the reporting of a prediction model study, regardless of the study methods used. The TRIPOD Statement is best used in conjunction with the TRIPOD explanation and elaboration document. To aid the editorial process and readers of prediction model studies, it is recommended that authors include a completed checklist in their submission (also available at www.tripod-statement.org). © 2015 Joint copyright. The Authors and Annals of Internal Medicine. Diabetic Medicine published by John Wiley Ltd. on behalf of Diabetes UK.
NASA Astrophysics Data System (ADS)
Tan, C. H.; Matjafri, M. Z.; Lim, H. S.
2015-10-01
This paper presents the prediction models which analyze and compute the CO2 emission in Malaysia. Each prediction model for CO2 emission will be analyzed based on three main groups which is transportation, electricity and heat production as well as residential buildings and commercial and public services. The prediction models were generated using data obtained from World Bank Open Data. Best subset method will be used to remove irrelevant data and followed by multi linear regression to produce the prediction models. From the results, high R-square (prediction) value was obtained and this implies that the models are reliable to predict the CO2 emission by using specific data. In addition, the CO2 emissions from these three groups are forecasted using trend analysis plots for observation purpose.
Food allergy animal models: an overview.
Helm, Ricki M
2002-05-01
Specific food allergy is characterized by sensitization to innocuous food proteins with production of allergen-specific IgE that binds to receptors on basophils and mast cells. Upon recurrent exposure to the same allergen, an allergic response is induced by mediator release following cross-linking of cell-bound allergen-specific IgE. The determination of what makes an innocuous food protein an allergen in predisposed individuals is unknown; however, mechanistic and protein allergen predictive models are being actively investigated in a number of animal models. Currently, there is no animal model that will actively profile known food allergens, predict the allergic potential of novel food proteins, or demonstrate clinically the human food allergic sensitization/allergic response. Animal models under investigation include mice, rats, the guinea pig, atopic dog, and neonatal swine. These models are being assessed for production of IgE, clinical responses to re-exposure, and a ranking of food allergens (based on potency) including a nonfood allergen protein source. A selection of animal models actively being investigated that will contribute to our understanding of what makes a protein an allergen and future predictive models for assessing the allergenicity of novel proteins is presented in this review.
A PBPK model for TCE with specificity for the male LE rat that accurately predicts TCE tissue time-course data has not been developed, although other PBPK models for TCE exist. Development of such a model was the present aim. The PBPK model consisted of 5 compartments: fat; slowl...
Lee II, Henry; Reusser, Deborah A.; Frazier, Melanie R; McCoy, Lee M; Clinton, Patrick J.; Clough, Jonathan S.
2014-01-01
The “Sea‐Level Affecting Marshes Model” (SLAMM) is a moderate resolution model used to predict the effects of sea level rise on marsh habitats (Craft et al. 2009). SLAMM has been used extensively on both the west coast (e.g., Glick et al., 2007) and east coast (e.g., Geselbracht et al., 2011) of the United States to evaluate potential changes in the distribution and extent of tidal marsh habitats. However, a limitation of the current version of SLAMM, (Version 6.2) is that it lacks the ability to model distribution changes in seagrass habitat resulting from sea level rise. Because of the ecological importance of SAV habitats, U.S. EPA, USGS, and USDA partnered with Warren Pinnacle Consulting to enhance the SLAMM modeling software to include new functionality in order to predict changes in Zostera marina distribution within Pacific Northwest estuaries in response to sea level rise. Specifically, the objective was to develop a SAV model that used generally available GIS data and parameters that were predictive and that could be customized for other estuaries that have GIS layers of existing SAV distribution. This report describes the procedure used to develop the SAV model for the Yaquina Bay Estuary, Oregon, appends a statistical script based on the open source R software to generate a similar SAV model for other estuaries that have data layers of existing SAV, and describes how to incorporate the model coefficients from the site‐specific SAV model into SLAMM to predict the effects of sea level rise on Zostera marina distributions. To demonstrate the applicability of the R tools, we utilize them to develop model coefficients for Willapa Bay, Washington using site‐specific SAV data.
Microarray-based cancer prediction using soft computing approach.
Wang, Xiaosheng; Gotoh, Osamu
2009-05-26
One of the difficulties in using gene expression profiles to predict cancer is how to effectively select a few informative genes to construct accurate prediction models from thousands or ten thousands of genes. We screen highly discriminative genes and gene pairs to create simple prediction models involved in single genes or gene pairs on the basis of soft computing approach and rough set theory. Accurate cancerous prediction is obtained when we apply the simple prediction models for four cancerous gene expression datasets: CNS tumor, colon tumor, lung cancer and DLBCL. Some genes closely correlated with the pathogenesis of specific or general cancers are identified. In contrast with other models, our models are simple, effective and robust. Meanwhile, our models are interpretable for they are based on decision rules. Our results demonstrate that very simple models may perform well on cancerous molecular prediction and important gene markers of cancer can be detected if the gene selection approach is chosen reasonably.
Identification of tissue-specific targeting peptide
NASA Astrophysics Data System (ADS)
Jung, Eunkyoung; Lee, Nam Kyung; Kang, Sang-Kee; Choi, Seung-Hoon; Kim, Daejin; Park, Kisoo; Choi, Kihang; Choi, Yun-Jaie; Jung, Dong Hyun
2012-11-01
Using phage display technique, we identified tissue-targeting peptide sets that recognize specific tissues (bone-marrow dendritic cell, kidney, liver, lung, spleen and visceral adipose tissue). In order to rapidly evaluate tissue-specific targeting peptides, we performed machine learning studies for predicting the tissue-specific targeting activity of peptides on the basis of peptide sequence information using four machine learning models and isolated the groups of peptides capable of mediating selective targeting to specific tissues. As a representative liver-specific targeting sequence, the peptide "DKNLQLH" was selected by the sequence similarity analysis. This peptide has a high degree of homology with protein ligands which can interact with corresponding membrane counterparts. We anticipate that our models will be applicable to the prediction of tissue-specific targeting peptides which can recognize the endothelial markers of target tissues.
Predicting protein-binding RNA nucleotides with consideration of binding partners.
Tuvshinjargal, Narankhuu; Lee, Wook; Park, Byungkyu; Han, Kyungsook
2015-06-01
In recent years several computational methods have been developed to predict RNA-binding sites in protein. Most of these methods do not consider interacting partners of a protein, so they predict the same RNA-binding sites for a given protein sequence even if the protein binds to different RNAs. Unlike the problem of predicting RNA-binding sites in protein, the problem of predicting protein-binding sites in RNA has received little attention mainly because it is much more difficult and shows a lower accuracy on average. In our previous study, we developed a method that predicts protein-binding nucleotides from an RNA sequence. In an effort to improve the prediction accuracy and usefulness of the previous method, we developed a new method that uses both RNA and protein sequence data. In this study, we identified effective features of RNA and protein molecules and developed a new support vector machine (SVM) model to predict protein-binding nucleotides from RNA and protein sequence data. The new model that used both protein and RNA sequence data achieved a sensitivity of 86.5%, a specificity of 86.2%, a positive predictive value (PPV) of 72.6%, a negative predictive value (NPV) of 93.8% and Matthews correlation coefficient (MCC) of 0.69 in a 10-fold cross validation; it achieved a sensitivity of 58.8%, a specificity of 87.4%, a PPV of 65.1%, a NPV of 84.2% and MCC of 0.48 in independent testing. For comparative purpose, we built another prediction model that used RNA sequence data alone and ran it on the same dataset. In a 10 fold-cross validation it achieved a sensitivity of 85.7%, a specificity of 80.5%, a PPV of 67.7%, a NPV of 92.2% and MCC of 0.63; in independent testing it achieved a sensitivity of 67.7%, a specificity of 78.8%, a PPV of 57.6%, a NPV of 85.2% and MCC of 0.45. In both cross-validations and independent testing, the new model that used both RNA and protein sequences showed a better performance than the model that used RNA sequence data alone in most performance measures. To the best of our knowledge, this is the first sequence-based prediction of protein-binding nucleotides in RNA which considers the binding partner of RNA. The new model will provide valuable information for designing biochemical experiments to find putative protein-binding sites in RNA with unknown structure. Copyright © 2015 Elsevier Ireland Ltd. All rights reserved.
Segev, G; Langston, C; Takada, K; Kass, P H; Cowgill, L D
2016-05-01
A scoring system for outcome prediction in dogs with acute kidney injury (AKI) recently has been developed but has not been validated. The scoring system previously developed for outcome prediction will accurately predict outcome in a validation cohort of dogs with AKI managed with hemodialysis. One hundred fifteen client-owned dogs with AKI. Medical records of dogs with AKI treated by hemodialysis between 2011 and 2015 were reviewed. Dogs were included only if all variables required to calculate the final predictive score were available, and the 30-day outcome was known. A predictive score for 3 models was calculated for each dog. Logistic regression was used to evaluate the association of the final predictive score with each model's outcome. Receiver operating curve (ROC) analyses were performed to determine sensitivity and specificity for each model based on previously established cut-off values. Higher scores for each model were associated with decreased survival probability (P < .001). Based on previously established cut-off values, 3 models (models A, B, C) were associated with sensitivities/specificities of 73/75%, 71/80%, and 75/86%, respectively, and correctly classified 74-80% of the dogs. All models were simple to apply and allowed outcome prediction that closely corresponded with actual outcome in an independent cohort. As expected, accuracies were slightly lower compared with those from the previously reported cohort used initially to develop the models. Copyright © 2016 The Authors. Journal of Veterinary Internal Medicine published by Wiley Periodicals, Inc. on behalf of the American College of Veterinary Internal Medicine.
2014-01-01
Background It is important to predict the quality of a protein structural model before its native structure is known. The method that can predict the absolute local quality of individual residues in a single protein model is rare, yet particularly needed for using, ranking and refining protein models. Results We developed a machine learning tool (SMOQ) that can predict the distance deviation of each residue in a single protein model. SMOQ uses support vector machines (SVM) with protein sequence and structural features (i.e. basic feature set), including amino acid sequence, secondary structures, solvent accessibilities, and residue-residue contacts to make predictions. We also trained a SVM model with two new additional features (profiles and SOV scores) on 20 CASP8 targets and found that including them can only improve the performance when real deviations between native and model are higher than 5Å. The SMOQ tool finally released uses the basic feature set trained on 85 CASP8 targets. Moreover, SMOQ implemented a way to convert predicted local quality scores into a global quality score. SMOQ was tested on the 84 CASP9 single-domain targets. The average difference between the residue-specific distance deviation predicted by our method and the actual distance deviation on the test data is 2.637Å. The global quality prediction accuracy of the tool is comparable to other good tools on the same benchmark. Conclusion SMOQ is a useful tool for protein single model quality assessment. Its source code and executable are available at: http://sysbio.rnet.missouri.edu/multicom_toolbox/. PMID:24776231
Cao, Renzhi; Wang, Zheng; Wang, Yiheng; Cheng, Jianlin
2014-04-28
It is important to predict the quality of a protein structural model before its native structure is known. The method that can predict the absolute local quality of individual residues in a single protein model is rare, yet particularly needed for using, ranking and refining protein models. We developed a machine learning tool (SMOQ) that can predict the distance deviation of each residue in a single protein model. SMOQ uses support vector machines (SVM) with protein sequence and structural features (i.e. basic feature set), including amino acid sequence, secondary structures, solvent accessibilities, and residue-residue contacts to make predictions. We also trained a SVM model with two new additional features (profiles and SOV scores) on 20 CASP8 targets and found that including them can only improve the performance when real deviations between native and model are higher than 5Å. The SMOQ tool finally released uses the basic feature set trained on 85 CASP8 targets. Moreover, SMOQ implemented a way to convert predicted local quality scores into a global quality score. SMOQ was tested on the 84 CASP9 single-domain targets. The average difference between the residue-specific distance deviation predicted by our method and the actual distance deviation on the test data is 2.637Å. The global quality prediction accuracy of the tool is comparable to other good tools on the same benchmark. SMOQ is a useful tool for protein single model quality assessment. Its source code and executable are available at: http://sysbio.rnet.missouri.edu/multicom_toolbox/.
Correlation Imaging Reveals Specific Crowding Dynamics of Kinesin Motor Proteins
NASA Astrophysics Data System (ADS)
Miedema, Daniël M.; Kushwaha, Vandana S.; Denisov, Dmitry V.; Acar, Seyda; Nienhuis, Bernard; Peterman, Erwin J. G.; Schall, Peter
2017-10-01
Molecular motor proteins fulfill the critical function of transporting organelles and other building blocks along the biopolymer network of the cell's cytoskeleton, but crowding effects are believed to crucially affect this motor-driven transport due to motor interactions. Physical transport models, like the paradigmatic, totally asymmetric simple exclusion process (TASEP), have been used to predict these crowding effects based on simple exclusion interactions, but verifying them in experiments remains challenging. Here, we introduce a correlation imaging technique to precisely measure the motor density, velocity, and run length along filaments under crowding conditions, enabling us to elucidate the physical nature of crowding and test TASEP model predictions. Using the kinesin motor proteins kinesin-1 and OSM-3, we identify crowding effects in qualitative agreement with TASEP predictions, and we achieve excellent quantitative agreement by extending the model with motor-specific interaction ranges and crowding-dependent detachment probabilities. These results confirm the applicability of basic nonequilibrium models to the intracellular transport and highlight motor-specific strategies to deal with crowding.
Ensemble-based methods for forecasting census in hospital units
2013-01-01
Background The ability to accurately forecast census counts in hospital departments has considerable implications for hospital resource allocation. In recent years several different methods have been proposed forecasting census counts, however many of these approaches do not use available patient-specific information. Methods In this paper we present an ensemble-based methodology for forecasting the census under a framework that simultaneously incorporates both (i) arrival trends over time and (ii) patient-specific baseline and time-varying information. The proposed model for predicting census has three components, namely: current census count, number of daily arrivals and number of daily departures. To model the number of daily arrivals, we use a seasonality adjusted Poisson Autoregressive (PAR) model where the parameter estimates are obtained via conditional maximum likelihood. The number of daily departures is predicted by modeling the probability of departure from the census using logistic regression models that are adjusted for the amount of time spent in the census and incorporate both patient-specific baseline and time varying patient-specific covariate information. We illustrate our approach using neonatal intensive care unit (NICU) data collected at Women & Infants Hospital, Providence RI, which consists of 1001 consecutive NICU admissions between April 1st 2008 and March 31st 2009. Results Our results demonstrate statistically significant improved prediction accuracy for 3, 5, and 7 day census forecasts and increased precision of our forecasting model compared to a forecasting approach that ignores patient-specific information. Conclusions Forecasting models that utilize patient-specific baseline and time-varying information make the most of data typically available and have the capacity to substantially improve census forecasts. PMID:23721123
Ensemble-based methods for forecasting census in hospital units.
Koestler, Devin C; Ombao, Hernando; Bender, Jesse
2013-05-30
The ability to accurately forecast census counts in hospital departments has considerable implications for hospital resource allocation. In recent years several different methods have been proposed forecasting census counts, however many of these approaches do not use available patient-specific information. In this paper we present an ensemble-based methodology for forecasting the census under a framework that simultaneously incorporates both (i) arrival trends over time and (ii) patient-specific baseline and time-varying information. The proposed model for predicting census has three components, namely: current census count, number of daily arrivals and number of daily departures. To model the number of daily arrivals, we use a seasonality adjusted Poisson Autoregressive (PAR) model where the parameter estimates are obtained via conditional maximum likelihood. The number of daily departures is predicted by modeling the probability of departure from the census using logistic regression models that are adjusted for the amount of time spent in the census and incorporate both patient-specific baseline and time varying patient-specific covariate information. We illustrate our approach using neonatal intensive care unit (NICU) data collected at Women & Infants Hospital, Providence RI, which consists of 1001 consecutive NICU admissions between April 1st 2008 and March 31st 2009. Our results demonstrate statistically significant improved prediction accuracy for 3, 5, and 7 day census forecasts and increased precision of our forecasting model compared to a forecasting approach that ignores patient-specific information. Forecasting models that utilize patient-specific baseline and time-varying information make the most of data typically available and have the capacity to substantially improve census forecasts.
2014-01-01
Spatial heterogeneity in the incidence of visceral leishmaniasis (VL) is an important aspect to be considered in planning control actions for the disease. The objective of this study was to predict areas at high risk for visceral leishmaniasis (VL) based on socioeconomic indicators and remote sensing data. We applied classification and regression trees to develop and validate prediction models. Performance of the models was assessed by means of sensitivity, specificity and area under the ROC curve. The model developed was able to discriminate 15 subsets of census tracts (CT) with different probabilities of containing CT with high risk of VL occurrence. The model presented, respectively, in the validation and learning samples, sensitivity of 79% and 52%, specificity of 75% and 66%, and area under the ROC curve of 83% and 66%. Considering the complex network of factors involved in the occurrence of VL in urban areas, the results of this study showed that the development of a predictive model for VL might be feasible and useful for guiding interventions against the disease, but it is still a challenge as demonstrated by the unsatisfactory predictive performance of the model developed. PMID:24885128
Baird, Jared A; Taylor, Lynne S
2011-06-01
The purpose of this study was to gain a better understanding of which factors contribute to the eutectic composition of drug-polyethylene glycol (PEG) blends and to compare experimental values with predictions from the semi-empirical model developed by Lacoulonche et al. Eutectic compositions of various drug-PEG 3350 solid dispersions were predicted, assuming athermal mixing, and compared to experimentally determined eutectic points. The presence or absence of specific interactions between the drug and PEG 3350 were investigated using Fourier transform infrared (FT-IR) spectroscopy. The eutectic composition for haloperidol-PEG and loratadine-PEG solid dispersions was accurately predicted using the model, while predictions for aceclofenac-PEG and chlorpropamide-PEG were very different from those experimentally observed. Deviations in the model prediction from ideal behavior for the systems evaluated were confirmed to be due to the presence of specific interactions between the drug and polymer, as demonstrated by IR spectroscopy. Detailed analysis showed that the eutectic composition prediction from the model is interdependent on the crystal lattice energy of the drug compound (evaluated from the melting temperature and the heat of fusion) as well as the nature of the drug-polymer interactions. In conclusion, for compounds with melting points less than 200°C, the model is ideally suited for predicting the eutectic composition of systems where there is an absence of drug-polymer interactions.
Seizure prediction in hippocampal and neocortical epilepsy using a model-based approach
Aarabi, Ardalan; He, Bin
2014-01-01
Objectives The aim of this study is to develop a model based seizure prediction method. Methods A neural mass model was used to simulate the macro-scale dynamics of intracranial EEG data. The model was composed of pyramidal cells, excitatory and inhibitory interneurons described through state equations. Twelve model’s parameters were estimated by fitting the model to the power spectral density of intracranial EEG signals and then integrated based on information obtained by investigating changes in the parameters prior to seizures. Twenty-one patients with medically intractable hippocampal and neocortical focal epilepsy were studied. Results Tuned to obtain maximum sensitivity, an average sensitivity of 87.07% and 92.6% with an average false prediction rate of 0.2 and 0.15/h were achieved using maximum seizure occurrence periods of 30 and 50 min and a minimum seizure prediction horizon of 10 s, respectively. Under maximum specificity conditions, the system sensitivity decreased to 82.9% and 90.05% and the false prediction rates were reduced to 0.16 and 0.12/h using maximum seizure occurrence periods of 30 and 50 min, respectively. Conclusions The spatio-temporal changes in the parameters demonstrated patient-specific preictal signatures that could be used for seizure prediction. Significance The present findings suggest that the model-based approach may aid prediction of seizures. PMID:24374087
Jennifer K. Rawlins; Bruce A. Roundy; Dennis Eggett; Nathan Cline
2011-01-01
Accurate prediction of germination for species used for semi-arid land revegetation would support selection of plant materials for specific climatic conditions and sites. Wet thermal-time models predict germination time by summing progress toward germination subpopulation percentages as a function of temperature across intermittent wet periods or within singular wet...
Event-based total suspended sediment particle size distribution model
NASA Astrophysics Data System (ADS)
Thompson, Jennifer; Sattar, Ahmed M. A.; Gharabaghi, Bahram; Warner, Richard C.
2016-05-01
One of the most challenging modelling tasks in hydrology is prediction of the total suspended sediment particle size distribution (TSS-PSD) in stormwater runoff generated from exposed soil surfaces at active construction sites and surface mining operations. The main objective of this study is to employ gene expression programming (GEP) and artificial neural networks (ANN) to develop a new model with the ability to more accurately predict the TSS-PSD by taking advantage of both event-specific and site-specific factors in the model. To compile the data for this study, laboratory scale experiments using rainfall simulators were conducted on fourteen different soils to obtain TSS-PSD. This data is supplemented with field data from three construction sites in Ontario over a period of two years to capture the effect of transport and deposition within the site. The combined data sets provide a wide range of key overlooked site-specific and storm event-specific factors. Both parent soil and TSS-PSD in runoff are quantified by fitting each to a lognormal distribution. Compared to existing regression models, the developed model more accurately predicted the TSS-PSD using a more comprehensive list of key model input parameters. Employment of the new model will increase the efficiency of deployment of required best management practices, designed based on TSS-PSD, to minimize potential adverse effects of construction site runoff on aquatic life in the receiving watercourses.
Corsini, Chiara; Baker, Catriona; Kung, Ethan; Schievano, Silvia; Arbia, Gregory; Baretta, Alessia; Biglino, Giovanni; Migliavacca, Francesco; Dubini, Gabriele; Pennati, Giancarlo; Marsden, Alison; Vignon-Clementel, Irene; Taylor, Andrew; Hsia, Tain-Yen; Dorfman, Adam
2014-01-01
In patients with congenital heart disease and a single ventricle (SV), ventricular support of the circulation is inadequate, and staged palliative surgery (usually 3 stages) is needed for treatment. In the various palliative surgical stages individual differences in the circulation are important and patient-specific surgical planning is ideal. In this study, an integrated approach between clinicians and engineers has been developed, based on patient-specific multi-scale models, and is here applied to predict stage 2 surgical outcomes. This approach involves four distinct steps: (1) collection of pre-operative clinical data from a patient presenting for SV palliation, (2) construction of the pre-operative model, (3) creation of feasible virtual surgical options which couple a three-dimensional model of the surgical anatomy with a lumped parameter model (LPM) of the remainder of the circulation and (4) performance of post-operative simulations to aid clinical decision making. The pre-operative model is described, agreeing well with clinical flow tracings and mean pressures. Two surgical options (bi-directional Glenn and hemi-Fontan operations) are virtually performed and coupled to the pre-operative LPM, with the hemodynamics of both options reported. Results are validated against postoperative clinical data. Ultimately, this work represents the first patient-specific predictive modeling of stage 2 palliation using virtual surgery and closed-loop multi-scale modeling.
CMAQ Model Evaluation Framework
CMAQ is tested to establish the modeling system’s credibility in predicting pollutants such as ozone and particulate matter. Evaluation of CMAQ has been designed to assess the model’s performance for specific time periods and for specific uses.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Huesemann, Michael H.; Crowe, Braden J.; Waller, Peter
Here, a microalgae biomass growth model was developed for screening novel strains for their potential to exhibit high biomass productivities under nutrient-replete conditions in outdoor ponds subjected to fluctuating light intensities and water temperatures. Growth is modeled by first estimating the light attenuation by biomass according to a scatter-corrected Beer-Lambert Law, and then calculating the specific growth rate in discretized culture volume slices that receive declining light intensities due to attenuation. The model requires the following experimentally determined strain-specific input parameters: specific growth rate as a function of light intensity and temperature, biomass loss rate in the dark as amore » function of temperature and average light intensity during the preceding light period, and the scatter-corrected biomass light absorption coefficient. The model was successful in predicting the growth performance and biomass productivity of three different microalgae species (Chlorella sorokiniana, Nannochloropsis salina, and Picochlorum sp.) in raceway pond cultures (batch and semi-continuous) subjected to diurnal sunlight intensity and water temperature variations. Model predictions were moderately sensitive to minor deviations in input parameters. To increase the predictive power of this and other microalgae biomass growth models, a better understanding of the effects of mixing-induced rapid light dark cycles on photo-inhibition and short-term biomass losses due to dark respiration in the aphotic zone of the pond is needed.« less
Huesemann, Michael H.; Crowe, Braden J.; Waller, Peter; ...
2015-12-11
Here, a microalgae biomass growth model was developed for screening novel strains for their potential to exhibit high biomass productivities under nutrient-replete conditions in outdoor ponds subjected to fluctuating light intensities and water temperatures. Growth is modeled by first estimating the light attenuation by biomass according to a scatter-corrected Beer-Lambert Law, and then calculating the specific growth rate in discretized culture volume slices that receive declining light intensities due to attenuation. The model requires the following experimentally determined strain-specific input parameters: specific growth rate as a function of light intensity and temperature, biomass loss rate in the dark as amore » function of temperature and average light intensity during the preceding light period, and the scatter-corrected biomass light absorption coefficient. The model was successful in predicting the growth performance and biomass productivity of three different microalgae species (Chlorella sorokiniana, Nannochloropsis salina, and Picochlorum sp.) in raceway pond cultures (batch and semi-continuous) subjected to diurnal sunlight intensity and water temperature variations. Model predictions were moderately sensitive to minor deviations in input parameters. To increase the predictive power of this and other microalgae biomass growth models, a better understanding of the effects of mixing-induced rapid light dark cycles on photo-inhibition and short-term biomass losses due to dark respiration in the aphotic zone of the pond is needed.« less
Austin, Peter C.; van Klaveren, David; Vergouwe, Yvonne; Nieboer, Daan; Lee, Douglas S.; Steyerberg, Ewout W.
2017-01-01
Objective Validation of clinical prediction models traditionally refers to the assessment of model performance in new patients. We studied different approaches to geographic and temporal validation in the setting of multicenter data from two time periods. Study Design and Setting We illustrated different analytic methods for validation using a sample of 14,857 patients hospitalized with heart failure at 90 hospitals in two distinct time periods. Bootstrap resampling was used to assess internal validity. Meta-analytic methods were used to assess geographic transportability. Each hospital was used once as a validation sample, with the remaining hospitals used for model derivation. Hospital-specific estimates of discrimination (c-statistic) and calibration (calibration intercepts and slopes) were pooled using random effects meta-analysis methods. I2 statistics and prediction interval width quantified geographic transportability. Temporal transportability was assessed using patients from the earlier period for model derivation and patients from the later period for model validation. Results Estimates of reproducibility, pooled hospital-specific performance, and temporal transportability were on average very similar, with c-statistics of 0.75. Between-hospital variation was moderate according to I2 statistics and prediction intervals for c-statistics. Conclusion This study illustrates how performance of prediction models can be assessed in settings with multicenter data at different time periods. PMID:27262237
van der Fels-Klerx, H J; Booij, C J H
2010-06-01
This article provides an overview of available systems for management of Fusarium mycotoxins in the cereal grain supply chain, with an emphasis on the use of predictive mathematical modeling. From the state of the art, it proposes future developments in modeling and management and their challenges. Mycotoxin contamination in cereal grain-based feed and food products is currently managed and controlled by good agricultural practices, good manufacturing practices, hazard analysis critical control points, and by checking and more recently by notification systems and predictive mathematical models. Most of the predictive models for Fusarium mycotoxins in cereal grains focus on deoxynivalenol in wheat and aim to help growers make decisions about the application of fungicides during cultivation. Future developments in managing Fusarium mycotoxins should include the linkage between predictive mathematical models and geographical information systems, resulting into region-specific predictions for mycotoxin occurrence. The envisioned geographically oriented decision support system may incorporate various underlying models for specific users' demands and regions and various related databases to feed the particular models with (geographically oriented) input data. Depending on the user requirements, the system selects the best fitting model and available input information. Future research areas include organizing data management in the cereal grain supply chain, developing predictive models for other stakeholders (taking into account the period up to harvest), other Fusarium mycotoxins, and cereal grain types, and understanding the underlying effects of the regional component in the models.
Robust functional regression model for marginal mean and subject-specific inferences.
Cao, Chunzheng; Shi, Jian Qing; Lee, Youngjo
2017-01-01
We introduce flexible robust functional regression models, using various heavy-tailed processes, including a Student t-process. We propose efficient algorithms in estimating parameters for the marginal mean inferences and in predicting conditional means as well as interpolation and extrapolation for the subject-specific inferences. We develop bootstrap prediction intervals (PIs) for conditional mean curves. Numerical studies show that the proposed model provides a robust approach against data contamination or distribution misspecification, and the proposed PIs maintain the nominal confidence levels. A real data application is presented as an illustrative example.
Sun, Baozhou; Lam, Dao; Yang, Deshan; Grantham, Kevin; Zhang, Tiezhi; Mutic, Sasa; Zhao, Tianyu
2018-05-01
Clinical treatment planning systems for proton therapy currently do not calculate monitor units (MUs) in passive scatter proton therapy due to the complexity of the beam delivery systems. Physical phantom measurements are commonly employed to determine the field-specific output factors (OFs) but are often subject to limited machine time, measurement uncertainties and intensive labor. In this study, a machine learning-based approach was developed to predict output (cGy/MU) and derive MUs, incorporating the dependencies on gantry angle and field size for a single-room proton therapy system. The goal of this study was to develop a secondary check tool for OF measurements and eventually eliminate patient-specific OF measurements. The OFs of 1754 fields previously measured in a water phantom with calibrated ionization chambers and electrometers for patient-specific fields with various range and modulation width combinations for 23 options were included in this study. The training data sets for machine learning models in three different methods (Random Forest, XGBoost and Cubist) included 1431 (~81%) OFs. Ten-fold cross-validation was used to prevent "overfitting" and to validate each model. The remaining 323 (~19%) OFs were used to test the trained models. The difference between the measured and predicted values from machine learning models was analyzed. Model prediction accuracy was also compared with that of the semi-empirical model developed by Kooy (Phys. Med. Biol. 50, 2005). Additionally, gantry angle dependence of OFs was measured for three groups of options categorized on the selection of the second scatters. Field size dependence of OFs was investigated for the measurements with and without patient-specific apertures. All three machine learning methods showed higher accuracy than the semi-empirical model which shows considerably large discrepancy of up to 7.7% for the treatment fields with full range and full modulation width. The Cubist-based solution outperformed all other models (P < 0.001) with the mean absolute discrepancy of 0.62% and maximum discrepancy of 3.17% between the measured and predicted OFs. The OFs showed a small dependence on gantry angle for small and deep options while they were constant for large options. The OF decreased by 3%-4% as the field radius was reduced to 2.5 cm. Machine learning methods can be used to predict OF for double-scatter proton machines with greater prediction accuracy than the most popular semi-empirical prediction model. By incorporating the gantry angle dependence and field size dependence, the machine learning-based methods can be used for a sanity check of OF measurements and bears the potential to eliminate the time-consuming patient-specific OF measurements. © 2018 American Association of Physicists in Medicine.
A Machine Learning Approach to Predict Gene Regulatory Networks in Seed Development in Arabidopsis
Ni, Ying; Aghamirzaie, Delasa; Elmarakeby, Haitham; Collakova, Eva; Li, Song; Grene, Ruth; Heath, Lenwood S.
2016-01-01
Gene regulatory networks (GRNs) provide a representation of relationships between regulators and their target genes. Several methods for GRN inference, both unsupervised and supervised, have been developed to date. Because regulatory relationships consistently reprogram in diverse tissues or under different conditions, GRNs inferred without specific biological contexts are of limited applicability. In this report, a machine learning approach is presented to predict GRNs specific to developing Arabidopsis thaliana embryos. We developed the Beacon GRN inference tool to predict GRNs occurring during seed development in Arabidopsis based on a support vector machine (SVM) model. We developed both global and local inference models and compared their performance, demonstrating that local models are generally superior for our application. Using both the expression levels of the genes expressed in developing embryos and prior known regulatory relationships, GRNs were predicted for specific embryonic developmental stages. The targets that are strongly positively correlated with their regulators are mostly expressed at the beginning of seed development. Potential direct targets were identified based on a match between the promoter regions of these inferred targets and the cis elements recognized by specific regulators. Our analysis also provides evidence for previously unknown inhibitory effects of three positive regulators of gene expression. The Beacon GRN inference tool provides a valuable model system for context-specific GRN inference and is freely available at https://github.com/BeaconProjectAtVirginiaTech/beacon_network_inference.git. PMID:28066488
Evaluating diagnosis-based risk-adjustment methods in a population with spinal cord dysfunction.
Warner, Grace; Hoenig, Helen; Montez, Maria; Wang, Fei; Rosen, Amy
2004-02-01
To examine performance of models in predicting health care utilization for individuals with spinal cord dysfunction. Regression models compared 2 diagnosis-based risk-adjustment methods, the adjusted clinical groups (ACGs) and diagnostic cost groups (DCGs). To improve prediction, we added to our model: (1) spinal cord dysfunction-specific diagnostic information, (2) limitations in self-care function, and (3) both 1 and 2. Models were replicated in 3 populations. Samples from 3 populations: (1) 40% of veterans using Veterans Health Administration services in fiscal year 1997 (FY97) (N=1,046,803), (2) veteran sample with spinal cord dysfunction identified by codes from the International Statistical Classification of Diseases, 9th Revision, Clinical Modifications (N=7666), and (3) veteran sample identified in Veterans Affairs Spinal Cord Dysfunction Registry (N=5888). Not applicable. Inpatient, outpatient, and total days of care in FY97. The DCG models (R(2) range,.22-.38) performed better than ACG models (R(2) range,.04-.34) for all outcomes. Spinal cord dysfunction-specific diagnostic information improved prediction more in the ACG model than in the DCG model (R(2) range for ACG,.14-.34; R(2) range for DCG,.24-.38). Information on self-care function slightly improved performance (R(2) range increased from 0 to.04). The DCG risk-adjustment models predicted health care utilization better than ACG models. ACG model prediction was improved by adding information.
Kasthurirathne, Suranga N; Vest, Joshua R; Menachemi, Nir; Halverson, Paul K; Grannis, Shaun J
2018-01-01
A growing variety of diverse data sources is emerging to better inform health care delivery and health outcomes. We sought to evaluate the capacity for clinical, socioeconomic, and public health data sources to predict the need for various social service referrals among patients at a safety-net hospital. We integrated patient clinical data and community-level data representing patients' social determinants of health (SDH) obtained from multiple sources to build random forest decision models to predict the need for any, mental health, dietitian, social work, or other SDH service referrals. To assess the impact of SDH on improving performance, we built separate decision models using clinical and SDH determinants and clinical data only. Decision models predicting the need for any, mental health, and dietitian referrals yielded sensitivity, specificity, and accuracy measures ranging between 60% and 75%. Specificity and accuracy scores for social work and other SDH services ranged between 67% and 77%, while sensitivity scores were between 50% and 63%. Area under the receiver operating characteristic curve values for the decision models ranged between 70% and 78%. Models for predicting the need for any services reported positive predictive values between 65% and 73%. Positive predictive values for predicting individual outcomes were below 40%. The need for various social service referrals can be predicted with considerable accuracy using a wide range of readily available clinical and community data that measure socioeconomic and public health conditions. While the use of SDH did not result in significant performance improvements, our approach represents a novel and important application of risk predictive modeling. © The Author 2017. Published by Oxford University Press on behalf of the American Medical Informatics Association. All rights reserved. For Permissions, please email: journals.permissions@oup.com
Population-based absolute risk estimation with survey data
Kovalchik, Stephanie A.; Pfeiffer, Ruth M.
2013-01-01
Absolute risk is the probability that a cause-specific event occurs in a given time interval in the presence of competing events. We present methods to estimate population-based absolute risk from a complex survey cohort that can accommodate multiple exposure-specific competing risks. The hazard function for each event type consists of an individualized relative risk multiplied by a baseline hazard function, which is modeled nonparametrically or parametrically with a piecewise exponential model. An influence method is used to derive a Taylor-linearized variance estimate for the absolute risk estimates. We introduce novel measures of the cause-specific influences that can guide modeling choices for the competing event components of the model. To illustrate our methodology, we build and validate cause-specific absolute risk models for cardiovascular and cancer deaths using data from the National Health and Nutrition Examination Survey. Our applications demonstrate the usefulness of survey-based risk prediction models for predicting health outcomes and quantifying the potential impact of disease prevention programs at the population level. PMID:23686614
A review of statistical updating methods for clinical prediction models.
Su, Ting-Li; Jaki, Thomas; Hickey, Graeme L; Buchan, Iain; Sperrin, Matthew
2018-01-01
A clinical prediction model is a tool for predicting healthcare outcomes, usually within a specific population and context. A common approach is to develop a new clinical prediction model for each population and context; however, this wastes potentially useful historical information. A better approach is to update or incorporate the existing clinical prediction models already developed for use in similar contexts or populations. In addition, clinical prediction models commonly become miscalibrated over time, and need replacing or updating. In this article, we review a range of approaches for re-using and updating clinical prediction models; these fall in into three main categories: simple coefficient updating, combining multiple previous clinical prediction models in a meta-model and dynamic updating of models. We evaluated the performance (discrimination and calibration) of the different strategies using data on mortality following cardiac surgery in the United Kingdom: We found that no single strategy performed sufficiently well to be used to the exclusion of the others. In conclusion, useful tools exist for updating existing clinical prediction models to a new population or context, and these should be implemented rather than developing a new clinical prediction model from scratch, using a breadth of complementary statistical methods.
Scott, Gregory G; Margulies, Susan S; Coats, Brittany
2016-10-01
Traumatic brain injury (TBI) is a leading cause of death and disability in the USA. To help understand and better predict TBI, researchers have developed complex finite element (FE) models of the head which incorporate many biological structures such as scalp, skull, meninges, brain (with gray/white matter differentiation), and vasculature. However, most models drastically simplify the membranes and substructures between the pia and arachnoid membranes. We hypothesize that substructures in the pia-arachnoid complex (PAC) contribute substantially to brain deformation following head rotation, and that when included in FE models accuracy of extra-axial hemorrhage prediction improves. To test these hypotheses, microscale FE models of the PAC were developed to span the variability of PAC substructure anatomy and regional density. The constitutive response of these models were then integrated into an existing macroscale FE model of the immature piglet brain to identify changes in cortical stress distribution and predictions of extra-axial hemorrhage (EAH). Incorporating regional variability of PAC substructures substantially altered the distribution of principal stress on the cortical surface of the brain compared to a uniform representation of the PAC. Simulations of 24 non-impact rapid head rotations in an immature piglet animal model resulted in improved accuracy of EAH prediction (to 94 % sensitivity, 100 % specificity), as well as a high accuracy in regional hemorrhage prediction (to 82-100 % sensitivity, 100 % specificity). We conclude that including a biofidelic PAC substructure variability in FE models of the head is essential for improved predictions of hemorrhage at the brain/skull interface.
Figueroa, Isabel; Leipold, Doug; Leong, Steve; Zheng, Bing; Triguero-Carrasco, Montserrat; Fourie-O'Donohue, Aimee; Kozak, Katherine R; Xu, Keyang; Schutten, Melissa; Wang, Hong; Polson, Andrew G; Kamath, Amrita V
2018-05-14
For antibody-drug conjugates (ADCs) that carry a cytotoxic drug, doses that can be administered in preclinical studies are typically limited by tolerability, leading to a narrow dose range that can be tested. For molecules with non-linear pharmacokinetics (PK), this limited dose range may be insufficient to fully characterize the PK of the ADC and limits translation to humans. Mathematical PK models are frequently used for molecule selection during preclinical drug development and for translational predictions to guide clinical study design. Here, we present a practical approach that uses limited PK and receptor occupancy (RO) data of the corresponding unconjugated antibody to predict ADC PK when conjugation does not alter the non-specific clearance or the antibody-target interaction. We used a 2-compartment model incorporating non-specific and specific (target mediated) clearances, where the latter is a function of RO, to describe the PK of anti-CD33 ADC with dose-limiting neutropenia in cynomolgus monkeys. We tested our model by comparing PK predictions based on the unconjugated antibody to observed ADC PK data that was not utilized for model development. Prospective prediction of human PK was performed by incorporating in vitro binding affinity differences between species for varying levels of CD33 target expression. Additionally, this approach was used to predict human PK of other previously tested anti-CD33 molecules with published clinical data. The findings showed that, for a cytotoxic ADC with non-linear PK and limited preclinical PK data, incorporating RO in the PK model and using data from the corresponding unconjugated antibody at higher doses allowed the identification of parameters to characterize monkey PK and enabled human PK predictions.
The current study examines predictions of transference ratios and related modeled parameters for oxidized sulfur and oxidized nitrogen using five years (2002-2006) of 12-km grid cell-specific annual estimates from EPA’s Community Air Quality Model (CMAQ) for five selected sub-re...
Godin, Bruno; Mayer, Frédéric; Agneessens, Richard; Gerin, Patrick; Dardenne, Pierre; Delfosse, Philippe; Delcarte, Jérôme
2015-01-01
The reliability of different models to predict the biochemical methane potential (BMP) of various plant biomasses using a multispecies dataset was compared. The most reliable prediction models of the BMP were those based on the near infrared (NIR) spectrum compared to those based on the chemical composition. The NIR predictions of local (specific regression and non-linear) models were able to estimate quantitatively, rapidly, cheaply and easily the BMP. Such a model could be further used for biomethanation plant management and optimization. The predictions of non-linear models were more reliable compared to those of linear models. The presentation form (green-dried, silage-dried and silage-wet form) of biomasses to the NIR spectrometer did not influence the performances of the NIR prediction models. The accuracy of the BMP method should be improved to enhance further the BMP prediction models. Copyright © 2014 Elsevier Ltd. All rights reserved.
Hassanpour, Saeed; Langlotz, Curtis P
2016-01-01
Imaging utilization has significantly increased over the last two decades, and is only recently showing signs of moderating. To help healthcare providers identify patients at risk for high imaging utilization, we developed a prediction model to recognize high imaging utilizers based on their initial imaging reports. The prediction model uses a machine learning text classification framework. In this study, we used radiology reports from 18,384 patients with at least one abdomen computed tomography study in their imaging record at Stanford Health Care as the training set. We modeled the radiology reports in a vector space and trained a support vector machine classifier for this prediction task. We evaluated our model on a separate test set of 4791 patients. In addition to high prediction accuracy, in our method, we aimed at achieving high specificity to identify patients at high risk for high imaging utilization. Our results (accuracy: 94.0%, sensitivity: 74.4%, specificity: 97.9%, positive predictive value: 87.3%, negative predictive value: 95.1%) show that a prediction model can enable healthcare providers to identify in advance patients who are likely to be high utilizers of imaging services. Machine learning classifiers developed from narrative radiology reports are feasible methods to predict imaging utilization. Such systems can be used to identify high utilizers, inform future image ordering behavior, and encourage judicious use of imaging. Copyright © 2016 The Association of University Radiologists. Published by Elsevier Inc. All rights reserved.
Estimates of the ionization association and dissociation constant (pKa) are vital to modeling the pharmacokinetic behavior of chemicals in vivo. Methodologies for the prediction of compound sequestration in specific tissues using partition coefficients require a parameter that ch...
Francy, Donna S.; Graham, Jennifer L.; Stelzer, Erin A.; Ecker, Christopher D.; Brady, Amie M. G.; Pam Struffolino,; Loftin, Keith A.
2015-11-06
The results of this study showed that water-quality and environmental variables are promising for use in site-specific daily or long-term predictive models. In order to develop more accurate models to predict toxin concentrations at freshwater lake sites, data need to be collected more frequently and for consecutive days in future studies.
Thiels, Cornelius A; Yu, Denny; Abdelrahman, Amro M; Habermann, Elizabeth B; Hallbeck, Susan; Pasupathy, Kalyan S; Bingener, Juliane
2017-01-01
Reliable prediction of operative duration is essential for improving patient and care team satisfaction, optimizing resource utilization and reducing cost. Current operative scheduling systems are unreliable and contribute to costly over- and underestimation of operative time. We hypothesized that the inclusion of patient-specific factors would improve the accuracy in predicting operative duration. We reviewed all elective laparoscopic cholecystectomies performed at a single institution between 01/2007 and 06/2013. Concurrent procedures were excluded. Univariate analysis evaluated the effect of age, gender, BMI, ASA, laboratory values, smoking, and comorbidities on operative duration. Multivariable linear regression models were constructed using the significant factors (p < 0.05). The patient factors model was compared to the traditional surgical scheduling system estimates, which uses historical surgeon-specific and procedure-specific operative duration. External validation was done using the ACS-NSQIP database (n = 11,842). A total of 1801 laparoscopic cholecystectomy patients met inclusion criteria. Female sex was associated with reduced operative duration (-7.5 min, p < 0.001 vs. male sex) while increasing BMI (+5.1 min BMI 25-29.9, +6.9 min BMI 30-34.9, +10.4 min BMI 35-39.9, +17.0 min BMI 40 + , all p < 0.05 vs. normal BMI), increasing ASA (+7.4 min ASA III, +38.3 min ASA IV, all p < 0.01 vs. ASA I), and elevated liver function tests (+7.9 min, p < 0.01 vs. normal) were predictive of increased operative duration on univariate analysis. A model was then constructed using these predictive factors. The traditional surgical scheduling system was poorly predictive of actual operative duration (R 2 = 0.001) compared to the patient factors model (R 2 = 0.08). The model remained predictive on external validation (R 2 = 0.14).The addition of surgeon as a variable in the institutional model further improved predictive ability of the model (R 2 = 0.18). The use of routinely available pre-operative patient factors improves the prediction of operative duration during cholecystectomy.
Naghibi Beidokhti, Hamid; Janssen, Dennis; van de Groes, Sebastiaan; Hazrati, Javad; Van den Boogaard, Ton; Verdonschot, Nico
2017-12-08
In finite element (FE) models knee ligaments can represented either by a group of one-dimensional springs, or by three-dimensional continuum elements based on segmentations. Continuum models closer approximate the anatomy, and facilitate ligament wrapping, while spring models are computationally less expensive. The mechanical properties of ligaments can be based on literature, or adjusted specifically for the subject. In the current study we investigated the effect of ligament modelling strategy on the predictive capability of FE models of the human knee joint. The effect of literature-based versus specimen-specific optimized material parameters was evaluated. Experiments were performed on three human cadaver knees, which were modelled in FE models with ligaments represented either using springs, or using continuum representations. In spring representation collateral ligaments were each modelled with three and cruciate ligaments with two single-element bundles. Stiffness parameters and pre-strains were optimized based on laxity tests for both approaches. Validation experiments were conducted to evaluate the outcomes of the FE models. Models (both spring and continuum) with subject-specific properties improved the predicted kinematics and contact outcome parameters. Models incorporating literature-based parameters, and particularly the spring models (with the representations implemented in this study), led to relatively high errors in kinematics and contact pressures. Using a continuum modelling approach resulted in more accurate contact outcome variables than the spring representation with two (cruciate ligaments) and three (collateral ligaments) single-element-bundle representations. However, when the prediction of joint kinematics is of main interest, spring ligament models provide a faster option with acceptable outcome. Copyright © 2017 Elsevier Ltd. All rights reserved.
Ayton, Ellyn; Porterfield, Katherine; Corley, Courtney D.
2017-01-01
This work is the first to take advantage of recurrent neural networks to predict influenza-like illness (ILI) dynamics from various linguistic signals extracted from social media data. Unlike other approaches that rely on timeseries analysis of historical ILI data and the state-of-the-art machine learning models, we build and evaluate the predictive power of neural network architectures based on Long Short Term Memory (LSTMs) units capable of nowcasting (predicting in “real-time”) and forecasting (predicting the future) ILI dynamics in the 2011 – 2014 influenza seasons. To build our models we integrate information people post in social media e.g., topics, embeddings, word ngrams, stylistic patterns, and communication behavior using hashtags and mentions. We then quantitatively evaluate the predictive power of different social media signals and contrast the performance of the-state-of-the-art regression models with neural networks using a diverse set of evaluation metrics. Finally, we combine ILI and social media signals to build a joint neural network model for ILI dynamics prediction. Unlike the majority of the existing work, we specifically focus on developing models for local rather than national ILI surveillance, specifically for military rather than general populations in 26 U.S. and six international locations., and analyze how model performance depends on the amount of social media data available per location. Our approach demonstrates several advantages: (a) Neural network architectures that rely on LSTM units trained on social media data yield the best performance compared to previously used regression models. (b) Previously under-explored language and communication behavior features are more predictive of ILI dynamics than stylistic and topic signals expressed in social media. (c) Neural network models learned exclusively from social media signals yield comparable or better performance to the models learned from ILI historical data, thus, signals from social media can be potentially used to accurately forecast ILI dynamics for the regions where ILI historical data is not available. (d) Neural network models learned from combined ILI and social media signals significantly outperform models that rely solely on ILI historical data, which adds to a great potential of alternative public sources for ILI dynamics prediction. (e) Location-specific models outperform previously used location-independent models e.g., U.S. only. (f) Prediction results significantly vary across geolocations depending on the amount of social media data available and ILI activity patterns. (g) Model performance improves with more tweets available per geo-location e.g., the error gets lower and the Pearson score gets higher for locations with more tweets. PMID:29244814
Volkova, Svitlana; Ayton, Ellyn; Porterfield, Katherine; Corley, Courtney D
2017-01-01
This work is the first to take advantage of recurrent neural networks to predict influenza-like illness (ILI) dynamics from various linguistic signals extracted from social media data. Unlike other approaches that rely on timeseries analysis of historical ILI data and the state-of-the-art machine learning models, we build and evaluate the predictive power of neural network architectures based on Long Short Term Memory (LSTMs) units capable of nowcasting (predicting in "real-time") and forecasting (predicting the future) ILI dynamics in the 2011 - 2014 influenza seasons. To build our models we integrate information people post in social media e.g., topics, embeddings, word ngrams, stylistic patterns, and communication behavior using hashtags and mentions. We then quantitatively evaluate the predictive power of different social media signals and contrast the performance of the-state-of-the-art regression models with neural networks using a diverse set of evaluation metrics. Finally, we combine ILI and social media signals to build a joint neural network model for ILI dynamics prediction. Unlike the majority of the existing work, we specifically focus on developing models for local rather than national ILI surveillance, specifically for military rather than general populations in 26 U.S. and six international locations., and analyze how model performance depends on the amount of social media data available per location. Our approach demonstrates several advantages: (a) Neural network architectures that rely on LSTM units trained on social media data yield the best performance compared to previously used regression models. (b) Previously under-explored language and communication behavior features are more predictive of ILI dynamics than stylistic and topic signals expressed in social media. (c) Neural network models learned exclusively from social media signals yield comparable or better performance to the models learned from ILI historical data, thus, signals from social media can be potentially used to accurately forecast ILI dynamics for the regions where ILI historical data is not available. (d) Neural network models learned from combined ILI and social media signals significantly outperform models that rely solely on ILI historical data, which adds to a great potential of alternative public sources for ILI dynamics prediction. (e) Location-specific models outperform previously used location-independent models e.g., U.S. only. (f) Prediction results significantly vary across geolocations depending on the amount of social media data available and ILI activity patterns. (g) Model performance improves with more tweets available per geo-location e.g., the error gets lower and the Pearson score gets higher for locations with more tweets.
Uncertainty in age-specific harvest estimates and consequences for white-tailed deer management
Collier, B.A.; Krementz, D.G.
2007-01-01
Age structure proportions (proportion of harvested individuals within each age class) are commonly used as support for regulatory restrictions and input for deer population models. Such use requires critical evaluation when harvest regulations force hunters to selectively harvest specific age classes, due to impact on the underlying population age structure. We used a stochastic population simulation model to evaluate the impact of using harvest proportions to evaluate changes in population age structure under a selective harvest management program at two scales. Using harvest proportions to parameterize the age-specific harvest segment of the model for the local scale showed that predictions of post-harvest age structure did not vary dependent upon whether selective harvest criteria were in use or not. At the county scale, yearling frequency in the post-harvest population increased, but model predictions indicated that post-harvest population size of 2.5 years old males would decline below levels found before implementation of the antler restriction, reducing the number of individuals recruited into older age classes. Across the range of age-specific harvest rates modeled, our simulation predicted that underestimation of age-specific harvest rates has considerable influence on predictions of post-harvest population age structure. We found that the consequence of uncertainty in harvest rates corresponds to uncertainty in predictions of residual population structure, and this correspondence is proportional to scale. Our simulations also indicate that regardless of use of harvest proportions or harvest rates, at either the local or county scale the modeled SHC had a high probability (>0.60 and >0.75, respectively) of eliminating recruitment into >2.5 years old age classes. Although frequently used to increase population age structure, our modeling indicated that selective harvest criteria can decrease or eliminate the number of white-tailed deer recruited into older age classes. Thus, we suggest that using harvest proportions for management planning and evaluation should be viewed with caution. In addition, we recommend that managers focus more attention on estimation of age-specific harvest rates, and modeling approaches which combine harvest rates with information from harvested individuals to further increase their ability to effectively manage deer populations under selective harvest programs. ?? 2006 Elsevier B.V. All rights reserved.
Are Subject-Specific Musculoskeletal Models Robust to the Uncertainties in Parameter Identification?
Valente, Giordano; Pitto, Lorenzo; Testi, Debora; Seth, Ajay; Delp, Scott L.; Stagni, Rita; Viceconti, Marco; Taddei, Fulvia
2014-01-01
Subject-specific musculoskeletal modeling can be applied to study musculoskeletal disorders, allowing inclusion of personalized anatomy and properties. Independent of the tools used for model creation, there are unavoidable uncertainties associated with parameter identification, whose effect on model predictions is still not fully understood. The aim of the present study was to analyze the sensitivity of subject-specific model predictions (i.e., joint angles, joint moments, muscle and joint contact forces) during walking to the uncertainties in the identification of body landmark positions, maximum muscle tension and musculotendon geometry. To this aim, we created an MRI-based musculoskeletal model of the lower limbs, defined as a 7-segment, 10-degree-of-freedom articulated linkage, actuated by 84 musculotendon units. We then performed a Monte-Carlo probabilistic analysis perturbing model parameters according to their uncertainty, and solving a typical inverse dynamics and static optimization problem using 500 models that included the different sets of perturbed variable values. Model creation and gait simulations were performed by using freely available software that we developed to standardize the process of model creation, integrate with OpenSim and create probabilistic simulations of movement. The uncertainties in input variables had a moderate effect on model predictions, as muscle and joint contact forces showed maximum standard deviation of 0.3 times body-weight and maximum range of 2.1 times body-weight. In addition, the output variables significantly correlated with few input variables (up to 7 out of 312) across the gait cycle, including the geometry definition of larger muscles and the maximum muscle tension in limited gait portions. Although we found subject-specific models not markedly sensitive to parameter identification, researchers should be aware of the model precision in relation to the intended application. In fact, force predictions could be affected by an uncertainty in the same order of magnitude of its value, although this condition has low probability to occur. PMID:25390896
Patient-specific dosimetric endpoints based treatment plan quality control in radiotherapy.
Song, Ting; Staub, David; Chen, Mingli; Lu, Weiguo; Tian, Zhen; Jia, Xun; Li, Yongbao; Zhou, Linghong; Jiang, Steve B; Gu, Xuejun
2015-11-07
In intensity modulated radiotherapy (IMRT), the optimal plan for each patient is specific due to unique patient anatomy. To achieve such a plan, patient-specific dosimetric goals reflecting each patient's unique anatomy should be defined and adopted in the treatment planning procedure for plan quality control. This study is to develop such a personalized treatment plan quality control tool by predicting patient-specific dosimetric endpoints (DEs). The incorporation of patient specific DEs is realized by a multi-OAR geometry-dosimetry model, capable of predicting optimal DEs based on the individual patient's geometry. The overall quality of a treatment plan is then judged with a numerical treatment plan quality indicator and characterized as optimal or suboptimal. Taking advantage of clinically available prostate volumetric modulated arc therapy (VMAT) treatment plans, we built and evaluated our proposed plan quality control tool. Using our developed tool, six of twenty evaluated plans were identified as sub-optimal plans. After plan re-optimization, these suboptimal plans achieved better OAR dose sparing without sacrificing the PTV coverage, and the dosimetric endpoints of the re-optimized plans agreed well with the model predicted values, which validate the predictability of the proposed tool. In conclusion, the developed tool is able to accurately predict optimally achievable DEs of multiple OARs, identify suboptimal plans, and guide plan optimization. It is a useful tool for achieving patient-specific treatment plan quality control.
Logistic Mixed Models to Investigate Implicit and Explicit Belief Tracking
Lages, Martin; Scheel, Anne
2016-01-01
We investigated the proposition of a two-systems Theory of Mind in adults’ belief tracking. A sample of N = 45 participants predicted the choice of one of two opponent players after observing several rounds in an animated card game. Three matches of this card game were played and initial gaze direction on target and subsequent choice predictions were recorded for each belief task and participant. We conducted logistic regressions with mixed effects on the binary data and developed Bayesian logistic mixed models to infer implicit and explicit mentalizing in true belief and false belief tasks. Although logistic regressions with mixed effects predicted the data well a Bayesian logistic mixed model with latent task- and subject-specific parameters gave a better account of the data. As expected explicit choice predictions suggested a clear understanding of true and false beliefs (TB/FB). Surprisingly, however, model parameters for initial gaze direction also indicated belief tracking. We discuss why task-specific parameters for initial gaze directions are different from choice predictions yet reflect second-order perspective taking. PMID:27853440
Explicit Pore Pressure Material Model in Carbon-Cloth Phenolic
NASA Technical Reports Server (NTRS)
Gutierrez-Lemini, Danton; Ehle, Curt
2003-01-01
An explicit material model that uses predicted pressure in the pores of a carbon-cloth phenolic (CCP) composite has been developed. This model is intended to be used within a finite-element model to predict phenomena specific to CCP components of solid-fuel-rocket nozzles subjected to high operating temperatures and to mechanical stresses that can be great enough to cause structural failures. Phenomena that can be predicted with the help of this model include failures of specimens in restrained-thermal-growth (RTG) tests, pocketing erosion, and ply lifting
Otgonsuren, Munkhzul; Estep, Michael J; Hossain, Nayeem; Younossi, Elena; Frost, Spencer; Henry, Linda; Hunt, Sharon; Fang, Yun; Goodman, Zachary; Younossi, Zobair M
2014-12-01
Non-alcoholic steatohepatitis (NASH) is the progressive form of non-alcoholic fatty liver disease (NAFLD). A liver biopsy is considered the "gold standard" for diagnosing/staging NASH. Identification of NAFLD/NASH using non-invasive tools is important for intervention. The study aims were to: develop/validate the predictive performance of a non-invasive model (index of NASH [ION]); assess the performance of a recognized non-invasive model (fatty liver index [FLI]) compared with ION for NAFLD diagnosis; determine which non-invasive model (FLI, ION, or NAFLD fibrosis score [NFS]) performed best in predicting age-adjusted mortality. From the National Health and Nutrition Examination Survey III database, anthropometric, clinical, ultrasound, laboratory, and mortality data were obtained (n = 4458; n = 861 [19.3%] NAFLD by ultrasound) and used to develop the ION model, and then to compare the ION and FLI models for NAFLD diagnosis. For validation and diagnosis of NASH, liver biopsy data were used (n = 152). Age-adjusted Cox proportional hazard modeling estimated the association among the three non-invasive tests (FLI, ION, and NFS) and mortality. FLI's threshold score > 60 and ION's threshold score > 22 had similar specificity (FLI = 80% vs ION = 82%) for NAFLD diagnosis; FLI < 30 (80% sensitivity) and ION < 11 (81% sensitivity) excluded NAFLD. An ION score > 50 predicted histological NASH (92% specificity); the FLI model did not predict NASH or mortality. The ION model was best in predicting cardiovascular/diabetes-related mortality; NFS predicted overall or diabetes-related mortality. The ION model was superior in predicting NASH and mortality compared with the FLI model. Studies are needed to validate ION. © 2014 Journal of Gastroenterology and Hepatology Foundation and Wiley Publishing Asia Pty Ltd.
Prediction of subjective ratings of emotional pictures by EEG features
NASA Astrophysics Data System (ADS)
McFarland, Dennis J.; Parvaz, Muhammad A.; Sarnacki, William A.; Goldstein, Rita Z.; Wolpaw, Jonathan R.
2017-02-01
Objective. Emotion dysregulation is an important aspect of many psychiatric disorders. Brain-computer interface (BCI) technology could be a powerful new approach to facilitating therapeutic self-regulation of emotions. One possible BCI method would be to provide stimulus-specific feedback based on subject-specific electroencephalographic (EEG) responses to emotion-eliciting stimuli. Approach. To assess the feasibility of this approach, we studied the relationships between emotional valence/arousal and three EEG features: amplitude of alpha activity over frontal cortex; amplitude of theta activity over frontal midline cortex; and the late positive potential over central and posterior mid-line areas. For each feature, we evaluated its ability to predict emotional valence/arousal on both an individual and a group basis. Twenty healthy participants (9 men, 11 women; ages 22-68) rated each of 192 pictures from the IAPS collection in terms of valence and arousal twice (96 pictures on each of 4 d over 2 weeks). EEG was collected simultaneously and used to develop models based on canonical correlation to predict subject-specific single-trial ratings. Separate models were evaluated for the three EEG features: frontal alpha activity; frontal midline theta; and the late positive potential. In each case, these features were used to simultaneously predict both the normed ratings and the subject-specific ratings. Main results. Models using each of the three EEG features with data from individual subjects were generally successful at predicting subjective ratings on training data, but generalization to test data was less successful. Sparse models performed better than models without regularization. Significance. The results suggest that the frontal midline theta is a better candidate than frontal alpha activity or the late positive potential for use in a BCI-based paradigm designed to modify emotional reactions.
Regression model estimation of early season crop proportions: North Dakota, some preliminary results
NASA Technical Reports Server (NTRS)
Lin, K. K. (Principal Investigator)
1982-01-01
To estimate crop proportions early in the season, an approach is proposed based on: use of a regression-based prediction equation to obtain an a priori estimate for specific major crop groups; modification of this estimate using current-year LANDSAT and weather data; and a breakdown of the major crop groups into specific crops by regression models. Results from the development and evaluation of appropriate regression models for the first portion of the proposed approach are presented. The results show that the model predicts 1980 crop proportions very well at both county and crop reporting district levels. In terms of planted acreage, the model underpredicted 9.1 percent of the 1980 published data on planted acreage at the county level. It predicted almost exactly the 1980 published data on planted acreage at the crop reporting district level and overpredicted the planted acreage by just 0.92 percent.
Modesto-Alapont, Vicente; Gonzalez-Marrachelli, Vannina; Vento-Rehues, Rosa; Jorda-Miñana, Angela; Blanquer-Olivas, Jose; Monleon, Daniel
2015-01-01
Early diagnosis and patient stratification may improve sepsis outcome by a timely start of the proper specific treatment. We aimed to identify metabolomic biomarkers of sepsis in urine by 1H-NMR spectroscopy to assess the severity and to predict outcomes. Urine samples were collected from 64 patients with severe sepsis or septic shock in the ICU for a 1H NMR spectra acquisition. A supervised analysis was performed on the processed spectra, and a predictive model for prognosis (30-days mortality/survival) of sepsis was constructed using partial least-squares discriminant analysis (PLS-DA). In addition, we compared the prediction power of metabolomics data respect the Sequential Organ Failure Assessment (SOFA) score. Supervised multivariate analysis afforded a good predictive model to distinguish the patient groups and detect specific metabolic patterns. Negative prognosis patients presented higher values of ethanol, glucose and hippurate, and on the contrary, lower levels of methionine, glutamine, arginine and phenylalanine. These metabolites could be part of a composite biopattern of the human metabolic response to sepsis shock and its mortality in ICU patients. The internal cross-validation showed robustness of the metabolic predictive model obtained and a better predictive ability in comparison with SOFA values. Our results indicate that NMR metabolic profiling might be helpful for determining the metabolomic phenotype of worst-prognosis septic patients in an early stage. A predictive model for the evolution of septic patients using these metabolites was able to classify cases with more sensitivity and specificity than the well-established organ dysfunction score SOFA. PMID:26565633
NASA Astrophysics Data System (ADS)
Wang, Yu; Guo, Yanzhi; Kuang, Qifan; Pu, Xuemei; Ji, Yue; Zhang, Zhihang; Li, Menglong
2015-04-01
The assessment of binding affinity between ligands and the target proteins plays an essential role in drug discovery and design process. As an alternative to widely used scoring approaches, machine learning methods have also been proposed for fast prediction of the binding affinity with promising results, but most of them were developed as all-purpose models despite of the specific functions of different protein families, since proteins from different function families always have different structures and physicochemical features. In this study, we proposed a random forest method to predict the protein-ligand binding affinity based on a comprehensive feature set covering protein sequence, binding pocket, ligand structure and intermolecular interaction. Feature processing and compression was respectively implemented for different protein family datasets, which indicates that different features contribute to different models, so individual representation for each protein family is necessary. Three family-specific models were constructed for three important protein target families of HIV-1 protease, trypsin and carbonic anhydrase respectively. As a comparison, two generic models including diverse protein families were also built. The evaluation results show that models on family-specific datasets have the superior performance to those on the generic datasets and the Pearson and Spearman correlation coefficients ( R p and Rs) on the test sets are 0.740, 0.874, 0.735 and 0.697, 0.853, 0.723 for HIV-1 protease, trypsin and carbonic anhydrase respectively. Comparisons with the other methods further demonstrate that individual representation and model construction for each protein family is a more reasonable way in predicting the affinity of one particular protein family.
Stylianou, Neophytos; Akbarov, Artur; Kontopantelis, Evangelos; Buchan, Iain; Dunn, Ken W
2015-08-01
Predicting mortality from burn injury has traditionally employed logistic regression models. Alternative machine learning methods have been introduced in some areas of clinical prediction as the necessary software and computational facilities have become accessible. Here we compare logistic regression and machine learning predictions of mortality from burn. An established logistic mortality model was compared to machine learning methods (artificial neural network, support vector machine, random forests and naïve Bayes) using a population-based (England & Wales) case-cohort registry. Predictive evaluation used: area under the receiver operating characteristic curve; sensitivity; specificity; positive predictive value and Youden's index. All methods had comparable discriminatory abilities, similar sensitivities, specificities and positive predictive values. Although some machine learning methods performed marginally better than logistic regression the differences were seldom statistically significant and clinically insubstantial. Random forests were marginally better for high positive predictive value and reasonable sensitivity. Neural networks yielded slightly better prediction overall. Logistic regression gives an optimal mix of performance and interpretability. The established logistic regression model of burn mortality performs well against more complex alternatives. Clinical prediction with a small set of strong, stable, independent predictors is unlikely to gain much from machine learning outside specialist research contexts. Copyright © 2015 Elsevier Ltd and ISBI. All rights reserved.
Chiesa, Marco; Cirasola, Antonella; Williams, Riccardo; Nassisi, Valentina; Fonagy, Peter
2017-04-01
Although several studies have highlighted the relationship between attachment states of mind and personality disorders, their findings have not been consistent, possibly due to the application of the traditional taxonomic classification model of attachment. A more recently developed dimensional classification of attachment representations, including more specific aspects of trauma-related representations, may have advantages. In this study, we compare specific associations and predictive power of the categorical attachment and dimensional models applied to 230 Adult Attachment Interview transcripts obtained from personality disordered and nonpsychiatric subjects. We also investigate the role that current levels of psychiatric distress may have in the prediction of PD. The results showed that both models predict the presence of PD, with the dimensional approach doing better in discriminating overall diagnosis of PD. However, both models are less helpful in discriminating specific PD diagnostic subtypes. Current psychiatric distress was found to be the most consistent predictor of PD capturing a large share of the variance and obscuring the role played by attachment variables. The results suggest that attachment parameters correlate with the presence of PD alone and have no specific associations with particular PD subtypes when current psychiatric distress is taken into account.
Feasibility of developing LSI microcircuit reliability prediction models
NASA Technical Reports Server (NTRS)
Ryerson, C. M.
1972-01-01
In the proposed modeling approach, when any of the essential key factors are not known initially, they can be approximated in various ways with a known impact on the accuracy of the final predictions. For example, on any program where reliability predictions are started at interim states of project completion, a-priori approximate estimates of the key factors are established for making preliminary predictions. Later these are refined for greater accuracy as subsequent program information of a more definitive nature becomes available. Specific steps to develop, validate and verify these new models are described.
Backenroth, Daniel; He, Zihuai; Kiryluk, Krzysztof; Boeva, Valentina; Pethukova, Lynn; Khurana, Ekta; Christiano, Angela; Buxbaum, Joseph D; Ionita-Laza, Iuliana
2018-05-03
We describe a method based on a latent Dirichlet allocation model for predicting functional effects of noncoding genetic variants in a cell-type- and/or tissue-specific way (FUN-LDA). Using this unsupervised approach, we predict tissue-specific functional effects for every position in the human genome in 127 different tissues and cell types. We demonstrate the usefulness of our predictions by using several validation experiments. Using eQTL data from several sources, including the GTEx project, Geuvadis project, and TwinsUK cohort, we show that eQTLs in specific tissues tend to be most enriched among the predicted functional variants in relevant tissues in Roadmap. We further show how these integrated functional scores can be used for (1) deriving the most likely cell or tissue type causally implicated for a complex trait by using summary statistics from genome-wide association studies and (2) estimating a tissue-based correlation matrix of various complex traits. We found large enrichment of heritability in functional components of relevant tissues for various complex traits, and FUN-LDA yielded higher enrichment estimates than existing methods. Finally, using experimentally validated functional variants from the literature and variants possibly implicated in disease by previous studies, we rigorously compare FUN-LDA with state-of-the-art functional annotation methods and show that FUN-LDA has better prediction accuracy and higher resolution than these methods. In particular, our results suggest that tissue- and cell-type-specific functional prediction methods tend to have substantially better prediction accuracy than organism-level prediction methods. Scores for each position in the human genome and for each ENCODE and Roadmap tissue are available online (see Web Resources). Copyright © 2018 American Society of Human Genetics. Published by Elsevier Inc. All rights reserved.
Estimating Model Prediction Error: Should You Treat Predictions as Fixed or Random?
NASA Technical Reports Server (NTRS)
Wallach, Daniel; Thorburn, Peter; Asseng, Senthold; Challinor, Andrew J.; Ewert, Frank; Jones, James W.; Rotter, Reimund; Ruane, Alexander
2016-01-01
Crop models are important tools for impact assessment of climate change, as well as for exploring management options under current climate. It is essential to evaluate the uncertainty associated with predictions of these models. We compare two criteria of prediction error; MSEP fixed, which evaluates mean squared error of prediction for a model with fixed structure, parameters and inputs, and MSEP uncertain( X), which evaluates mean squared error averaged over the distributions of model structure, inputs and parameters. Comparison of model outputs with data can be used to estimate the former. The latter has a squared bias term, which can be estimated using hindcasts, and a model variance term, which can be estimated from a simulation experiment. The separate contributions to MSEP uncertain (X) can be estimated using a random effects ANOVA. It is argued that MSEP uncertain (X) is the more informative uncertainty criterion, because it is specific to each prediction situation.
NASA Astrophysics Data System (ADS)
Ramakrishnan, N.; Tourdot, Richard W.; Eckmann, David M.; Ayyaswamy, Portonovo S.; Muzykantov, Vladimir R.; Radhakrishnan, Ravi
2016-06-01
In order to achieve selective targeting of affinity-ligand coated nanoparticles to the target tissue, it is essential to understand the key mechanisms that govern their capture by the target cell. Next-generation pharmacokinetic (PK) models that systematically account for proteomic and mechanical factors can accelerate the design, validation and translation of targeted nanocarriers (NCs) in the clinic. Towards this objective, we have developed a computational model to delineate the roles played by target protein expression and mechanical factors of the target cell membrane in determining the avidity of functionalized NCs to live cells. Model results show quantitative agreement with in vivo experiments when specific and non-specific contributions to NC binding are taken into account. The specific contributions are accounted for through extensive simulations of multivalent receptor-ligand interactions, membrane mechanics and entropic factors such as membrane undulations and receptor translation. The computed NC avidity is strongly dependent on ligand density, receptor expression, bending mechanics of the target cell membrane, as well as entropic factors associated with the membrane and the receptor motion. Our computational model can predict the in vivo targeting levels of the intracellular adhesion molecule-1 (ICAM1)-coated NCs targeted to the lung, heart, kidney, liver and spleen of mouse, when the contributions due to endothelial capture are accounted for. The effect of other cells (such as monocytes, etc.) do not improve the model predictions at steady state. We demonstrate the predictive utility of our model by predicting partitioning coefficients of functionalized NCs in mice and human tissues and report the statistical accuracy of our model predictions under different scenarios.
Financial Distress Prediction Using Discrete-time Hazard Model and Rating Transition Matrix Approach
NASA Astrophysics Data System (ADS)
Tsai, Bi-Huei; Chang, Chih-Huei
2009-08-01
Previous studies used constant cut-off indicator to distinguish distressed firms from non-distressed ones in the one-stage prediction models. However, distressed cut-off indicator must shift according to economic prosperity, rather than remains fixed all the time. This study focuses on Taiwanese listed firms and develops financial distress prediction models based upon the two-stage method. First, this study employs the firm-specific financial ratio and market factors to measure the probability of financial distress based on the discrete-time hazard models. Second, this paper further focuses on macroeconomic factors and applies rating transition matrix approach to determine the distressed cut-off indicator. The prediction models are developed by using the training sample from 1987 to 2004, and their levels of accuracy are compared with the test sample from 2005 to 2007. As for the one-stage prediction model, the model in incorporation with macroeconomic factors does not perform better than that without macroeconomic factors. This suggests that the accuracy is not improved for one-stage models which pool the firm-specific and macroeconomic factors together. In regards to the two stage models, the negative credit cycle index implies the worse economic status during the test period, so the distressed cut-off point is adjusted to increase based on such negative credit cycle index. After the two-stage models employ such adjusted cut-off point to discriminate the distressed firms from non-distressed ones, their error of misclassification becomes lower than that of one-stage ones. The two-stage models presented in this paper have incremental usefulness in predicting financial distress.
Peterson, A Townsend; Martínez-Campos, Carmen; Nakazawa, Yoshinori; Martínez-Meyer, Enrique
2005-09-01
Numerous human diseases-malaria, dengue, yellow fever and leishmaniasis, to name a few-are transmitted by insect vectors with brief life cycles and biting activity that varies in both space and time. Although the general geographic distributions of these epidemiologically important species are known, the spatiotemporal variation in their emergence and activity remains poorly understood. We used ecological niche modeling via a genetic algorithm to produce time-specific predictive models of monthly distributions of Aedes aegypti in Mexico in 1995. Significant predictions of monthly mosquito activity and distributions indicate that predicting spatiotemporal dynamics of disease vector species is feasible; significant coincidence with human cases of dengue indicate that these dynamics probably translate directly into transmission of dengue virus to humans. This approach provides new potential for optimizing use of resources for disease prevention and remediation via automated forecasting of disease transmission risk.
The capability of physiologically based pharmacokinetic models to incorporate age-appropriate physiological and chemical-specific parameters was utilized to predict changes in internal dosimetry for six volatile organic compounds (VOCs) across different ages of rats.
Knowledge-based grouping of modeled HLA peptide complexes.
Kangueane, P; Sakharkar, M K; Lim, K S; Hao, H; Lin, K; Chee, R E; Kolatkar, P R
2000-05-01
Human leukocyte antigens are the most polymorphic of human genes and multiple sequence alignment shows that such polymorphisms are clustered in the functional peptide binding domains. Because of such polymorphism among the peptide binding residues, the prediction of peptides that bind to specific HLA molecules is very difficult. In recent years two different types of computer based prediction methods have been developed and both the methods have their own advantages and disadvantages. The nonavailability of allele specific binding data restricts the use of knowledge-based prediction methods for a wide range of HLA alleles. Alternatively, the modeling scheme appears to be a promising predictive tool for the selection of peptides that bind to specific HLA molecules. The scoring of the modeled HLA-peptide complexes is a major concern. The use of knowledge based rules (van der Waals clashes and solvent exposed hydrophobic residues) to distinguish binders from nonbinders is applied in the present study. The rules based on (1) number of observed atomic clashes between the modeled peptide and the HLA structure, and (2) number of solvent exposed hydrophobic residues on the modeled peptide effectively discriminate experimentally known binders from poor/nonbinders. Solved crystal complexes show no vdW Clash (vdWC) in 95% cases and no solvent exposed hydrophobic peptide residues (SEHPR) were seen in 86% cases. In our attempt to compare experimental binding data with the predicted scores by this scoring scheme, 77% of the peptides are correctly grouped as good binders with a sensitivity of 71%.
Raymer, James; Abel, Guy J.; Rogers, Andrei
2012-01-01
Population projection models that introduce uncertainty are a growing subset of projection models in general. In this paper, we focus on the importance of decisions made with regard to the model specifications adopted. We compare the forecasts and prediction intervals associated with four simple regional population projection models: an overall growth rate model, a component model with net migration, a component model with in-migration and out-migration rates, and a multiregional model with destination-specific out-migration rates. Vector autoregressive models are used to forecast future rates of growth, birth, death, net migration, in-migration and out-migration, and destination-specific out-migration for the North, Midlands and South regions in England. They are also used to forecast different international migration measures. The base data represent a time series of annual data provided by the Office for National Statistics from 1976 to 2008. The results illustrate how both the forecasted subpopulation totals and the corresponding prediction intervals differ for the multiregional model in comparison to other simpler models, as well as for different assumptions about international migration. The paper ends end with a discussion of our results and possible directions for future research. PMID:23236221
Mousavi, S Jamaleddin; Farzaneh, Solmaz; Avril, Stéphane
2018-04-01
Dissections of ascending thoracic aortic aneurysms (ATAAs) cause significant morbidity and mortality worldwide. They occur when a tear in the intima-media of the aorta permits the penetration of the blood and the subsequent delamination and separation of the wall in 2 layers, forming a false channel. To predict computationally the risk of tear formation, stress analyses should be performed layer-specifically and they should consider internal or residual stresses that exist in the tissue. In the present paper, we propose a novel layer-specific damage model based on the constrained mixture theory, which intrinsically takes into account these internal stresses and can predict appropriately the tear formation. The model is implemented in finite-element commercial software Abaqus coupled with user material subroutine. Its capability is tested by applying it to the simulation of different exemplary situations, going from in vitro bulge inflation experiments on aortic samples to in vivo overpressurizing of patient-specific ATAAs. The simulations reveal that damage correctly starts from the intimal layer (luminal side) and propagates across the media as a tear but never hits the adventitia. This scenario is typically the first stage of development of an acute dissection, which is predicted for pressures of about 2.5 times the diastolic pressure by the model after calibrating the parameters against experimental data performed on collected ATAA samples. Further validations on a larger cohort of patients should hopefully confirm the potential of the model in predicting patient-specific damage evolution and possible risk of dissection during aneurysm growth for clinical applications. Copyright © 2017 John Wiley & Sons, Ltd.
NASA Astrophysics Data System (ADS)
Xiong, Chuan; Shi, Jiancheng
2014-01-01
To date, the light scattering models of snow consider very little about the real snow microstructures. The ideal spherical or other single shaped particle assumptions in previous snow light scattering models can cause error in light scattering modeling of snow and further cause errors in remote sensing inversion algorithms. This paper tries to build up a snow polarized reflectance model based on bicontinuous medium, with which the real snow microstructure is considered. The accurate specific surface area of bicontinuous medium can be analytically derived. The polarized Monte Carlo ray tracing technique is applied to the computer generated bicontinuous medium. With proper algorithms, the snow surface albedo, bidirectional reflectance distribution function (BRDF) and polarized BRDF can be simulated. The validation of model predicted spectral albedo and bidirectional reflectance factor (BRF) using experiment data shows good results. The relationship between snow surface albedo and snow specific surface area (SSA) were predicted, and this relationship can be used for future improvement of snow specific surface area (SSA) inversion algorithms. The model predicted polarized reflectance is validated and proved accurate, which can be further applied in polarized remote sensing.
Posterior Predictive Bayesian Phylogenetic Model Selection
Lewis, Paul O.; Xie, Wangang; Chen, Ming-Hui; Fan, Yu; Kuo, Lynn
2014-01-01
We present two distinctly different posterior predictive approaches to Bayesian phylogenetic model selection and illustrate these methods using examples from green algal protein-coding cpDNA sequences and flowering plant rDNA sequences. The Gelfand–Ghosh (GG) approach allows dissection of an overall measure of model fit into components due to posterior predictive variance (GGp) and goodness-of-fit (GGg), which distinguishes this method from the posterior predictive P-value approach. The conditional predictive ordinate (CPO) method provides a site-specific measure of model fit useful for exploratory analyses and can be combined over sites yielding the log pseudomarginal likelihood (LPML) which is useful as an overall measure of model fit. CPO provides a useful cross-validation approach that is computationally efficient, requiring only a sample from the posterior distribution (no additional simulation is required). Both GG and CPO add new perspectives to Bayesian phylogenetic model selection based on the predictive abilities of models and complement the perspective provided by the marginal likelihood (including Bayes Factor comparisons) based solely on the fit of competing models to observed data. [Bayesian; conditional predictive ordinate; CPO; L-measure; LPML; model selection; phylogenetics; posterior predictive.] PMID:24193892
Koneczny, Jarosław; Czekierdowski, Artur; Florczak, Marek; Poziemski, Paweł; Stachowicz, Norbert; Borowski, Dariusz
2017-01-01
Sonography based methods with various tumor markers are currently used to discriminate the type of adnexal masses. To compare the predictive value of selected sonography-based models along with subjective assessment in ovarian cancer prediction. We analyzed data of 271 women operated because of adnexal masses. All masses were verified by histological examination. Preoperative sonography was performed in all patients and various predictive models includ¬ing IOTA group logistic regression model LR1 (LR1), IOTA simple ultrasound-based rules by IOTA (SR), GI-RADS and risk of malignancy index (RMI3) were used. ROC curves were constructed and respective AUC's with 95% CI's were compared. Of 271 masses 78 proved to be malignant including 6 borderline tumors. LR1 had sensitivity of 91.0%, specificity of 91.2%, AUC = 0.95 (95% CI: 0.92-0.98). Sensitivity for GI-RADS for 271 patients was 88.5% with specificity of 85% and AUC = 0.91 (95% CI: 0.88-0.95). Subjective assessment yielded sensitivity and specificity of 85.9% and 96.9%, respectively with AUC = 0.97 (95% CI: 0.94-0.99). SR were applicable in 236 masses and had sensitivity of 90.6% with specificity of 95.3% and AUC = 0.93 (95% CI 0.89-0.97). RMI3 was calculated only in 104 women who had CA125 available and had sensitivity of 55.3%, specificity of 94% and AUC = 0.85 (95% CI: 0.77-0.93). Although subjective assessment by the ultrasound expert remains the best current method of adnexal tumors preoperative discrimination, the simplicity and high predictive value favor the IOTA SR method, and when not applicable, the IOTA LR1 or GI-RADS models to be primarily and effectively used.
Acosta-Pech, Rocío; Crossa, José; de Los Campos, Gustavo; Teyssèdre, Simon; Claustres, Bruno; Pérez-Elizalde, Sergio; Pérez-Rodríguez, Paulino
2017-07-01
A new genomic model that incorporates genotype × environment interaction gave increased prediction accuracy of untested hybrid response for traits such as percent starch content, percent dry matter content and silage yield of maize hybrids. The prediction of hybrid performance (HP) is very important in agricultural breeding programs. In plant breeding, multi-environment trials play an important role in the selection of important traits, such as stability across environments, grain yield and pest resistance. Environmental conditions modulate gene expression causing genotype × environment interaction (G × E), such that the estimated genetic correlations of the performance of individual lines across environments summarize the joint action of genes and environmental conditions. This article proposes a genomic statistical model that incorporates G × E for general and specific combining ability for predicting the performance of hybrids in environments. The proposed model can also be applied to any other hybrid species with distinct parental pools. In this study, we evaluated the predictive ability of two HP prediction models using a cross-validation approach applied in extensive maize hybrid data, comprising 2724 hybrids derived from 507 dent lines and 24 flint lines, which were evaluated for three traits in 58 environments over 12 years; analyses were performed for each year. On average, genomic models that include the interaction of general and specific combining ability with environments have greater predictive ability than genomic models without interaction with environments (ranging from 12 to 22%, depending on the trait). We concluded that including G × E in the prediction of untested maize hybrids increases the accuracy of genomic models.
Penny, Melissa A; Galactionova, Katya; Tarantino, Michael; Tanner, Marcel; Smith, Thomas A
2015-07-29
The RTS,S/AS01 malaria vaccine candidate recently completed Phase III trials in 11 African sites. Recommendations for its deployment will partly depend on predictions of public health impact in endemic countries. Previous predictions of these used only limited information on underlying vaccine properties and have not considered country-specific contextual data. Each Phase III trial cohort was simulated explicitly using an ensemble of individual-based stochastic models, and many hypothetical vaccine profiles. The true profile was estimated by Bayesian fitting of these models to the site- and time-specific incidence of clinical malaria in both trial arms over 18 months of follow-up. Health impacts of implementation via two vaccine schedules in 43 endemic sub-Saharan African countries, using country-specific prevalence, access to care, immunisation coverage and demography data, were predicted via weighted averaging over many simulations. The efficacy against infection of three doses of vaccine was initially approximately 65 % (when immunising 6-12 week old infants) and 80 % (children 5-17 months old), with a 1 year half-life (exponential decay). Either schedule will avert substantial disease, but predicted impact strongly depends on the decay rate of vaccine effects and average transmission intensity. For the first time Phase III site- and time-specific data were available to estimate both the underlying profile of RTS,S/AS01 and likely country-specific health impacts. Initial efficacy will probably be high, but decay rapidly. Adding RTS,S to existing control programs, assuming continuation of current levels of malaria exposure and of health system performance, will potentially avert 100-580 malaria deaths and 45,000 to 80,000 clinical episodes per 100,000 fully vaccinated children over an initial 10-year phase.
Predicting human chronically paralyzed muscle force: a comparison of three mathematical models.
Frey Law, Laura A; Shields, Richard K
2006-03-01
Chronic spinal cord injury (SCI) induces detrimental musculoskeletal adaptations that adversely affect health status, ranging from muscle paralysis and skin ulcerations to osteoporosis. SCI rehabilitative efforts may increasingly focus on preserving the integrity of paralyzed extremities to maximize health quality using electrical stimulation for isometric training and/or functional activities. Subject-specific mathematical muscle models could prove valuable for predicting the forces necessary to achieve therapeutic loading conditions in individuals with paralyzed limbs. Although numerous muscle models are available, three modeling approaches were chosen that can accommodate a variety of stimulation input patterns. To our knowledge, no direct comparisons between models using paralyzed muscle have been reported. The three models include 1) a simple second-order linear model with three parameters and 2) two six-parameter nonlinear models (a second-order nonlinear model and a Hill-derived nonlinear model). Soleus muscle forces from four individuals with complete, chronic SCI were used to optimize each model's parameters (using an increasing and decreasing frequency ramp) and to assess the models' predictive accuracies for constant and variable (doublet) stimulation trains at 5, 10, and 20 Hz in each individual. Despite the large differences in modeling approaches, the mean predicted force errors differed only moderately (8-15% error; P=0.0042), suggesting physiological force can be adequately represented by multiple mathematical constructs. The two nonlinear models predicted specific force characteristics better than the linear model in nearly all stimulation conditions, with minimal differences between the two nonlinear models. Either nonlinear mathematical model can provide reasonable force estimates; individual application needs may dictate the preferred modeling strategy.
Caprariello, Peter A.; Cuddy, Amy J. C.; Fiske, Susan T.
2013-01-01
The stereotype content model (SCM) posits that social structure predicts specific cultural stereotypes and associated emotional prejudices. No prior evidence at a societal level has manipulated both structural predictors and measured both stereotypes and prejudices. In the present study, participants (n = 120) responded to an immigration scenario depicting a high- or low-status group, competitive or not competitive, and rated their likely stereotype (on warmth and competence) and elicited emotional prejudices (admiration, contempt, envy, and pity). Seven of eight specific predictions are fully confirmed, supporting the SCM's predicted causality for social structural effects on cultural stereotypes and emotional prejudices. PMID:24285928
ADOT state-specific crash prediction models : an Arizona needs study.
DOT National Transportation Integrated Search
2016-12-01
The predictive method in the Highway Safety Manual (HSM) includes a safety performance function (SPF), : crash modification factors (CMFs), and a local calibration factor (C), if available. Two alternatives exist for : applying the HSM prediction met...
Forecasting peak asthma admissions in London: an application of quantile regression models.
Soyiri, Ireneous N; Reidpath, Daniel D; Sarran, Christophe
2013-07-01
Asthma is a chronic condition of great public health concern globally. The associated morbidity, mortality and healthcare utilisation place an enormous burden on healthcare infrastructure and services. This study demonstrates a multistage quantile regression approach to predicting excess demand for health care services in the form of asthma daily admissions in London, using retrospective data from the Hospital Episode Statistics, weather and air quality. Trivariate quantile regression models (QRM) of asthma daily admissions were fitted to a 14-day range of lags of environmental factors, accounting for seasonality in a hold-in sample of the data. Representative lags were pooled to form multivariate predictive models, selected through a systematic backward stepwise reduction approach. Models were cross-validated using a hold-out sample of the data, and their respective root mean square error measures, sensitivity, specificity and predictive values compared. Two of the predictive models were able to detect extreme number of daily asthma admissions at sensitivity levels of 76 % and 62 %, as well as specificities of 66 % and 76 %. Their positive predictive values were slightly higher for the hold-out sample (29 % and 28 %) than for the hold-in model development sample (16 % and 18 %). QRMs can be used in multistage to select suitable variables to forecast extreme asthma events. The associations between asthma and environmental factors, including temperature, ozone and carbon monoxide can be exploited in predicting future events using QRMs.
Forecasting peak asthma admissions in London: an application of quantile regression models
NASA Astrophysics Data System (ADS)
Soyiri, Ireneous N.; Reidpath, Daniel D.; Sarran, Christophe
2013-07-01
Asthma is a chronic condition of great public health concern globally. The associated morbidity, mortality and healthcare utilisation place an enormous burden on healthcare infrastructure and services. This study demonstrates a multistage quantile regression approach to predicting excess demand for health care services in the form of asthma daily admissions in London, using retrospective data from the Hospital Episode Statistics, weather and air quality. Trivariate quantile regression models (QRM) of asthma daily admissions were fitted to a 14-day range of lags of environmental factors, accounting for seasonality in a hold-in sample of the data. Representative lags were pooled to form multivariate predictive models, selected through a systematic backward stepwise reduction approach. Models were cross-validated using a hold-out sample of the data, and their respective root mean square error measures, sensitivity, specificity and predictive values compared. Two of the predictive models were able to detect extreme number of daily asthma admissions at sensitivity levels of 76 % and 62 %, as well as specificities of 66 % and 76 %. Their positive predictive values were slightly higher for the hold-out sample (29 % and 28 %) than for the hold-in model development sample (16 % and 18 %). QRMs can be used in multistage to select suitable variables to forecast extreme asthma events. The associations between asthma and environmental factors, including temperature, ozone and carbon monoxide can be exploited in predicting future events using QRMs.
Wang, Li-Ying; Zheng, Shu-Sen; Xu, Xiao; Wang, Wei-Lin; Wu, Jian; Zhang, Min; Shen, Yan; Yan, Sheng; Xie, Hai-Yang; Chen, Xin-Hua; Jiang, Tian-An; Chen, Fen
2015-02-01
The prognostic prediction of liver transplantation (LT) guides the donor organ allocation. However, there is currently no satisfactory model to predict the recipients' outcome, especially for the patients with HBV cirrhosis-related hepatocellular carcinoma (HCC). The present study was to develop a quantitative assessment model for predicting the post-LT survival in HBV-related HCC patients. Two hundred and thirty-eight LT recipients at the Liver Transplant Center, First Affiliated Hospital, Zhejiang University School of Medicine between 2008 and 2013 were included in this study. Their post-LT prognosis was recorded and multiple risk factors were analyzed using univariate and multivariate analyses in Cox regression. The score model was as follows: 0.114X(Child-Pugh score)-0.002X(positive HBV DNA detection time)+0.647X(number of tumor nodules)+0.055X(max diameter of tumor nodules)+0.231XlnAFP+0.437X(tumor differentiation grade). The receiver operating characteristic curve analysis showed that the area under the curve of the scoring model for predicting the post-LT survival was 0.887. The cut-off value was 1.27, which was associated with a sensitivity of 72.5% and a specificity of 90.7%, respectively. The quantitative score model for predicting post-LT survival proved to be sensitive and specific.
Characterization of Used Nuclear Fuel with Multivariate Analysis for Process Monitoring
DOE Office of Scientific and Technical Information (OSTI.GOV)
Dayman, Kenneth J.; Coble, Jamie B.; Orton, Christopher R.
2014-01-01
The Multi-Isotope Process (MIP) Monitor combines gamma spectroscopy and multivariate analysis to detect anomalies in various process streams in a nuclear fuel reprocessing system. Measured spectra are compared to models of nominal behavior at each measurement location to detect unexpected changes in system behavior. In order to improve the accuracy and specificity of process monitoring, fuel characterization may be used to more accurately train subsequent models in a full analysis scheme. This paper presents initial development of a reactor-type classifier that is used to select a reactor-specific partial least squares model to predict fuel burnup. Nuclide activities for prototypic usedmore » fuel samples were generated in ORIGEN-ARP and used to investigate techniques to characterize used nuclear fuel in terms of reactor type (pressurized or boiling water reactor) and burnup. A variety of reactor type classification algorithms, including k-nearest neighbors, linear and quadratic discriminant analyses, and support vector machines, were evaluated to differentiate used fuel from pressurized and boiling water reactors. Then, reactor type-specific partial least squares models were developed to predict the burnup of the fuel. Using these reactor type-specific models instead of a model trained for all light water reactors improved the accuracy of burnup predictions. The developed classification and prediction models were combined and applied to a large dataset that included eight fuel assembly designs, two of which were not used in training the models, and spanned the range of the initial 235U enrichment, cooling time, and burnup values expected of future commercial used fuel for reprocessing. Error rates were consistent across the range of considered enrichment, cooling time, and burnup values. Average absolute relative errors in burnup predictions for validation data both within and outside the training space were 0.0574% and 0.0597%, respectively. The errors seen in this work are artificially low, because the models were trained, optimized, and tested on simulated, noise-free data. However, these results indicate that the developed models may generalize well to new data and that the proposed approach constitutes a viable first step in developing a fuel characterization algorithm based on gamma spectra.« less
Kindergarten Predictors of Math Learning Disability
Mazzocco, Michèle M. M.; Thompson, Richard E.
2009-01-01
The aim of the present study was to address how to effectively predict mathematics learning disability (MLD). Specifically, we addressed whether cognitive data obtained during kindergarten can effectively predict which children will have MLD in third grade, whether an abbreviated test battery could be as effective as a standard psychoeducational assessment at predicting MLD, and whether the abbreviated battery corresponded to the literature on MLD characteristics. Participants were 226 children who enrolled in a 4-year prospective longitudinal study during kindergarten. We administered measures of mathematics achievement, formal and informal mathematics ability, visual-spatial reasoning, and rapid automatized naming and examined which test scores and test items from kindergarten best predicted MLD at grades 2 and 3. Statistical models using standardized scores from the entire test battery correctly classified ~80–83 percent of the participants as having, or not having, MLD. Regression models using scores from only individual test items were less predictive than models containing the standard scores, except for models using a specific subset of test items that dealt with reading numerals, number constancy, magnitude judgments of one-digit numbers, or mental addition of one-digit numbers. These models were as accurate in predicting MLD as was the model including the entire set of standard scores from the battery of tests examined. Our findings indicate that it is possible to effectively predict which kindergartners are at risk for MLD, and thus the findings have implications for early screening of MLD. PMID:20084182
Deter, Russell L.; Lee, Wesley; Yeo, Lami; Romero, Roberto
2012-01-01
Objectives To characterize 2nd and 3rd trimester fetal growth using Individualized Growth Assessment in a large cohort of fetuses with normal growth outcomes. Methods A prospective longitudinal study of 119 pregnancies was carried out from 18 weeks, MA, to delivery. Measurements of eleven fetal growth parameters were obtained from 3D scans at 3–4 week intervals. Regression analyses were used to determine Start Points [SP] and Rossavik model [P = c (t) k + st] coefficients c, k and s for each parameter in each fetus. Second trimester growth model specification functions were re-established. These functions were used to generate individual growth models and determine predicted s and s-residual [s = pred s + s-resid] values. Actual measurements were compared to predicted growth trajectories obtained from the growth models and Percent Deviations [% Dev = {{actual − predicted}/predicted} × 100] calculated. Age-specific reference standards for this statistic were defined using 2-level statistical modeling for the nine directly measured parameters and estimated weight. Results Rossavik models fit the data for all parameters very well [R2: 99%], with SP’s and k values similar to those found in a much smaller cohort. The c values were strongly related to the 2nd trimester slope [R2: 97%] as was predicted s to estimated c [R2: 95%]. The latter was negative for skeletal parameters and positive for soft tissue parameters. The s-residuals were unrelated to estimated c’s [R2: 0%], and had mean values of zero. Rossavik models predicted 3rd trimester growth with systematic errors close to 0% and random errors [95% range] of 5.7 – 10.9% and 20.0 – 24.3% for one and three dimensional parameters, respectively. Moderate changes in age-specific variability were seen in the 3rd trimester.. Conclusions IGA procedures for evaluating 2nd and 3rd trimester growth are now established based on a large cohort [4–6 fold larger than those used previously], thus permitting more reliable growth assessment with each fetus acting as its own control. New, more rigorously defined, age-specific standards for the evaluation of 3rd trimester growth deviations are now available for 10 anatomical parameters. Our results are also consistent with the predicted s and s-residual being representatives of growth controllers operating through the insulin-like growth factor [IGF] axis. PMID:23962305
Bridging the gap between computation and clinical biology: validation of cable theory in humans
Finlay, Malcolm C.; Xu, Lei; Taggart, Peter; Hanson, Ben; Lambiase, Pier D.
2013-01-01
Introduction: Computerized simulations of cardiac activity have significantly contributed to our understanding of cardiac electrophysiology, but techniques of simulations based on patient-acquired data remain in their infancy. We sought to integrate data acquired from human electrophysiological studies into patient-specific models, and validated this approach by testing whether electrophysiological responses to sequential premature stimuli could be predicted in a quantitatively accurate manner. Methods: Eleven patients with structurally normal hearts underwent electrophysiological studies. Semi-automated analysis was used to reconstruct activation and repolarization dynamics for each electrode. This S2 extrastimuli data was used to inform individualized models of cardiac conduction, including a novel derivation of conduction velocity restitution. Activation dynamics of multiple premature extrastimuli were then predicted from this model and compared against measured patient data as well as data derived from the ten-Tusscher cell-ionic model. Results: Activation dynamics following a premature S3 were significantly different from those after an S2. Patient specific models demonstrated accurate prediction of the S3 activation wave, (Pearson's R2 = 0.90, median error 4%). Examination of the modeled conduction dynamics allowed inferences into the spatial dispersion of activation delay. Further validation was performed against data from the ten-Tusscher cell-ionic model, with our model accurately recapitulating predictions of repolarization times (R2 = 0.99). Conclusions: Simulations based on clinically acquired data can be used to successfully predict complex activation patterns following sequential extrastimuli. Such modeling techniques may be useful as a method of incorporation of clinical data into predictive models. PMID:24027527
2013-01-01
Background The present study aimed to develop an artificial neural network (ANN) based prediction model for cardiovascular autonomic (CA) dysfunction in the general population. Methods We analyzed a previous dataset based on a population sample consisted of 2,092 individuals aged 30–80 years. The prediction models were derived from an exploratory set using ANN analysis. Performances of these prediction models were evaluated in the validation set. Results Univariate analysis indicated that 14 risk factors showed statistically significant association with CA dysfunction (P < 0.05). The mean area under the receiver-operating curve was 0.762 (95% CI 0.732–0.793) for prediction model developed using ANN analysis. The mean sensitivity, specificity, positive and negative predictive values were similar in the prediction models was 0.751, 0.665, 0.330 and 0.924, respectively. All HL statistics were less than 15.0. Conclusion ANN is an effective tool for developing prediction models with high value for predicting CA dysfunction among the general population. PMID:23902963
Burkhart, Katelyn A; Bruno, Alexander G; Bouxsein, Mary L; Bean, Jonathan F; Anderson, Dennis E
2018-01-01
Maximum muscle stress (MMS) is a critical parameter in musculoskeletal modeling, defining the maximum force that a muscle of given size can produce. However, a wide range of MMS values have been reported in literature, and few studies have estimated MMS in trunk muscles. Due to widespread use of musculoskeletal models in studies of the spine and trunk, there is a need to determine reasonable magnitude and range of trunk MMS. We measured trunk extension strength in 49 participants over 65 years of age, surveyed participants about low back pain, and acquired quantitative computed tomography (QCT) scans of their lumbar spines. Trunk muscle morphology was assessed from QCT scans and used to create a subject-specific musculoskeletal model for each participant. Model-predicted extension strength was computed using a trunk muscle MMS of 100 N/cm 2 . The MMS of each subject-specific model was then adjusted until the measured strength matched the model-predicted strength (±20 N). We found that measured trunk extension strength was significantly higher in men. With the initial constant MMS value, the musculoskeletal model generally over-predicted trunk extension strength. By adjusting MMS on a subject-specific basis, we found apparent MMS values ranging from 40 to 130 N/cm 2 , with an average of 75.5 N/cm 2 for both men and women. Subjects with low back pain had lower apparent MMS than subjects with no back pain. This work incorporates a unique approach to estimate subject-specific trunk MMS values via musculoskeletal modeling and provides a useful insight into MMS variation. © 2017 Orthopaedic Research Society. Published by Wiley Periodicals, Inc. J Orthop Res 36:498-505, 2018. © 2017 Orthopaedic Research Society. Published by Wiley Periodicals, Inc.
Adapting Price Predictions in TAC SCM
NASA Astrophysics Data System (ADS)
Pardoe, David; Stone, Peter
In agent-based markets, adapting to the behavior of other agents is often necessary for success. When it is not possible to directly model individual competitors, an agent may instead model and adapt to the market conditions that result from competitor behavior. Such an agent could still benefit from reasoning about specific competitor strategies by considering how various combinations of these strategies would impact the conditions being modeled. We present an application of such an approach to a specific prediction problem faced by the agent TacTex-06 in the Trading Agent Competition's Supply Chain Management scenario (TAC SCM).
The capability of physiologically-based pharmacokinetic (PBPK) models to incorporate ageappropriate physiological and chemical-specific parameters was utilized in this study to predict changes in internal dosimetry for six volatile organic compounds (VOCs) across different ages o...
Network model for thermal conductivities of unidirectional fiber-reinforced composites
NASA Astrophysics Data System (ADS)
Wang, Yang; Peng, Chaoyi; Zhang, Weihua
2014-12-01
An empirical network model has been developed to predict the in-plane thermal conductivities along arbitrary directions for unidirectional fiber-reinforced composites lamina. Measurements of thermal conductivities along different orientations were carried out. Good agreement was observed between values predicted by the network model and the experimental data; compared with the established analytical models, the newly proposed network model could give values with higher precision. Therefore, this network model is helpful to get a wider and more comprehensive understanding of heat transmission characteristics of fiber-reinforced composites and can be utilized as guidance to design and fabricate laminated composites with specific directional or specific locational thermal conductivities for structures that simultaneously perform mechanical and thermal functions, i.e. multifunctional structures (MFS).
The Lag Model, a Turbulence Model for Wall Bounded Flows Including Separation
NASA Technical Reports Server (NTRS)
Olsen, Michael E.; Coakley, Thomas J.; Kwak, Dochan (Technical Monitor)
2001-01-01
A new class of turbulence model is described for wall bounded, high Reynolds number flows. A specific turbulence model is demonstrated, with results for favorable and adverse pressure gradient flowfields. Separation predictions are as good or better than either Spalart Almaras or SST models, do not require specification of wall distance, and have similar or reduced computational effort compared with these models.
ERIC Educational Resources Information Center
Lippke, Sonia; Nigg, Claudio R.; Maddock, Jay E.
2007-01-01
This is the first study to test whether the stages of change of the transtheoretical model are qualitatively different through exploring discontinuity patterns in theory of planned behavior (TPB) variables using latent multigroup structural equation modeling (MSEM) with AMOS. Discontinuity patterns in terms of latent means and prediction patterns…
Youkhana, Adel H.; Ogoshi, Richard M.; Kiniry, James R.; ...
2017-05-02
Biomass is a promising renewable energy option that provides a more environmentally sustainable alternative to fossil resources by reducing the net flux of greenhouse gasses to the atmosphere. Yet, allometric models that allow the prediction of aboveground biomass (AGB), biomass carbon (C) stock non-destructively have not yet been developed for tropical perennial C 4 grasses currently under consideration as potential bioenergy feedstock in Hawaii and other subtropical and tropical locations. The objectives of this study were to develop optimal allometric relationships and site-specific models to predict AGB, biomass C stock of napiergrass, energycane, and sugarcane under cultivation practices for renewablemore » energy and validate these site-specific models against independent data sets generated from sites with widely different environments. Several allometric models were developed for each species from data at a low elevation field on the island of Maui, Hawaii. A simple power model with stalk diameter (D) was best related to AGB and biomass C stock for napiergrass, energycane, and sugarcane, (R 2 = 0.98, 0.96, and 0.97, respectively). The models were then tested against data collected from independent fields across an environmental gradient. For all crops, the models over-predicted AGB in plants with lower stalk D, but AGB was under-predicted in plants with higher stalk D. The models using stalk D were better for biomass prediction compared to dewlap H (Height from the base cut to most recently exposed leaf dewlap) models, which showed weak validation performance. Although stalk D model performed better, however, the mean square error (MSE)-systematic was ranged from 23 to 43 % of MSE for all crops. A strong relationship between model coefficient and rainfall was existed, although these were irrigated systems; suggesting a simple site-specific coefficient modulator for rainfall to reduce systematic errors in water-limited areas. These allometric equations provide a tool for farmers in the tropics to estimate perennial C4 grass biomass and C stock during decision-making for land management and as an environmental sustainability indicator within a renewable energy system.« less
Bianchi, Lorenzo; Schiavina, Riccardo; Borghesi, Marco; Bianchi, Federico Mineo; Briganti, Alberto; Carini, Marco; Terrone, Carlo; Mottrie, Alex; Gacci, Mauro; Gontero, Paolo; Imbimbo, Ciro; Marchioro, Giansilvio; Milanese, Giulio; Mirone, Vincenzo; Montorsi, Francesco; Morgia, Giuseppe; Novara, Giacomo; Porreca, Angelo; Volpe, Alessandro; Brunocilla, Eugenio
2018-04-06
To assess the predictive accuracy and the clinical value of a recent nomogram predicting cancer-specific mortality-free survival after surgery in pN1 prostate cancer patients through an external validation. We evaluated 518 prostate cancer patients treated with radical prostatectomy and pelvic lymph node dissection with evidence of nodal metastases at final pathology, at 10 tertiary centers. External validation was carried out using regression coefficients of the previously published nomogram. The performance characteristics of the model were assessed by quantifying predictive accuracy, according to the area under the curve in the receiver operating characteristic curve and model calibration. Furthermore, we systematically analyzed the specificity, sensitivity, positive predictive value and negative predictive value for each nomogram-derived probability cut-off. Finally, we implemented decision curve analysis, in order to quantify the nomogram's clinical value in routine practice. External validation showed inferior predictive accuracy as referred to in the internal validation (65.8% vs 83.3%, respectively). The discrimination (area under the curve) of the multivariable model was 66.7% (95% CI 60.1-73.0%) by testing with receiver operating characteristic curve analysis. The calibration plot showed an overestimation throughout the range of predicted cancer-specific mortality-free survival rates probabilities. However, in decision curve analysis, the nomogram's use showed a net benefit when compared with the scenarios of treating all patients or none. In an external setting, the nomogram showed inferior predictive accuracy and suboptimal calibration characteristics as compared to that reported in the original population. However, decision curve analysis showed a clinical net benefit, suggesting a clinical implication to correctly manage pN1 prostate cancer patients after surgery. © 2018 The Japanese Urological Association.
Coiera, Enrico; Wang, Ying; Magrabi, Farah; Concha, Oscar Perez; Gallego, Blanca; Runciman, William
2014-05-21
Current prognostic models factor in patient and disease specific variables but do not consider cumulative risks of hospitalization over time. We developed risk models of the likelihood of death associated with cumulative exposure to hospitalization, based on time-varying risks of hospitalization over any given day, as well as day of the week. Model performance was evaluated alone, and in combination with simple disease-specific models. Patients admitted between 2000 and 2006 from 501 public and private hospitals in NSW, Australia were used for training and 2007 data for evaluation. The impact of hospital care delivered over different days of the week and or times of the day was modeled by separating hospitalization risk into 21 separate time periods (morning, day, night across the days of the week). Three models were developed to predict death up to 7-days post-discharge: 1/a simple background risk model using age, gender; 2/a time-varying risk model for exposure to hospitalization (admission time, days in hospital); 3/disease specific models (Charlson co-morbidity index, DRG). Combining these three generated a full model. Models were evaluated by accuracy, AUC, Akaike and Bayesian information criteria. There was a clear diurnal rhythm to hospital mortality in the data set, peaking in the evening, as well as the well-known 'weekend-effect' where mortality peaks with weekend admissions. Individual models had modest performance on the test data set (AUC 0.71, 0.79 and 0.79 respectively). The combined model which included time-varying risk however yielded an average AUC of 0.92. This model performed best for stays up to 7-days (93% of admissions), peaking at days 3 to 5 (AUC 0.94). Risks of hospitalization vary not just with the day of the week but also time of the day, and can be used to make predictions about the cumulative risk of death associated with an individual's hospitalization. Combining disease specific models with such time varying- estimates appears to result in robust predictive performance. Such risk exposure models should find utility both in enhancing standard prognostic models as well as estimating the risk of continuation of hospitalization.
Maizlin, Ilan I; Redden, David T; Beierle, Elizabeth A; Chen, Mike K; Russell, Robert T
2017-04-01
Surgical wound classification, introduced in 1964, stratifies the risk of surgical site infection (SSI) based on a clinical estimate of the inoculum of bacteria encountered during the procedure. Recent literature has questioned the accuracy of predicting SSI risk based on wound classification. We hypothesized that a more specific model founded on specific patient and perioperative factors would more accurately predict the risk of SSI. Using all observations from the 2012 to 2014 pediatric National Surgical Quality Improvement Program-Pediatric (NSQIP-P) Participant Use File, patients were randomized into model creation and model validation datasets. Potential perioperative predictive factors were assessed with univariate analysis for each of 4 outcomes: wound dehiscence, superficial wound infection, deep wound infection, and organ space infection. A multiple logistic regression model with a step-wise backwards elimination was performed. A receiver operating characteristic curve with c-statistic was generated to assess the model discrimination for each outcome. A total of 183,233 patients were included. All perioperative NSQIP factors were evaluated for clinical pertinence. Of the original 43 perioperative predictive factors selected, 6 to 9 predictors for each outcome were significantly associated with postoperative SSI. The predictive accuracy level of our model compared favorably with the traditional wound classification in each outcome of interest. The proposed model from NSQIP-P demonstrated a significantly improved predictive ability for postoperative SSIs than the current wound classification system. This model will allow providers to more effectively counsel families and patients of these risks, and more accurately reflect true risks for individual surgical patients to hospitals and payers. Copyright © 2017 American College of Surgeons. Published by Elsevier Inc. All rights reserved.
Mulhearn, Tyler J; Watts, Logan L; Todd, E Michelle; Medeiros, Kelsey E; Connelly, Shane; Mumford, Michael D
2017-01-01
Although recent evidence suggests ethics education can be effective, the nature of specific training programs, and their effectiveness, varies considerably. Building on a recent path modeling effort, the present study developed and validated a predictive modeling tool for responsible conduct of research education. The predictive modeling tool allows users to enter ratings in relation to a given ethics training program and receive instantaneous evaluative information for course refinement. Validation work suggests the tool's predicted outcomes correlate strongly (r = 0.46) with objective course outcomes. Implications for training program development and refinement are discussed.
A Theory of Age-Dependent Mutation and Senescence
Moorad, Jacob A.; Promislow, Daniel E. L.
2008-01-01
Laboratory experiments show us that the deleterious character of accumulated novel age-specific mutations is reduced and made less variable with increased age. While theories of aging predict that the frequency of deleterious mutations at mutation–selection equilibrium will increase with the mutation's age of effect, they do not account for these age-related changes in the distribution of de novo mutational effects. Furthermore, no model predicts why this dependence of mutational effects upon age exists. Because the nature of mutational distributions plays a critical role in shaping patterns of senescence, we need to develop aging theory that explains and incorporates these effects. Here we propose a model that explains the age dependency of mutational effects by extending Fisher's geometrical model of adaptation to include a temporal dimension. Using a combination of simple analytical arguments and simulations, we show that our model predicts age-specific mutational distributions that are consistent with observations from mutation-accumulation experiments. Simulations show us that these age-specific mutational effects may generate patterns of senescence at mutation–selection equilibrium that are consistent with observed demographic patterns that are otherwise difficult to explain. PMID:18660535
DOE Office of Scientific and Technical Information (OSTI.GOV)
Draxl, C.; Churchfield, M.; Mirocha, J.
Wind plant aerodynamics are influenced by a combination of microscale and mesoscale phenomena. Incorporating mesoscale atmospheric forcing (e.g., diurnal cycles and frontal passages) into wind plant simulations can lead to a more accurate representation of microscale flows, aerodynamics, and wind turbine/plant performance. Our goal is to couple a numerical weather prediction model that can represent mesoscale flow [specifically the Weather Research and Forecasting model] with a microscale LES model (OpenFOAM) that can predict microscale turbulence and wake losses.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Youkhana, Adel H.; Ogoshi, Richard M.; Kiniry, James R.
Biomass is a promising renewable energy option that provides a more environmentally sustainable alternative to fossil resources by reducing the net flux of greenhouse gasses to the atmosphere. Yet, allometric models that allow the prediction of aboveground biomass (AGB), biomass carbon (C) stock non-destructively have not yet been developed for tropical perennial C 4 grasses currently under consideration as potential bioenergy feedstock in Hawaii and other subtropical and tropical locations. The objectives of this study were to develop optimal allometric relationships and site-specific models to predict AGB, biomass C stock of napiergrass, energycane, and sugarcane under cultivation practices for renewablemore » energy and validate these site-specific models against independent data sets generated from sites with widely different environments. Several allometric models were developed for each species from data at a low elevation field on the island of Maui, Hawaii. A simple power model with stalk diameter (D) was best related to AGB and biomass C stock for napiergrass, energycane, and sugarcane, (R 2 = 0.98, 0.96, and 0.97, respectively). The models were then tested against data collected from independent fields across an environmental gradient. For all crops, the models over-predicted AGB in plants with lower stalk D, but AGB was under-predicted in plants with higher stalk D. The models using stalk D were better for biomass prediction compared to dewlap H (Height from the base cut to most recently exposed leaf dewlap) models, which showed weak validation performance. Although stalk D model performed better, however, the mean square error (MSE)-systematic was ranged from 23 to 43 % of MSE for all crops. A strong relationship between model coefficient and rainfall was existed, although these were irrigated systems; suggesting a simple site-specific coefficient modulator for rainfall to reduce systematic errors in water-limited areas. These allometric equations provide a tool for farmers in the tropics to estimate perennial C4 grass biomass and C stock during decision-making for land management and as an environmental sustainability indicator within a renewable energy system.« less
De Carli, Margherita M; Baccarelli, Andrea A; Trevisi, Letizia; Pantic, Ivan; Brennan, Kasey JM; Hacker, Michele R; Loudon, Holly; Brunst, Kelly J; Wright, Robert O; Wright, Rosalind J; Just, Allan C
2017-01-01
Aim: We compared predictive modeling approaches to estimate placental methylation using cord blood methylation. Materials & methods: We performed locus-specific methylation prediction using both linear regression and support vector machine models with 174 matched pairs of 450k arrays. Results: At most CpG sites, both approaches gave poor predictions in spite of a misleading improvement in array-wide correlation. CpG islands and gene promoters, but not enhancers, were the genomic contexts where the correlation between measured and predicted placental methylation levels achieved higher values. We provide a list of 714 sites where both models achieved an R2 ≥0.75. Conclusion: The present study indicates the need for caution in interpreting cross-tissue predictions. Few methylation sites can be predicted between cord blood and placenta. PMID:28234020
Creep fatigue life prediction for engine hot section materials (ISOTROPIC)
NASA Technical Reports Server (NTRS)
Nelson, R. S.; Schoendorf, J. F.; Lin, L. S.
1986-01-01
The specific activities summarized include: verification experiments (base program); thermomechanical cycling model; multiaxial stress state model; cumulative loading model; screening of potential environmental and protective coating models; and environmental attack model.
Hochard, Kevin D; Heym, Nadja; Townsend, Ellen
2017-06-01
Heightened arousal significantly interacts with acquired capability to predict suicidality. We explore this interaction with insomnia and nightmares independently of waking state arousal symptoms, and test predictions of the Interpersonal Theory of Suicide (IPTS) and Escape Theory in relation to these sleep arousal symptoms. Findings from our e-survey (n = 540) supported the IPTS over models of Suicide as Escape. Sleep-specific measurements of arousal (insomnia and nightmares) showed no main effect, yet interacted with acquired capability to predict increased suicidality. The explained variance in suicidality by the interaction (1%-2%) using sleep-specific measures was comparable to variance explained by interactions previously reported in the literature using measurements composed of a mix of waking and sleep state arousal symptoms. Similarly, when entrapment (inability to escape) was included in models, main effects of sleep symptoms arousal were not detected yet interacted with entrapment to predict suicidality. We discuss findings in relation to treatment options suggesting that sleep-specific interventions be considered for the long-term management of at-risk individuals. © 2016 The American Association of Suicidology.
NASA Astrophysics Data System (ADS)
Davenport, F., IV; Harrison, L.; Shukla, S.; Husak, G. J.; Funk, C. C.
2017-12-01
We evaluate the predictive accuracy of an ensemble of empirical model specifications that use earth observation data to predict sub-national grain yields in Mexico and East Africa. Products that are actively used for seasonal drought monitoring are tested as yield predictors. Our research is driven by the fact that East Africa is a region where decisions regarding agricultural production are critical to preventing the loss of economic livelihoods and human life. Regional grain yield forecasts can be used to anticipate availability and prices of key staples, which can turn can inform decisions about targeting humanitarian response such as food aid. Our objective is to identify-for a given region, grain, and time year- what type of model and/or earth observation can most accurately predict end of season yields. We fit a set of models to county level panel data from Mexico, Kenya, Sudan, South Sudan, and Somalia. We then examine out of sample predicative accuracy using various linear and non-linear models that incorporate spatial and time varying coefficients. We compare accuracy within and across models that use predictor variables from remotely sensed measures of precipitation, temperature, soil moisture, and other land surface processes. We also examine at what point in the season a given model or product is most useful for determining predictive accuracy. Finally we compare predictive accuracy across a variety of agricultural regimes including high intensity irrigated commercial agricultural and rain fed subsistence level farms.
Munteanu, Cristian R; Gonzalez-Diaz, Humberto; Garcia, Rafael; Loza, Mabel; Pazos, Alejandro
2015-01-01
The molecular information encoding into molecular descriptors is the first step into in silico Chemoinformatics methods in Drug Design. The Machine Learning methods are a complex solution to find prediction models for specific biological properties of molecules. These models connect the molecular structure information such as atom connectivity (molecular graphs) or physical-chemical properties of an atom/group of atoms to the molecular activity (Quantitative Structure - Activity Relationship, QSAR). Due to the complexity of the proteins, the prediction of their activity is a complicated task and the interpretation of the models is more difficult. The current review presents a series of 11 prediction models for proteins, implemented as free Web tools on an Artificial Intelligence Model Server in Biosciences, Bio-AIMS (http://bio-aims.udc.es/TargetPred.php). Six tools predict protein activity, two models evaluate drug - protein target interactions and the other three calculate protein - protein interactions. The input information is based on the protein 3D structure for nine models, 1D peptide amino acid sequence for three tools and drug SMILES formulas for two servers. The molecular graph descriptor-based Machine Learning models could be useful tools for in silico screening of new peptides/proteins as future drug targets for specific treatments.
A Field Synopsis of Sex in Clinical Prediction Models for Cardiovascular Disease
Paulus, Jessica K.; Wessler, Benjamin S.; Lundquist, Christine; Lai, Lana L.Y.; Raman, Gowri; Lutz, Jennifer S.; Kent, David M.
2017-01-01
Background Several widely-used risk scores for cardiovascular disease (CVD) incorporate sex effects, yet there has been no systematic summary of the role of sex in clinical prediction models (CPMs). To better understand the potential of these models to support sex-specific care, we conducted a field synopsis of sex effects in CPMs for CVD. Methods and Results We identified CPMs in the Tufts Predictive Analytics and Comparative Effectiveness (PACE) CPM Registry, a comprehensive database of CVD CPMs published from 1/1990–5/2012. We report the proportion of models including sex effects on CVD incidence or prognosis, summarize the directionality of the predictive effects of sex, and explore factors influencing the inclusion of sex. Of 592 CVD-related CPMs, 193 (33%) included sex as a predictor or presented sex-stratified models. Sex effects were included in 78% (53/68) of models predicting incidence of CVD in a general population, versus only 35% (59/171), 21% (12/58) and 17% (12/72) of models predicting outcomes in patients with coronary artery disease (CAD), stroke, and heart failure, respectively. Among sex-including CPMs, women with heart failure were at lower mortality risk in 8/8 models; women undergoing revascularization for CAD were at higher mortality risk in 10/12 models. Factors associated with the inclusion of sex effects included the number of outcome events and using cohorts at-risk for CVD (rather than with established CVD). Conclusions While CPMs hold promise for supporting sex-specific decision making in CVD clinical care, sex effects are included in only one third of published CPMs. PMID:26908865
Developing a clinical utility framework to evaluate prediction models in radiogenomics
NASA Astrophysics Data System (ADS)
Wu, Yirong; Liu, Jie; Munoz del Rio, Alejandro; Page, David C.; Alagoz, Oguzhan; Peissig, Peggy; Onitilo, Adedayo A.; Burnside, Elizabeth S.
2015-03-01
Combining imaging and genetic information to predict disease presence and behavior is being codified into an emerging discipline called "radiogenomics." Optimal evaluation methodologies for radiogenomics techniques have not been established. We aim to develop a clinical decision framework based on utility analysis to assess prediction models for breast cancer. Our data comes from a retrospective case-control study, collecting Gail model risk factors, genetic variants (single nucleotide polymorphisms-SNPs), and mammographic features in Breast Imaging Reporting and Data System (BI-RADS) lexicon. We first constructed three logistic regression models built on different sets of predictive features: (1) Gail, (2) Gail+SNP, and (3) Gail+SNP+BI-RADS. Then, we generated ROC curves for three models. After we assigned utility values for each category of findings (true negative, false positive, false negative and true positive), we pursued optimal operating points on ROC curves to achieve maximum expected utility (MEU) of breast cancer diagnosis. We used McNemar's test to compare the predictive performance of the three models. We found that SNPs and BI-RADS features augmented the baseline Gail model in terms of the area under ROC curve (AUC) and MEU. SNPs improved sensitivity of the Gail model (0.276 vs. 0.147) and reduced specificity (0.855 vs. 0.912). When additional mammographic features were added, sensitivity increased to 0.457 and specificity to 0.872. SNPs and mammographic features played a significant role in breast cancer risk estimation (p-value < 0.001). Our decision framework comprising utility analysis and McNemar's test provides a novel framework to evaluate prediction models in the realm of radiogenomics.
Shen, Zhanlong; Lin, Yuanpei; Ye, Yingjiang; Jiang, Kewei; Xie, Qiwei; Gao, Zhidong; Wang, Shan
2018-04-01
To establish predicting models of surgical complications in elderly colorectal cancer patients. Surgical complications are usually critical and lethal in the elderly patients. However, none of the current models are specifically designed to predict surgical complications in elderly colorectal cancer patients. Details of 1008 cases of elderly colorectal cancer patients (age ≥ 65) were collected retrospectively from January 1998 to December 2013. Seventy-six clinicopathological variables which might affect postoperative complications in elderly patients were recorded. Multivariate stepwise logistic regression analysis was used to develop the risk model equations. The performance of the developed model was evaluated by measures of calibration (Hosmer-Lemeshow test) and discrimination (the area under the receiver-operator characteristic curve, AUC). The AUC of our established Surgical Complication Score for Elderly Colorectal Cancer patients (SCSECC) model was 0.743 (sensitivity, 82.1%; specificity, 78.3%). There was no significant discrepancy between observed and predicted incidence rates of surgical complications (AUC, 0.820; P = .812). The Surgical Site Infection Score for Elderly Colorectal Cancer patients (SSISECC) model showed significantly better prediction power compared to the National Nosocomial Infections Surveillance index (NNIS) (AUC, 0.732; P ˂ 0.001) and Efficacy of Nosocomial Infection Control index (SENIC) (AUC; 0.686; P˂0.001) models. The SCSECC and SSISECC models show good prediction power for postoperative surgical complication morbidity and surgical site infection in elderly colorectal cancer patients. Copyright © 2018 Elsevier Ltd, BASO ~ The Association for Cancer Surgery, and the European Society of Surgical Oncology. All rights reserved.
New prediction model for diagnosis of bacterial infection in febrile infants younger than 90 days.
Vujevic, Matea; Benzon, Benjamin; Markic, Josko
2017-01-01
Vujevic M, Benzon B, Markic J. New prediction model for diagnosis of bacterial infection in febrile infants younger than 90 days. Turk J Pediatr 2017; 59: 261-268. Due to non-specific clinical presentation in febrile infants, extensive laboratory testing is often carried out to distinguish simple viral disease from serious bacterial infection (SBI). Objective of this study was to compare efficacy of different biomarkers in early diagnosis of SBI in infants < 90 days old. Also, we developed prediction models with whom it will be possible to diagnose SBI with more accuracy than with any biomarkers independently. Febrile < 90-day-old infants hospitalized in 2-year-period at Department of Pediatrics, University Hospital Centre Split with suspicion of having SBI were included in this study. Retrospective cohort analysis of data acquired from medical records was performed. Out of 181 enrolled patients, SBI was confirmed in 70. Most common diagnosis was urinary tract infection (68.6%), followed by pneumonia (12.9%), sepsis (11.4%), gastroenterocolitis (5.7%) and meningitis (1.4%). Male gender was shown to be a risk factor for SBI in this population (p=0.008). White blood cell count (WBC), absolute neutrophil count (ANC) and C-reactive protein (CRP) were confirmed as the independent predictors of SBI, with CRP as the best one. Two prediction models built by combining biomarkers and clinical variables were selected as optimal with sensitivities of 74.3% and 75.7%, and specificities of 88.3% and 86%. Evidently, CRP is a more superior biomarker in diagnostics of SBI comparing to WBC and ANC. Prediction models were shown to be better in predicting SBI than independent biomarkers. Although both showed high sensitivity and specificity, their true strength should be determined using validation cohort.
PREDICTING POPULATION EXPOSURES TO PM10 AND PM 2.5
An improved model for human exposure to particulate matter (PM), specifically PM10 and PM2.5 is under development by the U.S. EPA/NERL. This model will incorporate data from new PM exposure measurement and exposure factors research. It is intended to be used to predict exposure...
Interest is increasing in using biological community data to provide information on the specific types of anthropogenic influences impacting streams. We built empirical models that predict the level of six different types of stress with fish and benthic macroinvertebrate data as...
Mizuno, Kiyonori; Andrish, Jack T; van den Bogert, Antonie J; McLean, Scott G
2009-12-01
While gender-based differences in knee joint anatomies/laxities are well documented, the potential for them to precipitate gender-dimorphic ACL loading and resultant injury risk has not been considered. To this end, we generated gender-specific models of ACL strain as a function of any six degrees of freedom (6DOF) knee joint load state via a combined cadaveric and analytical approach. Continuously varying joint forces and torques were applied to five male and five female cadaveric specimens and recorded along with synchronous knee flexion and ACL strain data. All data (approximately 10,000 samples) were submitted to specimen-specific regression analyses, affording ACL strain predictions as a function of the combined 6 DOF knee loads. Following individual model verifications, generalized gender-specific models were generated and subjected to 6 DOF external load scenarios consistent with both a clinical examination and a dynamic sports maneuver. The ensuing model-based strain predictions were subsequently examined for gender-based discrepancies. Male and female specimen-specific models predicted ACL strain within 0.51%+/-0.10% and 0.52%+/-0.07% of the measured data respectively, and explained more than 75% of the associated variance in each case. Predicted female ACL strains were also significantly larger than respective male values for both simulated 6 DOF load scenarios. Outcomes suggest that the female ACL will rupture in response to comparatively smaller external load applications. Future work must address the underlying anatomical/laxity contributions to knee joint mechanical and resultant ACL loading, ultimately affording prevention strategies that may cater to individual joint vulnerabilities.
The importance of understanding: Model space moderates goal specificity effects.
Kistner, Saskia; Burns, Bruce D; Vollmeyer, Regina; Kortenkamp, Ulrich
2016-01-01
The three-space theory of problem solving predicts that the quality of a learner's model and the goal specificity of a task interact on knowledge acquisition. In Experiment 1 participants used a computer simulation of a lever system to learn about torques. They either had to test hypotheses (nonspecific goal), or to produce given values for variables (specific goal). In the good- but not in the poor-model condition they saw torque depicted as an area. Results revealed the predicted interaction. A nonspecific goal only resulted in better learning when a good model of torques was provided. In Experiment 2 participants learned to manipulate the inputs of a system to control its outputs. A nonspecific goal to explore the system helped performance when compared to a specific goal to reach certain values when participants were given a good model, but not when given a poor model that suggested the wrong hypothesis space. Our findings support the three-space theory. They emphasize the importance of understanding for problem solving and stress the need to study underlying processes.
NASA Astrophysics Data System (ADS)
Rubinstein, Justin L.; Ellsworth, William L.; Beeler, Nicholas M.; Kilgore, Brian D.; Lockner, David A.; Savage, Heather M.
2012-02-01
The behavior of individual stick-slip events observed in three different laboratory experimental configurations is better explained by a "memoryless" earthquake model with fixed inter-event time or fixed slip than it is by the time- and slip-predictable models for earthquake occurrence. We make similar findings in the companion manuscript for the behavior of natural repeating earthquakes. Taken together, these results allow us to conclude that the predictions of a characteristic earthquake model that assumes either fixed slip or fixed recurrence interval should be preferred to the predictions of the time- and slip-predictable models for all earthquakes. Given that the fixed slip and recurrence models are the preferred models for all of the experiments we examine, we infer that in an event-to-event sense the elastic rebound model underlying the time- and slip-predictable models does not explain earthquake behavior. This does not indicate that the elastic rebound model should be rejected in a long-term-sense, but it should be rejected for short-term predictions. The time- and slip-predictable models likely offer worse predictions of earthquake behavior because they rely on assumptions that are too simple to explain the behavior of earthquakes. Specifically, the time-predictable model assumes a constant failure threshold and the slip-predictable model assumes that there is a constant minimum stress. There is experimental and field evidence that these assumptions are not valid for all earthquakes.
Lee, SoYean; Burns, G Leonard; Beauchaine, Theodore P; Becker, Stephen P
2016-08-01
The objective was to determine if the latent structure of attention-deficit/hyperactivity disorder (ADHD) and oppositional defiant disorder (ODD) symptoms is best explained by a general disruptive behavior factor along with specific inattention (IN), hyperactivity/impulsivity (HI), and ODD factors (a bifactor model) whereas the latent structure of sluggish cognitive tempo (SCT) symptoms is best explained by a first-order factor independent of the bifactor model of ADHD/ODD. Parents' (n = 703) and teachers' (n = 366) ratings of SCT, ADHD-IN, ADHD-HI, and ODD symptoms on the Child and Adolescent Disruptive Behavior Inventory (CADBI) in a community sample of children (ages 5-13; 55% girls) were used to evaluate 4 models of symptom organization. Results indicated that a bifactor model of ADHD/ODD symptoms, in conjunction with a separate first-order SCT factor, was the best model for both parent and teacher ratings. The first-order SCT factor showed discriminant validity with the general disruptive behavior and specific IN factors in the bifactor model. In addition, higher scores on the SCT factor predicted greater academic and social impairment, even after controlling for the general disruptive behavior and 3 specific factors. Consistent with predictions from the trait-impulsivity etiological model of externalizing liability, a single, general disruptive behavior factor accounted for nearly all common variance in ADHD/ODD symptoms, whereas SCT symptoms represented a factor different from the general disruptive behavior and specific IN factor. These results provide additional support for distinguishing between SCT and ADHD-IN. The study also demonstrates how etiological models can be used to predict specific latent structures of symptom organization. (PsycINFO Database Record (c) 2016 APA, all rights reserved).
Computational modeling of in vitro biological responses on polymethacrylate surfaces
Ghosh, Jayeeta; Lewitus, Dan Y; Chandra, Prafulla; Joy, Abraham; Bushman, Jared; Knight, Doyle; Kohn, Joachim
2011-01-01
The objective of this research was to examine the capabilities of QSPR (Quantitative Structure Property Relationship) modeling to predict specific biological responses (fibrinogen adsorption, cell attachment and cell proliferation index) on thin films of different polymethacrylates. Using 33 commercially available monomers it is theoretically possible to construct a library of over 40,000 distinct polymer compositions. A subset of these polymers were synthesized and solvent cast surfaces were prepared in 96 well plates for the measurement of fibrinogen adsorption. NIH 3T3 cell attachment and proliferation index were measured on spin coated thin films of these polymers. Based on the experimental results of these polymers, separate models were built for homo-, co-, and terpolymers in the library with good correlation between experiment and predicted values. The ability to predict biological responses by simple QSPR models for large numbers of polymers has important implications in designing biomaterials for specific biological or medical applications. PMID:21779132
Bowden, Joseph D; Bauerle, William L
2008-11-01
We investigated which parameters required by the MAESTRA model were most important in predicting leaf-area-based transpiration in 5-year-old trees of five deciduous hardwood species-yoshino cherry (Prunus x yedoensis Matsum.), red maple (Acer rubrum L. 'Autumn Flame'), trident maple (Acer buergeranum Miq.), Japanese flowering cherry (Prunus serrulata Lindl. 'Kwanzan') and London plane-tree (Platanus x acerifolia (Ait.) Willd.). Transpiration estimated from sap flow measured by the heat balance method in branches and trunks was compared with estimates predicted by the three-dimensional transpiration, photosynthesis and absorbed radiation model, MAESTRA. MAESTRA predicted species-specific transpiration from the interactions of leaf-level physiology and spatially explicit micro-scale weather patterns in a mixed deciduous hardwood plantation on a 15-min time step. The monthly differences between modeled mean daily transpiration estimates and measured mean daily sap flow ranged from a 35% underestimation for Acer buergeranum in June to a 25% overestimation for A. rubrum in July. The sensitivity of the modeled transpiration estimates was examined across a 30% error range for seven physiological input parameters. The minimum value of stomatal conductance as incident solar radiation tends to zero was determined to be eight times more influential than all other physiological model input parameters. This work quantified the major factors that influence modeled species-specific transpiration and confirmed the ability to scale leaf-level physiological attributes to whole-crown transpiration on a species-specific basis.
Fei, Yang; Hu, Jian; Gao, Kun; Tu, Jianfeng; Li, Wei-Qin; Wang, Wei
2017-06-01
To construct a radical basis function (RBF) artificial neural networks (ANNs) model to predict the incidence of acute pancreatitis (AP)-induced portal vein thrombosis. The analysis included 353 patients with AP who had admitted between January 2011 and December 2015. RBF ANNs model and logistic regression model were constructed based on eleven factors relevant to AP respectively. Statistical indexes were used to evaluate the value of the prediction in two models. The predict sensitivity, specificity, positive predictive value, negative predictive value and accuracy by RBF ANNs model for PVT were 73.3%, 91.4%, 68.8%, 93.0% and 87.7%, respectively. There were significant differences between the RBF ANNs and logistic regression models in these parameters (P<0.05). In addition, a comparison of the area under receiver operating characteristic curves of the two models showed a statistically significant difference (P<0.05). The RBF ANNs model is more likely to predict the occurrence of PVT induced by AP than logistic regression model. D-dimer, AMY, Hct and PT were important prediction factors of approval for AP-induced PVT. Copyright © 2017 Elsevier Inc. All rights reserved.
NASA Technical Reports Server (NTRS)
Spirka, T. A.; Myers, J. G.; Setser, R. M.; Halliburton, S. S.; White, R. D.; Chatzimavroudis, G. P.
2005-01-01
A priority of NASA is to identify and study possible risks to astronauts health during prolonged space missions [l]. The goal is to develop a procedure for a preflight evaluation of the cardiovascular system of an astronaut and to forecast how it will be affected during the mission. To predict these changes, a computational cardiovascular model must be constructed. Although physiology data can be used to make a general model, a more desirable subject-specific model requires anatomical, functional, and flow data from the specific astronaut. MRI has the unique advantage of providing images with all of the above information, including three-directional velocity data which can be used as boundary conditions in a computational fluid dynamics (CFD) program [2,3]. MRI-based CFD is very promising for reproduction of the flow patterns of a specific subject and prediction of changes in the absence of gravity. The aim of this study was to test the feasibility of this approach by reconstructing the geometry of MRI-scanned arterial models and reproducing the MRI-measured velocities using CFD simulations on these geometries.
NASA Astrophysics Data System (ADS)
Hallbauer-Zadorozhnaya, Valeriya; Santarato, Giovanni; Abu Zeid, Nasser
2015-08-01
In this paper, two separate but related goals are tackled. The first one is to demonstrate that in some saturated rock textures the non-linear behaviour of induced polarization (IP) and the violation of Ohm's law not only are real phenomena, but they can also be satisfactorily predicted by a suitable physical-mathematical model, which is our second goal. This model is based on Fick's second law. As the model links the specific dependence of resistivity and chargeability of a laboratory sample to the injected current and this in turn to its pore size distribution, it is able to predict pore size distribution from laboratory measurements, in good agreement with mercury injection capillary pressure test results. This fact opens up the possibility for hydrogeophysical applications on a macro scale. Mathematical modelling shows that the chargeability acquired in the field under normal conditions, that is at low current, will always be very small and approximately proportional to the applied current. A suitable field test site for demonstrating the possible reliance of both resistivity and chargeability on current was selected and a specific measuring strategy was established. Two data sets were acquired using different injected current strengths, while keeping the charging time constant. Observed variations of resistivity and chargeability are in agreement with those predicted by the mathematical model. These field test data should however be considered preliminary. If confirmed by further evidence, these facts may lead to changing the procedure of acquiring field measurements in future, and perhaps may encourage the design and building of a new specific geo-resistivity meter. This paper also shows that the well-known Marshall and Madden's equations based on Fick's law cannot be solved without specific boundary conditions.
Sparse Event Modeling with Hierarchical Bayesian Kernel Methods
2016-01-05
SECURITY CLASSIFICATION OF: The research objective of this proposal was to develop a predictive Bayesian kernel approach to model count data based on...several predictive variables. Such an approach, which we refer to as the Poisson Bayesian kernel model , is able to model the rate of occurrence of...which adds specificity to the model and can make nonlinear data more manageable. Early results show that the 1. REPORT DATE (DD-MM-YYYY) 4. TITLE
Jennifer A. Holm; H.H. Shugart; Skip J. Van Bloem; G.R. Larocque
2012-01-01
Because of human pressures, the need to understand and predict the long-term dynamics and development of subtropical dry forests is urgent. Through modifications to the ZELIG simulation model, including the development of species- and site-specific parameters and internal modifications, the capability to model and predict forest change within the 4500-ha Guanica State...
A three-dimensional inverse finite element analysis of the heel pad.
Chokhandre, Snehal; Halloran, Jason P; van den Bogert, Antonie J; Erdemir, Ahmet
2012-03-01
Quantification of plantar tissue behavior of the heel pad is essential in developing computational models for predictive analysis of preventive treatment options such as footwear for patients with diabetes. Simulation based studies in the past have generally adopted heel pad properties from the literature, in return using heel-specific geometry with material properties of a different heel. In exceptional cases, patient-specific material characterization was performed with simplified two-dimensional models, without further evaluation of a heel-specific response under different loading conditions. The aim of this study was to conduct an inverse finite element analysis of the heel in order to calculate heel-specific material properties in situ. Multidimensional experimental data available from a previous cadaver study by Erdemir et al. ("An Elaborate Data Set Characterizing the Mechanical Response of the Foot," ASME J. Biomech. Eng., 131(9), pp. 094502) was used for model development, optimization, and evaluation of material properties. A specimen-specific three-dimensional finite element representation was developed. Heel pad material properties were determined using inverse finite element analysis by fitting the model behavior to the experimental data. Compression dominant loading, applied using a spherical indenter, was used for optimization of the material properties. The optimized material properties were evaluated through simulations representative of a combined loading scenario (compression and anterior-posterior shear) with a spherical indenter and also of a compression dominant loading applied using an elevated platform. Optimized heel pad material coefficients were 0.001084 MPa (μ), 9.780 (α) (with an effective Poisson's ratio (ν) of 0.475), for a first-order nearly incompressible Ogden material model. The model predicted structural response of the heel pad was in good agreement for both the optimization (<1.05% maximum tool force, 0.9% maximum tool displacement) and validation cases (6.5% maximum tool force, 15% maximum tool displacement). The inverse analysis successfully predicted the material properties for the given specimen-specific heel pad using the experimental data for the specimen. The modeling framework and results can be used for accurate predictions of the three-dimensional interaction of the heel pad with its surroundings.
An Innovative Model to Predict Pediatric Emergency Department Return Visits.
Bergese, Ilaria; Frigerio, Simona; Clari, Marco; Castagno, Emanuele; De Clemente, Antonietta; Ponticelli, Elena; Scavino, Enrica; Berchialla, Paola
2016-10-06
Return visit (RV) to the emergency department (ED) is considered a benchmarking clinical indicator for health care quality. The purpose of this study was to develop a predictive model for early readmission risk in pediatric EDs comparing the performances of 2 learning machine algorithms. A retrospective study based on all children younger than 15 years spontaneously returning within 120 hours after discharge was conducted in an Italian university children's hospital between October 2012 and April 2013. Two predictive models, artificial neural network (ANN) and classification tree (CT), were used. Accuracy, specificity, and sensitivity were assessed. A total of 28,341 patient records were evaluated. Among them, 626 patients returned to the ED within 120 hours after their initial visit. Comparing ANN and CT, our analysis has shown that CT is the best model to predict RVs. The CT model showed an overall accuracy of 81%, slightly lower than the one achieved by the ANN (91.3%), but CT outperformed ANN with regard to sensitivity (79.8% vs 6.9%, respectively). The specificity was similar for the 2 models (CT, 97% vs ANN, 98.3%). In addition, the time of arrival and discharge along with the priority code assigned in triage, age, and diagnosis play a pivotal role to identify patients at high risk of RVs. These models provide a promising predictive tool for supporting the ED staff in preventing unnecessary RVs.
National Variation in Crop Yield Production Functions
NASA Astrophysics Data System (ADS)
Devineni, N.; Rising, J. A.
2017-12-01
A new multilevel model for yield prediction at the county scale using regional climate covariates is presented in this paper. A new crop specific water deficit index, growing degree days, extreme degree days, and time-trend as an approximation of technology improvements are used as predictors to estimate annual crop yields for each county from 1949 to 2009. Every county in the United States is allowed to have unique parameters describing how these weather predictors are related to yield outcomes. County-specific parameters are further modeled as varying according to climatic characteristics, allowing the prediction of parameters in regions where crops are not currently grown and into the future. The structural relationships between crop yield and regional climate as well as trends are estimated simultaneously. All counties are modeled in a single multilevel model with partial pooling to automatically group and reduce estimation uncertainties. The model captures up to 60% of the variability in crop yields after removing the effect of technology, does well in out of sample predictions and is useful in relating the climate responses to local bioclimatic factors. We apply the predicted growing models in a cost-benefit analysis to identify the most economically productive crop in each county.
Variable context Markov chains for HIV protease cleavage site prediction.
Oğul, Hasan
2009-06-01
Deciphering the knowledge of HIV protease specificity and developing computational tools for detecting its cleavage sites in protein polypeptide chain are very desirable for designing efficient and specific chemical inhibitors to prevent acquired immunodeficiency syndrome. In this study, we developed a generative model based on a generalization of variable order Markov chains (VOMC) for peptide sequences and adapted the model for prediction of their cleavability by certain proteases. The new method, called variable context Markov chains (VCMC), attempts to identify the context equivalence based on the evolutionary similarities between individual amino acids. It was applied for HIV-1 protease cleavage site prediction problem and shown to outperform existing methods in terms of prediction accuracy on a common dataset. In general, the method is a promising tool for prediction of cleavage sites of all proteases and encouraged to be used for any kind of peptide classification problem as well.
Hall, Charles B.; Lipton, Richard B.; Tennen, Howard; Haut, Sheryl R.
2014-01-01
Accurate prediction of seizures in persons with epilepsy offers opportunities for both precautionary measures and preemptive treatment. Previously identified predictors of seizures include patient-reported seizure anticipation, as well as stress, anxiety, and decreased sleep. In this study, we developed three models using 30 days of nightly seizure diary data in a cohort of 71 individuals with a history of uncontrolled seizures to predict subsequent seizures in the same cohort over a 30-day follow-up period. The best model combined the individual’s seizure history with that of the remainder of the cohort, resulting in 72% sensitivity for 80% specificity, and 0.83 area under the receiver operating characteristic curve. The possibility of clinically relevant prediction should be examined through electronic data capture and more specific and more frequent sampling, and with patient training to improve prediction. PMID:19138755
Estimating Glenoid Width for Instability-Related Bone Loss: A CT Evaluation of an MRI Formula.
Giles, Joshua W; Owens, Brett D; Athwal, George S
2015-07-01
Determining the magnitude of glenoid bone loss in cases of shoulder instability is an important step in selecting the optimal reconstructive procedure. Recently, a formula has been proposed that estimates native glenoid width based on magnetic resonance imaging (MRI) measurements of height (1/3 × glenoid height + 15 mm). This technique, however, has not been validated for use with computed tomography (CT), which is often the preferred imaging modality to assess bone deficiencies. The purpose of this project was 2-fold: (1) to determine if the MRI-based formula that predicts glenoid width from height is valid with CT and (2) to determine if a more accurate regression can be resolved for use specifically with CT data. Descriptive laboratory study. Ninety normal shoulder CT scans with preserved osseous anatomy were drawn from an existing database and analyzed. Measurements of glenoid height and width were performed by 2 observers on reconstructed 3-dimensional models. After assessment of reliability, the data were correlated, and regression models were created for male and female shoulders. The accuracy of the MRI-based model's predictions was then compared with that of the CT-based models. Intra- and interrater reliabilities were good to excellent for height and width, with intraclass correlation coefficients of 0.765 to 0.992. The height and width values had a strong correlation of 0.900 (P < .001). Regression analyses for male and female shoulders produced CT-specific formulas: for men, glenoid width = 2/3 × glenoid height + 5 mm; for women, glenoid width = 2/3 × glenoid height + 3 mm. Comparison of predictions from the MRI- and CT-specific formulas demonstrated good agreement (intraclass correlation coefficient = 0.818). The CT-specific formulas produced a root mean squared error of 1.2 mm, whereas application of the MRI-specific formula to CT images resulted in a root mean squared error of 1.5 mm. Use of the MRI-based formula on CT scans to predict glenoid width produced estimates that were nearly as accurate as the CT-specific formulas. The CT-specific formulas, however, are more accurate at predicting native glenoid width when applied to CT data. Imaging-specific (CT and MRI) formulas have been developed to estimate glenoid bone loss in patients with instability. The CT-specific formula can accurately predict native glenoid width, having an error of only 2.2% of average glenoid width. © 2015 The Author(s).
2013-08-01
surgeries, hospitalizations, etc). Once our model is developed we hope to apply our model at an outside institution, specifically University of...to build predictive models with the hope of improving disease management. It is difficult to find these factors in EMR systems as the...death, surgeries, hospitalizations, etc.) Once our model is developed, we hope to apply the model to de-identified data set from the University of
Lee, Yong Ju; Jung, Byeong Su; Kim, Kee-Tae; Paik, Hyun-Dong
2015-09-01
A predictive model was performed to describe the growth of Staphylococcus aureus in raw pork by using Integrated Pathogen Modeling Program 2013 and a polynomial model as a secondary predictive model. S. aureus requires approximately 180 h to reach 5-6 log CFU/g at 10 °C. At 15 °C and 25 °C, approximately 48 and 20 h, respectively, are required to cause food poisoning. Predicted data using the Gompertz model was the most accurate in this study. For lag time (LT) model, bias factor (Bf) and accuracy factor (Af) values were both 1.014, showing that the predictions were within a reliable range. For specific growth rate (SGR) model, Bf and Af were 1.188 and 1.190, respectively. Additionally, both Bf and Af values of the LT and SGR models were close to 1, indicating that IPMP Gompertz model is more adequate for predicting the growth of S. aureus on raw pork than other models. Copyright © 2015 Elsevier Ltd. All rights reserved.
Flassig, Robert J; Migal, Iryna; der Zalm, Esther van; Rihko-Struckmann, Liisa; Sundmacher, Kai
2015-01-16
Understanding the dynamics of biological processes can substantially be supported by computational models in the form of nonlinear ordinary differential equations (ODE). Typically, this model class contains many unknown parameters, which are estimated from inadequate and noisy data. Depending on the ODE structure, predictions based on unmeasured states and associated parameters are highly uncertain, even undetermined. For given data, profile likelihood analysis has been proven to be one of the most practically relevant approaches for analyzing the identifiability of an ODE structure, and thus model predictions. In case of highly uncertain or non-identifiable parameters, rational experimental design based on various approaches has shown to significantly reduce parameter uncertainties with minimal amount of effort. In this work we illustrate how to use profile likelihood samples for quantifying the individual contribution of parameter uncertainty to prediction uncertainty. For the uncertainty quantification we introduce the profile likelihood sensitivity (PLS) index. Additionally, for the case of several uncertain parameters, we introduce the PLS entropy to quantify individual contributions to the overall prediction uncertainty. We show how to use these two criteria as an experimental design objective for selecting new, informative readouts in combination with intervention site identification. The characteristics of the proposed multi-criterion objective are illustrated with an in silico example. We further illustrate how an existing practically non-identifiable model for the chlorophyll fluorescence induction in a photosynthetic organism, D. salina, can be rendered identifiable by additional experiments with new readouts. Having data and profile likelihood samples at hand, the here proposed uncertainty quantification based on prediction samples from the profile likelihood provides a simple way for determining individual contributions of parameter uncertainties to uncertainties in model predictions. The uncertainty quantification of specific model predictions allows identifying regions, where model predictions have to be considered with care. Such uncertain regions can be used for a rational experimental design to render initially highly uncertain model predictions into certainty. Finally, our uncertainty quantification directly accounts for parameter interdependencies and parameter sensitivities of the specific prediction.
Tree-based flood damage modeling of companies: Damage processes and model performance
NASA Astrophysics Data System (ADS)
Sieg, Tobias; Vogel, Kristin; Merz, Bruno; Kreibich, Heidi
2017-07-01
Reliable flood risk analyses, including the estimation of damage, are an important prerequisite for efficient risk management. However, not much is known about flood damage processes affecting companies. Thus, we conduct a flood damage assessment of companies in Germany with regard to two aspects. First, we identify relevant damage-influencing variables. Second, we assess the prediction performance of the developed damage models with respect to the gain by using an increasing amount of training data and a sector-specific evaluation of the data. Random forests are trained with data from two postevent surveys after flood events occurring in the years 2002 and 2013. For a sector-specific consideration, the data set is split into four subsets corresponding to the manufacturing, commercial, financial, and service sectors. Further, separate models are derived for three different company assets: buildings, equipment, and goods and stock. Calculated variable importance values reveal different variable sets relevant for the damage estimation, indicating significant differences in the damage process for various company sectors and assets. With an increasing number of data used to build the models, prediction errors decrease. Yet the effect is rather small and seems to saturate for a data set size of several hundred observations. In contrast, the prediction improvement achieved by a sector-specific consideration is more distinct, especially for damage to equipment and goods and stock. Consequently, sector-specific data acquisition and a consideration of sector-specific company characteristics in future flood damage assessments is expected to improve the model performance more than a mere increase in data.
Lerner, Zachary F; DeMers, Matthew S; Delp, Scott L; Browning, Raymond C
2015-02-26
Understanding degeneration of biological and prosthetic knee joints requires knowledge of the in-vivo loading environment during activities of daily living. Musculoskeletal models can estimate medial/lateral tibiofemoral compartment contact forces, yet anthropometric differences between individuals make accurate predictions challenging. We developed a full-body OpenSim musculoskeletal model with a knee joint that incorporates subject-specific tibiofemoral alignment (i.e. knee varus-valgus) and geometry (i.e. contact locations). We tested the accuracy of our model and determined the importance of these subject-specific parameters by comparing estimated to measured medial and lateral contact forces during walking in an individual with an instrumented knee replacement and post-operative genu valgum (6°). The errors in the predictions of the first peak medial and lateral contact force were 12.4% and 11.9%, respectively, for a model with subject-specific tibiofemoral alignment and contact locations determined through radiographic analysis, vs. 63.1% and 42.0%, respectively, for a model with generic parameters. We found that each degree of tibiofemoral alignment deviation altered the first peak medial compartment contact force by 51N (r(2)=0.99), while each millimeter of medial-lateral translation of the compartment contact point locations altered the first peak medial compartment contact force by 41N (r(2)=0.99). The model, available at www.simtk.org/home/med-lat-knee/, enables the specification of subject-specific joint alignment and compartment contact locations to more accurately estimate medial and lateral tibiofemoral contact forces in individuals with non-neutral alignment. Copyright © 2015 Elsevier Ltd. All rights reserved.
Lerner, Zachary F.; DeMers, Matthew S.; Delp, Scott L.; Browning, Raymond C.
2015-01-01
Understanding degeneration of biological and prosthetic knee joints requires knowledge of the in-vivo loading environment during activities of daily living. Musculoskeletal models can estimate medial/lateral tibiofemoral compartment contact forces, yet anthropometric differences between individuals make accurate predictions challenging. We developed a full-body OpenSim musculoskeletal model with a knee joint that incorporates subject-specific tibiofemoral alignment (i.e. knee varus-valgus) and geometry (i.e. contact locations). We tested the accuracy of our model and determined the importance of these subject-specific parameters by comparing estimated to measured medial and lateral contact forces during walking in an individual with an instrumented knee replacement and post-operative genu valgum (6°). The errors in the predictions of the first peak medial and lateral contact force were 12.4% and 11.9%, respectively, for a model with subject-specific tibiofemoral alignment and contact locations determined via radiographic analysis, vs. 63.1% and 42.0%, respectively, for a model with generic parameters. We found that each degree of tibiofemoral alignment deviation altered the first peak medial compartment contact force by 51N (r2=0.99), while each millimeter of medial-lateral translation of the compartment contact point locations altered the first peak medial compartment contact force by 41N (r2=0.99). The model, available at www.simtk.org/home/med-lat-knee/, enables the specification of subject-specific joint alignment and compartment contact locations to more accurately estimate medial and lateral tibiofemoral contact forces in individuals with non-neutral alignment. PMID:25595425
Candidacy for Kidney Transplantation of Older Adults
Grams, Morgan E.; Kucirka, Lauren M.; Hanrahan, Colleen F.; Montgomery, Robert A.; Massie, Allan B.; Segev, Dorry L.
2013-01-01
OBJECTIVES To develop a prediction model for kidney transplantation (KT) outcomes specific to older adults with end-stage renal disease (ESRD) and to use this model to estimate the number of excellent older KT candidates who lack access to KT. DESIGN Secondary analysis of data collected by the United Network for Organ Sharing and U.S. Renal Disease System. SETTING Retrospective analysis of national registry data. PARTICIPANTS Model development: Medicare-primary older recipients (aged ≥ 65) of a first KT between 1999 and 2006 (N = 6,988). Model application: incident Medicare-primary older adults with ESRD between 1999 and 2006 without an absolute or relative contraindication to transplantation (N = 128,850). MEASUREMENTS Comorbid conditions were extracted from U.S. Renal Disease System Form 2728 data and Medicare claims. RESULTS The prediction model used 19 variables to estimate post-KT outcome and showed good calibration (Hosmer–Lemeshow P = .44) and better prediction than previous population-average models (P < .001). Application of the model to the population with incident ESRD identified 11,756 excellent older transplant candidates (defined as >87% predicted 3-year post-KT survival, corresponding to the top 20% of transplanted older adults used in model development), of whom 76.3% (n = 8,966) lacked access. It was estimated that 11% of these candidates would have identified a suitable live donor had they been referred for KT. CONCLUSION A risk-prediction model specific to older adults can identify excellent KT candidates. Appropriate referral could result in significantly greater rates of KT in older adults. PMID:22239290
Small Engine Technology (SET) Task 23 ANOPP Noise Prediction for Small Engines, Wing Reflection Code
NASA Technical Reports Server (NTRS)
Lieber, Lysbeth; Brown, Daniel; Golub, Robert A. (Technical Monitor)
2000-01-01
The work performed under Task 23 consisted of the development and demonstration of improvements for the NASA Aircraft Noise Prediction Program (ANOPP), specifically targeted to the modeling of engine noise enhancement due to wing reflection. This report focuses on development of the model and procedure to predict the effects of wing reflection, and the demonstration of the procedure, using a representative wing/engine configuration.
Reagan, Andrew J; Dubief, Yves; Dodds, Peter Sheridan; Danforth, Christopher M
2016-01-01
A thermal convection loop is a annular chamber filled with water, heated on the bottom half and cooled on the top half. With sufficiently large forcing of heat, the direction of fluid flow in the loop oscillates chaotically, dynamics analogous to the Earth's weather. As is the case for state-of-the-art weather models, we only observe the statistics over a small region of state space, making prediction difficult. To overcome this challenge, data assimilation (DA) methods, and specifically ensemble methods, use the computational model itself to estimate the uncertainty of the model to optimally combine these observations into an initial condition for predicting the future state. Here, we build and verify four distinct DA methods, and then, we perform a twin model experiment with the computational fluid dynamics simulation of the loop using the Ensemble Transform Kalman Filter (ETKF) to assimilate observations and predict flow reversals. We show that using adaptively shaped localized covariance outperforms static localized covariance with the ETKF, and allows for the use of less observations in predicting flow reversals. We also show that a Dynamic Mode Decomposition (DMD) of the temperature and velocity fields recovers the low dimensional system underlying reversals, finding specific modes which together are predictive of reversal direction.
Reagan, Andrew J.; Dubief, Yves; Dodds, Peter Sheridan; Danforth, Christopher M.
2016-01-01
A thermal convection loop is a annular chamber filled with water, heated on the bottom half and cooled on the top half. With sufficiently large forcing of heat, the direction of fluid flow in the loop oscillates chaotically, dynamics analogous to the Earth’s weather. As is the case for state-of-the-art weather models, we only observe the statistics over a small region of state space, making prediction difficult. To overcome this challenge, data assimilation (DA) methods, and specifically ensemble methods, use the computational model itself to estimate the uncertainty of the model to optimally combine these observations into an initial condition for predicting the future state. Here, we build and verify four distinct DA methods, and then, we perform a twin model experiment with the computational fluid dynamics simulation of the loop using the Ensemble Transform Kalman Filter (ETKF) to assimilate observations and predict flow reversals. We show that using adaptively shaped localized covariance outperforms static localized covariance with the ETKF, and allows for the use of less observations in predicting flow reversals. We also show that a Dynamic Mode Decomposition (DMD) of the temperature and velocity fields recovers the low dimensional system underlying reversals, finding specific modes which together are predictive of reversal direction. PMID:26849061
Ji, Zhiwei; Wang, Bing; Yan, Ke; Dong, Ligang; Meng, Guanmin; Shi, Lei
2017-12-21
In recent years, the integration of 'omics' technologies, high performance computation, and mathematical modeling of biological processes marks that the systems biology has started to fundamentally impact the way of approaching drug discovery. The LINCS public data warehouse provides detailed information about cell responses with various genetic and environmental stressors. It can be greatly helpful in developing new drugs and therapeutics, as well as improving the situations of lacking effective drugs, drug resistance and relapse in cancer therapies, etc. In this study, we developed a Ternary status based Integer Linear Programming (TILP) method to infer cell-specific signaling pathway network and predict compounds' treatment efficacy. The novelty of our study is that phosphor-proteomic data and prior knowledge are combined for modeling and optimizing the signaling network. To test the power of our approach, a generic pathway network was constructed for a human breast cancer cell line MCF7; and the TILP model was used to infer MCF7-specific pathways with a set of phosphor-proteomic data collected from ten representative small molecule chemical compounds (most of them were studied in breast cancer treatment). Cross-validation indicated that the MCF7-specific pathway network inferred by TILP were reliable predicting a compound's efficacy. Finally, we applied TILP to re-optimize the inferred cell-specific pathways and predict the outcomes of five small compounds (carmustine, doxorubicin, GW-8510, daunorubicin, and verapamil), which were rarely used in clinic for breast cancer. In the simulation, the proposed approach facilitates us to identify a compound's treatment efficacy qualitatively and quantitatively, and the cross validation analysis indicated good accuracy in predicting effects of five compounds. In summary, the TILP model is useful for discovering new drugs for clinic use, and also elucidating the potential mechanisms of a compound to targets.
Identifying type 1 and type 2 diabetic cases using administrative data: a tree-structured model.
Lo-Ciganic, Weihsuan; Zgibor, Janice C; Ruppert, Kristine; Arena, Vincent C; Stone, Roslyn A
2011-05-01
To date, few administrative diabetes mellitus (DM) registries have distinguished type 1 diabetes mellitus (T1DM) from type 2 diabetes mellitus (T2DM). Using a classification tree model, a prediction rule was developed to distinguish T1DM from T2DM in a large administrative database. The Medical Archival Retrieval System at the University of Pittsburgh Medical Center included administrative and clinical data from January 1, 2000, through September 30, 2009, for 209,647 DM patients aged ≥18 years. Probable cases (8,173 T1DM and 125,111 T2DM) were identified by applying clinical criteria to administrative data. Nonparametric classification tree models were fit using TIBCO Spotfire S+ 8.1 (TIBCO Software), with model size based on 10-fold cross validation. Sensitivity, specificity, positive predictive value (PPV), and negative predictive value (NPV) of T1DM were estimated. The main predictors that distinguished T1DM from T2DM are age <40 years; International Classification of Disease, 9th revision, codes of T1DM or T2DM diagnosis; inpatient oral hypoglycemic agent use; inpatient insulin use; and episode(s) of diabetic ketoacidosis diagnosis. Compared with a complex clinical algorithm, the tree-structured model to predict T1DM had 92.8% sensitivity, 99.3% specificity, 89.5% PPV, and 99.5% NPV. The preliminary predictive rule appears to be promising. Being able to distinguish between DM subtypes in administrative databases will allow large-scale subtype-specific analyses of medical care costs, morbidity, and mortality. © 2011 Diabetes Technology Society.
Action perception as hypothesis testing.
Donnarumma, Francesco; Costantini, Marcello; Ambrosini, Ettore; Friston, Karl; Pezzulo, Giovanni
2017-04-01
We present a novel computational model that describes action perception as an active inferential process that combines motor prediction (the reuse of our own motor system to predict perceived movements) and hypothesis testing (the use of eye movements to disambiguate amongst hypotheses). The system uses a generative model of how (arm and hand) actions are performed to generate hypothesis-specific visual predictions, and directs saccades to the most informative places of the visual scene to test these predictions - and underlying hypotheses. We test the model using eye movement data from a human action observation study. In both the human study and our model, saccades are proactive whenever context affords accurate action prediction; but uncertainty induces a more reactive gaze strategy, via tracking the observed movements. Our model offers a novel perspective on action observation that highlights its active nature based on prediction dynamics and hypothesis testing. Copyright © 2017 The Authors. Published by Elsevier Ltd.. All rights reserved.
Towards Subject-Specific Strength Training Design through Predictive Use of Musculoskeletal Models.
Plüss, Michael; Schellenberg, Florian; Taylor, William R; Lorenzetti, Silvio
2018-01-01
Lower extremity dysfunction is often associated with hip muscle strength deficiencies. Detailed knowledge of the muscle forces generated in the hip under specific external loading conditions enables specific structures to be trained. The aim of this study was to find the most effective movement type and loading direction to enable the training of specific parts of the hip muscles using a standing posture and a pulley system. In a novel approach to release the predictive power of musculoskeletal modelling techniques based on inverse dynamics, flexion/extension and ab-/adduction movements were virtually created. To demonstrate the effectiveness of this approach, three hip orientations and an external loading force that was systematically rotated around the body were simulated using a state-of-the art OpenSim model in order to establish ideal designs for training of the anterior and posterior parts of the M. gluteus medius (GM). The external force direction as well as the hip orientation greatly influenced the muscle forces in the different parts of the GM. No setting was found for simultaneous training of the anterior and posterior parts with a muscle force higher than 50% of the maximum. Importantly, this study has demonstrated the use of musculoskeletal models as an approach to predict muscle force variations for different strength and rehabilitation exercise variations.
Towards Subject-Specific Strength Training Design through Predictive Use of Musculoskeletal Models
Plüss, Michael; Schellenberg, Florian
2018-01-01
Lower extremity dysfunction is often associated with hip muscle strength deficiencies. Detailed knowledge of the muscle forces generated in the hip under specific external loading conditions enables specific structures to be trained. The aim of this study was to find the most effective movement type and loading direction to enable the training of specific parts of the hip muscles using a standing posture and a pulley system. In a novel approach to release the predictive power of musculoskeletal modelling techniques based on inverse dynamics, flexion/extension and ab-/adduction movements were virtually created. To demonstrate the effectiveness of this approach, three hip orientations and an external loading force that was systematically rotated around the body were simulated using a state-of-the art OpenSim model in order to establish ideal designs for training of the anterior and posterior parts of the M. gluteus medius (GM). The external force direction as well as the hip orientation greatly influenced the muscle forces in the different parts of the GM. No setting was found for simultaneous training of the anterior and posterior parts with a muscle force higher than 50% of the maximum. Importantly, this study has demonstrated the use of musculoskeletal models as an approach to predict muscle force variations for different strength and rehabilitation exercise variations. PMID:29796082
Predictive information processing in music cognition. A critical review.
Rohrmeier, Martin A; Koelsch, Stefan
2012-02-01
Expectation and prediction constitute central mechanisms in the perception and cognition of music, which have been explored in theoretical and empirical accounts. We review the scope and limits of theoretical accounts of musical prediction with respect to feature-based and temporal prediction. While the concept of prediction is unproblematic for basic single-stream features such as melody, it is not straight-forward for polyphonic structures or higher-order features such as formal predictions. Behavioural results based on explicit and implicit (priming) paradigms provide evidence of priming in various domains that may reflect predictive behaviour. Computational learning models, including symbolic (fragment-based), probabilistic/graphical, or connectionist approaches, provide well-specified predictive models of specific features and feature combinations. While models match some experimental results, full-fledged music prediction cannot yet be modelled. Neuroscientific results regarding the early right-anterior negativity (ERAN) and mismatch negativity (MMN) reflect expectancy violations on different levels of processing complexity, and provide some neural evidence for different predictive mechanisms. At present, the combinations of neural and computational modelling methodologies are at early stages and require further research. Copyright © 2012 Elsevier B.V. All rights reserved.
Evaluation of scenario-specific modeling approaches to predict plane of array solar irradiation
Moslehi, Salim; Reddy, T. Agami; Katipamula, Srinivas
2017-12-20
Predicting thermal or electric power output from solar collectors requires knowledge of solar irradiance incident on the collector, known as plane of array irradiance. In the absence of such a measurement, plane of array irradiation can be predicted using relevant transposition models which essentially requires diffuse (or beam) radiation to be to be known along with total horizontal irradiation. The two main objectives of the current study are (1) to evaluate the extent to which the prediction of plane of array irradiance is improved when diffuse radiation is predicted using location-specific regression models developed from on-site measured data as againstmore » using generalized models; and (2) to estimate the expected uncertainties associated with plane of array irradiance predictions under different data collection scenarios likely to be encountered in practical situations. These issues have been investigated using monitored data for several U.S. locations in conjunction with the Typical Meteorological Year, version 3 database. An interesting behavior in the Typical Meteorological Year, version 3 data was also observed in correlation patterns between diffuse and total radiation taken from different years which seems to attest to a measurement problem. Furthermore, the current study was accomplished under a broader research agenda aimed at providing energy managers the necessary tools for predicting, scheduling, and controlling various sub-systems of an integrated energy system.« less
Evaluation of scenario-specific modeling approaches to predict plane of array solar irradiation
DOE Office of Scientific and Technical Information (OSTI.GOV)
Moslehi, Salim; Reddy, T. Agami; Katipamula, Srinivas
Predicting thermal or electric power output from solar collectors requires knowledge of solar irradiance incident on the collector, known as plane of array irradiance. In the absence of such a measurement, plane of array irradiation can be predicted using relevant transposition models which essentially requires diffuse (or beam) radiation to be to be known along with total horizontal irradiation. The two main objectives of the current study are (1) to evaluate the extent to which the prediction of plane of array irradiance is improved when diffuse radiation is predicted using location-specific regression models developed from on-site measured data as againstmore » using generalized models; and (2) to estimate the expected uncertainties associated with plane of array irradiance predictions under different data collection scenarios likely to be encountered in practical situations. These issues have been investigated using monitored data for several U.S. locations in conjunction with the Typical Meteorological Year, version 3 database. An interesting behavior in the Typical Meteorological Year, version 3 data was also observed in correlation patterns between diffuse and total radiation taken from different years which seems to attest to a measurement problem. Furthermore, the current study was accomplished under a broader research agenda aimed at providing energy managers the necessary tools for predicting, scheduling, and controlling various sub-systems of an integrated energy system.« less
Tertiary model of a plant cellulose synthase
Sethaphong, Latsavongsakda; Haigler, Candace H.; Kubicki, James D.; Zimmer, Jochen; Bonetta, Dario; DeBolt, Seth; Yingling, Yaroslava G.
2013-01-01
A 3D atomistic model of a plant cellulose synthase (CESA) has remained elusive despite over forty years of experimental effort. Here, we report a computationally predicted 3D structure of 506 amino acids of cotton CESA within the cytosolic region. Comparison of the predicted plant CESA structure with the solved structure of a bacterial cellulose-synthesizing protein validates the overall fold of the modeled glycosyltransferase (GT) domain. The coaligned plant and bacterial GT domains share a six-stranded β-sheet, five α-helices, and conserved motifs similar to those required for catalysis in other GT-2 glycosyltransferases. Extending beyond the cross-kingdom similarities related to cellulose polymerization, the predicted structure of cotton CESA reveals that plant-specific modules (plant-conserved region and class-specific region) fold into distinct subdomains on the periphery of the catalytic region. Computational results support the importance of the plant-conserved region and/or class-specific region in CESA oligomerization to form the multimeric cellulose–synthesis complexes that are characteristic of plants. Relatively high sequence conservation between plant CESAs allowed mapping of known mutations and two previously undescribed mutations that perturb cellulose synthesis in Arabidopsis thaliana to their analogous positions in the modeled structure. Most of these mutation sites are near the predicted catalytic region, and the confluence of other mutation sites supports the existence of previously undefined functional nodes within the catalytic core of CESA. Overall, the predicted tertiary structure provides a platform for the biochemical engineering of plant CESAs. PMID:23592721
A4 flavour model for Dirac neutrinos: Type I and inverse seesaw
NASA Astrophysics Data System (ADS)
Borah, Debasish; Karmakar, Biswajit
2018-05-01
We propose two different seesaw models namely, type I and inverse seesaw to realise light Dirac neutrinos within the framework of A4 discrete flavour symmetry. The additional fields and their transformations under the flavour symmetries are chosen in such a way that naturally predicts the hierarchies of different elements of the seesaw mass matrices in these two types of seesaw mechanisms. For generic choices of flavon alignments, both the models predict normal hierarchical light neutrino masses with the atmospheric mixing angle in the lower octant. Apart from predicting interesting correlations between different neutrino parameters as well as between neutrino and model parameters, the model also predicts the leptonic Dirac CP phase to lie in a specific range - π / 3 to π / 3. While the type I seesaw model predicts smaller values of absolute neutrino mass, the inverse seesaw predictions for the absolute neutrino masses can saturate the cosmological upper bound on sum of absolute neutrino masses for certain choices of model parameters.
Zielinski, Michal W; McGann, Locksley E; Nychka, John A; Elliott, Janet A W
2014-10-01
Thermodynamic solution theories allow the prediction of chemical potentials in solutions of known composition. In cryobiology, such models are a critical component of many mathematical models that are used to simulate the biophysical processes occurring in cells and tissues during cryopreservation. A number of solution theories, both thermodynamically ideal and non-ideal, have been proposed for use with cryobiological solutions. In this work, we have evaluated two non-ideal solution theories for predicting water chemical potential (i.e. osmolality) in multi-solute solutions relevant to cryobiology: the Elliott et al. form of the multi-solute osmotic virial equation, and the Kleinhans and Mazur freezing point summation model. These two solution theories require fitting to only single-solute data, although they can make predictions in multi-solute solutions. The predictions of these non-ideal solution theories were compared to predictions made using ideal dilute assumptions and to available literature multi-solute experimental osmometric data. A single, consistent set of literature single-solute solution data was used to fit for the required solute-specific coefficients for each of the non-ideal models. Our results indicate that the two non-ideal solution theories have similar overall performance, and both give more accurate predictions than ideal models. These results can be used to select between the non-ideal models for a specific multi-solute solution, and the updated coefficients provided in this work can be used to make the desired predictions. Copyright © 2014 The Authors. Published by Elsevier Inc. All rights reserved.
Communication Efficacy and Couples’ Cancer Management: Applying a Dyadic Appraisal Model
Magsamen-Conrad, Kate; Checton, Maria G.; Venetis, Maria K.; Greene, Kathryn
2014-01-01
The purpose of the present study was to apply Berg and Upchurch’s (2007) developmental-conceptual model to understand better how couples cope with cancer. Specifically, we hypothesized a dyadic appraisal model in which proximal factors (relational quality), dyadic appraisal (prognosis uncertainty), and dyadic coping (communication efficacy) predicted adjustment (cancer management). The study was cross-sectional and included 83 dyads in which one partner had been diagnosed with and/or treated for cancer. For both patients and partners, multilevel analyses using the actor-partner interdependence model (APIM) indicated that proximal contextual factors predicted dyadic appraisal and dyadic coping. Dyadic appraisal predicted dyadic coping, which then predicted dyadic adjustment. Patients’ confidence in their ability to talk about the cancer predicted their own cancer management. Partners’ confidence predicted their own and the patient’s ability to cope with cancer, which then predicted patients’ perceptions of their general health. Implications and future research are discussed. PMID:25983382
Communication Efficacy and Couples' Cancer Management: Applying a Dyadic Appraisal Model.
Magsamen-Conrad, Kate; Checton, Maria G; Venetis, Maria K; Greene, Kathryn
2015-06-01
The purpose of the present study was to apply Berg and Upchurch's (2007) developmental-conceptual model to understand better how couples cope with cancer. Specifically, we hypothesized a dyadic appraisal model in which proximal factors (relational quality), dyadic appraisal (prognosis uncertainty), and dyadic coping (communication efficacy) predicted adjustment (cancer management). The study was cross-sectional and included 83 dyads in which one partner had been diagnosed with and/or treated for cancer. For both patients and partners, multilevel analyses using the actor-partner interdependence model (APIM) indicated that proximal contextual factors predicted dyadic appraisal and dyadic coping. Dyadic appraisal predicted dyadic coping, which then predicted dyadic adjustment. Patients' confidence in their ability to talk about the cancer predicted their own cancer management. Partners' confidence predicted their own and the patient's ability to cope with cancer, which then predicted patients' perceptions of their general health. Implications and future research are discussed.
Task-Specific Response Strategy Selection on the Basis of Recent Training Experience
Fulvio, Jacqueline M.; Green, C. Shawn; Schrater, Paul R.
2014-01-01
The goal of training is to produce learning for a range of activities that are typically more general than the training task itself. Despite a century of research, predicting the scope of learning from the content of training has proven extremely difficult, with the same task producing narrowly focused learning strategies in some cases and broadly scoped learning strategies in others. Here we test the hypothesis that human subjects will prefer a decision strategy that maximizes performance and reduces uncertainty given the demands of the training task and that the strategy chosen will then predict the extent to which learning is transferable. To test this hypothesis, we trained subjects on a moving dot extrapolation task that makes distinct predictions for two types of learning strategy: a narrow model-free strategy that learns an input-output mapping for training stimuli, and a general model-based strategy that utilizes humans' default predictive model for a class of trajectories. When the number of distinct training trajectories is low, we predict better performance for the mapping strategy, but as the number increases, a predictive model is increasingly favored. Consonant with predictions, subject extrapolations for test trajectories were consistent with using a mapping strategy when trained on a small number of training trajectories and a predictive model when trained on a larger number. The general framework developed here can thus be useful both in interpreting previous patterns of task-specific versus task-general learning, as well as in building future training paradigms with certain desired outcomes. PMID:24391490
Chara, Liaskou; Eleftherios, Vouzounerakis; Maria, Moirasgenti; Anastasia, Trikoupi; Chryssoula, Staikou
2014-01-01
Background and Aims: Difficult airway assessment is based on various anatomic parameters of upper airway, much of it being concentrated on oral cavity and the pharyngeal structures. The diagnostic value of tests based on neck anatomy in predicting difficult laryngoscopy was assessed in this prospective, open cohort study. Methods: We studied 341 adult patients scheduled to receive general anaesthesia. Thyromental distance (TMD), sternomental distance (STMD), ratio of height to thyromental distance (RHTMD) and neck circumference (NC) were measured pre-operatively. The laryngoscopic view was classified according to the Cormack–Lehane Grade (1-4). Difficult laryngoscopy was defined as Cormack–Lehane Grade 3 or 4. The optimal cut-off points for each variable were identified by using receiver operating characteristic analysis. Sensitivity, specificity and positive predictive value and negative predictive value (NPV) were calculated for each test. Multivariate analysis with logistic regression, including all variables, was used to create a predictive model. Comparisons between genders were also performed. Results: Laryngoscopy was difficult in 12.6% of the patients. The cut-off values were: TMD ≤7 cm, STMD ≤15 cm, RHTMD >18.4 and NC >37.5 cm. The RHTMD had the highest sensitivity (88.4%) and NPV (95.2%), while TMD had the highest specificity (83.9%). The area under curve (AUC) for the TMD, STMD, RHTMD and NC was 0.63, 0.64, 0.62 and 0.54, respectively. The predictive model exhibited a higher and statistically significant diagnostic accuracy (AUC: 0.68, P < 0.001). Gender-specific cut-off points improved the predictive accuracy of NC in women (AUC: 0.65). Conclusions: The TMD, STMD, RHTMD and NC were found to be poor single predictors of difficult laryngoscopy, while a model including all four variables had a significant predictive accuracy. Among the studied tests, gender-specific cut-off points should be used for NC. PMID:24963183
Liaskou, Chara; Chara, Liaskou; Vouzounerakis, Eleftherios; Eleftherios, Vouzounerakis; Moirasgenti, Maria; Maria, Moirasgenti; Trikoupi, Anastasia; Anastasia, Trikoupi; Staikou, Chryssoula; Chryssoula, Staikou
2014-03-01
Difficult airway assessment is based on various anatomic parameters of upper airway, much of it being concentrated on oral cavity and the pharyngeal structures. The diagnostic value of tests based on neck anatomy in predicting difficult laryngoscopy was assessed in this prospective, open cohort study. We studied 341 adult patients scheduled to receive general anaesthesia. Thyromental distance (TMD), sternomental distance (STMD), ratio of height to thyromental distance (RHTMD) and neck circumference (NC) were measured pre-operatively. The laryngoscopic view was classified according to the Cormack-Lehane Grade (1-4). Difficult laryngoscopy was defined as Cormack-Lehane Grade 3 or 4. The optimal cut-off points for each variable were identified by using receiver operating characteristic analysis. Sensitivity, specificity and positive predictive value and negative predictive value (NPV) were calculated for each test. Multivariate analysis with logistic regression, including all variables, was used to create a predictive model. Comparisons between genders were also performed. Laryngoscopy was difficult in 12.6% of the patients. The cut-off values were: TMD ≤7 cm, STMD ≤15 cm, RHTMD >18.4 and NC >37.5 cm. The RHTMD had the highest sensitivity (88.4%) and NPV (95.2%), while TMD had the highest specificity (83.9%). The area under curve (AUC) for the TMD, STMD, RHTMD and NC was 0.63, 0.64, 0.62 and 0.54, respectively. The predictive model exhibited a higher and statistically significant diagnostic accuracy (AUC: 0.68, P < 0.001). Gender-specific cut-off points improved the predictive accuracy of NC in women (AUC: 0.65). The TMD, STMD, RHTMD and NC were found to be poor single predictors of difficult laryngoscopy, while a model including all four variables had a significant predictive accuracy. Among the studied tests, gender-specific cut-off points should be used for NC.
A dynamic spatio-temporal model for spatial data
Hefley, Trevor J.; Hooten, Mevin B.; Hanks, Ephraim M.; Russell, Robin; Walsh, Daniel P.
2017-01-01
Analyzing spatial data often requires modeling dependencies created by a dynamic spatio-temporal data generating process. In many applications, a generalized linear mixed model (GLMM) is used with a random effect to account for spatial dependence and to provide optimal spatial predictions. Location-specific covariates are often included as fixed effects in a GLMM and may be collinear with the spatial random effect, which can negatively affect inference. We propose a dynamic approach to account for spatial dependence that incorporates scientific knowledge of the spatio-temporal data generating process. Our approach relies on a dynamic spatio-temporal model that explicitly incorporates location-specific covariates. We illustrate our approach with a spatially varying ecological diffusion model implemented using a computationally efficient homogenization technique. We apply our model to understand individual-level and location-specific risk factors associated with chronic wasting disease in white-tailed deer from Wisconsin, USA and estimate the location the disease was first introduced. We compare our approach to several existing methods that are commonly used in spatial statistics. Our spatio-temporal approach resulted in a higher predictive accuracy when compared to methods based on optimal spatial prediction, obviated confounding among the spatially indexed covariates and the spatial random effect, and provided additional information that will be important for containing disease outbreaks.
RACER a Coarse-Grained RNA Model for Capturing Folding Free Energy in Molecular Dynamics Simulations
NASA Astrophysics Data System (ADS)
Cheng, Sara; Bell, David; Ren, Pengyu
RACER is a coarse-grained RNA model that can be used in molecular dynamics simulations to predict native structures and sequence-specific variation of free energy of various RNA structures. RACER is capable of accurate prediction of native structures of duplexes and hairpins (average RMSD of 4.15 angstroms), and RACER can capture sequence-specific variation of free energy in excellent agreement with experimentally measured stabilities (r-squared =0.98). The RACER model implements a new effective non-bonded potential and re-parameterization of hydrogen bond and Debye-Huckel potentials. Insights from the RACER model include the importance of treating pairing and stacking interactions separately in order to distinguish folded an unfolded states and identification of hydrogen-bonding, base stacking, and electrostatic interactions as essential driving forces for RNA folding. Future applications of the RACER model include predicting free energy landscapes of more complex RNA structures and use of RACER for multiscale simulations.
Nnoaham, Kelechi E.; Hummelshoj, Lone; Kennedy, Stephen H.; Jenkinson, Crispin; Zondervan, Krina T.
2012-01-01
Objective To generate and validate symptom-based models to predict endometriosis among symptomatic women prior to undergoing their first laparoscopy. Design Prospective, observational, two-phase study, in which women completed a 25-item questionnaire prior to surgery. Setting Nineteen hospitals in 13 countries. Patient(s) Symptomatic women (n = 1,396) scheduled for laparoscopy without a previous surgical diagnosis of endometriosis. Intervention(s) None. Main Outcome Measure(s) Sensitivity and specificity of endometriosis diagnosis predicted by symptoms and patient characteristics from optimal models developed using multiple logistic regression analyses in one data set (phase I), and independently validated in a second data set (phase II) by receiver operating characteristic (ROC) curve analysis. Result(s) Three hundred sixty (46.7%) women in phase I and 364 (58.2%) in phase II were diagnosed with endometriosis at laparoscopy. Menstrual dyschezia (pain on opening bowels) and a history of benign ovarian cysts most strongly predicted both any and stage III and IV endometriosis in both phases. Prediction of any-stage endometriosis, although improved by ultrasound scan evidence of cyst/nodules, was relatively poor (area under the curve [AUC] = 68.3). Stage III and IV disease was predicted with good accuracy (AUC = 84.9, sensitivity of 82.3% and specificity 75.8% at an optimal cut-off of 0.24). Conclusion(s) Our symptom-based models predict any-stage endometriosis relatively poorly and stage III and IV disease with good accuracy. Predictive tools based on such models could help to prioritize women for surgical investigation in clinical practice and thus contribute to reducing time to diagnosis. We invite other researchers to validate the key models in additional populations. PMID:22657249
Sexton, Nicholas J; Cooper, Richard P
2017-05-01
Task inhibition (also known as backward inhibition) is an hypothesised form of cognitive inhibition evident in multi-task situations, with the role of facilitating switching between multiple, competing tasks. This article presents a novel cognitive computational model of a backward inhibition mechanism. By combining aspects of previous cognitive models in task switching and conflict monitoring, the model instantiates the theoretical proposal that backward inhibition is the direct result of conflict between multiple task representations. In a first simulation, we demonstrate that the model produces two effects widely observed in the empirical literature, specifically, reaction time costs for both (n-1) task switches and n-2 task repeats. Through a systematic search of parameter space, we demonstrate that these effects are a general property of the model's theoretical content, and not specific parameter settings. We further demonstrate that the model captures previously reported empirical effects of inter-trial interval on n-2 switch costs. A final simulation extends the paradigm of switching between tasks of asymmetric difficulty to three tasks, and generates novel predictions for n-2 repetition costs. Specifically, the model predicts that n-2 repetition costs associated with hard-easy-hard alternations are greater than for easy-hard-easy alternations. Finally, we report two behavioural experiments testing this hypothesis, with results consistent with the model predictions. Copyright © 2017 The Authors. Published by Elsevier Inc. All rights reserved.
Akrami, Mohammad; Qian, Zhihui; Zou, Zhemin; Howard, David; Nester, Chris J; Ren, Lei
2018-04-01
The objective of this study was to develop and validate a subject-specific framework for modelling the human foot. This was achieved by integrating medical image-based finite element modelling, individualised multi-body musculoskeletal modelling and 3D gait measurements. A 3D ankle-foot finite element model comprising all major foot structures was constructed based on MRI of one individual. A multi-body musculoskeletal model and 3D gait measurements for the same subject were used to define loading and boundary conditions. Sensitivity analyses were used to investigate the effects of key modelling parameters on model predictions. Prediction errors of average and peak plantar pressures were below 10% in all ten plantar regions at five key gait events with only one exception (lateral heel, in early stance, error of 14.44%). The sensitivity analyses results suggest that predictions of peak plantar pressures are moderately sensitive to material properties, ground reaction forces and muscle forces, and significantly sensitive to foot orientation. The maximum region-specific percentage change ratios (peak stress percentage change over parameter percentage change) were 1.935-2.258 for ground reaction forces, 1.528-2.727 for plantar flexor muscles and 4.84-11.37 for foot orientations. This strongly suggests that loading and boundary conditions need to be very carefully defined based on personalised measurement data.
Normand, Frédéric; Lauri, Pierre-Éric
2012-03-01
Accurate and reliable predictive models are necessary to estimate nondestructively key variables for plant growth studies such as leaf area and leaf, stem, and total biomass. Predictive models are lacking at the current-year branch scale despite the importance of this scale in plant science. We calibrated allometric models to estimate leaf area and stem and branch (leaves + stem) mass of current-year branches, i.e., branches several months old studied at the end of the vegetative growth season, of four mango cultivars on the basis of their basal cross-sectional area. The effects of year, site, and cultivar were tested. Models were validated with independent data and prediction accuracy was evaluated with the appropriate statistics. Models revealed a positive allometry between dependent and independent variables, whose y-intercept but not the slope, was affected by the cultivar. The effects of year and site were negligible. For each branch characteristic, cultivar-specific models were more accurate than common models built with pooled data from the four cultivars. Prediction quality was satisfactory but with data dispersion around the models, particularly for large values. Leaf area and stem and branch mass of mango current-year branches could be satisfactorily estimated on the basis of branch basal cross-sectional area with cultivar-specific allometric models. The results suggested that, in addition to the heteroscedastic behavior of the variables studied, model accuracy was probably related to the functional plasticity of branches in relation to the light environment and/or to the number of growth units composing the branches.
ERIC Educational Resources Information Center
Betts, Lucy R.; Elder, Tracey J.; Hartley, James; Blurton, Anthony
2008-01-01
Recent initiatives to enhance retention and widen participation ensure it is crucial to understand the factors that predict students' performance during their undergraduate degree. The present research used Structural Equation Modeling (SEM) to test three separate models that examined the extent to which British Psychology students' A-level entry…
Species-specific predictive models of developmental toxicity using the ToxCast chemical library
EPA’s ToxCastTM project is profiling the in vitro bioactivity of chemicals to generate predictive models that correlate with observed in vivo toxicity. In vitro profiling methods are based on ToxCast data, consisting of over 600 high-throughput screening (HTS) and high-content sc...
ERIC Educational Resources Information Center
Armenta, Brian E.; Hautala, Dane S.; Whitbeck, Les B.
2015-01-01
In the present study, we considered the utility of the prototype/willingness model in predicting alcohol use among North-American Indigenous adolescents. Specifically, using longitudinal data, we examined the associations among subjective drinking norms, positive drinker prototypes, drinking expectations (as a proxy of drinking willingness), and…
Limiting conditions for decay in wood systems
Paul I. Morris; Jerrold E. Winandy
2002-01-01
Hygrothermal models can predict temperature and moisture conditions in wall components subjected to real weather data, but specific data and a fundamental understanding of how temperature and wood moisture content dictate the progression of decay under these conditions is required for modellers to predict consequences of decay on building performance. It is well...
Vibration analysis of the SA349/2 helicopter
NASA Technical Reports Server (NTRS)
Heffernan, Ruth; Precetti, Dominique; Johnson, Wayne
1991-01-01
Helicopter airframe vibration is examined using calculations and measurements for the SA349/2 research helicopter. The hub loads, which transmit excitations to the fuselage, are predicted using a comprehensive rotorcraft analysis and correlated with measuring hub loads. The predicted and measured hub loads are then coupled with finite element models representing the SA349/2 fuselage. The resulting vertical acceleration at the pilot seat is examined. Adjustments are made to the airframe structural models to examine the sensitivity of predicted vertical acceleration to the model. Changes of a few percent to the damping and frequency of specific models lead to large reductions in predicted vibration, and to major improvements in the correlations with measured pilot-seat vertical acceleration.
Reaction Time Correlations during Eye–Hand Coordination:Behavior and Modeling
Dean, Heather L.; Martí, Daniel; Tsui, Eva; Rinzel, John; Pesaran, Bijan
2011-01-01
During coordinated eye– hand movements, saccade reaction times (SRTs) and reach reaction times (RRTs) are correlated in humans and monkeys. Reaction times (RTs) measure the degree of movement preparation and can correlate with movement speed and accuracy. However, RTs can also reflect effector nonspecific influences, such as motivation and arousal. We use a combination of behavioral psychophysics and computational modeling to identify plausible mechanisms for correlations in SRTs and RRTs. To disambiguate nonspecific mechanisms from mechanisms specific to movement coordination, we introduce a dual-task paradigm in which a reach and a saccade are cued with a stimulus onset asynchrony (SOA). We then develop several variants of integrate-to-threshold models of RT, which postulate that responses are initiated when the neural activity encoding effector-specific movement preparation reaches a threshold. The integrator models formalize hypotheses about RT correlations and make predictions for how each RT should vary with SOA. To test these hypotheses, we trained three monkeys to perform the eye– hand SOA task and analyzed their SRTs and RRTs. In all three subjects, RT correlations decreased with increasing SOA duration. Additionally, mean SRT decreased with decreasing SOA, revealing facilitation of saccades with simultaneous reaches, as predicted by the model. These results are not consistent with the predictions of the models with common modulation or common input but are compatible with the predictions of a model with mutual excitation between two effector-specific integrators. We propose that RT correlations are not simply attributable to motivation and arousal and are a signature of coordination. PMID:21325507
Ripley, Beth; Kelil, Tatiana; Cheezum, Michael K.; Goncalves, Alexandra; Di Carli, Marcelo F.; Rybicki, Frank J.; Steigner, Mike; Mitsouras, Dimitrios; Blankstein, Ron
2017-01-01
Background 3D printing is a promising technique that may have applications in medicine, and there is expanding interest in the use of patient-specific 3D models to guide surgical interventions. Objective To determine the feasibility of using cardiac CT to print individual models of the aortic root complex for transcatheter aortic valve replacement (TAVR) planning as well as to determine the ability to predict paravalvular aortic regurgitation (PAR). Methods This retrospective study included 16 patients (9 with PAR identified on blinded interpretation of post-procedure trans-thoracic echocardiography and 7 age, sex, and valve size-matched controls with no PAR). 3D printed models of the aortic root were created from pre-TAVR cardiac computed tomography data. These models were fitted with printed valves and predictions regarding post-implant PAR were made using a light transmission test. Results Aortic root 3D models were highly accurate, with excellent agreement between annulus measurements made on 3D models and those made on corresponding 2D data (mean difference of −0.34 mm, 95% limits of agreement: ± 1.3 mm). The 3D printed valve models were within 0.1 mm of their designed dimensions. Examination of the fit of valves within patient-specific aortic root models correctly predicted PAR in 6 of 9 patients (6 true positive, 3 false negative) and absence of PAR in 5 of 7 patients (5 true negative, 2 false positive). Conclusions Pre-TAVR 3D-printing based on cardiac CT provides a unique patient-specific method to assess the physical interplay of the aortic root and implanted valves. With additional optimization, 3D models may complement traditional techniques used for predicting which patients are more likely to develop PAR. PMID:26732862
Beck, J D; Weintraub, J A; Disney, J A; Graves, R C; Stamm, J W; Kaste, L M; Bohannan, H M
1992-12-01
The purpose of this analysis is to compare three different statistical models for predicting children likely to be at risk of developing dental caries over a 3-yr period. Data are based on 4117 children who participated in the University of North Carolina Caries Risk Assessment Study, a longitudinal study conducted in the Aiken, South Carolina, and Portland, Maine areas. The three models differed with respect to either the types of variables included or the definition of disease outcome. The two "Prediction" models included both risk factor variables thought to cause dental caries and indicator variables that are associated with dental caries, but are not thought to be causal for the disease. The "Etiologic" model included only etiologic factors as variables. A dichotomous outcome measure--none or any 3-yr increment, was used in the "Any Risk Etiologic model" and the "Any Risk Prediction Model". Another outcome, based on a gradient measure of disease, was used in the "High Risk Prediction Model". The variables that are significant in these models vary across grades and sites, but are more consistent among the Etiologic model than the Predictor models. However, among the three sets of models, the Any Risk Prediction Models have the highest sensitivity and positive predictive values, whereas the High Risk Prediction Models have the highest specificity and negative predictive values. Considerations in determining model preference are discussed.
Wan, Eric Yuk Fai; Fong, Daniel Yee Tak; Fung, Colman Siu Cheung; Yu, Esther Yee Tak; Chin, Weng Yee; Chan, Anca Ka Chun; Lam, Cindy Lo Kuen
2017-06-01
This study aimed to develop and validate an all-cause mortality risk prediction model for Chinese primary care patients with type 2 diabetes mellitus(T2DM) in Hong Kong. A population-based retrospective cohort study was conducted on 132,462 Chinese patients who had received public primary care services during 2010. Each gender sample was randomly split on a 2:1 basis into derivation and validation cohorts and was followed-up for a median period of 5years. Gender-specific mortality risk prediction models showing the interaction effect between predictors and age were derived using Cox proportional hazards regression with forward stepwise approach. Developed models were compared with pre-existing models by Harrell's C-statistic and calibration plot using validation cohort. Common predictors of increased mortality risk in both genders included: age; smoking habit; diabetes duration; use of anti-hypertensive agents, insulin and lipid-lowering drugs; body mass index; hemoglobin A1c; systolic blood pressure(BP); total cholesterol to high-density lipoprotein-cholesterol ratio; urine albumin to creatinine ratio(urine ACR); and estimated glomerular filtration rate(eGFR). Prediction models showed better discrimination with Harrell"'s C-statistics of 0.768(males) and 0.782(females) and calibration power from the plots than previously established models. Our newly developed gender-specific models provide a more accurate predicted 5-year mortality risk for Chinese diabetic patients than other established models. Copyright © 2017 Elsevier Inc. All rights reserved.
Schoenmakers, Inez; Gousias, Petros; Jones, Kerry S; Prentice, Ann
2016-11-01
On a population basis, there is a gradual decline in vitamin D status (plasma 25(OH)D) throughout winter. We developed a mathematical model to predict the population winter plasma 25(OH)D concentration longitudinally, using age-specific values for 25(OH)D expenditure (25(OH)D 3 t 1/2 ), cross-sectional plasma 25(OH)D concentration and vitamin D intake (VDI) data from older (70+ years; n=492) and younger adults (18-69 years; n=448) participating in the UK National Diet and Nutrition Survey. From this model, the population VDI required to maintain the mean plasma 25(OH)D at a set concentration can be derived. As expected, both predicted and measured population 25(OH)D (mean (95%CI)) progressively declined from September to March (from 51 (40-61) to 38 (36-41)nmol/L (predicted) vs 38 (27-48)nmol/L (measured) in older people and from 59 (54-65) to 34 (31-37)nmol/L (predicted) vs 37 (31-44)nmol/L (measured) in younger people). The predicted and measured mean values closely matched. The predicted VDIs required to maintain mean winter plasma 25(OH)D at 50nmol/L at the population level were 10 (0-20) to 11 (9-14) and 11 (6-16) to 13(11-16)μg/d for older and younger adults, respectively dependent on the month. In conclusion, a prediction model accounting for 25(OH)D 3 t 1/2 , VDI and scaling factor for the 25(OH)D response to VDI, closely predicts measured population winter values. Refinements of this model may include specific scaling factors accounting for the 25(OH)D response at different VDIs and as influenced by body composition and specific values for 25(OH)D 3 t 1/2 dependent on host factors such as kidney function. This model may help to reduce the need for longitudinal measurements. Copyright © 2016 The Authors. Published by Elsevier Ltd.. All rights reserved.
Comparative Study of Shrinkage and Non-Shrinkage Model of Food Drying
NASA Astrophysics Data System (ADS)
Shahari, N.; Jamil, N.; Rasmani, KA.
2016-08-01
A single phase heat and mass model has always been used to represent the moisture and temperature distribution during the drying of food. Several effects of the drying process, such as physical and structural changes, have been considered in order to increase understanding of the movement of water and temperature. However, the comparison between the heat and mass equation with and without structural change (in terms of shrinkage), which can affect the accuracy of the prediction model, has been little investigated. In this paper, two mathematical models to describe the heat and mass transfer in food, with and without the assumption of structural change, were analysed. The equations were solved using the finite difference method. The converted coordinate system was introduced within the numerical computations for the shrinkage model. The result shows that the temperature with shrinkage predicts a higher temperature at a specific time compared to that of the non-shrinkage model. Furthermore, the predicted moisture content decreased faster at a specific time when the shrinkage effect was included in the model.
Predicting network modules of cell cycle regulators using relative protein abundance statistics.
Oguz, Cihan; Watson, Layne T; Baumann, William T; Tyson, John J
2017-02-28
Parameter estimation in systems biology is typically done by enforcing experimental observations through an objective function as the parameter space of a model is explored by numerical simulations. Past studies have shown that one usually finds a set of "feasible" parameter vectors that fit the available experimental data equally well, and that these alternative vectors can make different predictions under novel experimental conditions. In this study, we characterize the feasible region of a complex model of the budding yeast cell cycle under a large set of discrete experimental constraints in order to test whether the statistical features of relative protein abundance predictions are influenced by the topology of the cell cycle regulatory network. Using differential evolution, we generate an ensemble of feasible parameter vectors that reproduce the phenotypes (viable or inviable) of wild-type yeast cells and 110 mutant strains. We use this ensemble to predict the phenotypes of 129 mutant strains for which experimental data is not available. We identify 86 novel mutants that are predicted to be viable and then rank the cell cycle proteins in terms of their contributions to cumulative variability of relative protein abundance predictions. Proteins involved in "regulation of cell size" and "regulation of G1/S transition" contribute most to predictive variability, whereas proteins involved in "positive regulation of transcription involved in exit from mitosis," "mitotic spindle assembly checkpoint" and "negative regulation of cyclin-dependent protein kinase by cyclin degradation" contribute the least. These results suggest that the statistics of these predictions may be generating patterns specific to individual network modules (START, S/G2/M, and EXIT). To test this hypothesis, we develop random forest models for predicting the network modules of cell cycle regulators using relative abundance statistics as model inputs. Predictive performance is assessed by the areas under receiver operating characteristics curves (AUC). Our models generate an AUC range of 0.83-0.87 as opposed to randomized models with AUC values around 0.50. By using differential evolution and random forest modeling, we show that the model prediction statistics generate distinct network module-specific patterns within the cell cycle network.
Predicting herbicide and biocide concentrations in rivers across Switzerland
NASA Astrophysics Data System (ADS)
Wemyss, Devon; Honti, Mark; Stamm, Christian
2014-05-01
Pesticide concentrations vary strongly in space and time. Accordingly, intensive sampling is required to achieve a reliable quantification of pesticide pollution. As this requires substantial resources, loads and concentration ranges in many small and medium streams remain unknown. Here, we propose partially filling the information gap for herbicides and biocides by using a modelling approach that predicts stream concentrations without site-specific calibration simply based on generally available data like land use, discharge and nation-wide consumption data. The simple, conceptual model distinguishes herbicide losses from agricultural fields, private gardens and biocide losses from buildings (facades, roofs). The herbicide model is driven by river discharge and the applied herbicide mass; the biocide model requires precipitation and the footprint area of urban areas containing the biocide. The model approach allows for modelling concentrations across multiple catchments at the daily, or shorter, time scale and for small to medium-sized catchments (1 - 100 km2). Four high resolution sampling campaigns in the Swiss Plateau were used to calibrate the model parameters for six model compounds: atrazine, metolachlor, terbuthylazine, terbutryn, diuron and mecoprop. Five additional sampled catchments across Switzerland were used to directly compare the predicted to the measured concentrations. Analysis of the first results reveals a reasonable simulation of the concentration dynamics for specific rainfall events and across the seasons. Predicted concentration ranges are reasonable even without site-specific calibration. This indicates the transferability of the calibrated model directly to other areas. However, the results also demonstrate systematic biases in that the highest measured peaks were not attained by the model. Probable causes for these deviations are conceptual model limitations and input uncertainty (pesticide use intensity, local precipitation, etc.). Accordingly, the model will be conceptually improved. This presentation will present the model simulations and compare the performance of the original and the modified model versions. Finally, the model will be applied across approximately 50% of the catchments in the Swiss Plateau, where necessary input data is available and where the model concept can be reasonably applied.
UXO Burial Prediction Fidelity
2017-07-01
been developed to predict the initial penetration depth of underwater mines . SERDP would like to know if and how these existing mine models could be...designed for near-cylindrical mines —for munitions, however, projectile-specific drag, lift, and moment coefficients are needed for estimating...as inputs. Other models have been built to estimate these initial conditions for mines dropped into water. Can these mine models be useful for
An Interoceptive Predictive Coding Model of Conscious Presence
Seth, Anil K.; Suzuki, Keisuke; Critchley, Hugo D.
2011-01-01
We describe a theoretical model of the neurocognitive mechanisms underlying conscious presence and its disturbances. The model is based on interoceptive prediction error and is informed by predictive models of agency, general models of hierarchical predictive coding and dopaminergic signaling in cortex, the role of the anterior insular cortex (AIC) in interoception and emotion, and cognitive neuroscience evidence from studies of virtual reality and of psychiatric disorders of presence, specifically depersonalization/derealization disorder. The model associates presence with successful suppression by top-down predictions of informative interoceptive signals evoked by autonomic control signals and, indirectly, by visceral responses to afferent sensory signals. The model connects presence to agency by allowing that predicted interoceptive signals will depend on whether afferent sensory signals are determined, by a parallel predictive-coding mechanism, to be self-generated or externally caused. Anatomically, we identify the AIC as the likely locus of key neural comparator mechanisms. Our model integrates a broad range of previously disparate evidence, makes predictions for conjoint manipulations of agency and presence, offers a new view of emotion as interoceptive inference, and represents a step toward a mechanistic account of a fundamental phenomenological property of consciousness. PMID:22291673
An ensemble model of QSAR tools for regulatory risk assessment.
Pradeep, Prachi; Povinelli, Richard J; White, Shannon; Merrill, Stephen J
2016-01-01
Quantitative structure activity relationships (QSARs) are theoretical models that relate a quantitative measure of chemical structure to a physical property or a biological effect. QSAR predictions can be used for chemical risk assessment for protection of human and environmental health, which makes them interesting to regulators, especially in the absence of experimental data. For compatibility with regulatory use, QSAR models should be transparent, reproducible and optimized to minimize the number of false negatives. In silico QSAR tools are gaining wide acceptance as a faster alternative to otherwise time-consuming clinical and animal testing methods. However, different QSAR tools often make conflicting predictions for a given chemical and may also vary in their predictive performance across different chemical datasets. In a regulatory context, conflicting predictions raise interpretation, validation and adequacy concerns. To address these concerns, ensemble learning techniques in the machine learning paradigm can be used to integrate predictions from multiple tools. By leveraging various underlying QSAR algorithms and training datasets, the resulting consensus prediction should yield better overall predictive ability. We present a novel ensemble QSAR model using Bayesian classification. The model allows for varying a cut-off parameter that allows for a selection in the desirable trade-off between model sensitivity and specificity. The predictive performance of the ensemble model is compared with four in silico tools (Toxtree, Lazar, OECD Toolbox, and Danish QSAR) to predict carcinogenicity for a dataset of air toxins (332 chemicals) and a subset of the gold carcinogenic potency database (480 chemicals). Leave-one-out cross validation results show that the ensemble model achieves the best trade-off between sensitivity and specificity (accuracy: 83.8 % and 80.4 %, and balanced accuracy: 80.6 % and 80.8 %) and highest inter-rater agreement [kappa ( κ ): 0.63 and 0.62] for both the datasets. The ROC curves demonstrate the utility of the cut-off feature in the predictive ability of the ensemble model. This feature provides an additional control to the regulators in grading a chemical based on the severity of the toxic endpoint under study.
An ensemble model of QSAR tools for regulatory risk assessment
Pradeep, Prachi; Povinelli, Richard J.; White, Shannon; ...
2016-09-22
Quantitative structure activity relationships (QSARs) are theoretical models that relate a quantitative measure of chemical structure to a physical property or a biological effect. QSAR predictions can be used for chemical risk assessment for protection of human and environmental health, which makes them interesting to regulators, especially in the absence of experimental data. For compatibility with regulatory use, QSAR models should be transparent, reproducible and optimized to minimize the number of false negatives. In silico QSAR tools are gaining wide acceptance as a faster alternative to otherwise time-consuming clinical and animal testing methods. However, different QSAR tools often make conflictingmore » predictions for a given chemical and may also vary in their predictive performance across different chemical datasets. In a regulatory context, conflicting predictions raise interpretation, validation and adequacy concerns. To address these concerns, ensemble learning techniques in the machine learning paradigm can be used to integrate predictions from multiple tools. By leveraging various underlying QSAR algorithms and training datasets, the resulting consensus prediction should yield better overall predictive ability. We present a novel ensemble QSAR model using Bayesian classification. The model allows for varying a cut-off parameter that allows for a selection in the desirable trade-off between model sensitivity and specificity. The predictive performance of the ensemble model is compared with four in silico tools (Toxtree, Lazar, OECD Toolbox, and Danish QSAR) to predict carcinogenicity for a dataset of air toxins (332 chemicals) and a subset of the gold carcinogenic potency database (480 chemicals). Leave-one-out cross validation results show that the ensemble model achieves the best trade-off between sensitivity and specificity (accuracy: 83.8 % and 80.4 %, and balanced accuracy: 80.6 % and 80.8 %) and highest inter-rater agreement [kappa (κ): 0.63 and 0.62] for both the datasets. The ROC curves demonstrate the utility of the cut-off feature in the predictive ability of the ensemble model. In conclusion, this feature provides an additional control to the regulators in grading a chemical based on the severity of the toxic endpoint under study.« less
SU-F-R-46: Predicting Distant Failure in Lung SBRT Using Multi-Objective Radiomics Model
DOE Office of Scientific and Technical Information (OSTI.GOV)
Zhou, Z; Folkert, M; Iyengar, P
2016-06-15
Purpose: To predict distant failure in lung stereotactic body radiation therapy (SBRT) in early stage non-small cell lung cancer (NSCLC) by using a new multi-objective radiomics model. Methods: Currently, most available radiomics models use the overall accuracy as the objective function. However, due to data imbalance, a single object may not reflect the performance of a predictive model. Therefore, we developed a multi-objective radiomics model which considers both sensitivity and specificity as the objective functions simultaneously. The new model is used to predict distant failure in lung SBRT using 52 patients treated at our institute. Quantitative imaging features of PETmore » and CT as well as clinical parameters are utilized to build the predictive model. Image features include intensity features (9), textural features (12) and geometric features (8). Clinical parameters for each patient include demographic parameters (4), tumor characteristics (8), treatment faction schemes (4) and pretreatment medicines (6). The modelling procedure consists of two steps: extracting features from segmented tumors in PET and CT; and selecting features and training model parameters based on multi-objective. Support Vector Machine (SVM) is used as the predictive model, while a nondominated sorting-based multi-objective evolutionary computation algorithm II (NSGA-II) is used for solving the multi-objective optimization. Results: The accuracy for PET, clinical, CT, PET+clinical, PET+CT, CT+clinical, PET+CT+clinical are 71.15%, 84.62%, 84.62%, 85.54%, 82.69%, 84.62%, 86.54%, respectively. The sensitivities for the above seven combinations are 41.76%, 58.33%, 50.00%, 50.00%, 41.67%, 41.67%, 58.33%, while the specificities are 80.00%, 92.50%, 90.00%, 97.50%, 92.50%, 97.50%, 97.50%. Conclusion: A new multi-objective radiomics model for predicting distant failure in NSCLC treated with SBRT was developed. The experimental results show that the best performance can be obtained by combining all features.« less
Improved prediction of biochemical recurrence after radical prostatectomy by genetic polymorphisms.
Morote, Juan; Del Amo, Jokin; Borque, Angel; Ars, Elisabet; Hernández, Carlos; Herranz, Felipe; Arruza, Antonio; Llarena, Roberto; Planas, Jacques; Viso, María J; Palou, Joan; Raventós, Carles X; Tejedor, Diego; Artieda, Marta; Simón, Laureano; Martínez, Antonio; Rioja, Luis A
2010-08-01
Single nucleotide polymorphisms are inherited genetic variations that can predispose or protect individuals against clinical events. We hypothesized that single nucleotide polymorphism profiling may improve the prediction of biochemical recurrence after radical prostatectomy. We performed a retrospective, multi-institutional study of 703 patients treated with radical prostatectomy for clinically localized prostate cancer who had at least 5 years of followup after surgery. All patients were genotyped for 83 prostate cancer related single nucleotide polymorphisms using a low density oligonucleotide microarray. Baseline clinicopathological variables and single nucleotide polymorphisms were analyzed to predict biochemical recurrence within 5 years using stepwise logistic regression. Discrimination was measured by ROC curve AUC, specificity, sensitivity, predictive values, net reclassification improvement and integrated discrimination index. The overall biochemical recurrence rate was 35%. The model with the best fit combined 8 covariates, including the 5 clinicopathological variables prostate specific antigen, Gleason score, pathological stage, lymph node involvement and margin status, and 3 single nucleotide polymorphisms at the KLK2, SULT1A1 and TLR4 genes. Model predictive power was defined by 80% positive predictive value, 74% negative predictive value and an AUC of 0.78. The model based on clinicopathological variables plus single nucleotide polymorphisms showed significant improvement over the model without single nucleotide polymorphisms, as indicated by 23.3% net reclassification improvement (p = 0.003), integrated discrimination index (p <0.001) and likelihood ratio test (p <0.001). Internal validation proved model robustness (bootstrap corrected AUC 0.78, range 0.74 to 0.82). The calibration plot showed close agreement between biochemical recurrence observed and predicted probabilities. Predicting biochemical recurrence after radical prostatectomy based on clinicopathological data can be significantly improved by including patient genetic information. Copyright (c) 2010 American Urological Association Education and Research, Inc. Published by Elsevier Inc. All rights reserved.
Klimov, Sergey; Rida, Padmashree Cg; Aleskandarany, Mohammed A; Green, Andrew R; Ellis, Ian O; Janssen, Emiel Am; Rakha, Emad A; Aneja, Ritu
2017-09-05
Although distant metastasis (DM) in breast cancer (BC) is the most lethal form of recurrence and the most common underlying cause of cancer related deaths, the outcome following the development of DM is related to the site of metastasis. Triple negative BC (TNBC) is an aggressive form of BC characterised by early recurrences and high mortality. Athough multiple variables can be used to predict the risk of metastasis, few markers can predict the specific site of metastasis. This study aimed at identifying a biomarker signature to predict particular sites of DM in TNBC. A clinically annotated series of 322 TNBC were immunohistochemically stained with 133 biomarkers relevant to BC, to develop multibiomarker models for predicting metastasis to the bone, liver, lung and brain. Patients who experienced metastasis to each site were compared with those who did not, by gradually filtering the biomarker set via a two-tailed t-test and Cox univariate analyses. Biomarker combinations were finally ranked based on statistical significance, and evaluated in multivariable analyses. Our final models were able to stratify TNBC patients into high risk groups that showed over 5, 6, 7 and 8 times higher risk of developing metastasis to the bone, liver, lung and brain, respectively, than low-risk subgroups. These models for predicting site-specific metastasis retained significance following adjustment for tumour size, patient age and chemotherapy status. Our novel IHC-based biomarkers signatures, when assessed in primary TNBC tumours, enable prediction of specific sites of metastasis, and potentially unravel biomarkers previously unknown in site tropism.
The regionalization of national-scale SPARROW models for stream nutrients
Schwarz, Gregory E.; Alexander, Richard B.; Smith, Richard A.; Preston, Stephen D.
2011-01-01
This analysis modifies the parsimonious specification of recently published total nitrogen (TN) and total phosphorus (TP) national-scale SPAtially Referenced Regressions On Watershed attributes models to allow each model coefficient to vary geographically among three major river basins of the conterminous United States. Regionalization of the national models reduces the standard errors in the prediction of TN and TP loads, expressed as a percentage of the predicted load, by about 6 and 7%. We develop and apply a method for combining national-scale and regional-scale information to estimate a hybrid model that imposes cross-region constraints that limit regional variation in model coefficients, effectively reducing the number of free model parameters as compared to a collection of independent regional models. The hybrid TN and TP regional models have improved model fit relative to the respective national models, reducing the standard error in the prediction of loads, expressed as a percentage of load, by about 5 and 4%. Only 19% of the TN hybrid model coefficients and just 2% of the TP hybrid model coefficients show evidence of substantial regional specificity (more than ±100% deviation from the national model estimate). The hybrid models have much greater precision in the estimated coefficients than do the unconstrained regional models, demonstrating the efficacy of pooling information across regions to improve regional models.
Predictive and concurrent validity of the Braden scale in long-term care: a meta-analysis.
Wilchesky, Machelle; Lungu, Ovidiu
2015-01-01
Pressure ulcer prevention is an important long-term care (LTC) quality indicator. While the Braden Scale is a recommended risk assessment tool, there is a paucity of information specifically pertaining to its validity within the LTC setting. We, therefore, undertook a systematic review and meta-analysis comparing Braden Scale predictive and concurrent validity within this context. We searched the Medline, EMBASE, PsychINFO and PubMed databases from 1985-2014 for studies containing the requisite information to analyze tool validity. Our initial search yielded 3,773 articles. Eleven datasets emanating from nine published studies describing 40,361 residents met all meta-analysis inclusion criteria and were analyzed using random effects models. Pooled sensitivity, specificity, positive predictive value (PPV), and negative predictive values were 86%, 38%, 28%, and 93%, respectively. Specificity was poorer in concurrent samples as compared with predictive samples (38% vs. 72%), while PPV was low in both sample types (25 and 37%). Though random effects model results showed that the Scale had good overall predictive ability [RR, 4.33; 95% CI, 3.28-5.72], none of the concurrent samples were found to have "optimal" sensitivity and specificity. In conclusion, the appropriateness of the Braden Scale in LTC is questionable given its low specificity and PPV, in particular in concurrent validity studies. Future studies should further explore the extent to which the apparent low validity of the Scale in LTC is due to the choice of cutoff point and/or preventive strategies implemented by LTC staff as a matter of course. © 2015 by the Wound Healing Society.
Snitkin, Evan S; Dudley, Aimée M; Janse, Daniel M; Wong, Kaisheen; Church, George M; Segrè, Daniel
2008-01-01
Background Understanding the response of complex biochemical networks to genetic perturbations and environmental variability is a fundamental challenge in biology. Integration of high-throughput experimental assays and genome-scale computational methods is likely to produce insight otherwise unreachable, but specific examples of such integration have only begun to be explored. Results In this study, we measured growth phenotypes of 465 Saccharomyces cerevisiae gene deletion mutants under 16 metabolically relevant conditions and integrated them with the corresponding flux balance model predictions. We first used discordance between experimental results and model predictions to guide a stage of experimental refinement, which resulted in a significant improvement in the quality of the experimental data. Next, we used discordance still present in the refined experimental data to assess the reliability of yeast metabolism models under different conditions. In addition to estimating predictive capacity based on growth phenotypes, we sought to explain these discordances by examining predicted flux distributions visualized through a new, freely available platform. This analysis led to insight into the glycerol utilization pathway and the potential effects of metabolic shortcuts on model results. Finally, we used model predictions and experimental data to discriminate between alternative raffinose catabolism routes. Conclusions Our study demonstrates how a new level of integration between high throughput measurements and flux balance model predictions can improve understanding of both experimental and computational results. The added value of a joint analysis is a more reliable platform for specific testing of biological hypotheses, such as the catabolic routes of different carbon sources. PMID:18808699
Lunelli, Antonella; Pugliese, Andrea; Rizzo, Caterina
2009-07-01
Due to the recent emergence of H5N1 virus, the modelling of pandemic influenza has become a relevant issue. Here we present an SEIR model formulated to simulate a possible outbreak in Italy, analysing its structure and, more generally, the effect of including specific details into a model. These details regard population heterogeneities, such as age and spatial distribution, as well as stochasticity, that regulates the epidemic dynamics when the number of infectives is low. We discuss and motivate the specific modelling choices made when building the model and investigate how the model details influence the predicted dynamics. Our analysis may help in deciding which elements of complexity are worth including in the design of a deterministic model for pandemic influenza, in a balance between, on the one hand, keeping the model computationally efficient and the number of parameters low and, on the other hand, maintaining the necessary realistic features.
Psychopathy and Deviant Workplace Behavior: A Comparison of Two Psychopathy Models.
Carre, Jessica R; Mueller, Steven M; Schleicher, Karly M; Jones, Daniel N
2018-04-01
Although psychopathy is an interpersonally harmful construct, few studies have compared different psycho athy models in predicting different types of workplace deviance. We examined how the Triarchic Psychopathy Model (TRI-PM) and the Self-Report Psychopathy-Short Form (SRP-SF) predicted deviant workplace behaviors in two forms: sexual harassment and deviant work behaviors. Using structural equations modeling, the latent factor of psychopathy was predictive for both types of deviant workplace behavior. Specifically, the SRP-SF signif cantly predicted both measures of deviant workplace behavior. With respect to the TRI-PM, meanness and disinhibition significantly predicted higher scores of workplace deviance and workplace sexual harassment measures. Future research needs to investigate the influence of psychopathy on deviant workplace behaviors, and consider the measures they use when they investigate these constructs.
Franz, A; Triesch, J
2010-12-01
The perception of the unity of objects, their permanence when out of sight, and the ability to perceive continuous object trajectories even during occlusion belong to the first and most important capacities that infants have to acquire. Despite much research a unified model of the development of these abilities is still missing. Here we make an attempt to provide such a unified model. We present a recurrent artificial neural network that learns to predict the motion of stimuli occluding each other and that develops representations of occluded object parts. It represents completely occluded, moving objects for several time steps and successfully predicts their reappearance after occlusion. This framework allows us to account for a broad range of experimental data. Specifically, the model explains how the perception of object unity develops, the role of the width of the occluders, and it also accounts for differences between data for moving and stationary stimuli. We demonstrate that these abilities can be acquired by learning to predict the sensory input. The model makes specific predictions and provides a unifying framework that has the potential to be extended to other visual event categories. Copyright © 2010 Elsevier Inc. All rights reserved.
Beyond the three-component model of organizational commitment.
Solinger, Omar N; van Olffen, Woody; Roe, Robert A
2008-01-01
This article offers a conceptual critique of the three-component model (TCM) of organizational commitment (Allen & Meyer, 1990) and proposes a reconceptualization based on standard attitude theory. The authors use the attitude-behavior model by Eagly and Chaiken (1993) to demonstrate that the TCM combines fundamentally different attitudinal phenomena. They argue that general organizational commitment can best be understood as an attitude regarding the organization, while normative and continuance commitment are attitudes regarding specific forms of behavior (i.e., staying or leaving). The conceptual analysis shows that the TCM fails to qualify as general model of organizational commitment but instead represents a specific model for predicting turnover. The authors suggest that the use of the TCM be restricted to this purpose and that Eagly and Chaiken's model be adopted as a generic commitment model template from which a range of models for predicting specific organizational behaviors can be extracted. Finally, they discuss the definition and measurement of the organizational commitment attitude. Covering the affective, cognitive, and behavioral facets of this attitude helps to enhance construct validity and to differentiate the construct from other constructs. 2008 APA
Vyas, Urvi; Ghanouni, Pejman; Halpern, Casey H; Elias, Jeff; Pauly, Kim Butts
2016-09-01
In transcranial magnetic resonance-guided focused ultrasound (tcMRgFUS) treatments, the acoustic and spatial heterogeneity of the skull cause reflection, absorption, and scattering of the acoustic beams. These effects depend on skull-specific parameters and can lead to patient-specific thermal responses to the same transducer power. In this work, the authors develop a simulation tool to help predict these different experimental responses using 3D heterogeneous tissue models based on the subject CT images. The authors then validate and compare the predicted skull efficiencies to an experimental metric based on the subject thermal responses during tcMRgFUS treatments in a dataset of seventeen human subjects. Seventeen human head CT scans were used to create tissue acoustic models, simulating the effects of reflection, absorption, and scattering of the acoustic beam as it propagates through a heterogeneous skull. The hybrid angular spectrum technique was used to model the acoustic beam propagation of the InSightec ExAblate 4000 head transducer for each subject, yielding maps of the specific absorption rate (SAR). The simulation assumed the transducer was geometrically focused to the thalamus of each subject, and the focal SAR at the target was used as a measure of the simulated skull efficiency. Experimental skull efficiency for each subject was calculated using the thermal temperature maps from the tcMRgFUS treatments. Axial temperature images (with no artifacts) were reconstructed with a single baseline, corrected using a referenceless algorithm. The experimental skull efficiency was calculated by dividing the reconstructed temperature rise 8.8 s after sonication by the applied acoustic power. The simulated skull efficiency using individual-specific heterogeneous models predicts well (R(2) = 0.84) the experimental energy efficiency. This paper presents a simulation model to predict the variation in thermal responses measured in clinical ctMRGFYS treatments while being computationally feasible.
Vyas, Urvi; Ghanouni, Pejman; Halpern, Casey H.; Elias, Jeff; Pauly, Kim Butts
2016-01-01
Purpose: In transcranial magnetic resonance-guided focused ultrasound (tcMRgFUS) treatments, the acoustic and spatial heterogeneity of the skull cause reflection, absorption, and scattering of the acoustic beams. These effects depend on skull-specific parameters and can lead to patient-specific thermal responses to the same transducer power. In this work, the authors develop a simulation tool to help predict these different experimental responses using 3D heterogeneous tissue models based on the subject CT images. The authors then validate and compare the predicted skull efficiencies to an experimental metric based on the subject thermal responses during tcMRgFUS treatments in a dataset of seventeen human subjects. Methods: Seventeen human head CT scans were used to create tissue acoustic models, simulating the effects of reflection, absorption, and scattering of the acoustic beam as it propagates through a heterogeneous skull. The hybrid angular spectrum technique was used to model the acoustic beam propagation of the InSightec ExAblate 4000 head transducer for each subject, yielding maps of the specific absorption rate (SAR). The simulation assumed the transducer was geometrically focused to the thalamus of each subject, and the focal SAR at the target was used as a measure of the simulated skull efficiency. Experimental skull efficiency for each subject was calculated using the thermal temperature maps from the tcMRgFUS treatments. Axial temperature images (with no artifacts) were reconstructed with a single baseline, corrected using a referenceless algorithm. The experimental skull efficiency was calculated by dividing the reconstructed temperature rise 8.8 s after sonication by the applied acoustic power. Results: The simulated skull efficiency using individual-specific heterogeneous models predicts well (R2 = 0.84) the experimental energy efficiency. Conclusions: This paper presents a simulation model to predict the variation in thermal responses measured in clinical ctMRGFYS treatments while being computationally feasible. PMID:27587047
DOE Office of Scientific and Technical Information (OSTI.GOV)
Vyas, Urvi, E-mail: urvi.vyas@gmail.com; Ghanouni,
Purpose: In transcranial magnetic resonance-guided focused ultrasound (tcMRgFUS) treatments, the acoustic and spatial heterogeneity of the skull cause reflection, absorption, and scattering of the acoustic beams. These effects depend on skull-specific parameters and can lead to patient-specific thermal responses to the same transducer power. In this work, the authors develop a simulation tool to help predict these different experimental responses using 3D heterogeneous tissue models based on the subject CT images. The authors then validate and compare the predicted skull efficiencies to an experimental metric based on the subject thermal responses during tcMRgFUS treatments in a dataset of seventeen humanmore » subjects. Methods: Seventeen human head CT scans were used to create tissue acoustic models, simulating the effects of reflection, absorption, and scattering of the acoustic beam as it propagates through a heterogeneous skull. The hybrid angular spectrum technique was used to model the acoustic beam propagation of the InSightec ExAblate 4000 head transducer for each subject, yielding maps of the specific absorption rate (SAR). The simulation assumed the transducer was geometrically focused to the thalamus of each subject, and the focal SAR at the target was used as a measure of the simulated skull efficiency. Experimental skull efficiency for each subject was calculated using the thermal temperature maps from the tcMRgFUS treatments. Axial temperature images (with no artifacts) were reconstructed with a single baseline, corrected using a referenceless algorithm. The experimental skull efficiency was calculated by dividing the reconstructed temperature rise 8.8 s after sonication by the applied acoustic power. Results: The simulated skull efficiency using individual-specific heterogeneous models predicts well (R{sup 2} = 0.84) the experimental energy efficiency. Conclusions: This paper presents a simulation model to predict the variation in thermal responses measured in clinical ctMRGFYS treatments while being computationally feasible.« less
Manoharan, Prabu; Chennoju, Kiranmai; Ghoshal, Nanda
2015-07-01
BACE1 is an attractive target in Alzheimer's disease (AD) treatment. A rational drug design effort for the inhibition of BACE1 is actively pursued by researchers in both academic and pharmaceutical industries. This continued effort led to the steady accumulation of BACE1 crystal structures, co-complexed with different classes of inhibitors. This wealth of information is used in this study to develop target specific proteochemometric models and these models are exploited for predicting the prospective BACE1 inhibitors. The models developed in this study have performed excellently in predicting the computationally generated poses, separately obtained from single and ensemble docking approaches. The simple protein-ligand contact (SPLC) model outperforms other sophisticated high end models, in virtual screening performance, developed during this study. In an attempt to account for BACE1 protein active site flexibility information in predictive models, we included the change in the area of solvent accessible surface and the change in the volume of solvent accessible surface in our models. The ensemble and single receptor docking results obtained from this study indicate that the structural water mediated interactions improve the virtual screening results. Also, these waters are essential for recapitulating bioactive conformation during docking study. The proteochemometric models developed in this study can be used for the prediction of BACE1 inhibitors, during the early stage of AD drug discovery.
Mizuno, Kiyonori; Andrish, Jack T.; van den Bogert, Antonie J.; McLean, Scott G.
2009-01-01
While gender-based differences in knee joint anatomies/laxities are well documented, the potential for them to precipitate gender-dimorphic ACL loading and resultant injury risk has not been considered. To this end, we generated gender-specific models of ACL strain as a function of any six degrees of freedom (6DOF) knee joint load state via a combined cadaveric and analytical approach. Continuously varying joint forces and torques were applied to five male and five female cadaveric specimens and recorded along with synchronous knee flexion and ACL strain data. All data (~10,000 samples) were submitted to specimen-specific regression analyses, affording ACL strain predictions as a function of the combined 6 DOF knee loads. Following individual model verifications, generalized gender-specific models were generated and subjected to 6 DOF external load scenarios consistent with both a clinical examination and a dynamic sports maneuver. The ensuing model-based strain predictions were subsequently examined for gender-based discrepancies. Male and female specimen specific models predicted ACL strain within 0.51% ± 0.10% and 0.52% ± 0.07% of the measured data respectively, and explained more than 75% of the associated variance in each case. Predicted female ACL strains were also significantly larger than respective male values for both of simulated 6 DOF load scenarios. Outcomes suggest that the female ACL will rupture in response to comparatively smaller external load applications. Future work must address the underlying anatomical/laxity contributions to knee joint mechanical and resultant ACL loading, ultimately affording prevention strategies that may cater to individual joint vulnerabilities. PMID:19464897
Zador, Zsolt; Sperrin, Matthew; King, Andrew T
2016-01-01
Traumatic brain injury remains a global health problem. Understanding the relative importance of outcome predictors helps optimize our treatment strategies by informing assessment protocols, clinical decisions and trial designs. In this study we establish importance ranking for outcome predictors based on receiver operating indices to identify key predictors of outcome and create simple predictive models. We then explore the associations between key outcome predictors using Bayesian networks to gain further insight into predictor importance. We analyzed the corticosteroid randomization after significant head injury (CRASH) trial database of 10008 patients and included patients for whom demographics, injury characteristics, computer tomography (CT) findings and Glasgow Outcome Scale (GCS) were recorded (total of 13 predictors, which would be available to clinicians within a few hours following the injury in 6945 patients). Predictions of clinical outcome (death or severe disability at 6 months) were performed using logistic regression models with 5-fold cross validation. Predictive performance was measured using standardized partial area (pAUC) under the receiver operating curve (ROC) and we used Delong test for comparisons. Variable importance ranking was based on pAUC targeted at specificity (pAUCSP) and sensitivity (pAUCSE) intervals of 90-100%. Probabilistic associations were depicted using Bayesian networks. Complete AUC analysis showed very good predictive power (AUC = 0.8237, 95% CI: 0.8138-0.8336) for the complete model. Specificity focused importance ranking highlighted age, pupillary, motor responses, obliteration of basal cisterns/3rd ventricle and midline shift. Interestingly when targeting model sensitivity, the highest-ranking variables were age, severe extracranial injury, verbal response, hematoma on CT and motor response. Simplified models, which included only these key predictors, had similar performance (pAUCSP = 0.6523, 95% CI: 0.6402-0.6641 and pAUCSE = 0.6332, 95% CI: 0.62-0.6477) compared to the complete models (pAUCSP = 0.6664, 95% CI: 0.6543-0.679, pAUCSE = 0.6436, 95% CI: 0.6289-0.6585, de Long p value 0.1165 and 0.3448 respectively). Bayesian networks showed the predictors that did not feature in the simplified models were associated with those that did. We demonstrate that importance based variable selection allows simplified predictive models to be created while maintaining prediction accuracy. Variable selection targeting specificity confirmed key components of clinical assessment in TBI whereas sensitivity based ranking suggested extracranial injury as one of the important predictors. These results help refine our approach to head injury assessment, decision-making and outcome prediction targeted at model sensitivity and specificity. Bayesian networks proved to be a comprehensive tool for depicting probabilistic associations for key predictors giving insight into why the simplified model has maintained accuracy.
Cross-trial prediction of treatment outcome in depression: a machine learning approach.
Chekroud, Adam Mourad; Zotti, Ryan Joseph; Shehzad, Zarrar; Gueorguieva, Ralitza; Johnson, Marcia K; Trivedi, Madhukar H; Cannon, Tyrone D; Krystal, John Harrison; Corlett, Philip Robert
2016-03-01
Antidepressant treatment efficacy is low, but might be improved by matching patients to interventions. At present, clinicians have no empirically validated mechanisms to assess whether a patient with depression will respond to a specific antidepressant. We aimed to develop an algorithm to assess whether patients will achieve symptomatic remission from a 12-week course of citalopram. We used patient-reported data from patients with depression (n=4041, with 1949 completers) from level 1 of the Sequenced Treatment Alternatives to Relieve Depression (STAR*D; ClinicalTrials.gov, number NCT00021528) to identify variables that were most predictive of treatment outcome, and used these variables to train a machine-learning model to predict clinical remission. We externally validated the model in the escitalopram treatment group (n=151) of an independent clinical trial (Combining Medications to Enhance Depression Outcomes [COMED]; ClinicalTrials.gov, number NCT00590863). We identified 25 variables that were most predictive of treatment outcome from 164 patient-reportable variables, and used these to train the model. The model was internally cross-validated, and predicted outcomes in the STAR*D cohort with accuracy significantly above chance (64·6% [SD 3·2]; p<0·0001). The model was externally validated in the escitalopram treatment group (N=151) of COMED (accuracy 59·6%, p=0.043). The model also performed significantly above chance in a combined escitalopram-buproprion treatment group in COMED (n=134; accuracy 59·7%, p=0·023), but not in a combined venlafaxine-mirtazapine group (n=140; accuracy 51·4%, p=0·53), suggesting specificity of the model to underlying mechanisms. Building statistical models by mining existing clinical trial data can enable prospective identification of patients who are likely to respond to a specific antidepressant. Yale University. Copyright © 2016 Elsevier Ltd. All rights reserved.
Alishiri, Gholam Hossein; Bayat, Noushin; Fathi Ashtiani, Ali; Tavallaii, Seyed Abbas; Assari, Shervin; Moharamzad, Yashar
2008-01-01
The aim of this work was to develop two logistic regression models capable of predicting physical and mental health related quality of life (HRQOL) among rheumatoid arthritis (RA) patients. In this cross-sectional study which was conducted during 2006 in the outpatient rheumatology clinic of our university hospital, Short Form 36 (SF-36) was used for HRQOL measurements in 411 RA patients. A cutoff point to define poor versus good HRQOL was calculated using the first quartiles of SF-36 physical and mental component scores (33.4 and 36.8, respectively). Two distinct logistic regression models were used to derive predictive variables including demographic, clinical, and psychological factors. The sensitivity, specificity, and accuracy of each model were calculated. Poor physical HRQOL was positively associated with pain score, disease duration, monthly family income below 300 US$, comorbidity, patient global assessment of disease activity or PGA, and depression (odds ratios: 1.1; 1.004; 15.5; 1.1; 1.02; 2.08, respectively). The variables that entered into the poor mental HRQOL prediction model were monthly family income below 300 US$, comorbidity, PGA, and bodily pain (odds ratios: 6.7; 1.1; 1.01; 1.01, respectively). Optimal sensitivity and specificity were achieved at a cutoff point of 0.39 for the estimated probability of poor physical HRQOL and 0.18 for mental HRQOL. Sensitivity, specificity, and accuracy of the physical and mental models were 73.8, 87, 83.7% and 90.38, 70.36, 75.43%, respectively. The results show that the suggested models can be used to predict poor physical and mental HRQOL separately among RA patients using simple variables with acceptable accuracy. These models can be of use in the clinical decision-making of RA patients and to recognize patients with poor physical or mental HRQOL in advance, for better management.
Han, Dianwei; Zhang, Jun; Tang, Guiliang
2012-01-01
An accurate prediction of the pre-microRNA secondary structure is important in miRNA informatics. Based on a recently proposed model, nucleotide cyclic motifs (NCM), to predict RNA secondary structure, we propose and implement a Modified NCM (MNCM) model with a physics-based scoring strategy to tackle the problem of pre-microRNA folding. Our microRNAfold is implemented using a global optimal algorithm based on the bottom-up local optimal solutions. Our experimental results show that microRNAfold outperforms the current leading prediction tools in terms of True Negative rate, False Negative rate, Specificity, and Matthews coefficient ratio.
Fang, Yilin; Wilkins, Michael J; Yabusaki, Steven B; Lipton, Mary S; Long, Philip E
2012-12-01
Accurately predicting the interactions between microbial metabolism and the physical subsurface environment is necessary to enhance subsurface energy development, soil and groundwater cleanup, and carbon management. This study was an initial attempt to confirm the metabolic functional roles within an in silico model using environmental proteomic data collected during field experiments. Shotgun global proteomics data collected during a subsurface biostimulation experiment were used to validate a genome-scale metabolic model of Geobacter metallireducens-specifically, the ability of the metabolic model to predict metal reduction, biomass yield, and growth rate under dynamic field conditions. The constraint-based in silico model of G. metallireducens relates an annotated genome sequence to the physiological functions with 697 reactions controlled by 747 enzyme-coding genes. Proteomic analysis showed that 180 of the 637 G. metallireducens proteins detected during the 2008 experiment were associated with specific metabolic reactions in the in silico model. When the field-calibrated Fe(III) terminal electron acceptor process reaction in a reactive transport model for the field experiments was replaced with the genome-scale model, the model predicted that the largest metabolic fluxes through the in silico model reactions generally correspond to the highest abundances of proteins that catalyze those reactions. Central metabolism predicted by the model agrees well with protein abundance profiles inferred from proteomic analysis. Model discrepancies with the proteomic data, such as the relatively low abundances of proteins associated with amino acid transport and metabolism, revealed pathways or flux constraints in the in silico model that could be updated to more accurately predict metabolic processes that occur in the subsurface environment.
Downscaler Model for predicting daily air pollution
This model combines daily ozone and particulate matter monitoring and modeling data from across the U.S. to provide improved fine-scale estimates of air quality in communities and other specific locales.
NASA Technical Reports Server (NTRS)
Kirtman, Ben P.; Min, Dughong; Infanti, Johnna M.; Kinter, James L., III; Paolino, Daniel A.; Zhang, Qin; vandenDool, Huug; Saha, Suranjana; Mendez, Malaquias Pena; Becker, Emily;
2013-01-01
The recent US National Academies report "Assessment of Intraseasonal to Interannual Climate Prediction and Predictability" was unequivocal in recommending the need for the development of a North American Multi-Model Ensemble (NMME) operational predictive capability. Indeed, this effort is required to meet the specific tailored regional prediction and decision support needs of a large community of climate information users. The multi-model ensemble approach has proven extremely effective at quantifying prediction uncertainty due to uncertainty in model formulation, and has proven to produce better prediction quality (on average) then any single model ensemble. This multi-model approach is the basis for several international collaborative prediction research efforts, an operational European system and there are numerous examples of how this multi-model ensemble approach yields superior forecasts compared to any single model. Based on two NOAA Climate Test Bed (CTB) NMME workshops (February 18, and April 8, 2011) a collaborative and coordinated implementation strategy for a NMME prediction system has been developed and is currently delivering real-time seasonal-to-interannual predictions on the NOAA Climate Prediction Center (CPC) operational schedule. The hindcast and real-time prediction data is readily available (e.g., http://iridl.ldeo.columbia.edu/SOURCES/.Models/.NMME/) and in graphical format from CPC (http://origin.cpc.ncep.noaa.gov/products/people/wd51yf/NMME/index.html). Moreover, the NMME forecast are already currently being used as guidance for operational forecasters. This paper describes the new NMME effort, presents an overview of the multi-model forecast quality, and the complementary skill associated with individual models.
Collins, Gary S; Reitsma, Johannes B; Altman, Douglas G; Moons, Karel G M
2015-01-07
Prediction models are developed to aid health care providers in estimating the probability or risk that a specific disease or condition is present (diagnostic models) or that a specific event will occur in the future (prognostic models), to inform their decision making. However, the overwhelming evidence shows that the quality of reporting of prediction model studies is poor. Only with full and clear reporting of information on all aspects of a prediction model can risk of bias and potential usefulness of prediction models be adequately assessed. The Transparent Reporting of a multivariable prediction model for Individual Prognosis Or Diagnosis (TRIPOD) Initiative developed a set of recommendations for the reporting of studies developing, validating, or updating a prediction model, whether for diagnostic or prognostic purposes. This article describes how the TRIPOD Statement was developed. An extensive list of items based on a review of the literature was created, which was reduced after a Web based survey and revised during a three day meeting in June 2011 with methodologists, health care professionals, and journal editors. The list was refined during several meetings of the steering group and in e-mail discussions with the wider group of TRIPOD contributors. The resulting TRIPOD Statement is a checklist of 22 items, deemed essential for transparent reporting of a prediction model study. The TRIPOD Statement aims to improve the transparency of the reporting of a prediction model study regardless of the study methods used. The TRIPOD Statement is best used in conjunction with the TRIPOD explanation and elaboration document. To aid the editorial process and readers of prediction model studies, it is recommended that authors include a completed checklist in their submission (also available at www.tripod-statement.org).To encourage dissemination of the TRIPOD Statement, this article is freely accessible on the Annals of Internal Medicine Web site (www.annals.org) and will be also published in BJOG, British Journal of Cancer, British Journal of Surgery, BMC Medicine, The BMJ, Circulation, Diabetic Medicine, European Journal of Clinical Investigation, European Urology, and Journal of Clinical Epidemiology. The authors jointly hold the copyright of this article. An accompanying explanation and elaboration article is freely available only on www.annals.org; Annals of Internal Medicine holds copyright for that article. © BMJ Publishing Group Ltd 2014.
Wan, Eric Yuk Fai; Fong, Daniel Yee Tak; Fung, Colman Siu Cheung; Yu, Esther Yee Tak; Chin, Weng Yee; Chan, Anca Ka Chun; Lam, Cindy Lo Kuen
2017-08-01
Since diabetes mellitus (DM) is the leading cause of end stage renal disease (ESRD), this study aimed to develop a 5-year ESRD risk prediction model among Chinese patients with Type 2 DM (T2DM) in primary care. A retrospective cohort study was conducted on 149,333 Chinese adult T2DM primary care patients without ESRD in 2010. Using the derivation cohort over a median of 5 years follow-up, the gender-specific models including the interaction effect between predictors and age were derived using Cox regression with a forward stepwise approach. Harrell's C-statistic and calibration plot were applied to the validation cohort to assess discrimination and calibration of the models. Prediction models showed better discrimination with Harrell's C-statistics of 0.866 (males) and 0.862 (females) and calibration power from the plots than other established models. The predictors included age, usages of anti-hypertensive drugs, anti-glucose drugs, and Hemogloblin A1c, blood pressure, urine albumin/creatinine ratio (ACR) and estimated glomerular filtration rate (eGFR). Specific predictors for male were smoking and presence of sight threatening diabetic retinopathy while additional predictors for female included longer duration of diabetes and quadratic effect of body mass index. Interaction factors with age showed a greater weighting of insulin and urine ACR in younger males, and eGFR in younger females. Our newly developed gender-specific models provide a more accurate 5-year ESRD risk predictions for Chinese diabetic primary care patients than other existing models. The models included several modifiable risk factors that clinicians can use to counsel patients, and to target at in the delivery of care to patients.
Gupta, Rishi R; Gifford, Eric M; Liston, Ted; Waller, Chris L; Hohman, Moses; Bunin, Barry A; Ekins, Sean
2010-11-01
Ligand-based computational models could be more readily shared between researchers and organizations if they were generated with open source molecular descriptors [e.g., chemistry development kit (CDK)] and modeling algorithms, because this would negate the requirement for proprietary commercial software. We initially evaluated open source descriptors and model building algorithms using a training set of approximately 50,000 molecules and a test set of approximately 25,000 molecules with human liver microsomal metabolic stability data. A C5.0 decision tree model demonstrated that CDK descriptors together with a set of Smiles Arbitrary Target Specification (SMARTS) keys had good statistics [κ = 0.43, sensitivity = 0.57, specificity = 0.91, and positive predicted value (PPV) = 0.64], equivalent to those of models built with commercial Molecular Operating Environment 2D (MOE2D) and the same set of SMARTS keys (κ = 0.43, sensitivity = 0.58, specificity = 0.91, and PPV = 0.63). Extending the dataset to ∼193,000 molecules and generating a continuous model using Cubist with a combination of CDK and SMARTS keys or MOE2D and SMARTS keys confirmed this observation. When the continuous predictions and actual values were binned to get a categorical score we observed a similar κ statistic (0.42). The same combination of descriptor set and modeling method was applied to passive permeability and P-glycoprotein efflux data with similar model testing statistics. In summary, open source tools demonstrated predictive results comparable to those of commercial software with attendant cost savings. We discuss the advantages and disadvantages of open source descriptors and the opportunity for their use as a tool for organizations to share data precompetitively, avoiding repetition and assisting drug discovery.
Comas, Jorge; Benfeitas, Rui; Vilaprinyo, Ester; Sorribas, Albert; Solsona, Francesc; Farré, Gemma; Berman, Judit; Zorrilla, Uxue; Capell, Teresa; Sandmann, Gerhard; Zhu, Changfu; Christou, Paul; Alves, Rui
2016-09-01
Plant synthetic biology is still in its infancy. However, synthetic biology approaches have been used to manipulate and improve the nutritional and health value of staple food crops such as rice, potato and maize. With current technologies, production yields of the synthetic nutrients are a result of trial and error, and systematic rational strategies to optimize those yields are still lacking. Here, we present a workflow that combines gene expression and quantitative metabolomics with mathematical modeling to identify strategies for increasing production yields of nutritionally important carotenoids in the seed endosperm synthesized through alternative biosynthetic pathways in synthetic lines of white maize, which is normally devoid of carotenoids. Quantitative metabolomics and gene expression data are used to create and fit parameters of mathematical models that are specific to four independent maize lines. Sensitivity analysis and simulation of each model is used to predict which gene activities should be further engineered in order to increase production yields for carotenoid accumulation in each line. Some of these predictions (e.g. increasing Zmlycb/Gllycb will increase accumulated β-carotenes) are valid across the four maize lines and consistent with experimental observations in other systems. Other predictions are line specific. The workflow is adaptable to any other biological system for which appropriate quantitative information is available. Furthermore, we validate some of the predictions using experimental data from additional synthetic maize lines for which no models were developed. © 2016 The Authors The Plant Journal © 2016 John Wiley & Sons Ltd.
Predicting Trihalomethanes (THMs) in the New York City Water Supply
NASA Astrophysics Data System (ADS)
Mukundan, R.; Van Dreason, R.
2013-12-01
Chlorine, a commonly used disinfectant in most water supply systems, can combine with organic carbon to form disinfectant byproducts including carcinogenic trihalomethanes (THMs). We used water quality data from 24 monitoring sites within the New York City (NYC) water supply distribution system, measured between January 2009 and April 2012, to develop site-specific empirical models for predicting total trihalomethane (TTHM) levels. Terms in the model included various combinations of the following water quality parameters: total organic carbon, pH, specific conductivity, and water temperature. Reasonable estimates of TTHM levels were achieved with overall R2 of about 0.87 and predicted values within 5 μg/L of measured values. The relative importance of factors affecting TTHM formation was estimated by ranking the model regression coefficients. Site-specific models showed improved model performance statistics compared to a single model for the entire system most likely because the single model did not consider locational differences in the water treatment process. Although never out of compliance in 2011, the TTHM levels in the water supply increased following tropical storms Irene and Lee with 45% of the samples exceeding the 80 μg/L Maximum Contaminant Level (MCL) in October and November. This increase was explained by changes in water quality parameters, particularly by the increase in total organic carbon concentration and pH during this period.
Development of an accident duration prediction model on the Korean Freeway Systems.
Chung, Younshik
2010-01-01
Since duration prediction is one of the most important steps in an accident management process, there have been several approaches developed for modeling accident duration. This paper presents a model for the purpose of accident duration prediction based on accurately recorded and large accident dataset from the Korean Freeway Systems. To develop the duration prediction model, this study utilizes the log-logistic accelerated failure time (AFT) metric model and a 2-year accident duration dataset from 2006 to 2007. Specifically, the 2006 dataset is utilized to develop the prediction model and then, the 2007 dataset was employed to test the temporal transferability of the 2006 model. Although the duration prediction model has limitations such as large prediction error due to the individual differences of the accident treatment teams in terms of clearing similar accidents, the results from the 2006 model yielded a reasonable prediction based on the mean absolute percentage error (MAPE) scale. Additionally, the results of the statistical test for temporal transferability indicated that the estimated parameters in the duration prediction model are stable over time. Thus, this temporal stability suggests that the model may have potential to be used as a basis for making rational diversion and dispatching decisions in the event of an accident. Ultimately, such information will beneficially help in mitigating traffic congestion due to accidents.
On the predictive ability of mechanistic models for the Haitian cholera epidemic.
Mari, Lorenzo; Bertuzzo, Enrico; Finger, Flavio; Casagrandi, Renato; Gatto, Marino; Rinaldo, Andrea
2015-03-06
Predictive models of epidemic cholera need to resolve at suitable aggregation levels spatial data pertaining to local communities, epidemiological records, hydrologic drivers, waterways, patterns of human mobility and proxies of exposure rates. We address the above issue in a formal model comparison framework and provide a quantitative assessment of the explanatory and predictive abilities of various model settings with different spatial aggregation levels and coupling mechanisms. Reference is made to records of the recent Haiti cholera epidemics. Our intensive computations and objective model comparisons show that spatially explicit models accounting for spatial connections have better explanatory power than spatially disconnected ones for short-to-intermediate calibration windows, while parsimonious, spatially disconnected models perform better with long training sets. On average, spatially connected models show better predictive ability than disconnected ones. We suggest limits and validity of the various approaches and discuss the pathway towards the development of case-specific predictive tools in the context of emergency management. © 2015 The Author(s) Published by the Royal Society. All rights reserved.
Machine Learning Meta-analysis of Large Metagenomic Datasets: Tools and Biological Insights.
Pasolli, Edoardo; Truong, Duy Tin; Malik, Faizan; Waldron, Levi; Segata, Nicola
2016-07-01
Shotgun metagenomic analysis of the human associated microbiome provides a rich set of microbial features for prediction and biomarker discovery in the context of human diseases and health conditions. However, the use of such high-resolution microbial features presents new challenges, and validated computational tools for learning tasks are lacking. Moreover, classification rules have scarcely been validated in independent studies, posing questions about the generality and generalization of disease-predictive models across cohorts. In this paper, we comprehensively assess approaches to metagenomics-based prediction tasks and for quantitative assessment of the strength of potential microbiome-phenotype associations. We develop a computational framework for prediction tasks using quantitative microbiome profiles, including species-level relative abundances and presence of strain-specific markers. A comprehensive meta-analysis, with particular emphasis on generalization across cohorts, was performed in a collection of 2424 publicly available metagenomic samples from eight large-scale studies. Cross-validation revealed good disease-prediction capabilities, which were in general improved by feature selection and use of strain-specific markers instead of species-level taxonomic abundance. In cross-study analysis, models transferred between studies were in some cases less accurate than models tested by within-study cross-validation. Interestingly, the addition of healthy (control) samples from other studies to training sets improved disease prediction capabilities. Some microbial species (most notably Streptococcus anginosus) seem to characterize general dysbiotic states of the microbiome rather than connections with a specific disease. Our results in modelling features of the "healthy" microbiome can be considered a first step toward defining general microbial dysbiosis. The software framework, microbiome profiles, and metadata for thousands of samples are publicly available at http://segatalab.cibio.unitn.it/tools/metaml.
The circadian profile of epilepsy improves seizure forecasting.
Karoly, Philippa J; Ung, Hoameng; Grayden, David B; Kuhlmann, Levin; Leyde, Kent; Cook, Mark J; Freestone, Dean R
2017-08-01
It is now established that epilepsy is characterized by periodic dynamics that increase seizure likelihood at certain times of day, and which are highly patient-specific. However, these dynamics are not typically incorporated into seizure prediction algorithms due to the difficulty of estimating patient-specific rhythms from relatively short-term or unreliable data sources. This work outlines a novel framework to develop and assess seizure forecasts, and demonstrates that the predictive power of forecasting models is improved by circadian information. The analyses used long-term, continuous electrocorticography from nine subjects, recorded for an average of 320 days each. We used a large amount of out-of-sample data (a total of 900 days for algorithm training, and 2879 days for testing), enabling the most extensive post hoc investigation into seizure forecasting. We compared the results of an electrocorticography-based logistic regression model, a circadian probability, and a combined electrocorticography and circadian model. For all subjects, clinically relevant seizure prediction results were significant, and the addition of circadian information (combined model) maximized performance across a range of outcome measures. These results represent a proof-of-concept for implementing a circadian forecasting framework, and provide insight into new approaches for improving seizure prediction algorithms. The circadian framework adds very little computational complexity to existing prediction algorithms, and can be implemented using current-generation implant devices, or even non-invasively via surface electrodes using a wearable application. The ability to improve seizure prediction algorithms through straightforward, patient-specific modifications provides promise for increased quality of life and improved safety for patients with epilepsy. © The Author (2017). Published by Oxford University Press on behalf of the Guarantors of Brain. All rights reserved. For Permissions, please email: journals.permissions@oup.com.
Gröning, Flora; Jones, Marc E. H.; Curtis, Neil; Herrel, Anthony; O'Higgins, Paul; Evans, Susan E.; Fagan, Michael J.
2013-01-01
Computer-based simulation techniques such as multi-body dynamics analysis are becoming increasingly popular in the field of skull mechanics. Multi-body models can be used for studying the relationships between skull architecture, muscle morphology and feeding performance. However, to be confident in the modelling results, models need to be validated against experimental data, and the effects of uncertainties or inaccuracies in the chosen model attributes need to be assessed with sensitivity analyses. Here, we compare the bite forces predicted by a multi-body model of a lizard (Tupinambis merianae) with in vivo measurements, using anatomical data collected from the same specimen. This subject-specific model predicts bite forces that are very close to the in vivo measurements and also shows a consistent increase in bite force as the bite position is moved posteriorly on the jaw. However, the model is very sensitive to changes in muscle attributes such as fibre length, intrinsic muscle strength and force orientation, with bite force predictions varying considerably when these three variables are altered. We conclude that accurate muscle measurements are crucial to building realistic multi-body models and that subject-specific data should be used whenever possible. PMID:23614944
NASA Astrophysics Data System (ADS)
Loo, Lit-Hsin; Bougen-Zhukov, Nicola Michelle; Tan, Wei-Ling Cecilia
2017-03-01
Signaling pathways can generate different cellular responses to the same cytotoxic agents. Current quantitative models for predicting these differential responses are usually based on large numbers of intracellular gene products or signals at different levels of signaling cascades. Here, we report a study to predict cellular sensitivity to tumor necrosis factor alpha (TNFα) using high-throughput cellular imaging and machine-learning methods. We measured and compared 1170 protein phosphorylation events in a panel of human lung cancer cell lines based on different signals, subcellular regions, and time points within one hour of TNFα treatment. We found that two spatiotemporal-specific changes in an intermediate signaling protein, p90 ribosomal S6 kinase (RSK), are sufficient to predict the TNFα sensitivity of these cell lines. Our models could also predict the combined effects of TNFα and other kinase inhibitors, many of which are not known to target RSK directly. Therefore, early spatiotemporal-specific changes in intermediate signals are sufficient to represent the complex cellular responses to these perturbations. Our study provides a general framework for the development of rapid, signaling-based cytotoxicity screens that may be used to predict cellular sensitivity to a cytotoxic agent, or identify co-treatments that may sensitize or desensitize cells to the agent.
Loo, Lit-Hsin; Bougen-Zhukov, Nicola Michelle; Tan, Wei-Ling Cecilia
2017-01-01
Signaling pathways can generate different cellular responses to the same cytotoxic agents. Current quantitative models for predicting these differential responses are usually based on large numbers of intracellular gene products or signals at different levels of signaling cascades. Here, we report a study to predict cellular sensitivity to tumor necrosis factor alpha (TNFα) using high-throughput cellular imaging and machine-learning methods. We measured and compared 1170 protein phosphorylation events in a panel of human lung cancer cell lines based on different signals, subcellular regions, and time points within one hour of TNFα treatment. We found that two spatiotemporal-specific changes in an intermediate signaling protein, p90 ribosomal S6 kinase (RSK), are sufficient to predict the TNFα sensitivity of these cell lines. Our models could also predict the combined effects of TNFα and other kinase inhibitors, many of which are not known to target RSK directly. Therefore, early spatiotemporal-specific changes in intermediate signals are sufficient to represent the complex cellular responses to these perturbations. Our study provides a general framework for the development of rapid, signaling-based cytotoxicity screens that may be used to predict cellular sensitivity to a cytotoxic agent, or identify co-treatments that may sensitize or desensitize cells to the agent. PMID:28272488
Zeng, Fangfang; Li, Zhongtao; Yu, Xiaoling; Zhou, Linuo
2013-01-01
Background This study aimed to develop the artificial neural network (ANN) and multivariable logistic regression (LR) analyses for prediction modeling of cardiovascular autonomic (CA) dysfunction in the general population, and compare the prediction models using the two approaches. Methods and Materials We analyzed a previous dataset based on a Chinese population sample consisting of 2,092 individuals aged 30–80 years. The prediction models were derived from an exploratory set using ANN and LR analysis, and were tested in the validation set. Performances of these prediction models were then compared. Results Univariate analysis indicated that 14 risk factors showed statistically significant association with the prevalence of CA dysfunction (P<0.05). The mean area under the receiver-operating curve was 0.758 (95% CI 0.724–0.793) for LR and 0.762 (95% CI 0.732–0.793) for ANN analysis, but noninferiority result was found (P<0.001). The similar results were found in comparisons of sensitivity, specificity, and predictive values in the prediction models between the LR and ANN analyses. Conclusion The prediction models for CA dysfunction were developed using ANN and LR. ANN and LR are two effective tools for developing prediction models based on our dataset. PMID:23940593
Predicting the Emplacement of Improvised Explosive Devices: An Innovative Solution
ERIC Educational Resources Information Center
Lerner, Warren D.
2013-01-01
In this quantitative correlational study, simulated data were employed to examine artificial-intelligence techniques or, more specifically, artificial neural networks, as they relate to the location prediction of improvised explosive devices (IEDs). An ANN model was developed to predict IED placement, based upon terrain features and objects…
Descriptive modelling to predict deoxynivalenol in winter wheat in the Netherlands.
Van Der Fels-Klerx, H J; Burgers, S L G E; Booij, C J H
2010-05-01
Predictions of deoxynivalenol (DON) content in wheat at harvest can be useful for decision-making by stakeholders of the wheat feed and food supply chain. The objective of the current research was to develop quantitative predictive models for DON in mature winter wheat in the Netherlands for two specific groups of end-users. One model was developed for use by farmers in underpinning Fusarium spp. disease management, specifically the application of fungicides around wheat flowering (model A). The second model was developed for industry and food safety authorities, and considered the entire wheat cultivation period (model B). Model development was based on observational data collected from 425 fields throughout the Netherlands between 2001 and 2008. For each field, agronomical information, climatic data and DON levels in mature wheat were collected. Using multiple regression analyses, the set of biological relevant variables that provided the highest statistical performance was selected. The two final models include the following variables: region, wheat resistance level, spraying, flowering date, several climatic variables in the different stages of wheat growing, and length of the period between flowering and harvesting (model B only). The percentages of variance accounted for were 64.4% and 65.6% for models A and B, respectively. Model validation showed high correlation between the predicted and observed DON levels. The two models may be applied by various groups of end-users to reduce DON contamination in wheat-derived feed and food products and, ultimately, reduce animal and consumer health risks.
Fang, Xin; Reifman, Jaques; Wallqvist, Anders
2014-10-01
The human malaria parasite Plasmodium falciparum goes through a complex life cycle, including a roughly 48-hour-long intraerythrocytic developmental cycle (IDC) in human red blood cells. A better understanding of the metabolic processes required during the asexual blood-stage reproduction will enhance our basic knowledge of P. falciparum and help identify critical metabolic reactions and pathways associated with blood-stage malaria. We developed a metabolic network model that mechanistically links time-dependent gene expression, metabolism, and stage-specific growth, allowing us to predict the metabolic fluxes, the biomass production rates, and the timing of production of the different biomass components during the IDC. We predicted time- and stage-specific production of precursors and macromolecules for P. falciparum (strain HB3), allowing us to link specific metabolites to specific physiological functions. For example, we hypothesized that coenzyme A might be involved in late-IDC DNA replication and cell division. Moreover, the predicted ATP metabolism indicated that energy was mainly produced from glycolysis and utilized for non-metabolic processes. Finally, we used the model to classify the entire tricarboxylic acid cycle into segments, each with a distinct function, such as superoxide detoxification, glutamate/glutamine processing, and metabolism of fumarate as a byproduct of purine biosynthesis. By capturing the normal metabolic and growth progression in P. falciparum during the IDC, our model provides a starting point for further elucidation of strain-specific metabolic activity, host-parasite interactions, stress-induced metabolic responses, and metabolic responses to antimalarial drugs and drug candidates.
Predicting drug-induced liver injury in human with Naïve Bayes classifier approach.
Zhang, Hui; Ding, Lan; Zou, Yi; Hu, Shui-Qing; Huang, Hai-Guo; Kong, Wei-Bao; Zhang, Ji
2016-10-01
Drug-induced liver injury (DILI) is one of the major safety concerns in drug development. Although various toxicological studies assessing DILI risk have been developed, these methods were not sufficient in predicting DILI in humans. Thus, developing new tools and approaches to better predict DILI risk in humans has become an important and urgent task. In this study, we aimed to develop a computational model for assessment of the DILI risk with using a larger scale human dataset and Naïve Bayes classifier. The established Naïve Bayes prediction model was evaluated by 5-fold cross validation and an external test set. For the training set, the overall prediction accuracy of the 5-fold cross validation was 94.0 %. The sensitivity, specificity, positive predictive value and negative predictive value were 97.1, 89.2, 93.5 and 95.1 %, respectively. The test set with the concordance of 72.6 %, sensitivity of 72.5 %, specificity of 72.7 %, positive predictive value of 80.4 %, negative predictive value of 63.2 %. Furthermore, some important molecular descriptors related to DILI risk and some toxic/non-toxic fragments were identified. Thus, we hope the prediction model established here would be employed for the assessment of human DILI risk, and the obtained molecular descriptors and substructures should be taken into consideration in the design of new candidate compounds to help medicinal chemists rationally select the chemicals with the best prospects to be effective and safe.
NASA Technical Reports Server (NTRS)
Nese, Jon M.
1989-01-01
A dynamical systems approach is used to quantify the instantaneous and time-averaged predictability of a low-order moist general circulation model. Specifically, the effects on predictability of incorporating an active ocean circulation, implementing annual solar forcing, and asynchronously coupling the ocean and atmosphere are evaluated. The predictability and structure of the model attractors is compared using the Lyapunov exponents, the local divergence rates, and the correlation, fractal, and Lyapunov dimensions. The Lyapunov exponents measure the average rate of growth of small perturbations on an attractor, while the local divergence rates quantify phase-spatial variations of predictability. These local rates are exploited to efficiently identify and distinguish subtle differences in predictability among attractors. In addition, the predictability of monthly averaged and yearly averaged states is investigated by using attractor reconstruction techniques.
Proactive Supply Chain Performance Management with Predictive Analytics
Stefanovic, Nenad
2014-01-01
Today's business climate requires supply chains to be proactive rather than reactive, which demands a new approach that incorporates data mining predictive analytics. This paper introduces a predictive supply chain performance management model which combines process modelling, performance measurement, data mining models, and web portal technologies into a unique model. It presents the supply chain modelling approach based on the specialized metamodel which allows modelling of any supply chain configuration and at different level of details. The paper also presents the supply chain semantic business intelligence (BI) model which encapsulates data sources and business rules and includes the data warehouse model with specific supply chain dimensions, measures, and KPIs (key performance indicators). Next, the paper describes two generic approaches for designing the KPI predictive data mining models based on the BI semantic model. KPI predictive models were trained and tested with a real-world data set. Finally, a specialized analytical web portal which offers collaborative performance monitoring and decision making is presented. The results show that these models give very accurate KPI projections and provide valuable insights into newly emerging trends, opportunities, and problems. This should lead to more intelligent, predictive, and responsive supply chains capable of adapting to future business environment. PMID:25386605
Proactive supply chain performance management with predictive analytics.
Stefanovic, Nenad
2014-01-01
Today's business climate requires supply chains to be proactive rather than reactive, which demands a new approach that incorporates data mining predictive analytics. This paper introduces a predictive supply chain performance management model which combines process modelling, performance measurement, data mining models, and web portal technologies into a unique model. It presents the supply chain modelling approach based on the specialized metamodel which allows modelling of any supply chain configuration and at different level of details. The paper also presents the supply chain semantic business intelligence (BI) model which encapsulates data sources and business rules and includes the data warehouse model with specific supply chain dimensions, measures, and KPIs (key performance indicators). Next, the paper describes two generic approaches for designing the KPI predictive data mining models based on the BI semantic model. KPI predictive models were trained and tested with a real-world data set. Finally, a specialized analytical web portal which offers collaborative performance monitoring and decision making is presented. The results show that these models give very accurate KPI projections and provide valuable insights into newly emerging trends, opportunities, and problems. This should lead to more intelligent, predictive, and responsive supply chains capable of adapting to future business environment.
Jennings, Cecil A.; Sundmark, Aaron P.
2017-01-01
The relationships between environmental variables and the growth rates of fishes are important and rapidly expanding topics in fisheries ecology. We used an informationtheoretic approach to evaluate the influence of lake surface area and total phosphorus on the age-specific growth rates of Lepomis macrochirus (Bluegill) in 6 small impoundments in central Georgia. We used model averaging to create composite models and determine the relative importance of the variables within each model. Results indicated that surface area was the most important factor in the models predicting growth of Bluegills aged 1–4 years; total phosphorus was also an important predictor for the same age-classes. These results suggest that managers can use water quality and lake morphometry variables to create predictive models specific to their waterbody or region to help develop lake-specific management plans that select for and optimize local-level habitat factors for enhancing Bluegill growth.
Model Design for Military Advisors
2013-05-02
needs of their counterpart. This paper explores one area that would significantly improve advising outcomes; using advising models to match the...more specific. This paper develops three dominate models for advisors; the Stoic Acquaintance, the General Manger, and the Entertainer which can...then outcomes related to the individual counterpart’s developmental needs will be more predictable and specific. This paper will focus only on
Prediction of lake depth across a 17-state region in the United States
Oliver, Samantha K.; Soranno, Patricia A.; Fergus, C. Emi; Wagner, Tyler; Winslow, Luke A.; Scott, Caren E.; Webster, Katherine E.; Downing, John A.; Stanley, Emily H.
2016-01-01
Lake depth is an important characteristic for understanding many lake processes, yet it is unknown for the vast majority of lakes globally. Our objective was to develop a model that predicts lake depth using map-derived metrics of lake and terrestrial geomorphic features. Building on previous models that use local topography to predict lake depth, we hypothesized that regional differences in topography, lake shape, or sedimentation processes could lead to region-specific relationships between lake depth and the mapped features. We therefore used a mixed modeling approach that included region-specific model parameters. We built models using lake and map data from LAGOS, which includes 8164 lakes with maximum depth (Zmax) observations. The model was used to predict depth for all lakes ≥4 ha (n = 42 443) in the study extent. Lake surface area and maximum slope in a 100 m buffer were the best predictors of Zmax. Interactions between surface area and topography occurred at both the local and regional scale; surface area had a larger effect in steep terrain, so large lakes embedded in steep terrain were much deeper than those in flat terrain. Despite a large sample size and inclusion of regional variability, model performance (R2 = 0.29, RMSE = 7.1 m) was similar to other published models. The relative error varied by region, however, highlighting the importance of taking a regional approach to lake depth modeling. Additionally, we provide the largest known collection of observed and predicted lake depth values in the United States.
Denovan, Andrew; Dagnall, Neil; Drinkwater, Kenneth; Parker, Andrew; Clough, Peter
2017-01-01
The present study assessed the degree to which probabilistic reasoning performance and thinking style influenced perception of risk and self-reported levels of terrorism-related behavior change. A sample of 263 respondents, recruited via convenience sampling, completed a series of measures comprising probabilistic reasoning tasks (perception of randomness, base rate, probability, and conjunction fallacy), the Reality Testing subscale of the Inventory of Personality Organization (IPO-RT), the Domain-Specific Risk-Taking Scale, and a terrorism-related behavior change scale. Structural equation modeling examined three progressive models. Firstly, the Independence Model assumed that probabilistic reasoning, perception of risk and reality testing independently predicted terrorism-related behavior change. Secondly, the Mediation Model supposed that probabilistic reasoning and reality testing correlated, and indirectly predicted terrorism-related behavior change through perception of risk. Lastly, the Dual-Influence Model proposed that probabilistic reasoning indirectly predicted terrorism-related behavior change via perception of risk, independent of reality testing. Results indicated that performance on probabilistic reasoning tasks most strongly predicted perception of risk, and preference for an intuitive thinking style (measured by the IPO-RT) best explained terrorism-related behavior change. The combination of perception of risk with probabilistic reasoning ability in the Dual-Influence Model enhanced the predictive power of the analytical-rational route, with conjunction fallacy having a significant indirect effect on terrorism-related behavior change via perception of risk. The Dual-Influence Model possessed superior fit and reported similar predictive relations between intuitive-experiential and analytical-rational routes and terrorism-related behavior change. The discussion critically examines these findings in relation to dual-processing frameworks. This includes considering the limitations of current operationalisations and recommendations for future research that align outcomes and subsequent work more closely to specific dual-process models.
Wishart, Gordon C; Azzato, Elizabeth M; Greenberg, David C; Rashbass, Jem; Kearins, Olive; Lawrence, Gill; Caldas, Carlos; Pharoah, Paul D P
2010-01-01
The aim of this study was to develop and validate a prognostication model to predict overall and breast cancer specific survival for women treated for early breast cancer in the UK. Using the Eastern Cancer Registration and Information Centre (ECRIC) dataset, information was collated for 5,694 women who had surgery for invasive breast cancer in East Anglia from 1999 to 2003. Breast cancer mortality models for oestrogen receptor (ER) positive and ER negative tumours were derived from these data using Cox proportional hazards, adjusting for prognostic factors and mode of cancer detection (symptomatic versus screen-detected). An external dataset of 5,468 patients from the West Midlands Cancer Intelligence Unit (WMCIU) was used for validation. Differences in overall actual and predicted mortality were <1% at eight years for ECRIC (18.9% vs. 19.0%) and WMCIU (17.5% vs. 18.3%) with area under receiver-operator-characteristic curves (AUC) of 0.81 and 0.79 respectively. Differences in breast cancer specific actual and predicted mortality were <1% at eight years for ECRIC (12.9% vs. 13.5%) and <1.5% at eight years for WMCIU (12.2% vs. 13.6%) with AUC of 0.84 and 0.82 respectively. Model calibration was good for both ER positive and negative models although the ER positive model provided better discrimination (AUC 0.82) than ER negative (AUC 0.75). We have developed a prognostication model for early breast cancer based on UK cancer registry data that predicts breast cancer survival following surgery for invasive breast cancer and includes mode of detection for the first time. The model is well calibrated, provides a high degree of discrimination and has been validated in a second UK patient cohort.
NASA Astrophysics Data System (ADS)
Peng, Lanfang; Liu, Paiyu; Feng, Xionghan; Wang, Zimeng; Cheng, Tao; Liang, Yuzhen; Lin, Zhang; Shi, Zhenqing
2018-03-01
Predicting the kinetics of heavy metal adsorption and desorption in soil requires consideration of multiple heterogeneous soil binding sites and variations of reaction chemistry conditions. Although chemical speciation models have been developed for predicting the equilibrium of metal adsorption on soil organic matter (SOM) and important mineral phases (e.g. Fe and Al (hydr)oxides), there is still a lack of modeling tools for predicting the kinetics of metal adsorption and desorption reactions in soil. In this study, we developed a unified model for the kinetics of heavy metal adsorption and desorption in soil based on the equilibrium models WHAM 7 and CD-MUSIC, which specifically consider metal kinetic reactions with multiple binding sites of SOM and soil minerals simultaneously. For each specific binding site, metal adsorption and desorption rate coefficients were constrained by the local equilibrium partition coefficients predicted by WHAM 7 or CD-MUSIC, and, for each metal, the desorption rate coefficients of various binding sites were constrained by their metal binding constants with those sites. The model had only one fitting parameter for each soil binding phase, and all other parameters were derived from WHAM 7 and CD-MUSIC. A stirred-flow method was used to study the kinetics of Cd, Cu, Ni, Pb, and Zn adsorption and desorption in multiple soils under various pH and metal concentrations, and the model successfully reproduced most of the kinetic data. We quantitatively elucidated the significance of different soil components and important soil binding sites during the adsorption and desorption kinetic processes. Our model has provided a theoretical framework to predict metal adsorption and desorption kinetics, which can be further used to predict the dynamic behavior of heavy metals in soil under various natural conditions by coupling other important soil processes.
Denovan, Andrew; Dagnall, Neil; Drinkwater, Kenneth; Parker, Andrew; Clough, Peter
2017-01-01
The present study assessed the degree to which probabilistic reasoning performance and thinking style influenced perception of risk and self-reported levels of terrorism-related behavior change. A sample of 263 respondents, recruited via convenience sampling, completed a series of measures comprising probabilistic reasoning tasks (perception of randomness, base rate, probability, and conjunction fallacy), the Reality Testing subscale of the Inventory of Personality Organization (IPO-RT), the Domain-Specific Risk-Taking Scale, and a terrorism-related behavior change scale. Structural equation modeling examined three progressive models. Firstly, the Independence Model assumed that probabilistic reasoning, perception of risk and reality testing independently predicted terrorism-related behavior change. Secondly, the Mediation Model supposed that probabilistic reasoning and reality testing correlated, and indirectly predicted terrorism-related behavior change through perception of risk. Lastly, the Dual-Influence Model proposed that probabilistic reasoning indirectly predicted terrorism-related behavior change via perception of risk, independent of reality testing. Results indicated that performance on probabilistic reasoning tasks most strongly predicted perception of risk, and preference for an intuitive thinking style (measured by the IPO-RT) best explained terrorism-related behavior change. The combination of perception of risk with probabilistic reasoning ability in the Dual-Influence Model enhanced the predictive power of the analytical-rational route, with conjunction fallacy having a significant indirect effect on terrorism-related behavior change via perception of risk. The Dual-Influence Model possessed superior fit and reported similar predictive relations between intuitive-experiential and analytical-rational routes and terrorism-related behavior change. The discussion critically examines these findings in relation to dual-processing frameworks. This includes considering the limitations of current operationalisations and recommendations for future research that align outcomes and subsequent work more closely to specific dual-process models. PMID:29062288
A spectral clustering search algorithm for predicting shallow landslide size and location
Dino Bellugi; David G. Milledge; William E. Dietrich; Jim A. McKean; J. Taylor Perron; Erik B. Sudderth; Brian Kazian
2015-01-01
The potential hazard and geomorphic significance of shallow landslides depend on their location and size. Commonly applied one-dimensional stability models do not include lateral resistances and cannot predict landslide size. Multi-dimensional models must be applied to specific geometries, which are not known a priori, and testing all possible geometries is...
Morris, Melody K.; Saez-Rodriguez, Julio; Clarke, David C.; Sorger, Peter K.; Lauffenburger, Douglas A.
2011-01-01
Predictive understanding of cell signaling network operation based on general prior knowledge but consistent with empirical data in a specific environmental context is a current challenge in computational biology. Recent work has demonstrated that Boolean logic can be used to create context-specific network models by training proteomic pathway maps to dedicated biochemical data; however, the Boolean formalism is restricted to characterizing protein species as either fully active or inactive. To advance beyond this limitation, we propose a novel form of fuzzy logic sufficiently flexible to model quantitative data but also sufficiently simple to efficiently construct models by training pathway maps on dedicated experimental measurements. Our new approach, termed constrained fuzzy logic (cFL), converts a prior knowledge network (obtained from literature or interactome databases) into a computable model that describes graded values of protein activation across multiple pathways. We train a cFL-converted network to experimental data describing hepatocytic protein activation by inflammatory cytokines and demonstrate the application of the resultant trained models for three important purposes: (a) generating experimentally testable biological hypotheses concerning pathway crosstalk, (b) establishing capability for quantitative prediction of protein activity, and (c) prediction and understanding of the cytokine release phenotypic response. Our methodology systematically and quantitatively trains a protein pathway map summarizing curated literature to context-specific biochemical data. This process generates a computable model yielding successful prediction of new test data and offering biological insight into complex datasets that are difficult to fully analyze by intuition alone. PMID:21408212
Zastrow, Stefan; Brookman-May, Sabine; Cong, Thi Anh Phuong; Jurk, Stanislaw; von Bar, Immanuel; Novotny, Vladimir; Wirth, Manfred
2015-03-01
To predict outcome of patients with renal cell carcinoma (RCC) who undergo surgical therapy, risk models and nomograms are valuable tools. External validation on independent datasets is crucial for evaluating accuracy and generalizability of these models. The objective of the present study was to externally validate the postoperative nomogram developed by Karakiewicz et al. for prediction of cancer-specific survival. A total of 1,480 consecutive patients with a median follow-up of 82 months (IQR 46-128) were included into this analysis with 268 RCC-specific deaths. Nomogram-estimated survival probabilities were compared with survival probabilities of the actual cohort, and concordance indices were calculated. Calibration plots and decision curve analyses were used for evaluating calibration and clinical net benefit of the nomogram. Concordance between predictions of the nomogram and survival rates of the cohort was 0.911 after 12, 0.909 after 24 months and 0.896 after 60 months. Comparison of predicted probabilities and actual survival estimates with calibration plots showed an overestimation of tumor-specific survival based on nomogram predictions of high-risk patients, although calibration plots showed a reasonable calibration for probability ranges of interest. Decision curve analysis showed a positive net benefit of nomogram predictions for our patient cohort. The postoperative Karakiewicz nomogram provides a good concordance in this external cohort and is reasonably calibrated. It may overestimate tumor-specific survival in high-risk patients, which should be kept in mind when counseling patients. A positive net benefit of nomogram predictions was proven.
Carter, Jane V.; Roberts, Henry L.; Pan, Jianmin; Rice, Jonathan D.; Burton, James F.; Galbraith, Norman J.; Eichenberger, M. Robert; Jorden, Jeffery; Deveaux, Peter; Farmer, Russell; Williford, Anna; Kanaan, Ziad; Rai, Shesh N.; Galandiuk, Susan
2016-01-01
OBJECTIVE(S) Develop a plasma-based microRNA (miRNA) diagnostic assay specific for colorectal neoplasms, building upon our prior work. BACKGROUND Colorectal neoplasms (colorectal cancer [CRC] and colorectal advanced adenoma [CAA]) frequently develop in individuals at ages when other common cancers also occur. Current screening methods lack sensitivity, specificity, and have poor patient compliance. METHODS Plasma was screened for 380 miRNAs using microfluidic array technology from a “Training” cohort of 60 patients, (10 each) control, CRC, CAA, breast (BC), pancreatic (PC) and lung (LC) cancer. We identified uniquely dysregulated miRNAs specific for colorectal neoplasia (p<0.05, false discovery rate: 5%, adjusted α=0.0038). These miRNAs were evaluated using single assays in a “Test” cohort of 120 patients. A mathematical model was developed to predict blinded sample identity in a 150 patient “Validation” cohort using repeat-sub-sampling validation of the testing dataset with 1000 iterations each to assess model detection accuracy. RESULTS Seven miRNAs (miR-21, miR-29c, miR-122, miR-192, miR-346, miR-372, miR-374a) were selected based upon p-value, area-under-the-curve (AUC), fold-change, and biological plausibility. AUC (±95% CI) for “Test” cohort comparisons were 0.91 (0.85-0.96), 0.79 (0.70-0.88) and 0.98 (0.96-1.0), respectively. Our mathematical model predicted blinded sample identity with 69-77% accuracy between all neoplasia and controls, 67-76% accuracy between colorectal neoplasia and other cancers, and 86-90% accuracy between colorectal cancer and colorectal adenoma. CONCLUSIONS Our plasma miRNA assay and prediction model differentiates colorectal neoplasia from patients with other neoplasms and from controls with higher sensitivity and specificity compared to current clinical standards. PMID:27471839
R. B. Foltz; W. J. Elliot; N. S. Wagenbrenner
2011-01-01
Forested areas disturbed by access roads produce large amounts of sediment. One method to predict erosion and, hence, manage forest roads is the use of physically based soil erosion models. A perceived advantage of a physically based model is that it can be parameterized at one location and applied at another location with similar soil texture or geological parent...
Berthet, Pierre; Lansner, Anders
2014-01-01
Optogenetic stimulation of specific types of medium spiny neurons (MSNs) in the striatum has been shown to bias the selection of mice in a two choices task. This shift is dependent on the localisation and on the intensity of the stimulation but also on the recent reward history. We have implemented a way to simulate this increased activity produced by the optical flash in our computational model of the basal ganglia (BG). This abstract model features the direct and indirect pathways commonly described in biology, and a reward prediction pathway (RP). The framework is similar to Actor-Critic methods and to the ventral/dorsal distinction in the striatum. We thus investigated the impact on the selection caused by an added stimulation in each of the three pathways. We were able to reproduce in our model the bias in action selection observed in mice. Our results also showed that biasing the reward prediction is sufficient to create a modification in the action selection. However, we had to increase the percentage of trials with stimulation relative to that in experiments in order to impact the selection. We found that increasing only the reward prediction had a different effect if the stimulation in RP was action dependent (only for a specific action) or not. We further looked at the evolution of the change in the weights depending on the stage of learning within a block. A bias in RP impacts the plasticity differently depending on that stage but also on the outcome. It remains to experimentally test how the dopaminergic neurons are affected by specific stimulations of neurons in the striatum and to relate data to predictions of our model.
Genomic selection for crossbred performance accounting for breed-specific effects.
Lopes, Marcos S; Bovenhuis, Henk; Hidalgo, André M; van Arendonk, Johan A M; Knol, Egbert F; Bastiaansen, John W M
2017-06-26
Breed-specific effects are observed when the same allele of a given genetic marker has a different effect depending on its breed origin, which results in different allele substitution effects across breeds. In such a case, single-breed breeding values may not be the most accurate predictors of crossbred performance. Our aim was to estimate the contribution of alleles from each parental breed to the genetic variance of traits that are measured in crossbred offspring, and to compare the prediction accuracies of estimated direct genomic values (DGV) from a traditional genomic selection model (GS) that are trained on purebred or crossbred data, with accuracies of DGV from a model that accounts for breed-specific effects (BS), trained on purebred or crossbred data. The final dataset was composed of 924 Large White, 924 Landrace and 924 two-way cross (F1) genotyped and phenotyped animals. The traits evaluated were litter size (LS) and gestation length (GL) in pigs. The genetic correlation between purebred and crossbred performance was higher than 0.88 for both LS and GL. For both traits, the additive genetic variance was larger for alleles inherited from the Large White breed compared to alleles inherited from the Landrace breed (0.74 and 0.56 for LS, and 0.42 and 0.40 for GL, respectively). The highest prediction accuracies of crossbred performance were obtained when training was done on crossbred data. For LS, prediction accuracies were the same for GS and BS DGV (0.23), while for GL, prediction accuracy for BS DGV was similar to the accuracy of GS DGV (0.53 and 0.52, respectively). In this study, training on crossbred data resulted in higher prediction accuracy than training on purebred data and evidence of breed-specific effects for LS and GL was demonstrated. However, when training was done on crossbred data, both GS and BS models resulted in similar prediction accuracies. In future studies, traits with a lower genetic correlation between purebred and crossbred performance should be included to further assess the value of the BS model in genomic predictions.
Surgery on spinal epidural metastases (SEM) in renal cell carcinoma: a plea for a new paradigm.
Bakker, Nicolaas A; Coppes, Maarten H; Vergeer, Rob A; Kuijlen, Jos M A; Groen, Rob J M
2014-09-01
Prediction models for outcome of decompressive surgical resection of spinal epidural metastases (SEM) have in common that they have been developed for all types of SEM, irrespective of the type of primary tumor. It is our experience in clinical practice, however, that these models often fail to accurately predict outcome in the individual patient. To investigate whether decision making could be optimized by applying tumor-specific prediction models. For the proof of concept, we analyzed patients with SEM from renal cell carcinoma that we have operated on. Retrospective chart analysis 2006 to 2012. Twenty-one consecutive patients with symptomatic SEM of renal cell carcinoma. Predictive factors for survival. Next to established predictive factors for survival, we analyzed the predictive value of the Motzer criteria in these patients. The Motzer criteria comprise a specific and validated risk model for survival in patients with renal cell carcinoma. After multivariable analysis, only Motzer intermediate (hazard ratio [HR] 17.4, 95% confidence interval [CI] 1.82-166, p=.01) and high risk (HR 39.3, 95% CI 3.10-499, p=.005) turned out to be significantly associated with survival in patients with renal cell carcinoma that we have operated on. In this study, we have demonstrated that decision making could have been optimized by implementing the Motzer criteria next to established prediction models. We, therefore, suggest that in future, in patients with SEM from renal cell carcinoma, the Motzer criteria are also taken into account. Copyright © 2014 Elsevier Inc. All rights reserved.
Predictive Bcl-2 Family Binding Models Rooted in Experiment or Structure
DeBartolo, Joe; Dutta, Sanjib; Reich, Lothar; Keating, Amy E.
2013-01-01
Proteins of the Bcl-2 family either enhance or suppress programmed cell death and are centrally involved in cancer development and resistance to chemotherapy. BH3 (Bcl-2 homology 3)-only Bcl-2 proteins promote cell death by docking an α-helix into a hydrophobic groove on the surface of one or more of five pro-survival Bcl-2 receptor proteins. There is high structural homology within the pro-death and pro-survival families, yet a high degree of interaction specificity is nevertheless encoded, posing an interesting and important molecular recognition problem. Understanding protein features that dictate Bcl-2 interaction specificity is critical for designing peptide-based cancer therapeutics and diagnostics. In this study, we present peptide SPOT arrays and deep sequencing data from yeast display screening experiments that significantly expand the BH3 sequence space that has been experimentally tested for interaction with five human anti-apoptotic receptors. These data provide rich information about the determinants of Bcl-2 family specificity. To interpret and use the information, we constructed two simple data-based models that can predict affinity and specificity when evaluated on independent data sets within a limited sequence space. We also constructed a novel structure-based statistical potential, called STATIUM, which is remarkably good at predicting Bcl-2 affinity and specificity, especially considering it is not trained on experimental data. We compare the performance of our three models to each other and to alternative structure-based methods and discuss how such tools can guide prediction and design of new Bcl-2 family complexes. PMID:22617328
Stone, Wesley W.; Gilliom, Robert J.; Crawford, Charles G.
2008-01-01
Regression models were developed for predicting annual maximum and selected annual maximum moving-average concentrations of atrazine in streams using the Watershed Regressions for Pesticides (WARP) methodology developed by the National Water-Quality Assessment Program (NAWQA) of the U.S. Geological Survey (USGS). The current effort builds on the original WARP models, which were based on the annual mean and selected percentiles of the annual frequency distribution of atrazine concentrations. Estimates of annual maximum and annual maximum moving-average concentrations for selected durations are needed to characterize the levels of atrazine and other pesticides for comparison to specific water-quality benchmarks for evaluation of potential concerns regarding human health or aquatic life. Separate regression models were derived for the annual maximum and annual maximum 21-day, 60-day, and 90-day moving-average concentrations. Development of the regression models used the same explanatory variables, transformations, model development data, model validation data, and regression methods as those used in the original development of WARP. The models accounted for 72 to 75 percent of the variability in the concentration statistics among the 112 sampling sites used for model development. Predicted concentration statistics from the four models were within a factor of 10 of the observed concentration statistics for most of the model development and validation sites. Overall, performance of the models for the development and validation sites supports the application of the WARP models for predicting annual maximum and selected annual maximum moving-average atrazine concentration in streams and provides a framework to interpret the predictions in terms of uncertainty. For streams with inadequate direct measurements of atrazine concentrations, the WARP model predictions for the annual maximum and the annual maximum moving-average atrazine concentrations can be used to characterize the probable levels of atrazine for comparison to specific water-quality benchmarks. Sites with a high probability of exceeding a benchmark for human health or aquatic life can be prioritized for monitoring.
Musther, Helen; Harwood, Matthew D; Yang, Jiansong; Turner, David B; Rostami-Hodjegan, Amin; Jamei, Masoud
2017-09-01
The use of in vitro-in vivo extrapolation (IVIVE) techniques, mechanistically incorporated within physiologically based pharmacokinetic (PBPK) models, can harness in vitro drug data and enhance understanding of in vivo pharmacokinetics. This study's objective was to develop a user-friendly rat (250 g, male Sprague-Dawley) IVIVE-linked PBPK model. A 13-compartment PBPK model including mechanistic absorption models was developed, with required system data (anatomical, physiological, and relevant IVIVE scaling factors) collated from literature and analyzed. Overall, 178 system parameter values for the model are provided. This study also highlights gaps in available system data required for strain-specific rat PBPK model development. The model's functionality and performance were assessed using previous literature-sourced in vitro properties for diazepam, metoprolol, and midazolam. The results of simulations were compared against observed pharmacokinetic rat data. Predicted and observed concentration profiles in 10 tissues for diazepam after a single intravenous (i.v.) dose making use of either observed i.v. clearance (CL iv ) or in vitro hepatocyte intrinsic clearance (CL int ) for simulations generally led to good predictions in various tissue compartments. Overall, all i.v. plasma concentration profiles were successfully predicted. However, there were challenges in predicting oral plasma concentration profiles for metoprolol and midazolam, and the potential reasons and according solutions are discussed. Copyright © 2017 American Pharmacists Association®. Published by Elsevier Inc. All rights reserved.
A post audit of a model-designed ground water extraction system.
Andersen, Peter F; Lu, Silong
2003-01-01
Model post audits test the predictive capabilities of ground water models and shed light on their practical limitations. In the work presented here, ground water model predictions were used to design an extraction/treatment/injection system at a military ammunition facility and then were re-evaluated using site-specific water-level data collected approximately one year after system startup. The water-level data indicated that performance specifications for the design, i.e., containment, had been achieved over the required area, but that predicted water-level changes were greater than observed, particularly in the deeper zones of the aquifer. Probable model error was investigated by determining the changes that were required to obtain an improved match to observed water-level changes. This analysis suggests that the originally estimated hydraulic properties were in error by a factor of two to five. These errors may have resulted from attributing less importance to data from deeper zones of the aquifer and from applying pumping test results to a volume of material that was larger than the volume affected by the pumping test. To determine the importance of these errors to the predictions of interest, the models were used to simulate the capture zones resulting from the originally estimated and updated parameter values. The study suggests that, despite the model error, the ground water model contributed positively to the design of the remediation system.
Kowinsky, Amy M; Shovel, Judith; McLaughlin, Maribeth; Vertacnik, Lisa; Greenhouse, Pamela K; Martin, Susan Christie; Minnier, Tamra E
2012-01-01
Predictable and unpredictable patient care tasks compete for caregiver time and attention, making it difficult for patient care staff to reliably and consistently meet patient needs. We have piloted a redesigned care model that separates the work of patient care technicians based on task predictability and creates role specificity. This care model shows promise in improving the ability of staff to reliably complete tasks in a more consistent and timely manner.
Fear and Loving in Las Vegas: Evolution, Emotion, and Persuasion.
Griskevicius, Vladas; Goldstein, Noah J; Mortensen, Chad R; Sundie, Jill M; Cialdini, Robert B; Kenrick, Douglas T
2009-06-01
How do arousal-inducing contexts, such as frightening or romantic television programs, influence the effectiveness of basic persuasion heuristics? Different predictions are made by three theoretical models: A general arousal model predicts that arousal should increase effectiveness of heuristics; an affective valence model predicts that effectiveness should depend on whether the context elicits positive or negative affect; an evolutionary model predicts that persuasiveness should depend on both the specific emotion that is elicited and the content of the particular heuristic. Three experiments examined how fear-inducing versus romantic contexts influenced the effectiveness of two widely used heuristics-social proof (e.g., "most popular") and scarcity (e.g., "limited edition"). Results supported predictions from an evolutionary model, showing that fear can lead scarcity appeals to be counter-persuasive, and that romantic desire can lead social proof appeals to be counter-persuasive. The findings highlight how an evolutionary theoretical approach can lead to novel theoretical and practical marketing insights.
Badgett, Majors J; Boyes, Barry; Orlando, Ron
2018-02-16
A model that predicts retention for peptides using a HALO ® penta-HILIC column and gradient elution was created. Coefficients for each amino acid were derived using linear regression analysis and these coefficients can be summed to predict the retention of peptides. This model has a high correlation between experimental and predicted retention times (0.946), which is on par with previous RP and HILIC models. External validation of the model was performed using a set of H. pylori samples on the same LC-MS system used to create the model, and the deviation from actual to predicted times was low. Apart from amino acid composition, length and location of amino acid residues on a peptide were examined and two site-specific corrections for hydrophobic residues at the N-terminus as well as hydrophobic residues one spot over from the N-terminus were created. Copyright © 2017 Elsevier B.V. All rights reserved.
Baldwin, Mark A; Clary, Chadd; Maletsky, Lorin P; Rullkoetter, Paul J
2009-10-16
Verified computational models represent an efficient method for studying the relationship between articular geometry, soft-tissue constraint, and patellofemoral (PF) mechanics. The current study was performed to evaluate an explicit finite element (FE) modeling approach for predicting PF kinematics in the natural and implanted knee. Experimental three-dimensional kinematic data were collected on four healthy cadaver specimens in their natural state and after total knee replacement in the Kansas knee simulator during a simulated deep knee bend activity. Specimen-specific FE models were created from medical images and CAD implant geometry, and included soft-tissue structures representing medial-lateral PF ligaments and the quadriceps tendon. Measured quadriceps loads and prescribed tibiofemoral kinematics were used to predict dynamic kinematics of an isolated PF joint between 10 degrees and 110 degrees femoral flexion. Model sensitivity analyses were performed to determine the effect of rigid or deformable patellar representations and perturbed PF ligament mechanical properties (pre-tension and stiffness) on model predictions and computational efficiency. Predicted PF kinematics from the deformable analyses showed average root mean square (RMS) differences for the natural and implanted states of less than 3.1 degrees and 1.7 mm for all rotations and translations. Kinematic predictions with rigid bodies increased average RMS values slightly to 3.7 degrees and 1.9 mm with a five-fold decrease in computational time. Two-fold increases and decreases in PF ligament initial strain and linear stiffness were found to most adversely affect kinematic predictions for flexion, internal-external tilt and inferior-superior translation in both natural and implanted states. The verified models could be used to further investigate the effects of component alignment or soft-tissue variability on natural and implant PF mechanics.
Tredennick, Andrew T.; Bentley, Lisa Patrick; Hanan, Niall P.
2013-01-01
Theoretical models of allometric scaling provide frameworks for understanding and predicting how and why the morphology and function of organisms vary with scale. It remains unclear, however, if the predictions of ‘universal’ scaling models for vascular plants hold across diverse species in variable environments. Phenomena such as competition and disturbance may drive allometric scaling relationships away from theoretical predictions based on an optimized tree. Here, we use a hierarchical Bayesian approach to calculate tree-specific, species-specific, and ‘global’ (i.e. interspecific) scaling exponents for several allometric relationships using tree- and branch-level data harvested from three savanna sites across a rainfall gradient in Mali, West Africa. We use these exponents to provide a rigorous test of three plant scaling models (Metabolic Scaling Theory (MST), Geometric Similarity, and Stress Similarity) in savanna systems. For the allometric relationships we evaluated (diameter vs. length, aboveground mass, stem mass, and leaf mass) the empirically calculated exponents broadly overlapped among species from diverse environments, except for the scaling exponents for length, which increased with tree cover and density. When we compare empirical scaling exponents to the theoretical predictions from the three models we find MST predictions are most consistent with our observed allometries. In those situations where observations are inconsistent with MST we find that departure from theory corresponds with expected tradeoffs related to disturbance and competitive interactions. We hypothesize savanna trees have greater length-scaling exponents than predicted by MST due to an evolutionary tradeoff between fire escape and optimization of mechanical stability and internal resource transport. Future research on the drivers of systematic allometric variation could reconcile the differences between observed scaling relationships in variable ecosystems and those predicted by ideal models such as MST. PMID:23484003
Schneider, Markus; Rosam, Mathias; Glaser, Manuel; Patronov, Atanas; Shah, Harpreet; Back, Katrin Christiane; Daake, Marina Angelika; Buchner, Johannes; Antes, Iris
2016-10-01
Substrate binding to Hsp70 chaperones is involved in many biological processes, and the identification of potential substrates is important for a comprehensive understanding of these events. We present a multi-scale pipeline for an accurate, yet efficient prediction of peptides binding to the Hsp70 chaperone BiP by combining sequence-based prediction with molecular docking and MMPBSA calculations. First, we measured the binding of 15mer peptides from known substrate proteins of BiP by peptide array (PA) experiments and performed an accuracy assessment of the PA data by fluorescence anisotropy studies. Several sequence-based prediction models were fitted using this and other peptide binding data. A structure-based position-specific scoring matrix (SB-PSSM) derived solely from structural modeling data forms the core of all models. The matrix elements are based on a combination of binding energy estimations, molecular dynamics simulations, and analysis of the BiP binding site, which led to new insights into the peptide binding specificities of the chaperone. Using this SB-PSSM, peptide binders could be predicted with high selectivity even without training of the model on experimental data. Additional training further increased the prediction accuracies. Subsequent molecular docking (DynaDock) and MMGBSA/MMPBSA-based binding affinity estimations for predicted binders allowed the identification of the correct binding mode of the peptides as well as the calculation of nearly quantitative binding affinities. The general concept behind the developed multi-scale pipeline can readily be applied to other protein-peptide complexes with linearly bound peptides, for which sufficient experimental binding data for the training of classical sequence-based prediction models is not available. Proteins 2016; 84:1390-1407. © 2016 Wiley Periodicals, Inc. © 2016 Wiley Periodicals, Inc.
Carlisle, D.M.; Falcone, J.; Meador, M.R.
2009-01-01
We developed and evaluated empirical models to predict biological condition of wadeable streams in a large portion of the eastern USA, with the ultimate goal of prediction for unsampled basins. Previous work had classified (i.e., altered vs. unaltered) the biological condition of 920 streams based on a biological assessment of macroinvertebrate assemblages. Predictor variables were limited to widely available geospatial data, which included land cover, topography, climate, soils, societal infrastructure, and potential hydrologic modification. We compared the accuracy of predictions of biological condition class based on models with continuous and binary responses. We also evaluated the relative importance of specific groups and individual predictor variables, as well as the relationships between the most important predictors and biological condition. Prediction accuracy and the relative importance of predictor variables were different for two subregions for which models were created. Predictive accuracy in the highlands region improved by including predictors that represented both natural and human activities. Riparian land cover and road-stream intersections were the most important predictors. In contrast, predictive accuracy in the lowlands region was best for models limited to predictors representing natural factors, including basin topography and soil properties. Partial dependence plots revealed complex and nonlinear relationships between specific predictors and the probability of biological alteration. We demonstrate a potential application of the model by predicting biological condition in 552 unsampled basins across an ecoregion in southeastern Wisconsin (USA). Estimates of the likelihood of biological condition of unsampled streams could be a valuable tool for screening large numbers of basins to focus targeted monitoring of potentially unaltered or altered stream segments. ?? Springer Science+Business Media B.V. 2008.
Nazemi, S Majid; Amini, Morteza; Kontulainen, Saija A; Milner, Jaques S; Holdsworth, David W; Masri, Bassam A; Wilson, David R; Johnston, James D
2017-01-01
Quantitative computed tomography based subject-specific finite element modeling has potential to clarify the role of subchondral bone alterations in knee osteoarthritis initiation, progression, and pain. However, it is unclear what density-modulus equation(s) should be applied with subchondral cortical and subchondral trabecular bone when constructing finite element models of the tibia. Using a novel approach applying neural networks, optimization, and back-calculation against in situ experimental testing results, the objective of this study was to identify subchondral-specific equations that optimized finite element predictions of local structural stiffness at the proximal tibial subchondral surface. Thirteen proximal tibial compartments were imaged via quantitative computed tomography. Imaged bone mineral density was converted to elastic moduli using multiple density-modulus equations (93 total variations) then mapped to corresponding finite element models. For each variation, root mean squared error was calculated between finite element prediction and in situ measured stiffness at 47 indentation sites. Resulting errors were used to train an artificial neural network, which provided an unlimited number of model variations, with corresponding error, for predicting stiffness at the subchondral bone surface. Nelder-Mead optimization was used to identify optimum density-modulus equations for predicting stiffness. Finite element modeling predicted 81% of experimental stiffness variance (with 10.5% error) using optimized equations for subchondral cortical and trabecular bone differentiated with a 0.5g/cm 3 density. In comparison with published density-modulus relationships, optimized equations offered improved predictions of local subchondral structural stiffness. Further research is needed with anisotropy inclusion, a smaller voxel size and de-blurring algorithms to improve predictions. Copyright © 2016 Elsevier Ltd. All rights reserved.
A data-driven prediction method for fast-slow systems
NASA Astrophysics Data System (ADS)
Groth, Andreas; Chekroun, Mickael; Kondrashov, Dmitri; Ghil, Michael
2016-04-01
In this work, we present a prediction method for processes that exhibit a mixture of variability on low and fast scales. The method relies on combining empirical model reduction (EMR) with singular spectrum analysis (SSA). EMR is a data-driven methodology for constructing stochastic low-dimensional models that account for nonlinearity and serial correlation in the estimated noise, while SSA provides a decomposition of the complex dynamics into low-order components that capture spatio-temporal behavior on different time scales. Our study focuses on the data-driven modeling of partial observations from dynamical systems that exhibit power spectra with broad peaks. The main result in this talk is that the combination of SSA pre-filtering with EMR modeling improves, under certain circumstances, the modeling and prediction skill of such a system, as compared to a standard EMR prediction based on raw data. Specifically, it is the separation into "fast" and "slow" temporal scales by the SSA pre-filtering that achieves the improvement. We show, in particular that the resulting EMR-SSA emulators help predict intermittent behavior such as rapid transitions between specific regions of the system's phase space. This capability of the EMR-SSA prediction will be demonstrated on two low-dimensional models: the Rössler system and a Lotka-Volterra model for interspecies competition. In either case, the chaotic dynamics is produced through a Shilnikov-type mechanism and we argue that the latter seems to be an important ingredient for the good prediction skills of EMR-SSA emulators. Shilnikov-type behavior has been shown to arise in various complex geophysical fluid models, such as baroclinic quasi-geostrophic flows in the mid-latitude atmosphere and wind-driven double-gyre ocean circulation models. This pervasiveness of the Shilnikow mechanism of fast-slow transition opens interesting perspectives for the extension of the proposed EMR-SSA approach to more realistic situations.
Multivariate Statistical Models for Predicting Sediment Yields from Southern California Watersheds
Gartner, Joseph E.; Cannon, Susan H.; Helsel, Dennis R.; Bandurraga, Mark
2009-01-01
Debris-retention basins in Southern California are frequently used to protect communities and infrastructure from the hazards of flooding and debris flow. Empirical models that predict sediment yields are used to determine the size of the basins. Such models have been developed using analyses of records of the amount of material removed from debris retention basins, associated rainfall amounts, measures of watershed characteristics, and wildfire extent and history. In this study we used multiple linear regression methods to develop two updated empirical models to predict sediment yields for watersheds located in Southern California. The models are based on both new and existing measures of volume of sediment removed from debris retention basins, measures of watershed morphology, and characterization of burn severity distributions for watersheds located in Ventura, Los Angeles, and San Bernardino Counties. The first model presented reflects conditions in watersheds located throughout the Transverse Ranges of Southern California and is based on volumes of sediment measured following single storm events with known rainfall conditions. The second model presented is specific to conditions in Ventura County watersheds and was developed using volumes of sediment measured following multiple storm events. To relate sediment volumes to triggering storm rainfall, a rainfall threshold was developed to identify storms likely to have caused sediment deposition. A measured volume of sediment deposited by numerous storms was parsed among the threshold-exceeding storms based on relative storm rainfall totals. The predictive strength of the two models developed here, and of previously-published models, was evaluated using a test dataset consisting of 65 volumes of sediment yields measured in Southern California. The evaluation indicated that the model developed using information from single storm events in the Transverse Ranges best predicted sediment yields for watersheds in San Bernardino, Los Angeles, and Ventura Counties. This model predicts sediment yield as a function of the peak 1-hour rainfall, the watershed area burned by the most recent fire (at all severities), the time since the most recent fire, watershed area, average gradient, and relief ratio. The model that reflects conditions specific to Ventura County watersheds consistently under-predicted sediment yields and is not recommended for application. Some previously-published models performed reasonably well, while others either under-predicted sediment yields or had a larger range of errors in the predicted sediment yields.
Brown, Joshua W.
2009-01-01
The error likelihood computational model of anterior cingulate cortex (ACC) (Brown & Braver, 2005) has successfully predicted error likelihood effects, risk prediction effects, and how individual differences in conflict and error likelihood effects vary with trait differences in risk aversion. The same computational model now makes a further prediction that apparent conflict effects in ACC may result in part from an increasing number of simultaneously active responses, regardless of whether or not the cued responses are mutually incompatible. In Experiment 1, the model prediction was tested with a modification of the Eriksen flanker task, in which some task conditions require two otherwise mutually incompatible responses to be generated simultaneously. In that case, the two response processes are no longer in conflict with each other. The results showed small but significant medial PFC effects in the incongruent vs. congruent contrast, despite the absence of response conflict, consistent with model predictions. This is the multiple response effect. Nonetheless, actual response conflict led to greater ACC activation, suggesting that conflict effects are specific to particular task contexts. In Experiment 2, results from a change signal task suggested that the context dependence of conflict signals does not depend on error likelihood effects. Instead, inputs to ACC may reflect complex and task specific representations of motor acts, such as bimanual responses. Overall, the results suggest the existence of a richer set of motor signals monitored by medial PFC and are consistent with distinct effects of multiple responses, conflict, and error likelihood in medial PFC. PMID:19375509
Gonzalez-Buendia, Lucia; Delgado-Tirado, Santiago; Sanabria, M Rosa; Fernandez, Itziar; Coco, Rosa M
2017-08-18
To analyze predictors and develop predictive models of anatomic outcome in neovascular age-related macular degeneration (AMD) treated with as-needed ranibizumab after 4 years of follow-up. A multicenter consecutive case series non-interventional study was performed. Clinical, funduscopic and OCT characteristics of 194 treatment-naïve patients with AMD treated with as-needed ranibizumab for at least 2 years and up to 4 years were analyzed at baseline, 3 months and each year until the end of the follow-up. Baseline demographic and angiographic characteristics were also evaluated. R Statistical Software was used for statistical analysis. Main outcome measure was final anatomic status. Factors associated with less probability of preserved macula were diagnosis in 2009, older age, worse vision, presence of atrophy/fibrosis, pigment epithelium detachment, and geographic atrophy/fibrotic scar/neovascular AMD in the fellow eye. Factors associated with higher probability of GA were presence of atrophy and greater number of injections, whereas male sex, worse vision, lesser change in central macular thickness and presence of fibrosis were associated with less probability of GA as final macular status. Predictive model of preserved macula vs. GA/fibrotic scar showed sensibility of 77.78% and specificity of 69.09%. Predictive model of GA vs. fibrotic scar showed sensibility of 68.89% and specificity of 72.22%. We identified predictors of final macular status, and developed two predictive models. Predictive models that we propose are based on easily harvested variables, and, if validated, could be a useful tool for individual patient management and clinical research studies.
Testing mechanistic models of growth in insects.
Maino, James L; Kearney, Michael R
2015-11-22
Insects are typified by their small size, large numbers, impressive reproductive output and rapid growth. However, insect growth is not simply rapid; rather, insects follow a qualitatively distinct trajectory to many other animals. Here we present a mechanistic growth model for insects and show that increasing specific assimilation during the growth phase can explain the near-exponential growth trajectory of insects. The presented model is tested against growth data on 50 insects, and compared against other mechanistic growth models. Unlike the other mechanistic models, our growth model predicts energy reserves per biomass to increase with age, which implies a higher production efficiency and energy density of biomass in later instars. These predictions are tested against data compiled from the literature whereby it is confirmed that insects increase their production efficiency (by 24 percentage points) and energy density (by 4 J mg(-1)) between hatching and the attainment of full size. The model suggests that insects achieve greater production efficiencies and enhanced growth rates by increasing specific assimilation and increasing energy reserves per biomass, which are less costly to maintain than structural biomass. Our findings illustrate how the explanatory and predictive power of mechanistic growth models comes from their grounding in underlying biological processes. © 2015 The Author(s).
Quantitative Predictive Models for Systemic Toxicity (SOT)
Models to identify systemic and specific target organ toxicity were developed to help transition the field of toxicology towards computational models. By leveraging multiple data sources to incorporate read-across and machine learning approaches, a quantitative model of systemic ...
Kloog, Itai; Nordio, Francesco; Coull, Brent A; Schwartz, Joel
2012-11-06
Satellite-derived aerosol optical depth (AOD) measurements have the potential to provide spatiotemporally resolved predictions of both long and short-term exposures, but previous studies have generally shown moderate predictive power and lacked detailed high spatio- temporal resolution predictions across large domains. We aimed at extending our previous work by validating our model in another region with different geographical and metrological characteristics, and incorporating fine scale land use regression and nonrandom missingness to better predict PM(2.5) concentrations for days with or without satellite AOD measures. We start by calibrating AOD data for 2000-2008 across the Mid-Atlantic. We used mixed models regressing PM(2.5) measurements against day-specific random intercepts, and fixed and random AOD and temperature slopes. We used inverse probability weighting to account for nonrandom missingness of AOD, nested regions within days to capture spatial variation in the daily calibration, and introduced a penalization method that reduces the dimensionality of the large number of spatial and temporal predictors without selecting different predictors in different locations. We then take advantage of the association between grid-cell specific AOD values and PM(2.5) monitoring data, together with associations between AOD values in neighboring grid cells to develop grid cell predictions when AOD is missing. Finally to get local predictions (at the resolution of 50 m), we regressed the residuals from the predictions for each monitor from these previous steps against the local land use variables specific for each monitor. "Out-of-sample" 10-fold cross-validation was used to quantify the accuracy of our predictions at each step. For all days without AOD values, model performance was excellent (mean "out-of-sample" R(2) = 0.81, year-to-year variation 0.79-0.84). Upon removal of outliers in the PM(2.5) monitoring data, the results of the cross validation procedure was even better (overall mean "out of sample"R(2) of 0.85). Further, cross validation results revealed no bias in the predicted concentrations (Slope of observed vs predicted = 0.97-1.01). Our model allows one to reliably assess short-term and long-term human exposures in order to investigate both the acute and effects of ambient particles, respectively.
Predicting SVOC Emissions into Air and Foods in Support of ...
The release of semi-volatile organic compounds (SVOCs) from consumer articles may be a critical human exposure pathway. In addition, the migration of SVOCs from food packaging materials into foods may also be a dominant source of exposure for some chemicals. Here we describe recent efforts to characterize emission-related parameters for these exposure pathways to support prediction of aggregate exposures for thousands of chemicals For chemicals in consumer articles, Little et al. (2012) developed a screening-level indoor exposure prediction model which, for a given SVOC, principally depends on steady-state gas-phase concentrations (y0). We have developed a model that predicts y0 for SVOCs in consumer articles, allowing exposure predictions for 274 ToxCast chemicals. Published emissions data for 31 SVOCs found in flooring materials, provided a training set where both chemical-specific physicochemical properties, article specific formulation properties, and experimental design aspects were available as modeling descriptors. A linear regression yielded R2- and p- values of approximately 0.62 and 3.9E-05, respectively. A similar model was developed based upon physicochemical properties alone, since article information is often not available for a given SVOC or product. This latter model yielded R2 - and p- values of approximately 0.47 and 1.2E-10, respectively. Many SVOCs are also used as additives (e.g. plasticizers, antioxidants, lubricants) in plastic food pac
Poteat, V Paul; DiGiovanni, Craig D; Scheer, Jillian R
2013-03-01
As a form of bias-based harassment, homophobic behavior remains prominent in schools. Yet, little attention has been given to factors that underlie it, aside from bullying and sexual prejudice. Thus, we examined multiple domain general (empathy, perspective-taking, classroom respect norms) and sexual orientation-specific factors (sexual orientation identity importance, number of sexual minority friends, parents' sexual minority attitudes, media messages). We documented support for a model in which these sets of factors converged to predict homophobic behavior, mediated through bullying and prejudice, among 581 students in grades 9-12 (55 % female). The structural equation model indicated that, with the exception of media messages, these additional factors predicted levels of prejudice and bullying, which in turn predicted the likelihood of students to engage in homophobic behavior. These findings highlight the importance of addressing multiple interrelated factors in efforts to reduce bullying, prejudice, and discrimination among youth.
Tree-based modeling of complex interactions of phosphorus loadings and environmental factors.
Grunwald, S; Daroub, S H; Lang, T A; Diaz, O A
2009-06-01
Phosphorus (P) enrichment has been observed in the historic oligotrophic Greater Everglades in Florida mainly due to P influx from upstream, agriculturally dominated, low relief drainage basins of the Everglades Agricultural Area (EAA). Our specific objectives were to: (1) investigate relationships between various environmental factors and P loads in 10 farm basins within the EAA, (2) identify those environmental factors that impart major effects on P loads using three different tree-based modeling approaches, and (3) evaluate predictive models to assess P loads. We assembled thirteen environmental variable sets for all 10 sub-basins characterizing water level management, cropping practices, soils, hydrology, and farm-specific properties. Drainage flow and P concentrations were measured at each sub-basin outlet from 1992-2002 and aggregated to derive monthly P loads. We used three different tree-based models including single regression trees (ST), committee trees in Bagging (CTb) and ARCing (CTa) modes and ten-fold cross-validation to test prediction performances. The monthly P loads (MPL) during the monitoring period showed a maximum of 2528 kg (mean: 103 kg) and maximum monthly unit area P loads (UAL) of 4.88 kg P ha(-1) (mean: 0.16 kg P ha(-1)). Our results suggest that hydrologic/water management properties are the major controlling variables to predict MPL and UAL in the EAA. Tree-based modeling was successful in identifying relationships between P loads and environmental predictor variables on 10 farms in the EAA indicated by high R(2) (>0.80) and low prediction errors. Committee trees in ARCing mode generated the best performing models to predict P loads and P loads per unit area. Tree-based models had the ability to analyze complex, non-linear relationships between P loads and multiple variables describing hydrologic/water management, cropping practices, soil and farm-specific properties within the EAA.
Yang, Shaoyu; Chen, Xueqin; Pan, Yuelong; Yu, Jiekai; Li, Xin; Ma, Shenglin
2016-11-01
The present study aimed to identify potential serum biomarkers for predicting the clinical outcomes of patients with advanced non-small cell lung cancer (NSCLC) treated with epidermal growth factor receptor tyrosine kinase inhibitors (EGFR‑TKIs). A total of 61 samples were collected and analyzed using the integrated approach of magnetic bead‑based weak cation exchange chromatography and matrix‑assisted laser desorption/ionization‑time of flight‑mass spectrometry. The Zhejiang University Protein Chip Data Analysis system was used to identify the protein spectra of patients that are resistant and sensitive to EGFR‑TKIs. Furthermore, a support vector machine was used to construct a predictive model with high accuracy. The model was trained using 46 samples and tested with the remaining 15 samples. In addition, the ExPASy Bioinformatics Resource Portal was used to search potential candidate proteins for peaks in the predictive model. Seven mass/charge (m/z) peaks at 3,264, 9,156, 9,172, 3,964, 9,451, 4,295 and 3,983 Da, were identified as significantly different peaks between the EGFR‑TKIs sensitive and resistant groups. A predictive model was generated with three protein peaks at 3,264, 9,451 and 4,295 Da (m/z). This three‑peak model was capable of distinguishing EGFR‑TKIs resistant patients from sensitive patients with a specificity of 80% and a sensitivity of 80.77%. Furthermore, in a blind test, this model exhibited a high specificity (80%) and a high sensitivity (90%). Apelin, TYRO protein tyrosine kinase‑binding protein and big endothelin‑1 may be potential candidates for the proteins identified with an m/z of 3,264, 9,451 and 4,295 Da, respectively. The predictive model used in the present study may provide an improved understanding of the pathogenesis of NSCLC, and may provide insights for the development of TKI treatment plans tailored to specific patients.
Predictors of Cell Phone Use in Distracted Driving: Extending the Theory of Planned Behavior.
Tian, Yan; Robinson, James D
2017-09-01
This study examines the predictors of six distracted driving behaviors, and the survey data partially support Ajzen's (1991) Theory of Planned Behavior (TPB). The data suggest that the attitude variable predicted intention to engage in all six distracted driving behaviors (reading and sending text messages, making and answering cell phone calls, reading/viewing social media, and posting on social media while driving). Extending the model to include past experience and the variable perceived safety of technology yielded an improvement in the prediction of the distraction variables. Specifically, past experience predicted all six distracted driving behaviors, and the variable perceived safety of technology predicted intentions to read/view social media and intention to post on social media while driving. The study provides evidence for the importance of incorporating expanded variables into the original TPB model to predict cell phone use behaviors while driving, and it suggests that it is essential to tailor campaign materials for each specific cell phone use behavior to reduce distracted driving.
Residue-Specific Side-Chain Polymorphisms via Particle Belief Propagation.
Ghoraie, Laleh Soltan; Burkowski, Forbes; Li, Shuai Cheng; Zhu, Mu
2014-01-01
Protein side chains populate diverse conformational ensembles in crystals. Despite much evidence that there is widespread conformational polymorphism in protein side chains, most of the X-ray crystallography data are modeled by single conformations in the Protein Data Bank. The ability to extract or to predict these conformational polymorphisms is of crucial importance, as it facilitates deeper understanding of protein dynamics and functionality. In this paper, we describe a computational strategy capable of predicting side-chain polymorphisms. Our approach extends a particular class of algorithms for side-chain prediction by modeling the side-chain dihedral angles more appropriately as continuous rather than discrete variables. Employing a new inferential technique known as particle belief propagation, we predict residue-specific distributions that encode information about side-chain polymorphisms. Our predicted polymorphisms are in relatively close agreement with results from a state-of-the-art approach based on X-ray crystallography data, which characterizes the conformational polymorphisms of side chains using electron density information, and has successfully discovered previously unmodeled conformations.
Gupta, Sunil; Tran, Truyen; Luo, Wei; Phung, Dinh; Kennedy, Richard Lee; Broad, Adam; Campbell, David; Kipp, David; Singh, Madhu; Khasraw, Mustafa; Matheson, Leigh; Ashley, David M; Venkatesh, Svetha
2014-01-01
Objectives Using the prediction of cancer outcome as a model, we have tested the hypothesis that through analysing routinely collected digital data contained in an electronic administrative record (EAR), using machine-learning techniques, we could enhance conventional methods in predicting clinical outcomes. Setting A regional cancer centre in Australia. Participants Disease-specific data from a purpose-built cancer registry (Evaluation of Cancer Outcomes (ECO)) from 869 patients were used to predict survival at 6, 12 and 24 months. The model was validated with data from a further 94 patients, and results compared to the assessment of five specialist oncologists. Machine-learning prediction using ECO data was compared with that using EAR and a model combining ECO and EAR data. Primary and secondary outcome measures Survival prediction accuracy in terms of the area under the receiver operating characteristic curve (AUC). Results The ECO model yielded AUCs of 0.87 (95% CI 0.848 to 0.890) at 6 months, 0.796 (95% CI 0.774 to 0.823) at 12 months and 0.764 (95% CI 0.737 to 0.789) at 24 months. Each was slightly better than the performance of the clinician panel. The model performed consistently across a range of cancers, including rare cancers. Combining ECO and EAR data yielded better prediction than the ECO-based model (AUCs ranging from 0.757 to 0.997 for 6 months, AUCs from 0.689 to 0.988 for 12 months and AUCs from 0.713 to 0.973 for 24 months). The best prediction was for genitourinary, head and neck, lung, skin, and upper gastrointestinal tumours. Conclusions Machine learning applied to information from a disease-specific (cancer) database and the EAR can be used to predict clinical outcomes. Importantly, the approach described made use of digital data that is already routinely collected but underexploited by clinical health systems. PMID:24643167
Gupta, Sunil; Tran, Truyen; Luo, Wei; Phung, Dinh; Kennedy, Richard Lee; Broad, Adam; Campbell, David; Kipp, David; Singh, Madhu; Khasraw, Mustafa; Matheson, Leigh; Ashley, David M; Venkatesh, Svetha
2014-03-17
Using the prediction of cancer outcome as a model, we have tested the hypothesis that through analysing routinely collected digital data contained in an electronic administrative record (EAR), using machine-learning techniques, we could enhance conventional methods in predicting clinical outcomes. A regional cancer centre in Australia. Disease-specific data from a purpose-built cancer registry (Evaluation of Cancer Outcomes (ECO)) from 869 patients were used to predict survival at 6, 12 and 24 months. The model was validated with data from a further 94 patients, and results compared to the assessment of five specialist oncologists. Machine-learning prediction using ECO data was compared with that using EAR and a model combining ECO and EAR data. Survival prediction accuracy in terms of the area under the receiver operating characteristic curve (AUC). The ECO model yielded AUCs of 0.87 (95% CI 0.848 to 0.890) at 6 months, 0.796 (95% CI 0.774 to 0.823) at 12 months and 0.764 (95% CI 0.737 to 0.789) at 24 months. Each was slightly better than the performance of the clinician panel. The model performed consistently across a range of cancers, including rare cancers. Combining ECO and EAR data yielded better prediction than the ECO-based model (AUCs ranging from 0.757 to 0.997 for 6 months, AUCs from 0.689 to 0.988 for 12 months and AUCs from 0.713 to 0.973 for 24 months). The best prediction was for genitourinary, head and neck, lung, skin, and upper gastrointestinal tumours. Machine learning applied to information from a disease-specific (cancer) database and the EAR can be used to predict clinical outcomes. Importantly, the approach described made use of digital data that is already routinely collected but underexploited by clinical health systems.
A CFD Study on the Prediction of Cyclone Collection Efficiency
NASA Astrophysics Data System (ADS)
Gimbun, Jolius; Chuah, T. G.; Choong, Thomas S. Y.; Fakhru'L-Razi, A.
2005-09-01
This work presents a Computational Fluid Dynamics calculation to predict and to evaluate the effects of temperature, operating pressure and inlet velocity on the collection efficiency of gas cyclones. The numerical solutions were carried out using spreadsheet and commercial CFD code FLUENT 6.0. This paper also reviews four empirical models for the prediction of cyclone collection efficiency, namely Lapple [1], Koch and Licht [2], Li and Wang [3], and Iozia and Leith [4]. All the predictions proved to be satisfactory when compared with the presented experimental data. The CFD simulations predict the cyclone cut-off size for all operating conditions with a deviation of 3.7% from the experimental data. Specifically, results obtained from the computer modelling exercise have demonstrated that CFD model is the best method of modelling the cyclones collection efficiency.
Influence of nonelectrostatic ion-ion interactions on double-layer capacitance
NASA Astrophysics Data System (ADS)
Zhao, Hui
2012-11-01
Recently a Poisson-Helmholtz-Boltzmann (PHB) model [Bohinc , Phys. Rev. EPLEEE81539-375510.1103/PhysRevE.85.031130 85, 031130 (2012)] was developed by accounting for solvent-mediated nonelectrostatic ion-ion interactions. Nonelectrostatic interactions are described by a Yukawa-like pair potential. In the present work, we modify the PHB model by adding steric effects (finite ion size) into the free energy to derive governing equations. The modified PHB model is capable of capturing both ion specificity and ion crowding. This modified model is then employed to study the capacitance of the double layer. More specifically, we focus on the influence of nonelectrostatic ion-ion interactions on charging a double layer near a flat surface in the presence of steric effects. We numerically compute the differential capacitance as a function of the voltage under various conditions. At small voltages and low salt concentrations (dilute solution), we find out that the predictions from the modified PHB model are the same as those from the classical Poisson-Boltzmann theory, indicating that nonelectrostatic ion-ion interactions and steric effects are negligible. At moderate voltages, nonelectrostatic ion-ion interactions play an important role in determining the differential capacitance. Generally speaking, nonelectrostatic interactions decrease the capacitance because of additional nonelectrostatic repulsion among excess counterions inside the double layer. However, increasing the voltage gradually favors steric effects, which induce a condensed layer with crowding of counterions near the electrode. Accordingly, the predictions from the modified PHB model collapse onto those computed by the modified Poisson-Boltzmann theory considering steric effects alone. Finally, theoretical predictions are compared and favorably agree with experimental data, in particular, in concentrated solutions, leading one to conclude that the modified PHB model adequately predicts the diffuse-charge dynamics of the double layer with ion specificity and steric effects.
Radiatively induced neutrino mass model with flavor dependent gauge symmetry
NASA Astrophysics Data System (ADS)
Lee, SangJong; Nomura, Takaaki; Okada, Hiroshi
2018-06-01
We study a radiative seesaw model at one-loop level with a flavor dependent gauge symmetry U(1) μ - τ, in which we consider bosonic dark matter. We also analyze the constraints from lepton flavor violations, muon g - 2, relic density of dark matter, and collider physics, and carry out numerical analysis to search for allowed parameter region which satisfy all the constraints and to investigate some predictions. Furthermore we find that a simple but adhoc hypothesis induces specific two zero texture with inverse mass matrix, which provides us several predictions such as a specific pattern of Dirac CP phase.
Zhang, Li; Ai, Haixin; Chen, Wen; Yin, Zimo; Hu, Huan; Zhu, Junfeng; Zhao, Jian; Zhao, Qi; Liu, Hongsheng
2017-05-18
Carcinogenicity refers to a highly toxic end point of certain chemicals, and has become an important issue in the drug development process. In this study, three novel ensemble classification models, namely Ensemble SVM, Ensemble RF, and Ensemble XGBoost, were developed to predict carcinogenicity of chemicals using seven types of molecular fingerprints and three machine learning methods based on a dataset containing 1003 diverse compounds with rat carcinogenicity. Among these three models, Ensemble XGBoost is found to be the best, giving an average accuracy of 70.1 ± 2.9%, sensitivity of 67.0 ± 5.0%, and specificity of 73.1 ± 4.4% in five-fold cross-validation and an accuracy of 70.0%, sensitivity of 65.2%, and specificity of 76.5% in external validation. In comparison with some recent methods, the ensemble models outperform some machine learning-based approaches and yield equal accuracy and higher specificity but lower sensitivity than rule-based expert systems. It is also found that the ensemble models could be further improved if more data were available. As an application, the ensemble models are employed to discover potential carcinogens in the DrugBank database. The results indicate that the proposed models are helpful in predicting the carcinogenicity of chemicals. A web server called CarcinoPred-EL has been built for these models ( http://ccsipb.lnu.edu.cn/toxicity/CarcinoPred-EL/ ).
Toft-Petersen, A P; Ferrando-Vivas, P; Harrison, D A; Dunn, K; Rowan, K M
2018-05-15
In the UK, a network of specialist centres has been set up to provide critical care for burn patients. However, some burn patients are admitted to general intensive care units. Little is known about the casemix of these patients and how it compares with patients in specialist burn centres. It is not known whether burn-specific or generic risk prediction models perform better when applied to patients managed in intensive care units. We examined admissions for burns in the Case Mix Programme Database from April 2010 to March 2016. The casemix, activity and outcome in general and specialist burn intensive care units were compared and the fit of two burn-specific risk prediction models (revised Baux and Belgian Outcome in Burn Injury models) and one generic model (Intensive Care National Audit and Research Centre model) were compared. Patients in burn intensive care units had more extensive injuries compared with patients in general intensive care units (median (IQR [range]) burn surface area 16 (7-32 [0-98])% vs. 8 (1-18 [0-100])%, respectively) but in-hospital mortality was similar (22.8% vs. 19.0%, respectively). The discrimination and calibration of the generic Intensive Care National Audit and Research Centre model was superior to the revised Baux and Belgian Outcome in Burn Injury burn-specific models for patients managed on both specialist burn and general intensive care units. © 2018 The Association of Anaesthetists of Great Britain and Ireland.
Hannah, Iain; Montefiori, Erica; Modenese, Luca; Prinold, Joe; Viceconti, Marco; Mazzà, Claudia
2017-01-01
Subject-specific musculoskeletal modelling is especially useful in the study of juvenile and pathological subjects. However, such methodologies typically require a human operator to identify key landmarks from medical imaging data and are thus affected by unavoidable variability in the parameters defined and subsequent model predictions. The aim of this study was to thus quantify the inter- and intra-operator repeatability of a subject-specific modelling methodology developed for the analysis of subjects with juvenile idiopathic arthritis. Three operators each created subject-specific musculoskeletal foot and ankle models via palpation of bony landmarks, adjustment of geometrical muscle points and definition of joint coordinate systems. These models were then fused to a generic Arnold lower limb model for each of three modelled patients. The repeatability of each modelling operation was found to be comparable to those previously reported for the modelling of healthy, adult subjects. However, the inter-operator repeatability of muscle point definition was significantly greater than intra-operator repeatability (p < 0.05) and predicted ankle joint contact forces ranged by up to 24% and 10% of the peak force for the inter- and intra-operator analyses, respectively. Similarly, the maximum inter- and intra-operator variations in muscle force output were 64% and 23% of peak force, respectively. Our results suggest that subject-specific modelling is operator dependent at the foot and ankle, with the definition of muscle geometry the most significant source of output uncertainty. The development of automated procedures to prevent the misplacement of crucial muscle points should therefore be considered a particular priority for those developing subject-specific models. PMID:28427313
Hannah, Iain; Montefiori, Erica; Modenese, Luca; Prinold, Joe; Viceconti, Marco; Mazzà, Claudia
2017-05-01
Subject-specific musculoskeletal modelling is especially useful in the study of juvenile and pathological subjects. However, such methodologies typically require a human operator to identify key landmarks from medical imaging data and are thus affected by unavoidable variability in the parameters defined and subsequent model predictions. The aim of this study was to thus quantify the inter- and intra-operator repeatability of a subject-specific modelling methodology developed for the analysis of subjects with juvenile idiopathic arthritis. Three operators each created subject-specific musculoskeletal foot and ankle models via palpation of bony landmarks, adjustment of geometrical muscle points and definition of joint coordinate systems. These models were then fused to a generic Arnold lower limb model for each of three modelled patients. The repeatability of each modelling operation was found to be comparable to those previously reported for the modelling of healthy, adult subjects. However, the inter-operator repeatability of muscle point definition was significantly greater than intra-operator repeatability ( p < 0.05) and predicted ankle joint contact forces ranged by up to 24% and 10% of the peak force for the inter- and intra-operator analyses, respectively. Similarly, the maximum inter- and intra-operator variations in muscle force output were 64% and 23% of peak force, respectively. Our results suggest that subject-specific modelling is operator dependent at the foot and ankle, with the definition of muscle geometry the most significant source of output uncertainty. The development of automated procedures to prevent the misplacement of crucial muscle points should therefore be considered a particular priority for those developing subject-specific models.
A real-time prediction model for post-irradiation malignant cervical lymph nodes.
Lo, W-C; Cheng, P-W; Shueng, P-W; Hsieh, C-H; Chang, Y-L; Liao, L-J
2018-04-01
To establish a real-time predictive scoring model based on sonographic characteristics for identifying malignant cervical lymph nodes (LNs) in cancer patients after neck irradiation. One-hundred forty-four irradiation-treated patients underwent ultrasonography and ultrasound-guided fine-needle aspirations (USgFNAs), and the resultant data were used to construct a real-time and computerised predictive scoring model. This scoring system was further compared with our previously proposed prediction model. A predictive scoring model, 1.35 × (L axis) + 2.03 × (S axis) + 2.27 × (margin) + 1.48 × (echogenic hilum) + 3.7, was generated by stepwise multivariate logistic regression analysis. Neck LNs were considered to be malignant when the score was ≥ 7, corresponding to a sensitivity of 85.5%, specificity of 79.4%, positive predictive value (PPV) of 82.3%, negative predictive value (NPV) of 83.1%, and overall accuracy of 82.6%. When this new model and the original model were compared, the areas under the receiver operating characteristic curve (c-statistic) were 0.89 and 0.81, respectively (P < .05). A real-time sonographic predictive scoring model was constructed to provide prompt and reliable guidance for USgFNA biopsies to manage cervical LNs after neck irradiation. © 2017 John Wiley & Sons Ltd.
A Multivariate Model of Stakeholder Preference for Lethal Cat Management
Wald, Dara M.; Jacobson, Susan K.
2014-01-01
Identifying stakeholder beliefs and attitudes is critical for resolving management conflicts. Debate over outdoor cat management is often described as a conflict between two groups, environmental advocates and animal welfare advocates, but little is known about the variables predicting differences among these critical stakeholder groups. We administered a mail survey to randomly selected stakeholders representing both of these groups (n = 1,596) in Florida, where contention over the management of outdoor cats has been widespread. We used a structural equation model to evaluate stakeholder intention to support non-lethal management. The cognitive hierarchy model predicted that values influenced beliefs, which predicted general and specific attitudes, which in turn, influenced behavioral intentions. We posited that specific attitudes would mediate the effect of general attitudes, beliefs, and values on management support. Model fit statistics suggested that the final model fit the data well (CFI = 0.94, RMSEA = 0.062). The final model explained 74% of the variance in management support, and positive attitudes toward lethal management (humaneness) had the largest direct effect on management support. Specific attitudes toward lethal management and general attitudes toward outdoor cats mediated the relationship between positive (p<0.05) and negative cat-related impact beliefs (p<0.05) and support for management. These results supported the specificity hypothesis and the use of the cognitive hierarchy to assess stakeholder intention to support non-lethal cat management. Our findings suggest that stakeholders can simultaneously perceive both positive and negative beliefs about outdoor cats, which influence attitudes toward and support for non-lethal management. PMID:24736744
Calibration of a Land Subsidence Model Using InSAR Data via the Ensemble Kalman Filter.
Li, Liangping; Zhang, Meijing; Katzenstein, Kurt
2017-11-01
The application of interferometric synthetic aperture radar (InSAR) has been increasingly used to improve capabilities to model land subsidence in hydrogeologic studies. A number of investigations over the last decade show how spatially detailed time-lapse images of ground displacements could be utilized to advance our understanding for better predictions. In this work, we use simulated land subsidences as observed measurements, mimicking InSAR data to inversely infer inelastic specific storage in a stochastic framework. The inelastic specific storage is assumed as a random variable and modeled using a geostatistical method such that the detailed variations in space could be represented and also that the uncertainties of both characterization of specific storage and prediction of land subsidence can be assessed. The ensemble Kalman filter (EnKF), a real-time data assimilation algorithm, is used to inversely calibrate a land subsidence model by matching simulated subsidences with InSAR data. The performance of the EnKF is demonstrated in a synthetic example in which simulated surface deformations using a reference field are assumed as InSAR data for inverse modeling. The results indicate: (1) the EnKF can be used successfully to calibrate a land subsidence model with InSAR data; the estimation of inelastic specific storage is improved, and uncertainty of prediction is reduced, when all the data are accounted for; and (2) if the same ensemble is used to estimate Kalman gain, the analysis errors could cause filter divergence; thus, it is essential to include localization in the EnKF for InSAR data assimilation. © 2017, National Ground Water Association.
A multivariate model of stakeholder preference for lethal cat management.
Wald, Dara M; Jacobson, Susan K
2014-01-01
Identifying stakeholder beliefs and attitudes is critical for resolving management conflicts. Debate over outdoor cat management is often described as a conflict between two groups, environmental advocates and animal welfare advocates, but little is known about the variables predicting differences among these critical stakeholder groups. We administered a mail survey to randomly selected stakeholders representing both of these groups (n=1,596) in Florida, where contention over the management of outdoor cats has been widespread. We used a structural equation model to evaluate stakeholder intention to support non-lethal management. The cognitive hierarchy model predicted that values influenced beliefs, which predicted general and specific attitudes, which in turn, influenced behavioral intentions. We posited that specific attitudes would mediate the effect of general attitudes, beliefs, and values on management support. Model fit statistics suggested that the final model fit the data well (CFI=0.94, RMSEA=0.062). The final model explained 74% of the variance in management support, and positive attitudes toward lethal management (humaneness) had the largest direct effect on management support. Specific attitudes toward lethal management and general attitudes toward outdoor cats mediated the relationship between positive (p<0.05) and negative cat-related impact beliefs (p<0.05) and support for management. These results supported the specificity hypothesis and the use of the cognitive hierarchy to assess stakeholder intention to support non-lethal cat management. Our findings suggest that stakeholders can simultaneously perceive both positive and negative beliefs about outdoor cats, which influence attitudes toward and support for non-lethal management.
Barron, Daniel S; Fox, Peter T; Pardoe, Heath; Lancaster, Jack; Price, Larry R; Blackmon, Karen; Berry, Kristen; Cavazos, Jose E; Kuzniecky, Ruben; Devinsky, Orrin; Thesen, Thomas
2015-01-01
Noninvasive markers of brain function could yield biomarkers in many neurological disorders. Disease models constrained by coordinate-based meta-analysis are likely to increase this yield. Here, we evaluate a thalamic model of temporal lobe epilepsy that we proposed in a coordinate-based meta-analysis and extended in a diffusion tractography study of an independent patient population. Specifically, we evaluated whether thalamic functional connectivity (resting-state fMRI-BOLD) with temporal lobe areas can predict seizure onset laterality, as established with intracranial EEG. Twenty-four lesional and non-lesional temporal lobe epilepsy patients were studied. No significant differences in functional connection strength in patient and control groups were observed with Mann-Whitney Tests (corrected for multiple comparisons). Notwithstanding the lack of group differences, individual patient difference scores (from control mean connection strength) successfully predicted seizure onset zone as shown in ROC curves: discriminant analysis (two-dimensional) predicted seizure onset zone with 85% sensitivity and 91% specificity; logistic regression (four-dimensional) achieved 86% sensitivity and 100% specificity. The strongest markers in both analyses were left thalamo-hippocampal and right thalamo-entorhinal cortex functional connection strength. Thus, this study shows that thalamic functional connections are sensitive and specific markers of seizure onset laterality in individual temporal lobe epilepsy patients. This study also advances an overall strategy for the programmatic development of neuroimaging biomarkers in clinical and genetic populations: a disease model informed by coordinate-based meta-analysis was used to anatomically constrain individual patient analyses.
Ab initio state-specific N2 + O dissociation and exchange modeling for molecular simulations
NASA Astrophysics Data System (ADS)
Luo, Han; Kulakhmetov, Marat; Alexeenko, Alina
2017-02-01
Quasi-classical trajectory (QCT) calculations are used in this work to calculate state-specific N2(X1Σ ) +O(3P ) →2 N(4S ) +O(3P ) dissociation and N2(X1Σ ) +O(3P ) →NO(X2Π ) +N(4S ) exchange cross sections and rates based on the 13A″ and 13A' ab initio potential energy surface by Gamallo et al. [J. Chem. Phys. 119, 2545-2556 (2003)]. The calculations consider translational energies up to 23 eV and temperatures between 1000 K and 20 000 K. Vibrational favoring is observed for dissociation reaction at the whole range of collision energies and for exchange reaction around the dissociation limit. For the same collision energy, cross sections for v = 30 are 4 to 6 times larger than those for the ground state. The exchange reaction has an effective activation energy that is dependent on the initial rovibrational level, which is different from dissociation reaction. In addition, the exchange cross sections have a maximum when the total collision energy (TCE) approaches dissociation energy. The calculations are used to generate compact QCT-derived state-specific dissociation (QCT-SSD) and QCT-derived state-specific exchange (QCT-SSE) models, which describe over 1 × 106 cross sections with about 150 model parameters. The models can be used directly within direct simulation Monte Carlo and computational fluid dynamics simulations. Rate constants predicted by the new models are compared to the experimental measurements, direct QCT calculations and predictions by other models that include: TCE model, Bose-Candler QCT-based exchange model, Macheret-Fridman dissociation model, Macheret's exchange model, and Park's two-temperature model. The new models match QCT-calculated and experimental rates within 30% under nonequilibrium conditions while other models under predict by over an order of magnitude under vibrationally-cold conditions.
Cell design concepts for aqueous lithium-oxygen batteries: A model-based assessment
NASA Astrophysics Data System (ADS)
Grübl, Daniel; Bessler, Wolfgang G.
2015-11-01
Seven cell design concepts for aqueous (alkaline) lithium-oxygen batteries are investigated using a multi-physics continuum model for predicting cell behavior and performance in terms of the specific energy and specific power. Two different silver-based cathode designs (a gas diffusion electrode and a flooded cathode) and three different separator designs (a porous separator, a stirred separator chamber, and a redox-flow separator) are compared. Cathode and separator thicknesses are varied over a wide range (50 μm-20 mm) in order to identify optimum configurations. All designs show a considerable capacity-rate effect due to spatiotemporally inhomogeneous precipitation of solid discharge product LiOH·H2O. In addition, a cell design with flooded cathode and redox-flow separator including oxygen uptake within the external tank is suggested. For this design, the model predicts specific power up to 33 W/kg and specific energy up to 570 Wh/kg (gravimetric values of discharged cell including all cell components and catholyte except housing and piping).
Willoughby, Jessica Fitts; Myrick, Jessica Gall
2016-06-01
Research indicates that when people seek health information, they typically look for information about a specific symptom, preventive measure, disease, or treatment. It is unclear, however, whether general or disease-specific theoretical models best predict how people search for health information. We surveyed undergraduates (N = 963) at a large public southeastern university to examine health information seeking in two incongruent health contexts (sexual health and cancer) to test whether a general model would hold for specific topics that differed in their immediate personal relevance for the target population. We found that the planned risk information seeking model was statistically a good fit for the data. Yet multiple predicted paths were not supported in either data set. Certain variables, such as attitudes, norms, and affect, appear to be strong predictors of intentions to seek information across health contexts. Implications for theory building, research methodology, and applied work in health-related risk information seeking are discussed.
Li, Muyang; Williams, Daniel L.; Heckwolf, Marlies; ...
2016-10-04
In this paper, we explore the ability of several characterization approaches for phenotyping to extract information about plant cell wall properties in diverse maize genotypes with the goal of identifying approaches that could be used to predict the plant's response to deconstruction in a biomass-to-biofuel process. Specifically, a maize diversity panel was subjected to two high-throughput biomass characterization approaches, pyrolysis molecular beam mass spectrometry (py-MBMS) and near-infrared (NIR) spectroscopy, and chemometric models to predict a number of plant cell wall properties as well as enzymatic hydrolysis yields of glucose following either no pretreatment or with mild alkaline pretreatment. These weremore » compared to multiple linear regression (MLR) models developed from quantified properties. We were able to demonstrate that direct correlations to specific mass spectrometry ions from pyrolysis as well as characteristic regions of the second derivative of the NIR spectrum regions were comparable in their predictive capability to partial least squares (PLS) models for p-coumarate content, while the direct correlation to the spectral data was superior to the PLS for Klason lignin content and guaiacyl monomer release by thioacidolysis as assessed by cross-validation. The PLS models for prediction of hydrolysis yields using either py-MBMS or NIR spectra were superior to MLR models based on quantified properties for unpretreated biomass. However, the PLS models using the two high-throughput characterization approaches could not predict hydrolysis following alkaline pretreatment while MLR models based on quantified properties could. This is likely a consequence of quantified properties including some assessments of pretreated biomass, while the py-MBMS and NIR only utilized untreated biomass.« less
DOE Office of Scientific and Technical Information (OSTI.GOV)
Li, Muyang; Williams, Daniel L.; Heckwolf, Marlies
In this paper, we explore the ability of several characterization approaches for phenotyping to extract information about plant cell wall properties in diverse maize genotypes with the goal of identifying approaches that could be used to predict the plant's response to deconstruction in a biomass-to-biofuel process. Specifically, a maize diversity panel was subjected to two high-throughput biomass characterization approaches, pyrolysis molecular beam mass spectrometry (py-MBMS) and near-infrared (NIR) spectroscopy, and chemometric models to predict a number of plant cell wall properties as well as enzymatic hydrolysis yields of glucose following either no pretreatment or with mild alkaline pretreatment. These weremore » compared to multiple linear regression (MLR) models developed from quantified properties. We were able to demonstrate that direct correlations to specific mass spectrometry ions from pyrolysis as well as characteristic regions of the second derivative of the NIR spectrum regions were comparable in their predictive capability to partial least squares (PLS) models for p-coumarate content, while the direct correlation to the spectral data was superior to the PLS for Klason lignin content and guaiacyl monomer release by thioacidolysis as assessed by cross-validation. The PLS models for prediction of hydrolysis yields using either py-MBMS or NIR spectra were superior to MLR models based on quantified properties for unpretreated biomass. However, the PLS models using the two high-throughput characterization approaches could not predict hydrolysis following alkaline pretreatment while MLR models based on quantified properties could. This is likely a consequence of quantified properties including some assessments of pretreated biomass, while the py-MBMS and NIR only utilized untreated biomass.« less
GPU-based RFA simulation for minimally invasive cancer treatment of liver tumours.
Mariappan, Panchatcharam; Weir, Phil; Flanagan, Ronan; Voglreiter, Philip; Alhonnoro, Tuomas; Pollari, Mika; Moche, Michael; Busse, Harald; Futterer, Jurgen; Portugaller, Horst Rupert; Sequeiros, Roberto Blanco; Kolesnik, Marina
2017-01-01
Radiofrequency ablation (RFA) is one of the most popular and well-standardized minimally invasive cancer treatments (MICT) for liver tumours, employed where surgical resection has been contraindicated. Less-experienced interventional radiologists (IRs) require an appropriate planning tool for the treatment to help avoid incomplete treatment and so reduce the tumour recurrence risk. Although a few tools are available to predict the ablation lesion geometry, the process is computationally expensive. Also, in our implementation, a few patient-specific parameters are used to improve the accuracy of the lesion prediction. Advanced heterogeneous computing using personal computers, incorporating the graphics processing unit (GPU) and the central processing unit (CPU), is proposed to predict the ablation lesion geometry. The most recent GPU technology is used to accelerate the finite element approximation of Penne's bioheat equation and a three state cell model. Patient-specific input parameters are used in the bioheat model to improve accuracy of the predicted lesion. A fast GPU-based RFA solver is developed to predict the lesion by doing most of the computational tasks in the GPU, while reserving the CPU for concurrent tasks such as lesion extraction based on the heat deposition at each finite element node. The solver takes less than 3 min for a treatment duration of 26 min. When the model receives patient-specific input parameters, the deviation between real and predicted lesion is below 3 mm. A multi-centre retrospective study indicates that the fast RFA solver is capable of providing the IR with the predicted lesion in the short time period before the intervention begins when the patient has been clinically prepared for the treatment.
Nazemi, S Majid; Kalajahi, S Mehrdad Hosseini; Cooper, David M L; Kontulainen, Saija A; Holdsworth, David W; Masri, Bassam A; Wilson, David R; Johnston, James D
2017-07-05
Previously, a finite element (FE) model of the proximal tibia was developed and validated against experimentally measured local subchondral stiffness. This model indicated modest predictions of stiffness (R 2 =0.77, normalized root mean squared error (RMSE%)=16.6%). Trabecular bone though was modeled with isotropic material properties despite its orthotropic anisotropy. The objective of this study was to identify the anisotropic FE modeling approach which best predicted (with largest explained variance and least amount of error) local subchondral bone stiffness at the proximal tibia. Local stiffness was measured at the subchondral surface of 13 medial/lateral tibial compartments using in situ macro indentation testing. An FE model of each specimen was generated assuming uniform anisotropy with 14 different combinations of cortical- and tibial-specific density-modulus relationships taken from the literature. Two FE models of each specimen were also generated which accounted for the spatial variation of trabecular bone anisotropy directly from clinical CT images using grey-level structure tensor and Cowin's fabric-elasticity equations. Stiffness was calculated using FE and compared to measured stiffness in terms of R 2 and RMSE%. The uniform anisotropic FE model explained 53-74% of the measured stiffness variance, with RMSE% ranging from 12.4 to 245.3%. The models which accounted for spatial variation of trabecular bone anisotropy predicted 76-79% of the variance in stiffness with RMSE% being 11.2-11.5%. Of the 16 evaluated finite element models in this study, the combination of Synder and Schneider (for cortical bone) and Cowin's fabric-elasticity equations (for trabecular bone) best predicted local subchondral bone stiffness. Copyright © 2017 Elsevier Ltd. All rights reserved.
Subliminal action priming modulates the perceived intensity of sensory action consequences.
Stenner, Max-Philipp; Bauer, Markus; Sidarus, Nura; Heinze, Hans-Jochen; Haggard, Patrick; Dolan, Raymond J
2014-02-01
The sense of control over the consequences of one's actions depends on predictions about these consequences. According to an influential computational model, consistency between predicted and observed action consequences attenuates perceived stimulus intensity, which might provide a marker of agentic control. An important assumption of this model is that these predictions are generated within the motor system. However, previous studies of sensory attenuation have typically confounded motor-specific perceptual modulation with perceptual effects of stimulus predictability that are not specific to motor action. As a result, these studies cannot unambiguously attribute sensory attenuation to a motor locus. We present a psychophysical experiment on auditory attenuation that avoids this pitfall. Subliminal masked priming of motor actions with compatible prime-target pairs has previously been shown to modulate both reaction times and the explicit feeling of control over action consequences. Here, we demonstrate reduced perceived loudness of tones caused by compatibly primed actions. Importantly, this modulation results from a manipulation of motor processing and is not confounded by stimulus predictability. We discuss our results with respect to theoretical models of the mechanisms underlying sensory attenuation and subliminal motor priming. Copyright © 2013 The Authors. Published by Elsevier B.V. All rights reserved.
Subliminal action priming modulates the perceived intensity of sensory action consequences☆
Stenner, Max-Philipp; Bauer, Markus; Sidarus, Nura; Heinze, Hans-Jochen; Haggard, Patrick; Dolan, Raymond J.
2014-01-01
The sense of control over the consequences of one’s actions depends on predictions about these consequences. According to an influential computational model, consistency between predicted and observed action consequences attenuates perceived stimulus intensity, which might provide a marker of agentic control. An important assumption of this model is that these predictions are generated within the motor system. However, previous studies of sensory attenuation have typically confounded motor-specific perceptual modulation with perceptual effects of stimulus predictability that are not specific to motor action. As a result, these studies cannot unambiguously attribute sensory attenuation to a motor locus. We present a psychophysical experiment on auditory attenuation that avoids this pitfall. Subliminal masked priming of motor actions with compatible prime–target pairs has previously been shown to modulate both reaction times and the explicit feeling of control over action consequences. Here, we demonstrate reduced perceived loudness of tones caused by compatibly primed actions. Importantly, this modulation results from a manipulation of motor processing and is not confounded by stimulus predictability. We discuss our results with respect to theoretical models of the mechanisms underlying sensory attenuation and subliminal motor priming. PMID:24333539
The fraction of quiescent massive galaxies in the early Universe
NASA Astrophysics Data System (ADS)
Fontana, A.; Santini, P.; Grazian, A.; Pentericci, L.; Fiore, F.; Castellano, M.; Giallongo, E.; Menci, N.; Salimbeni, S.; Cristiani, S.; Nonino, M.; Vanzella, E.
2009-07-01
Aims: We attempt to compile a complete, mass-selected sample of galaxies with low specific star-formation rates, and compare their properties with theoretical model predictions. Methods: We use the f(24 μ m})/f(K) flux ratio and the SED fitting to the 0.35-8.0 μm spectral distribution, to select quiescent galaxies from z≃ 0.4 to z≃ 4 in the GOODS-MUSIC sample. Our observational selection can be translated into thresholds in specific star-formation rate dot{M}/M_*, which can be compared with theoretical predictions. Results: In the framework of the well-known global decline in quiescent galaxy fraction with redshift, we find that a non-negligible fraction {≃ 15-20% of massive galaxies with low specific star-formation rate exists up to z≃ 4, including a tail of “red and dead” galaxies with dot{M}/M_*<10-11 yr-1. Theoretical models vary to a large extent in their predictions for the fraction of galaxies with low specific star-formation rates, but are unable to provide a global match to our data.
Can plantar soft tissue mechanics enhance prognosis of diabetic foot ulcer?
Naemi, R; Chatzistergos, P; Suresh, S; Sundar, L; Chockalingam, N; Ramachandran, A
2017-04-01
To investigate if the assessment of the mechanical properties of plantar soft tissue can increase the accuracy of predicting Diabetic Foot Ulceration (DFU). 40 patients with diabetic neuropathy and no DFU were recruited. Commonly assessed clinical parameters along with plantar soft tissue stiffness and thickness were measured at baseline using ultrasound elastography technique. 7 patients developed foot ulceration during a 12months follow-up. Logistic regression was used to identify parameters that contribute to predicting the DFU incidence. The effect of using parameters related to the mechanical behaviour of plantar soft tissue on the specificity, sensitivity, prediction strength and accuracy of the predicting models for DFU was assessed. Patients with higher plantar soft tissue thickness and lower stiffness at the 1st Metatarsal head area showed an increased risk of DFU. Adding plantar soft tissue stiffness and thickness to the model improved its specificity (by 3%), sensitivity (by 14%), prediction accuracy (by 5%) and prognosis strength (by 1%). The model containing all predictors was able to effectively (χ 2 (8, N=40)=17.55, P<0.05) distinguish between the patients with and without DFU incidence. The mechanical properties of plantar soft tissue can be used to improve the predictability of DFU in moderate/high risk patients. Copyright © 2017 Elsevier B.V. All rights reserved.
Harris, Peter R; Sillence, Elizabeth; Briggs, Pam
2011-07-27
How do people decide which sites to use when seeking health advice online? We can assume, from related work in e-commerce, that general design factors known to affect trust in the site are important, but in this paper we also address the impact of factors specific to the health domain. The current study aimed to (1) assess the factorial structure of a general measure of Web trust, (2) model how the resultant factors predicted trust in, and readiness to act on, the advice found on health-related websites, and (3) test whether adding variables from social cognition models to capture elements of the response to threatening, online health-risk information enhanced the prediction of these outcomes. Participants were asked to recall a site they had used to search for health-related information and to think of that site when answering an online questionnaire. The questionnaire consisted of a general Web trust questionnaire plus items assessing appraisals of the site, including threat appraisals, information checking, and corroboration. It was promoted on the hungersite.com website. The URL was distributed via Yahoo and local print media. We assessed the factorial structure of the measures using principal components analysis and modeled how well they predicted the outcome measures using structural equation modeling (SEM) with EQS software. We report an analysis of the responses of participants who searched for health advice for themselves (N = 561). Analysis of the general Web trust questionnaire revealed 4 factors: information quality, personalization, impartiality, and credible design. In the final SEM model, information quality and impartiality were direct predictors of trust. However, variables specific to eHealth (perceived threat, coping, and corroboration) added substantially to the ability of the model to predict variance in trust and readiness to act on advice on the site. The final model achieved a satisfactory fit: χ(2) (5) = 10.8 (P = .21), comparative fit index = .99, root mean square error of approximation = .052. The model accounted for 66% of the variance in trust and 49% of the variance in readiness to act on the advice. Adding variables specific to eHealth enhanced the ability of a model of trust to predict trust and readiness to act on advice.
Harris, Peter R; Briggs, Pam
2011-01-01
Background How do people decide which sites to use when seeking health advice online? We can assume, from related work in e-commerce, that general design factors known to affect trust in the site are important, but in this paper we also address the impact of factors specific to the health domain. Objective The current study aimed to (1) assess the factorial structure of a general measure of Web trust, (2) model how the resultant factors predicted trust in, and readiness to act on, the advice found on health-related websites, and (3) test whether adding variables from social cognition models to capture elements of the response to threatening, online health-risk information enhanced the prediction of these outcomes. Methods Participants were asked to recall a site they had used to search for health-related information and to think of that site when answering an online questionnaire. The questionnaire consisted of a general Web trust questionnaire plus items assessing appraisals of the site, including threat appraisals, information checking, and corroboration. It was promoted on the hungersite.com website. The URL was distributed via Yahoo and local print media. We assessed the factorial structure of the measures using principal components analysis and modeled how well they predicted the outcome measures using structural equation modeling (SEM) with EQS software. Results We report an analysis of the responses of participants who searched for health advice for themselves (N = 561). Analysis of the general Web trust questionnaire revealed 4 factors: information quality, personalization, impartiality, and credible design. In the final SEM model, information quality and impartiality were direct predictors of trust. However, variables specific to eHealth (perceived threat, coping, and corroboration) added substantially to the ability of the model to predict variance in trust and readiness to act on advice on the site. The final model achieved a satisfactory fit: χ2 5 = 10.8 (P = .21), comparative fit index = .99, root mean square error of approximation = .052. The model accounted for 66% of the variance in trust and 49% of the variance in readiness to act on the advice. Conclusions Adding variables specific to eHealth enhanced the ability of a model of trust to predict trust and readiness to act on advice. PMID:21795237
Dinov, Ivo D; Heavner, Ben; Tang, Ming; Glusman, Gustavo; Chard, Kyle; Darcy, Mike; Madduri, Ravi; Pa, Judy; Spino, Cathie; Kesselman, Carl; Foster, Ian; Deutsch, Eric W; Price, Nathan D; Van Horn, John D; Ames, Joseph; Clark, Kristi; Hood, Leroy; Hampstead, Benjamin M; Dauer, William; Toga, Arthur W
2016-01-01
A unique archive of Big Data on Parkinson's Disease is collected, managed and disseminated by the Parkinson's Progression Markers Initiative (PPMI). The integration of such complex and heterogeneous Big Data from multiple sources offers unparalleled opportunities to study the early stages of prevalent neurodegenerative processes, track their progression and quickly identify the efficacies of alternative treatments. Many previous human and animal studies have examined the relationship of Parkinson's disease (PD) risk to trauma, genetics, environment, co-morbidities, or life style. The defining characteristics of Big Data-large size, incongruency, incompleteness, complexity, multiplicity of scales, and heterogeneity of information-generating sources-all pose challenges to the classical techniques for data management, processing, visualization and interpretation. We propose, implement, test and validate complementary model-based and model-free approaches for PD classification and prediction. To explore PD risk using Big Data methodology, we jointly processed complex PPMI imaging, genetics, clinical and demographic data. Collective representation of the multi-source data facilitates the aggregation and harmonization of complex data elements. This enables joint modeling of the complete data, leading to the development of Big Data analytics, predictive synthesis, and statistical validation. Using heterogeneous PPMI data, we developed a comprehensive protocol for end-to-end data characterization, manipulation, processing, cleaning, analysis and validation. Specifically, we (i) introduce methods for rebalancing imbalanced cohorts, (ii) utilize a wide spectrum of classification methods to generate consistent and powerful phenotypic predictions, and (iii) generate reproducible machine-learning based classification that enables the reporting of model parameters and diagnostic forecasting based on new data. We evaluated several complementary model-based predictive approaches, which failed to generate accurate and reliable diagnostic predictions. However, the results of several machine-learning based classification methods indicated significant power to predict Parkinson's disease in the PPMI subjects (consistent accuracy, sensitivity, and specificity exceeding 96%, confirmed using statistical n-fold cross-validation). Clinical (e.g., Unified Parkinson's Disease Rating Scale (UPDRS) scores), demographic (e.g., age), genetics (e.g., rs34637584, chr12), and derived neuroimaging biomarker (e.g., cerebellum shape index) data all contributed to the predictive analytics and diagnostic forecasting. Model-free Big Data machine learning-based classification methods (e.g., adaptive boosting, support vector machines) can outperform model-based techniques in terms of predictive precision and reliability (e.g., forecasting patient diagnosis). We observed that statistical rebalancing of cohort sizes yields better discrimination of group differences, specifically for predictive analytics based on heterogeneous and incomplete PPMI data. UPDRS scores play a critical role in predicting diagnosis, which is expected based on the clinical definition of Parkinson's disease. Even without longitudinal UPDRS data, however, the accuracy of model-free machine learning based classification is over 80%. The methods, software and protocols developed here are openly shared and can be employed to study other neurodegenerative disorders (e.g., Alzheimer's, Huntington's, amyotrophic lateral sclerosis), as well as for other predictive Big Data analytics applications.
Two-Pendulum Model of Propellant Slosh in Europa Clipper PMD Tank
NASA Technical Reports Server (NTRS)
Ng, Wanyi; Benson, David
2017-01-01
Model propellant slosh for Europa Clipper using two pendulums such that controls engineers can predict slosh behavior during the mission. Importance of predicting propellant slosh; (1) Sloshing changes CM (center of mass) of spacecraft and exerts forces and torques on spacecraft. (2) Avoid natural frequencies of structures. (3) Size ACS (Attitude Control Systems) thrusters to counteract forces and torques. Can model sloshing fluid as two pendulums with specific parameters (mass, length, damping),
Dazard, Jean-Eudes; Choe, Michael; LeBlanc, Michael; Rao, J. Sunil
2015-01-01
PRIMsrc is a novel implementation of a non-parametric bump hunting procedure, based on the Patient Rule Induction Method (PRIM), offering a unified treatment of outcome variables, including censored time-to-event (Survival), continuous (Regression) and discrete (Classification) responses. To fit the model, it uses a recursive peeling procedure with specific peeling criteria and stopping rules depending on the response. To validate the model, it provides an objective function based on prediction-error or other specific statistic, as well as two alternative cross-validation techniques, adapted to the task of decision-rule making and estimation in the three types of settings. PRIMsrc comes as an open source R package, including at this point: (i) a main function for fitting a Survival Bump Hunting model with various options allowing cross-validated model selection to control model size (#covariates) and model complexity (#peeling steps) and generation of cross-validated end-point estimates; (ii) parallel computing; (iii) various S3-generic and specific plotting functions for data visualization, diagnostic, prediction, summary and display of results. It is available on CRAN and GitHub. PMID:26798326
Bazeley, Peter S; Prithivi, Sridevi; Struble, Craig A; Povinelli, Richard J; Sem, Daniel S
2006-01-01
Cytochrome P450 2D6 (CYP2D6) is used to develop an approach for predicting affinity and relevant binding conformation(s) for highly flexible binding sites. The approach combines the use of docking scores and compound properties as attributes in building a neural network (NN) model. It begins by identifying segments of CYP2D6 that are important for binding specificity, based on structural variability among diverse CYP enzymes. A family of distinct, low-energy conformations of CYP2D6 are generated using simulated annealing (SA) and a collection of 82 compounds with known CYP2D6 affinities are docked. Interestingly, docking poses are observed on the backside of the heme as well as in the known active site. Docking scores for the active site binders, along with compound-specific attributes, are used to train a neural network model to properly bin compounds as strong binders, moderate binders, or nonbinders. Attribute selection is used to preselect the most important scores and compound-specific attributes for the model. A prediction accuracy of 85+/-6% is achieved. Dominant attributes include docking scores for three of the 20 conformations in the ensemble as well as the compound's formal charge, number of aromatic rings, and AlogP. Although compound properties were highly predictive attributes (12% improvement over baseline) in the NN-based prediction of CYP2D6 binders, their combined use with docking score attributes is synergistic (net increase of 23% above baseline). Beyond prediction of affinity, attribute selection provides a way to identify the most relevant protein conformation(s), in terms of binding competence. In the case of CYP2D6, three out of the ensemble of 20 SA-generated structures are found to be the most predictive for binding.
ten Haaf, Kevin; Tammemägi, Martin C.; Han, Summer S.; Kong, Chung Yin; Plevritis, Sylvia K.; de Koning, Harry J.; Steyerberg, Ewout W.
2017-01-01
Background Selection of candidates for lung cancer screening based on individual risk has been proposed as an alternative to criteria based on age and cumulative smoking exposure (pack-years). Nine previously established risk models were assessed for their ability to identify those most likely to develop or die from lung cancer. All models considered age and various aspects of smoking exposure (smoking status, smoking duration, cigarettes per day, pack-years smoked, time since smoking cessation) as risk predictors. In addition, some models considered factors such as gender, race, ethnicity, education, body mass index, chronic obstructive pulmonary disease, emphysema, personal history of cancer, personal history of pneumonia, and family history of lung cancer. Methods and findings Retrospective analyses were performed on 53,452 National Lung Screening Trial (NLST) participants (1,925 lung cancer cases and 884 lung cancer deaths) and 80,672 Prostate, Lung, Colorectal and Ovarian Cancer Screening Trial (PLCO) ever-smoking participants (1,463 lung cancer cases and 915 lung cancer deaths). Six-year lung cancer incidence and mortality risk predictions were assessed for (1) calibration (graphically) by comparing the agreement between the predicted and the observed risks, (2) discrimination (area under the receiver operating characteristic curve [AUC]) between individuals with and without lung cancer (death), and (3) clinical usefulness (net benefit in decision curve analysis) by identifying risk thresholds at which applying risk-based eligibility would improve lung cancer screening efficacy. To further assess performance, risk model sensitivities and specificities in the PLCO were compared to those based on the NLST eligibility criteria. Calibration was satisfactory, but discrimination ranged widely (AUCs from 0.61 to 0.81). The models outperformed the NLST eligibility criteria over a substantial range of risk thresholds in decision curve analysis, with a higher sensitivity for all models and a slightly higher specificity for some models. The PLCOm2012, Bach, and Two-Stage Clonal Expansion incidence models had the best overall performance, with AUCs >0.68 in the NLST and >0.77 in the PLCO. These three models had the highest sensitivity and specificity for predicting 6-y lung cancer incidence in the PLCO chest radiography arm, with sensitivities >79.8% and specificities >62.3%. In contrast, the NLST eligibility criteria yielded a sensitivity of 71.4% and a specificity of 62.2%. Limitations of this study include the lack of identification of optimal risk thresholds, as this requires additional information on the long-term benefits (e.g., life-years gained and mortality reduction) and harms (e.g., overdiagnosis) of risk-based screening strategies using these models. In addition, information on some predictor variables included in the risk prediction models was not available. Conclusions Selection of individuals for lung cancer screening using individual risk is superior to selection criteria based on age and pack-years alone. The benefits, harms, and feasibility of implementing lung cancer screening policies based on risk prediction models should be assessed and compared with those of current recommendations. PMID:28376113
Route Prediction on Tracking Data to Location-Based Services
NASA Astrophysics Data System (ADS)
Petróczi, Attila István; Gáspár-Papanek, Csaba
Wireless networks have become so widespread, it is beneficial to determine the ability of cellular networks for localization. This property enables the development of location-based services, providing useful information. These services can be improved by route prediction under the condition of using simple algorithms, because of the limited capabilities of mobile stations. This study gives alternative solutions for this problem of route prediction based on a specific graph model. Our models provide the opportunity to reach our destinations with less effort.
Thomason, Elizabeth; Volling, Brenda L.; Flynn, Heather A.; McDonough, Susan C.; Marcus, Sheila M.; Lopez, Juan F.; Vazquez, Delia M.
2015-01-01
Despite the consistent link between parenting stress and postpartum depressive symptoms, few studies have explored the relationships longitudinally. The purpose of this study was to test bidirectional and unidirectional models of depressive symptoms and parenting stress. Uniquely, three specific domains of parenting stress were examined: parental distress, difficult child stress, and parent–child dysfunctional interaction (PCDI). One hundred and five women completed the Beck Depression Inventory and the Parenting Stress Index–Short Form at 3, 7, and 14 months after giving birth. Structural equation modeling revealed that total parenting stress predicted later depressive symptoms, however, there were different patterns between postpartum depressive symptoms and different types of parenting stress. A unidirectional model of parental distress predicting depressive symptoms best fit the data, with significant stability paths but non-significant cross-lagged paths. A unidirectional model of depressive symptoms predicted significant later difficult child stress. No model fit well with PCDI. Future research should continue to explore the specific nature of the associations of postpartum depression and different types of parenting stress on infant development and the infant–mother relationship. PMID:24956500
Respiratory morbidity of pattern and model makers exposed to wood, plastic, and metal products
DOE Office of Scientific and Technical Information (OSTI.GOV)
Robins, T.G.; Haboubi, G.; Demers, R.Y.
Pattern and model makers are skilled tradespersons who may be exposed to hardwoods, softwoods, phenol-formaldehyde resin-impregnated woods, epoxy and polyester/styrene resin systems, and welding and metal-casting fumes. The relationship of respiratory symptoms (wheezing, chronic bronchitis, dyspnea) and pulmonary function (FVC% predicted, FEV1% predicted, FEV1/FVC% predicted) with interview-derived cumulative exposure estimates to specific workplace agents and to all work with wood, plastic, or metal products was investigated in 751 pattern and model makers in southeast Michigan. In stratified analyses and age- and smoking-adjusted linear and logistic regression models, measures of cumulative wood exposures were associated with decrements in pulmonary function andmore » dyspnea, but not with other symptoms. In similar analyses, measures of cumulative plastic exposures were associated with wheezing, chronic bronchitis, and dyspnea, but not with decrements in pulmonary function. Prior studies of exposure levels among pattern and model makers and of respiratory health effects of specific agents among other occupational groups support the plausibility of wood-related effects more strongly than that of plastic-related effects.« less
Computer Models of Personality: Implications for Measurement
ERIC Educational Resources Information Center
Cranton, P. A.
1976-01-01
Current research on computer models of personality is reviewed and categorized under five headings: (1) models of belief systems; (2) models of interpersonal behavior; (3) models of decision-making processes; (4) prediction models; and (5) theory-based simulations of specific processes. The use of computer models in personality measurement is…
Blinded Validation of Breath Biomarkers of Lung Cancer, a Potential Ancillary to Chest CT Screening
Phillips, Michael; Bauer, Thomas L.; Cataneo, Renee N.; Lebauer, Cassie; Mundada, Mayur; Pass, Harvey I.; Ramakrishna, Naren; Rom, William N.; Vallières, Eric
2015-01-01
Background Breath volatile organic compounds (VOCs) have been reported as biomarkers of lung cancer, but it is not known if biomarkers identified in one group can identify disease in a separate independent cohort. Also, it is not known if combining breath biomarkers with chest CT has the potential to improve the sensitivity and specificity of lung cancer screening. Methods Model-building phase (unblinded): Breath VOCs were analyzed with gas chromatography mass spectrometry in 82 asymptomatic smokers having screening chest CT, 84 symptomatic high-risk subjects with a tissue diagnosis, 100 without a tissue diagnosis, and 35 healthy subjects. Multiple Monte Carlo simulations identified breath VOC mass ions with greater than random diagnostic accuracy for lung cancer, and these were combined in a multivariate predictive algorithm. Model-testing phase (blinded validation): We analyzed breath VOCs in an independent cohort of similar subjects (n = 70, 51, 75 and 19 respectively). The algorithm predicted discriminant function (DF) values in blinded replicate breath VOC samples analyzed independently at two laboratories (A and B). Outcome modeling: We modeled the expected effects of combining breath biomarkers with chest CT on the sensitivity and specificity of lung cancer screening. Results Unblinded model-building phase. The algorithm identified lung cancer with sensitivity 74.0%, specificity 70.7% and C-statistic 0.78. Blinded model-testing phase: The algorithm identified lung cancer at Laboratory A with sensitivity 68.0%, specificity 68.4%, C-statistic 0.71; and at Laboratory B with sensitivity 70.1%, specificity 68.0%, C-statistic 0.70, with linear correlation between replicates (r = 0.88). In a projected outcome model, breath biomarkers increased the sensitivity, specificity, and positive and negative predictive values of chest CT for lung cancer when the tests were combined in series or parallel. Conclusions Breath VOC mass ion biomarkers identified lung cancer in a separate independent cohort, in a blinded replicated study. Combining breath biomarkers with chest CT could potentially improve the sensitivity and specificity of lung cancer screening. Trial Registration ClinicalTrials.gov NCT00639067 PMID:26698306
Predicting tidal currents in San Francisco Bay using a spectral model
Burau, Jon R.; Cheng, Ralph T.
1988-01-01
This paper describes the formulation of a spectral (or frequency based) model which solves the linearized shallow water equations. To account for highly variable basin bathymetry, spectral solutions are obtained using the finite element method which allows the strategic placement of the computation points in the specific areas of interest or in areas where the gradients of the dependent variables are expected to be large. Model results are compared with data using simple statistics to judge overall model performance in the San Francisco Bay estuary. Once the model is calibrated and verified, prediction of the tides and tidal currents in San Francisco Bay is accomplished by applying astronomical tides (harmonic constants deduced from field data) at the prediction time along the model boundaries.
DOT National Transportation Integrated Search
2013-07-01
The study provides estimation of site specific variation in environmental factors that can be : used in predicting seasonal and long-term variations in moduli of unbound materials. Using : these site specific estimates, the EICM climatic input files ...
Material Stream Strategy for Lithium and Inorganics (U)
DOE Office of Scientific and Technical Information (OSTI.GOV)
Safarik, Douglas Joseph; Dunn, Paul Stanton; Korzekwa, Deniece Rochelle
Design Agency Responsibilities: Manufacturing Support to meet Stockpile Stewardship goals for maintaining the nuclear stockpile through experimental and predictive modeling capability. Development and maintenance of Manufacturing Science expertise to assess material specifications and performance boundaries, and their relationship to processing parameters. Production Engineering Evaluations with competence in design requirements, material specifications, and manufacturing controls. Maintenance and enhancement of Aging Science expertise to support Stockpile Stewardship predictive science capability.
Douglas, Steven; Dixon, Barnali; Griffin, Dale W.
2018-01-01
With continued population growth and increasing use of fresh groundwater resources, protection of this valuable resource is critical. A cost effective means to assess risk of groundwater contamination potential will provide a useful tool to protect these resources. Integrating geospatial methods offers a means to quantify the risk of contaminant potential in cost effective and spatially explicit ways. This research was designed to compare the ability of intrinsic (DRASTIC) and specific (Attenuation Factor; AF) vulnerability models to indicate groundwater vulnerability areas by comparing model results to the presence of pesticides from groundwater sample datasets. A logistic regression was used to assess the relationship between the environmental variables and the presence or absence of pesticides within regions of varying vulnerability. According to the DRASTIC model, more than 20% of the study area is very highly vulnerable. Approximately 30% is very highly vulnerable according to the AF model. When groundwater concentrations of individual pesticides were compared to model predictions, the results were mixed. Model predictability improved when concentrations of the group of similar pesticides were compared to model results. Compared to the DRASTIC model, the AF model more accurately predicts the distribution of the number of contaminated wells within each vulnerability class.
Kontodimopoulos, Nick; Bozios, Panagiotis; Yfantopoulos, John; Niakas, Dimitris
2013-04-01
The purpose of this methodological study was to to provide insight into the under-addressed issue of the longitudinal predictive ability of mapping models. Post-intervention predicted and reported utilities were compared, and the effect of disease severity on the observed differences was examined. A cohort of 120 rheumatoid arthritis (RA) patients (60.0% female, mean age 59.0) embarking on therapy with biological agents completed the Modified Health Assessment Questionnaire (MHAQ) and the EQ-5D at baseline, and at 3, 6 and 12 months post-intervention. OLS regression produced a mapping equation to estimate post-intervention EQ-5D utilities from baseline MHAQ data. Predicted and reported utilities were compared with t test, and the prediction error was modeled, using fixed effects, in terms of covariates such as age, gender, time, disease duration, treatment, RF, DAS28 score, predicted and reported EQ-5D. The OLS model (RMSE = 0.207, R(2) = 45.2%) consistently underestimated future utilities, with a mean prediction error of 6.5%. Mean absolute differences between reported and predicted EQ-5D utilities at 3, 6 and 12 months exceeded the typically reported MID of the EQ-5D (0.03). According to the fixed-effects model, time, lower predicted EQ-5D and higher DAS28 scores had a significant impact on prediction errors, which appeared increasingly negative for lower reported EQ-5D scores, i.e., predicted utilities tended to be lower than reported ones in more severe health states. This study builds upon existing research having demonstrated the potential usefulness of mapping disease-specific instruments onto utility measures. The specific issue of longitudinal validity is addressed, as mapping models derived from baseline patients need to be validated on post-therapy samples. The underestimation of post-treatment utilities in the present study, at least in more severe patients, warrants further research before it is prudent to conduct cost-utility analyses in the context of RA by means of the MHAQ alone.
Haase, Claudia M.; Holley, Sarah; Bloch, Lian; Verstaen, Alice; Levenson, Robert W.
2016-01-01
Objectively coded interpersonal emotional behaviors that emerged during a 15-minute marital conflict interaction predicted the development of physical symptoms in a 20-year longitudinal study of long-term marriages. Dyadic latent growth curve modeling showed that anger behavior predicted increases in cardiovascular symptoms and stonewalling behavior predicted increases in musculoskeletal symptoms. Both associations were found for husbands (although cross-lagged path models also showed some support for wives) and were controlled for sociodemographic characteristics (age, education) and behaviors (i.e., exercise, smoking, alcohol consumption, caffeine consumption) known to influence health. Both associations did not exist at the start of the study, but only emerged over the ensuing 20 years. There was some support for the specificity of these relationships (i.e., stonewalling behavior did not predict cardiovascular symptoms; anger behavior did not predict musculoskeletal symptoms; neither symptom was predicted by fear nor sadness behavior), with the anger-cardiovascular relationship emerging as most robust. Using cross-lagged path models to probe directionality of these associations, emotional behaviors predicted physical health symptoms over time (with some reverse associations found as well). These findings illuminate longstanding theoretical and applied issues concerning the association between interpersonal emotional behaviors and physical health and suggest opportunities for preventive interventions focused on specific emotions to help address major public health problems. PMID:27213730
Genome-Wide Prediction of the Performance of Three-Way Hybrids in Barley.
Li, Zuo; Philipp, Norman; Spiller, Monika; Stiewe, Gunther; Reif, Jochen C; Zhao, Yusheng
2017-03-01
Predicting the grain yield performance of three-way hybrids is challenging. Three-way crosses are relevant for hybrid breeding in barley ( L.) and maize ( L.) adapted to East Africa. The main goal of our study was to implement and evaluate genome-wide prediction approaches of the performance of three-way hybrids using data of single-cross hybrids for a scenario in which parental lines of the three-way hybrids originate from three genetically distinct subpopulations. We extended the ridge regression best linear unbiased prediction (RRBLUP) and devised a genomic selection model allowing for subpopulation-specific marker effects (GSA-RRBLUP: general and subpopulation-specific additive RRBLUP). Using an empirical barley data set, we showed that applying GSA-RRBLUP tripled the prediction ability of three-way hybrids from 0.095 to 0.308 compared with RRBLUP, modeling one additive effect for all three subpopulations. The experimental findings were further substantiated with computer simulations. Our results emphasize the potential of GSA-RRBLUP to improve genome-wide hybrid prediction of three-way hybrids for scenarios of genetically diverse parental populations. Because of the advantages of the GSA-RRBLUP model in dealing with hybrids from different parental populations, it may also be a promising approach to boost the prediction ability for hybrid breeding programs based on genetically diverse heterotic groups. Copyright © 2017 Crop Science Society of America.
Predicting selective drug targets in cancer through metabolic networks
Folger, Ori; Jerby, Livnat; Frezza, Christian; Gottlieb, Eyal; Ruppin, Eytan; Shlomi, Tomer
2011-01-01
The interest in studying metabolic alterations in cancer and their potential role as novel targets for therapy has been rejuvenated in recent years. Here, we report the development of the first genome-scale network model of cancer metabolism, validated by correctly identifying genes essential for cellular proliferation in cancer cell lines. The model predicts 52 cytostatic drug targets, of which 40% are targeted by known, approved or experimental anticancer drugs, and the rest are new. It further predicts combinations of synthetic lethal drug targets, whose synergy is validated using available drug efficacy and gene expression measurements across the NCI-60 cancer cell line collection. Finally, potential selective treatments for specific cancers that depend on cancer type-specific downregulation of gene expression and somatic mutations are compiled. PMID:21694718
Neural networks to predict exosphere temperature corrections
NASA Astrophysics Data System (ADS)
Choury, Anna; Bruinsma, Sean; Schaeffer, Philippe
2013-10-01
Precise orbit prediction requires a forecast of the atmospheric drag force with a high degree of accuracy. Artificial neural networks are universal approximators derived from artificial intelligence and are widely used for prediction. This paper presents a method of artificial neural networking for prediction of the thermosphere density by forecasting exospheric temperature, which will be used by the semiempirical thermosphere Drag Temperature Model (DTM) currently developed. Artificial neural network has shown to be an effective and robust forecasting model for temperature prediction. The proposed model can be used for any mission from which temperature can be deduced accurately, i.e., it does not require specific training. Although the primary goal of the study was to create a model for 1 day ahead forecast, the proposed architecture has been generalized to 2 and 3 days prediction as well. The impact of artificial neural network predictions has been quantified for the low-orbiting satellite Gravity Field and Steady-State Ocean Circulation Explorer in 2011, and an order of magnitude smaller orbit errors were found when compared with orbits propagated using the thermosphere model DTM2009.
TU-F-17A-03: An Analytical Respiratory Perturbation Model for Lung Motion Prediction
DOE Office of Scientific and Technical Information (OSTI.GOV)
Li, G; Yuan, A; Wei, J
2014-06-15
Purpose: Breathing irregularity is common, causing unreliable prediction in tumor motion for correlation-based surrogates. Both tidal volume (TV) and breathing pattern (BP=ΔVthorax/TV, where TV=ΔVthorax+ΔVabdomen) affect lung motion in anterior-posterior and superior-inferior directions. We developed a novel respiratory motion perturbation (RMP) model in analytical form to account for changes in TV and BP in motion prediction from simulation to treatment. Methods: The RMP model is an analytical function of patient-specific anatomic and physiologic parameters. It contains a base-motion trajectory d(x,y,z) derived from a 4-dimensional computed tomography (4DCT) at simulation and a perturbation term Δd(ΔTV,ΔBP) accounting for deviation at treatment from simulation.more » The perturbation is dependent on tumor-specific location and patient-specific anatomy. Eleven patients with simulation and treatment 4DCT images were used to assess the RMP method in motion prediction from 4DCT1 to 4DCT2, and vice versa. For each patient, ten motion trajectories of corresponding points in the lower lobes were measured in both 4DCTs: one served as the base-motion trajectory and the other as the ground truth for comparison. In total, 220 motion trajectory predictions were assessed. The motion discrepancy between two 4DCTs for each patient served as a control. An established 5D motion model was used for comparison. Results: The average absolute error of RMP model prediction in superior-inferior direction is 1.6±1.8 mm, similar to 1.7±1.6 mm from the 5D model (p=0.98). Some uncertainty is associated with limited spatial resolution (2.5mm slice thickness) and temporal resolution (10-phases). Non-corrected motion discrepancy between two 4DCTs is 2.6±2.7mm, with the maximum of ±20mm, and correction is necessary (p=0.01). Conclusion: The analytical motion model predicts lung motion with accuracy similar to the 5D model. The analytical model is based on physical relationships, requires no training, and therefore is potentially more resilient to breathing irregularities. On-going investigation introduces airflow into the RMP model for improvement. This research is in part supported by NIH (U54CA137788/132378). AY would like to thank MSKCC summer medical student research program supported by National Cancer Institute and hosted by Department of Medical Physics at MSKCC.« less
Xu, Jingting; Hu, Hong; Dai, Yang
The identification of enhancers is a challenging task. Various types of epigenetic information including histone modification have been utilized in the construction of enhancer prediction models based on a diverse panel of machine learning schemes. However, DNA methylation profiles generated from the whole genome bisulfite sequencing (WGBS) have not been fully explored for their potential in enhancer prediction despite the fact that low methylated regions (LMRs) have been implied to be distal active regulatory regions. In this work, we propose a prediction framework, LMethyR-SVM, using LMRs identified from cell-type-specific WGBS DNA methylation profiles and a weighted support vector machine learning framework. In LMethyR-SVM, the set of cell-type-specific LMRs is further divided into three sets: reliable positive, like positive and likely negative, according to their resemblance to a small set of experimentally validated enhancers in the VISTA database based on an estimated non-parametric density distribution. Then, the prediction model is obtained by solving a weighted support vector machine. We demonstrate the performance of LMethyR-SVM by using the WGBS DNA methylation profiles derived from the human embryonic stem cell type (H1) and the fetal lung fibroblast cell type (IMR90). The predicted enhancers are highly conserved with a reasonable validation rate based on a set of commonly used positive markers including transcription factors, p300 binding and DNase-I hypersensitive sites. In addition, we show evidence that the large fraction of the LMethyR-SVM predicted enhancers are not predicted by ChromHMM in H1 cell type and they are more enriched for the FANTOM5 enhancers. Our work suggests that low methylated regions detected from the WGBS data are useful as complementary resources to histone modification marks in developing models for the prediction of cell-type-specific enhancers.
Evaluation of puberty by verifying spontaneous and stimulated gonadotropin values in girls.
Chin, Vivian L; Cai, Ziyong; Lam, Leslie; Shah, Bina; Zhou, Ping
2015-03-01
Changes in pharmacological agents and advancements in laboratory assays have changed the gonadotropin-releasing hormone analog stimulation test. To determine the best predictive model for detecting puberty in girls. Thirty-five girls, aged 2 years 7 months to 9 years 3 months, with central precocious puberty (CPP) (n=20) or premature thelarche/premature adrenarche (n=15). Diagnoses were based on clinical information, baseline hormones, bone age, and pelvic sonogram. Gonadotropins and E2 were analyzed using immunochemiluminometric assay. Logistic regression for CPP was performed. The best predictor of CPP is the E2-change model based on 3- to 24-h values, providing 80% sensitivity and 87% specificity. Three-hour luteinizing hormone (LH) provided 75% sensitivity and 87% specificity. Basal LH lowered sensitivity to 65% and specificity to 53%. The E2-change model provided the best predictive power; however, 3-h LH was more practical and convenient when evaluating puberty in girls.
Prediction of Surface and pH-Specific Binding of Peptides to Metal and Oxide Nanoparticles
NASA Astrophysics Data System (ADS)
Heinz, Hendrik; Lin, Tzu-Jen; Emami, Fateme Sadat; Ramezani-Dakhel, Hadi; Naik, Rajesh; Knecht, Marc; Perry, Carole C.; Huang, Yu
2015-03-01
The mechanism of specific peptide adsorption onto metallic and oxidic nanostructures has been elucidated in atomic resolution using novel force fields and surface models in comparison to measurements. As an example, variations in peptide adsorption on Pd and Pt nanoparticles depending on shape, size, and location of peptides on specific bounding facets are explained. Accurate computational predictions of reaction rates in C-C coupling reactions using particle models derived from HE-XRD and PDF data illustrate the utility of computational methods for the rational design of new catalysts. On oxidic nanoparticles such as silica and apatites, it is revealed how changes in pH lead to similarity scores of attracted peptides lower than 20%, supported by appropriate model surfaces and data from adsorption isotherms. The results demonstrate how new computational methods can support the design of nanoparticle carriers for drug release and the understanding of calcification mechanisms in the human body.
An application of hybrid downscaling model to forecast summer precipitation at stations in China
NASA Astrophysics Data System (ADS)
Liu, Ying; Fan, Ke
2014-06-01
A pattern prediction hybrid downscaling method was applied to predict summer (June-July-August) precipitation at China 160 stations. The predicted precipitation from the downscaling scheme is available one month before. Four predictors were chosen to establish the hybrid downscaling scheme. The 500-hPa geopotential height (GH5) and 850-hPa specific humidity (q85) were from the skillful predicted output of three DEMETER (Development of a European Multi-model Ensemble System for Seasonal to Interannual Prediction) general circulation models (GCMs). The 700-hPa geopotential height (GH7) and sea level pressure (SLP) were from reanalysis datasets. The hybrid downscaling scheme (HD-4P) has better prediction skill than a conventional statistical downscaling model (SD-2P) which contains two predictors derived from the output of GCMs, although two downscaling schemes were performed to improve the seasonal prediction of summer rainfall in comparison with the original output of the DEMETER GCMs. In particular, HD-4P downscaling predictions showed lower root mean square errors than those based on the SD-2P model. Furthermore, the HD-4P downscaling model reproduced the China summer precipitation anomaly centers more accurately than the scenario of the SD-2P model in 1998. A hybrid downscaling prediction should be effective to improve the prediction skill of summer rainfall at stations in China.
Genomic-Enabled Prediction in Maize Using Kernel Models with Genotype × Environment Interaction
Bandeira e Sousa, Massaine; Cuevas, Jaime; de Oliveira Couto, Evellyn Giselly; Pérez-Rodríguez, Paulino; Jarquín, Diego; Fritsche-Neto, Roberto; Burgueño, Juan; Crossa, Jose
2017-01-01
Multi-environment trials are routinely conducted in plant breeding to select candidates for the next selection cycle. In this study, we compare the prediction accuracy of four developed genomic-enabled prediction models: (1) single-environment, main genotypic effect model (SM); (2) multi-environment, main genotypic effects model (MM); (3) multi-environment, single variance G×E deviation model (MDs); and (4) multi-environment, environment-specific variance G×E deviation model (MDe). Each of these four models were fitted using two kernel methods: a linear kernel Genomic Best Linear Unbiased Predictor, GBLUP (GB), and a nonlinear kernel Gaussian kernel (GK). The eight model-method combinations were applied to two extensive Brazilian maize data sets (HEL and USP data sets), having different numbers of maize hybrids evaluated in different environments for grain yield (GY), plant height (PH), and ear height (EH). Results show that the MDe and the MDs models fitted with the Gaussian kernel (MDe-GK, and MDs-GK) had the highest prediction accuracy. For GY in the HEL data set, the increase in prediction accuracy of SM-GK over SM-GB ranged from 9 to 32%. For the MM, MDs, and MDe models, the increase in prediction accuracy of GK over GB ranged from 9 to 49%. For GY in the USP data set, the increase in prediction accuracy of SM-GK over SM-GB ranged from 0 to 7%. For the MM, MDs, and MDe models, the increase in prediction accuracy of GK over GB ranged from 34 to 70%. For traits PH and EH, gains in prediction accuracy of models with GK compared to models with GB were smaller than those achieved in GY. Also, these gains in prediction accuracy decreased when a more difficult prediction problem was studied. PMID:28455415
Genomic-Enabled Prediction in Maize Using Kernel Models with Genotype × Environment Interaction.
Bandeira E Sousa, Massaine; Cuevas, Jaime; de Oliveira Couto, Evellyn Giselly; Pérez-Rodríguez, Paulino; Jarquín, Diego; Fritsche-Neto, Roberto; Burgueño, Juan; Crossa, Jose
2017-06-07
Multi-environment trials are routinely conducted in plant breeding to select candidates for the next selection cycle. In this study, we compare the prediction accuracy of four developed genomic-enabled prediction models: (1) single-environment, main genotypic effect model (SM); (2) multi-environment, main genotypic effects model (MM); (3) multi-environment, single variance G×E deviation model (MDs); and (4) multi-environment, environment-specific variance G×E deviation model (MDe). Each of these four models were fitted using two kernel methods: a linear kernel Genomic Best Linear Unbiased Predictor, GBLUP (GB), and a nonlinear kernel Gaussian kernel (GK). The eight model-method combinations were applied to two extensive Brazilian maize data sets (HEL and USP data sets), having different numbers of maize hybrids evaluated in different environments for grain yield (GY), plant height (PH), and ear height (EH). Results show that the MDe and the MDs models fitted with the Gaussian kernel (MDe-GK, and MDs-GK) had the highest prediction accuracy. For GY in the HEL data set, the increase in prediction accuracy of SM-GK over SM-GB ranged from 9 to 32%. For the MM, MDs, and MDe models, the increase in prediction accuracy of GK over GB ranged from 9 to 49%. For GY in the USP data set, the increase in prediction accuracy of SM-GK over SM-GB ranged from 0 to 7%. For the MM, MDs, and MDe models, the increase in prediction accuracy of GK over GB ranged from 34 to 70%. For traits PH and EH, gains in prediction accuracy of models with GK compared to models with GB were smaller than those achieved in GY. Also, these gains in prediction accuracy decreased when a more difficult prediction problem was studied. Copyright © 2017 Bandeira e Sousa et al.
Towards a generalized energy prediction model for machine tools
Bhinge, Raunak; Park, Jinkyoo; Law, Kincho H.; Dornfeld, David A.; Helu, Moneer; Rachuri, Sudarsan
2017-01-01
Energy prediction of machine tools can deliver many advantages to a manufacturing enterprise, ranging from energy-efficient process planning to machine tool monitoring. Physics-based, energy prediction models have been proposed in the past to understand the energy usage pattern of a machine tool. However, uncertainties in both the machine and the operating environment make it difficult to predict the energy consumption of the target machine reliably. Taking advantage of the opportunity to collect extensive, contextual, energy-consumption data, we discuss a data-driven approach to develop an energy prediction model of a machine tool in this paper. First, we present a methodology that can efficiently and effectively collect and process data extracted from a machine tool and its sensors. We then present a data-driven model that can be used to predict the energy consumption of the machine tool for machining a generic part. Specifically, we use Gaussian Process (GP) Regression, a non-parametric machine-learning technique, to develop the prediction model. The energy prediction model is then generalized over multiple process parameters and operations. Finally, we apply this generalized model with a method to assess uncertainty intervals to predict the energy consumed to machine any part using a Mori Seiki NVD1500 machine tool. Furthermore, the same model can be used during process planning to optimize the energy-efficiency of a machining process. PMID:28652687
Towards a generalized energy prediction model for machine tools.
Bhinge, Raunak; Park, Jinkyoo; Law, Kincho H; Dornfeld, David A; Helu, Moneer; Rachuri, Sudarsan
2017-04-01
Energy prediction of machine tools can deliver many advantages to a manufacturing enterprise, ranging from energy-efficient process planning to machine tool monitoring. Physics-based, energy prediction models have been proposed in the past to understand the energy usage pattern of a machine tool. However, uncertainties in both the machine and the operating environment make it difficult to predict the energy consumption of the target machine reliably. Taking advantage of the opportunity to collect extensive, contextual, energy-consumption data, we discuss a data-driven approach to develop an energy prediction model of a machine tool in this paper. First, we present a methodology that can efficiently and effectively collect and process data extracted from a machine tool and its sensors. We then present a data-driven model that can be used to predict the energy consumption of the machine tool for machining a generic part. Specifically, we use Gaussian Process (GP) Regression, a non-parametric machine-learning technique, to develop the prediction model. The energy prediction model is then generalized over multiple process parameters and operations. Finally, we apply this generalized model with a method to assess uncertainty intervals to predict the energy consumed to machine any part using a Mori Seiki NVD1500 machine tool. Furthermore, the same model can be used during process planning to optimize the energy-efficiency of a machining process.
van Dijk, Lisanne V; Brouwer, Charlotte L; van der Schaaf, Arjen; Burgerhof, Johannes G M; Beukinga, Roelof J; Langendijk, Johannes A; Sijtsema, Nanna M; Steenbakkers, Roel J H M
2017-02-01
Current models for the prediction of late patient-rated moderate-to-severe xerostomia (XER 12m ) and sticky saliva (STIC 12m ) after radiotherapy are based on dose-volume parameters and baseline xerostomia (XER base ) or sticky saliva (STIC base ) scores. The purpose is to improve prediction of XER 12m and STIC 12m with patient-specific characteristics, based on CT image biomarkers (IBMs). Planning CT-scans and patient-rated outcome measures were prospectively collected for 249 head and neck cancer patients treated with definitive radiotherapy with or without systemic treatment. The potential IBMs represent geometric, CT intensity and textural characteristics of the parotid and submandibular glands. Lasso regularisation was used to create multivariable logistic regression models, which were internally validated by bootstrapping. The prediction of XER 12m could be improved significantly by adding the IBM "Short Run Emphasis" (SRE), which quantifies heterogeneity of parotid tissue, to a model with mean contra-lateral parotid gland dose and XER base . For STIC 12m , the IBM maximum CT intensity of the submandibular gland was selected in addition to STIC base and mean dose to submandibular glands. Prediction of XER 12m and STIC 12m was improved by including IBMs representing heterogeneity and density of the salivary glands, respectively. These IBMs could guide additional research to the patient-specific response of healthy tissue to radiation dose. Copyright © 2016 The Authors. Published by Elsevier B.V. All rights reserved.
Brubaker, Douglas; Difeo, Analisa; Chen, Yanwen; Pearl, Taylor; Zhai, Kaide; Bebek, Gurkan; Chance, Mark; Barnholtz-Sloan, Jill
2014-01-01
The revolution in sequencing techniques in the past decade has provided an extensive picture of the molecular mechanisms behind complex diseases such as cancer. The Cancer Cell Line Encyclopedia (CCLE) and The Cancer Genome Project (CGP) have provided an unprecedented opportunity to examine copy number, gene expression, and mutational information for over 1000 cell lines of multiple tumor types alongside IC50 values for over 150 different drugs and drug related compounds. We present a novel pipeline called DIRPP, Drug Intervention Response Predictions with PARADIGM7, which predicts a cell line's response to a drug intervention from molecular data. PARADIGM (Pathway Recognition Algorithm using Data Integration on Genomic Models) is a probabilistic graphical model used to infer patient specific genetic activity by integrating copy number and gene expression data into a factor graph model of a cellular network. We evaluated the performance of DIRPP on endometrial, ovarian, and breast cancer related cell lines from the CCLE and CGP for nine drugs. The pipeline is sensitive enough to predict the response of a cell line with accuracy and precision across datasets as high as 80 and 88% respectively. We then classify drugs by the specific pathway mechanisms governing drug response. This classification allows us to compare drugs by cellular response mechanisms rather than simply by their specific gene targets. This pipeline represents a novel approach for predicting clinical drug response and generating novel candidates for drug repurposing and repositioning.
Clegg, Lindsay Wendel; Mac Gabhann, Feilim
2015-01-01
Matrix-binding isoforms and non-matrix-binding isoforms of vascular endothelial growth factor (VEGF) are both capable of stimulating vascular remodeling, but the resulting blood vessel networks are structurally and functionally different. Here, we develop and validate a computational model of the binding of soluble and immobilized ligands to VEGF receptor 2 (VEGFR2), the endosomal trafficking of VEGFR2, and site-specific VEGFR2 tyrosine phosphorylation to study differences in induced signaling between these VEGF isoforms. In capturing essential features of VEGFR2 signaling and trafficking, our model suggests that VEGFR2 trafficking parameters are largely consistent across multiple endothelial cell lines. Simulations demonstrate distinct localization of VEGFR2 phosphorylated on Y1175 and Y1214. This is the first model to clearly show that differences in site-specific VEGFR2 activation when stimulated with immobilized VEGF compared to soluble VEGF can be accounted for by altered trafficking of VEGFR2 without an intrinsic difference in receptor activation. The model predicts that Neuropilin-1 can induce differences in the surface-to-internal distribution of VEGFR2. Simulations also show that ligated VEGFR2 and phosphorylated VEGFR2 levels diverge over time following stimulation. Using this model, we identify multiple key levers that alter how VEGF binding to VEGFR2 results in different coordinated patterns of multiple downstream signaling pathways. Specifically, simulations predict that VEGF immobilization, interactions with Neuropilin-1, perturbations of VEGFR2 trafficking, and changes in expression or activity of phosphatases acting on VEGFR2 all affect the magnitude, duration, and relative strength of VEGFR2 phosphorylation on tyrosines 1175 and 1214, and they do so predictably within our single consistent model framework. PMID:26067165
Speiser, Jaime Lynn; Lee, William M; Karvellas, Constantine J
2015-01-01
Assessing prognosis for acetaminophen-induced acute liver failure (APAP-ALF) patients often presents significant challenges. King's College (KCC) has been validated on hospital admission, but little has been published on later phases of illness. We aimed to improve determinations of prognosis both at the time of and following admission for APAP-ALF using Classification and Regression Tree (CART) models. CART models were applied to US ALFSG registry data to predict 21-day death or liver transplant early (on admission) and post-admission (days 3-7) for 803 APAP-ALF patients enrolled 01/1998-09/2013. Accuracy in prediction of outcome (AC), sensitivity (SN), specificity (SP), and area under receiver-operating curve (AUROC) were compared between 3 models: KCC (INR, creatinine, coma grade, pH), CART analysis using only KCC variables (KCC-CART) and a CART model using new variables (NEW-CART). Traditional KCC yielded 69% AC, 90% SP, 27% SN, and 0.58 AUROC on admission, with similar performance post-admission. KCC-CART at admission offered predictive 66% AC, 65% SP, 67% SN, and 0.74 AUROC. Post-admission, KCC-CART had predictive 82% AC, 86% SP, 46% SN and 0.81 AUROC. NEW-CART models using MELD (Model for end stage liver disease), lactate and mechanical ventilation on admission yielded predictive 72% AC, 71% SP, 77% SN and AUROC 0.79. For later stages, NEW-CART (MELD, lactate, coma grade) offered predictive AC 86%, SP 91%, SN 46%, AUROC 0.73. CARTs offer simple prognostic models for APAP-ALF patients, which have higher AUROC and SN than KCC, with similar AC and negligibly worse SP. Admission and post-admission predictions were developed. • Prognostication in acetaminophen-induced acute liver failure (APAP-ALF) is challenging beyond admission • Little has been published regarding the use of King's College Criteria (KCC) beyond admission and KCC has shown limited sensitivity in subsequent studies • Classification and Regression Tree (CART) methodology allows the development of predictive models using binary splits and offers an intuitive method for predicting outcome, using processes familiar to clinicians • Data from the ALFSG registry suggested that CART prognosis models for the APAP population offer improved sensitivity and model performance over traditional regression-based KCC, while maintaining similar accuracy and negligibly worse specificity • KCC-CART models offered modest improvement over traditional KCC, with NEW-CART models performing better than KCC-CART particularly at late time points.
Predicting Ideological Prejudice
Brandt, Mark J.
2017-01-01
A major shortcoming of current models of ideological prejudice is that although they can anticipate the direction of the association between participants’ ideology and their prejudice against a range of target groups, they cannot predict the size of this association. I developed and tested models that can make specific size predictions for this association. A quantitative model that used the perceived ideology of the target group as the primary predictor of the ideology-prejudice relationship was developed with a representative sample of Americans (N = 4,940) and tested against models using the perceived status of and choice to belong to the target group as predictors. In four studies (total N = 2,093), ideology-prejudice associations were estimated, and these observed estimates were compared with the models’ predictions. The model that was based only on perceived ideology was the most parsimonious with the smallest errors. PMID:28394693
Uzun, Harun; Yıldız, Zeynep; Goldfarb, Jillian L; Ceylan, Selim
2017-06-01
As biomass becomes more integrated into our energy feedstocks, the ability to predict its combustion enthalpies from routine data such as carbon, ash, and moisture content enables rapid decisions about utilization. The present work constructs a novel artificial neural network model with a 3-3-1 tangent sigmoid architecture to predict biomasses' higher heating values from only their proximate analyses, requiring minimal specificity as compared to models based on elemental composition. The model presented has a considerably higher correlation coefficient (0.963) and lower root mean square (0.375), mean absolute (0.328), and mean bias errors (0.010) than other models presented in the literature which, at least when applied to the present data set, tend to under-predict the combustion enthalpy. Copyright © 2017 Elsevier Ltd. All rights reserved.
Potential impact of remote sensing data on sea-state analysis and prediction
NASA Technical Reports Server (NTRS)
Cardone, V. J.
1983-01-01
The severe North Atlantic storm which damaged the ocean liner Queen Elizabeth 2 (QE2) was studied to assess the impact of remotely sensed marine surface wind data obtained by SEASAT-A, on sea state specifications and forecasts. Alternate representations of the surface wind field in the QE2 storm were produced from the SEASAT enhanced data base, and from operational analyses based upon conventional data. The wind fields were used to drive a high resolution spectral ocean surface wave prediction model. Results show that sea state analyses would have been vastly improved during the period of storm formation and explosive development had remote sensing wind data been available in real time. A modest improvement in operational 12 to 24 hour wave forecasts would have followed automatically from the improved initial state specification made possible by the remote sensing data in both numerical and sea state prediction models. Significantly improved 24 to 48 hour wave forecasts require in addition to remote sensing data, refinement in the numerical and physical aspects of weather prediction models.
Garcia, Luís Filipe; de Oliveira, Luís Caldas; de Matos, David Martins
2016-01-01
This study compared the performance of two statistical location-aware pictogram prediction mechanisms, with an all-purpose (All) pictogram prediction mechanism, having no location knowledge. The All approach had a unique language model under all locations. One of the location-aware alternatives, the location-specific (Spec) approach, made use of specific language models for pictogram prediction in each location of interest. The other location-aware approach resulted from combining the Spec and the All approaches, and was designated the mixed approach (Mix). In this approach, the language models acquired knowledge from all locations, but a higher relevance was assigned to the vocabulary from the associated location. Results from simulations showed that the Mix and Spec approaches could only outperform the baseline in a statistically significant way if pictogram users reuse more than 50% and 75% of their sentences, respectively. Under low sentence reuse conditions there were no statistically significant differences between the location-aware approaches and the All approach. Under these conditions, the Mix approach performed better than the Spec approach in a statistically significant way.
Lippke, Sonia; Plotnikoff, Ronald C
2009-05-01
Two different theories of health behaviour have been chosen with the aim of theory integration: a continuous theory (protection motivation theory, PMT) and a stage model (transtheoretical model, TTM). This is the first study to test whether the stages of the TTM moderate the interrelation of PMT-variables and the mediation of motivation, as well as PMT-variables' interactions in predicting stage transitions. Hypotheses were tested regarding (1) mean patterns, stage pair-comparisons and nonlinear trends using ANOVAs; (2) prediction-patterns for the different stage groups employing multi-group structural equation modelling (MSEM) and nested model analyses; and (3) stage transitions using binary logistic regression analyses. Adults (N=1,602) were assessed over a 6 month period on their physical activity stages, PMT-variables and subsequent behaviour. (1) Particular mean differences and nonlinear trends in all test variables were found. (2) The PMT adequately fitted the five stage groups. The MSEM revealed that covariances within threat appraisal and coping appraisal were invariant and all other constrains were stage-specific, i.e. stage was a moderator. Except for self-efficacy, motivation fully mediated the relationship between the social-cognitive variables and behaviour. (3) Predicting stage transitions with the PMT-variables underscored the importance of self-efficacy. Only when threat appraisal and coping appraisal were high, stage movement was more likely in the preparation stage. Results emphasize stage-specific differences of the PMT mechanisms, and hence, support the stage construct. The findings may guide further theory building and research integrating different theoretical approaches.
Shen, Weidong; Sakamoto, Naoko; Yang, Limin
2016-07-07
The objectives of this study were to evaluate and model the probability of melanoma-specific death and competing causes of death for patients with melanoma by competing risk analysis, and to build competing risk nomograms to provide individualized and accurate predictive tools. Melanoma data were obtained from the Surveillance Epidemiology and End Results program. All patients diagnosed with primary non-metastatic melanoma during the years 2004-2007 were potentially eligible for inclusion. The cumulative incidence function (CIF) was used to describe the probability of melanoma mortality and competing risk mortality. We used Gray's test to compare differences in CIF between groups. The proportional subdistribution hazard approach by Fine and Gray was used to model CIF. We built competing risk nomograms based on the models that we developed. The 5-year cumulative incidence of melanoma death was 7.1 %, and the cumulative incidence of other causes of death was 7.4 %. We identified that variables associated with an elevated probability of melanoma-specific mortality included older age, male sex, thick melanoma, ulcerated cancer, and positive lymph nodes. The nomograms were well calibrated. C-indexes were 0.85 and 0.83 for nomograms predicting the probability of melanoma mortality and competing risk mortality, which suggests good discriminative ability. This large study cohort enabled us to build a reliable competing risk model and nomogram for predicting melanoma prognosis. Model performance proved to be good. This individualized predictive tool can be used in clinical practice to help treatment-related decision making.
Regression models to predict hip joint centers in pathological hip population.
Mantovani, Giulia; Ng, K C Geoffrey; Lamontagne, Mario
2016-02-01
The purpose was to investigate the validity of Harrington's and Davis's hip joint center (HJC) regression equations on a population affected by a hip deformity, (i.e., femoroacetabular impingement). Sixty-seven participants (21 healthy controls, 46 with a cam-type deformity) underwent pelvic CT imaging. Relevant bony landmarks and geometric HJCs were digitized from the images, and skin thickness was measured for the anterior and posterior superior iliac spines. Non-parametric statistical and Bland-Altman tests analyzed differences between the predicted HJC (from regression equations) and the actual HJC (from CT images). The error from Davis's model (25.0 ± 6.7 mm) was larger than Harrington's (12.3 ± 5.9 mm, p<0.001). There were no differences between groups, thus, studies on femoroacetabular impingement can implement conventional regression models. Measured skin thickness was 9.7 ± 7.0mm and 19.6 ± 10.9 mm for the anterior and posterior bony landmarks, respectively, and correlated with body mass index. Skin thickness estimates can be considered to reduce the systematic error introduced by surface markers. New adult-specific regression equations were developed from the CT dataset, with the hypothesis that they could provide better estimates when tuned to a larger adult-specific dataset. The linear models were validated on external datasets and using leave-one-out cross-validation techniques; Prediction errors were comparable to those of Harrington's model, despite the adult-specific population and the larger sample size, thus, prediction accuracy obtained from these parameters could not be improved. Copyright © 2015 Elsevier B.V. All rights reserved.
Bootstrap Prediction Intervals in Non-Parametric Regression with Applications to Anomaly Detection
NASA Technical Reports Server (NTRS)
Kumar, Sricharan; Srivistava, Ashok N.
2012-01-01
Prediction intervals provide a measure of the probable interval in which the outputs of a regression model can be expected to occur. Subsequently, these prediction intervals can be used to determine if the observed output is anomalous or not, conditioned on the input. In this paper, a procedure for determining prediction intervals for outputs of nonparametric regression models using bootstrap methods is proposed. Bootstrap methods allow for a non-parametric approach to computing prediction intervals with no specific assumptions about the sampling distribution of the noise or the data. The asymptotic fidelity of the proposed prediction intervals is theoretically proved. Subsequently, the validity of the bootstrap based prediction intervals is illustrated via simulations. Finally, the bootstrap prediction intervals are applied to the problem of anomaly detection on aviation data.
Fenlon, Caroline; O'Grady, Luke; Butler, Stephen; Doherty, Michael L; Dunnion, John
2017-01-01
Herd fertility in pasture-based dairy farms is a key driver of farm economics. Models for predicting nulliparous reproductive outcomes are rare, but age, genetics, weight, and BCS have been identified as factors influencing heifer conception. The aim of this study was to create a simulation model of heifer conception to service with thorough evaluation. Artificial Insemination service records from two research herds and ten commercial herds were provided to build and evaluate the models. All were managed as spring-calving pasture-based systems. The factors studied were related to age, genetics, and time of service. The data were split into training and testing sets and bootstrapping was used to train the models. Logistic regression (with and without random effects) and generalised additive modelling were selected as the model-building techniques. Two types of evaluation were used to test the predictive ability of the models: discrimination and calibration. Discrimination, which includes sensitivity, specificity, accuracy and ROC analysis, measures a model's ability to distinguish between positive and negative outcomes. Calibration measures the accuracy of the predicted probabilities with the Hosmer-Lemeshow goodness-of-fit, calibration plot and calibration error. After data cleaning and the removal of services with missing values, 1396 services remained to train the models and 597 were left for testing. Age, breed, genetic predicted transmitting ability for calving interval, month and year were significant in the multivariate models. The regression models also included an interaction between age and month. Year within herd was a random effect in the mixed regression model. Overall prediction accuracy was between 77.1% and 78.9%. All three models had very high sensitivity, but low specificity. The two regression models were very well-calibrated. The mean absolute calibration errors were all below 4%. Because the models were not adept at identifying unsuccessful services, they are not suggested for use in predicting the outcome of individual heifer services. Instead, they are useful for the comparison of services with different covariate values or as sub-models in whole-farm simulations. The mixed regression model was identified as the best model for prediction, as the random effects can be ignored and the other variables can be easily obtained or simulated.
NASA Astrophysics Data System (ADS)
Singh, Savita; Singh, Alok; Sharma, Sudhir Kumar
2017-06-01
In this paper, an analytical modeling and prediction of tensile and flexural strength of three dimensional micro-scaled novel coconut shell powder (CSP) reinforced epoxy polymer composites have been reported. The novel CSP has a specific mixing ratio of different coconut shell particle size. A comparison is made between obtained experimental strength and modified Guth model. The result shows a strong evidence for non-validation of modified Guth model for strength prediction. Consequently, a constitutive modeled equation named Singh model has been developed to predict the tensile and flexural strength of this novel CSP reinforced epoxy composite. Moreover, high resolution Raman spectrum shows that 40 % CSP reinforced epoxy composite has high dielectric constant to become an alternative material for capacitance whereas fractured surface morphology revealed that a strong bonding between novel CSP and epoxy polymer for the application as light weight composite materials in engineering.
A productivity model for parasitized, multibrooded songbirds
Powell, L.A.; Knutson, M.G.
2006-01-01
We present an enhancement of a simulation model to predict annual productivity for Wood Thrushes (Hylocichla mustelina) and American Redstarts (Setophaga ruticilla); the model includes effects of Brown-headed Cowbird (Molothrus ater) parasitism. We used species-specific data from the Driftless Area Ecoregion of Wisconsin, Minnesota, and Iowa to parameterize the model as a case study. The simulation model predicted annual productivity of 2.03 ?? 1.60 SD for Wood Thrushes and 1.56 ?? 1.31 SD for American Redstarts. Our sensitivity analysis showed that high parasitism lowered Wood Thrush annual productivity more than American Redstart productivity, even though parasitism affected individual nests of redstarts more severely. Annual productivity predictions are valuable for habitat managers, but productivity is not easily obtained from field studies. Our model provides a useful means of integrating complex life history parameters to predict productivity for songbirds that experience nest parasitism. ?? The Cooper Ornithological Society 2006.
Brown, Timothy A.; Naragon-Gainey, Kristin
2013-01-01
The triple vulnerability model (Barlow, 2000, 2002) posits that three vulnerabilities contribute to the etiology of emotional disorders: (1) general biological vulnerability (i.e., dimensions of temperament such as neuroticism and extraversion); (2) general psychological vulnerability (i.e., perceived control over life stress and emotional states); (3) disorder-specific psychological vulnerability (e.g., thought-action fusion for obsessive-compulsive disorder, OCD). Despite the prominence of this model, a comprehensive empirical evaluation has not yet been undertaken. The current study used structural equation modeling to test the triple vulnerability model in a large clinical sample (N = 700), focusing on vulnerabilities for depression, social phobia, generalized anxiety disorder (GAD), and OCD. Specifically, we examined the incremental prediction of each level of the triple vulnerability model for each disorder, with the following putative disorder-specific psychological vulnerabilities: thought-action fusion (TAF) for OCD, the dysfunctional attitudes (DAS) for depression, and intolerance of uncertainty (IoU) for GAD. In the final model that included all three levels of vulnerabilities, neuroticism had significant direct effects on all four disorder constructs, and extraversion was inversely associated with depression and social phobia. However, perceived control was significantly associated with GAD and OCD only. Of the disorder-specific psychological vulnerabilities, TAF was significantly and specifically related to OCD. In contrast, DAS and IoU were not significant predictors of depression and GAD respectively, instead contributing to other disorders. The results are discussed in regard to structural models of the emotional disorders and the various roles of general and specific vulnerability dimensions in the onset, severity, and temporal course of psychopathology. PMID:23611077
DOE Office of Scientific and Technical Information (OSTI.GOV)
Brothers, Michael C; Nesbitt, Anna E; Hallock, Michael J
2011-01-01
Homology modeling is a powerful tool for predicting protein structures, whose success depends on obtaining a reasonable alignment between a given structural template and the protein sequence being analyzed. In order to leverage greater predictive power for proteins with few structural templates, we have developed a method to rank homology models based upon their compliance to secondary structure derived from experimental solid-state NMR (SSNMR) data. Such data is obtainable in a rapid manner by simple SSNMR experiments (e.g., (13)C-(13)C 2D correlation spectra). To test our homology model scoring procedure for various amino acid labeling schemes, we generated a library ofmore » 7,474 homology models for 22 protein targets culled from the TALOS+/SPARTA+ training set of protein structures. Using subsets of amino acids that are plausibly assigned by SSNMR, we discovered that pairs of the residues Val, Ile, Thr, Ala and Leu (VITAL) emulate an ideal dataset where all residues are site specifically assigned. Scoring the models with a predicted VITAL site-specific dataset and calculating secondary structure with the Chemical Shift Index resulted in a Pearson correlation coefficient (-0.75) commensurate to the control (-0.77), where secondary structure was scored site specifically for all amino acids (ALL 20) using STRIDE. This method promises to accelerate structure procurement by SSNMR for proteins with unknown folds through guiding the selection of remotely homologous protein templates and assessing model quality.« less
Shah, Neomi; Hanna, David B; Teng, Yanping; Sotres-Alvarez, Daniela; Hall, Martica; Loredo, Jose S; Zee, Phyllis; Kim, Mimi; Yaggi, H Klar; Redline, Susan; Kaplan, Robert C
2016-06-01
We developed and validated the first-ever sleep apnea (SA) risk calculator in a large population-based cohort of Hispanic/Latino subjects. Cross-sectional data on adults from the Hispanic Community Health Study/Study of Latinos (2008-2011) were analyzed. Subjective and objective sleep measurements were obtained. Clinically significant SA was defined as an apnea-hypopnea index ≥ 15 events per hour. Using logistic regression, four prediction models were created: three sex-specific models (female-only, male-only, and a sex × covariate interaction model to allow differential predictor effects), and one overall model with sex included as a main effect only. Models underwent 10-fold cross-validation and were assessed by using the C statistic. SA and its predictive variables; a total of 17 variables were considered. A total of 12,158 participants had complete sleep data available; 7,363 (61%) were women. The population-weighted prevalence of SA (apnea-hypopnea index ≥ 15 events per hour) was 6.1% in female subjects and 13.5% in male subjects. Male-only (C statistic, 0.808) and female-only (C statistic, 0.836) prediction models had the same predictor variables (ie, age, BMI, self-reported snoring). The sex-interaction model (C statistic, 0.836) contained sex, age, age × sex, BMI, BMI × sex, and self-reported snoring. The final overall model (C statistic, 0.832) contained age, BMI, snoring, and sex. We developed two websites for our SA risk calculator: one in English (https://www.montefiore.org/sleepapneariskcalc.html) and another in Spanish (http://www.montefiore.org/sleepapneariskcalc-es.html). We created an internally validated, highly discriminating, well-calibrated, and parsimonious prediction model for SA. Contrary to the study hypothesis, the variables did not have different predictive magnitudes in male and female subjects. Copyright © 2016 American College of Chest Physicians. Published by Elsevier Inc. All rights reserved.
Ng, Kenney; Ghoting, Amol; Steinhubl, Steven R.; Stewart, Walter F.; Malin, Bradley; Sun, Jimeng
2014-01-01
Objective Healthcare analytics research increasingly involves the construction of predictive models for disease targets across varying patient cohorts using electronic health records (EHRs). To facilitate this process, it is critical to support a pipeline of tasks: 1) cohort construction, 2) feature construction, 3) cross-validation, 4) feature selection, and 5) classification. To develop an appropriate model, it is necessary to compare and refine models derived from a diversity of cohorts, patient-specific features, and statistical frameworks. The goal of this work is to develop and evaluate a predictive modeling platform that can be used to simplify and expedite this process for health data. Methods To support this goal, we developed a PARAllel predictive MOdeling (PARAMO) platform which 1) constructs a dependency graph of tasks from specifications of predictive modeling pipelines, 2) schedules the tasks in a topological ordering of the graph, and 3) executes those tasks in parallel. We implemented this platform using Map-Reduce to enable independent tasks to run in parallel in a cluster computing environment. Different task scheduling preferences are also supported. Results We assess the performance of PARAMO on various workloads using three datasets derived from the EHR systems in place at Geisinger Health System and Vanderbilt University Medical Center and an anonymous longitudinal claims database. We demonstrate significant gains in computational efficiency against a standard approach. In particular, PARAMO can build 800 different models on a 300,000 patient data set in 3 hours in parallel compared to 9 days if running sequentially. Conclusion This work demonstrates that an efficient parallel predictive modeling platform can be developed for EHR data. This platform can facilitate large-scale modeling endeavors and speed-up the research workflow and reuse of health information. This platform is only a first step and provides the foundation for our ultimate goal of building analytic pipelines that are specialized for health data researchers. PMID:24370496
Ng, Kenney; Ghoting, Amol; Steinhubl, Steven R; Stewart, Walter F; Malin, Bradley; Sun, Jimeng
2014-04-01
Healthcare analytics research increasingly involves the construction of predictive models for disease targets across varying patient cohorts using electronic health records (EHRs). To facilitate this process, it is critical to support a pipeline of tasks: (1) cohort construction, (2) feature construction, (3) cross-validation, (4) feature selection, and (5) classification. To develop an appropriate model, it is necessary to compare and refine models derived from a diversity of cohorts, patient-specific features, and statistical frameworks. The goal of this work is to develop and evaluate a predictive modeling platform that can be used to simplify and expedite this process for health data. To support this goal, we developed a PARAllel predictive MOdeling (PARAMO) platform which (1) constructs a dependency graph of tasks from specifications of predictive modeling pipelines, (2) schedules the tasks in a topological ordering of the graph, and (3) executes those tasks in parallel. We implemented this platform using Map-Reduce to enable independent tasks to run in parallel in a cluster computing environment. Different task scheduling preferences are also supported. We assess the performance of PARAMO on various workloads using three datasets derived from the EHR systems in place at Geisinger Health System and Vanderbilt University Medical Center and an anonymous longitudinal claims database. We demonstrate significant gains in computational efficiency against a standard approach. In particular, PARAMO can build 800 different models on a 300,000 patient data set in 3h in parallel compared to 9days if running sequentially. This work demonstrates that an efficient parallel predictive modeling platform can be developed for EHR data. This platform can facilitate large-scale modeling endeavors and speed-up the research workflow and reuse of health information. This platform is only a first step and provides the foundation for our ultimate goal of building analytic pipelines that are specialized for health data researchers. Copyright © 2013 Elsevier Inc. All rights reserved.
Integrating in silico models to enhance predictivity for developmental toxicity.
Marzo, Marco; Kulkarni, Sunil; Manganaro, Alberto; Roncaglioni, Alessandra; Wu, Shengde; Barton-Maclaren, Tara S; Lester, Cathy; Benfenati, Emilio
2016-08-31
Application of in silico models to predict developmental toxicity has demonstrated limited success particularly when employed as a single source of information. It is acknowledged that modelling the complex outcomes related to this endpoint is a challenge; however, such models have been developed and reported in the literature. The current study explored the possibility of integrating the selected public domain models (CAESAR, SARpy and P&G model) with the selected commercial modelling suites (Multicase, Leadscope and Derek Nexus) to assess if there is an increase in overall predictive performance. The results varied according to the data sets used to assess performance which improved upon model integration relative to individual models. Moreover, because different models are based on different specific developmental toxicity effects, integration of these models increased the applicable chemical and biological spaces. It is suggested that this approach reduces uncertainty associated with in silico predictions by achieving a consensus among a battery of models. The use of tools to assess the applicability domain also improves the interpretation of the predictions. This has been verified in the case of the software VEGA, which makes freely available QSAR models with a measurement of the applicability domain. Copyright © 2016 Elsevier Ireland Ltd. All rights reserved.
Large-scale structure prediction by improved contact predictions and model quality assessment.
Michel, Mirco; Menéndez Hurtado, David; Uziela, Karolis; Elofsson, Arne
2017-07-15
Accurate contact predictions can be used for predicting the structure of proteins. Until recently these methods were limited to very big protein families, decreasing their utility. However, recent progress by combining direct coupling analysis with machine learning methods has made it possible to predict accurate contact maps for smaller families. To what extent these predictions can be used to produce accurate models of the families is not known. We present the PconsFold2 pipeline that uses contact predictions from PconsC3, the CONFOLD folding algorithm and model quality estimations to predict the structure of a protein. We show that the model quality estimation significantly increases the number of models that reliably can be identified. Finally, we apply PconsFold2 to 6379 Pfam families of unknown structure and find that PconsFold2 can, with an estimated 90% specificity, predict the structure of up to 558 Pfam families of unknown structure. Out of these, 415 have not been reported before. Datasets as well as models of all the 558 Pfam families are available at http://c3.pcons.net/ . All programs used here are freely available. arne@bioinfo.se. © The Author 2017. Published by Oxford University Press. All rights reserved. For Permissions, please e-mail: journals.permissions@oup.com
Gunalan, Kabilar; Chaturvedi, Ashutosh; Howell, Bryan; Duchin, Yuval; Lempka, Scott F; Patriat, Remi; Sapiro, Guillermo; Harel, Noam; McIntyre, Cameron C
2017-01-01
Deep brain stimulation (DBS) is an established clinical therapy and computational models have played an important role in advancing the technology. Patient-specific DBS models are now common tools in both academic and industrial research, as well as clinical software systems. However, the exact methodology for creating patient-specific DBS models can vary substantially and important technical details are often missing from published reports. Provide a detailed description of the assembly workflow and parameterization of a patient-specific DBS pathway-activation model (PAM) and predict the response of the hyperdirect pathway to clinical stimulation. Integration of multiple software tools (e.g. COMSOL, MATLAB, FSL, NEURON, Python) enables the creation and visualization of a DBS PAM. An example DBS PAM was developed using 7T magnetic resonance imaging data from a single unilaterally implanted patient with Parkinson's disease (PD). This detailed description implements our best computational practices and most elaborate parameterization steps, as defined from over a decade of technical evolution. Pathway recruitment curves and strength-duration relationships highlight the non-linear response of axons to changes in the DBS parameter settings. Parameterization of patient-specific DBS models can be highly detailed and constrained, thereby providing confidence in the simulation predictions, but at the expense of time demanding technical implementation steps. DBS PAMs represent new tools for investigating possible correlations between brain pathway activation patterns and clinical symptom modulation.
Determination of photovoltaic concentrator optical design specifications using performance modeling
NASA Astrophysics Data System (ADS)
Kerschen, Kevin A.; Levy, Sheldon L.
The strategy used to develop an optical design specification for a 500X concentration photovoltaic module to be used with a 28-percent-efficient concentrator photovoltaic cell is reported. The computer modeling code (PVOPTICS) developed for this purpose, a Fresnel lens design strategy, and optical component specification procedures are described. Comparisons are made between the predicted performance and the measured performance of components fabricated to those specifications. An acrylic lens and a reflective secondary optical element have been tested, showing efficiencies exceeding 88 percent.
NASA Technical Reports Server (NTRS)
Eck, Marshall; Mukunda, Meera
1988-01-01
A calculational method is described which provides a powerful tool for predicting solid rocket motor (SRM) casing and liquid rocket tankage fragmentation response. The approach properly partitions the available impulse to each major system-mass component. It uses the Pisces code developed by Physics International to couple the forces generated by an Eulerian-modeled gas flow field to a Lagrangian-modeled fuel and casing system. The details of the predictive analytical modeling process and the development of normalized relations for momentum partition as a function of SRM burn time and initial geometry are discussed. Methods for applying similar modeling techniques to liquid-tankage-overpressure failures are also discussed. Good agreement between predictions and observations are obtained for five specific events.
Chan, Kuang-Lim; Rosli, Rozana; Tatarinova, Tatiana V; Hogan, Michael; Firdaus-Raih, Mohd; Low, Eng-Ti Leslie
2017-01-27
Gene prediction is one of the most important steps in the genome annotation process. A large number of software tools and pipelines developed by various computing techniques are available for gene prediction. However, these systems have yet to accurately predict all or even most of the protein-coding regions. Furthermore, none of the currently available gene-finders has a universal Hidden Markov Model (HMM) that can perform gene prediction for all organisms equally well in an automatic fashion. We present an automated gene prediction pipeline, Seqping that uses self-training HMM models and transcriptomic data. The pipeline processes the genome and transcriptome sequences of the target species using GlimmerHMM, SNAP, and AUGUSTUS pipelines, followed by MAKER2 program to combine predictions from the three tools in association with the transcriptomic evidence. Seqping generates species-specific HMMs that are able to offer unbiased gene predictions. The pipeline was evaluated using the Oryza sativa and Arabidopsis thaliana genomes. Benchmarking Universal Single-Copy Orthologs (BUSCO) analysis showed that the pipeline was able to identify at least 95% of BUSCO's plantae dataset. Our evaluation shows that Seqping was able to generate better gene predictions compared to three HMM-based programs (MAKER2, GlimmerHMM and AUGUSTUS) using their respective available HMMs. Seqping had the highest accuracy in rice (0.5648 for CDS, 0.4468 for exon, and 0.6695 nucleotide structure) and A. thaliana (0.5808 for CDS, 0.5955 for exon, and 0.8839 nucleotide structure). Seqping provides researchers a seamless pipeline to train species-specific HMMs and predict genes in newly sequenced or less-studied genomes. We conclude that the Seqping pipeline predictions are more accurate than gene predictions using the other three approaches with the default or available HMMs.
Hypoglycemia prediction with subject-specific recursive time-series models.
Eren-Oruklu, Meriyan; Cinar, Ali; Quinn, Lauretta
2010-01-01
Avoiding hypoglycemia while keeping glucose within the narrow normoglycemic range (70-120 mg/dl) is a major challenge for patients with type 1 diabetes. Continuous glucose monitors can provide hypoglycemic alarms when the measured glucose decreases below a threshold. However, a better approach is to provide an early alarm that predicts a hypoglycemic episode before it occurs, allowing enough time for the patient to take the necessary precaution to avoid hypoglycemia. We have previously proposed subject-specific recursive models for the prediction of future glucose concentrations and evaluated their prediction performance. In this work, our objective was to evaluate this algorithm further to predict hypoglycemia and provide early hypoglycemic alarms. Three different methods were proposed for alarm decision, where (A) absolute predicted glucose values, (B) cumulative-sum (CUSUM) control chart, and (C) exponentially weighted moving-average (EWMA) control chart were used. Each method was validated using data from the Diabetes Research in Children Network, which consist of measurements from a continuous glucose sensor during an insulin-induced hypoglycemia. Reference serum glucose measurements were used to determine the sensitivity to predict hypoglycemia and the false alarm rate. With the hypoglycemic threshold set to 60 mg/dl, sensitivity of 89, 87.5, and 89% and specificity of 67, 74, and 78% were reported for methods A, B, and C, respectively. Mean values for time to detection were 30 +/- 5.51 (A), 25.8 +/- 6.46 (B), and 27.7 +/- 5.32 (C) minutes. Compared to the absolute value method, both CUSUM and EWMA methods behaved more conservatively before raising an alarm (reduced time to detection), which significantly decreased the false alarm rate and increased the specificity. 2010 Diabetes Technology Society.
Ahadi, Alireza; Sablok, Gaurav; Hutvagner, Gyorgy
2017-04-07
MicroRNAs (miRNAs) are ∼19-22 nucleotides (nt) long regulatory RNAs that regulate gene expression by recognizing and binding to complementary sequences on mRNAs. The key step in revealing the function of a miRNA, is the identification of miRNA target genes. Recent biochemical advances including PAR-CLIP and HITS-CLIP allow for improved miRNA target predictions and are widely used to validate miRNA targets. Here, we present miRTar2GO, which is a model, trained on the common rules of miRNA-target interactions, Argonaute (Ago) CLIP-Seq data and experimentally validated miRNA target interactions. miRTar2GO is designed to predict miRNA target sites using more relaxed miRNA-target binding characteristics. More importantly, miRTar2GO allows for the prediction of cell-type specific miRNA targets. We have evaluated miRTar2GO against other widely used miRNA target prediction algorithms and demonstrated that miRTar2GO produced significantly higher F1 and G scores. Target predictions, binding specifications, results of the pathway analysis and gene ontology enrichment of miRNA targets are freely available at http://www.mirtar2go.org. © The Author(s) 2016. Published by Oxford University Press on behalf of Nucleic Acids Research.
Bejaei, M; Wiseman, K; Cheng, K M
2015-01-01
Consumers' interest in specialty eggs appears to be growing in Europe and North America. The objective of this research was to develop logistic regression models that utilise purchaser attributes and demographics to predict the probability of a consumer purchasing a specific type of table egg including regular (white and brown), non-caged (free-run, free-range and organic) or nutrient-enhanced eggs. These purchase prediction models, together with the purchasers' attributes, can be used to assess market opportunities of different egg types specifically in British Columbia (BC). An online survey was used to gather data for the models. A total of 702 completed questionnaires were submitted by BC residents. Selected independent variables included in the logistic regression to develop models for different egg types to predict the probability of a consumer purchasing a specific type of table egg. The variables used in the model accounted for 54% and 49% of variances in the purchase of regular and non-caged eggs, respectively. Research results indicate that consumers of different egg types exhibit a set of unique and statistically significant characteristics and/or demographics. For example, consumers of regular eggs were less educated, older, price sensitive, major chain store buyers, and store flyer users, and had lower awareness about different types of eggs and less concern regarding animal welfare issues. However, most of the non-caged egg consumers were less concerned about price, had higher awareness about different types of table eggs, purchased their eggs from local/organic grocery stores, farm gates or farmers markets, and they were more concerned about care and feeding of hens compared to consumers of other eggs types.
Circulating tumour DNA methylation markers for diagnosis and prognosis of hepatocellular carcinoma
NASA Astrophysics Data System (ADS)
Xu, Rui-Hua; Wei, Wei; Krawczyk, Michal; Wang, Wenqiu; Luo, Huiyan; Flagg, Ken; Yi, Shaohua; Shi, William; Quan, Qingli; Li, Kang; Zheng, Lianghong; Zhang, Heng; Caughey, Bennett A.; Zhao, Qi; Hou, Jiayi; Zhang, Runze; Xu, Yanxin; Cai, Huimin; Li, Gen; Hou, Rui; Zhong, Zheng; Lin, Danni; Fu, Xin; Zhu, Jie; Duan, Yaou; Yu, Meixing; Ying, Binwu; Zhang, Wengeng; Wang, Juan; Zhang, Edward; Zhang, Charlotte; Li, Oulan; Guo, Rongping; Carter, Hannah; Zhu, Jian-Kang; Hao, Xiaoke; Zhang, Kang
2017-11-01
An effective blood-based method for the diagnosis and prognosis of hepatocellular carcinoma (HCC) has not yet been developed. Circulating tumour DNA (ctDNA) carrying cancer-specific genetic and epigenetic aberrations may enable a noninvasive `liquid biopsy' for diagnosis and monitoring of cancer. Here, we identified an HCC-specific methylation marker panel by comparing HCC tissue and normal blood leukocytes and showed that methylation profiles of HCC tumour DNA and matched plasma ctDNA are highly correlated. Using cfDNA samples from a large cohort of 1,098 HCC patients and 835 normal controls, we constructed a diagnostic prediction model that showed high diagnostic specificity and sensitivity (P < 0.001) and was highly correlated with tumour burden, treatment response, and stage. Additionally, we constructed a prognostic prediction model that effectively predicted prognosis and survival (P < 0.001). Together, these findings demonstrate in a large clinical cohort the utility of ctDNA methylation markers in the diagnosis, surveillance, and prognosis of HCC.
A microRNA-based prediction model for lymph node metastasis in hepatocellular carcinoma.
Zhang, Li; Xiang, Zuo-Lin; Zeng, Zhao-Chong; Fan, Jia; Tang, Zhao-You; Zhao, Xiao-Mei
2016-01-19
We developed an efficient microRNA (miRNA) model that could predict the risk of lymph node metastasis (LNM) in hepatocellular carcinoma (HCC). We first evaluated a training cohort of 192 HCC patients after hepatectomy and found five LNM associated predictive factors: vascular invasion, Barcelona Clinic Liver Cancer stage, miR-145, miR-31, and miR-92a. The five statistically independent factors were used to develop a predictive model. The predictive value of the miRNA-based model was confirmed in a validation cohort of 209 consecutive HCC patients. The prediction model was scored for LNM risk from 0 to 8. The cutoff value 4 was used to distinguish high-risk and low-risk groups. The model sensitivity and specificity was 69.6 and 80.2%, respectively, during 5 years in the validation cohort. And the area under the curve (AUC) for the miRNA-based prognostic model was 0.860. The 5-year positive and negative predictive values of the model in the validation cohort were 30.3 and 95.5%, respectively. Cox regression analysis revealed that the LNM hazard ratio of the high-risk versus low-risk groups was 11.751 (95% CI, 5.110-27.021; P < 0.001) in the validation cohort. In conclusion, the miRNA-based model is reliable and accurate for the early prediction of LNM in patients with HCC.
Dewey, Daniel; Schuldberg, David; Madathil, Renee
2014-08-01
This study investigated whether specific peritraumatic emotions differentially predict PTSD symptom clusters in individuals who have experienced stressful life events. Hypotheses were developed based on the SPAARS model of PTSD. It was predicted that the peritraumatic emotions of anger, disgust, guilt, and fear would significantly predict re-experiencing and avoidance symptoms, while only fear would predict hyperarousal. Undergraduate students (N = 144) participated in this study by completing a packet of self-report questionnaires. Multiple regression analyses were conducted with PCL-S symptom cluster scores as dependent variables and peritraumatic fear, guilt, anger, shame, and disgust as predictor variables. As hypothesized, peritraumatic anger, guilt, and fear all significantly predicted re-experiencing. However, only fear predicted avoidance, and anger significantly predicted hyperarousal. Results are discussed in relation to the theoretical role of emotions in the etiology of PTSD following the experience of a stressful life event.
Tropical forecasting - Predictability perspective
NASA Technical Reports Server (NTRS)
Shukla, J.
1989-01-01
Results are presented of classical predictability studies and forecast experiments with observed initial conditions to show the nature of initial error growth and final error equilibration for the tropics and midlatitudes, separately. It is found that the theoretical upper limit of tropical circulation predictability is far less than for midlatitudes. The error growth for a complete general circulation model is compared to a dry version of the same model in which there is no prognostic equation for moisture, and diabatic heat sources are prescribed. It is found that the growth rate of synoptic-scale errors for the dry model is significantly smaller than for the moist model, suggesting that the interactions between dynamics and moist processes are among the important causes of atmospheric flow predictability degradation. Results are then presented of numerical experiments showing that correct specification of the slowly varying boundary condition of SST produces significant improvement in the prediction of time-averaged circulation and rainfall over the tropics.
Real-time assessments of water quality: expanding nowcasting throughout the Great Lakes
,
2013-01-01
Nowcasts are systems that inform the public of current bacterial water-quality conditions at beaches on the basis of predictive models. During 2010–12, the U.S. Geological Survey (USGS) worked with 23 local and State agencies to improve existing operational beach nowcast systems at 4 beaches and expand the use of predictive models in nowcasts at an additional 45 beaches throughout the Great Lakes. The predictive models were specific to each beach, and the best model for each beach was based on a unique combination of environmental and water-quality explanatory variables. The variables used most often in models to predict Escherichia coli (E. coli) concentrations or the probability of exceeding a State recreational water-quality standard included turbidity, day of the year, wave height, wind direction and speed, antecedent rainfall for various time periods, and change in lake level over 24 hours. During validation of 42 beach models during 2012, the models performed better than the current method to assess recreational water quality (previous day's E. coli concentration). The USGS will continue to work with local agencies to improve nowcast predictions, enable technology transfer of predictive model development procedures, and implement more operational systems during 2013 and beyond.
20171015 - Predicting Exposure Pathways with Machine Learning (ISES)
Prioritizing the risk posed to human health from the thousands of chemicals in the environment requires tools that can estimate exposure rates from limited information. High throughput models exist to make predictions of exposure via specific, important pathways such as residenti...
Kong, Ru; Li, Jingwei; Orban, Csaba; Sabuncu, Mert R; Liu, Hesheng; Schaefer, Alexander; Sun, Nanbo; Zuo, Xi-Nian; Holmes, Avram J; Eickhoff, Simon B; Yeo, B T Thomas
2018-06-06
Resting-state functional magnetic resonance imaging (rs-fMRI) offers the opportunity to delineate individual-specific brain networks. A major question is whether individual-specific network topography (i.e., location and spatial arrangement) is behaviorally relevant. Here, we propose a multi-session hierarchical Bayesian model (MS-HBM) for estimating individual-specific cortical networks and investigate whether individual-specific network topography can predict human behavior. The multiple layers of the MS-HBM explicitly differentiate intra-subject (within-subject) from inter-subject (between-subject) network variability. By ignoring intra-subject variability, previous network mappings might confuse intra-subject variability for inter-subject differences. Compared with other approaches, MS-HBM parcellations generalized better to new rs-fMRI and task-fMRI data from the same subjects. More specifically, MS-HBM parcellations estimated from a single rs-fMRI session (10 min) showed comparable generalizability as parcellations estimated by 2 state-of-the-art methods using 5 sessions (50 min). We also showed that behavioral phenotypes across cognition, personality, and emotion could be predicted by individual-specific network topography with modest accuracy, comparable to previous reports predicting phenotypes based on connectivity strength. Network topography estimated by MS-HBM was more effective for behavioral prediction than network size, as well as network topography estimated by other parcellation approaches. Thus, similar to connectivity strength, individual-specific network topography might also serve as a fingerprint of human behavior.
Weather Research and Forecasting Model Sensitivity Comparisons for Warm Season Convective Initiation
NASA Technical Reports Server (NTRS)
Watson, Leela R.
2007-01-01
This report describes the work done by the Applied Meteorology Unit (AMU) in assessing the success of different model configurations in predicting warm season convection over East-Central Florida. The Weather Research and Forecasting Environmental Modeling System (WRF EMS) software allows users to choose among two dynamical cores - the Advanced Research WRF (ARW) and the Non-hydrostatic Mesoscale Model (NMM). There are also data assimilation analysis packages available for the initialization of the WRF model - the Local Analysis and Prediction System (LAPS) and the Advanced Regional Prediction System (ARPS) Data Analysis System (ADAS). Besides model core and initialization options, the WRF model can be run with one- or two-way nesting. Having a series of initialization options and WRF cores, as well as many options within each core, creates challenges for local forecasters, such as determining which configuration options are best to address specific forecast concerns. This project assessed three different model intializations available to determine which configuration best predicts warm season convective initiation in East-Central Florida. The project also examined the use of one- and two-way nesting in predicting warm season convection.
Longitudinal histories as predictors of future diagnoses of domestic abuse: modelling study
Kohane, Isaac S; Mandl, Kenneth D
2009-01-01
Objective To determine whether longitudinal data in patients’ historical records, commonly available in electronic health record systems, can be used to predict a patient’s future risk of receiving a diagnosis of domestic abuse. Design Bayesian models, known as intelligent histories, used to predict a patient’s risk of receiving a future diagnosis of abuse, based on the patient’s diagnostic history. Retrospective evaluation of the model’s predictions using an independent testing set. Setting A state-wide claims database covering six years of inpatient admissions to hospital, admissions for observation, and encounters in emergency departments. Population All patients aged over 18 who had at least four years between their earliest and latest visits recorded in the database (561 216 patients). Main outcome measures Timeliness of detection, sensitivity, specificity, positive predictive values, and area under the ROC curve. Results 1.04% (5829) of the patients met the narrow case definition for abuse, while 3.44% (19 303) met the broader case definition for abuse. The model achieved sensitive, specific (area under the ROC curve of 0.88), and early (10-30 months in advance, on average) prediction of patients’ future risk of receiving a diagnosis of abuse. Analysis of model parameters showed important differences between sexes in the risks associated with certain diagnoses. Conclusions Commonly available longitudinal diagnostic data can be useful for predicting a patient’s future risk of receiving a diagnosis of abuse. This modelling approach could serve as the basis for an early warning system to help doctors identify high risk patients for further screening. PMID:19789406
Li, John; Maclehose, Rich; Smith, Kirk; Kaehler, Dawn; Hedberg, Craig
2011-01-01
Foodborne illness surveillance based on consumer complaints detects outbreaks by finding common exposures among callers, but this process is often difficult. Laboratory testing of ill callers could also help identify potential outbreaks. However, collection of stool samples from all callers is not feasible. Methods to help screen calls for etiology are needed to increase the efficiency of complaint surveillance systems and increase the likelihood of detecting foodborne outbreaks caused by Salmonella. Data from the Minnesota Department of Health foodborne illness surveillance database (2000 to 2008) were analyzed. Complaints with identified etiologies were examined to create a predictive model for Salmonella. Bootstrap methods were used to internally validate the model. Seventy-one percent of complaints in the foodborne illness database with known etiologies were due to norovirus. The predictive model had a good discriminatory ability to identify Salmonella calls. Three cutoffs for the predictive model were tested: one that maximized sensitivity, one that maximized specificity, and one that maximized predictive ability, providing sensitivities and specificities of 32 and 96%, 100 and 54%, and 89 and 72%, respectively. Development of a predictive model for Salmonella could help screen calls for etiology. The cutoff that provided the best predictive ability for Salmonella corresponded to a caller reporting diarrhea and fever with no vomiting, and five or fewer people ill. Screening calls for etiology would help identify complaints for further follow-up and result in identifying Salmonella cases that would otherwise go unconfirmed; in turn, this could lead to the identification of more outbreaks.
Specific Coping Behaviors in Relation to Adolescent Depression and Suicidal Ideation
Horwitz, Adam G.; Hill, Ryan M.; King, Cheryl A.
2010-01-01
The coping strategies used by adolescents to deal with stress may have implications for the development of depression and suicidal ideation. This study examined coping categories and specific coping behaviors used by adolescents to assess the relation of coping to depression and suicidal ideation. In hierarchical regression models, the specific coping behaviors of behavioral disengagement and self-blame were predictive of higher levels of depression; depression and using emotional support were predictive of suicidal ideation. Results suggest that specific behaviors within the broad coping categories of emotion-focused coping (e.g., self-blame) and avoidant coping (e.g., behavioral disengagement) account for these categories’ associations with depression and suicidal ideation. Specific problem-focused coping strategies did not independently predict lower levels of depression or suicidal ideation. It may be beneficial for interventions to focus on eliminating maladaptive coping behaviors in addition to introducing or enhancing positive coping behaviors. PMID:21074841
Webb, Samuel J; Hanser, Thierry; Howlin, Brendan; Krause, Paul; Vessey, Jonathan D
2014-03-25
A new algorithm has been developed to enable the interpretation of black box models. The developed algorithm is agnostic to learning algorithm and open to all structural based descriptors such as fragments, keys and hashed fingerprints. The algorithm has provided meaningful interpretation of Ames mutagenicity predictions from both random forest and support vector machine models built on a variety of structural fingerprints.A fragmentation algorithm is utilised to investigate the model's behaviour on specific substructures present in the query. An output is formulated summarising causes of activation and deactivation. The algorithm is able to identify multiple causes of activation or deactivation in addition to identifying localised deactivations where the prediction for the query is active overall. No loss in performance is seen as there is no change in the prediction; the interpretation is produced directly on the model's behaviour for the specific query. Models have been built using multiple learning algorithms including support vector machine and random forest. The models were built on public Ames mutagenicity data and a variety of fingerprint descriptors were used. These models produced a good performance in both internal and external validation with accuracies around 82%. The models were used to evaluate the interpretation algorithm. Interpretation was revealed that links closely with understood mechanisms for Ames mutagenicity. This methodology allows for a greater utilisation of the predictions made by black box models and can expedite further study based on the output for a (quantitative) structure activity model. Additionally the algorithm could be utilised for chemical dataset investigation and knowledge extraction/human SAR development.
Cook, Benjamin L; Progovac, Ana M; Chen, Pei; Mullin, Brian; Hou, Sherry; Baca-Garcia, Enrique
2016-01-01
Natural language processing (NLP) and machine learning were used to predict suicidal ideation and heightened psychiatric symptoms among adults recently discharged from psychiatric inpatient or emergency room settings in Madrid, Spain. Participants responded to structured mental and physical health instruments at multiple follow-up points. Outcome variables of interest were suicidal ideation and psychiatric symptoms (GHQ-12). Predictor variables included structured items (e.g., relating to sleep and well-being) and responses to one unstructured question, "how do you feel today?" We compared NLP-based models using the unstructured question with logistic regression prediction models using structured data. The PPV, sensitivity, and specificity for NLP-based models of suicidal ideation were 0.61, 0.56, and 0.57, respectively, compared to 0.73, 0.76, and 0.62 of structured data-based models. The PPV, sensitivity, and specificity for NLP-based models of heightened psychiatric symptoms (GHQ-12 ≥ 4) were 0.56, 0.59, and 0.60, respectively, compared to 0.79, 0.79, and 0.85 in structured models. NLP-based models were able to generate relatively high predictive values based solely on responses to a simple general mood question. These models have promise for rapidly identifying persons at risk of suicide or psychological distress and could provide a low-cost screening alternative in settings where lengthy structured item surveys are not feasible.
The Study of Rain Specific Attenuation for the Prediction of Satellite Propagation in Malaysia
NASA Astrophysics Data System (ADS)
Mandeep, J. S.; Ng, Y. Y.; Abdullah, H.; Abdullah, M.
2010-06-01
Specific attenuation is the fundamental quantity in the calculation of rain attenuation for terrestrial path and slant paths representing as rain attenuation per unit distance (dB/km). Specific attenuation is an important element in developing the predicted rain attenuation model. This paper deals with the empirical determination of the power law coefficients which allow calculating the specific attenuation in dB/km from the knowledge of the rain rate in mm/h. The main purpose of the paper is to obtain the coefficients of k and α of power law relationship between specific attenuation. Three years (from 1st January 2006 until 31st December 2008) rain gauge and beacon data taken from USM, Nibong Tebal have been used to do the empirical procedure analysis of rain specific attenuation. The data presented are semi-empirical in nature. A year-to-year variation of the coefficients has been indicated and the empirical measured data was compared with ITU-R provided regression coefficient. The result indicated that the USM empirical measured data was significantly vary from ITU-R predicted value. Hence, ITU-R recommendation for regression coefficients of rain specific attenuation is not suitable for predicting rain attenuation at Malaysia.
Distribution drivers and physiological responses in geothermal bryophyte communities.
García, Estefanía Llaneza; Rosenstiel, Todd N; Graves, Camille; Shortlidge, Erin E; Eppley, Sarah M
2016-04-01
Our ability to explain community structure rests on our ability to define the importance of ecological niches, including realized ecological niches, in shaping communities, but few studies of plant distributions have combined predictive models with physiological measures. Using field surveys and statistical modeling, we predicted distribution drivers in geothermal bryophyte (moss) communities of Lassen Volcanic National Park (California, USA). In the laboratory, we used drying and rewetting experiments to test whether the strong species-specific effects of relative humidity on distributions predicted by the models were correlated with physiological characters. We found that the three most common bryophytes in geothermal communities were significantly affected by three distinct distribution drivers: temperature, light, and relative humidity. Aulacomnium palustre, whose distribution is significantly affected by relative humidity according to our model, and which occurs in high-humidity sites, showed extreme signs of stress after drying and never recovered optimal values of PSII efficiency after rewetting. Campylopus introflexus, whose distribution is not affected by humidity according to our model, was able to maintain optimal values of PSII efficiency for 48 hr at 50% water loss and recovered optimal values of PSII efficiency after rewetting. Our results suggest that species-specific environmental stressors tightly constrain the ecological niches of geothermal bryophytes. Tests of tolerance to drying in two bryophyte species corresponded with model predictions of the comparative importance of relative humidity as distribution drivers for these species. © 2016 Botanical Society of America.
Gao, Chao; Sun, Hanbo; Wang, Tuo; Tang, Ming; Bohnen, Nicolaas I; Müller, Martijn L T M; Herman, Talia; Giladi, Nir; Kalinin, Alexandr; Spino, Cathie; Dauer, William; Hausdorff, Jeffrey M; Dinov, Ivo D
2018-05-08
In this study, we apply a multidisciplinary approach to investigate falls in PD patients using clinical, demographic and neuroimaging data from two independent initiatives (University of Michigan and Tel Aviv Sourasky Medical Center). Using machine learning techniques, we construct predictive models to discriminate fallers and non-fallers. Through controlled feature selection, we identified the most salient predictors of patient falls including gait speed, Hoehn and Yahr stage, postural instability and gait difficulty-related measurements. The model-based and model-free analytical methods we employed included logistic regression, random forests, support vector machines, and XGboost. The reliability of the forecasts was assessed by internal statistical (5-fold) cross validation as well as by external out-of-bag validation. Four specific challenges were addressed in the study: Challenge 1, develop a protocol for harmonizing and aggregating complex, multisource, and multi-site Parkinson's disease data; Challenge 2, identify salient predictive features associated with specific clinical traits, e.g., patient falls; Challenge 3, forecast patient falls and evaluate the classification performance; and Challenge 4, predict tremor dominance (TD) vs. posture instability and gait difficulty (PIGD). Our findings suggest that, compared to other approaches, model-free machine learning based techniques provide a more reliable clinical outcome forecasting of falls in Parkinson's patients, for example, with a classification accuracy of about 70-80%.
Are prediction models for Lynch syndrome valid for probands with endometrial cancer?
Backes, Floor J; Hampel, Heather; Backes, Katherine A; Vaccarello, Luis; Lewandowski, George; Bell, Jeffrey A; Reid, Gary C; Copeland, Larry J; Fowler, Jeffrey M; Cohn, David E
2009-01-01
Currently, three prediction models are used to predict a patient's risk of having Lynch syndrome (LS). These models have been validated in probands with colorectal cancer (CRC), but not in probands presenting with endometrial cancer (EMC). Thus, the aim was to determine the performance of these prediction models in women with LS presenting with EMC. Probands with EMC and LS were identified. Personal and family history was entered into three prediction models, PREMM(1,2), MMRpro, and MMRpredict. Probabilities of mutations in the mismatch repair genes were recorded. Accurate prediction was defined as a model predicting at least a 5% chance of a proband carrying a mutation. From 562 patients prospectively enrolled in a clinical trial of patients with EMC, 13 (2.2%) were shown to have LS. Nine patients had a mutation in MSH6, three in MSH2, and one in MLH1. MMRpro predicted that 3 of 9 patients with an MSH6, 3 of 3 with an MSH2, and 1 of 1 patient with an MLH1 mutation could have LS. For MMRpredict, EMC coded as "proximal CRC" predicted 5 of 5, and as "distal CRC" three of five. PREMM(1,2) predicted that 4 of 4 with an MLH1 or MSH2 could have LS. Prediction of LS in probands presenting with EMC using current models for probands with CRC works reasonably well. Further studies are needed to develop models that include questions specific to patients with EMC with a greater age range, as well as placing increased emphasis on prediction of LS in probands with MSH6 mutations.