Single-case synthesis tools I: Comparing tools to evaluate SCD quality and rigor.
Zimmerman, Kathleen N; Ledford, Jennifer R; Severini, Katherine E; Pustejovsky, James E; Barton, Erin E; Lloyd, Blair P
2018-03-03
Tools for evaluating the quality and rigor of single case research designs (SCD) are often used when conducting SCD syntheses. Preferred components include evaluations of design features related to the internal validity of SCD to obtain quality and/or rigor ratings. Three tools for evaluating the quality and rigor of SCD (Council for Exceptional Children, What Works Clearinghouse, and Single-Case Analysis and Design Framework) were compared to determine if conclusions regarding the effectiveness of antecedent sensory-based interventions for young children changed based on choice of quality evaluation tool. Evaluation of SCD quality differed across tools, suggesting selection of quality evaluation tools impacts evaluation findings. Suggestions for selecting an appropriate quality and rigor assessment tool are provided and across-tool conclusions are drawn regarding the quality and rigor of studies. Finally, authors provide guidance for using quality evaluations in conjunction with outcome analyses when conducting syntheses of interventions evaluated in the context of SCD. Copyright © 2018 Elsevier Ltd. All rights reserved.
SMAP validation of soil moisture products
USDA-ARS?s Scientific Manuscript database
The Soil Moisture Active Passive (SMAP) satellite will be launched by the National Aeronautics and Space Administration in October 2014. SMAP will also incorporate a rigorous calibration and validation program that will support algorithm refinement and provide users with information on the accuracy ...
Peer Review of EPA's Draft BMDS Document: Exponential ...
BMDS is one of the Agency's premier tools for estimating risk assessments, therefore the validity and reliability of its statistical models are of paramount importance. This page provides links to peer review of the BMDS applications and its models as they were developed and eventually released documenting the rigorous review process taken to provide the best science tools available for statistical modeling. This page provides links to peer review of the BMDS applications and its models as they were developed and eventually released documenting the rigorous review process taken to provide the best science tools available for statistical modeling.
DESCQA: Synthetic Sky Catalog Validation Framework
NASA Astrophysics Data System (ADS)
Mao, Yao-Yuan; Uram, Thomas D.; Zhou, Rongpu; Kovacs, Eve; Ricker, Paul M.; Kalmbach, J. Bryce; Padilla, Nelson; Lanusse, François; Zu, Ying; Tenneti, Ananth; Vikraman, Vinu; DeRose, Joseph
2018-04-01
The DESCQA framework provides rigorous validation protocols for assessing the quality of high-quality simulated sky catalogs in a straightforward and comprehensive way. DESCQA enables the inspection, validation, and comparison of an inhomogeneous set of synthetic catalogs via the provision of a common interface within an automated framework. An interactive web interface is also available at portal.nersc.gov/project/lsst/descqa.
Cypress, Brigitte S
Issues are still raised even now in the 21st century by the persistent concern with achieving rigor in qualitative research. There is also a continuing debate about the analogous terms reliability and validity in naturalistic inquiries as opposed to quantitative investigations. This article presents the concept of rigor in qualitative research using a phenomenological study as an exemplar to further illustrate the process. Elaborating on epistemological and theoretical conceptualizations by Lincoln and Guba, strategies congruent with qualitative perspective for ensuring validity to establish the credibility of the study are described. A synthesis of the historical development of validity criteria evident in the literature during the years is explored. Recommendations are made for use of the term rigor instead of trustworthiness and the reconceptualization and renewed use of the concept of reliability and validity in qualitative research, that strategies for ensuring rigor must be built into the qualitative research process rather than evaluated only after the inquiry, and that qualitative researchers and students alike must be proactive and take responsibility in ensuring the rigor of a research study. The insights garnered here will move novice researchers and doctoral students to a better conceptual grasp of the complexity of reliability and validity and its ramifications for qualitative inquiry.
Learning from Science and Sport - How we, Safety, "Engage with Rigor"
NASA Astrophysics Data System (ADS)
Herd, A.
2012-01-01
As the world of spaceflight safety is relatively small and potentially inward-looking, we need to be aware of the "outside world". We should then try to remind ourselves to be open to the possibility that data, knowledge or experience from outside of the spaceflight community may provide some constructive alternate perspectives. This paper will assess aspects from two seemingly tangential fields, science and sport, and align these with the world of safety. In doing so some useful insights will be given to the challenges we face and may provide solutions relevant in our everyday (of safety engineering). Sport, particularly a contact sport such as rugby union, requires direct interaction between members of two (opposing) teams. Professional, accurately timed and positioned interaction for a desired outcome. These interactions, whilst an essential part of the game, are however not without their constraints. The rugby scrum has constraints as to the formation and engagement of the two teams. The controlled engagement provides for an interaction between the two teams in a safe manner. The constraints arising from the reality that an incorrect engagement could cause serious injury to members of either team. In academia, scientific rigor is applied to assure that the arguments provided and the conclusions drawn in academic papers presented for publication are valid, legitimate and credible. The scientific goal of the need for rigor may be expressed in the example of achieving a statistically relevant sample size, n, in order to assure analysis validity of the data pool. A failure to apply rigor could then place the entire study at risk of failing to have the respective paper published. This paper will consider the merits of these two different aspects, scientific rigor and sports engagement, and offer a reflective look at how this may provide a "modus operandi" for safety engineers at any level whether at their desks (creating or reviewing safety assessments) or in a safety review meeting (providing a verbal critique of the presented safety case).
Walach, Harald; Falkenberg, Torkel; Fønnebø, Vinjar; Lewith, George; Jonas, Wayne B
2006-01-01
Background The reasoning behind evaluating medical interventions is that a hierarchy of methods exists which successively produce improved and therefore more rigorous evidence based medicine upon which to make clinical decisions. At the foundation of this hierarchy are case studies, retrospective and prospective case series, followed by cohort studies with historical and concomitant non-randomized controls. Open-label randomized controlled studies (RCTs), and finally blinded, placebo-controlled RCTs, which offer most internal validity are considered the most reliable evidence. Rigorous RCTs remove bias. Evidence from RCTs forms the basis of meta-analyses and systematic reviews. This hierarchy, founded on a pharmacological model of therapy, is generalized to other interventions which may be complex and non-pharmacological (healing, acupuncture and surgery). Discussion The hierarchical model is valid for limited questions of efficacy, for instance for regulatory purposes and newly devised products and pharmacological preparations. It is inadequate for the evaluation of complex interventions such as physiotherapy, surgery and complementary and alternative medicine (CAM). This has to do with the essential tension between internal validity (rigor and the removal of bias) and external validity (generalizability). Summary Instead of an Evidence Hierarchy, we propose a Circular Model. This would imply a multiplicity of methods, using different designs, counterbalancing their individual strengths and weaknesses to arrive at pragmatic but equally rigorous evidence which would provide significant assistance in clinical and health systems innovation. Such evidence would better inform national health care technology assessment agencies and promote evidence based health reform. PMID:16796762
THE USE OF RESEARCH RESULTS IN TEACHING SOCIAL WORK PRACTICE.
ERIC Educational Resources Information Center
LAWRENCE, RICHARD G.
BECAUSE THE SUCCESS OF INTERVENTION DEPENDS UPON THE VALIDITY OF THE PROPOSITIONS EMPLOYED, AND BECAUSE SCIENTIFIC RESEARCH ASSURES VALIDITY BY PROVIDING THE MOST SYSTEMATIC AND RIGOROUS ATTENTION TO PROBLEMS, THE UTILIZATION OF RESEARCH IS IMPORTANT TO SOCIAL WORK PRACTICE. SEVERAL FACTORS LIMIT ITS USE--(1) ALTHOUGH CONCEPTS ARE CLEARLY DEFINED…
Validating the Use of AP® Exam Scores for College Course Placement. Research Report 2013-2
ERIC Educational Resources Information Center
Patterson, Brian F.; Ewing, Maureen
2013-01-01
The Advanced Placement Program® (AP®) was created to provide access to rigorous, college-level curricula to motivated and prepared high school students. This study evaluated whether the AP Exam scores from the summative exams associated with 10 courses were valid for the placement of students into higher-level college courses in the subject area…
Banna, Jinan C; Vera Becerra, Luz E; Kaiser, Lucia L; Townsend, Marilyn S
2010-01-01
Development of outcome measures relevant to health nutrition behaviors requires a rigorous process of testing and revision. Whereas researchers often report performance of quantitative data collection to assess questionnaire validity and reliability, qualitative testing procedures are often overlooked. This report outlines a procedure for assessing face validity of a Spanish-language dietary assessment tool. Reviewing the literature produced no rigorously validated Spanish-language food behavior assessment tools for the US Department of Agriculture's food assistance and education programs. In response to this need, this study evaluated the face validity of a Spanish-language food behavior checklist adapted from a 16-item English version of a food behavior checklist shown to be valid and reliable for limited-resource English speakers. The English version was translated using rigorous methods involving initial translation by one party and creation of five possible versions. Photos were modified based on client input and new photos were taken as necessary. A sample of low-income, Spanish-speaking women completed cognitive interviews (n=20). Spanish translation experts (n=7) fluent in both languages and familiar with both cultures made minor modifications but essentially approved client preferences. The resulting checklist generated a readability score of 93, indicating low reading difficulty. The Spanish-language checklist has adequate face validity in the target population and is ready for further validation using convergent measures. At the conclusion of testing, this instrument may be used to evaluate nutrition education interventions in California. These qualitative procedures provide a framework for designing evaluation tools for low-literate audiences participating in the US Department of Agriculture food assistance and education programs. Copyright 2010 American Dietetic Association. Published by Elsevier Inc. All rights reserved.
BANNA, JINAN C.; VERA BECERRA, LUZ E.; KAISER, LUCIA L.; TOWNSEND, MARILYN S.
2015-01-01
Development of outcome measures relevant to health nutrition behaviors requires a rigorous process of testing and revision. Whereas researchers often report performance of quantitative data collection to assess questionnaire validity and reliability, qualitative testing procedures are often overlooked. This report outlines a procedure for assessing face validity of a Spanish-language dietary assessment tool. Reviewing the literature produced no rigorously validated Spanish-language food behavior assessment tools for the US Department of Agriculture’s food assistance and education programs. In response to this need, this study evaluated the face validity of a Spanish-language food behavior checklist adapted from a 16-item English version of a food behavior checklist shown to be valid and reliable for limited-resource English speakers. The English version was translated using rigorous methods involving initial translation by one party and creation of five possible versions. Photos were modified based on client input and new photos were taken as necessary. A sample of low-income, Spanish-speaking women completed cognitive interviews (n=20). Spanish translation experts (n=7) fluent in both languages and familiar with both cultures made minor modifications but essentially approved client preferences. The resulting checklist generated a readability score of 93, indicating low reading difficulty. The Spanish-language checklist has adequate face validity in the target population and is ready for further validation using convergent measures. At the conclusion of testing, this instrument may be used to evaluate nutrition education interventions in California. These qualitative procedures provide a framework for designing evaluation tools for low-literate audiences participating in the US Department of Agriculture food assistance and education programs. PMID:20102831
ERIC Educational Resources Information Center
Geiser, Saul; Santelices, Maria Veronica
2007-01-01
High-school grades are often viewed as an unreliable criterion for college admissions, owing to differences in grading standards across high schools, while standardized tests are seen as methodologically rigorous, providing a more uniform and valid yardstick for assessing student ability and achievement. The present study challenges that…
Item-Based Psychometrics of the Preschool Behavioral and Emotional Rating Scale
ERIC Educational Resources Information Center
Cress, Cynthia J.; Lambert, Matthew C.; Epstein, Michael H.
2014-01-01
The Preschool Behavioral and Emotional Rating Scale (PreBERS) is an assessment of emotional and behavioral strengths in preschoolers with well-established reliability and validity for educational and clinical application in children with and without disabilities. The present study provides further evidence of psychometric rigor for items and…
Federal Register 2010, 2011, 2012, 2013, 2014
2010-03-22
... on rigorous scientifically based research methods to assess the effectiveness of a particular... activities and programs; and (B) Includes research that-- (i) Employs systematic, empirical methods that draw... or observational methods that provide reliable and valid data across evaluators and observers, across...
Federal Register 2010, 2011, 2012, 2013, 2014
2010-01-15
... that is based on rigorous scientifically based research methods to assess the effectiveness of a...) Relies on measurements or observational methods that provide reliable and valid data across evaluators... of innovative, cohesive models that are based on research and have demonstrated that they effectively...
When Assessment Data Are Words: Validity Evidence for Qualitative Educational Assessments.
Cook, David A; Kuper, Ayelet; Hatala, Rose; Ginsburg, Shiphra
2016-10-01
Quantitative scores fail to capture all important features of learner performance. This awareness has led to increased use of qualitative data when assessing health professionals. Yet the use of qualitative assessments is hampered by incomplete understanding of their role in forming judgments, and lack of consensus in how to appraise the rigor of judgments therein derived. The authors articulate the role of qualitative assessment as part of a comprehensive program of assessment, and translate the concept of validity to apply to judgments arising from qualitative assessments. They first identify standards for rigor in qualitative research, and then use two contemporary assessment validity frameworks to reorganize these standards for application to qualitative assessment.Standards for rigor in qualitative research include responsiveness, reflexivity, purposive sampling, thick description, triangulation, transparency, and transferability. These standards can be reframed using Messick's five sources of validity evidence (content, response process, internal structure, relationships with other variables, and consequences) and Kane's four inferences in validation (scoring, generalization, extrapolation, and implications). Evidence can be collected and evaluated for each evidence source or inference. The authors illustrate this approach using published research on learning portfolios.The authors advocate a "methods-neutral" approach to assessment, in which a clearly stated purpose determines the nature of and approach to data collection and analysis. Increased use of qualitative assessments will necessitate more rigorous judgments of the defensibility (validity) of inferences and decisions. Evidence should be strategically sought to inform a coherent validity argument.
Snodgrass, Melinda R; Chung, Moon Y; Meadan, Hedda; Halle, James W
2018-03-01
Single-case research (SCR) has been a valuable methodology in special education research. Montrose Wolf (1978), an early pioneer in single-case methodology, coined the term "social validity" to refer to the social importance of the goals selected, the acceptability of procedures employed, and the effectiveness of the outcomes produced in applied investigations. Since 1978, many contributors to SCR have included social validity as a feature of their articles and several authors have examined the prevalence and role of social validity in SCR. We systematically reviewed all SCR published in six highly-ranked special education journals from 2005 to 2016 to establish the prevalence of social validity assessments and to evaluate their scientific rigor. We found relatively low, but stable prevalence with only 28 publications addressing all three factors of the social validity construct (i.e., goals, procedures, outcomes). We conducted an in-depth analysis of the scientific rigor of these 28 publications. Social validity remains an understudied construct in SCR, and the scientific rigor of social validity assessments is often lacking. Implications and future directions are discussed. Copyright © 2018 Elsevier Ltd. All rights reserved.
A computational framework for automation of point defect calculations
DOE Office of Scientific and Technical Information (OSTI.GOV)
Goyal, Anuj; Gorai, Prashun; Peng, Haowei
We have developed a complete and rigorously validated open-source Python framework to automate point defect calculations using density functional theory. Furthermore, the framework provides an effective and efficient method for defect structure generation, and creation of simple yet customizable workflows to analyze defect calculations. This package provides the capability to compute widely-accepted correction schemes to overcome finite-size effects, including (1) potential alignment, (2) image-charge correction, and (3) band filling correction to shallow defects. Using Si, ZnO and In2O3 as test examples, we demonstrate the package capabilities and validate the methodology.
A computational framework for automation of point defect calculations
Goyal, Anuj; Gorai, Prashun; Peng, Haowei; ...
2017-01-13
We have developed a complete and rigorously validated open-source Python framework to automate point defect calculations using density functional theory. Furthermore, the framework provides an effective and efficient method for defect structure generation, and creation of simple yet customizable workflows to analyze defect calculations. This package provides the capability to compute widely-accepted correction schemes to overcome finite-size effects, including (1) potential alignment, (2) image-charge correction, and (3) band filling correction to shallow defects. Using Si, ZnO and In2O3 as test examples, we demonstrate the package capabilities and validate the methodology.
Schwartz, Lisa A; Hamilton, Jessica L; Brumley, Lauren D; Barakat, Lamia P; Deatrick, Janet A; Szalda, Dava E; Bevans, Katherine B; Tucker, Carole A; Daniel, Lauren C; Butler, Eliana; Kazak, Anne E; Hobbie, Wendy L; Ginsberg, Jill P; Psihogios, Alexandra M; Ver Hoeve, Elizabeth; Tuchman, Lisa K
2017-10-01
The development of the Transition Readiness Inventory (TRI) item pool for adolescent and young adult childhood cancer survivors is described, aiming to both advance transition research and provide an example of the application of NIH Patient Reported Outcomes Information System methods. Using rigorous measurement development methods including mixed methods, patient and parent versions of the TRI item pool were created based on the Social-ecological Model of Adolescent and young adult Readiness for Transition (SMART). Each stage informed development and refinement of the item pool. Content validity ratings and cognitive interviews resulted in 81 content valid items for the patient version and 85 items for the parent version. TRI represents the first multi-informant, rigorously developed transition readiness item pool that comprehensively measures the social-ecological components of transition readiness. Discussion includes clinical implications, the application of TRI and the methods to develop the item pool to other populations, and next steps for further validation and refinement. © The Author 2017. Published by Oxford University Press on behalf of the Society of Pediatric Psychology. All rights reserved. For permissions, please e-mail: journals.permissions@oup.com
The Youth Diversity Acceptance Scale: Development and Validity
ERIC Educational Resources Information Center
Beck, Kara L.; Acevedo-Polakovich, Ignacio D.; Lyons, Eileen; Estevez, Jessica; Sevecke, Jessica R.; Rossman, Danielle L.; Barnett, Miya L.; Fisher, Heidi R.
2018-01-01
To facilitate rigorous research on community programs to promote positive intergroup relations among youth, 2 studies were conducted to establish the psychometric properties of the Youth Diversity Acceptance Scale (YoDA). In Study 1, a unifactorial structure was identified for the YoDA using data provided by 126 high school youth involved in a…
Studying Sexual Aggression: A Review of the Evolution and Validity of Laboratory Paradigms
Davis, Kelly Cue; George, William H.; Nagayama Hall, Gordon C.; Parrott, Dominic J.; Tharp, Andra Teten; Stappenbeck, Cynthia A.
2018-01-01
Objective Researchers have endeavored for decades to develop and implement experimental assessments of sexual aggression and its precursors to capitalize on the many scientific advantages offered by laboratory experiments, such as rigorous control of key variables and identification of causal relationships. The purpose of this review is to provide an overview of and commentary on the evolution of these laboratory-based methods. Conclusions To date, two primary types of sexual aggression laboratory studies have been developed: those that involve behavioral analogues of sexual aggression and those that assess postulated precursors to sexually aggressive behavior. Although the study of sexual aggression in the laboratory is fraught with methodological challenges, validity concerns, and ethical considerations, advances in the field have resulted in greater methodological rigor, more precise dependent measures, and improved experimental validity, reliability, and realism. Because highly effective sexual aggression prevention strategies remain elusive, continued laboratory-based investigation of sexual aggression coupled with translation of critical findings to the development and modification of sexual aggression prevention programs remains an important task for the field. PMID:29675289
Development and Validation of an Instrument to Measure University Students' Biotechnology Attitude
NASA Astrophysics Data System (ADS)
Erdogan, Mehmet; Özel, Murat; Uşak, Muhammet; Prokop, Pavol
2009-06-01
The impact of biotechnologies on peoples' everyday lives continuously increases. Measuring young peoples' attitudes toward biotechnologies is therefore very important and its results are useful not only for science curriculum developers and policy makers, but also for producers and distributors of genetically modified products. Despite of substantial number of instruments which focused on measuring student attitudes toward biotechnology, a majority of them were not rigorously validated. This study deals with the development and validation of an attitude questionnaire toward biotechnology. Detailed information on development and validation process of the instrument is provided. Data gathered from 326 university students provided evidence for the validity and reliability of the new instrument which consists of 28 attitude items on a five point likert type scale. It is believed that the instrument will serve as a valuable tool for both instructors and researchers in science education to assess students' biotechnology attitudes.
How to Map Theory: Reliable Methods Are Fruitless Without Rigorous Theory.
Gray, Kurt
2017-09-01
Good science requires both reliable methods and rigorous theory. Theory allows us to build a unified structure of knowledge, to connect the dots of individual studies and reveal the bigger picture. Some have criticized the proliferation of pet "Theories," but generic "theory" is essential to healthy science, because questions of theory are ultimately those of validity. Although reliable methods and rigorous theory are synergistic, Action Identification suggests psychological tension between them: The more we focus on methodological details, the less we notice the broader connections. Therefore, psychology needs to supplement training in methods (how to design studies and analyze data) with training in theory (how to connect studies and synthesize ideas). This article provides a technique for visually outlining theory: theory mapping. Theory mapping contains five elements, which are illustrated with moral judgment and with cars. Also included are 15 additional theory maps provided by experts in emotion, culture, priming, power, stress, ideology, morality, marketing, decision-making, and more (see all at theorymaps.org ). Theory mapping provides both precision and synthesis, which helps to resolve arguments, prevent redundancies, assess the theoretical contribution of papers, and evaluate the likelihood of surprising effects.
Hickey, Graeme L; Blackstone, Eugene H
2016-08-01
Clinical risk-prediction models serve an important role in healthcare. They are used for clinical decision-making and measuring the performance of healthcare providers. To establish confidence in a model, external model validation is imperative. When designing such an external model validation study, thought must be given to patient selection, risk factor and outcome definitions, missing data, and the transparent reporting of the analysis. In addition, there are a number of statistical methods available for external model validation. Execution of a rigorous external validation study rests in proper study design, application of suitable statistical methods, and transparent reporting. Copyright © 2016 The American Association for Thoracic Surgery. Published by Elsevier Inc. All rights reserved.
ERIC Educational Resources Information Center
Rebell, Michael
2007-01-01
Background/Context: In recent years, state legislatures, state education departments, and advocacy groups in over 30 states have sponsored education adequacy studies, which aim to determine objectively the amount of funding needed to provide all students a meaningful opportunity for an adequate education. Not surprisingly, because of their growing…
Characteristics of School Districts That Participate in Rigorous National Educational Evaluations
Stuart, Elizabeth A.; Bell, Stephen H.; Ebnesajjad, Cyrus; Olsen, Robert B.; Orr, Larry L.
2017-01-01
Given increasing interest in evidence-based policy, there is growing attention to how well the results from rigorous program evaluations may inform policy decisions. However, little attention has been paid to documenting the characteristics of schools or districts that participate in rigorous educational evaluations, and how they compare to potential target populations for the interventions that were evaluated. Utilizing a list of the actual districts that participated in 11 large-scale rigorous educational evaluations, we compare those districts to several different target populations of districts that could potentially be affected by policy decisions regarding the interventions under study. We find that school districts that participated in the 11 rigorous educational evaluations differ from the interventions’ target populations in several ways, including size, student performance on state assessments, and location (urban/rural). These findings raise questions about whether, as currently implemented, the results from rigorous impact studies in education are likely to generalize to the larger set of school districts—and thus schools and students—of potential interest to policymakers, and how we can improve our study designs to retain strong internal validity while also enhancing external validity. PMID:29276552
Predictive QSAR modeling workflow, model applicability domains, and virtual screening.
Tropsha, Alexander; Golbraikh, Alexander
2007-01-01
Quantitative Structure Activity Relationship (QSAR) modeling has been traditionally applied as an evaluative approach, i.e., with the focus on developing retrospective and explanatory models of existing data. Model extrapolation was considered if only in hypothetical sense in terms of potential modifications of known biologically active chemicals that could improve compounds' activity. This critical review re-examines the strategy and the output of the modern QSAR modeling approaches. We provide examples and arguments suggesting that current methodologies may afford robust and validated models capable of accurate prediction of compound properties for molecules not included in the training sets. We discuss a data-analytical modeling workflow developed in our laboratory that incorporates modules for combinatorial QSAR model development (i.e., using all possible binary combinations of available descriptor sets and statistical data modeling techniques), rigorous model validation, and virtual screening of available chemical databases to identify novel biologically active compounds. Our approach places particular emphasis on model validation as well as the need to define model applicability domains in the chemistry space. We present examples of studies where the application of rigorously validated QSAR models to virtual screening identified computational hits that were confirmed by subsequent experimental investigations. The emerging focus of QSAR modeling on target property forecasting brings it forward as predictive, as opposed to evaluative, modeling approach.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Ricci, P., E-mail: paolo.ricci@epfl.ch; Riva, F.; Theiler, C.
In the present work, a Verification and Validation procedure is presented and applied showing, through a practical example, how it can contribute to advancing our physics understanding of plasma turbulence. Bridging the gap between plasma physics and other scientific domains, in particular, the computational fluid dynamics community, a rigorous methodology for the verification of a plasma simulation code is presented, based on the method of manufactured solutions. This methodology assesses that the model equations are correctly solved, within the order of accuracy of the numerical scheme. The technique to carry out a solution verification is described to provide a rigorousmore » estimate of the uncertainty affecting the numerical results. A methodology for plasma turbulence code validation is also discussed, focusing on quantitative assessment of the agreement between experiments and simulations. The Verification and Validation methodology is then applied to the study of plasma turbulence in the basic plasma physics experiment TORPEX [Fasoli et al., Phys. Plasmas 13, 055902 (2006)], considering both two-dimensional and three-dimensional simulations carried out with the GBS code [Ricci et al., Plasma Phys. Controlled Fusion 54, 124047 (2012)]. The validation procedure allows progress in the understanding of the turbulent dynamics in TORPEX, by pinpointing the presence of a turbulent regime transition, due to the competition between the resistive and ideal interchange instabilities.« less
Rigor, vigor, and the study of health disparities
Adler, Nancy; Bush, Nicole R.; Pantell, Matthew S.
2012-01-01
Health disparities research spans multiple fields and methods and documents strong links between social disadvantage and poor health. Associations between socioeconomic status (SES) and health are often taken as evidence for the causal impact of SES on health, but alternative explanations, including the impact of health on SES, are plausible. Studies showing the influence of parents’ SES on their children’s health provide evidence for a causal pathway from SES to health, but have limitations. Health disparities researchers face tradeoffs between “rigor” and “vigor” in designing studies that demonstrate how social disadvantage becomes biologically embedded and results in poorer health. Rigorous designs aim to maximize precision in the measurement of SES and health outcomes through methods that provide the greatest control over temporal ordering and causal direction. To achieve precision, many studies use a single SES predictor and single disease. However, doing so oversimplifies the multifaceted, entwined nature of social disadvantage and may overestimate the impact of that one variable and underestimate the true impact of social disadvantage on health. In addition, SES effects on overall health and functioning are likely to be greater than effects on any one disease. Vigorous designs aim to capture this complexity and maximize ecological validity through more complete assessment of social disadvantage and health status, but may provide less-compelling evidence of causality. Newer approaches to both measurement and analysis may enable enhanced vigor as well as rigor. Incorporating both rigor and vigor into studies will provide a fuller understanding of the causes of health disparities. PMID:23045672
Quasi-experimental study designs series-paper 6: risk of bias assessment.
Waddington, Hugh; Aloe, Ariel M; Becker, Betsy Jane; Djimeu, Eric W; Hombrados, Jorge Garcia; Tugwell, Peter; Wells, George; Reeves, Barney
2017-09-01
Rigorous and transparent bias assessment is a core component of high-quality systematic reviews. We assess modifications to existing risk of bias approaches to incorporate rigorous quasi-experimental approaches with selection on unobservables. These are nonrandomized studies using design-based approaches to control for unobservable sources of confounding such as difference studies, instrumental variables, interrupted time series, natural experiments, and regression-discontinuity designs. We review existing risk of bias tools. Drawing on these tools, we present domains of bias and suggest directions for evaluation questions. The review suggests that existing risk of bias tools provide, to different degrees, incomplete transparent criteria to assess the validity of these designs. The paper then presents an approach to evaluating the internal validity of quasi-experiments with selection on unobservables. We conclude that tools for nonrandomized studies of interventions need to be further developed to incorporate evaluation questions for quasi-experiments with selection on unobservables. Copyright © 2017 Elsevier Inc. All rights reserved.
Accuracy and performance of 3D mask models in optical projection lithography
NASA Astrophysics Data System (ADS)
Agudelo, Viviana; Evanschitzky, Peter; Erdmann, Andreas; Fühner, Tim; Shao, Feng; Limmer, Steffen; Fey, Dietmar
2011-04-01
Different mask models have been compared: rigorous electromagnetic field (EMF) modeling, rigorous EMF modeling with decomposition techniques and the thin mask approach (Kirchhoff approach) to simulate optical diffraction from different mask patterns in projection systems for lithography. In addition, each rigorous model was tested for two different formulations for partially coherent imaging: The Hopkins assumption and rigorous simulation of mask diffraction orders for multiple illumination angles. The aim of this work is to closely approximate results of the rigorous EMF method by the thin mask model enhanced with pupil filtering techniques. The validity of this approach for different feature sizes, shapes and illumination conditions is investigated.
The Coverage of the Holocaust in High School History Textbooks
ERIC Educational Resources Information Center
Lindquist, David
2009-01-01
The Holocaust is now a regular part of high school history curricula throughout the United States and, as a result, coverage of the Holocaust has become a standard feature of high school textbooks. As with any major event, it is important for textbooks to provide a rigorously accurate and valid historical account. In dealing with the Holocaust,…
Wild, Diane; Furtado, Tamzin; Angalakuditi, Mallik
2012-01-01
Background The Child Behavior Checklist (CBCL) is a caregiver rating scale for assessing the behavioral profile of children. It was developed in the US, and has been extensively translated and used in a large number of studies internationally. Objective The objective of this study was to translate the CBCL into six languages using a rigorous translation methodology, placing particular emphasis on cultural adaptation and ensuring that the measure has content validity with carers of children with epilepsy. Methods A rigorous translation and cultural adaptation methodology was used. This is a process which includes two forward translations, reconciliation, two back-translations, and cognitive debriefing interviews with five carers of children with epilepsy in each country. In addition, a series of open-ended questions were asked of the carers in order to provide evidence of content validity. Results A number of cultural adaptations were made during the translation process. This included adaptations to the examples of sports and hobbies. An addition of “milk delivery” was made to the job examples in the Malayalam translation. In addition, two sexual problem items were removed from the Hebrew translation for Israel. Conclusion An additional six translations of the CBCL are now available for use in multinational studies. These translations have evidence of content validity for use with parents of children with epilepsy and have been appropriately culturally adapted so that they are acceptable for use in the target countries. The study highlights the importance of a rigorous translation process and the process of cultural adaptation. PMID:22715318
Self-report: psychology's four-letter word.
Haeffel, Gerald J; Howard, George S
2010-01-01
Self-report continues to be one of the most widely used measurement strategies in psychology despite longstanding concerns about its validity and scientific rigor. In this article, the merits of self-report are examined from a philosophy of science perspective. A framework is also provided for evaluating self-report measures. Specifically, four issues are presented that can be used as a decision aid when making choices about measurement.
Peer Review Documents Related to the Evaluation of ...
BMDS is one of the Agency's premier tools for estimating risk assessments, therefore the validity and reliability of its statistical models are of paramount importance. This page provides links to peer review and expert summaries of the BMDS application and its models as they were developed and eventually released documenting the rigorous review process taken to provide the best science tools available for statistical modeling. This page provides links to peer reviews and expert summaries of the BMDS applications and its models as they were developed and eventually released.
Empirical validation of an agent-based model of wood markets in Switzerland
Hilty, Lorenz M.; Lemm, Renato; Thees, Oliver
2018-01-01
We present an agent-based model of wood markets and show our efforts to validate this model using empirical data from different sources, including interviews, workshops, experiments, and official statistics. Own surveys closed gaps where data was not available. Our approach to model validation used a variety of techniques, including the replication of historical production amounts, prices, and survey results, as well as a historical case study of a large sawmill entering the market and becoming insolvent only a few years later. Validating the model using this case provided additional insights, showing how the model can be used to simulate scenarios of resource availability and resource allocation. We conclude that the outcome of the rigorous validation qualifies the model to simulate scenarios concerning resource availability and allocation in our study region. PMID:29351300
NASA Astrophysics Data System (ADS)
Eisfeld, Eugen; Roth, Johannes
2018-05-01
Based on hybrid molecular dynamics/two-temperature simulations, we study the validity of the application of Lambert-Beer's law, which is conveniently used in various modeling approaches of ultra-short pulse laser ablation of metals. The method is compared to a more rigorous treatment, which involves solving the Helmholtz wave equation for different pulse durations ranging from 100 fs to 5 ps and a wavelength of 800 nm. Our simulations show a growing agreement with increasing pulse durations, and we provide appropriate optical parameters for all investigated pulse durations.
Impact of AMS-02 Measurements on Reducing GCR Model Uncertainties
NASA Technical Reports Server (NTRS)
Slaba, T. C.; O'Neill, P. M.; Golge, S.; Norbury, J. W.
2015-01-01
For vehicle design, shield optimization, mission planning, and astronaut risk assessment, the exposure from galactic cosmic rays (GCR) poses a significant and complex problem both in low Earth orbit and in deep space. To address this problem, various computational tools have been developed to quantify the exposure and risk in a wide range of scenarios. Generally, the tool used to describe the ambient GCR environment provides the input into subsequent computational tools and is therefore a critical component of end-to-end procedures. Over the past few years, several researchers have independently and very carefully compared some of the widely used GCR models to more rigorously characterize model differences and quantify uncertainties. All of the GCR models studied rely heavily on calibrating to available near-Earth measurements of GCR particle energy spectra, typically over restricted energy regions and short time periods. In this work, we first review recent sensitivity studies quantifying the ions and energies in the ambient GCR environment of greatest importance to exposure quantities behind shielding. Currently available measurements used to calibrate and validate GCR models are also summarized within this context. It is shown that the AMS-II measurements will fill a critically important gap in the measurement database. The emergence of AMS-II measurements also provides a unique opportunity to validate existing models against measurements that were not used to calibrate free parameters in the empirical descriptions. Discussion is given regarding rigorous approaches to implement the independent validation efforts, followed by recalibration of empirical parameters.
Araújo, Luciano V; Malkowski, Simon; Braghetto, Kelly R; Passos-Bueno, Maria R; Zatz, Mayana; Pu, Calton; Ferreira, João E
2011-12-22
Recent medical and biological technology advances have stimulated the development of new testing systems that have been providing huge, varied amounts of molecular and clinical data. Growing data volumes pose significant challenges for information processing systems in research centers. Additionally, the routines of genomics laboratory are typically characterized by high parallelism in testing and constant procedure changes. This paper describes a formal approach to address this challenge through the implementation of a genetic testing management system applied to human genome laboratory. We introduced the Human Genome Research Center Information System (CEGH) in Brazil, a system that is able to support constant changes in human genome testing and can provide patients updated results based on the most recent and validated genetic knowledge. Our approach uses a common repository for process planning to ensure reusability, specification, instantiation, monitoring, and execution of processes, which are defined using a relational database and rigorous control flow specifications based on process algebra (ACP). The main difference between our approach and related works is that we were able to join two important aspects: 1) process scalability achieved through relational database implementation, and 2) correctness of processes using process algebra. Furthermore, the software allows end users to define genetic testing without requiring any knowledge about business process notation or process algebra. This paper presents the CEGH information system that is a Laboratory Information Management System (LIMS) based on a formal framework to support genetic testing management for Mendelian disorder studies. We have proved the feasibility and showed usability benefits of a rigorous approach that is able to specify, validate, and perform genetic testing using easy end user interfaces.
2011-01-01
Background Recent medical and biological technology advances have stimulated the development of new testing systems that have been providing huge, varied amounts of molecular and clinical data. Growing data volumes pose significant challenges for information processing systems in research centers. Additionally, the routines of genomics laboratory are typically characterized by high parallelism in testing and constant procedure changes. Results This paper describes a formal approach to address this challenge through the implementation of a genetic testing management system applied to human genome laboratory. We introduced the Human Genome Research Center Information System (CEGH) in Brazil, a system that is able to support constant changes in human genome testing and can provide patients updated results based on the most recent and validated genetic knowledge. Our approach uses a common repository for process planning to ensure reusability, specification, instantiation, monitoring, and execution of processes, which are defined using a relational database and rigorous control flow specifications based on process algebra (ACP). The main difference between our approach and related works is that we were able to join two important aspects: 1) process scalability achieved through relational database implementation, and 2) correctness of processes using process algebra. Furthermore, the software allows end users to define genetic testing without requiring any knowledge about business process notation or process algebra. Conclusions This paper presents the CEGH information system that is a Laboratory Information Management System (LIMS) based on a formal framework to support genetic testing management for Mendelian disorder studies. We have proved the feasibility and showed usability benefits of a rigorous approach that is able to specify, validate, and perform genetic testing using easy end user interfaces. PMID:22369688
Investigation of the Thermomechanical Response of Shape Memory Alloy Hybrid Composite Beams
NASA Technical Reports Server (NTRS)
Davis, Brian A.
2005-01-01
Previous work at NASA Langley Research Center (LaRC) involved fabrication and testing of composite beams with embedded, pre-strained shape memory alloy (SMA) ribbons. That study also provided comparison of experimental results with numerical predictions from a research code making use of a new thermoelastic model for shape memory alloy hybrid composite (SMAHC) structures. The previous work showed qualitative validation of the numerical model. However, deficiencies in the experimental-numerical correlation were noted and hypotheses for the discrepancies were given for further investigation. The goal of this work is to refine the experimental measurement and numerical modeling approaches in order to better understand the discrepancies, improve the correlation between prediction and measurement, and provide rigorous quantitative validation of the numerical model. Thermal buckling, post-buckling, and random responses to thermal and inertial (base acceleration) loads are studied. Excellent agreement is achieved between the predicted and measured results, thereby quantitatively validating the numerical tool.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Ramanathan, Arvind; Steed, Chad A; Pullum, Laura L
Compartmental models in epidemiology are widely used as a means to model disease spread mechanisms and understand how one can best control the disease in case an outbreak of a widespread epidemic occurs. However, a significant challenge within the community is in the development of approaches that can be used to rigorously verify and validate these models. In this paper, we present an approach to rigorously examine and verify the behavioral properties of compartmen- tal epidemiological models under several common modeling scenarios including birth/death rates and multi-host/pathogen species. Using metamorphic testing, a novel visualization tool and model checking, we buildmore » a workflow that provides insights into the functionality of compartmental epidemiological models. Our initial results indicate that metamorphic testing can be used to verify the implementation of these models and provide insights into special conditions where these mathematical models may fail. The visualization front-end allows the end-user to scan through a variety of parameters commonly used in these models to elucidate the conditions under which an epidemic can occur. Further, specifying these models using a process algebra allows one to automatically construct behavioral properties that can be rigorously verified using model checking. Taken together, our approach allows for detecting implementation errors as well as handling conditions under which compartmental epidemiological models may fail to provide insights into disease spread dynamics.« less
Kirkpatrick, Beth D; Colgate, E Ross; Mychaleckyj, Josyf C; Haque, Rashidul; Dickson, Dorothy M; Carmolli, Marya P; Nayak, Uma; Taniuchi, Mami; Naylor, Caitlin; Qadri, Firdausi; Ma, Jennie Z; Alam, Masud; Walsh, Mary Claire; Diehl, Sean A; Petri, William A
2015-04-01
Oral vaccines appear less effective in children in the developing world. Proposed biologic reasons include concurrent enteric infections, malnutrition, breast milk interference, and environmental enteropathy (EE). Rigorous study design and careful data management are essential to begin to understand this complex problem while assuring research subject safety. Herein, we describe the methodology and lessons learned in the PROVIDE study (Dhaka, Bangladesh). A randomized clinical trial platform evaluated the efficacy of delayed-dose oral rotavirus vaccine as well as the benefit of an injectable polio vaccine replacing one dose of oral polio vaccine. This rigorous infrastructure supported the additional examination of hypotheses of vaccine underperformance. Primary and secondary efficacy and immunogenicity measures for rotavirus and polio vaccines were measured, as well as the impact of EE and additional exploratory variables. Methods for the enrollment and 2-year follow-up of a 700 child birth cohort are described, including core laboratory, safety, regulatory, and data management practices. Intense efforts to standardize clinical, laboratory, and data management procedures in a developing world setting provide clinical trials rigor to all outcomes. Although this study infrastructure requires extensive time and effort, it allows optimized safety and confidence in the validity of data gathered in complex, developing country settings. © The American Society of Tropical Medicine and Hygiene.
Assessing Sensitivity of Early Head Start Study Findings to Manipulated Randomization Threats
ERIC Educational Resources Information Center
Green, Sheridan
2013-01-01
Increasing demands for design rigor and an emphasis on evidence-based practice on a national level indicated a need for further guidance related to successful implementation of randomized studies in education. Rigorous and meaningful experimental research and its conclusions help establish a valid theoretical and evidence base for educational…
Quantitative validation of carbon-fiber laminate low velocity impact simulations
English, Shawn A.; Briggs, Timothy M.; Nelson, Stacy M.
2015-09-26
Simulations of low velocity impact with a flat cylindrical indenter upon a carbon fiber fabric reinforced polymer laminate are rigorously validated. Comparison of the impact energy absorption between the model and experiment is used as the validation metric. Additionally, non-destructive evaluation, including ultrasonic scans and three-dimensional computed tomography, provide qualitative validation of the models. The simulations include delamination, matrix cracks and fiber breaks. An orthotropic damage and failure constitutive model, capable of predicting progressive damage and failure, is developed in conjunction and described. An ensemble of simulations incorporating model parameter uncertainties is used to predict a response distribution which ismore » then compared to experimental output using appropriate statistical methods. Lastly, the model form errors are exposed and corrected for use in an additional blind validation analysis. The result is a quantifiable confidence in material characterization and model physics when simulating low velocity impact in structures of interest.« less
Validation of Fatigue Modeling Predictions in Aviation Operations
NASA Technical Reports Server (NTRS)
Gregory, Kevin; Martinez, Siera; Flynn-Evans, Erin
2017-01-01
Bio-mathematical fatigue models that predict levels of alertness and performance are one potential tool for use within integrated fatigue risk management approaches. A number of models have been developed that provide predictions based on acute and chronic sleep loss, circadian desynchronization, and sleep inertia. Some are publicly available and gaining traction in settings such as commercial aviation as a means of evaluating flight crew schedules for potential fatigue-related risks. Yet, most models have not been rigorously evaluated and independently validated for the operations to which they are being applied and many users are not fully aware of the limitations in which model results should be interpreted and applied.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Reynolds, William; Weber, Marta S.; Farber, Robert M.
Social Media provide an exciting and novel view into social phenomena. The vast amounts of data that can be gathered from the Internet coupled with massively parallel supercomputers such as the Cray XMT open new vistas for research. Conclusions drawn from such analysis must recognize that social media are distinct from the underlying social reality. Rigorous validation is essential. This paper briefly presents results obtained from computational analysis of social media - utilizing both blog and twitter data. Validation of these results is discussed in the context of a framework of established methodologies from the social sciences. Finally, an outlinemore » for a set of supporting studies is proposed.« less
2014-05-01
propagations CoCs Contaminants of concern GC Gas chromatography DNAPL Dense nonaqueous phase liquid ISCO In situ chemical oxidation HCA...used for the design and scale-up of air strippers, ion exchange systems, precipitation reactors , and many other treatment processes. Such treatability...studies provide definitive data on system dimensions and reagent dosages using linear or non -linear scale-up. Designing these processes without the
NASA Technical Reports Server (NTRS)
Jenkins, Michael G.; Salem, Jonathan A.
2016-01-01
Physical and mechanical properties and performance of advanced ceramics and glasses are difficult to measure correctly without the proper techniques. For over three decades, ASTM Committee C28 on Advanced Ceramics, has developed high quality, rigorous, full-consensus standards (e.g., test methods, practices, guides, terminology) to measure properties and performance of monolithic and composite ceramics that may be applied to glasses in some cases. These standards testing particulars for many mechanical, physical, thermal, properties and performance of these materials. As a result these standards provide accurate, reliable, repeatable and complete data. Within Committee C28 users, producers, researchers, designers, academicians, etc. have written, continually updated, and validated through round-robin test programs, nearly 50 standards since the Committees founding in 1986. This paper provides a retrospective review of the 30 years of ASTM Committee C28 including a graphical pictogram listing of C28 standards along with examples of the tangible benefits of advanced ceramics standards to demonstrate their practical applications.
Educational Research as Disciplined Inquiry: Examining the Facets of Rigor in Our Work.
ERIC Educational Resources Information Center
Munby, Hugh
This paper explores how facets of the concept "rigor" might be applied to questions about the validity and reliability of research independently of the research modes. The focus of the critical lens could then be on how to assess the contribution of various forms of research rather than on the "paradigm wars" and arguments…
A Computational Framework for Automation of Point Defect Calculations
NASA Astrophysics Data System (ADS)
Goyal, Anuj; Gorai, Prashun; Peng, Haowei; Lany, Stephan; Stevanovic, Vladan; National Renewable Energy Laboratory, Golden, Colorado 80401 Collaboration
A complete and rigorously validated open-source Python framework to automate point defect calculations using density functional theory has been developed. The framework provides an effective and efficient method for defect structure generation, and creation of simple yet customizable workflows to analyze defect calculations. The package provides the capability to compute widely accepted correction schemes to overcome finite-size effects, including (1) potential alignment, (2) image-charge correction, and (3) band filling correction to shallow defects. Using Si, ZnO and In2O3as test examples, we demonstrate the package capabilities and validate the methodology. We believe that a robust automated tool like this will enable the materials by design community to assess the impact of point defects on materials performance. National Renewable Energy Laboratory, Golden, Colorado 80401.
Using Focus Groups to Validate a Pharmacy Vaccination Training Program.
Bushell, Mary; Morrissey, Hana; Ball, Patrick
2015-06-12
Introduction: Focus group methodology is commonly used to quickly collate, integrated views from a variety of different stakeholders. This paper provides an example of how focus groups can be employed to collate expert opinion informing amendments on a newly developed training program for integration into undergraduate pharmacy curricula. Materials and methods: Four focus groups were conducted, across three continents, to determine the appropriateness and reliability of a developed vaccination training program with nested injection skills training. All focus groups were comprised of legitimate experts in the field of vaccination, medicine and/or pharmacy. Results: Themes that emerged across focus groups informed amendments giving rise to a validated version of a training program. Discussion : The rigorous validation of the vaccination training program offers generalizable lessons to inform the design and validation of future training programs intended for the health sector and or pharmacy curricula. Using the knowledge and experience of focus group participants fostered collaborative problem solving and validation of material and concept development. The group dynamics of a focus group allowed synthesis of feedback in an inter-professional manner. Conclusions : This paper provides a demonstration of how focus groups can be structured and used by health researchers to validate a newly developed training program.
The MINERVA Software Development Process
NASA Technical Reports Server (NTRS)
Narkawicz, Anthony; Munoz, Cesar A.; Dutle, Aaron M.
2017-01-01
This paper presents a software development process for safety-critical software components of cyber-physical systems. The process is called MINERVA, which stands for Mirrored Implementation Numerically Evaluated against Rigorously Verified Algorithms. The process relies on formal methods for rigorously validating code against its requirements. The software development process uses: (1) a formal specification language for describing the algorithms and their functional requirements, (2) an interactive theorem prover for formally verifying the correctness of the algorithms, (3) test cases that stress the code, and (4) numerical evaluation on these test cases of both the algorithm specifications and their implementations in code. The MINERVA process is illustrated in this paper with an application to geo-containment algorithms for unmanned aircraft systems. These algorithms ensure that the position of an aircraft never leaves a predetermined polygon region and provide recovery maneuvers when the region is inadvertently exited.
The Importance of Method Selection in Determining Product Integrity for Nutrition Research1234
Mudge, Elizabeth M; Brown, Paula N
2016-01-01
The American Herbal Products Association estimates that there as many as 3000 plant species in commerce. The FDA estimates that there are about 85,000 dietary supplement products in the marketplace. The pace of product innovation far exceeds that of analytical methods development and validation, with new ingredients, matrixes, and combinations resulting in an analytical community that has been unable to keep up. This has led to a lack of validated analytical methods for dietary supplements and to inappropriate method selection where methods do exist. Only after rigorous validation procedures to ensure that methods are fit for purpose should they be used in a routine setting to verify product authenticity and quality. By following systematic procedures and establishing performance requirements for analytical methods before method development and validation, methods can be developed that are both valid and fit for purpose. This review summarizes advances in method selection, development, and validation regarding herbal supplement analysis and provides several documented examples of inappropriate method selection and application. PMID:26980823
The Importance of Method Selection in Determining Product Integrity for Nutrition Research.
Mudge, Elizabeth M; Betz, Joseph M; Brown, Paula N
2016-03-01
The American Herbal Products Association estimates that there as many as 3000 plant species in commerce. The FDA estimates that there are about 85,000 dietary supplement products in the marketplace. The pace of product innovation far exceeds that of analytical methods development and validation, with new ingredients, matrixes, and combinations resulting in an analytical community that has been unable to keep up. This has led to a lack of validated analytical methods for dietary supplements and to inappropriate method selection where methods do exist. Only after rigorous validation procedures to ensure that methods are fit for purpose should they be used in a routine setting to verify product authenticity and quality. By following systematic procedures and establishing performance requirements for analytical methods before method development and validation, methods can be developed that are both valid and fit for purpose. This review summarizes advances in method selection, development, and validation regarding herbal supplement analysis and provides several documented examples of inappropriate method selection and application. © 2016 American Society for Nutrition.
ERIC Educational Resources Information Center
Micceri, Theodore; Brigman, Leellen; Spatig, Robert
2009-01-01
An extensive, internally cross-validated analytical study using nested (within academic disciplines) Multilevel Modeling (MLM) on 4,560 students identified functional criteria for defining high school curriculum rigor and further determined which measures could best be used to help guide decision making for marginal applicants. The key outcome…
Critical Analysis of Strategies for Determining Rigor in Qualitative Inquiry.
Morse, Janice M
2015-09-01
Criteria for determining the trustworthiness of qualitative research were introduced by Guba and Lincoln in the 1980s when they replaced terminology for achieving rigor, reliability, validity, and generalizability with dependability, credibility, and transferability. Strategies for achieving trustworthiness were also introduced. This landmark contribution to qualitative research remains in use today, with only minor modifications in format. Despite the significance of this contribution over the past four decades, the strategies recommended to achieve trustworthiness have not been critically examined. Recommendations for where, why, and how to use these strategies have not been developed, and how well they achieve their intended goal has not been examined. We do not know, for example, what impact these strategies have on the completed research. In this article, I critique these strategies. I recommend that qualitative researchers return to the terminology of social sciences, using rigor, reliability, validity, and generalizability. I then make recommendations for the appropriate use of the strategies recommended to achieve rigor: prolonged engagement, persistent observation, and thick, rich description; inter-rater reliability, negative case analysis; peer review or debriefing; clarifying researcher bias; member checking; external audits; and triangulation. © The Author(s) 2015.
DOE Office of Scientific and Technical Information (OSTI.GOV)
English, Shawn A.; Briggs, Timothy M.; Nelson, Stacy M.
Simulations of low velocity impact with a flat cylindrical indenter upon a carbon fiber fabric reinforced polymer laminate are rigorously validated. Comparison of the impact energy absorption between the model and experiment is used as the validation metric. Additionally, non-destructive evaluation, including ultrasonic scans and three-dimensional computed tomography, provide qualitative validation of the models. The simulations include delamination, matrix cracks and fiber breaks. An orthotropic damage and failure constitutive model, capable of predicting progressive damage and failure, is developed in conjunction and described. An ensemble of simulations incorporating model parameter uncertainties is used to predict a response distribution which ismore » then compared to experimental output using appropriate statistical methods. Lastly, the model form errors are exposed and corrected for use in an additional blind validation analysis. The result is a quantifiable confidence in material characterization and model physics when simulating low velocity impact in structures of interest.« less
USDA-ARS?s Scientific Manuscript database
Background: A review of the literature produced no rigorously tested and validated Spanish-language physical activity survey or evaluation tools for use by USDA’s food assistance and education programs. The purpose of the current study was to develop and evaluate the face validity of a visually enha...
Quality and rigor of the concept mapping methodology: a pooled study analysis.
Rosas, Scott R; Kane, Mary
2012-05-01
The use of concept mapping in research and evaluation has expanded dramatically over the past 20 years. Researchers in academic, organizational, and community-based settings have applied concept mapping successfully without the benefit of systematic analyses across studies to identify the features of a methodologically sound study. Quantitative characteristics and estimates of quality and rigor that may guide for future studies are lacking. To address this gap, we conducted a pooled analysis of 69 concept mapping studies to describe characteristics across study phases, generate specific indicators of validity and reliability, and examine the relationship between select study characteristics and quality indicators. Individual study characteristics and estimates were pooled and quantitatively summarized, describing the distribution, variation and parameters for each. In addition, variation in the concept mapping data collection in relation to characteristics and estimates was examined. Overall, results suggest concept mapping yields strong internal representational validity and very strong sorting and rating reliability estimates. Validity and reliability were consistently high despite variation in participation and task completion percentages across data collection modes. The implications of these findings as a practical reference to assess the quality and rigor for future concept mapping studies are discussed. Copyright © 2011 Elsevier Ltd. All rights reserved.
Validation of a 30-year-old process for the manufacture of L-asparaginase from Erwinia chrysanthemi.
Gervais, David; Allison, Nigel; Jennings, Alan; Jones, Shane; Marks, Trevor
2013-04-01
A 30-year-old manufacturing process for the biologic product L-asparaginase from the plant pathogen Erwinia chrysanthemi was rigorously qualified and validated, with a high level of agreement between validation data and the 6-year process database. L-Asparaginase exists in its native state as a tetrameric protein and is used as a chemotherapeutic agent in the treatment regimen for Acute Lymphoblastic Leukaemia (ALL). The manufacturing process involves fermentation of the production organism, extraction and purification of the L-asparaginase to make drug substance (DS), and finally formulation and lyophilisation to generate drug product (DP). The extensive manufacturing experience with the product was used to establish ranges for all process parameters and product quality attributes. The product and in-process intermediates were rigorously characterised, and new assays, such as size-exclusion and reversed-phase UPLC, were developed, validated, and used to analyse several pre-validation batches. Finally, three prospective process validation batches were manufactured and product quality data generated using both the existing and the new analytical methods. These data demonstrated the process to be robust, highly reproducible and consistent, and the validation was successful, contributing to the granting of an FDA product license in November, 2011.
Salipur, Zdravko; Bertocci, Gina
2010-01-01
It has been shown that ANSI WC19 transit wheelchairs that are crashworthy in frontal impact exhibit catastrophic failures in rear impact and may not be able to provide stable seating support and thus occupant protection for the wheelchair occupant. Thus far only limited sled test and computer simulation data have been available to study rear impact wheelchair safety. Computer modeling can be used as an economic and comprehensive tool to gain critical knowledge regarding wheelchair integrity and occupant safety. This study describes the development and validation of a computer model simulating an adult wheelchair-seated occupant subjected to a rear impact event. The model was developed in MADYMO and validated rigorously using the results of three similar sled tests conducted to specifications provided in the draft ISO/TC 173 standard. Outcomes from the model can provide critical wheelchair loading information to wheelchair and tiedown manufacturers, resulting in safer wheelchair designs for rear impact conditions. (c) 2009 IPEM. Published by Elsevier Ltd. All rights reserved.
Conflict: Operational Realism versus Analytical Rigor in Defense Modeling and Simulation
2012-06-14
Campbell, Experimental and Quasi- Eperimental Designs for Generalized Causal Inference, Boston: Houghton Mifflin Company, 2002. [7] R. T. Johnson, G...experimentation? In order for an experiment to be considered rigorous, and the results valid, the experiment should be designed using established...addition to the interview, the pilots were administered a written survey, designed to capture their reactions regarding the level of realism present
DESCQA: An Automated Validation Framework for Synthetic Sky Catalogs
NASA Astrophysics Data System (ADS)
Mao, Yao-Yuan; Kovacs, Eve; Heitmann, Katrin; Uram, Thomas D.; Benson, Andrew J.; Campbell, Duncan; Cora, Sofía A.; DeRose, Joseph; Di Matteo, Tiziana; Habib, Salman; Hearin, Andrew P.; Bryce Kalmbach, J.; Krughoff, K. Simon; Lanusse, François; Lukić, Zarija; Mandelbaum, Rachel; Newman, Jeffrey A.; Padilla, Nelson; Paillas, Enrique; Pope, Adrian; Ricker, Paul M.; Ruiz, Andrés N.; Tenneti, Ananth; Vega-Martínez, Cristian A.; Wechsler, Risa H.; Zhou, Rongpu; Zu, Ying; The LSST Dark Energy Science Collaboration
2018-02-01
The use of high-quality simulated sky catalogs is essential for the success of cosmological surveys. The catalogs have diverse applications, such as investigating signatures of fundamental physics in cosmological observables, understanding the effect of systematic uncertainties on measured signals and testing mitigation strategies for reducing these uncertainties, aiding analysis pipeline development and testing, and survey strategy optimization. The list of applications is growing with improvements in the quality of the catalogs and the details that they can provide. Given the importance of simulated catalogs, it is critical to provide rigorous validation protocols that enable both catalog providers and users to assess the quality of the catalogs in a straightforward and comprehensive way. For this purpose, we have developed the DESCQA framework for the Large Synoptic Survey Telescope Dark Energy Science Collaboration as well as for the broader community. The goal of DESCQA is to enable the inspection, validation, and comparison of an inhomogeneous set of synthetic catalogs via the provision of a common interface within an automated framework. In this paper, we present the design concept and first implementation of DESCQA. In order to establish and demonstrate its full functionality we use a set of interim catalogs and validation tests. We highlight several important aspects, both technical and scientific, that require thoughtful consideration when designing a validation framework, including validation metrics and how these metrics impose requirements on the synthetic sky catalogs.
Validation study and routine control monitoring of moist heat sterilization procedures.
Shintani, Hideharu
2012-06-01
The proposed approach to validation of steam sterilization in autoclaves follows the basic life cycle concepts applicable to all validation programs. Understand the function of sterilization process, develop and understand the cycles to carry out the process, and define a suitable test or series of tests to confirm that the function of the process is suitably ensured by the structure provided. Sterilization of product and components and parts that come in direct contact with sterilized product is the most critical of pharmaceutical processes. Consequently, this process requires a most rigorous and detailed approach to validation. An understanding of the process requires a basic understanding of microbial death, the parameters that facilitate that death, the accepted definition of sterility, and the relationship between the definition and sterilization parameters. Autoclaves and support systems need to be designed, installed, and qualified in a manner that ensures their continued reliability. Lastly, the test program must be complete and definitive. In this paper, in addition to validation study, documentation of IQ, OQ and PQ concretely were described.
CFD Validation Experiment of a Mach 2.5 Axisymmetric Shock-Wave/Boundary-Layer Interaction
NASA Technical Reports Server (NTRS)
Davis, David O.
2015-01-01
Experimental investigations of specific flow phenomena, e.g., Shock Wave Boundary-Layer Interactions (SWBLI), provide great insight to the flow behavior but often lack the necessary details to be useful as CFD validation experiments. Reasons include: 1.Undefined boundary conditions Inconsistent results 2.Undocumented 3D effects (CL only measurements) 3.Lack of uncertainty analysis While there are a number of good subsonic experimental investigations that are sufficiently documented to be considered test cases for CFD and turbulence model validation, the number of supersonic and hypersonic cases is much less. This was highlighted by Settles and Dodsons [1] comprehensive review of available supersonic and hypersonic experimental studies. In all, several hundred studies were considered for their database.Of these, over a hundred were subjected to rigorous acceptance criteria. Based on their criteria, only 19 (12 supersonic, 7 hypersonic) were considered of sufficient quality to be used for validation purposes. Aeschliman and Oberkampf [2] recognized the need to develop a specific methodology for experimental studies intended specifically for validation purposes.
Pedagogy and Academic Success in Prelicensure Nursing Education.
Murray, Teri A
2016-01-01
The purpose of this article is to provide a brief description of the New Careers in Nursing (NCIN) program; highlight the features of the NCIN Preentry Immersion program designed to help students achieve academic success; introduce two NCIN innovation teaching projects that used active learning strategies to foster student engagement; and conduct an integrative review on the pedagogies used to foster academic success in nursing education. The integrative review revealed that interactive pedagogies fostered student engagement and increased the students' knowledge acquisition, competence, confidence, and satisfaction. Significant variations in the methodological rigor for the studies included in this review were noted in addition to nebulousness between nursing education research and evaluation. The review validated the need for more rigorous research in nursing education to improve the students' academic experience and subsequent success of all nursing students, including those from underrepresented or disadvantaged backgrounds, enrolled in prelicensure nursing education programs. Copyright © 2016 Elsevier Inc. All rights reserved.
Measuring coherence with entanglement concurrence
NASA Astrophysics Data System (ADS)
Qi, Xianfei; Gao, Ting; Yan, Fengli
2017-07-01
Quantum coherence is a fundamental manifestation of the quantum superposition principle. Recently, Baumgratz et al (2014 Phys. Rev. Lett. 113 140401) presented a rigorous framework to quantify coherence from the view of theory of physical resource. Here we propose a new valid quantum coherence measure which is a convex roof measure, for a quantum system of arbitrary dimension, essentially using the generalized Gell-Mann matrices. Rigorous proof shows that the proposed coherence measure, coherence concurrence, fulfills all the requirements dictated by the resource theory of quantum coherence measures. Moreover, strong links between the resource frameworks of coherence concurrence and entanglement concurrence is derived, which shows that any degree of coherence with respect to some reference basis can be converted to entanglement via incoherent operations. Our work provides a clear quantitative and operational connection between coherence and entanglement based on two kinds of concurrence. This new coherence measure, coherence concurrence, may also be beneficial to the study of quantum coherence.
Effective Recruitment of Schools for Randomized Clinical Trials: Role of School Nurses.
Petosa, R L; Smith, L
2017-01-01
In school settings, nurses lead efforts to improve the student health and well-being to support academic success. Nurses are guided by evidenced-based practice and data to inform care decisions. The randomized controlled trial (RCT) is considered the gold standard of scientific rigor for clinical trials. RCTs are critical to the development of evidence-based health promotion programs in schools. The purpose of this article is to present practical solutions to implementing principles of randomization to RCT trials conducted in school settings. Randomization is a powerful sampling method used to build internal and external validity. The school's daily organization and educational mission provide several barriers to randomization. Based on the authors' experience in conducting school-based RCTs, they offer a host of practical solutions to working with schools to successfully implement randomization procedures. Nurses play a critical role in implementing RCTs in schools to promote rigorous science in support of evidence-based practice.
LIHE Spectral Dynamics and Jaguar Data Acquisition System Measurement Assurance Results 2014.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Covert, Timothy T.; Willis, Michael David; Radtke, Gregg Arthur
2015-06-01
The Light Initiated High Explosive (LIHE) facility performs high rigor, high consequence impulse testing for the nuclear weapons (NW) community. To support the facility mission, LIHE's extensive data acquisition system (DAS) is comprised of several discrete components as well as a fully integrated system. Due to the high consequence and high rigor of the testing performed at LIHE, a measurement assurance plan (MAP) was developed in collaboration with NW system customers to meet their data quality needs and to provide assurance of the robustness of the LIHE DAS. While individual components of the DAS have been calibrated by the SNLmore » Primary Standards Laboratory (PSL), the integrated nature of this complex system requires verification of the complete system, from end-to-end. This measurement assurance plan (MAP) report documents the results of verification and validation procedures used to ensure that the data quality meets customer requirements.« less
Ray-optical theory of broadband partially coherent emission
NASA Astrophysics Data System (ADS)
Epstein, Ariel; Tessler, Nir; Einziger, Pinchas D.
2013-04-01
We present a rigorous formulation of the effects of spectral broadening on emission of partially coherent source ensembles embedded in multilayered formations with arbitrarily shaped interfaces, provided geometrical optics is valid. The resulting ray-optical theory, applicable to a variety of optical systems from terahertz lenses to photovoltaic cells, quantifies the fundamental interplay between bandwidth and layer dimensions, and sheds light on common practices in optical analysis of statistical fields, e.g., disregarding multiple reflections or neglecting interference cross terms.
The Holy Trinity of Methodological Rigor: A Skeptical View
ERIC Educational Resources Information Center
Coryn, Chris L. S.
2007-01-01
The author discusses validation hierarchies grounded in the tradition of quantitative research that generally consists of the criteria of validity, reliability and objectivity and compares this with similar criteria developed by the qualitative tradition, described as trustworthiness, dependability and confirmability. Although these quantitative…
Control group design: enhancing rigor in research of mind-body therapies for depression.
Kinser, Patricia Anne; Robins, Jo Lynne
2013-01-01
Although a growing body of research suggests that mind-body therapies may be appropriate to integrate into the treatment of depression, studies consistently lack methodological sophistication particularly in the area of control groups. In order to better understand the relationship between control group selection and methodological rigor, we provide a brief review of the literature on control group design in yoga and tai chi studies for depression, and we discuss challenges we have faced in the design of control groups for our recent clinical trials of these mind-body complementary therapies for women with depression. To address the multiple challenges of research about mind-body therapies, we suggest that researchers should consider 4 key questions: whether the study design matches the research question; whether the control group addresses performance, expectation, and detection bias; whether the control group is ethical, feasible, and attractive; and whether the control group is designed to adequately control for nonspecific intervention effects. Based on these questions, we provide specific recommendations about control group design with the goal of minimizing bias and maximizing validity in future research.
Control Group Design: Enhancing Rigor in Research of Mind-Body Therapies for Depression
Kinser, Patricia Anne; Robins, Jo Lynne
2013-01-01
Although a growing body of research suggests that mind-body therapies may be appropriate to integrate into the treatment of depression, studies consistently lack methodological sophistication particularly in the area of control groups. In order to better understand the relationship between control group selection and methodological rigor, we provide a brief review of the literature on control group design in yoga and tai chi studies for depression, and we discuss challenges we have faced in the design of control groups for our recent clinical trials of these mind-body complementary therapies for women with depression. To address the multiple challenges of research about mind-body therapies, we suggest that researchers should consider 4 key questions: whether the study design matches the research question; whether the control group addresses performance, expectation, and detection bias; whether the control group is ethical, feasible, and attractive; and whether the control group is designed to adequately control for nonspecific intervention effects. Based on these questions, we provide specific recommendations about control group design with the goal of minimizing bias and maximizing validity in future research. PMID:23662111
Soil moisture and temperature algorithms and validation
USDA-ARS?s Scientific Manuscript database
Passive microwave remote sensing of soil moisture has matured over the past decade as a result of the Advanced Microwave Scanning Radiometer (AMSR) program of JAXA. This program has resulted in improved algorithms that have been supported by rigorous validation. Access to the products and the valida...
A Critical Review of Methods to Evaluate the Impact of FDA Regulatory Actions
Briesacher, Becky A.; Soumerai, Stephen B.; Zhang, Fang; Toh, Sengwee; Andrade, Susan E.; Wagner, Joann L.; Shoaibi, Azadeh; Gurwitz, Jerry H.
2013-01-01
Purpose To conduct a synthesis of the literature on methods to evaluate the impacts of FDA regulatory actions, and identify best practices for future evaluations. Methods We searched MEDLINE for manuscripts published between January 1948 and August 2011 that included terms related to FDA, regulatory actions, and empirical evaluation; the review additionally included FDA-identified literature. We used a modified Delphi method to identify preferred methodologies. We included studies with explicit methods to address threats to validity, and identified designs and analytic methods with strong internal validity that have been applied to other policy evaluations. Results We included 18 studies out of 243 abstracts and papers screened. Overall, analytic rigor in prior evaluations of FDA regulatory actions varied considerably; less than a quarter of studies (22%) included control groups. Only 56% assessed changes in the use of substitute products/services, and 11% examined patient health outcomes. Among studies meeting minimal criteria of rigor, 50% found no impact or weak/modest impacts of FDA actions and 33% detected unintended consequences. Among those studies finding significant intended effects of FDA actions, all cited the importance of intensive communication efforts. There are preferred methods with strong internal validity that have yet to be applied to evaluations of FDA regulatory actions. Conclusions Rigorous evaluations of the impact of FDA regulatory actions have been limited and infrequent. Several methods with strong internal validity are available to improve trustworthiness of future evaluations of FDA policies. PMID:23847020
Moura, Lidia Mvr; Westover, M Brandon; Kwasnik, David; Cole, Andrew J; Hsu, John
2017-01-01
The elderly population faces an increasing number of cases of chronic neurological conditions, such as epilepsy and Alzheimer's disease. Because the elderly with epilepsy are commonly excluded from randomized controlled clinical trials, there are few rigorous studies to guide clinical practice. When the elderly are eligible for trials, they either rarely participate or frequently have poor adherence to therapy, thus limiting both generalizability and validity. In contrast, large observational data sets are increasingly available, but are susceptible to bias when using common analytic approaches. Recent developments in causal inference-analytic approaches also introduce the possibility of emulating randomized controlled trials to yield valid estimates. We provide a practical example of the application of the principles of causal inference to a large observational data set of patients with epilepsy. This review also provides a framework for comparative-effectiveness research in chronic neurological conditions.
Perspectives on Validation of High-Throughput Assays Supporting 21st Century Toxicity Testing
In vitro high-throughput screening (HTS) assays are seeing increasing use in toxicity testing. HTS assays can simultaneously test many chemicals but have seen limited use in the regulatory arena, in part because of the need to undergo rigorous, time-consuming formal validation. ...
GCOM-W soil moisture and temperature algorithms and validation
USDA-ARS?s Scientific Manuscript database
Passive microwave remote sensing of soil moisture has matured over the past decade as a result of the Advanced Microwave Scanning Radiometer (AMSR) program of JAXA. This program has resulted in improved algorithms that have been supported by rigorous validation. Access to the products and the valida...
Nine Criteria for a Measure of Scientific Output
Kreiman, Gabriel; Maunsell, John H. R.
2011-01-01
Scientific research produces new knowledge, technologies, and clinical treatments that can lead to enormous returns. Often, the path from basic research to new paradigms and direct impact on society takes time. Precise quantification of scientific output in the short-term is not an easy task but is critical for evaluating scientists, laboratories, departments, and institutions. While there have been attempts to quantifying scientific output, we argue that current methods are not ideal and suffer from solvable difficulties. Here we propose criteria that a metric should have to be considered a good index of scientific output. Specifically, we argue that such an index should be quantitative, based on robust data, rapidly updated and retrospective, presented with confidence intervals, normalized by number of contributors, career stage and discipline, impractical to manipulate, and focused on quality over quantity. Such an index should be validated through empirical testing. The purpose of quantitatively evaluating scientific output is not to replace careful, rigorous review by experts but rather to complement those efforts. Because it has the potential to greatly influence the efficiency of scientific research, we have a duty to reflect upon and implement novel and rigorous ways of evaluating scientific output. The criteria proposed here provide initial steps toward the systematic development and validation of a metric to evaluate scientific output. PMID:22102840
Cognitive Support During High-Consequence Episodes of Care in Cardiovascular Surgery.
Conboy, Heather M; Avrunin, George S; Clarke, Lori A; Osterweil, Leon J; Christov, Stefan C; Goldman, Julian M; Yule, Steven J; Zenati, Marco A
2017-03-01
Despite significant efforts to reduce preventable adverse events in medical processes, such events continue to occur at unacceptable rates. This paper describes a computer science approach that uses formal process modeling to provide situationally aware monitoring and management support to medical professionals performing complex processes. These process models represent both normative and non-normative situations, and are validated by rigorous automated techniques such as model checking and fault tree analysis, in addition to careful review by experts. Context-aware Smart Checklists are then generated from the models, providing cognitive support during high-consequence surgical episodes. The approach is illustrated with a case study in cardiovascular surgery.
QSAR modeling: where have you been? Where are you going to?
Cherkasov, Artem; Muratov, Eugene N; Fourches, Denis; Varnek, Alexandre; Baskin, Igor I; Cronin, Mark; Dearden, John; Gramatica, Paola; Martin, Yvonne C; Todeschini, Roberto; Consonni, Viviana; Kuz'min, Victor E; Cramer, Richard; Benigni, Romualdo; Yang, Chihae; Rathman, James; Terfloth, Lothar; Gasteiger, Johann; Richard, Ann; Tropsha, Alexander
2014-06-26
Quantitative structure-activity relationship modeling is one of the major computational tools employed in medicinal chemistry. However, throughout its entire history it has drawn both praise and criticism concerning its reliability, limitations, successes, and failures. In this paper, we discuss (i) the development and evolution of QSAR; (ii) the current trends, unsolved problems, and pressing challenges; and (iii) several novel and emerging applications of QSAR modeling. Throughout this discussion, we provide guidelines for QSAR development, validation, and application, which are summarized in best practices for building rigorously validated and externally predictive QSAR models. We hope that this Perspective will help communications between computational and experimental chemists toward collaborative development and use of QSAR models. We also believe that the guidelines presented here will help journal editors and reviewers apply more stringent scientific standards to manuscripts reporting new QSAR studies, as well as encourage the use of high quality, validated QSARs for regulatory decision making.
QSAR Modeling: Where have you been? Where are you going to?
Cherkasov, Artem; Muratov, Eugene N.; Fourches, Denis; Varnek, Alexandre; Baskin, Igor I.; Cronin, Mark; Dearden, John; Gramatica, Paola; Martin, Yvonne C.; Todeschini, Roberto; Consonni, Viviana; Kuz'min, Victor E.; Cramer, Richard; Benigni, Romualdo; Yang, Chihae; Rathman, James; Terfloth, Lothar; Gasteiger, Johann; Richard, Ann; Tropsha, Alexander
2014-01-01
Quantitative Structure-Activity Relationship modeling is one of the major computational tools employed in medicinal chemistry. However, throughout its entire history it has drawn both praise and criticism concerning its reliability, limitations, successes, and failures. In this paper, we discuss: (i) the development and evolution of QSAR; (ii) the current trends, unsolved problems, and pressing challenges; and (iii) several novel and emerging applications of QSAR modeling. Throughout this discussion, we provide guidelines for QSAR development, validation, and application, which are summarized in best practices for building rigorously validated and externally predictive QSAR models. We hope that this Perspective will help communications between computational and experimental chemists towards collaborative development and use of QSAR models. We also believe that the guidelines presented here will help journal editors and reviewers apply more stringent scientific standards to manuscripts reporting new QSAR studies, as well as encourage the use of high quality, validated QSARs for regulatory decision making. PMID:24351051
Study design elements for rigorous quasi-experimental comparative effectiveness research.
Maciejewski, Matthew L; Curtis, Lesley H; Dowd, Bryan
2013-03-01
Quasi-experiments are likely to be the workhorse study design used to generate evidence about the comparative effectiveness of alternative treatments, because of their feasibility, timeliness, affordability and external validity compared with randomized trials. In this review, we outline potential sources of discordance in results between quasi-experiments and experiments, review study design choices that can improve the internal validity of quasi-experiments, and outline innovative data linkage strategies that may be particularly useful in quasi-experimental comparative effectiveness research. There is an urgent need to resolve the debate about the evidentiary value of quasi-experiments since equal consideration of rigorous quasi-experiments will broaden the base of evidence that can be brought to bear in clinical decision-making and governmental policy-making.
DESCQA: An Automated Validation Framework for Synthetic Sky Catalogs
DOE Office of Scientific and Technical Information (OSTI.GOV)
Mao, Yao-Yuan; Kovacs, Eve; Heitmann, Katrin
The use of high-quality simulated sky catalogs is essential for the success of cosmological surveys. The catalogs have diverse applications, such as investigating signatures of fundamental physics in cosmological observables, understanding the effect of systematic uncertainties on measured signals and testing mitigation strategies for reducing these uncertainties, aiding analysis pipeline development and testing, and survey strategy optimization. The list of applications is growing with improvements in the quality of the catalogs and the details that they can provide. Given the importance of simulated catalogs, it is critical to provide rigorous validation protocols that enable both catalog providers and users tomore » assess the quality of the catalogs in a straightforward and comprehensive way. For this purpose, we have developed the DESCQA framework for the Large Synoptic Survey Telescope Dark Energy Science Collaboration as well as for the broader community. The goal of DESCQA is to enable the inspection, validation, and comparison of an inhomogeneous set of synthetic catalogs via the provision of a common interface within an automated framework. Here in this paper, we present the design concept and first implementation of DESCQA. In order to establish and demonstrate its full functionality we use a set of interim catalogs and validation tests. We highlight several important aspects, both technical and scientific, that require thoughtful consideration when designing a validation framework, including validation metrics and how these metrics impose requirements on the synthetic sky catalogs.« less
DESCQA: An Automated Validation Framework for Synthetic Sky Catalogs
Mao, Yao-Yuan; Kovacs, Eve; Heitmann, Katrin; ...
2018-02-08
The use of high-quality simulated sky catalogs is essential for the success of cosmological surveys. The catalogs have diverse applications, such as investigating signatures of fundamental physics in cosmological observables, understanding the effect of systematic uncertainties on measured signals and testing mitigation strategies for reducing these uncertainties, aiding analysis pipeline development and testing, and survey strategy optimization. The list of applications is growing with improvements in the quality of the catalogs and the details that they can provide. Given the importance of simulated catalogs, it is critical to provide rigorous validation protocols that enable both catalog providers and users tomore » assess the quality of the catalogs in a straightforward and comprehensive way. For this purpose, we have developed the DESCQA framework for the Large Synoptic Survey Telescope Dark Energy Science Collaboration as well as for the broader community. The goal of DESCQA is to enable the inspection, validation, and comparison of an inhomogeneous set of synthetic catalogs via the provision of a common interface within an automated framework. Here in this paper, we present the design concept and first implementation of DESCQA. In order to establish and demonstrate its full functionality we use a set of interim catalogs and validation tests. We highlight several important aspects, both technical and scientific, that require thoughtful consideration when designing a validation framework, including validation metrics and how these metrics impose requirements on the synthetic sky catalogs.« less
Proceedings of the 1999 U.S. DOE Hydrogen Program Review
DOE Office of Scientific and Technical Information (OSTI.GOV)
NREL
2000-08-28
The Proceedings of the 1999 US Department of Energy (DOE) Hydrogen Program Review serve as an important technology reference for the DOE Hydrogen Program. This document contains technical progress reports on 60 research and technology validation projects funded by the DOE Hydrogen Program in Fiscal Year 1999, in support of its mission to make hydrogen a cost-effective energy carrier for utility, building, and transportation applications. Each year, the Program conducts a rigorous review of its portfolio of projects, utilizing teams of experts to provide vital feedback on the progress of research.
Fluid moments of the nonlinear Landau collision operator
Hirvijoki, E.; Lingam, M.; Pfefferle, D.; ...
2016-08-09
An important problem in plasma physics is the lack of an accurate and complete description of Coulomb collisions in associated fluid models. To shed light on the problem, this Letter introduces an integral identity involving the multivariate Hermite tensor polynomials and presents a method for computing exact expressions for the fluid moments of the nonlinear Landau collision operator. In conclusion, the proposed methodology provides a systematic and rigorous means of extending the validity of fluid models that have an underlying inverse-square force particle dynamics to arbitrary collisionality and flow.
Co-Occurrence of ADHD and High IQ: A Case Series Empirical Study
ERIC Educational Resources Information Center
Cordeiro, Mara L.; Farias, Antonio C.; Cunha, Alexandre; Benko, Cassia R.; Farias, Lucilene G.; Costa, Maria T.; Martins, Leandra F.; McCracken, James T.
2011-01-01
Objective: The validity of a diagnosis of ADHD in children with a high intelligence quotient (IQ) remains controversial. Using a multidisciplinary approach, rigorous diagnostic criteria, and worldwide-validated psychometric instruments, we identified a group of children attending public schools in southern Brazil for co-occurrence of high IQ and…
Learning Transfer--Validation of the Learning Transfer System Inventory in Portugal
ERIC Educational Resources Information Center
Velada, Raquel; Caetano, Antonio; Bates, Reid; Holton, Ed
2009-01-01
Purpose: The purpose of this paper is to analyze the construct validity of learning transfer system inventory (LTSI) for use in Portugal. Furthermore, it also aims to analyze whether LTSI dimensions differ across individual variables such as gender, age, educational level and job tenure. Design/methodology/approach: After a rigorous translation…
Validation of the GCOM-W SCA and JAXA soil moisture algorithms
USDA-ARS?s Scientific Manuscript database
Satellite-based remote sensing of soil moisture has matured over the past decade as a result of the Global Climate Observing Mission-Water (GCOM-W) program of JAXA. This program has resulted in improved algorithms that have been supported by rigorous validation. Access to the products and the valida...
PRO development: rigorous qualitative research as the crucial foundation.
Lasch, Kathryn Eilene; Marquis, Patrick; Vigneux, Marc; Abetz, Linda; Arnould, Benoit; Bayliss, Martha; Crawford, Bruce; Rosa, Kathleen
2010-10-01
Recently published articles have described criteria to assess qualitative research in the health field in general, but very few articles have delineated qualitative methods to be used in the development of Patient-Reported Outcomes (PROs). In fact, how PROs are developed with subject input through focus groups and interviews has been given relatively short shrift in the PRO literature when compared to the plethora of quantitative articles on the psychometric properties of PROs. If documented at all, most PRO validation articles give little for the reader to evaluate the content validity of the measures and the credibility and trustworthiness of the methods used to develop them. Increasingly, however, scientists and authorities want to be assured that PRO items and scales have meaning and relevance to subjects. This article was developed by an international, interdisciplinary group of psychologists, psychometricians, regulatory experts, a physician, and a sociologist. It presents rigorous and appropriate qualitative research methods for developing PROs with content validity. The approach described combines an overarching phenomenological theoretical framework with grounded theory data collection and analysis methods to yield PRO items and scales that have content validity.
PRO development: rigorous qualitative research as the crucial foundation
Marquis, Patrick; Vigneux, Marc; Abetz, Linda; Arnould, Benoit; Bayliss, Martha; Crawford, Bruce; Rosa, Kathleen
2010-01-01
Recently published articles have described criteria to assess qualitative research in the health field in general, but very few articles have delineated qualitative methods to be used in the development of Patient-Reported Outcomes (PROs). In fact, how PROs are developed with subject input through focus groups and interviews has been given relatively short shrift in the PRO literature when compared to the plethora of quantitative articles on the psychometric properties of PROs. If documented at all, most PRO validation articles give little for the reader to evaluate the content validity of the measures and the credibility and trustworthiness of the methods used to develop them. Increasingly, however, scientists and authorities want to be assured that PRO items and scales have meaning and relevance to subjects. This article was developed by an international, interdisciplinary group of psychologists, psychometricians, regulatory experts, a physician, and a sociologist. It presents rigorous and appropriate qualitative research methods for developing PROs with content validity. The approach described combines an overarching phenomenological theoretical framework with grounded theory data collection and analysis methods to yield PRO items and scales that have content validity. PMID:20512662
Pediatric Issues in Sports Concussions
Giza, Christopher C.
2014-01-01
Purpose of Review: Sports-related concussions are receiving increasing attention in both the lay press and medical literature. While most media attention has been on high-profile collegiate or professional athletes, the vast majority of individuals participating in contact and collision sports are adolescents and children. This review provides a practical approach toward youth sports-related concussion with a foundation in the recent guidelines, but including specific considerations when applying these management principles to children and adolescents. Recent Findings: Objective measurement of early signs and symptoms is challenging in younger patients, and many commonly used assessment tools await rigorous validation for younger patients. Excellent evidence-based guidelines exist for CT evaluation of mild traumatic brain injury presenting to the emergency department. Evidence suggests that recovery from sports-related concussion takes longer in high school athletes compared with collegiate or professionals; however, rigorous studies below high school age are still lacking. Summary: Proper care for concussion in youth requires a delicate balance of clinical skills, age-appropriate assessment, and individualized management to achieve optimal outcomes. PMID:25470161
Surrogate endpoints in randomized cardiovascular clinical trials.
Domanski, Michael; Pocock, Stuart; Bernaud, Corine; Borer, Jeffrey; Geller, Nancy; Revkin, James; Zannad, Faiez
2011-08-01
Surrogate endpoints predict the occurrence and timing of a clinical endpoint of interest (CEI). Substitution of a surrogate endpoint for a CEI can dramatically reduce the time and cost necessary to complete a Phase III clinical trial. However, assurance that use of a surrogate endpoint will result in a correct conclusion regarding treatment effect on a CEI requires prior rigorous validation of the surrogate. Surrogate endpoints can also be of substantial use in Phase I and II studies to assess whether the intended therapeutic pathway is operative, thus providing assurance regarding the reasonableness of proceeding to a Phase III trial. This paper discusses the uses and validation of surrogate endpoints. © 2010 The Authors Fundamental and Clinical Pharmacology © 2010 Société Française de Pharmacologie et de Thérapeutique.
Increasing rigor in NMR-based metabolomics through validated and open source tools
Eghbalnia, Hamid R; Romero, Pedro R; Westler, William M; Baskaran, Kumaran; Ulrich, Eldon L; Markley, John L
2016-01-01
The metabolome, the collection of small molecules associated with an organism, is a growing subject of inquiry, with the data utilized for data-intensive systems biology, disease diagnostics, biomarker discovery, and the broader characterization of small molecules in mixtures. Owing to their close proximity to the functional endpoints that govern an organism’s phenotype, metabolites are highly informative about functional states. The field of metabolomics identifies and quantifies endogenous and exogenous metabolites in biological samples. Information acquired from nuclear magnetic spectroscopy (NMR), mass spectrometry (MS), and the published literature, as processed by statistical approaches, are driving increasingly wider applications of metabolomics. This review focuses on the role of databases and software tools in advancing the rigor, robustness, reproducibility, and validation of metabolomics studies. PMID:27643760
Increasing rigor in NMR-based metabolomics through validated and open source tools.
Eghbalnia, Hamid R; Romero, Pedro R; Westler, William M; Baskaran, Kumaran; Ulrich, Eldon L; Markley, John L
2017-02-01
The metabolome, the collection of small molecules associated with an organism, is a growing subject of inquiry, with the data utilized for data-intensive systems biology, disease diagnostics, biomarker discovery, and the broader characterization of small molecules in mixtures. Owing to their close proximity to the functional endpoints that govern an organism's phenotype, metabolites are highly informative about functional states. The field of metabolomics identifies and quantifies endogenous and exogenous metabolites in biological samples. Information acquired from nuclear magnetic spectroscopy (NMR), mass spectrometry (MS), and the published literature, as processed by statistical approaches, are driving increasingly wider applications of metabolomics. This review focuses on the role of databases and software tools in advancing the rigor, robustness, reproducibility, and validation of metabolomics studies. Copyright © 2016. Published by Elsevier Ltd.
Rigor Made Easy: Getting Started
ERIC Educational Resources Information Center
Blackburn, Barbara R.
2012-01-01
Bestselling author and noted rigor expert Barbara Blackburn shares the secrets to getting started, maintaining momentum, and reaching your goals. Learn what rigor looks like in the classroom, understand what it means for your students, and get the keys to successful implementation. Learn how to use rigor to raise expectations, provide appropriate…
NASA Technical Reports Server (NTRS)
Pulkkinen, A.; Rastaetter, L.; Kuznetsova, M.; Singer, H.; Balch, C.; Weimer, D.; Toth, G.; Ridley, A.; Gombosi, T.; Wiltberger, M.;
2013-01-01
In this paper we continue the community-wide rigorous modern space weather model validation efforts carried out within GEM, CEDAR and SHINE programs. In this particular effort, in coordination among the Community Coordinated Modeling Center (CCMC), NOAA Space Weather Prediction Center (SWPC), modelers, and science community, we focus on studying the models' capability to reproduce observed ground magnetic field fluctuations, which are closely related to geomagnetically induced current phenomenon. One of the primary motivations of the work is to support NOAA SWPC in their selection of the next numerical model that will be transitioned into operations. Six geomagnetic events and 12 geomagnetic observatories were selected for validation.While modeled and observed magnetic field time series are available for all 12 stations, the primary metrics analysis is based on six stations that were selected to represent the high-latitude and mid-latitude locations. Events-based analysis and the corresponding contingency tables were built for each event and each station. The elements in the contingency table were then used to calculate Probability of Detection (POD), Probability of False Detection (POFD) and Heidke Skill Score (HSS) for rigorous quantification of the models' performance. In this paper the summary results of the metrics analyses are reported in terms of POD, POFD and HSS. More detailed analyses can be carried out using the event by event contingency tables provided as an online appendix. An online interface built at CCMC and described in the supporting information is also available for more detailed time series analyses.
Validation of Bioreactor and Human-on-a-Chip Devices for Chemical Safety Assessment.
Rebelo, Sofia P; Dehne, Eva-Maria; Brito, Catarina; Horland, Reyk; Alves, Paula M; Marx, Uwe
2016-01-01
Equipment and device qualification and test assay validation in the field of tissue engineered human organs for substance assessment remain formidable tasks with only a few successful examples so far. The hurdles seem to increase with the growing complexity of the biological systems, emulated by the respective models. Controlled single tissue or organ culture in bioreactors improves the organ-specific functions and maintains their phenotypic stability for longer periods of time. The reproducibility attained with bioreactor operations is, per se, an advantage for the validation of safety assessment. Regulatory agencies have gradually altered the validation concept from exhaustive "product" to rigorous and detailed process characterization, valuing reproducibility as a standard for validation. "Human-on-a-chip" technologies applying micro-physiological systems to the in vitro combination of miniaturized human organ equivalents into functional human micro-organisms are nowadays thought to be the most elaborate solution created to date. They target the replacement of the current most complex models-laboratory animals. Therefore, we provide here a road map towards the validation of such "human-on-a-chip" models and qualification of their respective bioreactor and microchip equipment along a path currently used for the respective animal models.
ERIC Educational Resources Information Center
Naidoo, Devika
2014-01-01
This article explores aspects of Pierre Bourdieu's theory of social science research practice and discusses their relevance for enhancing validity when researching the "other." Aspects such as: a relational way of thinking about concepts, epistemology and methodology; the rigorous construction of the object of research; and…
Validation of the ROMI-RIP rough mill simulator
Edward R. Thomas; Urs Buehlmann
2002-01-01
The USDA Forest Service's ROMI-RIP rough mill rip-first simulation program is a popular tool for analyzing rough mill conditions, determining more efficient rough mill practices, and finding optimal lumber board cut-up patterns. However, until now, the results generated by ROMI-RIP have not been rigorously compared to those of an actual rough mill. Validating the...
ERIC Educational Resources Information Center
Acharya, Sushil; Manohar, Priyadarshan; Wu, Peter; Schilling, Walter
2017-01-01
Imparting real world experiences in a software verification and validation (SV&V) course is often a challenge due to the lack of effective active learning tools. This pedagogical requirement is important because graduates are expected to develop software that meets rigorous quality standards in functional and application domains. Realizing the…
Quantifying falsifiability of scientific theories
NASA Astrophysics Data System (ADS)
Nemenman, Ilya
I argue that the notion of falsifiability, a key concept in defining a valid scientific theory, can be quantified using Bayesian Model Selection, which is a standard tool in modern statistics. This relates falsifiability to the quantitative version of the statistical Occam's razor, and allows transforming some long-running arguments about validity of scientific theories from philosophical discussions to rigorous mathematical calculations.
David, Hamilton P; Carey, Cayelan C.; Arvola, Lauri; Arzberger, Peter; Brewer, Carol A.; Cole, Jon J; Gaiser, Evelyn; Hanson, Paul C.; Ibelings, Bas W; Jennings, Eleanor; Kratz, Tim K; Lin, Fang-Pang; McBride, Christopher G.; de Motta Marques, David; Muraoka, Kohji; Nishri, Ami; Qin, Boqiang; Read, Jordan S.; Rose, Kevin C.; Ryder, Elizabeth; Weathers, Kathleen C.; Zhu, Guangwei; Trolle, Dennis; Brookes, Justin D
2014-01-01
A Global Lake Ecological Observatory Network (GLEON; www.gleon.org) has formed to provide a coordinated response to the need for scientific understanding of lake processes, utilising technological advances available from autonomous sensors. The organisation embraces a grassroots approach to engage researchers from varying disciplines, sites spanning geographic and ecological gradients, and novel sensor and cyberinfrastructure to synthesise high-frequency lake data at scales ranging from local to global. The high-frequency data provide a platform to rigorously validate process- based ecological models because model simulation time steps are better aligned with sensor measurements than with lower-frequency, manual samples. Two case studies from Trout Bog, Wisconsin, USA, and Lake Rotoehu, North Island, New Zealand, are presented to demonstrate that in the past, ecological model outputs (e.g., temperature, chlorophyll) have been relatively poorly validated based on a limited number of directly comparable measurements, both in time and space. The case studies demonstrate some of the difficulties of mapping sensor measurements directly to model state variable outputs as well as the opportunities to use deviations between sensor measurements and model simulations to better inform process understanding. Well-validated ecological models provide a mechanism to extrapolate high-frequency sensor data in space and time, thereby potentially creating a fully 3-dimensional simulation of key variables of interest.
Rational selection of training and test sets for the development of validated QSAR models
NASA Astrophysics Data System (ADS)
Golbraikh, Alexander; Shen, Min; Xiao, Zhiyan; Xiao, Yun-De; Lee, Kuo-Hsiung; Tropsha, Alexander
2003-02-01
Quantitative Structure-Activity Relationship (QSAR) models are used increasingly to screen chemical databases and/or virtual chemical libraries for potentially bioactive molecules. These developments emphasize the importance of rigorous model validation to ensure that the models have acceptable predictive power. Using k nearest neighbors ( kNN) variable selection QSAR method for the analysis of several datasets, we have demonstrated recently that the widely accepted leave-one-out (LOO) cross-validated R2 (q2) is an inadequate characteristic to assess the predictive ability of the models [Golbraikh, A., Tropsha, A. Beware of q2! J. Mol. Graphics Mod. 20, 269-276, (2002)]. Herein, we provide additional evidence that there exists no correlation between the values of q 2 for the training set and accuracy of prediction ( R 2) for the test set and argue that this observation is a general property of any QSAR model developed with LOO cross-validation. We suggest that external validation using rationally selected training and test sets provides a means to establish a reliable QSAR model. We propose several approaches to the division of experimental datasets into training and test sets and apply them in QSAR studies of 48 functionalized amino acid anticonvulsants and a series of 157 epipodophyllotoxin derivatives with antitumor activity. We formulate a set of general criteria for the evaluation of predictive power of QSAR models.
All That Glitters: A Glimpse into the Future of Cancer Screening
Developing new screening approaches and rigorously establishing their validity is challenging. Researchers are actively searching for new screening tests that improve the benefits of screening while limiting the harms.
NASA Technical Reports Server (NTRS)
Davis, Brian; Turner, Travis L.; Seelecke, Stefan
2008-01-01
An experimental and numerical investigation into the static and dynamic responses of shape memory alloy hybrid composite (SMAHC) beams is performed to provide quantitative validation of a recently commercialized numerical analysis/design tool for SMAHC structures. The SMAHC beam specimens consist of a composite matrix with embedded pre-strained SMA actuators, which act against the mechanical boundaries of the structure when thermally activated to adaptively stiffen the structure. Numerical results are produced from the numerical model as implemented into the commercial finite element code ABAQUS. A rigorous experimental investigation is undertaken to acquire high fidelity measurements including infrared thermography and projection moire interferometry for full-field temperature and displacement measurements, respectively. High fidelity numerical results are also obtained from the numerical model and include measured parameters, such as geometric imperfection and thermal load. Excellent agreement is achieved between the predicted and measured results of the static and dynamic thermomechanical response, thereby providing quantitative validation of the numerical tool.
DOE Office of Scientific and Technical Information (OSTI.GOV)
None
2015-08-01
Since 1990, the National Renewable Energy Laboratory’s (NREL's) National Wind Technology Center (NWTC) has tested more than 150 wind turbine blades. NWTC researchers can test full-scale and subcomponent articles, conduct data analyses, and provide engineering expertise on best design practices. Structural testing of wind turbine blades enables designers, manufacturers, and owners to validate designs and assess structural performance to specific load conditions. Rigorous structural testing can reveal design and manufacturing problems at an early stage of development that can lead to overall improvements in design and increase system reliability.
Monitoring programs to assess reintroduction efforts: A critical component in recovery
Muths, E.; Dreitz, V.
2008-01-01
Reintroduction is a powerful tool in our conservation toolbox. However, the necessary follow-up, i.e. long-term monitoring, is not commonplace and if instituted may lack rigor. We contend that valid monitoring is possible, even with sparse data. We present a means to monitor based on demographic data and a projection model using the Wyoming toad (Bufo baxten) as an example. Using an iterative process, existing data is built upon gradually such that demographic estimates and subsequent inferences increase in reliability. Reintroduction and defensible monitoring may become increasingly relevant as the outlook for amphibians, especially in tropical regions, continues to deteriorate and emergency collection, captive breeding, and reintroduction become necessary. Rigorous use of appropriate modeling and an adaptive approach can validate the use of reintroduction and substantially increase its value to recovery programs. ?? 2008 Museu de Cie??ncies Naturals.
Testability of evolutionary game dynamics based on experimental economics data
NASA Astrophysics Data System (ADS)
Wang, Yijia; Chen, Xiaojie; Wang, Zhijian
2017-11-01
Understanding the dynamic processes of a real game system requires an appropriate dynamics model, and rigorously testing a dynamics model is nontrivial. In our methodological research, we develop an approach to testing the validity of game dynamics models that considers the dynamic patterns of angular momentum and speed as measurement variables. Using Rock-Paper-Scissors (RPS) games as an example, we illustrate the geometric patterns in the experiment data. We then derive the related theoretical patterns from a series of typical dynamics models. By testing the goodness-of-fit between the experimental and theoretical patterns, we show that the validity of these models can be evaluated quantitatively. Our approach establishes a link between dynamics models and experimental systems, which is, to the best of our knowledge, the most effective and rigorous strategy for ascertaining the testability of evolutionary game dynamics models.
ERIC Educational Resources Information Center
Song, Ji Hoon; Chermack, Thomas J.
2008-01-01
The primary purpose of this research is to adapt the Dimensions of the Learning Organization Questionnaire developed by Watkins and Marsick, and examine its validity and reliability in a Korean context. Although the concept of the learning organization is well established, there has not been rigorous validation of the application of the…
ERIC Educational Resources Information Center
Ashour, Rateb; Khasawneh, Samer; Abu-Alruz, Jamal; Alsharqawi, Subhi
2012-01-01
The primary purpose of this study was to determine the curriculum orientations of pre-service teachers at a university in Jordan. Rigorous translation procedures were utilized to validate an Arabic version of the Curriculum Orientation Inventory (COI) for use in Jordan. The validated COI was administered to a sample of 259 pre-service teachers who…
NASA Astrophysics Data System (ADS)
Bellare, Mihir; Ristenpart, Thomas; Rogaway, Phillip; Stegers, Till
Format-preserving encryption (FPE) encrypts a plaintext of some specified format into a ciphertext of identical format—for example, encrypting a valid credit-card number into a valid credit-card number. The problem has been known for some time, but it has lacked a fully general and rigorous treatment. We provide one, starting off by formally defining FPE and security goals for it. We investigate the natural approach for achieving FPE on complex domains, the “rank-then-encipher” approach, and explore what it can and cannot do. We describe two flavors of unbalanced Feistel networks that can be used for achieving FPE, and we prove new security results for each. We revisit the cycle-walking approach for enciphering on a non-sparse subset of an encipherable domain, showing that the timing information that may be divulged by cycle walking is not a damaging thing to leak.
Koh, Keumseok; Reno, Rebecca; Hyder, Ayaz
2018-04-01
Recent advances in computing resources have increased interest in systems modeling and population health. While group model building (GMB) has been effectively applied in developing system dynamics models (SD), few studies have used GMB for developing an agent-based model (ABM). This article explores the use of a GMB approach to develop an ABM focused on food insecurity. In our GMB workshops, we modified a set of the standard GMB scripts to develop and validate an ABM in collaboration with local experts and stakeholders. Based on this experience, we learned that GMB is a useful collaborative modeling platform for modelers and community experts to address local population health issues. We also provide suggestions for increasing the use of the GMB approach to develop rigorous, useful, and validated ABMs.
Enhancing patient autonomy through peer review to replace the FDA's rigorous approval process.
Caplan, Arthur
2012-10-01
There may once have been a time when doctors unquestioningly accepted the government's declaration of a drug's effectiveness and when patients unquestioningly accepted the prescriptions of their doctors. That time has passed. Now, information--good and bad--showers from all directions on patients and physicians alike. A filter is needed, and peer review provides the best one. But who or what is this validated information for? Ethically, its primary purpose is to enable patients to make decisions consistent with their values. Providing vetted information in a form that is useful to patients requires an emphasis on comprehensible, comprehensive, trustworthy, verifiable, and transparent communication. The hypothetical comparative effectiveness case study in this month's Health Affairs does not appear to rise to the level that would be helpful to providers or patients.
The Rigor Mortis of Education: Rigor Is Required in a Dying Educational System
ERIC Educational Resources Information Center
Mixon, Jason; Stuart, Jerry
2009-01-01
In an effort to answer the "Educational Call to Arms", our national public schools have turned to Advanced Placement (AP) courses as the predominate vehicle used to address the lack of academic rigor in our public high schools. Advanced Placement is believed by many to provide students with the rigor and work ethic necessary to…
PCA as a practical indicator of OPLS-DA model reliability.
Worley, Bradley; Powers, Robert
Principal Component Analysis (PCA) and Orthogonal Projections to Latent Structures Discriminant Analysis (OPLS-DA) are powerful statistical modeling tools that provide insights into separations between experimental groups based on high-dimensional spectral measurements from NMR, MS or other analytical instrumentation. However, when used without validation, these tools may lead investigators to statistically unreliable conclusions. This danger is especially real for Partial Least Squares (PLS) and OPLS, which aggressively force separations between experimental groups. As a result, OPLS-DA is often used as an alternative method when PCA fails to expose group separation, but this practice is highly dangerous. Without rigorous validation, OPLS-DA can easily yield statistically unreliable group separation. A Monte Carlo analysis of PCA group separations and OPLS-DA cross-validation metrics was performed on NMR datasets with statistically significant separations in scores-space. A linearly increasing amount of Gaussian noise was added to each data matrix followed by the construction and validation of PCA and OPLS-DA models. With increasing added noise, the PCA scores-space distance between groups rapidly decreased and the OPLS-DA cross-validation statistics simultaneously deteriorated. A decrease in correlation between the estimated loadings (added noise) and the true (original) loadings was also observed. While the validity of the OPLS-DA model diminished with increasing added noise, the group separation in scores-space remained basically unaffected. Supported by the results of Monte Carlo analyses of PCA group separations and OPLS-DA cross-validation metrics, we provide practical guidelines and cross-validatory recommendations for reliable inference from PCA and OPLS-DA models.
ERIC Educational Resources Information Center
Ross, Steven M.; Morrison, Jennifer R.
2014-01-01
In a paper published 25 years ago, Ross and Morrison ("Educ Technol Res Dev" 37(1):19-33, 1989) called for a "happy medium" in educational technology research, to be achieved by balancing high rigor of studies (internal validity) with relevance to real-world applications (external validity). In this paper, we argue that,…
Covariance propagation in spectral indices
Griffin, P. J.
2015-01-09
In this study, the dosimetry community has a history of using spectral indices to support neutron spectrum characterization and cross section validation efforts. An important aspect to this type of analysis is the proper consideration of the contribution of the spectrum uncertainty to the total uncertainty in calculated spectral indices (SIs). This study identifies deficiencies in the traditional treatment of the SI uncertainty, provides simple bounds to the spectral component in the SI uncertainty estimates, verifies that these estimates are reflected in actual applications, details a methodology that rigorously captures the spectral contribution to the uncertainty in the SI, andmore » provides quantified examples that demonstrate the importance of the proper treatment the spectral contribution to the uncertainty in the SI.« less
Klassen, Anne F; Riff, Karen Wy Wong; Longmire, Natasha M; Albert, Asteria; Allen, Gregory C; Aydin, Mustafa Asim; Baker, Stephen B; Cano, Stefan J; Chan, Andrew J; Courtemanche, Douglas J; Dreise, Marieke M; Goldstein, Jesse A; Goodacre, Timothy E E; Harman, Karen E; Munill, Montserrat; Mahony, Aisling O; Aguilera, Mirta Palomares; Peterson, Petra; Pusic, Andrea L; Slator, Rona; Stiernman, Mia; Tsangaris, Elena; Tholpady, Sunil S; Vargas, Federico; Forrest, Christopher R
2018-04-16
Patients with cleft lip and/or palate can undergo numerous procedures to improve appearance, speech, dentition and hearing. We developed a cleft-specific patient-reported outcome instrument to facilitate rigorous international measurement and benchmarking. Data were collected from patients aged 8-29 years with cleft lip and/or palate at 30 hospitals in 12 countries between October 2014 and November 2016. Rasch measurement theory analysis was used to refine the scales and to examine reliability and validity. Normative CLEFT-Q values were computed for age, sex and cleft type. Analysis led to the refinement of an eating and drinking checklist and 12 scales measuring appearance (of the face, nose, nostrils, teeth, lips, jaws and cleft lip scar), health-related quality of life (psychological, social, school, speech distress) and speech function. All scales met the requirements of the Rasch model. Analysis to explore differential item functioning by age, sex and country provided evidence to support the use of a common scoring algorithm for each scale for international use. Lower (worse) scores on CLEFT-Q scales were associated with having a speech problem, being unhappy with facial appearance, and needing future cleft-related treatments, providing evidence of construct validity. Normative values for age, sex and cleft type showed poorer outcomes associated with older age, female sex and having a visible cleft. The CLEFT-Q represents a rigorously developed instrument that can be used internationally to collect and compare evidence-based outcomes data from patients aged 8-29 years of age with cleft lip and/or palate. © 2018 Joule Inc. or its licensors.
Riff, Karen WY Wong; Longmire, Natasha M.; Albert, Asteria; Allen, Gregory C.; Aydin, Mustafa Asim; Baker, Stephen B.; Cano, Stefan J.; Chan, Andrew J.; Courtemanche, Douglas J.; Dreise, Marieke M.; Goldstein, Jesse A.; Goodacre, Timothy E.E.; Harman, Karen E.; Munill, Montserrat; Mahony, Aisling O.; Aguilera, Mirta Palomares; Peterson, Petra; Pusic, Andrea L.; Slator, Rona; Stiernman, Mia; Tsangaris, Elena; Tholpady, Sunil S.; Vargas, Federico; Forrest, Christopher R.
2018-01-01
BACKGROUND: Patients with cleft lip and/or palate can undergo numerous procedures to improve appearance, speech, dentition and hearing. We developed a cleft-specific patient-reported outcome instrument to facilitate rigorous international measurement and benchmarking. METHODS: Data were collected from patients aged 8–29 years with cleft lip and/or palate at 30 hospitals in 12 countries between October 2014 and November 2016. Rasch measurement theory analysis was used to refine the scales and to examine reliability and validity. Normative CLEFT-Q values were computed for age, sex and cleft type. RESULTS: Analysis led to the refinement of an eating and drinking checklist and 12 scales measuring appearance (of the face, nose, nostrils, teeth, lips, jaws and cleft lip scar), health-related quality of life (psychological, social, school, speech distress) and speech function. All scales met the requirements of the Rasch model. Analysis to explore differential item functioning by age, sex and country provided evidence to support the use of a common scoring algorithm for each scale for international use. Lower (worse) scores on CLEFT-Q scales were associated with having a speech problem, being unhappy with facial appearance, and needing future cleft-related treatments, providing evidence of construct validity. Normative values for age, sex and cleft type showed poorer outcomes associated with older age, female sex and having a visible cleft. INTERPRETATION: The CLEFT-Q represents a rigorously developed instrument that can be used internationally to collect and compare evidence-based outcomes data from patients aged 8–29 years of age with cleft lip and/or palate. PMID:29661814
Instrument Selection for Randomized Controlled Trials Why This and Not That?
Records, Kathie; Keller, Colleen; Ainsworth, Barbara; Permana, Paska
2011-01-01
A fundamental linchpin for obtaining rigorous findings in quantitative research involves the selection of survey instruments. Psychometric recommendations are available for the processes for scale development and testing and guidance for selection of established scales. These processes are necessary to address the validity link between the phenomena under investigation, the empirical measures and, ultimately, the theoretical ties between these and the world views of the participants. Detailed information is most often provided about study design and protocols, but far less frequently is a detailed theoretical explanation provided for why specific instruments are chosen. Guidance to inform choices is often difficult to find when scales are needed for specific cultural, ethnic, or racial groups. This paper details the rationale underlying instrument selection for measurement of the major processes (intervention, mediator and moderator variables, outcome variables) in an ongoing study of postpartum Latinas, Madres para la Salud [Mothers for Health]. The rationale underpinning our choices includes a discussion of alternatives, when appropriate. These exemplars may provide direction for other intervention researchers who are working with specific cultural, racial, or ethnic groups or for other investigators who are seeking to select the ‘best’ instrument. Thoughtful consideration of measurement and articulation of the rationale underlying our choices facilitates the maintenance of rigor within the study design and improves our ability to assess study outcomes. PMID:21986392
NASA Technical Reports Server (NTRS)
Jenkins, Michael G.; Salem, Jonathan A.
2016-01-01
Physical and mechanical properties and performance of advanced ceramics and glasses are difficult to measure correctly without the proper techniques. For over three decades, ASTM Committee C28 on Advanced Ceramics, has developed high-quality, technically-rigorous, full-consensus standards (e.g., test methods, practices, guides, terminology) to measure properties and performance of monolithic and composite ceramics that may be applied to glasses in some cases. These standards contain testing particulars for many mechanical, physical, thermal, properties and performance of these materials. As a result these standards are used to generate accurate, reliable, repeatable and complete data. Within Committee C28, users, producers, researchers, designers, academicians, etc. have written, continually updated, and validated through round-robin test programs, 50 standards since the Committee's founding in 1986. This paper provides a detailed retrospective of the 30 years of ASTM Committee C28 including a graphical pictogram listing of C28 standards along with examples of the tangible benefits of standards for advanced ceramics to demonstrate their practical applications.
IEEE/NASA Workshop on Leveraging Applications of Formal Methods, Verification, and Validation
NASA Technical Reports Server (NTRS)
Margaria, Tiziana (Editor); Steffen, Bernhard (Editor); Hichey, Michael G.
2005-01-01
This volume contains the Preliminary Proceedings of the 2005 IEEE ISoLA Workshop on Leveraging Applications of Formal Methods, Verification, and Validation, with a special track on the theme of Formal Methods in Human and Robotic Space Exploration. The workshop was held on 23-24 September 2005 at the Loyola College Graduate Center, Columbia, MD, USA. The idea behind the Workshop arose from the experience and feedback of ISoLA 2004, the 1st International Symposium on Leveraging Applications of Formal Methods held in Paphos (Cyprus) last October-November. ISoLA 2004 served the need of providing a forum for developers, users, and researchers to discuss issues related to the adoption and use of rigorous tools and methods for the specification, analysis, verification, certification, construction, test, and maintenance of systems from the point of view of their different application domains.
Monga, Ash K; Tracey, Michael R; Subbaroyan, Jeyakumar
2012-08-01
The aim of this manuscript was to provide a systematic literature review of clinical trial evidence for a range of electrical stimulation therapies in the treatment of lower urinary tract symptoms (LUTS). The databases MEDLINE, BIOSIS Previews, Inside Conferences, and EMBASE were searched. Original clinical studies with greater than 15 subjects were included. Seventy-three studies were included, representing implanted sacral nerve stimulation (SNS), percutaneous posterior tibial nerve stimulation (PTNS), and transcutaneous electrical stimulation (TENS) therapy modalities. Median mean reductions in incontinence episodes and voiding frequency were similar for implanted SNS and PTNS. However, long-term follow-up data to validate the sustained benefit of PTNS are lacking. Despite a substantial body of research devoted to SNS validation, it is not possible to definitively define the appropriate role of this therapy owing largely to study design flaws that inhibited rigorous intention to treat analyses for the majority of these studies.
Behavior and emotional disturbance in Prader-Willi syndrome.
Einfeld, S L; Smith, A; Durvasula, S; Florio, T; Tonge, B J
1999-01-15
To determine if persons with the Prader-Willi syndrome (PWS) have increased psychopathology when compared with matched controls, and whether there is a specific behavior phenotype in PWS, the behavior of 46 persons with PWS was compared with that of control individuals derived from a community sample (N = 454) of persons with mental retardation (MR). Behaviors were studied using the Developmental Behaviour Checklist, an instrument of established validity in the evaluation of behavioral disturbance in individuals with MR. PWS subjects were found to be more behaviorally disturbed than controls overall, and especially in antisocial behavior. In addition, some individual behaviors were more common in PWS subjects than controls. When these behaviors are considered together with findings from other studies using acceptably rigorous methods, a consensus behavior phenotype for PWS can be formulated. This will provide a valid foundation for studies of the mechanism of genetic pathogenesis of behavior in PWS.
FDA 2011 process validation guidance: lifecycle compliance model.
Campbell, Cliff
2014-01-01
This article has been written as a contribution to the industry's efforts in migrating from a document-driven to a data-driven compliance mindset. A combination of target product profile, control engineering, and general sum principle techniques is presented as the basis of a simple but scalable lifecycle compliance model in support of modernized process validation. Unit operations and significant variables occupy pole position within the model, documentation requirements being treated as a derivative or consequence of the modeling process. The quality system is repositioned as a subordinate of system quality, this being defined as the integral of related "system qualities". The article represents a structured interpretation of the U.S. Food and Drug Administration's 2011 Guidance for Industry on Process Validation and is based on the author's educational background and his manufacturing/consulting experience in the validation field. The U.S. Food and Drug Administration's Guidance for Industry on Process Validation (2011) provides a wide-ranging and rigorous outline of compliant drug manufacturing requirements relative to its 20(th) century predecessor (1987). Its declared focus is patient safety, and it identifies three inter-related (and obvious) stages of the compliance lifecycle. Firstly, processes must be designed, both from a technical and quality perspective. Secondly, processes must be qualified, providing evidence that the manufacturing facility is fully "roadworthy" and fit for its intended purpose. Thirdly, processes must be verified, meaning that commercial batches must be monitored to ensure that processes remain in a state of control throughout their lifetime.
Rigor and Responsiveness in Classroom Activity
ERIC Educational Resources Information Center
Thomspon, Jessica; Hagenah, Sara; Kang, Hosun; Stroupe, David; Braaten, Melissa; Colley, Carolyn; Windschitl, Mark
2016-01-01
Background/Context: There are few examples from classrooms or the literature that provide a clear vision of teaching that simultaneously promotes rigorous disciplinary activity and is responsive to all students. Maintaining rigorous and equitable classroom discourse is a worthy goal, yet there is no clear consensus of how this actually works in a…
NASA Technical Reports Server (NTRS)
Glytsis, Elias N.; Brundrett, David L.; Gaylord, Thomas K.
1993-01-01
A review of the rigorous coupled-wave analysis as applied to the diffraction of electro-magnetic waves by gratings is presented. The analysis is valid for any polarization, angle of incidence, and conical diffraction. Cascaded and/or multiplexed gratings as well as material anisotropy can be incorporated under the same formalism. Small period rectangular groove gratings can also be modeled using approximately equivalent uniaxial homogeneous layers (effective media). The ordinary and extraordinary refractive indices of these layers depend on the gratings filling factor, the refractive indices of the substrate and superstrate, and the ratio of the freespace wavelength to grating period. Comparisons of the homogeneous effective medium approximations with the rigorous coupled-wave analysis are presented. Antireflection designs (single-layer or multilayer) using the effective medium models are presented and compared. These ultra-short period antireflection gratings can also be used to produce soft x-rays. Comparisons of the rigorous coupled-wave analysis with experimental results on soft x-ray generation by gratings are also included.
Higher order temporal finite element methods through mixed formalisms.
Kim, Jinkyu
2014-01-01
The extended framework of Hamilton's principle and the mixed convolved action principle provide new rigorous weak variational formalism for a broad range of initial boundary value problems in mathematical physics and mechanics. In this paper, their potential when adopting temporally higher order approximations is investigated. The classical single-degree-of-freedom dynamical systems are primarily considered to validate and to investigate the performance of the numerical algorithms developed from both formulations. For the undamped system, all the algorithms are symplectic and unconditionally stable with respect to the time step. For the damped system, they are shown to be accurate with good convergence characteristics.
Nonequilibrium itinerant-electron magnetism: A time-dependent mean-field theory
NASA Astrophysics Data System (ADS)
Secchi, A.; Lichtenstein, A. I.; Katsnelson, M. I.
2016-08-01
We study the dynamical magnetic susceptibility of a strongly correlated electronic system in the presence of a time-dependent hopping field, deriving a generalized Bethe-Salpeter equation that is valid also out of equilibrium. Focusing on the single-orbital Hubbard model within the time-dependent Hartree-Fock approximation, we solve the equation in the nonequilibrium adiabatic regime, obtaining a closed expression for the transverse magnetic susceptibility. From this, we provide a rigorous definition of nonequilibrium (time-dependent) magnon frequencies and exchange parameters, expressed in terms of nonequilibrium single-electron Green's functions and self-energies. In the particular case of equilibrium, we recover previously known results.
NASA Astrophysics Data System (ADS)
Katsoulakis, Markos A.; Vlachos, Dionisios G.
2003-11-01
We derive a hierarchy of successively coarse-grained stochastic processes and associated coarse-grained Monte Carlo (CGMC) algorithms directly from the microscopic processes as approximations in larger length scales for the case of diffusion of interacting particles on a lattice. This hierarchy of models spans length scales between microscopic and mesoscopic, satisfies a detailed balance, and gives self-consistent fluctuation mechanisms whose noise is asymptotically identical to the microscopic MC. Rigorous, detailed asymptotics justify and clarify these connections. Gradient continuous time microscopic MC and CGMC simulations are compared under far from equilibrium conditions to illustrate the validity of our theory and delineate the errors obtained by rigorous asymptotics. Information theory estimates are employed for the first time to provide rigorous error estimates between the solutions of microscopic MC and CGMC, describing the loss of information during the coarse-graining process. Simulations under periodic boundary conditions are used to verify the information theory error estimates. It is shown that coarse-graining in space leads also to coarse-graining in time by q2, where q is the level of coarse-graining, and overcomes in part the hydrodynamic slowdown. Operation counting and CGMC simulations demonstrate significant CPU savings in continuous time MC simulations that vary from q3 for short potentials to q4 for long potentials. Finally, connections of the new coarse-grained stochastic processes to stochastic mesoscopic and Cahn-Hilliard-Cook models are made.
Treatment of Neuromyelitis Optica: Review and Recommendations
Kimbrough, Dorlan J; Fujihara, Kazuo; Jacob, Anu; Lana-Peixoto, Marco A; Leite, Maria Isabel; Levy, Michael; Marignier, Romain; Nakashima, Ichiro; Palace, Jacqueline; de Seze, Jérôme; Stuve, Olaf; Tenembaum, Silvia N; Traboulsee, Anthony; Waubant, Emmanuelle; Weinshenker, Brian G; Wingerchuk, Dean M
2014-01-01
Neuromyelitis optica (NMO) is an autoimmune demyelinating disease preferentially targeting the optic nerves and spinal cord. Once regarded as a variant of multiple sclerosis (MS), NMO is now recognized to be a different disease with unique pathology and immunopathogenesis that does not respond to traditional MS immunomodulators such as interferons. Preventive therapy in NMO has focused on a range of immunosuppressive medications, none of which have been validated in a rigorous randomized trial. However, multiple retrospective and a few recent prospective studies have provided evidence for the use of six medications for the prevention of NMO exacerbations: azathioprine, rituximab, mycophenolate mofetil, prednisone, methotrexate and mitoxantrone. This review provides a comprehensive analysis of each of these medications in NMO and concludes with a set of recommended consensus practices. PMID:24555176
Chen, Po-Yu
2014-01-01
The validness of the expiration dates (validity period) that manufacturers provide on food product labels is a crucial food safety problem. Governments must study how to use their authority by implementing fair awards and punishments to prompt manufacturers into adopting rigorous considerations, such as the effect of adopting new storage methods for extending product validity periods on expected costs. Assuming that a manufacturer sells fresh food or drugs, this manufacturer must respond to current stochastic demands at each unit of time to determine the purchase amount of products for sale. If this decision maker is capable and an opportunity arises, new packaging methods (e.g., aluminum foil packaging, vacuum packaging, high-temperature sterilization after glass packaging, or packaging with various degrees of dryness) or storage methods (i.e., adding desiccants or various antioxidants) can be chosen to extend the validity periods of products. To minimize expected costs, this decision maker must be aware of the processing costs of new storage methods, inventory standards, inventory cycle lengths, and changes in relationships between factors such as stochastic demand functions in a cycle. Based on these changes in relationships, this study established a mathematical model as a basis for discussing the aforementioned topics.
Pavell, Anthony; Hughes, Keith A
2010-01-01
This article describes a method for achieving the load equivalence model, described in Parenteral Drug Association Technical Report 1, using a mass-based approach. The item and load bracketing approach allows for mixed equipment load size variation for operational flexibility along with decreased time to introduce new items to the operation. The article discusses the utilization of approximately 67 items/components (Table IV) identified for routine sterilization with varying quantities required weekly. The items were assessed for worst-case identification using four temperature-related criteria. The criteria were used to provide a data-based identification of worst-case items, and/or item equivalence, to carry forward into cycle validation using a variable load pattern. The mass approach to maximum load determination was used to bracket routine production use and allows for variable loading patterns. The result of the item mapping and load bracketing data is "a proven acceptable range" of sterilizing conditions including loading configuration and location. The application of these approaches, while initially more time/test-intensive than alternate approaches, provides a method of cycle validation with long-term benefit of ease of ongoing qualification, minimizing time and requirements for new equipment qualification for similar loads/use, and for rapid and rigorous assessment of new items for sterilization.
Buschmann, Dominik; Haberberger, Anna; Kirchner, Benedikt; Spornraft, Melanie; Riedmaier, Irmgard; Schelling, Gustav; Pfaffl, Michael W.
2016-01-01
Small RNA-Seq has emerged as a powerful tool in transcriptomics, gene expression profiling and biomarker discovery. Sequencing cell-free nucleic acids, particularly microRNA (miRNA), from liquid biopsies additionally provides exciting possibilities for molecular diagnostics, and might help establish disease-specific biomarker signatures. The complexity of the small RNA-Seq workflow, however, bears challenges and biases that researchers need to be aware of in order to generate high-quality data. Rigorous standardization and extensive validation are required to guarantee reliability, reproducibility and comparability of research findings. Hypotheses based on flawed experimental conditions can be inconsistent and even misleading. Comparable to the well-established MIQE guidelines for qPCR experiments, this work aims at establishing guidelines for experimental design and pre-analytical sample processing, standardization of library preparation and sequencing reactions, as well as facilitating data analysis. We highlight bottlenecks in small RNA-Seq experiments, point out the importance of stringent quality control and validation, and provide a primer for differential expression analysis and biomarker discovery. Following our recommendations will encourage better sequencing practice, increase experimental transparency and lead to more reproducible small RNA-Seq results. This will ultimately enhance the validity of biomarker signatures, and allow reliable and robust clinical predictions. PMID:27317696
Utility of distributed hydrologic and water quality models for watershed management and sustainability studies should be accompanied by rigorous model uncertainty analysis. However, the use of complex watershed models primarily follows the traditional {calibrate/validate/predict}...
[Experimental study of restiffening of the rigor mortis].
Wang, X; Li, M; Liao, Z G; Yi, X F; Peng, X M
2001-11-01
To observe changes of the length of sarcomere of rat when restiffening. We measured the length of sarcomere of quadriceps in 40 rats in different condition by scanning electron microscope. The length of sarcomere of rigor mortis without destroy is obviously shorter than that of restiffening. The length of sarcomere is negatively correlative to the intensity of rigor mortis. Measuring the length of sarcomere can determine the intensity of rigor mortis and provide evidence for estimation of time since death.
Validity of Cognitive Load Measures in Simulation-Based Training: A Systematic Review.
Naismith, Laura M; Cavalcanti, Rodrigo B
2015-11-01
Cognitive load theory (CLT) provides a rich framework to inform instructional design. Despite the applicability of CLT to simulation-based medical training, findings from multimedia learning have not been consistently replicated in this context. This lack of transferability may be related to issues in measuring cognitive load (CL) during simulation. The authors conducted a review of CLT studies across simulation training contexts to assess the validity evidence for different CL measures. PRISMA standards were followed. For 48 studies selected from a search of MEDLINE, EMBASE, PsycInfo, CINAHL, and ERIC databases, information was extracted about study aims, methods, validity evidence of measures, and findings. Studies were categorized on the basis of findings and prevalence of validity evidence collected, and statistical comparisons between measurement types and research domains were pursued. CL during simulation training has been measured in diverse populations including medical trainees, pilots, and university students. Most studies (71%; 34) used self-report measures; others included secondary task performance, physiological indices, and observer ratings. Correlations between CL and learning varied from positive to negative. Overall validity evidence for CL measures was low (mean score 1.55/5). Studies reporting greater validity evidence were more likely to report that high CL impaired learning. The authors found evidence that inconsistent correlations between CL and learning may be related to issues of validity in CL measures. Further research would benefit from rigorous documentation of validity and from triangulating measures of CL. This can better inform CLT instructional design for simulation-based medical training.
NASA Astrophysics Data System (ADS)
Benguria, Rafael D.; Depassier, M. Cristina; Loss, Michael
2012-12-01
We study the effect of a cutoff on the speed of pulled fronts of the one-dimensional reaction diffusion equation. To accomplish this, we first use variational techniques to prove the existence of a heteroclinic orbit in phase space for traveling wave solutions of the corresponding reaction diffusion equation under conditions that include discontinuous reaction profiles. This existence result allows us to prove rigorous upper and lower bounds on the minimal speed of monotonic fronts in terms of the cut-off parameter ɛ. From these bounds we estimate the range of validity of the Brunet-Derrida formula for a general class of reaction terms.
NASA Astrophysics Data System (ADS)
Jaboulay, Jean-Charles; Brun, Emeric; Hugot, François-Xavier; Huynh, Tan-Dat; Malouch, Fadhel; Mancusi, Davide; Tsilanizara, Aime
2017-09-01
After fission or fusion reactor shutdown the activated structure emits decay photons. For maintenance operations the radiation dose map must be established in the reactor building. Several calculation schemes have been developed to calculate the shutdown dose rate. These schemes are widely developed in fusion application and more precisely for the ITER tokamak. This paper presents the rigorous-two-steps scheme implemented at CEA. It is based on the TRIPOLI-4® Monte Carlo code and the inventory code MENDEL. The ITER shutdown dose rate benchmark has been carried out, results are in a good agreement with the other participant.
Formative Assessment as Mediation
ERIC Educational Resources Information Center
De Vos, Mark; Belluigi, Dina Zoe
2011-01-01
Whilst principles of validity, reliability and fairness should be central concerns for the assessment of student learning in higher education, simplistic notions of "transparency" and "explicitness" in terms of assessment criteria should be critiqued more rigorously. This article examines the inherent tensions resulting from CRA's links to both…
Accurate Biomass Estimation via Bayesian Adaptive Sampling
NASA Technical Reports Server (NTRS)
Wheeler, Kevin R.; Knuth, Kevin H.; Castle, Joseph P.; Lvov, Nikolay
2005-01-01
The following concepts were introduced: a) Bayesian adaptive sampling for solving biomass estimation; b) Characterization of MISR Rahman model parameters conditioned upon MODIS landcover. c) Rigorous non-parametric Bayesian approach to analytic mixture model determination. d) Unique U.S. asset for science product validation and verification.
Elder Abuse: Global Situation, Risk Factors, and Prevention Strategies
Pillemer, Karl; Burnes, David; Riffin, Catherine; Lachs, Mark S.
2016-01-01
Purpose: Elder mistreatment is now recognized internationally as a pervasive and growing problem, urgently requiring the attention of health care systems, social welfare agencies, policymakers, and the general public. In this article, we provide an overview of global issues in the field of elder abuse, with a focus on prevention. Design and Methods: This article provides a scoping review of key issues in the field from an international perspective. Results: By drawing primarily on population-based studies, this scoping review provided a more valid and reliable synthesis of current knowledge about prevalence and risk factors than has been available. Despite the lack of scientifically rigorous intervention research on elder abuse, the review also identified 5 promising strategies for prevention. Implications: The findings highlight a growing consensus across studies regarding the extent and causes of elder mistreatment, as well as the urgent need for efforts to make elder mistreatment prevention programs more effective and evidence based. PMID:26994260
The International Arctic Buoy Programme (IABP)
NASA Astrophysics Data System (ADS)
Rigor, I. G.; Ortmeyer, M.
2003-12-01
The Arctic has undergone dramatic changes in weather, climate and environment. It should be noted that many of these changes were first observed and studied using data from the International Arctic Buoy Programme (IABP). For example, IABP data were fundamental to Walsh et al. (1996) showing that atmospheric pressure has decreased, Rigor et al. (2000) showing that air temperatures have increased, and to Proshutinsky and Johnson (1997); Steele and Boyd, (1998); Kwok, (2000); and Rigor et al. (2002) showing that the clockwise circulation of sea ice and the ocean has weakened. All these results relied heavily on data from the IABP. In addition to supporting these studies of climate change, the IABP observations are also used to forecast weather and ice conditions, validate satellite retrievals of environmental variables, to force, validate and initialize numerical models. Over 350 papers have been written using data from the IABP. The observations and datasets of the IABP data are one of the cornerstones for environmental forecasting and research in the Arctic.
Coverage Metrics for Requirements-Based Testing: Evaluation of Effectiveness
NASA Technical Reports Server (NTRS)
Staats, Matt; Whalen, Michael W.; Heindahl, Mats P. E.; Rajan, Ajitha
2010-01-01
In black-box testing, the tester creates a set of tests to exercise a system under test without regard to the internal structure of the system. Generally, no objective metric is used to measure the adequacy of black-box tests. In recent work, we have proposed three requirements coverage metrics, allowing testers to objectively measure the adequacy of a black-box test suite with respect to a set of requirements formalized as Linear Temporal Logic (LTL) properties. In this report, we evaluate the effectiveness of these coverage metrics with respect to fault finding. Specifically, we conduct an empirical study to investigate two questions: (1) do test suites satisfying a requirements coverage metric provide better fault finding than randomly generated test suites of approximately the same size?, and (2) do test suites satisfying a more rigorous requirements coverage metric provide better fault finding than test suites satisfying a less rigorous requirements coverage metric? Our results indicate (1) only one coverage metric proposed -- Unique First Cause (UFC) coverage -- is sufficiently rigorous to ensure test suites satisfying the metric outperform randomly generated test suites of similar size and (2) that test suites satisfying more rigorous coverage metrics provide better fault finding than test suites satisfying less rigorous coverage metrics.
Lyon, Aaron R; Connors, Elizabeth; Jensen-Doss, Amanda; Landes, Sara J; Lewis, Cara C; McLeod, Bryce D; Rutt, Christopher; Stanick, Cameo; Weiner, Bryan J
2017-09-01
The advancement of implementation science is dependent on identifying assessment strategies that can address implementation and clinical outcome variables in ways that are valid, relevant to stakeholders, and scalable. This paper presents a measurement agenda for implementation science that integrates the previously disparate assessment traditions of idiographic and nomothetic approaches. Although idiographic and nomothetic approaches are both used in implementation science, a review of the literature on this topic suggests that their selection can be indiscriminate, driven by convenience, and not explicitly tied to research study design. As a result, they are not typically combined deliberately or effectively. Thoughtful integration may simultaneously enhance both the rigor and relevance of assessments across multiple levels within health service systems. Background on nomothetic and idiographic assessment is provided as well as their potential to support research in implementation science. Drawing from an existing framework, seven structures (of various sequencing and weighting options) and five functions (Convergence, Complementarity, Expansion, Development, Sampling) for integrating conceptually distinct research methods are articulated as they apply to the deliberate, design-driven integration of nomothetic and idiographic assessment approaches. Specific examples and practical guidance are provided to inform research consistent with this framework. Selection and integration of idiographic and nomothetic assessments for implementation science research designs can be improved. The current paper argues for the deliberate application of a clear framework to improve the rigor and relevance of contemporary assessment strategies.
Current to the ionosphere following a lightning stroke
NASA Technical Reports Server (NTRS)
Hale, L. C.; Baginski, M. E.
1987-01-01
A simple analytical expression for calculating the total current waveform to the ionosphere after a lightning stroke is derived. The validity of this expression is demonstrated by comparison with a more rigorous computer solution of Maxwell's equations. The analytic model demonstrates that the temporal variation of the current induced in the ionosphere and global circuit and the corresponding return current in the earth depends on the conductivity profile at intervening altitudes in the middle atmosphere. A conclusion is that capacitative coupling may provide tighter coupling between the lower atmosphere and the ionosphere than usually considered, in both directions, which may help to explain observations which seem to indicate that magnetospheric phenomena may in some instances trigger lightning.
NASA-STD-7009 Guidance Document for Human Health and Performance Models and Simulations
NASA Technical Reports Server (NTRS)
Walton, Marlei; Mulugeta, Lealem; Nelson, Emily S.; Myers, Jerry G.
2014-01-01
Rigorous verification, validation, and credibility (VVC) processes are imperative to ensure that models and simulations (MS) are sufficiently reliable to address issues within their intended scope. The NASA standard for MS, NASA-STD-7009 (7009) [1] was a resultant outcome of the Columbia Accident Investigation Board (CAIB) to ensure MS are developed, applied, and interpreted appropriately for making decisions that may impact crew or mission safety. Because the 7009 focus is engineering systems, a NASA-STD-7009 Guidance Document is being developed to augment the 7009 and provide information, tools, and techniques applicable to the probabilistic and deterministic biological MS more prevalent in human health and performance (HHP) and space biomedical research and operations.
Rost, Christina M.; Sachet, Edward; Borman, Trent; Moballegh, Ali; Dickey, Elizabeth C.; Hou, Dong; Jones, Jacob L.; Curtarolo, Stefano; Maria, Jon-Paul
2015-01-01
Configurational disorder can be compositionally engineered into mixed oxide by populating a single sublattice with many distinct cations. The formulations promote novel and entropy-stabilized forms of crystalline matter where metal cations are incorporated in new ways. Here, through rigorous experiments, a simple thermodynamic model, and a five-component oxide formulation, we demonstrate beyond reasonable doubt that entropy predominates the thermodynamic landscape, and drives a reversible solid-state transformation between a multiphase and single-phase state. In the latter, cation distributions are proven to be random and homogeneous. The findings validate the hypothesis that deliberate configurational disorder provides an orthogonal strategy to imagine and discover new phases of crystalline matter and untapped opportunities for property engineering. PMID:26415623
Handling missing values in the MDS-UPDRS.
Goetz, Christopher G; Luo, Sheng; Wang, Lu; Tilley, Barbara C; LaPelle, Nancy R; Stebbins, Glenn T
2015-10-01
This study was undertaken to define the number of missing values permissible to render valid total scores for each Movement Disorder Society Unified Parkinson's Disease Rating Scale (MDS-UPDRS) part. To handle missing values, imputation strategies serve as guidelines to reject an incomplete rating or create a surrogate score. We tested a rigorous, scale-specific, data-based approach to handling missing values for the MDS-UPDRS. From two large MDS-UPDRS datasets, we sequentially deleted item scores, either consistently (same items) or randomly (different items) across all subjects. Lin's Concordance Correlation Coefficient (CCC) compared scores calculated without missing values with prorated scores based on sequentially increasing missing values. The maximal number of missing values retaining a CCC greater than 0.95 determined the threshold for rendering a valid prorated score. A second confirmatory sample was selected from the MDS-UPDRS international translation program. To provide valid part scores applicable across all Hoehn and Yahr (H&Y) stages when the same items are consistently missing, one missing item from Part I, one from Part II, three from Part III, but none from Part IV can be allowed. To provide valid part scores applicable across all H&Y stages when random item entries are missing, one missing item from Part I, two from Part II, seven from Part III, but none from Part IV can be allowed. All cutoff values were confirmed in the validation sample. These analyses are useful for constructing valid surrogate part scores for MDS-UPDRS when missing items fall within the identified threshold and give scientific justification for rejecting partially completed ratings that fall below the threshold. © 2015 International Parkinson and Movement Disorder Society.
Does condom social marketing improve health outcomes and increase usage and equitable access?
Knerr, Wendy
2011-05-01
Condom social marketing (CSM) has increased condom supplies, broadened commercial markets for condoms and introduced marketing innovations in developing countries. Yet rigorous and reliable evidence of the impact on condom usage and disease prevention is limited, as is evidence of the impact on equity of access to condoms for poor populations, women and people living with HIV. One strand of research on CSM reports mostly on output (e.g. sales and processes) and market growth; but these have been found to be highly unreliable measures of condom usage. Another strand of research reports primarily on changes in sexual behaviour, attitude or condom usage, using survey data. While random sampling is rare, these studies often use representative samples, which provide some measure of validity. There have been attempts to improve the reliability or results to good effect, but challenges remain for researchers, scholars and donors, including the need to supplement output data with measures of behaviour change, use rigorous designs which are built into programmes a priori, report on equity measures, report on potential harms of CSM programmes, and encourage external and systematic reviews. Copyright © 2011 Reproductive Health Matters. Published by Elsevier Ltd. All rights reserved.
Concrete ensemble Kalman filters with rigorous catastrophic filter divergence
Kelly, David; Majda, Andrew J.; Tong, Xin T.
2015-01-01
The ensemble Kalman filter and ensemble square root filters are data assimilation methods used to combine high-dimensional, nonlinear dynamical models with observed data. Ensemble methods are indispensable tools in science and engineering and have enjoyed great success in geophysical sciences, because they allow for computationally cheap low-ensemble-state approximation for extremely high-dimensional turbulent forecast models. From a theoretical perspective, the dynamical properties of these methods are poorly understood. One of the central mysteries is the numerical phenomenon known as catastrophic filter divergence, whereby ensemble-state estimates explode to machine infinity, despite the true state remaining in a bounded region. In this article we provide a breakthrough insight into the phenomenon, by introducing a simple and natural forecast model that transparently exhibits catastrophic filter divergence under all ensemble methods and a large set of initializations. For this model, catastrophic filter divergence is not an artifact of numerical instability, but rather a true dynamical property of the filter. The divergence is not only validated numerically but also proven rigorously. The model cleanly illustrates mechanisms that give rise to catastrophic divergence and confirms intuitive accounts of the phenomena given in past literature. PMID:26261335
Concrete ensemble Kalman filters with rigorous catastrophic filter divergence.
Kelly, David; Majda, Andrew J; Tong, Xin T
2015-08-25
The ensemble Kalman filter and ensemble square root filters are data assimilation methods used to combine high-dimensional, nonlinear dynamical models with observed data. Ensemble methods are indispensable tools in science and engineering and have enjoyed great success in geophysical sciences, because they allow for computationally cheap low-ensemble-state approximation for extremely high-dimensional turbulent forecast models. From a theoretical perspective, the dynamical properties of these methods are poorly understood. One of the central mysteries is the numerical phenomenon known as catastrophic filter divergence, whereby ensemble-state estimates explode to machine infinity, despite the true state remaining in a bounded region. In this article we provide a breakthrough insight into the phenomenon, by introducing a simple and natural forecast model that transparently exhibits catastrophic filter divergence under all ensemble methods and a large set of initializations. For this model, catastrophic filter divergence is not an artifact of numerical instability, but rather a true dynamical property of the filter. The divergence is not only validated numerically but also proven rigorously. The model cleanly illustrates mechanisms that give rise to catastrophic divergence and confirms intuitive accounts of the phenomena given in past literature.
Aerial photography flight quality assessment with GPS/INS and DEM data
NASA Astrophysics Data System (ADS)
Zhao, Haitao; Zhang, Bing; Shang, Jiali; Liu, Jiangui; Li, Dong; Chen, Yanyan; Zuo, Zhengli; Chen, Zhengchao
2018-01-01
The flight altitude, ground coverage, photo overlap, and other acquisition specifications of an aerial photography flight mission directly affect the quality and accuracy of the subsequent mapping tasks. To ensure smooth post-flight data processing and fulfill the pre-defined mapping accuracy, flight quality assessments should be carried out in time. This paper presents a novel and rigorous approach for flight quality evaluation of frame cameras with GPS/INS data and DEM, using geometric calculation rather than image analysis as in the conventional methods. This new approach is based mainly on the collinearity equations, in which the accuracy of a set of flight quality indicators is derived through a rigorous error propagation model and validated with scenario data. Theoretical analysis and practical flight test of an aerial photography mission using an UltraCamXp camera showed that the calculated photo overlap is accurate enough for flight quality assessment of 5 cm ground sample distance image, using the SRTMGL3 DEM and the POSAV510 GPS/INS data. An even better overlap accuracy could be achieved for coarser-resolution aerial photography. With this new approach, the flight quality evaluation can be conducted on site right after landing, providing accurate and timely information for decision making.
A surface spherical harmonic expansion of gravity anomalies on the ellipsoid
NASA Astrophysics Data System (ADS)
Claessens, S. J.; Hirt, C.
2015-10-01
A surface spherical harmonic expansion of gravity anomalies with respect to a geodetic reference ellipsoid can be used to model the global gravity field and reveal its spectral properties. In this paper, a direct and rigorous transformation between solid spherical harmonic coefficients of the Earth's disturbing potential and surface spherical harmonic coefficients of gravity anomalies in ellipsoidal approximation with respect to a reference ellipsoid is derived. This transformation cannot rigorously be achieved by the Hotine-Jekeli transformation between spherical and ellipsoidal harmonic coefficients. The method derived here is used to create a surface spherical harmonic model of gravity anomalies with respect to the GRS80 ellipsoid from the EGM2008 global gravity model. Internal validation of the model shows a global RMS precision of 1 nGal. This is significantly more precise than previous solutions based on spherical approximation or approximations to order or , which are shown to be insufficient for the generation of surface spherical harmonic coefficients with respect to a geodetic reference ellipsoid. Numerical results of two applications of the new method (the computation of ellipsoidal corrections to gravimetric geoid computation, and area means of gravity anomalies in ellipsoidal approximation) are provided.
Karkar, Ravi; Schroeder, Jessica; Epstein, Daniel A; Pina, Laura R; Scofield, Jeffrey; Fogarty, James; Kientz, Julie A; Munson, Sean A; Vilardaga, Roger; Zia, Jasmine
2017-05-02
Diagnostic self-tracking, the recording of personal information to diagnose or manage a health condition, is a common practice, especially for people with chronic conditions. Unfortunately, many who attempt diagnostic self-tracking have trouble accomplishing their goals. People often lack knowledge and skills needed to design and conduct scientifically rigorous experiments, and current tools provide little support. To address these shortcomings and explore opportunities for diagnostic self-tracking, we designed, developed, and evaluated a mobile app that applies a self-experimentation framework to support patients suffering from irritable bowel syndrome (IBS) in identifying their personal food triggers. TummyTrials aids a person in designing, executing, and analyzing self-experiments to evaluate whether a specific food triggers their symptoms. We examined the feasibility of this approach in a field study with 15 IBS patients, finding that participants could use the tool to reliably undergo a self-experiment. However, we also discovered an underlying tension between scientific validity and the lived experience of self-experimentation. We discuss challenges of applying clinical research methods in everyday life, motivating a need for the design of self-experimentation systems to balance rigor with the uncertainties of everyday life.
Nonlinear saturation of tearing mode islands.
Hastie, R J; Militello, F; Porcelli, F
2005-08-05
New, rigorous results for the tearing island saturation problem are presented. These results are valid for the realistic case where the magnetic island structure is non-symmetric about the reconnection surface and the electron temperature, on which the electrical resistivity depends, is evolved self-consistently with the island growth.
2014-09-01
The NATO Science and Technology Organization Science & Technology (S& T ) in the NATO context is defined as the selective and rigorous...generation and application of state-of-the-art, validated knowledge for defence and security purposes. S& T activities embrace scientific research...engineering, operational research and analysis, synthesis, integration and validation of knowledge derived through the scientific method. In NATO, S& T is
Is the Simple Shoulder Test a valid outcome instrument for shoulder arthroplasty?
Hsu, Jason E; Russ, Stacy M; Somerson, Jeremy S; Tang, Anna; Warme, Winston J; Matsen, Frederick A
2017-10-01
The Simple Shoulder Test (SST) is a brief, inexpensive, and widely used patient-reported outcome tool, but it has not been rigorously evaluated for patients having shoulder arthroplasty. The goal of this study was to rigorously evaluate the validity of the SST for outcome assessment in shoulder arthroplasty using a systematic review of the literature and an analysis of its properties in a series of 408 surgical cases. SST scores, 36-Item Short Form Health Survey scores, and satisfaction scores were collected preoperatively and 2 years postoperatively. Responsiveness was assessed by comparing preoperative and 2-year postoperative scores. Criterion validity was determined by correlating the SST with the 36-Item Short Form Health Survey. Construct validity was tested through 5 clinical hypotheses regarding satisfaction, comorbidities, insurance status, previous failed surgery, and narcotic use. Scores after arthroplasty improved from 3.9 ± 2.8 to 10.2 ± 2.3 (P < .001). The change in SST correlated strongly with patient satisfaction (P < .001). The SST had large Cohen's d effect sizes and standardized response means. Criterion validity was supported by significant differences between satisfied and unsatisfied patients, those with more severe and less severe comorbidities, those with workers' compensation or Medicaid and other types of insurance, those with and without previous failed shoulder surgery, and those taking and those not taking narcotic pain medication before surgery (P < .005). These data combined with a systematic review of the literature demonstrate that the SST is a valid and responsive patient-reported outcome measure for assessing the outcomes of shoulder arthroplasty. Copyright © 2017 Journal of Shoulder and Elbow Surgery Board of Trustees. Published by Elsevier Inc. All rights reserved.
Keeley, Thomas; Al-Janabi, Hareth; Lorgelly, Paula; Coast, Joanna
2013-01-01
Purpose The ICECAP-A and EQ-5D-5L are two index measures appropriate for use in health research. Assessment of content validity allows understanding of whether a measure captures the most relevant and important aspects of a concept. This paper reports a qualitative assessment of the content validity and appropriateness for use of the eq-5D-5L and ICECAP-A measures, using novel methodology. Methods In-depth semi-structured interviews were conducted with research professionals in the UK and Australia. Informants were purposively sampled based on their professional role. Data were analysed in an iterative, thematic and constant comparative manner. A two stage investigation - the comparative direct approach - was developed to address the methodological challenges of the content validity research and allow rigorous assessment. Results Informants viewed the ICECAP-A as an assessment of the broader determinants of quality of life, but lacking in assessment of health-related determinants. The eq-5D-5L was viewed as offering good coverage of health determinants, but as lacking in assessment of these broader determinants. Informants held some concerns about the content or wording of the Self-care, Pain/Discomfort and Anxiety/Depression items (EQ-5D-5L) and the Enjoyment, Achievement and attachment items (ICECAP-A). Conclusion Using rigorous qualitative methodology the results suggest that the ICECAP-A and EQ-5D-5L hold acceptable levels of content validity and are appropriate for use in health research. This work adds expert opinion to the emerging body of research using patients and public to validate these measures. PMID:24367708
NASA Astrophysics Data System (ADS)
Davis, Brian; Turner, Travis L.; Seelecke, Stefan
2005-05-01
Previous work at NASA Langley Research Center (LaRC) involved fabrication and testing of composite beams with embedded, pre-strained shape memory alloy (SMA) ribbons within the beam structures. That study also provided comparison of experimental results with numerical predictions from a research code making use of a new thermoelastic model for shape memory alloy hybrid composite (SMAHC) structures. The previous work showed qualitative validation of the numerical model. However, deficiencies in the experimental-numerical correlation were noted and hypotheses for the discrepancies were given for further investigation. The goal of this work is to refine the experimental measurement and numerical modeling approaches in order to better understand the discrepancies, improve the correlation between prediction and measurement, and provide rigorous quantitative validation of the numerical analysis/design tool. The experimental investigation is refined by a more thorough test procedure and incorporation of higher fidelity measurements such as infrared thermography and projection moire interferometry. The numerical results are produced by a recently commercialized version of the constitutive model as implemented in ABAQUS and are refined by incorporation of additional measured parameters such as geometric imperfection. Thermal buckling, post-buckling, and random responses to thermal and inertial (base acceleration) loads are studied. The results demonstrate the effectiveness of SMAHC structures in controlling static and dynamic responses by adaptive stiffening. Excellent agreement is achieved between the predicted and measured results of the static and dynamic thermomechanical response, thereby providing quantitative validation of the numerical tool.
NASA Technical Reports Server (NTRS)
Davis, Brian; Turner, Travis L.; Seelecke, Stefan
2005-01-01
Previous work at NASA Langley Research Center (LaRC) involved fabrication and testing of composite beams with embedded, pre-strained shape memory alloy (SMA) ribbons within the beam structures. That study also provided comparison of experimental results with numerical predictions from a research code making use of a new thermoelastic model for shape memory alloy hybrid composite (SMAHC) structures. The previous work showed qualitative validation of the numerical model. However, deficiencies in the experimental-numerical correlation were noted and hypotheses for the discrepancies were given for further investigation. The goal of this work is to refine the experimental measurement and numerical modeling approaches in order to better understand the discrepancies, improve the correlation between prediction and measurement, and provide rigorous quantitative validation of the numerical analysis/design tool. The experimental investigation is refined by a more thorough test procedure and incorporation of higher fidelity measurements such as infrared thermography and projection moire interferometry. The numerical results are produced by a recently commercialized version of the constitutive model as implemented in ABAQUS and are refined by incorporation of additional measured parameters such as geometric imperfection. Thermal buckling, post-buckling, and random responses to thermal and inertial (base acceleration) loads are studied. The results demonstrate the effectiveness of SMAHC structures in controlling static and dynamic responses by adaptive stiffening. Excellent agreement is achieved between the predicted and measured results of the static and dynamic thermomechanical response, thereby providing quantitative validation of the numerical tool.
Complicating Methodological Transparency
ERIC Educational Resources Information Center
Bridges-Rhoads, Sarah; Van Cleave, Jessica; Hughes, Hilary E.
2016-01-01
A historical indicator of the quality, validity, and rigor of qualitative research has been the documentation and disclosure of the behind-the-scenes work of the researcher. In this paper, we use what we call "methodological data" as a tool to complicate the possibility and desirability of such transparency. Specifically, we draw on our…
Prompt assessment and management actions are required if we are to reduce the current rapid loss of habitat and biodiversity worldwide. Statistically valid quantification of the biota and habitat condition in water bodies are prerequisites for rigorous assessment of aquatic biodi...
Wisconsin's Model Academic Standards for Music.
ERIC Educational Resources Information Center
Nikolay, Pauli; Grady, Susan; Stefonek, Thomas
To assist parents and educators in preparing students for the 21st century, Wisconsin citizens have become involved in the development of challenging academic standards in 12 curricular areas. Having clear standards for students and teachers makes it possible to develop rigorous local curricula and valid, reliable assessments. This model of…
Given the relatively high cost of mapping impervious surfaces at regional scales, substantial effort is being expended in the development of moderate-resolution, satellite-based methods for estimating impervious surface area (ISA). To rigorously assess the accuracy of these data ...
DOT National Transportation Integrated Search
1994-04-01
Between January 1986 and March 1992, the Federal Aviation Administration's 42-day Nonradar Screen was used to identify Air Traffic Control Specialist (ATCS) candidates with the highest potential to succeed in the rigorous ATCS field training program....
Learning, Judgment, and the Rooted Particular
ERIC Educational Resources Information Center
McCabe, David
2012-01-01
This article begins by acknowledging the general worry that scholarship in the humanities lacks the rigor and objectivity of other scholarly fields. In considering the validity of that criticism, I distinguish two models of learning: the covering law model exemplified by the natural sciences, and the model of rooted particularity that…
Challenges and Innovations in a Community-Based Participatory Randomized Controlled Trial
ERIC Educational Resources Information Center
Goodkind, Jessica R.; Amer, Suha; Christian, Charlisa; Hess, Julia Meredith; Bybee, Deborah; Isakson, Brian L.; Baca, Brandon; Ndayisenga, Martin; Greene, R. Neil; Shantzek, Cece
2017-01-01
Randomized controlled trials (RCTs) are a long-standing and important design for conducting rigorous tests of the effectiveness of health interventions. However, many questions have been raised about the external validity of RCTs, their utility in explicating mechanisms of intervention and participants' intervention experiences, and their…
ERIC Educational Resources Information Center
McCormick, Meghan P.; Cappella, Elise; Hughes, Diane L.; Gallagher, Emily K.
2015-01-01
Peers become increasingly influential in children's development during late childhood and early adolescence. A large body of research has documented children's proclivity for forming friendships with peers who share similar attributes to themselves, a phenomenon termed homophily. Researchers have used multiple procedures to operationalize…
Alves, Vinicius M.; Muratov, Eugene; Fourches, Denis; Strickland, Judy; Kleinstreuer, Nicole; Andrade, Carolina H.; Tropsha, Alexander
2015-01-01
Repetitive exposure to a chemical agent can induce an immune reaction in inherently susceptible individuals that leads to skin sensitization. Although many chemicals have been reported as skin sensitizers, there have been very few rigorously validated QSAR models with defined applicability domains (AD) that were developed using a large group of chemically diverse compounds. In this study, we have aimed to compile, curate, and integrate the largest publicly available dataset related to chemically-induced skin sensitization, use this data to generate rigorously validated and QSAR models for skin sensitization, and employ these models as a virtual screening tool for identifying putative sensitizers among environmental chemicals. We followed best practices for model building and validation implemented with our predictive QSAR workflow using random forest modeling technique in combination with SiRMS and Dragon descriptors. The Correct Classification Rate (CCR) for QSAR models discriminating sensitizers from non-sensitizers were 71–88% when evaluated on several external validation sets, within a broad AD, with positive (for sensitizers) and negative (for non-sensitizers) predicted rates of 85% and 79% respectively. When compared to the skin sensitization module included in the OECD QSAR toolbox as well as to the skin sensitization model in publicly available VEGA software, our models showed a significantly higher prediction accuracy for the same sets of external compounds as evaluated by Positive Predicted Rate, Negative Predicted Rate, and CCR. These models were applied to identify putative chemical hazards in the ScoreCard database of possible skin or sense organ toxicants as primary candidates for experimental validation. PMID:25560674
Clinical relevance in anesthesia journals.
Lauritsen, Jakob; Møller, Ann M
2006-04-01
The purpose of this review is to present the latest knowledge and research on the definition and distribution of clinically relevant articles in anesthesia journals. It will also discuss the importance of the chosen methodology and outcome of articles. In the last few years, more attention has been paid to evidence-based medicine in anesthesia. Several articles on the subject have focused on the need to base clinical decisions on sound research employing both methodological rigor and clinically relevant outcomes. The number of systematic reviews in anesthesia literature is increasing as well as the focus on diminishing the number of surrogate outcomes. It has been shown that the impact factor is not a valid measure of establishing the level of clinical relevance to a journal. This review presents definitions of clinically relevant anesthesia articles. A clinically relevant article employs both methodological rigor and a clinically relevant outcome. The terms methodological rigor and clinical outcomes are fully discussed in the review as well as problems with journal impact factors.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Wilson, Rachel M; Tfaily, Malak M
These data are provided in support of the Commentary, Advanced molecular techniques provide a rigorous method for characterizing organic matter quality in complex systems, Wilson and Tfaily (2018). Measurement results demonstrate that optical characterization of peatland dissolved organic matter (DOM) may not fully capture classically identified chemical characteristics and may, therefore, not be the best measure of organic matter quality.
Fernández-Domínguez, Juan Carlos; de Pedro-Gómez, Joan Ernest; Morales-Asencio, José Miguel; Sastre-Fullana, Pedro; Sesé-Abad, Albert
2017-01-01
Introduction Most of the EBP measuring instruments available to date present limitations both in the operationalisation of the construct and also in the rigour of their psychometric development, as revealed in the literature review performed. The aim of this paper is to provide rigorous and adequate reliability and validity evidence of the scores of a new transdisciplinary psychometric tool, the Health Sciences Evidence-Based Practice (HS-EBP), for measuring the construct EBP in Health Sciences professionals. Methods A pilot study and a subsequent two-stage validation test sample were conducted to progressively refine the instrument until a reduced 60-item version with a five-factor latent structure. Reliability was analysed through both Cronbach’s alpha coefficient and intraclass correlations (ICC). Latent structure was contrasted using confirmatory factor analysis (CFA) following a model comparison aproach. Evidence of criterion validity of the scores obtained was achieved by considering attitudinal resistance to change, burnout, and quality of professional life as criterion variables; while convergent validity was assessed using the Spanish version of the Evidence-Based Practice Questionnaire (EBPQ-19). Results Adequate evidence of both reliability and ICC was obtained for the five dimensions of the questionnaire. According to the CFA model comparison, the best fit corresponded to the five-factor model (RMSEA = 0.049; CI 90% RMSEA = [0.047; 0.050]; CFI = 0.99). Adequate criterion and convergent validity evidence was also provided. Finally, the HS-EBP showed the capability to find differences between EBP training levels as an important evidence of decision validity. Conclusions Reliability and validity evidence obtained regarding the HS-EBP confirm the adequate operationalisation of the EBP construct as a process put into practice to respond to every clinical situation arising in the daily practice of professionals in health sciences (transprofessional). The tool could be useful for EBP individual assessment and for evaluating the impact of specific interventions to improve EBP. PMID:28486533
Fernández-Domínguez, Juan Carlos; de Pedro-Gómez, Joan Ernest; Morales-Asencio, José Miguel; Bennasar-Veny, Miquel; Sastre-Fullana, Pedro; Sesé-Abad, Albert
2017-01-01
Most of the EBP measuring instruments available to date present limitations both in the operationalisation of the construct and also in the rigour of their psychometric development, as revealed in the literature review performed. The aim of this paper is to provide rigorous and adequate reliability and validity evidence of the scores of a new transdisciplinary psychometric tool, the Health Sciences Evidence-Based Practice (HS-EBP), for measuring the construct EBP in Health Sciences professionals. A pilot study and a subsequent two-stage validation test sample were conducted to progressively refine the instrument until a reduced 60-item version with a five-factor latent structure. Reliability was analysed through both Cronbach's alpha coefficient and intraclass correlations (ICC). Latent structure was contrasted using confirmatory factor analysis (CFA) following a model comparison aproach. Evidence of criterion validity of the scores obtained was achieved by considering attitudinal resistance to change, burnout, and quality of professional life as criterion variables; while convergent validity was assessed using the Spanish version of the Evidence-Based Practice Questionnaire (EBPQ-19). Adequate evidence of both reliability and ICC was obtained for the five dimensions of the questionnaire. According to the CFA model comparison, the best fit corresponded to the five-factor model (RMSEA = 0.049; CI 90% RMSEA = [0.047; 0.050]; CFI = 0.99). Adequate criterion and convergent validity evidence was also provided. Finally, the HS-EBP showed the capability to find differences between EBP training levels as an important evidence of decision validity. Reliability and validity evidence obtained regarding the HS-EBP confirm the adequate operationalisation of the EBP construct as a process put into practice to respond to every clinical situation arising in the daily practice of professionals in health sciences (transprofessional). The tool could be useful for EBP individual assessment and for evaluating the impact of specific interventions to improve EBP.
Augmented assessment as a means to augmented reality.
Bergeron, Bryan
2006-01-01
Rigorous scientific assessment of educational technologies typically lags behind the availability of the technologies by years because of the lack of validated instruments and benchmarks. Even when the appropriate assessment instruments are available, they may not be applied because of time and monetary constraints. Work in augmented reality, instrumented mannequins, serious gaming, and similar promising educational technologies that haven't undergone timely, rigorous evaluation, highlights the need for assessment methodologies that address the limitations of traditional approaches. The most promising augmented assessment solutions incorporate elements of rapid prototyping used in the software industry, simulation-based assessment techniques modeled after methods used in bioinformatics, and object-oriented analysis methods borrowed from object oriented programming.
Scale/TSUNAMI Sensitivity Data for ICSBEP Evaluations
DOE Office of Scientific and Technical Information (OSTI.GOV)
Rearden, Bradley T; Reed, Davis Allan; Lefebvre, Robert A
2011-01-01
The Tools for Sensitivity and Uncertainty Analysis Methodology Implementation (TSUNAMI) software developed at Oak Ridge National Laboratory (ORNL) as part of the Scale code system provide unique methods for code validation, gap analysis, and experiment design. For TSUNAMI analysis, sensitivity data are generated for each application and each existing or proposed experiment used in the assessment. The validation of diverse sets of applications requires potentially thousands of data files to be maintained and organized by the user, and a growing number of these files are available through the International Handbook of Evaluated Criticality Safety Benchmark Experiments (IHECSBE) distributed through themore » International Criticality Safety Benchmark Evaluation Program (ICSBEP). To facilitate the use of the IHECSBE benchmarks in rigorous TSUNAMI validation and gap analysis techniques, ORNL generated SCALE/TSUNAMI sensitivity data files (SDFs) for several hundred benchmarks for distribution with the IHECSBE. For the 2010 edition of IHECSBE, the sensitivity data were generated using 238-group cross-section data based on ENDF/B-VII.0 for 494 benchmark experiments. Additionally, ORNL has developed a quality assurance procedure to guide the generation of Scale inputs and sensitivity data, as well as a graphical user interface to facilitate the use of sensitivity data in identifying experiments and applying them in validation studies.« less
Dunne, Carmel L; Fraser, Jennifer; Gardner, Glenn E
2014-07-01
to develop and test the reliability and validity of a research instrument to measure women's perceptions of social support provided during labour by at least one lay birth companion. a cross-sectional study was carried out from April 2009 to February 2010. non-tertiary hospital in the outer western region of Brisbane, Australia. six registered midwives and 10 postnatal women reviewed the instrument. The instrument was then completed by 293 inpatient women who had experienced a vaginal birth. the Birth Companion Support Questionnaire (BCSQ) was developed and its reliability and validity were evaluated in this study. An exploratory factor analysis was performed on the final instrument using principal component analysis with an oblique (Promax) rotation. This process suggested two subscales: emotional support and tangible support. The questionnaire was found to be reliable and valid for use in midwifery research. the BCSQ is an appropriate instrument to measure women's perceptions of lay birth companion support during labour. this is the first rigorous study to develop and test a measure of social support in labour which is critical at a time when policy makers and health planners need to consider the needs of birthing women and their network of support friends and family. © 2013 Elsevier Ltd. All rights reserved.
Simulation-based assessment in anesthesiology: requirements for practical implementation.
Boulet, John R; Murray, David J
2010-04-01
Simulations have taken a central role in the education and assessment of medical students, residents, and practicing physicians. The introduction of simulation-based assessments in anesthesiology, especially those used to establish various competencies, has demanded fairly rigorous studies concerning the psychometric properties of the scores. Most important, major efforts have been directed at identifying, and addressing, potential threats to the validity of simulation-based assessment scores. As a result, organizations that wish to incorporate simulation-based assessments into their evaluation practices can access information regarding effective test development practices, the selection of appropriate metrics, the minimization of measurement errors, and test score validation processes. The purpose of this article is to provide a broad overview of the use of simulation for measuring physician skills and competencies. For simulations used in anesthesiology, studies that describe advances in scenario development, the development of scoring rubrics, and the validation of assessment results are synthesized. Based on the summary of relevant research, psychometric requirements for practical implementation of simulation-based assessments in anesthesiology are forwarded. As technology expands, and simulation-based education and evaluation takes on a larger role in patient safety initiatives, the groundbreaking work conducted to date can serve as a model for those individuals and organizations that are responsible for developing, scoring, or validating simulation-based education and assessment programs in anesthesiology.
Tuomivaara, Sami T; Yaoi, Katsuro; O'Neill, Malcolm A; York, William S
2015-01-30
Xyloglucans are structurally complex plant cell wall polysaccharides that are involved in cell growth and expansion, energy metabolism, and signaling. Determining the structure-function relationships of xyloglucans would benefit from the availability of a comprehensive and structurally diverse collection of rigorously characterized xyloglucan oligosaccharides. Here, we present a workflow for the semi-preparative scale generation and purification of neutral and acidic xyloglucan oligosaccharides using a combination of enzymatic and chemical treatments and size-exclusion chromatography. Twenty-six of these oligosaccharides were purified to near homogeneity and their structures validated using a combination of matrix-assisted laser desorption/ionization mass spectrometry, high-performance anion exchange chromatography, and 1H nuclear magnetic resonance spectroscopy. Mass spectrometry and analytical chromatography were compared as methods for xyloglucan oligosaccharide quantification. 1H chemical shifts were assigned using two-dimensional correlation spectroscopy. A comprehensive update of the nomenclature describing xyloglucan side-chain structures is provided for reference. Copyright © 2014 Elsevier Ltd. All rights reserved.
Dommert, M; Reginatto, M; Zboril, M; Fiedler, F; Helmbrecht, S; Enghardt, W; Lutz, B
2017-11-28
Bonner sphere measurements are typically analyzed using unfolding codes. It is well known that it is difficult to get reliable estimates of uncertainties for standard unfolding procedures. An alternative approach is to analyze the data using Bayesian parameter estimation. This method provides reliable estimates of the uncertainties of neutron spectra leading to rigorous estimates of uncertainties of the dose. We extend previous Bayesian approaches and apply the method to stray neutrons in proton therapy environments by introducing a new parameterized model which describes the main features of the expected neutron spectra. The parameterization is based on information that is available from measurements and detailed Monte Carlo simulations. The validity of this approach has been validated with results of an experiment using Bonner spheres carried out at the experimental hall of the OncoRay proton therapy facility in Dresden. © The Author 2017. Published by Oxford University Press. All rights reserved. For Permissions, please email: journals.permissions@oup.com.
Mathematical Rigor vs. Conceptual Change: Some Early Results
NASA Astrophysics Data System (ADS)
Alexander, W. R.
2003-05-01
Results from two different pedagogical approaches to teaching introductory astronomy at the college level will be presented. The first of these approaches is a descriptive, conceptually based approach that emphasizes conceptual change. This descriptive class is typically an elective for non-science majors. The other approach is a mathematically rigorous treatment that emphasizes problem solving and is designed to prepare students for further study in astronomy. The mathematically rigorous class is typically taken by science majors. It also fulfills an elective science requirement for these science majors. The Astronomy Diagnostic Test version 2 (ADT 2.0) was used as an assessment instrument since the validity and reliability have been investigated by previous researchers. The ADT 2.0 was administered as both a pre-test and post-test to both groups. Initial results show no significant difference between the two groups in the post-test. However, there is a slightly greater improvement for the descriptive class between the pre and post testing compared to the mathematically rigorous course. There was great care to account for variables. These variables included: selection of text, class format as well as instructor differences. Results indicate that the mathematically rigorous model, doesn't improve conceptual understanding any better than the conceptual change model. Additional results indicate that there is a similar gender bias in favor of males that has been measured by previous investigators. This research has been funded by the College of Science and Mathematics at James Madison University.
A methodology for the rigorous verification of plasma simulation codes
NASA Astrophysics Data System (ADS)
Riva, Fabio
2016-10-01
The methodology used to assess the reliability of numerical simulation codes constitutes the Verification and Validation (V&V) procedure. V&V is composed by two separate tasks: the verification, which is a mathematical issue targeted to assess that the physical model is correctly solved, and the validation, which determines the consistency of the code results, and therefore of the physical model, with experimental data. In the present talk we focus our attention on the verification, which in turn is composed by the code verification, targeted to assess that a physical model is correctly implemented in a simulation code, and the solution verification, that quantifies the numerical error affecting a simulation. Bridging the gap between plasma physics and other scientific domains, we introduced for the first time in our domain a rigorous methodology for the code verification, based on the method of manufactured solutions, as well as a solution verification based on the Richardson extrapolation. This methodology was applied to GBS, a three-dimensional fluid code based on a finite difference scheme, used to investigate the plasma turbulence in basic plasma physics experiments and in the tokamak scrape-off layer. Overcoming the difficulty of dealing with a numerical method intrinsically affected by statistical noise, we have now generalized the rigorous verification methodology to simulation codes based on the particle-in-cell algorithm, which are employed to solve Vlasov equation in the investigation of a number of plasma physics phenomena.
Stem Cell Research and Clinical Translation: A Roadmap about Good Clinical Practice and Patient Care
Scopetti, Matteo; Gatto, Vittorio
2017-01-01
The latest research achievements in the field of stem cells led in 2016 to the publication of “Guidelines for Stem Cell Research and Clinical Translation” by the International Society for Stem Cell Research (ISSCR). Updating the topics covered in previous publications, the new recommendations offer interesting ethical and scientific insights. Under the common principles of research integrity, protection of patient's welfare, respect for the research subjects, transparency and social justice, the centrality of good clinical practice, and informed consent in research and translational medicine is supported. The guidelines implement the abovementioned publications, requiring rigor in all areas of research, promoting the validity of the scientific activity results and emphasizing the need for an accurate and efficient public communication. This paper aims to analyze the aforementioned guidelines in order to provide a valid interpretive tool for experts. In particular, a research activity focused on the bioethical, scientific, and social implications of the new recommendations is carried out in order to provide food for thought. Finally, as an emerging issue of potential impact of current guidelines, an overview on implications of compensation for egg donation is offered. PMID:29090010
Frati, Paola; Scopetti, Matteo; Santurro, Alessandro; Gatto, Vittorio; Fineschi, Vittorio
2017-01-01
The latest research achievements in the field of stem cells led in 2016 to the publication of "Guidelines for Stem Cell Research and Clinical Translation" by the International Society for Stem Cell Research (ISSCR). Updating the topics covered in previous publications, the new recommendations offer interesting ethical and scientific insights. Under the common principles of research integrity, protection of patient's welfare, respect for the research subjects, transparency and social justice, the centrality of good clinical practice, and informed consent in research and translational medicine is supported. The guidelines implement the abovementioned publications, requiring rigor in all areas of research, promoting the validity of the scientific activity results and emphasizing the need for an accurate and efficient public communication. This paper aims to analyze the aforementioned guidelines in order to provide a valid interpretive tool for experts. In particular, a research activity focused on the bioethical, scientific, and social implications of the new recommendations is carried out in order to provide food for thought. Finally, as an emerging issue of potential impact of current guidelines, an overview on implications of compensation for egg donation is offered.
History and Future for the Happy Marriage between the MODIS Land team and Fluxnet
NASA Astrophysics Data System (ADS)
Running, S. W.
2015-12-01
When I wrote the proposal to NASA in 1988 for daily global evapotranspiration and gross primary production algorithms for the MODIS sensor, I had no validation plan. Fluxnet probably saved my MODIS career by developing a global network of rigorously calibrated towers measuring water and carbon fluxes over a wide variety of ecosystems that I could not even envision at the time that first proposal was written. However my enthusiasm for Fluxnet was not reciprocated by the Fluxnet community until we began providing 7 x 7 pixel MODIS Land datasets exactly over each of their towers every 8 days, without them having to crawl thru the global datasets and make individual orders. This system, known informally as the MODIS ASCII cutouts, began in 2002 and operates at the Oak Ridge DAAC to this day, cementing a mutually beneficial data interchange between the Fluxnet and remote sensing communities. This talk will briefly discuss the history of MODIS validation with flux towers, and flux spatial scaling with MODIS data. More importantly I will detail the future continuity of global biophysical datasets in the post-MODIS era, and what next generation sensors will provide.
NASA Technical Reports Server (NTRS)
Jacklin, Stephen; Schumann, Johann; Gupta, Pramod; Richard, Michael; Guenther, Kurt; Soares, Fola
2005-01-01
Adaptive control technologies that incorporate learning algorithms have been proposed to enable automatic flight control and vehicle recovery, autonomous flight, and to maintain vehicle performance in the face of unknown, changing, or poorly defined operating environments. In order for adaptive control systems to be used in safety-critical aerospace applications, they must be proven to be highly safe and reliable. Rigorous methods for adaptive software verification and validation must be developed to ensure that control system software failures will not occur. Of central importance in this regard is the need to establish reliable methods that guarantee convergent learning, rapid convergence (learning) rate, and algorithm stability. This paper presents the major problems of adaptive control systems that use learning to improve performance. The paper then presents the major procedures and tools presently developed or currently being developed to enable the verification, validation, and ultimate certification of these adaptive control systems. These technologies include the application of automated program analysis methods, techniques to improve the learning process, analytical methods to verify stability, methods to automatically synthesize code, simulation and test methods, and tools to provide on-line software assurance.
QUADAS and STARD: evaluating the quality of diagnostic accuracy studies.
Oliveira, Maria Regina Fernandes de; Gomes, Almério de Castro; Toscano, Cristiana Maria
2011-04-01
To compare the performance of two approaches, one based on the Quality Assessment of Diagnostic Accuracy Studies (QUADAS) and another on the Standards for Reporting Studies of Diagnostic Accuracy (STARD), in evaluating the quality of studies validating the OptiMal® rapid malaria diagnostic test. Articles validating the rapid test published until 2007 were searched in the Medline/PubMed database. This search retrieved 13 articles. A combination of 12 QUADAS criteria and three STARD criteria were compared with the 12 QUADAS criteria alone. Articles that fulfilled at least 50% of QUADAS criteria were considered as regular to good quality. Of the 13 articles retrieved, 12 fulfilled at least 50% of QUADAS criteria, and only two fulfilled the STARD/QUADAS criteria combined. Considering the two criteria combination (> 6 QUADAS and > 3 STARD), two studies (15.4%) showed good methodological quality. The articles selection using the proposed combination resulted in two to eight articles, depending on the number of items assumed as cutoff point. The STARD/QUADAS combination has the potential to provide greater rigor when evaluating the quality of studies validating malaria diagnostic tests, given that it incorporates relevant information not contemplated in the QUADAS criteria alone.
Testability of evolutionary game dynamics based on experimental economics data
NASA Astrophysics Data System (ADS)
Wang, Yijia; Chen, Xiaojie; Wang, Zhijian
In order to better understand the dynamic processes of a real game system, we need an appropriate dynamics model, so to evaluate the validity of a model is not a trivial task. Here, we demonstrate an approach, considering the dynamical macroscope patterns of angular momentum and speed as the measurement variables, to evaluate the validity of various dynamics models. Using the data in real time Rock-Paper-Scissors (RPS) games experiments, we obtain the experimental dynamic patterns, and then derive the related theoretical dynamic patterns from a series of typical dynamics models respectively. By testing the goodness-of-fit between the experimental and theoretical patterns, the validity of the models can be evaluated. One of the results in our study case is that, among all the nonparametric models tested, the best-known Replicator dynamics model performs almost worst, while the Projection dynamics model performs best. Besides providing new empirical macroscope patterns of social dynamics, we demonstrate that the approach can be an effective and rigorous tool to test game dynamics models. Fundamental Research Funds for the Central Universities (SSEYI2014Z) and the National Natural Science Foundation of China (Grants No. 61503062).
Validation of Extended MHD Models using MST RFP Plasmas
NASA Astrophysics Data System (ADS)
Jacobson, C. M.; Chapman, B. E.; Craig, D.; McCollam, K. J.; Sovinec, C. R.
2016-10-01
Significant effort has been devoted to improvement of computational models used in fusion energy sciences. Rigorous validation of these models is necessary in order to increase confidence in their ability to predict the performance of future devices. MST is a well diagnosed reversed-field pinch (RFP) capable of operation over a wide range of parameters. In particular, the Lundquist number S, a key parameter in resistive magnetohydrodynamics (MHD), can be varied over a wide range and provide substantial overlap with MHD RFP simulations. MST RFP plasmas are simulated using both DEBS, a nonlinear single-fluid visco-resistive MHD code, and NIMROD, a nonlinear extended MHD code, with S ranging from 104 to 5 ×104 for single-fluid runs, with the magnetic Prandtl number Pm = 1 . Experiments with plasma current IP ranging from 60 kA to 500 kA result in S from 4 ×104 to 8 ×106 . Validation metric comparisons are presented, focusing on how magnetic fluctuations b scale with S. Single-fluid NIMROD results give S b - 0.21 , and experiments give S b - 0.28 for the dominant m = 1 , n = 6 mode. Preliminary two-fluid NIMROD results are also presented. Work supported by US DOE.
Validation of educational assessments: a primer for simulation and beyond.
Cook, David A; Hatala, Rose
2016-01-01
Simulation plays a vital role in health professions assessment. This review provides a primer on assessment validation for educators and education researchers. We focus on simulation-based assessment of health professionals, but the principles apply broadly to other assessment approaches and topics. Validation refers to the process of collecting validity evidence to evaluate the appropriateness of the interpretations, uses, and decisions based on assessment results. Contemporary frameworks view validity as a hypothesis, and validity evidence is collected to support or refute the validity hypothesis (i.e., that the proposed interpretations and decisions are defensible). In validation, the educator or researcher defines the proposed interpretations and decisions, identifies and prioritizes the most questionable assumptions in making these interpretations and decisions (the "interpretation-use argument"), empirically tests those assumptions using existing or newly-collected evidence, and then summarizes the evidence as a coherent "validity argument." A framework proposed by Messick identifies potential evidence sources: content, response process, internal structure, relationships with other variables, and consequences. Another framework proposed by Kane identifies key inferences in generating useful interpretations: scoring, generalization, extrapolation, and implications/decision. We propose an eight-step approach to validation that applies to either framework: Define the construct and proposed interpretation, make explicit the intended decision(s), define the interpretation-use argument and prioritize needed validity evidence, identify candidate instruments and/or create/adapt a new instrument, appraise existing evidence and collect new evidence as needed, keep track of practical issues, formulate the validity argument, and make a judgment: does the evidence support the intended use? Rigorous validation first prioritizes and then empirically evaluates key assumptions in the interpretation and use of assessment scores. Validation science would be improved by more explicit articulation and prioritization of the interpretation-use argument, greater use of formal validation frameworks, and more evidence informing the consequences and implications of assessment.
Measuring Patient-Reported Outcomes: Key Metrics in Reconstructive Surgery.
Voineskos, Sophocles H; Nelson, Jonas A; Klassen, Anne F; Pusic, Andrea L
2018-01-29
Satisfaction and improved quality of life are among the most important outcomes for patients undergoing plastic and reconstructive surgery for a variety of diseases and conditions. Patient-reported outcome measures (PROMs) are essential tools for evaluating the benefits of newly developed surgical techniques. Modern PROMs are being developed with new psychometric approaches, such as Rasch Measurement Theory, and their measurement properties (validity, reliability, responsiveness) are rigorously tested. These advances have resulted in the availability of PROMs that provide clinically meaningful data and effectively measure functional as well as psychosocial outcomes. This article guides the reader through the steps of creating a PROM and highlights the potential research and clinical uses of such instruments. Limitations of PROMs and anticipated future directions in this field are discussed.
Characterizing (rating) the performance of large photovoltaic arrays for all operating conditions
DOE Office of Scientific and Technical Information (OSTI.GOV)
King, D.L.; Eckert, P.E.
1996-06-01
A new method has been developed for characterizing the electrical performance of photovoltaic arrays. The method provides both a ``rating`` at standard reporting conditions and a rigorous yet straightforward model for predicting array performance at all operating conditions. For the first time, the performance model handles the influences of irradiance, module temperature, solar spectrum, solar angle-of-incidence, and temperature coefficients, in a practical way. Validity of the procedure was confirmed during field testing of a 25-kW array recently installed by Arizona Public Service Co. on Carol Spring Mountain (which powers microwave, ceullular phone, and TV communictions equipment). This paper describes themore » characterization procedure, measured array performance, and the predictive model.« less
Exclusive processes and the fundamental structure of hadrons
Brodsky, Stanley J.
2015-01-20
I review the historical development of QCD predictions for exclusive hadronic processes, beginning with constituent counting rules and the quark interchange mechanism, phenomena which gave early validation for the quark structure of hadrons. The subsequent development of pQCD factorization theorems for hard exclusive amplitudes and the development of evolution equations for the hadron distribution amplitudes provided a rigorous framework for calculating hadronic form factors and hard scattering exclusive scattering processes at high momentum transfer. I also give a brief introduction to the field of "light-front holography" and the insights it brings to quark confinement, the behavior of the QCD couplingmore » in the nonperturbative domain, as well as hadron spectroscopy and the dynamics of exclusive processes.« less
Exclusive processes and the fundamental structure of hadrons
DOE Office of Scientific and Technical Information (OSTI.GOV)
Brodsky, Stanley J.
I review the historical development of QCD predictions for exclusive hadronic processes, beginning with constituent counting rules and the quark interchange mechanism, phenomena which gave early validation for the quark structure of hadrons. The subsequent development of pQCD factorization theorems for hard exclusive amplitudes and the development of evolution equations for the hadron distribution amplitudes provided a rigorous framework for calculating hadronic form factors and hard scattering exclusive scattering processes at high momentum transfer. I also give a brief introduction to the field of "light-front holography" and the insights it brings to quark confinement, the behavior of the QCD couplingmore » in the nonperturbative domain, as well as hadron spectroscopy and the dynamics of exclusive processes.« less
Determining the Measurement Quality of a Montessori High School Teacher Evaluation Survey
ERIC Educational Resources Information Center
Setari, Anthony Philip; Bradley, Kelly D.
2017-01-01
The purpose of this study was to conduct a psychometric validation of a course evaluation instrument, known as a student evaluation of teaching (SET), implemented in a Montessori high school. The authors demonstrate to the Montessori community how to rigorously examine the measurement and assessment quality of instruments used within Montessori…
Psychometric Assessment of the Mindful Attention Awareness Scale (MAAS) among Chinese Adolescents
ERIC Educational Resources Information Center
Black, David S.; Sussman, Steve; Johnson, C. Anderson; Milam, Joel
2012-01-01
The Mindful Attention Awareness Scale (MAAS) has the longest empirical track record as a valid measure of trait mindfulness. Most of what is understood about trait mindfulness comes from administering the MAAS to relatively homogenous samples of Caucasian adults. This study rigorously evaluates the psychometric properties of the MAAS among Chinese…
Meeting the needs of an ever-demanding market.
Rigby, Richard
2002-04-01
Balancing cost and performance in packaging is critical. This article outlines techniques to assist in this whilst delivering added value and product differentiation. The techniques include a rigorous statistical process capable of delivering cost reduction and improved quality and a computer modelling process that can save time when validating new packaging options.
ERIC Educational Resources Information Center
Wipawayangkool, Kamphol
2011-01-01
The notion of knowledge internalization (KI), albeit a critical link in Nonaka's (1994) organizational knowledge creation theory, has not been rigorously conceptualized and defined, let alone operationalized. To strengthen the foundation for knowledge management (KM) research, we attempt to fulfill the following research objectives in the three…
Video Games for Neuro-Cognitive Optimization.
Mishra, Jyoti; Anguera, Joaquin A; Gazzaley, Adam
2016-04-20
Sophisticated video games that integrate engaging cognitive training with real-time biosensing and neurostimulation have the potential to optimize cognitive performance in health and disease. We argue that technology development must be paired with rigorous scientific validation and discuss academic and industry opportunities in this field. Copyright © 2016 Elsevier Inc. All rights reserved.
A More Rigorous Quasi-Experimental Alternative to the One-Group Pretest-Posttest Design.
ERIC Educational Resources Information Center
Johnson, Craig W.
1986-01-01
A simple quasi-experimental design is described which may have utility in a variety of applied and laboratory research settings where ordinarily the one-group pretest-posttest pre-experimental design might otherwise be the procedure of choice. The design approaches the internal validity of true experimental designs while optimizing external…
ERIC Educational Resources Information Center
Johnson, Lawrence J.; LaMontagne, M. J.
1993-01-01
This paper describes content analysis as a data analysis technique useful for examining written or verbal communication within early intervention. The article outlines the use of referential or thematic recording units derived from interview data, identifies procedural guidelines, and addresses issues of rigor and validity. (Author/JDD)
Complexity, Representation and Practice: Case Study as Method and Methodology
ERIC Educational Resources Information Center
Miles, Rebecca
2015-01-01
While case study is considered a common approach to examining specific and particular examples in research disciplines such as law, medicine and psychology, in the social sciences case study is often treated as a lesser, flawed or undemanding methodology which is less valid, reliable or theoretically rigorous than other methodologies. Building on…
Alarms about structural alerts.
Alves, Vinicius; Muratov, Eugene; Capuzzi, Stephen; Politi, Regina; Low, Yen; Braga, Rodolpho; Zakharov, Alexey V; Sedykh, Alexander; Mokshyna, Elena; Farag, Sherif; Andrade, Carolina; Kuz'min, Victor; Fourches, Denis; Tropsha, Alexander
2016-08-21
Structural alerts are widely accepted in chemical toxicology and regulatory decision support as a simple and transparent means to flag potential chemical hazards or group compounds into categories for read-across. However, there has been a growing concern that alerts disproportionally flag too many chemicals as toxic, which questions their reliability as toxicity markers. Conversely, the rigorously developed and properly validated statistical QSAR models can accurately and reliably predict the toxicity of a chemical; however, their use in regulatory toxicology has been hampered by the lack of transparency and interpretability. We demonstrate that contrary to the common perception of QSAR models as "black boxes" they can be used to identify statistically significant chemical substructures (QSAR-based alerts) that influence toxicity. We show through several case studies, however, that the mere presence of structural alerts in a chemical, irrespective of the derivation method (expert-based or QSAR-based), should be perceived only as hypotheses of possible toxicological effect. We propose a new approach that synergistically integrates structural alerts and rigorously validated QSAR models for a more transparent and accurate safety assessment of new chemicals.
Miyata, Hiroaki; Kai, Ichiro
2006-05-01
Debate about the relationship between quantitative and qualitative paradigms is often muddled and confused and the clutter of terms and arguments has resulted in the concepts becoming obscure and unrecognizable. It is therefore very important to reconsider evaluation criteria regarding rigor in social science. As Lincoln & Guba have already compared quantitative paradigms (validity, reliability, neutrality, generalizability) with qualitative paradigms (credibility, dependability, confirmability, transferability), we have discuss use of evaluation criteria based on pragmatic perspective. Validity/Credibility is the paradigm concerned to observational framework, while Reliability/Dependability refer to the range of stability in observations, Neutrality/Confirmability reflect influences between observers and subjects, Generalizability/Transferability have epistemological difference in the way findings are applied. Qualitative studies, however, does not always chose the qualitative paradigms. If we assume the stability to some extent, it is better to use the quantitative paradigm (reliability). Moreover as a quantitative study can not always guarantee a perfect observational framework, with stability in all phases of observations, it is useful to use qualitative paradigms to enhance the rigor in the study.
Best practices for evaluating single nucleotide variant calling methods for microbial genomics
Olson, Nathan D.; Lund, Steven P.; Colman, Rebecca E.; Foster, Jeffrey T.; Sahl, Jason W.; Schupp, James M.; Keim, Paul; Morrow, Jayne B.; Salit, Marc L.; Zook, Justin M.
2015-01-01
Innovations in sequencing technologies have allowed biologists to make incredible advances in understanding biological systems. As experience grows, researchers increasingly recognize that analyzing the wealth of data provided by these new sequencing platforms requires careful attention to detail for robust results. Thus far, much of the scientific Communit’s focus for use in bacterial genomics has been on evaluating genome assembly algorithms and rigorously validating assembly program performance. Missing, however, is a focus on critical evaluation of variant callers for these genomes. Variant calling is essential for comparative genomics as it yields insights into nucleotide-level organismal differences. Variant calling is a multistep process with a host of potential error sources that may lead to incorrect variant calls. Identifying and resolving these incorrect calls is critical for bacterial genomics to advance. The goal of this review is to provide guidance on validating algorithms and pipelines used in variant calling for bacterial genomics. First, we will provide an overview of the variant calling procedures and the potential sources of error associated with the methods. We will then identify appropriate datasets for use in evaluating algorithms and describe statistical methods for evaluating algorithm performance. As variant calling moves from basic research to the applied setting, standardized methods for performance evaluation and reporting are required; it is our hope that this review provides the groundwork for the development of these standards. PMID:26217378
Applicability Analysis of Validation Evidence for Biomedical Computational Models
Pathmanathan, Pras; Gray, Richard A.; Romero, Vicente J.; ...
2017-09-07
Computational modeling has the potential to revolutionize medicine the way it transformed engineering. However, despite decades of work, there has only been limited progress to successfully translate modeling research to patient care. One major difficulty which often occurs with biomedical computational models is an inability to perform validation in a setting that closely resembles how the model will be used. For example, for a biomedical model that makes in vivo clinically relevant predictions, direct validation of predictions may be impossible for ethical, technological, or financial reasons. Unavoidable limitations inherent to the validation process lead to challenges in evaluating the credibilitymore » of biomedical model predictions. Therefore, when evaluating biomedical models, it is critical to rigorously assess applicability, that is, the relevance of the computational model, and its validation evidence to the proposed context of use (COU). However, there are no well-established methods for assessing applicability. In this paper, we present a novel framework for performing applicability analysis and demonstrate its use with a medical device computational model. The framework provides a systematic, step-by-step method for breaking down the broad question of applicability into a series of focused questions, which may be addressed using supporting evidence and subject matter expertise. The framework can be used for model justification, model assessment, and validation planning. While motivated by biomedical models, it is relevant to a broad range of disciplines and underlying physics. Finally, the proposed applicability framework could help overcome some of the barriers inherent to validation of, and aid clinical implementation of, biomedical models.« less
The analytical validation of the Oncotype DX Recurrence Score assay
Baehner, Frederick L
2016-01-01
In vitro diagnostic multivariate index assays are highly complex molecular assays that can provide clinically actionable information regarding the underlying tumour biology and facilitate personalised treatment. These assays are only useful in clinical practice if all of the following are established: analytical validation (i.e., how accurately/reliably the assay measures the molecular characteristics), clinical validation (i.e., how consistently/accurately the test detects/predicts the outcomes of interest), and clinical utility (i.e., how likely the test is to significantly improve patient outcomes). In considering the use of these assays, clinicians often focus primarily on the clinical validity/utility; however, the analytical validity of an assay (e.g., its accuracy, reproducibility, and standardisation) should also be evaluated and carefully considered. This review focuses on the rigorous analytical validation and performance of the Oncotype DX® Breast Cancer Assay, which is performed at the Central Clinical Reference Laboratory of Genomic Health, Inc. The assay process includes tumour tissue enrichment (if needed), RNA extraction, gene expression quantitation (using a gene panel consisting of 16 cancer genes plus 5 reference genes and quantitative real-time RT-PCR), and an automated computer algorithm to produce a Recurrence Score® result (scale: 0–100). This review presents evidence showing that the Recurrence Score result reported for each patient falls within a tight clinically relevant confidence interval. Specifically, the review discusses how the development of the assay was designed to optimise assay performance, presents data supporting its analytical validity, and describes the quality control and assurance programmes that ensure optimal test performance over time. PMID:27729940
The analytical validation of the Oncotype DX Recurrence Score assay.
Baehner, Frederick L
2016-01-01
In vitro diagnostic multivariate index assays are highly complex molecular assays that can provide clinically actionable information regarding the underlying tumour biology and facilitate personalised treatment. These assays are only useful in clinical practice if all of the following are established: analytical validation (i.e., how accurately/reliably the assay measures the molecular characteristics), clinical validation (i.e., how consistently/accurately the test detects/predicts the outcomes of interest), and clinical utility (i.e., how likely the test is to significantly improve patient outcomes). In considering the use of these assays, clinicians often focus primarily on the clinical validity/utility; however, the analytical validity of an assay (e.g., its accuracy, reproducibility, and standardisation) should also be evaluated and carefully considered. This review focuses on the rigorous analytical validation and performance of the Oncotype DX ® Breast Cancer Assay, which is performed at the Central Clinical Reference Laboratory of Genomic Health, Inc. The assay process includes tumour tissue enrichment (if needed), RNA extraction, gene expression quantitation (using a gene panel consisting of 16 cancer genes plus 5 reference genes and quantitative real-time RT-PCR), and an automated computer algorithm to produce a Recurrence Score ® result (scale: 0-100). This review presents evidence showing that the Recurrence Score result reported for each patient falls within a tight clinically relevant confidence interval. Specifically, the review discusses how the development of the assay was designed to optimise assay performance, presents data supporting its analytical validity, and describes the quality control and assurance programmes that ensure optimal test performance over time.
Applicability Analysis of Validation Evidence for Biomedical Computational Models
DOE Office of Scientific and Technical Information (OSTI.GOV)
Pathmanathan, Pras; Gray, Richard A.; Romero, Vicente J.
Computational modeling has the potential to revolutionize medicine the way it transformed engineering. However, despite decades of work, there has only been limited progress to successfully translate modeling research to patient care. One major difficulty which often occurs with biomedical computational models is an inability to perform validation in a setting that closely resembles how the model will be used. For example, for a biomedical model that makes in vivo clinically relevant predictions, direct validation of predictions may be impossible for ethical, technological, or financial reasons. Unavoidable limitations inherent to the validation process lead to challenges in evaluating the credibilitymore » of biomedical model predictions. Therefore, when evaluating biomedical models, it is critical to rigorously assess applicability, that is, the relevance of the computational model, and its validation evidence to the proposed context of use (COU). However, there are no well-established methods for assessing applicability. In this paper, we present a novel framework for performing applicability analysis and demonstrate its use with a medical device computational model. The framework provides a systematic, step-by-step method for breaking down the broad question of applicability into a series of focused questions, which may be addressed using supporting evidence and subject matter expertise. The framework can be used for model justification, model assessment, and validation planning. While motivated by biomedical models, it is relevant to a broad range of disciplines and underlying physics. Finally, the proposed applicability framework could help overcome some of the barriers inherent to validation of, and aid clinical implementation of, biomedical models.« less
NASA Technical Reports Server (NTRS)
Arnold, S. M.; Binienda, W. K.; Tan, H. Q.; Xu, M. H.
1992-01-01
Analytical derivations of stress intensity factors (SIF's) of a multicracked plate can be complex and tedious. Recent advances, however, in intelligent application of symbolic computation can overcome these difficulties and provide the means to rigorously and efficiently analyze this class of problems. Here, the symbolic algorithm required to implement the methodology described in Part 1 is presented. The special problem-oriented symbolic functions to derive the fundamental kernels are described, and the associated automatically generated FORTRAN subroutines are given. As a result, a symbolic/FORTRAN package named SYMFRAC, capable of providing accurate SIF's at each crack tip, was developed and validated. Simple illustrative examples using SYMFRAC show the potential of the present approach for predicting the macrocrack propagation path due to existing microcracks in the vicinity of a macrocrack tip, when the influence of the microcrack's location, orientation, size, and interaction are taken into account.
Boundary layer simulator improvement
NASA Technical Reports Server (NTRS)
Praharaj, Sarat C.; Schmitz, Craig P.; Nouri, Joseph A.
1989-01-01
Boundary Layer Integral Matrix Procedure (BLIMPJ) has been identified by the propulsion community as the rigorous boundary layer program in connection with the existing JANNAF reference programs. The improvements made to BLIMPJ and described herein have potential applications in the design of the future Orbit Transfer Vehicle engines. The turbulence model is validated to include the effects of wall roughness and a way is devised to treat multiple smooth-rough surfaces. A prediction of relaminarization regions is examined as is the combined effects of wall cooling and surface roughness on relaminarization. A turbulence model to represent the effects of constant condensed phase loading is given. A procedure is described for thrust decrement calculation in thick boundary layers by coupling the T-D Kinetics Program and BLIMPJ and a way is provided for thrust loss optimization. Potential experimental studies in rocket nozzles are identified along with the required instrumentation to provide accurate measurements in support of the presented new analytical models.
Eremenco, Sonya; Pease, Sheryl; Mann, Sarah; Berry, Pamela
2017-01-01
This paper describes the rationale and goals of the Patient-Reported Outcome (PRO) Consortium's instrument translation process. The PRO Consortium has developed a number of novel PRO measures which are in the process of qualification by the U.S. Food and Drug Administration (FDA) for use in clinical trials where endpoints based on these measures would support product labeling claims. Given the importance of FDA qualification of these measures, the PRO Consortium's Process Subcommittee determined that a detailed linguistic validation (LV) process was necessary to ensure that all translations of Consortium-developed PRO measures are performed using a standardized approach with the rigor required to meet regulatory and pharmaceutical industry expectations, as well as having a clearly defined instrument translation process that the translation industry can support. The consensus process involved gathering information about current best practices from 13 translation companies with expertise in LV, consolidating the findings to generate a proposed process, and obtaining iterative feedback from the translation companies and PRO Consortium member firms on the proposed process in two rounds of review in order to update existing principles of good practice in LV and to provide sufficient detail for the translation process to ensure consistency across PRO Consortium measures, sponsors, and translation companies. The consensus development resulted in a 12-step process that outlines universal and country-specific new translation approaches, as well as country-specific adaptations of existing translations. The PRO Consortium translation process will play an important role in maintaining the validity of the data generated through these measures by ensuring that they are translated by qualified linguists following a standardized and rigorous process that reflects best practice.
Huebner, David M; Perry, Nicholas S
2015-10-01
Behavioral interventions to reduce sexual risk behavior depend on strong health behavior theory. By identifying the psychosocial variables that lead causally to sexual risk, theories provide interventionists with a guide for how to change behavior. However, empirical research is critical to determining whether a particular theory adequately explains sexual risk behavior. A large body of cross-sectional evidence, which has been reviewed elsewhere, supports the notion that certain theory-based constructs (e.g., self-efficacy) are correlates of sexual behavior. However, given the limitations of inferring causality from correlational research, it is essential that we review the evidence from more methodologically rigorous studies (i.e., longitudinal and experimental designs). This systematic review identified 44 longitudinal studies in which investigators attempted to predict sexual risk from psychosocial variables over time. We also found 134 experimental studies (i.e., randomized controlled trials of HIV interventions), but of these only 9 (6.7 %) report the results of mediation analyses that might provide evidence for the validity of health behavior theories in predicting sexual behavior. Results show little convergent support across both types of studies for most traditional, theoretical predictors of sexual behavior. This suggests that the field must expand the body of empirical work that utilizes the most rigorous study designs to test our theoretical assumptions. The inconsistent results of existing research would indicate that current theoretical models of sexual risk behavior are inadequate, and may require expansion or adaptation.
A toolbox of immunoprecipitation-grade monoclonal antibodies to human transcription factors.
Venkataraman, Anand; Yang, Kun; Irizarry, Jose; Mackiewicz, Mark; Mita, Paolo; Kuang, Zheng; Xue, Lin; Ghosh, Devlina; Liu, Shuang; Ramos, Pedro; Hu, Shaohui; Bayron Kain, Diane; Keegan, Sarah; Saul, Richard; Colantonio, Simona; Zhang, Hongyan; Behn, Florencia Pauli; Song, Guang; Albino, Edisa; Asencio, Lillyann; Ramos, Leonardo; Lugo, Luvir; Morell, Gloriner; Rivera, Javier; Ruiz, Kimberly; Almodovar, Ruth; Nazario, Luis; Murphy, Keven; Vargas, Ivan; Rivera-Pacheco, Zully Ann; Rosa, Christian; Vargas, Moises; McDade, Jessica; Clark, Brian S; Yoo, Sooyeon; Khambadkone, Seva G; de Melo, Jimmy; Stevanovic, Milanka; Jiang, Lizhi; Li, Yana; Yap, Wendy Y; Jones, Brittany; Tandon, Atul; Campbell, Elliot; Montelione, Gaetano T; Anderson, Stephen; Myers, Richard M; Boeke, Jef D; Fenyö, David; Whiteley, Gordon; Bader, Joel S; Pino, Ignacio; Eichinger, Daniel J; Zhu, Heng; Blackshaw, Seth
2018-03-19
A key component of efforts to address the reproducibility crisis in biomedical research is the development of rigorously validated and renewable protein-affinity reagents. As part of the US National Institutes of Health (NIH) Protein Capture Reagents Program (PCRP), we have generated a collection of 1,406 highly validated immunoprecipitation- and/or immunoblotting-grade mouse monoclonal antibodies (mAbs) to 737 human transcription factors, using an integrated production and validation pipeline. We used HuProt human protein microarrays as a primary validation tool to identify mAbs with high specificity for their cognate targets. We further validated PCRP mAbs by means of multiple experimental applications, including immunoprecipitation, immunoblotting, chromatin immunoprecipitation followed by sequencing (ChIP-seq), and immunohistochemistry. We also conducted a meta-analysis that identified critical variables that contribute to the generation of high-quality mAbs. All validation data, protocols, and links to PCRP mAb suppliers are available at http://proteincapture.org.
ERIC Educational Resources Information Center
Manthey, George
2005-01-01
The most effective teaching strategies require higher order thinking, but the most used strategies seem to involve lower order thinking. If a comparison could be made between the cognitive rigor of the content standards that students are to be learning and the cognitive rigor of the actual work students are doing, then these kind of data are…
Inter-Comparison of SMAP, SMOS and GCOM-W Soil Moisture Products
NASA Astrophysics Data System (ADS)
Bindlish, R.; Jackson, T. J.; Chan, S.; Burgin, M. S.; Colliander, A.; Cosh, M. H.
2016-12-01
The Soil Moisture Active Passive (SMAP) mission was launched on Jan 31, 2015. The goal of the SMAP mission is to produce soil moisture with accuracy better than 0.04 m3/m3 with a revisit frequency of 2-3 days. The validated standard SMAP passive soil moisture product (L2SMP) with a spatial resolution of 36 km was released in May 2016. Soil moisture observations from in situ sensors are typically used to validate the satellite estimates. But, in situ observations provide ground truth for limited amount of landcover and climatic conditions. Although each mission will have its own issues, observations by other satellite instruments can be play a role in the calibration and validation of SMAP. SMAP, SMOS and GCOM-W missions share some commonnalities because they are currently providing operational brightness temperature and soil moisture products. SMAP and SMOS operate at L-band but GCOM-W uses X-band observations for soil moisture estimation. All these missions use different ancillary data sources, parameterization and algorithm to retrieve soil moisture. Therefore, it is important to validate and to compare the consistency of these products. Soil moisture products from the different missions will be compared with the in situ observations. SMAP soil moisture products will be inter-compared at global scales with SMOS and GCOM-W soil moisture products. The major contribution of satellite product inter-comparison is that it allows the assessment of the quality of the products over wider geographical and climate domains. Rigorous assessment will lead to a more reliable and accurate soil moisture product from all the missions.
Prejudice reduction: what works? A review and assessment of research and practice.
Paluck, Elizabeth Levy; Green, Donald P
2009-01-01
This article reviews the observational, laboratory, and field experimental literatures on interventions for reducing prejudice. Our review places special emphasis on assessing the methodological rigor of existing research, calling attention to problems of design and measurement that threaten both internal and external validity. Of the hundreds of studies we examine, a small fraction speak convincingly to the questions of whether, why, and under what conditions a given type of intervention works. We conclude that the causal effects of many widespread prejudice-reduction interventions, such as workplace diversity training and media campaigns, remain unknown. Although some intergroup contact and cooperation interventions appear promising, a much more rigorous and broad-ranging empirical assessment of prejudice-reduction strategies is needed to determine what works.
Mundy, Lily R; Miller, H Catherine; Klassen, Anne F; Cano, Stefan J; Pusic, Andrea L
2016-10-01
Patient-reported outcomes (PROs) are of growing importance in research and clinical care and may be used as primary outcomes or as compliments to traditional surgical outcomes. In assessing the impact of surgical and traumatic scars, PROs are often the most meaningful. To assess outcomes from the patient perspective, rigorously developed and validated PRO instruments are essential. The authors conducted a systematic literature review to identify PRO instruments developed and/or validated for patients with surgical and/or non-burn traumatic scars. Identified instruments were assessed for content, development process, and validation under recommended guidelines for PRO instrument development. The systematic review identified 6534 articles. After review, we identified four PRO instruments meeting inclusion criteria: patient and observer scar assessment scale (POSAS), bock quality of life questionnaire for patients with keloid and hypertrophic scarring (Bock), patient scar assessment questionnaire (PSAQ), and patient-reported impact of scars measure (PRISM). Common concepts measured were symptoms and psychosocial well-being. Only PSAQ had a dedicated appearance domain. Qualitative data were used to inform content for the PSAQ and PRISM, and a modern psychometric approach (Rasch Measurement Theory) was used to develop PRISM and to test POSAS. Overall, PRISM demonstrated the most rigorous design and validation process, however, was limited by the lack of a dedicated appearance domain. PRO instruments to evaluate outcomes in scars exist but vary in terms of concepts measured and psychometric soundness. This review discusses the strengths and weaknesses of existing instruments, highlighting the need for future scar-focused PRO instrument development. This journal requires that authors assign a level of evidence to each article. For a full description of these Evidence-Based Medicine ratings, please refer to Table of Contents or the online Instructions to Authors www.springer.com/00266 .
Trends: Rigor Mortis in the Arts.
ERIC Educational Resources Information Center
Blodget, Alden S.
1991-01-01
Outlines how past art education provided a refuge for students from the rigors of other academic subjects. Observes that in recent years art education has become "discipline based." Argues that art educators need to reaffirm their commitment to a humanistic way of knowing. (KM)
Mobile mental health: a challenging research agenda.
Olff, Miranda
2015-01-01
The field of mobile health ("m-Health") is evolving rapidly and there is an explosive growth of psychological tools on the market. Exciting high-tech developments may identify symptoms, help individuals manage their own mental health, encourage help seeking, and provide both preventive and therapeutic interventions. This development has the potential to be an efficient cost-effective approach reducing waiting lists and serving a considerable portion of people globally ("g-Health"). However, few of the mobile applications (apps) have been rigorously evaluated. There is little information on how valid screening and assessment tools are, which of the mobile intervention apps are effective, or how well mobile apps compare to face-to-face treatments. But how feasible is rigorous scientific evaluation with the rising demands from policy makers, business partners, and users for their quick release? In this paper, developments in m-Health tools-targeting screening, assessment, prevention, and treatment-are reviewed with examples from the field of trauma and posttraumatic stress disorder. The academic challenges in developing and evaluating m-Health tools are being addressed. Evidence-based guidance is needed on appropriate research designs that may overcome some of the public and ethical challenges (e.g., equity, availability) and the market-driven wish to have mobile apps in the "App Store" yesterday rather than tomorrow.
Enhancing patient-doctor-computer communication in primary care: towards measurement construction.
Assis-Hassid, Shiri; Reychav, Iris; Heart, Tsipi; Pliskin, Joseph S; Reis, Shmuel
2015-01-01
The traditional dyadic dynamics of the medical encounter has been altered into a triadic relationship by introducing the computer into the examination room. This study defines Patient-Doctor-Computer Communication (PDCC) as a new construct and provides an initial validation process of an instrument for assessing PDCC in the computerized exam room: the e-SEGUE. Based on the existing literature, a new construct, PDCC, is defined as the physician's ability to provide patient-centered care while using the computer during the medical encounter. This study elucidates 27 PDCC-related behaviors from the relevant literature and state of the art models of PDCC. These were embedded in the SEGUE communication assessment framework to form the e-SEGUE, a communication skills assessment tool that integrates computer-related communication skills. Based on Mackenzie et al.'s methodological approach of measurement construction, we conducted a two-phased content validity analysis by a general and expert panels of the PDCC behaviors represented in the e-SEGUE. This study was carried out in an environment where EMR use is universal and fully integrated in the physicians' workflow. The panels consisted of medical students, residents, primary care physicians, healthcare leaders and faculty of medicine members, who rated and provided input regarding the 27 behaviors. Overall, results show high level of agreement with 23 PDCC-related behaviors. The PDCC instrument developed in this study, the e-SEGUE, fared well in a rigorous, albeit initial, validation process has a unique potential for training and enhancing patient-doctor communication (PDC) in the computerized examination room pending further development.
Hartman, Nicholas; Wittler, Mary; Askew, Kim; Manthey, David
2016-01-01
Placement of ultrasound-guided central lines is a critical skill for physicians in several specialties. Improving the quality of care delivered surrounding this procedure demands rigorous measurement of competency, and validated tools to assess performance are essential. Using the iterative, modified Delphi technique and experts in multiple disciplines across the United States, the study team created a 30-item checklist designed to assess competency in the placement of ultrasound-guided internal jugular central lines. Cronbach α was .94, indicating an excellent degree of internal consistency. Further validation of this checklist will require its implementation in simulated and clinical environments. © The Author(s) 2014.
Guidance for updating clinical practice guidelines: a systematic review of methodological handbooks.
Vernooij, Robin W M; Sanabria, Andrea Juliana; Solà, Ivan; Alonso-Coello, Pablo; Martínez García, Laura
2014-01-02
Updating clinical practice guidelines (CPGs) is a crucial process for maintaining the validity of recommendations. Methodological handbooks should provide guidance on both developing and updating CPGs. However, little is known about the updating guidance provided by these handbooks. We conducted a systematic review to identify and describe the updating guidance provided by CPG methodological handbooks and included handbooks that provide updating guidance for CPGs. We searched in the Guidelines International Network library, US National Guidelines Clearinghouse and MEDLINE (PubMed) from 1966 to September 2013. Two authors independently selected the handbooks and extracted the data. We used descriptive statistics to analyze the extracted data and conducted a narrative synthesis. We included 35 handbooks. Most handbooks (97.1%) focus mainly on developing CPGs, including variable degrees of information about updating. Guidance on identifying new evidence and the methodology of assessing the need for an update is described in 11 (31.4%) and eight handbooks (22.8%), respectively. The period of time between two updates is described in 25 handbooks (71.4%), two to three years being the most frequent (40.0%). The majority of handbooks do not provide guidance for the literature search, evidence selection, assessment, synthesis, and external review of the updating process. Guidance for updating CPGs is poorly described in methodological handbooks. This guidance should be more rigorous and explicit. This could lead to a more optimal updating process, and, ultimately to valid trustworthy guidelines.
A Rigorous Curriculum Really Matters
ERIC Educational Resources Information Center
Cook, Erika
2013-01-01
As every good secondary administrator knows, rigorous curricula matter. Challenging curricula is the factor in lifting each student to reach their potential: "the academic intensity of the student's high school curriculum still counts more than anything else...in providing momentum toward completing a bachelor's degree" (Adelman, 2006,…
Blended Learning Improves Science Education.
Stockwell, Brent R; Stockwell, Melissa S; Cennamo, Michael; Jiang, Elise
2015-08-27
Blended learning is an emerging paradigm for science education but has not been rigorously assessed. We performed a randomized controlled trial of blended learning. We found that in-class problem solving improved exam performance, and video assignments increased attendance and satisfaction. This validates a new model for science communication and education. Copyright © 2015 Elsevier Inc. All rights reserved.
ERIC Educational Resources Information Center
Sanetti, Lisa M. Hagermoser; Collier-Meek, Melissa A.
2014-01-01
Although it is widely accepted that procedural fidelity data are important for making valid decisions about intervention effectiveness, there is little empirical guidance for researchers and practitioners regarding how to assess procedural fidelity. A first step in moving procedural fidelity assessment research forward is to develop a…
The Validity of the Academic Rigor Index (ARI) for Predicting FYGPA. Research Report 2012-5
ERIC Educational Resources Information Center
Mattern, Krista D.; Wyatt, Jeffrey N.
2012-01-01
A recurrent trend in higher education research has been to identify additional predictors of college success beyond the traditional measures of high school grade point average (HSGPA) and standardized test scores, given that a large percentage of unaccounted variance in college performance remains. A recent study by Wyatt, Wiley, Camara, and…
ERIC Educational Resources Information Center
Cole, Russell; Deke, John; Seftor, Neil
2016-01-01
The What Works Clearinghouse (WWC) maintains design standards to identify rigorous, internally valid education research. As education researchers advance new methodologies, the WWC must revise its standards to include an assessment of the new designs. Recently, the WWC has revised standards for two emerging study designs: regression discontinuity…
ERIC Educational Resources Information Center
Gallagher, Carole; Rabinowitz, Stanley; Yeagley, Pamela
2011-01-01
Researchers recommend that policymakers use data from multiple sources when making decisions that have high-stakes consequences (Herman, Baker, & Linn, 2004; Linn, 2007; Stone & Lane, 2003). For this reason, a fair but rigorous teacher-effectiveness rating process relies on evidence collected from different sources (Goe, Bell, & Little, 2008;…
ERIC Educational Resources Information Center
Acharya, Sushil; Manohar, Priyadarshan Anant; Wu, Peter; Maxim, Bruce; Hansen, Mary
2018-01-01
Active learning tools are critical in imparting real world experiences to the students within a classroom environment. This is important because graduates are expected to develop software that meets rigorous quality standards in functional and application domains with little to no training. However, there is a well-recognized need for the…
Opportunity Knocks! Inquiry, the New National Social Studies and Science Standards, and You
ERIC Educational Resources Information Center
Ratzer, Mary Boyd
2014-01-01
To be recognized and adopted, the promise and potential of national standards must validate the instructional role of the school librarian and build formative knowledge through authentic process and products. Inquiry is a direct link to those dimensions. Standards that are rigorous and relevant optimize the developing brain's affinity for building…
Good Evaluation Measures: More than Their Psychometric Properties
ERIC Educational Resources Information Center
Weitzman, Beth C.; Silver, Diana
2013-01-01
In this commentary, we examine Braverman's insights into the trade-offs between feasibility and rigor in evaluation measures and reject his assessment of the trade-off as a zero-sum game. We, argue that feasibility and policy salience are, like reliability and validity, intrinsic to the definition of a good measure. To reduce the tension between…
Validation of Self-Image of Aging Scale for Chinese Elders
ERIC Educational Resources Information Center
Bai, Xue; Chan, K. S.; Chow, Nelson
2012-01-01
Researchers are increasingly interested in the "image of aging" concept. Models on the image of aging abound, but few have rigorously tested measures that are culturally sensitive and domain-specific. This study first translates Levy et al.'s (2004) Image of Aging Scale into the Chinese language and revises it into the Chinese Version of…
ERIC Educational Resources Information Center
Flanagan, Dawn P.; Alfonso, Vincent C.; Reynolds, Matthew R.
2013-01-01
In this commentary, we reviewed two clinical validation studies on the Wechsler Scales conducted by Weiss and colleagues. These researchers used a rigorous within-battery model-fitting approach that demonstrated the factorial invariance of the Wechsler Intelligence Scale for Children-Fourth Edition (WISC-IV) and Wechsler Adult Intelligence…
Validation Engine for Observational Protocols. Measures of Effective Teaching (MET) Project
ERIC Educational Resources Information Center
Bill & Melinda Gates Foundation, 2010
2010-01-01
In the fall of 2009, the Bill and Melinda Gates Foundation launched the two-year Measures of Effective Teaching (MET) project to rigorously develop and test multiple measures of teacher effectiveness. As part of the project, partners from more than a dozen reputable academic, non-profit and for-profit organizations collected and analyzed data from…
An analytical method to identify and quantify trace levels of C5 to C12 perfluorocarboxylic acids (PFCAs) in articles of commerce (AOC) is developed and rigorously validated. Solid samples were extracted in methanol, and liquid samples were diluted with a solvent consisting of 60...
ERIC Educational Resources Information Center
Condon, Christopher; Clifford, Matthew
2012-01-01
Assessing school principal performance is both necessary and challenging. During the past five years, many states have begun using validated measures in summative assessments of novice principal competency as a basis for certification decisions. Although standardized tests are used for certification purposes, other types of assessments are being…
HA, Mei; QIAN, Xiaoling; YANG, Hong; HUANG, Jichun; LIU, Changjiang
2016-01-01
Background: The public’s cognition of stroke and responses to stroke symptoms are important to prevent complications and decrease the mortality when stroke occurs. The aim of study was to develop and validate the Chinese version of the Stroke Action Test (C-STAT) in a Chinese population. Methods: This study was rigorously implemented with the published guideline for the translation, adaptation and validation of instruments for the cross-cultural use in healthcare care research. A cross-sectional study was performed among 328 stroke patients and family members in the Department of Neurology in the Second Hospital of Lanzhou University, Gansu province, China in 2014. Results: The Chinese version of the instrument showed favorable content equivalence with the source version. Values of Cronbach’s alpha and test-retest reliability of the C-STAT were 0.88 and 0.86, respectively. Principal component analysis supported four-factor solutions of the C-STAT. Criterion-related validity showed that the C-STAT was a significant predictor of the 7-item stroke symptom scores (R = 0.77; t = 21.74, P< 0.001). Conclusion: The C-STAT is an intelligible and brief psychometrical tool to assess individuals’ knowledge of the appropriate responses to stroke symptoms in Chinese populations. It could also be used by health care providers to assess educational programs on stroke prevention. PMID:28053925
NASA Astrophysics Data System (ADS)
Yu, Hesheng; Thé, Jesse
2016-11-01
The prediction of the dispersion of air pollutants in urban areas is of great importance to public health, homeland security, and environmental protection. Computational Fluid Dynamics (CFD) emerges as an effective tool for pollutant dispersion modelling. This paper reports and quantitatively validates the shear stress transport (SST) k-ω turbulence closure model and its transitional variant for pollutant dispersion under complex urban environment for the first time. Sensitivity analysis is performed to establish recommendation for the proper use of turbulence models in urban settings. The current SST k-ω simulation is validated rigorously by extensive experimental data using hit rate for velocity components, and the "factor of two" of observations (FAC2) and fractional bias (FB) for concentration field. The simulation results show that current SST k-ω model can predict flow field nicely with an overall hit rate of 0.870, and concentration dispersion with FAC2 = 0.721 and FB = 0.045. The flow simulation of the current SST k-ω model is slightly inferior to that of a detached eddy simulation (DES), but better than that of standard k-ε model. However, the current study is the best among these three model approaches, when validated against measurements of pollutant dispersion in the atmosphere. This work aims to provide recommendation for proper use of CFD to predict pollutant dispersion in urban environment.
Update on simulation-based surgical training and assessment in ophthalmology: a systematic review.
Thomsen, Ann Sofia S; Subhi, Yousif; Kiilgaard, Jens Folke; la Cour, Morten; Konge, Lars
2015-06-01
This study reviews the evidence behind simulation-based surgical training of ophthalmologists to determine (1) the validity of the reported models and (2) the ability to transfer skills to the operating room. Simulation-based training is established widely within ophthalmology, although it often lacks a scientific basis for implementation. We conducted a systematic review of trials involving simulation-based training or assessment of ophthalmic surgical skills among health professionals. The search included 5 databases (PubMed, EMBASE, PsycINFO, Cochrane Library, and Web of Science) and was completed on March 1, 2014. Overall, the included trials were divided into animal, cadaver, inanimate, and virtual-reality models. Risk of bias was assessed using the Cochrane Collaboration's tool. Validity evidence was evaluated using a modern validity framework (Messick's). We screened 1368 reports for eligibility and included 118 trials. The most common surgery simulated was cataract surgery. Most validity trials investigated only 1 or 2 of 5 sources of validity (87%). Only 2 trials (48 participants) investigated transfer of skills to the operating room; 4 trials (65 participants) evaluated the effect of simulation-based training on patient-related outcomes. Because of heterogeneity of the studies, it was not possible to conduct a quantitative analysis. The methodologic rigor of trials investigating simulation-based surgical training in ophthalmology is inadequate. To ensure effective implementation of training models, evidence-based knowledge of validity and efficacy is needed. We provide a useful tool for implementation and evaluation of research in simulation-based training. Copyright © 2015 American Academy of Ophthalmology. Published by Elsevier Inc. All rights reserved.
Earth as an Extrasolar Planet: Earth Model Validation Using EPOXI Earth Observations
NASA Astrophysics Data System (ADS)
Robinson, Tyler D.; Meadows, Victoria S.; Crisp, David; Deming, Drake; A'Hearn, Michael F.; Charbonneau, David; Livengood, Timothy A.; Seager, Sara; Barry, Richard K.; Hearty, Thomas; Hewagama, Tilak; Lisse, Carey M.; McFadden, Lucy A.; Wellnitz, Dennis D.
2011-06-01
The EPOXI Discovery Mission of Opportunity reused the Deep Impact flyby spacecraft to obtain spatially and temporally resolved visible photometric and moderate resolution near-infrared (NIR) spectroscopic observations of Earth. These remote observations provide a rigorous validation of whole-disk Earth model simulations used to better understand remotely detectable extrasolar planet characteristics. We have used these data to upgrade, correct, and validate the NASA Astrobiology Institute's Virtual Planetary Laboratory three-dimensional line-by-line, multiple-scattering spectral Earth model. This comprehensive model now includes specular reflectance from the ocean and explicitly includes atmospheric effects such as Rayleigh scattering, gas absorption, and temperature structure. We have used this model to generate spatially and temporally resolved synthetic spectra and images of Earth for the dates of EPOXI observation. Model parameters were varied to yield an optimum fit to the data. We found that a minimum spatial resolution of ∼100 pixels on the visible disk, and four categories of water clouds, which were defined by using observed cloud positions and optical thicknesses, were needed to yield acceptable fits. The validated model provides a simultaneous fit to Earth's lightcurve, absolute brightness, and spectral data, with a root-mean-square (RMS) error of typically less than 3% for the multiwavelength lightcurves and residuals of ∼10% for the absolute brightness throughout the visible and NIR spectral range. We have extended our validation into the mid-infrared by comparing the model to high spectral resolution observations of Earth from the Atmospheric Infrared Sounder, obtaining a fit with residuals of ∼7% and brightness temperature errors of less than 1 K in the atmospheric window. For the purpose of understanding the observable characteristics of the distant Earth at arbitrary viewing geometry and observing cadence, our validated forward model can be used to simulate Earth's time-dependent brightness and spectral properties for wavelengths from the far ultraviolet to the far infrared.
Earth as an extrasolar planet: Earth model validation using EPOXI earth observations.
Robinson, Tyler D; Meadows, Victoria S; Crisp, David; Deming, Drake; A'hearn, Michael F; Charbonneau, David; Livengood, Timothy A; Seager, Sara; Barry, Richard K; Hearty, Thomas; Hewagama, Tilak; Lisse, Carey M; McFadden, Lucy A; Wellnitz, Dennis D
2011-06-01
The EPOXI Discovery Mission of Opportunity reused the Deep Impact flyby spacecraft to obtain spatially and temporally resolved visible photometric and moderate resolution near-infrared (NIR) spectroscopic observations of Earth. These remote observations provide a rigorous validation of whole-disk Earth model simulations used to better understand remotely detectable extrasolar planet characteristics. We have used these data to upgrade, correct, and validate the NASA Astrobiology Institute's Virtual Planetary Laboratory three-dimensional line-by-line, multiple-scattering spectral Earth model. This comprehensive model now includes specular reflectance from the ocean and explicitly includes atmospheric effects such as Rayleigh scattering, gas absorption, and temperature structure. We have used this model to generate spatially and temporally resolved synthetic spectra and images of Earth for the dates of EPOXI observation. Model parameters were varied to yield an optimum fit to the data. We found that a minimum spatial resolution of ∼100 pixels on the visible disk, and four categories of water clouds, which were defined by using observed cloud positions and optical thicknesses, were needed to yield acceptable fits. The validated model provides a simultaneous fit to Earth's lightcurve, absolute brightness, and spectral data, with a root-mean-square (RMS) error of typically less than 3% for the multiwavelength lightcurves and residuals of ∼10% for the absolute brightness throughout the visible and NIR spectral range. We have extended our validation into the mid-infrared by comparing the model to high spectral resolution observations of Earth from the Atmospheric Infrared Sounder, obtaining a fit with residuals of ∼7% and brightness temperature errors of less than 1 K in the atmospheric window. For the purpose of understanding the observable characteristics of the distant Earth at arbitrary viewing geometry and observing cadence, our validated forward model can be used to simulate Earth's time-dependent brightness and spectral properties for wavelengths from the far ultraviolet to the far infrared. Key Words: Astrobiology-Extrasolar terrestrial planets-Habitability-Planetary science-Radiative transfer. Astrobiology 11, 393-408.
Earth as an Extrasolar Planet: Earth Model Validation Using EPOXI Earth Observations
NASA Technical Reports Server (NTRS)
Robinson, Tyler D.; Meadows, Victoria S.; Crisp, David; Deming, Drake; A'Hearn, Michael F.; Charbonneau, David; Livengood, Timothy A.; Seager, Sara; Barry, Richard; Hearty, Thomas;
2011-01-01
The EPOXI Discovery Mission of Opportunity reused the Deep Impact flyby spacecraft to obtain spatially and temporally resolved visible photometric and moderate resolution near-infrared (NIR) spectroscopic observations of Earth. These remote observations provide a rigorous validation of whole disk Earth model simulations used to better under- stand remotely detectable extrasolar planet characteristics. We have used these data to upgrade, correct, and validate the NASA Astrobiology Institute s Virtual Planetary Laboratory three-dimensional line-by-line, multiple-scattering spectral Earth model (Tinetti et al., 2006a,b). This comprehensive model now includes specular reflectance from the ocean and explicitly includes atmospheric effects such as Rayleigh scattering, gas absorption, and temperature structure. We have used this model to generate spatially and temporally resolved synthetic spectra and images of Earth for the dates of EPOXI observation. Model parameters were varied to yield an optimum fit to the data. We found that a minimum spatial resolution of approx.100 pixels on the visible disk, and four categories of water clouds, which were defined using observed cloud positions and optical thicknesses, were needed to yield acceptable fits. The validated model provides a simultaneous fit to the Earth s lightcurve, absolute brightness, and spectral data, with a root-mean-square error of typically less than 3% for the multiwavelength lightcurves, and residuals of approx.10% for the absolute brightness throughout the visible and NIR spectral range. We extend our validation into the mid-infrared by comparing the model to high spectral resolution observations of Earth from the Atmospheric Infrared Sounder, obtaining a fit with residuals of approx.7%, and temperature errors of less than 1K in the atmospheric window. For the purpose of understanding the observable characteristics of the distant Earth at arbitrary viewing geometry and observing cadence, our validated forward model can be used to simulate Earth s time dependent brightness and spectral properties for wavelengths from the far ultraviolet to the far infrared.brightness
Clarity versus complexity: land-use modeling as a practical tool for decision-makers
Sohl, Terry L.; Claggett, Peter
2013-01-01
The last decade has seen a remarkable increase in the number of modeling tools available to examine future land-use and land-cover (LULC) change. Integrated modeling frameworks, agent-based models, cellular automata approaches, and other modeling techniques have substantially improved the representation of complex LULC systems, with each method using a different strategy to address complexity. However, despite the development of new and better modeling tools, the use of these tools is limited for actual planning, decision-making, or policy-making purposes. LULC modelers have become very adept at creating tools for modeling LULC change, but complicated models and lack of transparency limit their utility for decision-makers. The complicated nature of many LULC models also makes it impractical or even impossible to perform a rigorous analysis of modeling uncertainty. This paper provides a review of land-cover modeling approaches and the issues causes by the complicated nature of models, and provides suggestions to facilitate the increased use of LULC models by decision-makers and other stakeholders. The utility of LULC models themselves can be improved by 1) providing model code and documentation, 2) through the use of scenario frameworks to frame overall uncertainties, 3) improving methods for generalizing key LULC processes most important to stakeholders, and 4) adopting more rigorous standards for validating models and quantifying uncertainty. Communication with decision-makers and other stakeholders can be improved by increasing stakeholder participation in all stages of the modeling process, increasing the transparency of model structure and uncertainties, and developing user-friendly decision-support systems to bridge the link between LULC science and policy. By considering these options, LULC science will be better positioned to support decision-makers and increase real-world application of LULC modeling results.
NASA Technical Reports Server (NTRS)
Straight, Christian L.; Bubenheim, David L.; Bates, Maynard E.; Flynn, Michael T.
1994-01-01
CELSS Antarctic Analog Project (CAAP) represents a logical solution to the multiple objectives of both the NASA and the National Science Foundation (NSF). CAAP will result in direct transfer of proven technologies and systems, proven under the most rigorous of conditions, to the NSF and to society at large. This project goes beyond, as it must, the generally accepted scope of CELSS and life support systems including the issues of power generation, human dynamics, community systems, and training. CAAP provides a vivid and starkly realistic testbed of Controlled Ecological Life Support System (CELSS) and life support systems and methods. CAAP will also be critical in the development and validation of performance parameters for future advanced life support systems.
Numerical Approach to Spatial Deterministic-Stochastic Models Arising in Cell Biology.
Schaff, James C; Gao, Fei; Li, Ye; Novak, Igor L; Slepchenko, Boris M
2016-12-01
Hybrid deterministic-stochastic methods provide an efficient alternative to a fully stochastic treatment of models which include components with disparate levels of stochasticity. However, general-purpose hybrid solvers for spatially resolved simulations of reaction-diffusion systems are not widely available. Here we describe fundamentals of a general-purpose spatial hybrid method. The method generates realizations of a spatially inhomogeneous hybrid system by appropriately integrating capabilities of a deterministic partial differential equation solver with a popular particle-based stochastic simulator, Smoldyn. Rigorous validation of the algorithm is detailed, using a simple model of calcium 'sparks' as a testbed. The solver is then applied to a deterministic-stochastic model of spontaneous emergence of cell polarity. The approach is general enough to be implemented within biologist-friendly software frameworks such as Virtual Cell.
Functional differentiability in time-dependent quantum mechanics.
Penz, Markus; Ruggenthaler, Michael
2015-03-28
In this work, we investigate the functional differentiability of the time-dependent many-body wave function and of derived quantities with respect to time-dependent potentials. For properly chosen Banach spaces of potentials and wave functions, Fréchet differentiability is proven. From this follows an estimate for the difference of two solutions to the time-dependent Schrödinger equation that evolve under the influence of different potentials. Such results can be applied directly to the one-particle density and to bounded operators, and present a rigorous formulation of non-equilibrium linear-response theory where the usual Lehmann representation of the linear-response kernel is not valid. Further, the Fréchet differentiability of the wave function provides a new route towards proving basic properties of time-dependent density-functional theory.
Does McNemar's test compare the sensitivities and specificities of two diagnostic tests?
Kim, Soeun; Lee, Woojoo
2017-02-01
McNemar's test is often used in practice to compare the sensitivities and specificities for the evaluation of two diagnostic tests. For correct evaluation of accuracy, an intuitive recommendation is to test the diseased and the non-diseased groups separately so that the sensitivities can be compared among the diseased, and specificities can be compared among the healthy group of people. This paper provides a rigorous theoretical framework for this argument and study the validity of McNemar's test regardless of the conditional independence assumption. We derive McNemar's test statistic under the null hypothesis considering both assumptions of conditional independence and conditional dependence. We then perform power analyses to show how the result is affected by the amount of the conditional dependence under alternative hypothesis.
Regenerative medicine blueprint.
Terzic, Andre; Harper, C Michel; Gores, Gregory J; Pfenning, Michael A
2013-12-01
Regenerative medicine, a paragon of future healthcare, holds unprecedented potential in extending the reach of treatment modalities for individuals across diseases and lifespan. Emerging regenerative technologies, focused on structural repair and functional restoration, signal a radical transformation in medical and surgical practice. Regenerative medicine is poised to provide innovative solutions in addressing major unmet needs for patients, ranging from congenital disease and trauma to degenerative conditions. Realization of the regenerative model of care predicates a stringent interdisciplinary paradigm that will drive validated science into standardized clinical options. Designed as a catalyst in advancing rigorous new knowledge on disease causes and cures into informed delivery of quality care, the Mayo Clinic regenerative medicine blueprint offers a patient-centered, team-based strategy that optimizes the discovery-translation-application roadmap for the express purpose of science-supported practice advancement.
Burrus, Barri B.; Scott, Alicia Richmond
2012-01-01
Adolescent parents and their children are at increased risk for adverse short- and long-term health and social outcomes. Effective interventions are needed to support these young families. We studied the evidence base and found a dearth of rigorously evaluated programs. Strategies from successful interventions are needed to inform both intervention design and policies affecting these adolescents. The lack of rigorous evaluations may be attributable to inadequate emphasis on and sufficient funding for evaluation, as well as to challenges encountered by program evaluators working with this population. More rigorous program evaluations are urgently needed to provide scientifically sound guidance for programming and policy decisions. Evaluation lessons learned have implications for other vulnerable populations. PMID:22897541
A mind you can count on: validating breath counting as a behavioral measure of mindfulness.
Levinson, Daniel B; Stoll, Eli L; Kindy, Sonam D; Merry, Hillary L; Davidson, Richard J
2014-01-01
Mindfulness practice of present moment awareness promises many benefits, but has eluded rigorous behavioral measurement. To date, research has relied on self-reported mindfulness or heterogeneous mindfulness trainings to infer skillful mindfulness practice and its effects. In four independent studies with over 400 total participants, we present the first construct validation of a behavioral measure of mindfulness, breath counting. We found it was reliable, correlated with self-reported mindfulness, differentiated long-term meditators from age-matched controls, and was distinct from sustained attention and working memory measures. In addition, we employed breath counting to test the nomological network of mindfulness. As theorized, we found skill in breath counting associated with more meta-awareness, less mind wandering, better mood, and greater non-attachment (i.e., less attentional capture by distractors formerly paired with reward). We also found in a randomized online training study that 4 weeks of breath counting training improved mindfulness and decreased mind wandering relative to working memory training and no training controls. Together, these findings provide the first evidence for breath counting as a behavioral measure of mindfulness.
Predictive 5-Year Survivorship Model of Cystic Fibrosis
Liou, Theodore G.; Adler, Frederick R.; FitzSimmons, Stacey C.; Cahill, Barbara C.; Hibbs, Jonathan R.; Marshall, Bruce C.
2007-01-01
The objective of this study was to create a 5-year survivorship model to identify key clinical features of cystic fibrosis. Such a model could help researchers and clinicians to evaluate therapies, improve the design of prospective studies, monitor practice patterns, counsel individual patients, and determine the best candidates for lung transplantation. The authors used information from the Cystic Fibrosis Foundation Patient Registry (CFFPR), which has collected longitudinal data on approximately 90% of cystic fibrosis patients diagnosed in the United States since 1986. They developed multivariate logistic regression models by using data on 5,820 patients randomly selected from 11,630 in the CFFPR in 1993. Models were tested for goodness of fit and were validated for the remaining 5,810 patients for 1993. The validated 5-year survivorship model included age, forced expiratory volume in 1 second as a percentage of predicted normal, gender, weight-for-age z score, pancreatic sufficiency, diabetes mellitus, Staphylococcus aureus infection, Burkerholderia cepacia infection, and annual number of acute pulmonary exacerbations. The model provides insights into the complex nature of cystic fibrosis and supplies a rigorous tool for clinical practice and research. PMID:11207152
Prediction of protein subcellular locations by GO-FunD-PseAA predictor.
Chou, Kuo-Chen; Cai, Yu-Dong
2004-08-06
The localization of a protein in a cell is closely correlated with its biological function. With the explosion of protein sequences entering into DataBanks, it is highly desired to develop an automated method that can fast identify their subcellular location. This will expedite the annotation process, providing timely useful information for both basic research and industrial application. In view of this, a powerful predictor has been developed by hybridizing the gene ontology approach [Nat. Genet. 25 (2000) 25], functional domain composition approach [J. Biol. Chem. 277 (2002) 45765], and the pseudo-amino acid composition approach [Proteins Struct. Funct. Genet. 43 (2001) 246; Erratum: ibid. 44 (2001) 60]. As a showcase, the recently constructed dataset [Bioinformatics 19 (2003) 1656] was used for demonstration. The dataset contains 7589 proteins classified into 12 subcellular locations: chloroplast, cytoplasmic, cytoskeleton, endoplasmic reticulum, extracellular, Golgi apparatus, lysosomal, mitochondrial, nuclear, peroxisomal, plasma membrane, and vacuolar. The overall success rate of prediction obtained by the jackknife cross-validation was 92%. This is so far the highest success rate performed on this dataset by following an objective and rigorous cross-validation procedure.
Formalization of Generalized Constraint Language: A Crucial Prelude to Computing With Words.
Khorasani, Elham S; Rahimi, Shahram; Calvert, Wesley
2013-02-01
The generalized constraint language (GCL), introduced by Zadeh, serves as a basis for computing with words (CW). It provides an agenda to express the imprecise and fuzzy information embedded in natural language and allows reasoning with perceptions. Despite its fundamental role, the definition of GCL has remained informal since its introduction by Zadeh, and to our knowledge, no attempt has been made to formulate a rigorous theoretical framework for GCL. Such formalization is necessary for further theoretical and practical advancement of CW for two important reasons. First, it provides the underlying infrastructure for the development of useful inference patterns based on sound theories. Second, it determines the scope of GCL and hence facilitates the translation of natural language expressions into GCL. This paper is an attempt to step in this direction by providing a formal syntax together with a compositional semantics for GCL. A soundness theorem is defined, and Zadeh's deduction rules are proved to be valid in the defined semantics. Furthermore, a discussion is provided on how the proposed language may be used in practice.
Barriers to healthcare for transgender individuals.
Safer, Joshua D; Coleman, Eli; Feldman, Jamie; Garofalo, Robert; Hembree, Wylie; Radix, Asa; Sevelius, Jae
2016-04-01
Transgender persons suffer significant health disparities and may require medical intervention as part of their care. The purpose of this manuscript is to briefly review the literature characterizing barriers to healthcare for transgender individuals and to propose research priorities to understand mechanisms of those barriers and interventions to overcome them. Current research emphasizes sexual minorities' self-report of barriers, rather than using direct methods. The biggest barrier to healthcare reported by transgender individuals is lack of access because of lack of providers who are sufficiently knowledgeable on the topic. Other barriers include: financial barriers, discrimination, lack of cultural competence by providers, health systems barriers, and socioeconomic barriers. National research priorities should include rigorous determination of the capacity of the US healthcare system to provide adequate care for transgender individuals. Studies should determine knowledge and biases of the medical workforce across the spectrum of medical training with regard to transgender medical care; adequacy of sufficient providers for the care required, larger social structural barriers, and status of a framework to pay for appropriate care. As well, studies should propose and validate potential solutions to address identified gaps.
Chaabene, Helmi; Negra, Yassine; Bouguezzi, Raja; Capranica, Laura; Franchini, Emerson; Prieske, Olaf; Hbacha, Hamdi; Granacher, Urs
2018-01-01
The regular monitoring of physical fitness and sport-specific performance is important in elite sports to increase the likelihood of success in competition. This study aimed to systematically review and to critically appraise the methodological quality, validation data, and feasibility of the sport-specific performance assessment in Olympic combat sports like amateur boxing, fencing, judo, karate, taekwondo, and wrestling. A systematic search was conducted in the electronic databases PubMed, Google-Scholar, and Science-Direct up to October 2017. Studies in combat sports were included that reported validation data (e.g., reliability, validity, sensitivity) of sport-specific tests. Overall, 39 studies were eligible for inclusion in this review. The majority of studies (74%) contained sample sizes <30 subjects. Nearly, 1/3 of the reviewed studies lacked a sufficient description (e.g., anthropometrics, age, expertise level) of the included participants. Seventy-two percent of studies did not sufficiently report inclusion/exclusion criteria of their participants. In 62% of the included studies, the description and/or inclusion of a familiarization session (s) was either incomplete or not existent. Sixty-percent of studies did not report any details about the stability of testing conditions. Approximately half of the studies examined reliability measures of the included sport-specific tests (intraclass correlation coefficient [ICC] = 0.43-1.00). Content validity was addressed in all included studies, criterion validity (only the concurrent aspect of it) in approximately half of the studies with correlation coefficients ranging from r = -0.41 to 0.90. Construct validity was reported in 31% of the included studies and predictive validity in only one. Test sensitivity was addressed in 13% of the included studies. The majority of studies (64%) ignored and/or provided incomplete information on test feasibility and methodological limitations of the sport-specific test. In 28% of the included studies, insufficient information or a complete lack of information was provided in the respective field of the test application. Several methodological gaps exist in studies that used sport-specific performance tests in Olympic combat sports. Additional research should adopt more rigorous validation procedures in the application and description of sport-specific performance tests in Olympic combat sports.
Multiscale sagebrush rangeland habitat modeling in southwest Wyoming
Homer, Collin G.; Aldridge, Cameron L.; Meyer, Debra K.; Coan, Michael J.; Bowen, Zachary H.
2009-01-01
Sagebrush-steppe ecosystems in North America have experienced dramatic elimination and degradation since European settlement. As a result, sagebrush-steppe dependent species have experienced drastic range contractions and population declines. Coordinated ecosystem-wide research, integrated with monitoring and management activities, would improve the ability to maintain existing sagebrush habitats. However, current data only identify resource availability locally, with rigorous spatial tools and models that accurately model and map sagebrush habitats over large areas still unavailable. Here we report on an effort to produce a rigorous large-area sagebrush-habitat classification and inventory with statistically validated products and estimates of precision in the State of Wyoming. This research employs a combination of significant new tools, including (1) modeling sagebrush rangeland as a series of independent continuous field components that can be combined and customized by any user at multiple spatial scales; (2) collecting ground-measured plot data on 2.4-meter imagery in the same season the satellite imagery is acquired; (3) effective modeling of ground-measured data on 2.4-meter imagery to maximize subsequent extrapolation; (4) acquiring multiple seasons (spring, summer, and fall) of an additional two spatial scales of imagery (30 meter and 56 meter) for optimal large-area modeling; (5) using regression tree classification technology that optimizes data mining of multiple image dates, ratios, and bands with ancillary data to extrapolate ground training data to coarser resolution sensors; and (6) employing rigorous accuracy assessment of model predictions to enable users to understand the inherent uncertainties. First-phase results modeled eight rangeland components (four primary targets and four secondary targets) as continuous field predictions. The primary targets included percent bare ground, percent herbaceousness, percent shrub, and percent litter. The four secondary targets included percent sagebrush (Artemisia spp.), percent big sagebrush (Artemisia tridentata), percent Wyoming sagebrush (Artemisia tridentata wyomingensis), and sagebrush height (centimeters). Results were validated by an independent accuracy assessment with root mean square error (RMSE) values ranging from 6.38 percent for bare ground to 2.99 percent for sagebrush at the QuickBird scale and RMSE values ranging from 12.07 percent for bare ground to 6.34 percent for sagebrush at the full Landsat scale. Subsequent project phases are now in progress, with plans to deliver products that improve accuracies of existing components, model new components, complete models over larger areas, track changes over time (from 1988 to 2007), and ultimately model wildlife population trends against these changes. We believe these results offer significant improvement in sagebrush rangeland quantification at multiple scales and offer users products that have been rigorously validated.
Reduced and Validated Kinetic Mechanisms for Hydrogen-CO-sir Combustion in Gas Turbines
DOE Office of Scientific and Technical Information (OSTI.GOV)
Yiguang Ju; Frederick Dryer
2009-02-07
Rigorous experimental, theoretical, and numerical investigation of various issues relevant to the development of reduced, validated kinetic mechanisms for synthetic gas combustion in gas turbines was carried out - including the construction of new radiation models for combusting flows, improvement of flame speed measurement techniques, measurements and chemical kinetic analysis of H{sub 2}/CO/CO{sub 2}/O{sub 2}/diluent mixtures, revision of the H{sub 2}/O{sub 2} kinetic model to improve flame speed prediction capabilities, and development of a multi-time scale algorithm to improve computational efficiency in reacting flow simulations.
Analyses of Alternatives: Toward a More Rigorous Determination of Scope
2014-04-30
justification is required to avoid three months of wasted effort, the change is unlikely to happen. Applying a Systems View Systems Thinking and...Defense Acquisition Guidebook para. 10.5.2 IOC FOC TRL 1‐3 TRL 4 TRL 7 TRL 8 TRL 9 Compon‐ ent and/or Bread ‐ board Validation In a Relevant Environ‐ ment
ERIC Educational Resources Information Center
Hirano, Kara A.; Shanley, Lina; Garbacz, S. Andrew; Rowe, Dawn A.; Lindstrom, Lauren; Leve, Leslie D.
2018-01-01
Parent involvement is a predictor of postsecondary education and employment outcomes, but rigorous measures of parent involvement for youth with disabilities are lacking. Hirano, Garbacz, Shanley, and Rowe adapted scales based on Hoover-Dempsey and Sandler model of parent involvement for use with parents of youth with disabilities aged 14 to 23.…
Student-Level Analysis of Year 1 (2003-2004) Achievement Outcomes for Tennessee Charter Schools
ERIC Educational Resources Information Center
Ross, Steven M.; McDonald, Aaron J.; Gallagher, Brenda McSparrin
2005-01-01
This report presents student-level achievement results for the four charter schools that began operation in Tennessee during the 2003-04 academic year. To conduct a rigorous and valid analysis of student achievement outcomes at these schools, we employed a matched program-control design at the student level, whereby each charter school student was…
ERIC Educational Resources Information Center
Colbert, Peta; Wyatt-Smith, Claire; Klenowski, Val
2012-01-01
This article considers the conditions that are necessary at system and local levels for teacher assessment to be valid, reliable and rigorous. With sustainable assessment cultures as a goal, the article examines how education systems can support local-level efforts for quality learning and dependable teacher assessment. This is achieved through…
Evaluation of Delaware Stars for Early Success: Year 1 Report. Research Report
ERIC Educational Resources Information Center
Schwartz, Heather L.; Karoly, Lynn A.; Le, Vi-Nhuan; Tamargo, Jennifer; Setodji, Claude Messan
2014-01-01
Delaware was in the first group of states to receive a federal grant in 2012 to improve early care and education services and increase the number of infants, toddlers, and preschool-age children in high-quality programs. One component of the state's grant is a rigorous validation process for Delaware Stars for Early Success, a voluntary quality…
ERIC Educational Resources Information Center
Gibson, Jenny; Hussain, Jamilla; Holsgrove, Samina; Adams, Catherine; Green, Jonathan
2011-01-01
Direct observation of peer relating is potentially a sensitive and ecologically valid measure of child social functioning, but there has been a lack of standardised methods. The Manchester Inventory for Playground Observation (MIPO) was developed as a practical yet rigorous assessment of this kind for 5-11 year olds. We report on the initial…
ERIC Educational Resources Information Center
Barton, Erin E.; Pustejovsky, James E.; Maggin, Daniel M.; Reichow, Brian
2017-01-01
The adoption of methods and strategies validated through rigorous, experimentally oriented research is a core professional value of special education. We conducted a systematic review and meta-analysis examining the experimental literature on Technology-Aided Instruction and Intervention (TAII) using research identified as part of the National…
ERIC Educational Resources Information Center
Herlihy, Corinne; Karger, Ezra; Pollard, Cynthia; Hill, Heather C.; Kraft, Matthew A.; Williams, Megan; Howard, Sarah
2014-01-01
Context: In the past two years, states have implemented sweeping reforms to their teacher evaluation systems in response to Race to the Top legislation and, more recently, NCLB waivers. With these new systems, policymakers hope to make teacher evaluation both more rigorous and more grounded in specific job performance domains such as teaching…
Sun, Jiangming; Carlsson, Lars; Ahlberg, Ernst; Norinder, Ulf; Engkvist, Ola; Chen, Hongming
2017-07-24
Conformal prediction has been proposed as a more rigorous way to define prediction confidence compared to other application domain concepts that have earlier been used for QSAR modeling. One main advantage of such a method is that it provides a prediction region potentially with multiple predicted labels, which contrasts to the single valued (regression) or single label (classification) output predictions by standard QSAR modeling algorithms. Standard conformal prediction might not be suitable for imbalanced data sets. Therefore, Mondrian cross-conformal prediction (MCCP) which combines the Mondrian inductive conformal prediction with cross-fold calibration sets has been introduced. In this study, the MCCP method was applied to 18 publicly available data sets that have various imbalance levels varying from 1:10 to 1:1000 (ratio of active/inactive compounds). Our results show that MCCP in general performed well on bioactivity data sets with various imbalance levels. More importantly, the method not only provides confidence of prediction and prediction regions compared to standard machine learning methods but also produces valid predictions for the minority class. In addition, a compound similarity based nonconformity measure was investigated. Our results demonstrate that although it gives valid predictions, its efficiency is much worse than that of model dependent metrics.
Solomon, Howard M; Makris, Susan L; Alsaid, Hasan; Bermudez, Oscar; Beyer, Bruce K; Chen, Antong; Chen, Connie L; Chen, Zhou; Chmielewski, Gary; DeLise, Anthony M; de Schaepdrijver, Luc; Dogdas, Belma; French, Julian; Harrouk, Wafa; Helfgott, Jonathan; Henkelman, R Mark; Hesterman, Jacob; Hew, Kok-Wah; Hoberman, Alan; Lo, Cecilia W; McDougal, Andrew; Minck, Daniel R; Scott, Lelia; Stewart, Jane; Sutherland, Vicki; Tatiparthi, Arun K; Winkelmann, Christopher T; Wise, L David; Wood, Sandra L; Ying, Xiaoyou
2016-06-01
During the past two decades the use and refinements of imaging modalities have markedly increased making it possible to image embryos and fetuses used in pivotal nonclinical studies submitted to regulatory agencies. Implementing these technologies into the Good Laboratory Practice environment requires rigorous testing, validation, and documentation to ensure the reproducibility of data. A workshop on current practices and regulatory requirements was held with the goal of defining minimal criteria for the proper implementation of these technologies and subsequent submission to regulatory agencies. Micro-computed tomography (micro-CT) is especially well suited for high-throughput evaluations, and is gaining popularity to evaluate fetal skeletons to assess the potential developmental toxicity of test agents. This workshop was convened to help scientists in the developmental toxicology field understand and apply micro-CT technology to nonclinical toxicology studies and facilitate the regulatory acceptance of imaging data. Presentations and workshop discussions covered: (1) principles of micro-CT fetal imaging; (2) concordance of findings with conventional skeletal evaluations; and (3) regulatory requirements for validating the system. Establishing these requirements for micro-CT examination can provide a path forward for laboratories considering implementing this technology and provide regulatory agencies with a basis to consider the acceptability of data generated via this technology. Published by Elsevier Inc.
Closed loop statistical performance analysis of N-K knock controllers
NASA Astrophysics Data System (ADS)
Peyton Jones, James C.; Shayestehmanesh, Saeed; Frey, Jesse
2017-09-01
The closed loop performance of engine knock controllers cannot be rigorously assessed from single experiments or simulations because knock behaves as a random process and therefore the response belongs to a random distribution also. In this work a new method is proposed for computing the distributions and expected values of the closed loop response, both in steady state and in response to disturbances. The method takes as its input the control law, and the knock propensity characteristic of the engine which is mapped from open loop steady state tests. The method is applicable to the 'n-k' class of knock controllers in which the control action is a function only of the number of cycles n since the last control move, and the number k of knock events that have occurred in this time. A Cumulative Summation (CumSum) based controller falls within this category, and the method is used to investigate the performance of the controller in a deeper and more rigorous way than has previously been possible. The results are validated using onerous Monte Carlo simulations, which confirm both the validity of the method and its high computational efficiency.
Developing and validating a sham cupping device.
Lee, Myeong Soo; Kim, Jong-In; Kong, Jae Cheol; Lee, Dong-Hyo; Shin, Byung-Cheul
2010-12-01
The aims of this study were to develop a sham cupping device and to validate its use as a placebo control for healthy volunteers. A sham cupping device was developed by establishing a small hole to reduce the negative pressure after suction such that inner pressure could not be maintained in the cup. We enrolled 34 healthy participants to evaluate the validity of the sham cupping device as a placebo control. The participants were informed that they would receive either real or sham cupping and were asked which treatment they thought they had received. Other sensations and adverse events related to cupping therapy were investigated. 17 patients received real cupping therapy and 17 received sham cupping. The two groups felt similar sensations. There was a tendency for subjects to feel that real cupping created a stronger sensation than sham cupping (48.9±21.4 vs 33.3±20.3 on a 100mm visual analogue scale). There were only mild to moderate adverse events observed in both groups. We developed a new sham cupping device that seems to provide a credible control for real cupping therapy by producing little or no negative pressure. This conclusion was supported by a pilot study, but more rigorous research is warranted regarding the use of this device.
D’Costa, Jamie J.; Goldsmith, James C.; Wilson, Jayne S.; Bryan, Richard T.; Ward, Douglas G.
2016-01-01
For over 80 years, cystoscopy has remained the gold-standard for detecting tumours of the urinary bladder. Since bladder tumours have a tendency to recur and progress, many patients are subjected to repeated cystoscopies during long-term surveillance, with the procedure being both unpleasant for the patient and expensive for healthcare providers. The identification and validation of bladder tumour specific molecular markers in urine could enable tumour detection and reduce reliance on cystoscopy, and numerous classes of biomarkers have been studied. Proteins represent the most intensively studied class of biomolecule in this setting. As an aid to researchers searching for better urinary biomarkers, we report a comprehensive systematic review of the literature and a searchable database of proteins that have been investigated to date. Our objective was to classify these proteins as: 1) those with robustly characterised sensitivity and specificity for bladder cancer detection; 2) those that show potential but further investigation is required; 3) those unlikely to warrant further investigation; and 4) those investigated as prognostic markers. This work should help to prioritise certain biomarkers for rigorous validation, whilst preventing wasted effort on proteins that have shown no association whatsoever with the disease, or only modest biomarker performance despite large-scale efforts at validation. PMID:27500198
NASA Astrophysics Data System (ADS)
Israel, Maya; Wherfel, Quentin M.; Shehab, Saadeddine; Ramos, Evan A.; Metzger, Adam; Reese, George C.
2016-07-01
This paper describes the development, validation, and uses of the Collaborative Computing Observation Instrument (C-COI), a web-based analysis instrument that classifies individual and/or collaborative behaviors of students during computing problem-solving (e.g. coding, programming). The C-COI analyzes data gathered through video and audio screen recording software that captures students' computer screens as they program, and their conversations with their peers or adults. The instrument allows researchers to organize and quantify these data to track behavioral patterns that could be further analyzed for deeper understanding of persistence and/or collaborative interactions. The article provides a rationale for the C-COI including the development of a theoretical framework for measuring collaborative interactions in computer-mediated environments. This theoretical framework relied on the computer-supported collaborative learning literature related to adaptive help seeking, the joint problem-solving space in which collaborative computing occurs, and conversations related to outcomes and products of computational activities. Instrument development and validation also included ongoing advisory board feedback from experts in computer science, collaborative learning, and K-12 computing as well as classroom observations to test out the constructs in the C-COI. These processes resulted in an instrument with rigorous validation procedures and a high inter-rater reliability.
Validation of MHD Models using MST RFP Plasmas
NASA Astrophysics Data System (ADS)
Jacobson, C. M.; Chapman, B. E.; den Hartog, D. J.; McCollam, K. J.; Sarff, J. S.; Sovinec, C. R.
2017-10-01
Rigorous validation of computational models used in fusion energy sciences over a large parameter space and across multiple magnetic configurations can increase confidence in their ability to predict the performance of future devices. MST is a well diagnosed reversed-field pinch (RFP) capable of operation with plasma current ranging from 60 kA to 500 kA. The resulting Lundquist number S, a key parameter in resistive magnetohydrodynamics (MHD), ranges from 4 ×104 to 8 ×106 for standard RFP plasmas and provides substantial overlap with MHD RFP simulations. MST RFP plasmas are simulated using both DEBS, a nonlinear single-fluid visco-resistive MHD code, and NIMROD, a nonlinear extended MHD code, with S ranging from 104 to 105 for single-fluid runs, and the magnetic Prandtl number Pm = 1 . Validation metric comparisons are presented, focusing on how normalized magnetic fluctuations at the edge b scale with S. Preliminary results for the dominant n = 6 mode are b S - 0 . 20 +/- 0 . 02 for single-fluid NIMROD, b S - 0 . 25 +/- 0 . 05 for DEBS, and b S - 0 . 20 +/- 0 . 02 for experimental measurements, however there is a significant discrepancy in mode amplitudes. Preliminary two-fluid NIMROD results are also presented. Work supported by US DOE.
Identification and validation of nebulized aerosol devices for sputum induction
Davidson, Warren J; Dennis, John; The, Stephanie; Litoski, Belinda; Pieron, Cora; Leigh, Richard
2014-01-01
Induced sputum cell count measurement has proven reliability for evaluating airway inflammation in patients with asthma and other airway diseases. Although the use of nebulizer devices for sputum induction is commonplace, they are generally labelled as single-patient devices by the manufacturer and, therefore, cannot be used for multiple patients in large clinical sputum induction programs due to infect ion-control requirements. Accordingly, this study investigated the aerosol characteristics of alternative devices that could be used in such programs. BACKGROUND: Induced sputum cell counts are a noninvasive and reliable method for evaluating the presence, type and degree of airway inflammation in patients with asthma. Currently, standard nebulizer devices used for sputum induction in multiple patients are labelled as single-patient devices by the manufacturer, which conflicts with infection prevention and control requirements. As such, these devices cannot feasibly be used in a clinical sputum induction program. Therefore, there is a need to identify alternative nebulizer devices that are either disposable or labelled for multi-patient use. OBJECTIVE: To apply validated rigorous, scientific testing methods to identify and validate commercially available nebulizer devices appropriate for use in a clinical sputum induction program. METHODS: Measurement of nebulized aerosol output and size for the selected nebulizer designs followed robust International Organization for Standardization methods. Sputum induction using two of these nebulizers was successfully performed on 10 healthy adult subjects. The cytotechnologist performing sputum cell counts was blinded to the type of nebulizer used. RESULTS: The studied nebulizers had variable aerosol outputs. The AeroNeb Solo (Aerogen, Ireland), Omron NE-U17 (Omron, Japan) and EASYneb II (Flaem Nuova, Italy) systems were found to have similar measurements of aerosol size. There was no significant difference in induced sputum cell results between the AeroNeb Solo and EASYneb II devices. DISCUSSION: There is a need for rigorous, scientific evaluation of nebulizer devices for clinical applications, including sputum induction, for measurement of cell counts. CONCLUSION: The present study was the most comprehensive analysis of different nebulizer devices for sputum induction to measure cell counts, and provides a framework for appropriate evaluation of nebulizer devices for induced sputum testing. PMID:24288700
A Model-Based Approach to Support Validation of Medical Cyber-Physical Systems.
Silva, Lenardo C; Almeida, Hyggo O; Perkusich, Angelo; Perkusich, Mirko
2015-10-30
Medical Cyber-Physical Systems (MCPS) are context-aware, life-critical systems with patient safety as the main concern, demanding rigorous processes for validation to guarantee user requirement compliance and specification-oriented correctness. In this article, we propose a model-based approach for early validation of MCPS, focusing on promoting reusability and productivity. It enables system developers to build MCPS formal models based on a library of patient and medical device models, and simulate the MCPS to identify undesirable behaviors at design time. Our approach has been applied to three different clinical scenarios to evaluate its reusability potential for different contexts. We have also validated our approach through an empirical evaluation with developers to assess productivity and reusability. Finally, our models have been formally verified considering functional and safety requirements and model coverage.
A Model-Based Approach to Support Validation of Medical Cyber-Physical Systems
Silva, Lenardo C.; Almeida, Hyggo O.; Perkusich, Angelo; Perkusich, Mirko
2015-01-01
Medical Cyber-Physical Systems (MCPS) are context-aware, life-critical systems with patient safety as the main concern, demanding rigorous processes for validation to guarantee user requirement compliance and specification-oriented correctness. In this article, we propose a model-based approach for early validation of MCPS, focusing on promoting reusability and productivity. It enables system developers to build MCPS formal models based on a library of patient and medical device models, and simulate the MCPS to identify undesirable behaviors at design time. Our approach has been applied to three different clinical scenarios to evaluate its reusability potential for different contexts. We have also validated our approach through an empirical evaluation with developers to assess productivity and reusability. Finally, our models have been formally verified considering functional and safety requirements and model coverage. PMID:26528982
Treatment of charge singularities in implicit solvent models.
Geng, Weihua; Yu, Sining; Wei, Guowei
2007-09-21
This paper presents a novel method for solving the Poisson-Boltzmann (PB) equation based on a rigorous treatment of geometric singularities of the dielectric interface and a Green's function formulation of charge singularities. Geometric singularities, such as cusps and self-intersecting surfaces, in the dielectric interfaces are bottleneck in developing highly accurate PB solvers. Based on an advanced mathematical technique, the matched interface and boundary (MIB) method, we have recently developed a PB solver by rigorously enforcing the flux continuity conditions at the solvent-molecule interface where geometric singularities may occur. The resulting PB solver, denoted as MIBPB-II, is able to deliver second order accuracy for the molecular surfaces of proteins. However, when the mesh size approaches half of the van der Waals radius, the MIBPB-II cannot maintain its accuracy because the grid points that carry the interface information overlap with those that carry distributed singular charges. In the present Green's function formalism, the charge singularities are transformed into interface flux jump conditions, which are treated on an equal footing as the geometric singularities in our MIB framework. The resulting method, denoted as MIBPB-III, is able to provide highly accurate electrostatic potentials at a mesh as coarse as 1.2 A for proteins. Consequently, at a given level of accuracy, the MIBPB-III is about three times faster than the APBS, a recent multigrid PB solver. The MIBPB-III has been extensively validated by using analytically solvable problems, molecular surfaces of polyatomic systems, and 24 proteins. It provides reliable benchmark numerical solutions for the PB equation.
Treatment of charge singularities in implicit solvent models
NASA Astrophysics Data System (ADS)
Geng, Weihua; Yu, Sining; Wei, Guowei
2007-09-01
This paper presents a novel method for solving the Poisson-Boltzmann (PB) equation based on a rigorous treatment of geometric singularities of the dielectric interface and a Green's function formulation of charge singularities. Geometric singularities, such as cusps and self-intersecting surfaces, in the dielectric interfaces are bottleneck in developing highly accurate PB solvers. Based on an advanced mathematical technique, the matched interface and boundary (MIB) method, we have recently developed a PB solver by rigorously enforcing the flux continuity conditions at the solvent-molecule interface where geometric singularities may occur. The resulting PB solver, denoted as MIBPB-II, is able to deliver second order accuracy for the molecular surfaces of proteins. However, when the mesh size approaches half of the van der Waals radius, the MIBPB-II cannot maintain its accuracy because the grid points that carry the interface information overlap with those that carry distributed singular charges. In the present Green's function formalism, the charge singularities are transformed into interface flux jump conditions, which are treated on an equal footing as the geometric singularities in our MIB framework. The resulting method, denoted as MIBPB-III, is able to provide highly accurate electrostatic potentials at a mesh as coarse as 1.2Å for proteins. Consequently, at a given level of accuracy, the MIBPB-III is about three times faster than the APBS, a recent multigrid PB solver. The MIBPB-III has been extensively validated by using analytically solvable problems, molecular surfaces of polyatomic systems, and 24 proteins. It provides reliable benchmark numerical solutions for the PB equation.
Validation (not just verification) of Deep Space Missions
NASA Technical Reports Server (NTRS)
Duren, Riley M.
2006-01-01
ion & Validation (V&V) is a widely recognized and critical systems engineering function. However, the often used definition 'Verification proves the design is right; validation proves it is the right design' is rather vague. And while Verification is a reasonably well standardized systems engineering process, Validation is a far more abstract concept and the rigor and scope applied to it varies widely between organizations and individuals. This is reflected in the findings in recent Mishap Reports for several NASA missions, in which shortfalls in Validation (not just Verification) were cited as root- or contributing-factors in catastrophic mission loss. Furthermore, although there is strong agreement in the community that Test is the preferred method for V&V, many people equate 'V&V' with 'Test', such that Analysis and Modeling aren't given comparable attention. Another strong motivator is a realization that the rapid growth in complexity of deep-space missions (particularly Planetary Landers and Space Observatories given their inherent unknowns) is placing greater demands on systems engineers to 'get it right' with Validation.
Methodological challenges of validating a clinical decision-making tool in the practice environment.
Brennan, Caitlin W; Daly, Barbara J
2015-04-01
Validating a measurement tool intended for use in the practice environment poses challenges that may not be present when validating a tool intended solely for research purposes. The aim of this article is to describe the methodological challenges of validating a clinical decision-making tool, the Oncology Acuity Tool, which nurses use to make nurse assignment and staffing decisions prospectively each shift. Data were derived from a larger validation study, during which several methodological challenges arose. Revisions to the tool, including conducting iterative feedback cycles with end users, were necessary before the validation study was initiated. The "true" value of patient acuity is unknown, and thus, two approaches to inter-rater reliability assessment were used. Discordant perspectives existed between experts and end users. Balancing psychometric rigor with clinical relevance may be achieved through establishing research-practice partnerships, seeking active and continuous feedback with end users, and weighing traditional statistical rules of thumb with practical considerations. © The Author(s) 2014.
Climate Change: Providing Equitable Access to a Rigorous and Engaging Curriculum
ERIC Educational Resources Information Center
Cardichon, Jessica; Roc, Martens
2013-01-01
This report examines how implementing rigorous and engaging curriculum aligned with college- and career-ready standards fosters positive school climates in which students are motivated to succeed, achievement gaps narrow, and learning and outcomes improve. It includes federal, state, and local recommendations for increasing access to high-quality,…
Challenges in Building Usable Knowledge in Education
ERIC Educational Resources Information Center
Hedges, Larry V.
2018-01-01
The scientific rigor of education research has improved dramatically since the year 2000. Much of the credit for this improvement is deserved by Institute of Education Sciences (IES) policies that helped create a demand for rigorous research; increased human capital capacity to carry out such work; provided funding for the work itself; and…
Strategies Leaders Can Use to Improve Rigor in Their Schools
ERIC Educational Resources Information Center
Williamson, Ronald; Blackburn, Barbara R.
2009-01-01
Concern about rigor is not new. Since the release of "A Nation At Risk" (National Commission on Excellence in Education, 1983) the debate about the quality of America's schools has grown exponentially. This debate calls for dramatically different schools, schools that are much more responsive to student need, and provide a rigorous…
Provencher, Steeve; Archer, Stephen L; Ramirez, F Daniel; Hibbert, Benjamin; Paulin, Roxane; Boucherat, Olivier; Lacasse, Yves; Bonnet, Sébastien
2018-03-30
Despite advances in our understanding of the pathophysiology and the management of pulmonary arterial hypertension (PAH), significant therapeutic gaps remain for this devastating disease. Yet, few innovative therapies beyond the traditional pathways of endothelial dysfunction have reached clinical trial phases in PAH. Although there are inherent limitations of the currently available models of PAH, the leaky pipeline of innovative therapies relates, in part, to flawed preclinical research methodology, including lack of rigour in trial design, incomplete invasive hemodynamic assessment, and lack of careful translational studies that replicate randomized controlled trials in humans with attention to adverse effects and benefits. Rigorous methodology should include the use of prespecified eligibility criteria, sample sizes that permit valid statistical analysis, randomization, blinded assessment of standardized outcomes, and transparent reporting of results. Better design and implementation of preclinical studies can minimize inherent flaws in the models of PAH, reduce the risk of bias, and enhance external validity and our ability to distinguish truly promising therapies form many false-positive or overstated leads. Ideally, preclinical studies should use advanced imaging, study several preclinical pulmonary hypertension models, or correlate rodent and human findings and consider the fate of the right ventricle, which is the major determinant of prognosis in human PAH. Although these principles are widely endorsed, empirical evidence suggests that such rigor is often lacking in pulmonary hypertension preclinical research. The present article discusses the pitfalls in the design of preclinical pulmonary hypertension trials and discusses opportunities to create preclinical trials with improved predictive value in guiding early-phase drug development in patients with PAH, which will need support not only from researchers, peer reviewers, and editors but also from academic institutions, funding agencies, and animal ethics authorities. © 2018 American Heart Association, Inc.
Genetic and environmental effects on the muscle structure response post-mortem.
Thompson, J M; Perry, D; Daly, B; Gardner, G E; Johnston, D J; Pethick, D W
2006-09-01
This paper reviewed the mechanisms by which glycolytic rate and pre-rigor stretching of muscle impact on meat quality. If muscle is free to shorten during the rigor process extremes in glycolytic rate can impact negatively on meat quality by inducing either cold or rigor shortening. Factors that contribute to variation in glycolytic rate include the glycogen concentration at slaughter and fibre type of the muscle. Glycolysis is highly sensitive to temperature, which is an important factor in heavy grain fed carcasses. An alternative solution to controlling glycolysis is to stretch the muscle pre-rigor so that it cannot shorten, thus providing an insurance against extremes in processing conditions. Results are presented which show a large reduction in variance (both additive and phenotypic) in tenderness caused by pre-rigor stretching. Whilst this did not impact on the heritability of shear force, it did reduce genotype differences. The implications of these results on the magnitude of genotype effects on tenderness is discussed.
Developing nurse and physician questionnaires to assess primary work areas in intensive care units.
Rashid, Mahbub; Boyle, Diane K; Crosser, Michael
2014-01-01
The objective of the study was to develop instruments for describing and assessing some aspects of design of the primary work areas of nurses and physicians in intensive care units (ICUs). Separate questionnaires for ICU physicians and nurses were developed. Items related to individual- and unit-level design features of the primary work areas of nurses and physicians were organized using constructs found in the literature. Items related to staff satisfaction and staff use of time in relation to primary work area design were also included. All items and constructs were reviewed by experts for content validity and were modified as needed before use. The final questionnaires were administered to a convenience sample of 4 ICUs in 2 large urban hospitals. A total of 55 nurses and 29 physicians completed the survey. The Cronbach α was used to measure internal consistency, and factor analysis was used to provide construct-related validity. Convergent and discriminant validity were assessed through examining bivariate correlations between relevant scales/items. Analysis of variance was used to identify whether the between-group member responses were significant among the 4 units. The Cronbach α values for all except 3 preliminary scales indicated acceptable reliability. Factor analysis indicated that some preliminary scales could be partitioned into subscales for finer descriptions of the primary work areas. Correlational analysis provided strong evidence of convergent and discriminant validity of all the scales and subscales. The significance level of F-statistics showed that the units were significantly different from each other, providing evidence of more between-unit variance than within-unit variance. Therefore, the questionnaires developed in the study offer a promising departure point for rigorous description and evaluation of the primary work areas in relation to staff satisfaction and use of time in ICUs at a time when the importance of such studies is growing.
German translation, cultural adaptation, and validation of the Health Literacy Questionnaire (HLQ).
Nolte, Sandra; Osborne, Richard H; Dwinger, Sarah; Elsworth, Gerald R; Conrad, Melanie L; Rose, Matthias; Härter, Martin; Dirmaier, Jörg; Zill, Jördis M
2017-01-01
The Health Literacy Questionnaire (HLQ), developed in Australia in 2012 using a 'validity-driven' approach, has been rapidly adopted and is being applied in many countries and languages. It is a multidimensional measure comprising nine distinct domains that may be used for surveys, needs assessment, evaluation and outcomes assessment as well as for informing service improvement and the development of interventions. The aim of this paper is to describe the German translation of the HLQ and to present the results of the validation of the culturally adapted version. The HLQ comprises 44 items, which were translated and culturally adapted to the German context. This study uses data collected from a sample of 1,058 persons with chronic conditions. Statistical analyses include descriptive and confirmatory factor analyses. In one-factor congeneric models, all scales demonstrated good fit after few model adjustments. In a single, highly restrictive nine-factor model (no cross-loadings, no correlated errors) replication of the original English-language version was achieved with fit indices and psychometric properties similar to the original HLQ. Reliability for all scales was excellent, with a Cronbach's Alpha of at least 0.77. High to very high correlations between some HLQ factors were observed, suggesting that higher order factors may be present. Our rigorous development and validation protocol, as well as strict adaptation processes, have generated a remarkable reproduction of the HLQ in German. The results of this validation provide evidence that the HLQ is robust and can be recommended for use in German-speaking populations. German Clinical Trial Registration (DRKS): DRKS00000584. Registered 23 March 2011.
Vanquelef, Enguerran; Simon, Sabrina; Marquant, Gaelle; Garcia, Elodie; Klimerak, Geoffroy; Delepine, Jean Charles; Cieplak, Piotr; Dupradeau, François-Yves
2011-07-01
R.E.D. Server is a unique, open web service, designed to derive non-polarizable RESP and ESP charges and to build force field libraries for new molecules/molecular fragments. It provides to computational biologists the means to derive rigorously molecular electrostatic potential-based charges embedded in force field libraries that are ready to be used in force field development, charge validation and molecular dynamics simulations. R.E.D. Server interfaces quantum mechanics programs, the RESP program and the latest version of the R.E.D. tools. A two step approach has been developed. The first one consists of preparing P2N file(s) to rigorously define key elements such as atom names, topology and chemical equivalencing needed when building a force field library. Then, P2N files are used to derive RESP or ESP charges embedded in force field libraries in the Tripos mol2 format. In complex cases an entire set of force field libraries or force field topology database is generated. Other features developed in R.E.D. Server include help services, a demonstration, tutorials, frequently asked questions, Jmol-based tools useful to construct PDB input files and parse R.E.D. Server outputs as well as a graphical queuing system allowing any user to check the status of R.E.D. Server jobs.
Psychometric analysis of the Brisbane Practice Environment Measure (B-PEM).
Flint, Anndrea; Farrugia, Charles; Courtney, Mary; Webster, Joan
2010-03-01
To undertake rigorous psychometric testing of the newly developed contemporary work environment measure (the Brisbane Practice Environment Measure [B-PEM]) using exploratory factor analysis and confirmatory factor analysis. Content validity of the 33-item measure was established by a panel of experts. Initial testing involved 195 nursing staff using principal component factor analysis with varimax rotation (orthogonal) and Cronbach's alpha coefficients. Confirmatory factor analysis was conducted using data from a further 983 nursing staff. Principal component factor analysis yielded a four-factor solution with eigenvalues greater than 1 that explained 52.53% of the variance. These factors were then verified using confirmatory factor analysis. Goodness-of-fit indices showed an acceptable fit overall with the full model, explaining 21% to 73% of the variance. Deletion of items took place throughout the evolution of the instrument, resulting in a 26-item, four-factor measure called the Brisbane Practice Environment Measure-Tested. The B-PEM has undergone rigorous psychometric testing, providing evidence of internal consistency and goodness-of-fit indices within acceptable ranges. The measure can be utilised as a subscale or total score reflective of a contemporary nursing work environment. An up-to-date instrument to measure practice environment may be useful for nursing leaders to monitor the workplace and to assist in identifying areas for improvement, facilitating greater job satisfaction and retention.
Rigorous quantum limits on monitoring free masses and harmonic oscillators
NASA Astrophysics Data System (ADS)
Roy, S. M.
2018-03-01
There are heuristic arguments proposing that the accuracy of monitoring position of a free mass m is limited by the standard quantum limit (SQL): σ2( X (t ) ) ≥σ2( X (0 ) ) +(t2/m2) σ2( P (0 ) ) ≥ℏ t /m , where σ2( X (t ) ) and σ2( P (t ) ) denote variances of the Heisenberg representation position and momentum operators. Yuen [Phys. Rev. Lett. 51, 719 (1983), 10.1103/PhysRevLett.51.719] discovered that there are contractive states for which this result is incorrect. Here I prove universally valid rigorous quantum limits (RQL), viz. rigorous upper and lower bounds on σ2( X (t ) ) in terms of σ2( X (0 ) ) and σ2( P (0 ) ) , given by Eq. (12) for a free mass and by Eq. (36) for an oscillator. I also obtain the maximally contractive and maximally expanding states which saturate the RQL, and use the contractive states to set up an Ozawa-type measurement theory with accuracies respecting the RQL but beating the standard quantum limit. The contractive states for oscillators improve on the Schrödinger coherent states of constant variance and may be useful for gravitational wave detection and optical communication.
Rigorous Combination of GNSS and VLBI: How it Improves Earth Orientation and Reference Frames
NASA Astrophysics Data System (ADS)
Lambert, S. B.; Richard, J. Y.; Bizouard, C.; Becker, O.
2017-12-01
Current reference series (C04) of the International Earth Rotation and Reference Systems Service (IERS) are produced by a weighted combination of Earth orientation parameters (EOP) time series built up by combination centers of each technique (VLBI, GNSS, Laser ranging, DORIS). In the future, we plan to derive EOP from a rigorous combination of the normal equation systems of the four techniques.We present here the results of a rigorous combination of VLBI and GNSS pre-reduced, constraint-free, normal equations with the DYNAMO geodetic analysis software package developed and maintained by the French GRGS (Groupe de Recherche en GeÌodeÌsie Spatiale). The used normal equations are those produced separately by the IVS and IGS combination centers to which we apply our own minimal constraints.We address the usefulness of such a method with respect to the classical, a posteriori, combination method, and we show whether EOP determinations are improved.Especially, we implement external validations of the EOP series based on comparison with geophysical excitation and examination of the covariance matrices. Finally, we address the potential of the technique for the next generation celestial reference frames, which are currently determined by VLBI only.
Current clinical research in orthodontics: a perspective.
Baumrind, Sheldon
2006-10-01
This essay explores briefly the approach of the Craniofacial Research Instrumentation Laboratory to the systematic and rigorous investigation of the usual outcome of orthodontic treatment in the practices of experienced clinicians. CRIL's goal is to produce a shareable electronic database of reliable, valid, and representative data on clinical practice as an aid in the production of an improved environment for truly evidence-based orthodontic treatment.
Coordination and Data Management of the International Arctic Buoy Programme (IABP)
2002-09-30
for forcing, validation and assimilation into numerical climate models , and for forecasting weather and ice conditions. TRANSITIONS Using IABP ...Coordination and Data Management of the International Arctic Buoy Programme ( IABP ) Ignatius G. Rigor 1013 NE 40th Street Polar Science Center...analyzed geophysical fields. APPROACH The IABP is a collaboration between 25 different institutions from 8 different countries, which work together
ERIC Educational Resources Information Center
Bull, Rebecca; Yao, Shih-Ying; Ng, Ee Lynn
2017-01-01
The early childhood sector in Singapore has witnessed vast changes in the past two decades. One of the key policy aims is to improve classroom quality. To ensure a rigorous evaluation of the quality of early childhood environments in Singapore, it is important to determine whether commonly used assessments of quality are valid indicators across…
A. M. S. Smith; L. B. Lenilte; A. T. Hudak; P. Morgan
2007-01-01
The Differenced Normalized Burn Ratio (deltaNBR) is widely used to map post-fire effects in North America from multispectral satellite imagery, but has not been rigorously validated across the great diversity in vegetation types. The importance of these maps to fire rehabilitation crews highlights the need for continued assessment of alternative remote sensing...
Argument within a Scientific Debate: The Case of the DRD2 A1 Allele as a Gene for Alcoholism.
ERIC Educational Resources Information Center
Wastyn, Ronald O.; Wastyn, M. Linda
1997-01-01
Investigates how opposing parties advanced arguments to the scientific community about the validity of DRD2 A1 allele as a gene causing alcoholism. Demonstrates to what extent scientists debate each other in journals by advancing opposing viewpoints with rigor and insight. Reveals what it means when scientists label a discovery in terms of finding…
Trends in Study Methods Used in Undergraduate Medical Education Research, 1969–2007
Baernstein, Amy; Liss, Hillary K.; Carney, Patricia A.; Elmore, Joann G.
2011-01-01
Context Evidence-based medical education requires rigorous studies appraising educational efficacy. Objectives To assess trends over time in methods used to evaluate undergraduate medical education interventions and to identify whether participation of medical education departments or centers is associated with more rigorous methods. Data Sources The PubMed, Cochrane Controlled Trials Registry, Campbell Collaboration, and ERIC databases (January 1966–March 2007) were searched using terms equivalent to students, medical and education, medical crossed with all relevant study designs. Study Selection We selected publications in all languages from every fifth year, plus the most recent 12 months, that evaluated an educational intervention for undergraduate medical students. Four hundred seventy-two publications met criteria for review. Data Extraction Data were abstracted on number of participants; types of comparison groups; whether outcomes assessed were objective, subjective, and/or validated; timing of outcome assessments; funding; and participation of medical education departments and centers. Ten percent of publications were independently abstracted by 2 authors to assess validity of the data abstraction. Results The annual number of publications increased over time from 1 (1969–1970) to 147 (2006–2007). In the most recent year, there was a mean of 145 medical student participants; 9 (6%) recruited participants from multiple institutions; 80 (54%) used comparison groups; 37 (25%) used randomized control groups; 91 (62%) had objective outcomes; 23 (16%) had validated outcomes; 35 (24%) assessed an outcome more than 1 month later; 21 (14%) estimated statistical power; and 66 (45%) reported funding. In 2006–2007, medical education department or center participation, reported in 46 (31%) of the recent publications, was associated only with enrolling more medical student participants (P = .04); for all studies from 1969 to 2007, it was associated only with measuring an objective outcome (P = .048). Between 1969 and 2007, the percentage of publications reporting statistical power and funding increased; percentages did not change for other study features. Conclusions The annual number of published studies of undergraduate medical education interventions demonstrating methodological rigor has been increasing. However, considerable opportunities for improvement remain. PMID:17785648
Rivard, Justin D; Vergis, Ashley S; Unger, Bertram J; Hardy, Krista M; Andrew, Chris G; Gillman, Lawrence M; Park, Jason
2014-06-01
Computer-based surgical simulators capture a multitude of metrics based on different aspects of performance, such as speed, accuracy, and movement efficiency. However, without rigorous assessment, it may be unclear whether all, some, or none of these metrics actually reflect technical skill, which can compromise educational efforts on these simulators. We assessed the construct validity of individual performance metrics on the LapVR simulator (Immersion Medical, San Jose, CA, USA) and used these data to create task-specific summary metrics. Medical students with no prior laparoscopic experience (novices, N = 12), junior surgical residents with some laparoscopic experience (intermediates, N = 12), and experienced surgeons (experts, N = 11) all completed three repetitions of four LapVR simulator tasks. The tasks included three basic skills (peg transfer, cutting, clipping) and one procedural skill (adhesiolysis). We selected 36 individual metrics on the four tasks that assessed six different aspects of performance, including speed, motion path length, respect for tissue, accuracy, task-specific errors, and successful task completion. Four of seven individual metrics assessed for peg transfer, six of ten metrics for cutting, four of nine metrics for clipping, and three of ten metrics for adhesiolysis discriminated between experience levels. Time and motion path length were significant on all four tasks. We used the validated individual metrics to create summary equations for each task, which successfully distinguished between the different experience levels. Educators should maintain some skepticism when reviewing the plethora of metrics captured by computer-based simulators, as some but not all are valid. We showed the construct validity of a limited number of individual metrics and developed summary metrics for the LapVR. The summary metrics provide a succinct way of assessing skill with a single metric for each task, but require further validation.
Ustün, B; Compton, W; Mager, D; Babor, T; Baiyewu, O; Chatterji, S; Cottler, L; Göğüş, A; Mavreas, V; Peters, L; Pull, C; Saunders, J; Smeets, R; Stipec, M R; Vrasti, R; Hasin, D; Room, R; Van den Brink, W; Regier, D; Blaine, J; Grant, B F; Sartorius, N
1997-09-25
The WHO Study on the reliability and validity of the alcohol and drug use disorder instruments in an international study which has taken place in centres in ten countries, aiming to test the reliability and validity of three diagnostic instruments for alcohol and drug use disorders: the Composite International Diagnostic Interview (CIDI), the Schedules for Clinical Assessment in Neuropsychiatry (SCAN) and a special version of the Alcohol Use Disorder and Associated Disabilities Interview schedule-alcohol/drug-revised (AUDADIS-ADR). The purpose of the reliability and validity (R&V) study is to further develop the alcohol and drug sections of these instruments so that a range of substance-related diagnoses can be made in a systematic, consistent, and reliable way. The study focuses on new criteria proposed in the tenth revision of the International Classification of Diseases (ICD-10) and the fourth revision of the diagnostic and statistical manual of mental disorders (DSM-IV) for dependence, harmful use and abuse categories for alcohol and psychoactive substance use disorders. A systematic study including a scientifically rigorous measure of reliability (i.e. 1 week test-retest reliability) and validity (i.e. comparison between clinical and non-clinical measures) has been undertaken. Results have yielded useful information on reliability and validity of these instruments at diagnosis, criteria and question level. Overall the diagnostic concordance coefficients (kappa, kappa) were very good for dependence disorders (0.7-0.9), but were somewhat lower for the abuse and harmful use categories. The comparisons among instruments and independent clinical evaluations and debriefing interviews gave important information about possible sources of unreliability, and provided useful clues on the applicability and consistency of nosological concepts across cultures.
Jalil, Rozh; Soukup, Tayana; Akhter, Waseem; Sevdalis, Nick; Green, James S A
2018-03-03
High-quality leadership and chairing skills are vital for good performance in multidisciplinary tumor boards (MTBs), but no instruments currently exist for assessing and improving these skills. To construct and validate a robust instrument for assessment of MTB leading and chairing skills. We developed an observational MTB leadership assessment instrument (ATLAS). ATLAS includes 12 domains that assess the leadership and chairing skills of the MTB chairperson. ATLAS has gone through a rigorous process of refinement and content validation prior to use to assess the MTB lead by two urological surgeons (blinded to each other) in 7 real-live (n = 286 cases) and 10 video-recorded (n = 131 cases) MTBs. ATLAS domains were analyzed via descriptive statistics. Instrument content was evaluated for validity using the content validation index (CVI). Intraclass correlation coefficients (ICCs) were used to assess inter-observer reliability. Instrument refining resulted in ATLAS including the following 12 domains: time management, communication, encouraging contribution, ability to summarize, ensuring all patients have treatment plan, case prioritization, keeping meeting focused, facilitate discussion, conflict management, leadership, creating good working atmosphere, and recruitment for clinical trials. CVI was acceptable and inter-rater agreement adequate to high for all domains. Agreement was somewhat higher in real-time MTBs compared to video ratings. Concurrent validation evidence was derived via positive and significant correlations between ATLAS and an established validated brief MTB leadership assessment scale. ATLAS is an observational assessment instrument that can be reliably used for assessing leadership and chairing skills in cancer MTBs (both live and video-recorded). The ability to assess and feedback on team leader performance provides the ground for promotion of good practice and continuing professional development of tumor board leaders.
How to Help Students Conceptualize the Rigorous Definition of the Limit of a Sequence
ERIC Educational Resources Information Center
Roh, Kyeong Hah
2010-01-01
This article suggests an activity, called the epsilon-strip activity, as an instructional method for conceptualization of the rigorous definition of the limit of a sequence via visualization. The article also describes the learning objectives of each instructional step of the activity, and then provides detailed instructional methods to guide…
Finite machines, mental procedures, and modern physics.
Lupacchini, Rossella
2007-01-01
A Turing machine provides a mathematical definition of the natural process of calculating. It rests on trust that a procedure of reason can be reproduced mechanically. Turing's analysis of the concept of mechanical procedure in terms of a finite machine convinced Gödel of the validity of the Church thesis. And yet, Gödel's later concern was that, insofar as Turing's work shows that "mental procedure cannot go beyond mechanical procedures", it would imply the same kind of limitation on human mind. He therefore deems Turing's argument to be inconclusive. The question then arises as to which extent a computing machine operating by finite means could provide an adequate model of human intelligence. It is argued that a rigorous answer to this question can be given by developing Turing's considerations on the nature of mental processes. For Turing such processes are the consequence of physical processes and he seems to be led to the conclusion that quantum mechanics could help to find a more comprehensive explanation of them.
Experiment for validation of fluid-structure interaction models and algorithms.
Hessenthaler, A; Gaddum, N R; Holub, O; Sinkus, R; Röhrle, O; Nordsletten, D
2017-09-01
In this paper a fluid-structure interaction (FSI) experiment is presented. The aim of this experiment is to provide a challenging yet easy-to-setup FSI test case that addresses the need for rigorous testing of FSI algorithms and modeling frameworks. Steady-state and periodic steady-state test cases with constant and periodic inflow were established. Focus of the experiment is on biomedical engineering applications with flow being in the laminar regime with Reynolds numbers 1283 and 651. Flow and solid domains were defined using computer-aided design (CAD) tools. The experimental design aimed at providing a straightforward boundary condition definition. Material parameters and mechanical response of a moderately viscous Newtonian fluid and a nonlinear incompressible solid were experimentally determined. A comprehensive data set was acquired by using magnetic resonance imaging to record the interaction between the fluid and the solid, quantifying flow and solid motion. Copyright © 2016 The Authors. International Journal for Numerical Methods in Biomedical Engineering published by John Wiley & Sons Ltd.
Measuring teamwork in health care settings: a review of survey instruments.
Valentine, Melissa A; Nembhard, Ingrid M; Edmondson, Amy C
2015-04-01
Teamwork in health care settings is widely recognized as an important factor in providing high-quality patient care. However, the behaviors that comprise effective teamwork, the organizational factors that support teamwork, and the relationship between teamwork and patient outcomes remain empirical questions in need of rigorous study. To identify and review survey instruments used to assess dimensions of teamwork so as to facilitate high-quality research on this topic. We conducted a systematic review of articles published before September 2012 to identify survey instruments used to measure teamwork and to assess their conceptual content, psychometric validity, and relationships to outcomes of interest. We searched the ISI Web of Knowledge database, and identified relevant articles using the search terms team, teamwork, or collaboration in combination with survey, scale, measure, or questionnaire. We found 39 surveys that measured teamwork. Surveys assessed different dimensions of teamwork. The most commonly assessed dimensions were communication, coordination, and respect. Of the 39 surveys, 10 met all of the criteria for psychometric validity, and 14 showed significant relationships to nonself-report outcomes. Evidence of psychometric validity is lacking for many teamwork survey instruments. However, several psychometrically valid instruments are available. Researchers aiming to advance research on teamwork in health care should consider using or adapting one of these instruments before creating a new one. Because instruments vary considerably in the behavioral processes and emergent states of teamwork that they capture, researchers must carefully evaluate the conceptual consistency between instrument, research question, and context.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Alves, Vinicius M.; Laboratory for Molecular Modeling, Division of Chemical Biology and Medicinal Chemistry, Eshelman School of Pharmacy, University of North Carolina, Chapel Hill, NC 27599; Muratov, Eugene
Repetitive exposure to a chemical agent can induce an immune reaction in inherently susceptible individuals that leads to skin sensitization. Although many chemicals have been reported as skin sensitizers, there have been very few rigorously validated QSAR models with defined applicability domains (AD) that were developed using a large group of chemically diverse compounds. In this study, we have aimed to compile, curate, and integrate the largest publicly available dataset related to chemically-induced skin sensitization, use this data to generate rigorously validated and QSAR models for skin sensitization, and employ these models as a virtual screening tool for identifying putativemore » sensitizers among environmental chemicals. We followed best practices for model building and validation implemented with our predictive QSAR workflow using Random Forest modeling technique in combination with SiRMS and Dragon descriptors. The Correct Classification Rate (CCR) for QSAR models discriminating sensitizers from non-sensitizers was 71–88% when evaluated on several external validation sets, within a broad AD, with positive (for sensitizers) and negative (for non-sensitizers) predicted rates of 85% and 79% respectively. When compared to the skin sensitization module included in the OECD QSAR Toolbox as well as to the skin sensitization model in publicly available VEGA software, our models showed a significantly higher prediction accuracy for the same sets of external compounds as evaluated by Positive Predicted Rate, Negative Predicted Rate, and CCR. These models were applied to identify putative chemical hazards in the Scorecard database of possible skin or sense organ toxicants as primary candidates for experimental validation. - Highlights: • It was compiled the largest publicly-available skin sensitization dataset. • Predictive QSAR models were developed for skin sensitization. • Developed models have higher prediction accuracy than OECD QSAR Toolbox. • Putative chemical hazards in the Scorecard database were found using our models.« less
Toker, Lilah; Rocco, Brad; Sibille, Etienne
2017-01-01
Establishing the molecular diversity of cell types is crucial for the study of the nervous system. We compiled a cross-laboratory database of mouse brain cell type-specific transcriptomes from 36 major cell types from across the mammalian brain using rigorously curated published data from pooled cell type microarray and single-cell RNA-sequencing (RNA-seq) studies. We used these data to identify cell type-specific marker genes, discovering a substantial number of novel markers, many of which we validated using computational and experimental approaches. We further demonstrate that summarized expression of marker gene sets (MGSs) in bulk tissue data can be used to estimate the relative cell type abundance across samples. To facilitate use of this expanding resource, we provide a user-friendly web interface at www.neuroexpresso.org. PMID:29204516
Numerical Approach to Spatial Deterministic-Stochastic Models Arising in Cell Biology
Gao, Fei; Li, Ye; Novak, Igor L.; Slepchenko, Boris M.
2016-01-01
Hybrid deterministic-stochastic methods provide an efficient alternative to a fully stochastic treatment of models which include components with disparate levels of stochasticity. However, general-purpose hybrid solvers for spatially resolved simulations of reaction-diffusion systems are not widely available. Here we describe fundamentals of a general-purpose spatial hybrid method. The method generates realizations of a spatially inhomogeneous hybrid system by appropriately integrating capabilities of a deterministic partial differential equation solver with a popular particle-based stochastic simulator, Smoldyn. Rigorous validation of the algorithm is detailed, using a simple model of calcium ‘sparks’ as a testbed. The solver is then applied to a deterministic-stochastic model of spontaneous emergence of cell polarity. The approach is general enough to be implemented within biologist-friendly software frameworks such as Virtual Cell. PMID:27959915
Parallel labeling experiments for pathway elucidation and (13)C metabolic flux analysis.
Antoniewicz, Maciek R
2015-12-01
Metabolic pathway models provide the foundation for quantitative studies of cellular physiology through the measurement of intracellular metabolic fluxes. For model organisms metabolic models are well established, with many manually curated genome-scale model reconstructions, gene knockout studies and stable-isotope tracing studies. However, for non-model organisms a similar level of knowledge is often lacking. Compartmentation of cellular metabolism in eukaryotic systems also presents significant challenges for quantitative (13)C-metabolic flux analysis ((13)C-MFA). Recently, innovative (13)C-MFA approaches have been developed based on parallel labeling experiments, the use of multiple isotopic tracers and integrated data analysis, that allow more rigorous validation of pathway models and improved quantification of metabolic fluxes. Applications of these approaches open new research directions in metabolic engineering, biotechnology and medicine. Copyright © 2015 Elsevier Ltd. All rights reserved.
Development of a fourth generation predictive capability maturity model.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Hills, Richard Guy; Witkowski, Walter R.; Urbina, Angel
2013-09-01
The Predictive Capability Maturity Model (PCMM) is an expert elicitation tool designed to characterize and communicate completeness of the approaches used for computational model definition, verification, validation, and uncertainty quantification associated for an intended application. The primary application of this tool at Sandia National Laboratories (SNL) has been for physics-based computational simulations in support of nuclear weapons applications. The two main goals of a PCMM evaluation are 1) the communication of computational simulation capability, accurately and transparently, and 2) the development of input for effective planning. As a result of the increasing importance of computational simulation to SNLs mission, themore » PCMM has evolved through multiple generations with the goal to provide more clarity, rigor, and completeness in its application. This report describes the approach used to develop the fourth generation of the PCMM.« less
A Thermal Management Systems Model for the NASA GTX RBCC Concept
NASA Technical Reports Server (NTRS)
Traci, Richard M.; Farr, John L., Jr.; Laganelli, Tony; Walker, James (Technical Monitor)
2002-01-01
The Vehicle Integrated Thermal Management Analysis Code (VITMAC) was further developed to aid the analysis, design, and optimization of propellant and thermal management concepts for advanced propulsion systems. The computational tool is based on engineering level principles and models. A graphical user interface (GUI) provides a simple and straightforward method to assess and evaluate multiple concepts before undertaking more rigorous analysis of candidate systems. The tool incorporates the Chemical Equilibrium and Applications (CEA) program and the RJPA code to permit heat transfer analysis of both rocket and air breathing propulsion systems. Key parts of the code have been validated with experimental data. The tool was specifically tailored to analyze rocket-based combined-cycle (RBCC) propulsion systems being considered for space transportation applications. This report describes the computational tool and its development and verification for NASA GTX RBCC propulsion system applications.
Bayesian operational modal analysis with asynchronous data, part I: Most probable value
NASA Astrophysics Data System (ADS)
Zhu, Yi-Chen; Au, Siu-Kui
2018-01-01
In vibration tests, multiple sensors are used to obtain detailed mode shape information about the tested structure. Time synchronisation among data channels is required in conventional modal identification approaches. Modal identification can be more flexibly conducted if this is not required. Motivated by the potential gain in feasibility and economy, this work proposes a Bayesian frequency domain method for modal identification using asynchronous 'output-only' ambient data, i.e. 'operational modal analysis'. It provides a rigorous means for identifying the global mode shape taking into account the quality of the measured data and their asynchronous nature. This paper (Part I) proposes an efficient algorithm for determining the most probable values of modal properties. The method is validated using synthetic and laboratory data. The companion paper (Part II) investigates identification uncertainty and challenges in applications to field vibration data.
Chen, Zheng; Liu, Liu; Mu, Lin
2017-05-03
In this paper, we consider the linear transport equation under diffusive scaling and with random inputs. The method is based on the generalized polynomial chaos approach in the stochastic Galerkin framework. Several theoretical aspects will be addressed. Additionally, a uniform numerical stability with respect to the Knudsen number ϵ, and a uniform in ϵ error estimate is given. For temporal and spatial discretizations, we apply the implicit–explicit scheme under the micro–macro decomposition framework and the discontinuous Galerkin method, as proposed in Jang et al. (SIAM J Numer Anal 52:2048–2072, 2014) for deterministic problem. Lastly, we provide a rigorous proof ofmore » the stochastic asymptotic-preserving (sAP) property. Extensive numerical experiments that validate the accuracy and sAP of the method are conducted.« less
Premenstrual Dysphoric Disorder: Evidence for a New Category for DSM-5
Epperson, C. Neill; Steiner, Meir; Hartlage, S. Ann; Eriksson, Elias; Schmidt, Peter J.; Jones, Ian; Yonkers, Kimberly A.
2012-01-01
Premenstrual dysphoric disorder, which affects 2%–5% of premenopausal women, was included in Appendix B of DSM-IV, “Criterion Sets and Axes Provided for Further Study.” Since then, aided by the inclusion of specific and rigorous criteria in DSM-IV, there has been an explosion of research on the epidemiology, phenomenology, pathogenesis, and treatment of the disorder. In 2009, the Mood Disorders Work Group for DSM-5 convened a group of experts to examine the literature on premenstrual dysphoric disorder and provide recommendations regarding the appropriate criteria and placement for the disorder in DSM-5. Based on thorough review and lengthy discussion, the work group proposed that the information on the diagnosis, treatment, and validation of the disorder has matured sufficiently for it to qualify as a full category in DSM-5. A move to the position of category, rather than a criterion set in need of further study, will provide greater legitimacy for the disorder and encourage the growth of evidence-based research, ultimately leading to new treatments. PMID:22764360
Elder Abuse: Global Situation, Risk Factors, and Prevention Strategies.
Pillemer, Karl; Burnes, David; Riffin, Catherine; Lachs, Mark S
2016-04-01
Elder mistreatment is now recognized internationally as a pervasive and growing problem, urgently requiring the attention of health care systems, social welfare agencies, policymakers, and the general public. In this article, we provide an overview of global issues in the field of elder abuse, with a focus on prevention. This article provides a scoping review of key issues in the field from an international perspective. By drawing primarily on population-based studies, this scoping review provided a more valid and reliable synthesis of current knowledge about prevalence and risk factors than has been available. Despite the lack of scientifically rigorous intervention research on elder abuse, the review also identified 5 promising strategies for prevention. The findings highlight a growing consensus across studies regarding the extent and causes of elder mistreatment, as well as the urgent need for efforts to make elder mistreatment prevention programs more effective and evidence based. © The Author 2016. Published by Oxford University Press on behalf of The Gerontological Society of America. All rights reserved. For permissions, please e-mail: journals.permissions@oup.com.
Cerebrospinal Fluid Biomarkers for Huntington's Disease.
Byrne, Lauren M; Wild, Edward J
2016-01-01
Cerebrospinal fluid (CSF) is enriched in brain-derived components and represents an accessible and appealing means of interrogating the CNS milieu to study neurodegenerative diseases and identify biomarkers to facilitate the development of novel therapeutics. Many such CSF biomarkers have been proposed for Huntington's disease (HD) but none has been validated for clinical trial use. Across many studies proposing dozens of biomarker candidates, there is a notable lack of statistical power, consistency, rigor and validation. Here we review proposed CSF biomarkers including neurotransmitters, transglutaminase activity, kynurenine pathway metabolites, oxidative stress markers, inflammatory markers, neuroendocrine markers, protein markers of neuronal death, proteomic approaches and mutant huntingtin protein itself. We reflect on the need for large-scale, standardized CSF collections with detailed phenotypic data to validate and qualify much-needed CSF biomarkers for clinical trial use in HD.
Educational Testing and Validity of Conclusions in the Scholarship of Teaching and Learning
Beltyukova, Svetlana A.; Martin, Beth A.
2013-01-01
Validity and its integral evidence of reliability are fundamentals for educational and psychological measurement, and standards of educational testing. Herein, we describe these standards of educational testing, along with their subtypes including internal consistency, inter-rater reliability, and inter-rater agreement. Next, related issues of measurement error and effect size are discussed. This article concludes with a call for future authors to improve reporting of psychometrics and practical significance with educational testing in the pharmacy education literature. By increasing the scientific rigor of educational research and reporting, the overall quality and meaningfulness of SoTL will be improved. PMID:24249848
Haslam, Divna; Filus, Ania; Morawska, Alina; Sanders, Matthew R; Fletcher, Renee
2015-06-01
This paper outlines the development and validation of the Work-Family Conflict Scale (WAFCS) designed to measure work-to-family conflict (WFC) and family-to-work conflict (FWC) for use with parents of young children. An expert informant and consumer feedback approach was utilised to develop and refine 20 items, which were subjected to a rigorous validation process using two separate samples of parents of 2-12 year old children (n = 305 and n = 264). As a result of statistical analyses several items were dropped resulting in a brief 10-item scale comprising two subscales assessing theoretically distinct but related constructs: FWC (five items) and WFC (five items). Analyses revealed both subscales have good internal consistency, construct validity as well as concurrent and predictive validity. The results indicate the WAFCS is a promising brief measure for the assessment of work-family conflict in parents. Benefits of the measure as well as potential uses are discussed.
Perspectives on Validation of High-Throughput Assays Supporting 21st Century Toxicity Testing1
Judson, Richard; Kavlock, Robert; Martin, Matt; Reif, David; Houck, Keith; Knudsen, Thomas; Richard, Ann; Tice, Raymond R.; Whelan, Maurice; Xia, Menghang; Huang, Ruili; Austin, Christopher; Daston, George; Hartung, Thomas; Fowle, John R.; Wooge, William; Tong, Weida; Dix, David
2014-01-01
Summary In vitro, high-throughput screening (HTS) assays are seeing increasing use in toxicity testing. HTS assays can simultaneously test many chemicals, but have seen limited use in the regulatory arena, in part because of the need to undergo rigorous, time-consuming formal validation. Here we discuss streamlining the validation process, specifically for prioritization applications in which HTS assays are used to identify a high-concern subset of a collection of chemicals. The high-concern chemicals could then be tested sooner rather than later in standard guideline bioassays. The streamlined validation process would continue to ensure the reliability and relevance of assays for this application. We discuss the following practical guidelines: (1) follow current validation practice to the extent possible and practical; (2) make increased use of reference compounds to better demonstrate assay reliability and relevance; (3) deemphasize the need for cross-laboratory testing, and; (4) implement a web-based, transparent and expedited peer review process. PMID:23338806
Biomarkers of exposure to new and emerging tobacco delivery products.
Schick, Suzaynn F; Blount, Benjamin C; Jacob, Peyton; Saliba, Najat A; Bernert, John T; El Hellani, Ahmad; Jatlow, Peter; Pappas, R Steven; Wang, Lanqing; Foulds, Jonathan; Ghosh, Arunava; Hecht, Stephen S; Gomez, John C; Martin, Jessica R; Mesaros, Clementina; Srivastava, Sanjay; St Helen, Gideon; Tarran, Robert; Lorkiewicz, Pawel K; Blair, Ian A; Kimmel, Heather L; Doerschuk, Claire M; Benowitz, Neal L; Bhatnagar, Aruni
2017-09-01
Accurate and reliable measurements of exposure to tobacco products are essential for identifying and confirming patterns of tobacco product use and for assessing their potential biological effects in both human populations and experimental systems. Due to the introduction of new tobacco-derived products and the development of novel ways to modify and use conventional tobacco products, precise and specific assessments of exposure to tobacco are now more important than ever. Biomarkers that were developed and validated to measure exposure to cigarettes are being evaluated to assess their use for measuring exposure to these new products. Here, we review current methods for measuring exposure to new and emerging tobacco products, such as electronic cigarettes, little cigars, water pipes, and cigarillos. Rigorously validated biomarkers specific to these new products have not yet been identified. Here, we discuss the strengths and limitations of current approaches, including whether they provide reliable exposure estimates for new and emerging products. We provide specific guidance for choosing practical and economical biomarkers for different study designs and experimental conditions. Our goal is to help both new and experienced investigators measure exposure to tobacco products accurately and avoid common experimental errors. With the identification of the capacity gaps in biomarker research on new and emerging tobacco products, we hope to provide researchers, policymakers, and funding agencies with a clear action plan for conducting and promoting research on the patterns of use and health effects of these products.
Tissues from population-based cancer registries: a novel approach to increasing research potential.
Goodman, Marc T; Hernandez, Brenda Y; Hewitt, Stephen; Lynch, Charles F; Coté, Timothy R; Frierson, Henry F; Moskaluk, Christopher A; Killeen, Jeffrey L; Cozen, Wendy; Key, Charles R; Clegg, Limin; Reichman, Marsha; Hankey, Benjamin F; Edwards, Brenda
2005-07-01
Population-based cancer registries, such as those included in the Surveillance, Epidemiology, and End-Results (SEER) Program, offer tremendous research potential beyond traditional surveillance activities. We describe the expansion of SEER registries to gather formalin-fixed, paraffin-embedded tissue from cancer patients on a population basis. Population-based tissue banks have the advantage of providing an unbiased sampling frame for evaluating the public health impact of genes or protein targets that may be used for therapeutic or diagnostic purposes in defined communities. Such repositories provide a unique resource for testing new molecular classification schemes for cancer, validating new biologic markers of malignancy, prognosis and progression, assessing therapeutic targets, and measuring allele frequencies of cancer-associated genetic polymorphisms or germline mutations in representative samples. The assembly of tissue microarrays will allow for the use of rapid, large-scale protein-expression profiling of tumor samples while limiting depletion of this valuable resource. Access to biologic specimens through SEER registries will provide researchers with demographic, clinical, and risk factor information on cancer patients with assured data quality and completeness. Clinical outcome data, such as disease-free survival, can be correlated with previously validated prognostic markers. Furthermore, the anonymity of the study subject can be protected through rigorous standards of confidentiality. SEER-based tissue resources represent a step forward in true, population-based tissue repositories of tumors from US patients and may serve as a foundation for molecular epidemiology studies of cancer in this country.
Validity of Meta-analysis in Diabetes: We Need to Be Aware of Its Limitations
Home, Philip D.
2013-01-01
To deliver high-quality clinical care to patients with diabetes and other chronic conditions, clinicians must understand the evidence available from studies that have been performed to address important clinical management questions. In an evidence-based approach to clinical care, the evidence from clinical research should be integrated with clinical expertise, pathophysiological knowledge, and an understanding of patient values. As such, in an effort to provide information from many studies, the publication of diabetes meta-analyses has increased markedly in the recent past, using either observational or clinical trial data. In this regard, guidelines have been developed to direct the performance of meta-analysis to provide consistency among contributions. Thus, when done appropriately, meta-analysis can provide estimates from clinically and statistically homogeneous but underpowered studies and is useful in supporting clinical decisions, guidelines, and cost-effectiveness analysis. However, often these conditions are not met, the data considered are unreliable, and the results should not be assumed to be any more valid than the data underlying the included studies. To provide an understanding of both sides of the argument, we provide a discussion of this topic as part of this two-part point-counterpoint narrative. In the point narrative as presented below, Dr. Home provides his opinion and review of the data to date showing that we need to carefully evaluate meta-analyses and to learn what results are reliable. In the counterpoint narrative following Dr. Home’s contribution, Drs. Golden and Bass emphasize that an effective system exists to guide meta-analysis and that rigorously conducted, high-quality systematic reviews and meta-analyses are an indispensable tool in evidence synthesis despite their limitations. —William T. Cefalu, MD Editor in Chief, Diabetes Care PMID:24065844
Rigor and Relevance Redux: Director's Biennial Report to Congress. IES 2009-6010
ERIC Educational Resources Information Center
Whitehurst, Grover J.
2008-01-01
The mission of the Institute of Education Sciences (IES) is to provide rigorous evidence on which to ground education practice and policy and to encourage its use. The Education Sciences Reform Act of 2002 (ESRA) requires that the Director of IES, on a biennial basis, transmit to the President, the National Board for Education Sciences, and the…
Uncertainty information in climate data records from Earth observation
NASA Astrophysics Data System (ADS)
Merchant, Christopher J.; Paul, Frank; Popp, Thomas; Ablain, Michael; Bontemps, Sophie; Defourny, Pierre; Hollmann, Rainer; Lavergne, Thomas; Laeng, Alexandra; de Leeuw, Gerrit; Mittaz, Jonathan; Poulsen, Caroline; Povey, Adam C.; Reuter, Max; Sathyendranath, Shubha; Sandven, Stein; Sofieva, Viktoria F.; Wagner, Wolfgang
2017-07-01
The question of how to derive and present uncertainty information in climate data records (CDRs) has received sustained attention within the European Space Agency Climate Change Initiative (CCI), a programme to generate CDRs addressing a range of essential climate variables (ECVs) from satellite data. Here, we review the nature, mathematics, practicalities, and communication of uncertainty information in CDRs from Earth observations. This review paper argues that CDRs derived from satellite-based Earth observation (EO) should include rigorous uncertainty information to support the application of the data in contexts such as policy, climate modelling, and numerical weather prediction reanalysis. Uncertainty, error, and quality are distinct concepts, and the case is made that CDR products should follow international metrological norms for presenting quantified uncertainty. As a baseline for good practice, total standard uncertainty should be quantified per datum in a CDR, meaning that uncertainty estimates should clearly discriminate more and less certain data. In this case, flags for data quality should not duplicate uncertainty information, but instead describe complementary information (such as the confidence in the uncertainty estimate provided or indicators of conditions violating the retrieval assumptions). The paper discusses the many sources of error in CDRs, noting that different errors may be correlated across a wide range of timescales and space scales. Error effects that contribute negligibly to the total uncertainty in a single-satellite measurement can be the dominant sources of uncertainty in a CDR on the large space scales and long timescales that are highly relevant for some climate applications. For this reason, identifying and characterizing the relevant sources of uncertainty for CDRs is particularly challenging. The characterization of uncertainty caused by a given error effect involves assessing the magnitude of the effect, the shape of the error distribution, and the propagation of the uncertainty to the geophysical variable in the CDR accounting for its error correlation properties. Uncertainty estimates can and should be validated as part of CDR validation when possible. These principles are quite general, but the approach to providing uncertainty information appropriate to different ECVs is varied, as confirmed by a brief review across different ECVs in the CCI. User requirements for uncertainty information can conflict with each other, and a variety of solutions and compromises are possible. The concept of an ensemble CDR as a simple means of communicating rigorous uncertainty information to users is discussed. Our review concludes by providing eight concrete recommendations for good practice in providing and communicating uncertainty in EO-based climate data records.
Styles, Maggie; Cheyne, Helen; O'Carroll, Ronan; Greig, Fiona; Dagge-Bell, Fiona; Niven, Catherine
2011-10-01
to develop appropriate tools to assess midwives' attitudes and behaviour in relation to decision making involving risk. a questionnaire and series of vignettes were developed and testes to explore midwives' intrapartum decision making in relation to their attitudes towards risk. An innovative online computer package was developed specifically for use in the STORK Study which enabled the programme to be very tightly controlled with limited functions accessible to participants. a pilot study was conducted with over 50 midwives and nurses to ensure face and content validity of the vignettes and questionnaire. Initially designed to be a paper-based study, rigorous piloting highlighted the many difficulties in presenting it in that particular format. The solution to this problem was to develop the study as a secure online package. online data collection provided the researchers with a greater degree of control of the data collection process, not achievable using traditional paper survey methods. Another example of this control is the immediate entry of data from participants' responses to a background database which automatically stores and backs up data this means that no additional time is required for data entry. The cost of employing an information technology professional was easily offset by the financial savings made through the limited use of stationery and postage. although the development and testing of the research tools for the STORK Study was labour and time intensive, ultimately a questionnaire and vignette package was produced that had been rigorously tested by over 50 midwives and nurses. The researchers are confident in the reliability of the questionnaire and vignettes, as well as the validity of the data collected. The use of an online survey is clearly indicated when the population has readily available internet access, and where controlling the process of data collection is required, as such control cannot be achieved in traditional survey and questionnaire implementation. Copyright © 2010 Elsevier Ltd. All rights reserved.
Assessments of species' vulnerability to climate change: From pseudo to science
Wade, Alisa A.; Hand, Brian K.; Kovach, Ryan; Muhlfeld, Clint C.; Waples, Robin S.; Luikart, Gordon
2017-01-01
Climate change vulnerability assessments (CCVAs) are important tools to plan for and mitigate potential impacts of climate change. However, CCVAs often lack scientific rigor, which can ultimately lead to poor conservation prioritization and associated ecological and economic costs. We discuss the need to improve comparability and consistency of CCVAs and either validate their findings or improve assessment of CCVA uncertainty and sensitivity to methodological assumptions.
Equivalence of the Kelvin-Planck statement of the second law and the principle of entropy increase
NASA Astrophysics Data System (ADS)
Sarasua, L. G.; Abal, G.
2016-09-01
We present a demonstration of the equivalence between the Kelvin-Planck statement of the second law and the principle of entropy increase. Despite the fundamental importance of these two statements, a rigorous treatment to establish their equivalence is missing in standard physics textbooks. The argument is valid under very general conditions, but is simple and suited to an undergraduate course.
When is good, good enough? Methodological pragmatism for sustainable guideline development.
Browman, George P; Somerfield, Mark R; Lyman, Gary H; Brouwers, Melissa C
2015-03-06
Continuous escalation in methodological and procedural rigor for evidence-based processes in guideline development is associated with increasing costs and production delays that threaten sustainability. While health research methodologists are appropriately responsible for promoting increasing rigor in guideline development, guideline sponsors are responsible for funding such processes. This paper acknowledges that other stakeholders in addition to methodologists should be more involved in negotiating trade-offs between methodological procedures and efficiency in guideline production to produce guidelines that are 'good enough' to be trustworthy and affordable under specific circumstances. The argument for reasonable methodological compromise to meet practical circumstances is consistent with current implicit methodological practice. This paper proposes a conceptual tool as a framework to be used by different stakeholders in negotiating, and explicitly reporting, reasonable compromises for trustworthy as well as cost-worthy guidelines. The framework helps fill a transparency gap in how methodological choices in guideline development are made. The principle, 'when good is good enough' can serve as a basis for this approach. The conceptual tool 'Efficiency-Validity Methodological Continuum' acknowledges trade-offs between validity and efficiency in evidence-based guideline development and allows for negotiation, guided by methodologists, of reasonable methodological compromises among stakeholders. Collaboration among guideline stakeholders in the development process is necessary if evidence-based guideline development is to be sustainable.
A comment on the use of flushing time, residence time, and age as transport time scales
Monsen, N.E.; Cloern, J.E.; Lucas, L.V.; Monismith, Stephen G.
2002-01-01
Applications of transport time scales are pervasive in biological, hydrologic, and geochemical studies yet these times scales are not consistently defined and applied with rigor in the literature. We compare three transport time scales (flushing time, age, and residence time) commonly used to measure the retention of water or scalar quantities transported with water. We identify the underlying assumptions associated with each time scale, describe procedures for computing these time scales in idealized cases, and identify pitfalls when real-world systems deviate from these idealizations. We then apply the time scale definitions to a shallow 378 ha tidal lake to illustrate how deviations between real water bodies and the idealized examples can result from: (1) non-steady flow; (2) spatial variability in bathymetry, circulation, and transport time scales; and (3) tides that introduce complexities not accounted for in the idealized cases. These examples illustrate that no single transport time scale is valid for all time periods, locations, and constituents, and no one time scale describes all transport processes. We encourage aquatic scientists to rigorously define the transport time scale when it is applied, identify the underlying assumptions in the application of that concept, and ask if those assumptions are valid in the application of that approach for computing transport time scales in real systems.
Sutton, Patrice
2014-01-01
Background: Synthesizing what is known about the environmental drivers of health is instrumental to taking prevention-oriented action. Methods of research synthesis commonly used in environmental health lag behind systematic review methods developed in the clinical sciences over the past 20 years. Objectives: We sought to develop a proof of concept of the “Navigation Guide,” a systematic and transparent method of research synthesis in environmental health. Discussion: The Navigation Guide methodology builds on best practices in research synthesis in evidence-based medicine and environmental health. Key points of departure from current methods of expert-based narrative review prevalent in environmental health include a prespecified protocol, standardized and transparent documentation including expert judgment, a comprehensive search strategy, assessment of “risk of bias,” and separation of the science from values and preferences. Key points of departure from evidence-based medicine include assigning a “moderate” quality rating to human observational studies and combining diverse evidence streams. Conclusions: The Navigation Guide methodology is a systematic and rigorous approach to research synthesis that has been developed to reduce bias and maximize transparency in the evaluation of environmental health information. Although novel aspects of the method will require further development and validation, our findings demonstrated that improved methods of research synthesis under development at the National Toxicology Program and under consideration by the U.S. Environmental Protection Agency are fully achievable. The institutionalization of robust methods of systematic and transparent review would provide a concrete mechanism for linking science to timely action to prevent harm. Citation: Woodruff TJ, Sutton P. 2014. The Navigation Guide systematic review methodology: a rigorous and transparent method for translating environmental health science into better health outcomes. Environ Health Perspect 122:1007–1014; http://dx.doi.org/10.1289/ehp.1307175 PMID:24968373
NASA Astrophysics Data System (ADS)
Sarkar, Biplab; Adhikari, Satrajit
If a coupled three-state electronic manifold forms a sub-Hilbert space, it is possible to express the non-adiabatic coupling (NAC) elements in terms of adiabatic-diabatic transformation (ADT) angles. Consequently, we demonstrate: (a) Those explicit forms of the NAC terms satisfy the Curl conditions with non-zero Divergences; (b) The formulation of extended Born-Oppenheimer (EBO) equation for any three-state BO system is possible only when there exists coordinate independent ratio of the gradients for each pair of ADT angles leading to zero Curls at and around the conical intersection(s). With these analytic advancements, we formulate a rigorous EBO equation and explore its validity as well as necessity with respect to the approximate one (Sarkar and Adhikari, J Chem Phys 2006, 124, 074101) by performing numerical calculations on two different models constructed with different chosen forms of the NAC elements.
Impact of topographic mask models on scanner matching solutions
NASA Astrophysics Data System (ADS)
Tyminski, Jacek K.; Pomplun, Jan; Renwick, Stephen P.
2014-03-01
Of keen interest to the IC industry are advanced computational lithography applications such as Optical Proximity Correction of IC layouts (OPC), scanner matching by optical proximity effect matching (OPEM), and Source Optimization (SO) and Source-Mask Optimization (SMO) used as advanced reticle enhancement techniques. The success of these tasks is strongly dependent on the integrity of the lithographic simulators used in computational lithography (CL) optimizers. Lithographic mask models used by these simulators are key drivers impacting the accuracy of the image predications, and as a consequence, determine the validity of these CL solutions. Much of the CL work involves Kirchhoff mask models, a.k.a. thin masks approximation, simplifying the treatment of the mask near-field images. On the other hand, imaging models for hyper-NA scanner require that the interactions of the illumination fields with the mask topography be rigorously accounted for, by numerically solving Maxwell's Equations. The simulators used to predict the image formation in the hyper-NA scanners must rigorously treat the masks topography and its interaction with the scanner illuminators. Such imaging models come at a high computational cost and pose challenging accuracy vs. compute time tradeoffs. Additional complication comes from the fact that the performance metrics used in computational lithography tasks show highly non-linear response to the optimization parameters. Finally, the number of patterns used for tasks such as OPC, OPEM, SO, or SMO range from tens to hundreds. These requirements determine the complexity and the workload of the lithography optimization tasks. The tools to build rigorous imaging optimizers based on first-principles governing imaging in scanners are available, but the quantifiable benefits they might provide are not very well understood. To quantify the performance of OPE matching solutions, we have compared the results of various imaging optimization trials obtained with Kirchhoff mask models to those obtained with rigorous models involving solutions of Maxwell's Equations. In both sets of trials, we used sets of large numbers of patterns, with specifications representative of CL tasks commonly encountered in hyper-NA imaging. In this report we present OPEM solutions based on various mask models and discuss the models' impact on hyper- NA scanner matching accuracy. We draw conclusions on the accuracy of results obtained with thin mask models vs. the topographic OPEM solutions. We present various examples representative of the scanner image matching for patterns representative of the current generation of IC designs.
Surgical simulation: a urological perspective.
Wignall, Geoffrey R; Denstedt, John D; Preminger, Glenn M; Cadeddu, Jeffrey A; Pearle, Margaret S; Sweet, Robert M; McDougall, Elspeth M
2008-05-01
Surgical education is changing rapidly as several factors including budget constraints and medicolegal concerns limit opportunities for urological trainees. New methods of skills training such as low fidelity bench trainers and virtual reality simulators offer new avenues for surgical education. In addition, surgical simulation has the potential to allow practicing surgeons to develop new skills and maintain those they already possess. We provide a review of the background, current status and future directions of surgical simulators as they pertain to urology. We performed a literature review and an overview of surgical simulation in urology. Surgical simulators are in various stages of development and validation. Several simulators have undergone extensive validation studies and are in use in surgical curricula. While virtual reality simulators offer the potential to more closely mimic reality and present entire operations, low fidelity simulators remain useful in skills training, particularly for novices and junior trainees. Surgical simulation remains in its infancy. However, the potential to shorten learning curves for difficult techniques and practice surgery without risk to patients continues to drive the development of increasingly more advanced and realistic models. Surgical simulation is an exciting area of surgical education. The future is bright as advancements in computing and graphical capabilities offer new innovations in simulator technology. Simulators must continue to undergo rigorous validation studies to ensure that time spent by trainees on bench trainers and virtual reality simulators will translate into improved surgical skills in the operating room.
Wang, Shirley V; Schneeweiss, Sebastian; Berger, Marc L; Brown, Jeffrey; de Vries, Frank; Douglas, Ian; Gagne, Joshua J; Gini, Rosa; Klungel, Olaf; Mullins, C Daniel; Nguyen, Michael D; Rassen, Jeremy A; Smeeth, Liam; Sturkenboom, Miriam
2017-09-01
Defining a study population and creating an analytic dataset from longitudinal healthcare databases involves many decisions. Our objective was to catalogue scientific decisions underpinning study execution that should be reported to facilitate replication and enable assessment of validity of studies conducted in large healthcare databases. We reviewed key investigator decisions required to operate a sample of macros and software tools designed to create and analyze analytic cohorts from longitudinal streams of healthcare data. A panel of academic, regulatory, and industry experts in healthcare database analytics discussed and added to this list. Evidence generated from large healthcare encounter and reimbursement databases is increasingly being sought by decision-makers. Varied terminology is used around the world for the same concepts. Agreeing on terminology and which parameters from a large catalogue are the most essential to report for replicable research would improve transparency and facilitate assessment of validity. At a minimum, reporting for a database study should provide clarity regarding operational definitions for key temporal anchors and their relation to each other when creating the analytic dataset, accompanied by an attrition table and a design diagram. A substantial improvement in reproducibility, rigor and confidence in real world evidence generated from healthcare databases could be achieved with greater transparency about operational study parameters used to create analytic datasets from longitudinal healthcare databases. © 2017 The Authors. Pharmacoepidemiology & Drug Safety Published by John Wiley & Sons Ltd.
Cook, James L; Smith, Pat; Stannard, James P; Pfeiffer, Ferris; Kuroki, Keiichi; Bozynski, Chantelle C; Cook, Cristi
2017-09-01
Novel graft types, fixation methods, and means for augmenting anterior cruciate ligament (ACL) reconstructions require preclinical validation prior to safe and effective clinical application. The objective of this study was to describe and validate a translational canine model for all-inside arthroscopic complete ACL reconstruction using a quadriceps tendon allograft with internal brace (QTIB). With institutional approval, adult research hounds underwent complete transection of the native ACL followed by all-inside ACL reconstruction using the novel QTIB construct with suspensory fixation ( n = 10). Contralateral knees were used as nonoperated controls ( n = 10). Dogs were assessed over a 6-month period using functional, diagnostic imaging, gross, biomechanical, and histologic outcome measures required for preclinical animal models. Study results suggest that the novel QTIB construct used for complete ACL reconstruction can provide sustained knee stability and function without the development of premature osteoarthritis in a rigorous and valid preclinical model. The unique configuration of the QTIB construct-the combination of a tendon allograft with a synthetic suture tape internal brace-allowed for an effective biologic-synthetic load-sharing ACL construct. It prevented early failure, allowed for direct, four-zone graft-to-bone healing, and functional graft remodeling while avoiding problems noted with use of all-synthetic grafts. Thieme Medical Publishers 333 Seventh Avenue, New York, NY 10001, USA.
Yasuda, Kazuhiro; Muroi, Daisuke; Ohira, Masahiro; Iwata, Hiroyasu
2017-10-01
Unilateral spatial neglect (USN) is defined as impaired ability to attend and see on one side, and when present, it interferes seriously with daily life. These symptoms can exist for near and far spaces combined or independently, and it is important to provide effective intervention for near and far space neglect. The purpose of this pilot study was to propose an immersive virtual reality (VR) rehabilitation program using a head-mounted display that is able to train both near and far space neglect, and to validate the immediate effect of the VR program in both near and far space neglect. Ten USN patients underwent the VR program with a pre-post design and no control. In the virtual environment, we developed visual searching and reaching tasks using an immersive VR system. Behavioral inattention test (BIT) scores obtained pre- and immediate post-VR program were compared. BIT scores obtained pre- and post-VR program revealed that far space neglect but not near space neglect improved promptly after the VR program. This effect for far space neglect was observed in the cancelation task, but not in the line bisection task. Positive effects of the immersive VR program for far space neglect are suggested by the results of the present pilot study. However, further studies with rigorous designs are needed to validate its clinical effectiveness.
One-month validation of the Space Weather Modeling Framework geospace model
NASA Astrophysics Data System (ADS)
Haiducek, J. D.; Welling, D. T.; Ganushkina, N. Y.; Morley, S.; Ozturk, D. S.
2017-12-01
The Space Weather Modeling Framework (SWMF) geospace model consists of a magnetohydrodynamic (MHD) simulation coupled to an inner magnetosphere model and an ionosphere model. This provides a predictive capability for magnetopsheric dynamics, including ground-based and space-based magnetic fields, geomagnetic indices, currents and densities throughout the magnetosphere, cross-polar cap potential, and magnetopause and bow shock locations. The only inputs are solar wind parameters and F10.7 radio flux. We have conducted a rigorous validation effort consisting of a continuous simulation covering the month of January, 2005 using three different model configurations. This provides a relatively large dataset for assessment of the model's predictive capabilities. We find that the model does an excellent job of predicting the Sym-H index, and performs well at predicting Kp and CPCP during active times. Dayside magnetopause and bow shock positions are also well predicted. The model tends to over-predict Kp and CPCP during quiet times and under-predicts the magnitude of AL during disturbances. The model under-predicts the magnitude of night-side geosynchronous Bz, and over-predicts the radial distance to the flank magnetopause and bow shock. This suggests that the model over-predicts stretching of the magnetotail and the overall size of the magnetotail. With the exception of the AL index and the nightside geosynchronous magnetic field, we find the results to be insensitive to grid resolution.
Earth as an Extrasolar Planet: Earth Model Validation Using EPOXI Earth Observations
Meadows, Victoria S.; Crisp, David; Deming, Drake; A'Hearn, Michael F.; Charbonneau, David; Livengood, Timothy A.; Seager, Sara; Barry, Richard K.; Hearty, Thomas; Hewagama, Tilak; Lisse, Carey M.; McFadden, Lucy A.; Wellnitz, Dennis D.
2011-01-01
Abstract The EPOXI Discovery Mission of Opportunity reused the Deep Impact flyby spacecraft to obtain spatially and temporally resolved visible photometric and moderate resolution near-infrared (NIR) spectroscopic observations of Earth. These remote observations provide a rigorous validation of whole-disk Earth model simulations used to better understand remotely detectable extrasolar planet characteristics. We have used these data to upgrade, correct, and validate the NASA Astrobiology Institute's Virtual Planetary Laboratory three-dimensional line-by-line, multiple-scattering spectral Earth model. This comprehensive model now includes specular reflectance from the ocean and explicitly includes atmospheric effects such as Rayleigh scattering, gas absorption, and temperature structure. We have used this model to generate spatially and temporally resolved synthetic spectra and images of Earth for the dates of EPOXI observation. Model parameters were varied to yield an optimum fit to the data. We found that a minimum spatial resolution of ∼100 pixels on the visible disk, and four categories of water clouds, which were defined by using observed cloud positions and optical thicknesses, were needed to yield acceptable fits. The validated model provides a simultaneous fit to Earth's lightcurve, absolute brightness, and spectral data, with a root-mean-square (RMS) error of typically less than 3% for the multiwavelength lightcurves and residuals of ∼10% for the absolute brightness throughout the visible and NIR spectral range. We have extended our validation into the mid-infrared by comparing the model to high spectral resolution observations of Earth from the Atmospheric Infrared Sounder, obtaining a fit with residuals of ∼7% and brightness temperature errors of less than 1 K in the atmospheric window. For the purpose of understanding the observable characteristics of the distant Earth at arbitrary viewing geometry and observing cadence, our validated forward model can be used to simulate Earth's time-dependent brightness and spectral properties for wavelengths from the far ultraviolet to the far infrared. Key Words: Astrobiology—Extrasolar terrestrial planets—Habitability—Planetary science—Radiative transfer. Astrobiology 11, 393–408. PMID:21631250
Gong, Ting; Szustakowski, Joseph D
2013-04-15
For heterogeneous tissues, measurements of gene expression through mRNA-Seq data are confounded by relative proportions of cell types involved. In this note, we introduce an efficient pipeline: DeconRNASeq, an R package for deconvolution of heterogeneous tissues based on mRNA-Seq data. It adopts a globally optimized non-negative decomposition algorithm through quadratic programming for estimating the mixing proportions of distinctive tissue types in next-generation sequencing data. We demonstrated the feasibility and validity of DeconRNASeq across a range of mixing levels and sources using mRNA-Seq data mixed in silico at known concentrations. We validated our computational approach for various benchmark data, with high correlation between our predicted cell proportions and the real fractions of tissues. Our study provides a rigorous, quantitative and high-resolution tool as a prerequisite to use mRNA-Seq data. The modularity of package design allows an easy deployment of custom analytical pipelines for data from other high-throughput platforms. DeconRNASeq is written in R, and is freely available at http://bioconductor.org/packages. Supplementary data are available at Bioinformatics online.
Stewart, Regan W; Tuerk, Peter W; Metzger, Isha W; Davidson, Tatiana M; Young, John
2016-02-01
Structured diagnostic interviews are widely considered to be the optimal method of assessing symptoms of posttraumatic stress; however, few clinicians report using structured assessments to guide clinical practice. One commonly cited impediment to these assessment approaches is the amount of time required for test administration and interpretation. Empirically keyed methods to reduce the administration time of structured assessments may be a viable solution to increase the use of standardized and reliable diagnostic tools. Thus, the present research conducted an initial feasibility study using a sample of treatment-seeking military veterans (N = 1,517) to develop a truncated assessment protocol based on the Clinician-Administered Posttraumatic Stress Disorder (PTSD) Scale (CAPS). Decision-tree analysis was utilized to identify a subset of predictor variables among the CAPS items that were most predictive of a diagnosis of PTSD. The algorithm-driven, atheoretical sequence of questions reduced the number of items administered by more than 75% and classified the validation sample at 92% accuracy. These results demonstrated the feasibility of developing a protocol to assess PTSD in a way that imposes little assessment burden while still providing a reliable categorization. (c) 2016 APA, all rights reserved).
Confirmatory factor analysis for two questionnaires of caregiving in eating disorders
Hibbs, Rebecca; Rhind, Charlotte; Sallis, Hannah; Goddard, Elizabeth; Raenker, Simone; Ayton, Agnes; Bamford, Bryony; Arcelus, Jon; Boughton, Nicky; Connan, Frances; Goss, Ken; Lazlo, Bert; Morgan, John; Moore, Kim; Robertson, David; Schreiber-Kounine, Christa; Sharma, Sonu; Whitehead, Linette; Lacey, Hubert; Schmidt, Ulrike; Treasure, Janet
2014-01-01
Objective: Caring for someone diagnosed with an eating disorder (ED) is associated with a high level of burden and psychological distress which can inadvertently contribute to the maintenance of the illness. The Eating Disorders Symptom Impact Scale (EDSIS) and Accommodation and Enabling Scale for Eating Disorders (AESED) are self-report scales to assess elements of caregiving theorised to contribute to the maintenance of an ED. Further validation and confirmation of the factor structures for these scales are necessary for rigorous evaluation of complex interventions which target these modifiable elements of caregiving. Method: EDSIS and AESED data from 268 carers of people with anorexia nervosa (AN), recruited from consecutive admissions to 15 UK inpatient or day patient hospital units, were subjected to confirmatory factor analysis to test model fit by applying the existing factor structures: (a) four-factor structure for the EDSIS and (b) five-factor structure for the AESED. Results: Confirmatory factor analytic results support the existing four-factor and five-factor structures for the EDSIS and the AESED, respectively. Discussion: The present findings provide further validation of the EDSIS and the AESED as tools to assess modifiable elements of caregiving for someone with an ED. PMID:25750785
NASA Technical Reports Server (NTRS)
Hale, Joseph P.
2006-01-01
Models and simulations (M&S) are critical resources in the exploration of space. They support program management, systems engineering, integration, analysis, test, and operations and provide critical information and data supporting key analyses and decisions (technical, cost and schedule). Consequently, there is a clear need to establish a solid understanding of M&S strengths and weaknesses, and the bounds within which they can credibly support decision-making. Their usage requires the implementation of a rigorous approach to verification, validation and accreditation (W&A) and establishment of formal process and practices associated with their application. To ensure decision-making is suitably supported by information (data, models, test beds) from activities (studies, exercises) from M&S applications that are understood and characterized, ESMD is establishing formal, tailored W&A processes and practices. In addition, to ensure the successful application of M&S within ESMD, a formal process for the certification of analysts that use M&S is being implemented. This presentation will highlight NASA's Exploration Systems Mission Directorate (ESMD) management approach for M&S W&A to ensure decision-makers receive timely information on the model's fidelity, credibility, and quality.
Using eye tracking to identify faking attempts during penile plethysmography assessment.
Trottier, Dominique; Rouleau, Joanne-Lucine; Renaud, Patrice; Goyette, Mathieu
2014-01-01
Penile plethysmography (PPG) is considered the most rigorous method for sexual interest assessment. Nevertheless, it is subject to faking attempts by participants, which compromises the internal validity of the instrument. To date, various attempts have been made to limit voluntary control of sexual response during PPG assessments, without satisfactory results. This exploratory research examined eye-tracking technologies' ability to identify the presence of cognitive strategies responsible for erectile inhibition during PPG assessment. Eye movements and penile responses for 20 subjects were recorded while exploring animated human-like computer-generated stimuli in a virtual environment under three distinct viewing conditions: (a) the free visual exploration of a preferred sexual stimulus without erectile inhibition; (b) the viewing of a preferred sexual stimulus with erectile inhibition; and (c) the free visual exploration of a non-preferred sexual stimulus. Results suggest that attempts to control erectile responses generate specific eye-movement variations, characterized by a general deceleration of the exploration process and limited exploration of the erogenous zone. Findings indicate that recording eye movements can provide significant information on the presence of competing covert processes responsible for erectile inhibition. The use of eye-tracking technologies during PPG could therefore lead to improved internal validity of the plethysmographic procedure.
Conceptualizing Rigor and Its Implications for Education in the Era of the Common Core
ERIC Educational Resources Information Center
Paige, David D.; Smith, Grant S.; Sizemore, John M.
2015-01-01
The adoption of Common Core State Standards in the USA by 46 states and the District of Columbia has provided several new foci for K-12 instruction, not the least of which is the reading and understanding of complex text, a higher order thinking process. Closely associated with this is the notion of rigor, the focus of the present study. As…
Wallach, Joshua D; Ross, Joseph S; Naci, Huseyin
2018-06-01
The US Food and Drug Administration has several regulatory programs and pathways to expedite the development and approval of therapeutic agents aimed at treating serious or life-debilitating conditions. A common feature of these programs is the regulatory flexibility, which allows for a customized approval approach that enables market authorization on the basis of less rigorous evidence, in exchange for requiring postmarket evidence generation. An increasing share of therapeutic agents approved by the Food and Drug Administration in recent years are associated with expedited programs. In this article, we provide an overview of the evidentiary standards required by the Food and Drug Administration's expedited development and review programs, summarize the findings of the recent academic literature demonstrating some of the limitations of these programs, and outline potential opportunities to address these limitations. Recent evidence suggests that therapeutic agents in the Food and Drug Administration's expedited programs are approved on the basis of fewer and smaller studies that may lack comparator groups and random allocation, and rather than focusing on clinical outcomes for study endpoints, rely instead on surrogate markers of disease. Once on the market, agents receiving expedited approvals are often quickly incorporated into clinical practice, and evidence generated in the postmarket period may not necessarily address the evidentiary limitations at the time of market entry. Furthermore, not all pathways require additional postmarket studies. Evidence suggests that drugs in expedited approval programs are associated with a greater likelihood that the Food and Drug Administration will take a safety action following market entry. There are several opportunities to improve the timeliness, information value, and validity of the pre- and postmarket studies of therapeutic agents receiving expedited approvals. When use of nonrandomized and uncontrolled studies cannot be avoided prior to market entry, randomized trials should be mandatory in the postmarket period, unless there are strong justifications for not carrying out such studies. In the premarket period, validity of the surrogate markers can be improved by more rigorously evaluating their correlation with patient-relevant clinical outcomes. Opportunities to reduce the duration, complexity, and cost of postmarket randomized trials should not compromise their validity and instead incorporate pragmatic "real-world" design elements. Despite recent enthusiasm for widely using real-world evidence, adaptive designs, and pragmatic trials in the regulatory setting, caution is warranted until large-scale empirical evaluations demonstrate their validity compared to more traditional trial designs.
Chaabene, Helmi; Negra, Yassine; Bouguezzi, Raja; Capranica, Laura; Franchini, Emerson; Prieske, Olaf; Hbacha, Hamdi; Granacher, Urs
2018-01-01
The regular monitoring of physical fitness and sport-specific performance is important in elite sports to increase the likelihood of success in competition. This study aimed to systematically review and to critically appraise the methodological quality, validation data, and feasibility of the sport-specific performance assessment in Olympic combat sports like amateur boxing, fencing, judo, karate, taekwondo, and wrestling. A systematic search was conducted in the electronic databases PubMed, Google-Scholar, and Science-Direct up to October 2017. Studies in combat sports were included that reported validation data (e.g., reliability, validity, sensitivity) of sport-specific tests. Overall, 39 studies were eligible for inclusion in this review. The majority of studies (74%) contained sample sizes <30 subjects. Nearly, 1/3 of the reviewed studies lacked a sufficient description (e.g., anthropometrics, age, expertise level) of the included participants. Seventy-two percent of studies did not sufficiently report inclusion/exclusion criteria of their participants. In 62% of the included studies, the description and/or inclusion of a familiarization session (s) was either incomplete or not existent. Sixty-percent of studies did not report any details about the stability of testing conditions. Approximately half of the studies examined reliability measures of the included sport-specific tests (intraclass correlation coefficient [ICC] = 0.43–1.00). Content validity was addressed in all included studies, criterion validity (only the concurrent aspect of it) in approximately half of the studies with correlation coefficients ranging from r = −0.41 to 0.90. Construct validity was reported in 31% of the included studies and predictive validity in only one. Test sensitivity was addressed in 13% of the included studies. The majority of studies (64%) ignored and/or provided incomplete information on test feasibility and methodological limitations of the sport-specific test. In 28% of the included studies, insufficient information or a complete lack of information was provided in the respective field of the test application. Several methodological gaps exist in studies that used sport-specific performance tests in Olympic combat sports. Additional research should adopt more rigorous validation procedures in the application and description of sport-specific performance tests in Olympic combat sports. PMID:29692739
Volume Holograms in Photopolymers: Comparison between Analytical and Rigorous Theories
Gallego, Sergi; Neipp, Cristian; Estepa, Luis A.; Ortuño, Manuel; Márquez, Andrés; Francés, Jorge; Pascual, Inmaculada; Beléndez, Augusto
2012-01-01
There is no doubt that the concept of volume holography has led to an incredibly great amount of scientific research and technological applications. One of these applications is the use of volume holograms as optical memories, and in particular, the use of a photosensitive medium like a photopolymeric material to record information in all its volume. In this work we analyze the applicability of Kogelnik’s Coupled Wave theory to the study of volume holograms recorded in photopolymers. Some of the theoretical models in the literature describing the mechanism of hologram formation in photopolymer materials use Kogelnik’s theory to analyze the gratings recorded in photopolymeric materials. If Kogelnik’s theory cannot be applied is necessary to use a more general Coupled Wave theory (CW) or the Rigorous Coupled Wave theory (RCW). The RCW does not incorporate any approximation and thus, since it is rigorous, permits judging the accurateness of the approximations included in Kogelnik’s and CW theories. In this article, a comparison between the predictions of the three theories for phase transmission diffraction gratings is carried out. We have demonstrated the agreement in the prediction of CW and RCW and the validity of Kogelnik’s theory only for gratings with spatial frequencies higher than 500 lines/mm for the usual values of the refractive index modulations obtained in photopolymers.
Rosella, L; Bowman, C; Pach, B; Morgan, S; Fitzpatrick, T; Goel, V
2016-07-01
Most quality appraisal tools were developed for clinical medicine and tend to be study-specific with a strong emphasis on risk of bias. In order to be more relevant to public health, an appropriate quality appraisal tool needs to be less reliant on the evidence hierarchy and consider practice applicability. Given the broad range of study designs used in public health, the objective of this study was to develop and validate a meta-tool that combines public health-focused principles of appraisal coupled with a set of design-specific companion tools. Several design methods were used to develop and validate the tool including literature review, synthesis, and validation with a reference standard. A search of critical appraisal tools relevant to public health was conducted; core concepts were collated. The resulting framework was piloted during three feedback sessions with public health practitioners. Following subsequent revisions, the final meta-tool, the Meta Quality Appraisal Tool (MetaQAT), was then validated through a content analysis of appraisals conducted by two groups of experienced public health researchers (MetaQAT vs generic appraisal form). The MetaQAT framework consists of four domains: relevancy, reliability, validity, and applicability. In addition, a companion tool was assembled from existing critical appraisal tools to provide study design-specific guidance on validity appraisal. Content analysis showed similar methodological and generalizability concerns were raised by both groups; however, the MetaQAT appraisers commented more extensively on applicability to public health practice. Critical appraisal tools designed for clinical medicine have limitations for use in the context of public health. The meta-tool structure of the MetaQAT allows for rigorous appraisal, while allowing users to simultaneously appraise the multitude of study designs relevant to public health research and assess non-standard domains, such as applicability. Copyright © 2015 The Authors. Published by Elsevier Ltd.. All rights reserved.
Buckling of a stiff thin film on an elastic graded compliant substrate.
Chen, Zhou; Chen, Weiqiu; Song, Jizhou
2017-12-01
The buckling of a stiff film on a compliant substrate has attracted much attention due to its wide applications such as thin-film metrology, surface patterning and stretchable electronics. An analytical model is established for the buckling of a stiff thin film on a semi-infinite elastic graded compliant substrate subjected to in-plane compression. The critical compressive strain and buckling wavelength for the sinusoidal mode are obtained analytically for the case with the substrate modulus decaying exponentially. The rigorous finite element analysis (FEA) is performed to validate the analytical model and investigate the postbuckling behaviour of the system. The critical buckling strain for the period-doubling mode is obtained numerically. The influences of various material parameters on the results are investigated. These results are helpful to provide physical insights on the buckling of elastic graded substrate-supported thin film.
Reflections on experimental research in medical education.
Cook, David A; Beckman, Thomas J
2010-08-01
As medical education research advances, it is important that education researchers employ rigorous methods for conducting and reporting their investigations. In this article we discuss several important yet oft neglected issues in designing experimental research in education. First, randomization controls for only a subset of possible confounders. Second, the posttest-only design is inherently stronger than the pretest-posttest design, provided the study is randomized and the sample is sufficiently large. Third, demonstrating the superiority of an educational intervention in comparison to no intervention does little to advance the art and science of education. Fourth, comparisons involving multifactorial interventions are hopelessly confounded, have limited application to new settings, and do little to advance our understanding of education. Fifth, single-group pretest-posttest studies are susceptible to numerous validity threats. Finally, educational interventions (including the comparison group) must be described in detail sufficient to allow replication.
Jeffries, D J; Donkor, S; Brookes, R H; Fox, A; Hill, P C
2004-09-01
The data requirements of a large multidisciplinary tuberculosis case contact study are complex. We describe an ACCESS-based relational database system that meets our rigorous requirements for data entry and validation, while being user-friendly, flexible, exportable, and easy to install on a network or stand alone system. This includes the development of a double data entry package for epidemiology and laboratory data, semi-automated entry of ELISPOT data directly from the plate reader, and a suite of new programmes for the manipulation and integration of flow cytometry data. The double entered epidemiology and immunology databases are combined into a separate database, providing a near-real-time analysis of immuno-epidemiological data, allowing important trends to be identified early and major decisions about the study to be made and acted on. This dynamic data management model is portable and can easily be applied to other studies.
From Newton's Law to the Linear Boltzmann Equation Without Cut-Off
NASA Astrophysics Data System (ADS)
Ayi, Nathalie
2017-03-01
We provide a rigorous derivation of the linear Boltzmann equation without cut-off starting from a system of particles interacting via a potential with infinite range as the number of particles N goes to infinity under the Boltzmann-Grad scaling. More particularly, we will describe the motion of a tagged particle in a gas close to global equilibrium. The main difficulty in our context is that, due to the infinite range of the potential, a non-integrable singularity appears in the angular collision kernel, making no longer valid the single-use of Lanford's strategy. Our proof relies then on a combination of Lanford's strategy, of tools developed recently by Bodineau, Gallagher and Saint-Raymond to study the collision process, and of new duality arguments to study the additional terms associated with the long-range interaction, leading to some explicit weak estimates.
Gómez, Aina G; Ondiviela, Bárbara; Puente, Araceli; Juanes, José A
2015-05-15
This work presents a standard and unified procedure for assessment of environmental risks at the contaminant source level in port aquatic systems. Using this method, port managers and local authorities will be able to hierarchically classify environmental hazards and proceed with the most suitable management actions. This procedure combines rigorously selected parameters and indicators to estimate the environmental risk of each contaminant source based on its probability, consequences and vulnerability. The spatio-temporal variability of multiple stressors (agents) and receptors (endpoints) is taken into account to provide accurate estimations for application of precisely defined measures. The developed methodology is tested on a wide range of different scenarios via application in six European ports. The validation process confirms its usefulness, versatility and adaptability as a management tool for port water quality in Europe and worldwide. Copyright © 2015 Elsevier Ltd. All rights reserved.
Too many targets, not enough patients: rethinking neuroblastoma clinical trials.
Fletcher, Jamie I; Ziegler, David S; Trahair, Toby N; Marshall, Glenn M; Haber, Michelle; Norris, Murray D
2018-06-01
Neuroblastoma is a rare solid tumour of infancy and early childhood with a disproportionate contribution to paediatric cancer mortality and morbidity. Combination chemotherapy, radiation therapy and immunotherapy remains the standard approach to treat high-risk disease, with few recurrent, actionable genetic aberrations identified at diagnosis. However, recent studies indicate that actionable aberrations are far more common in relapsed neuroblastoma, possibly as a result of clonal expansion. In addition, although the major validated disease driver, MYCN, is not currently directly targetable, multiple promising approaches to target MYCN indirectly are in development. We propose that clinical trial design needs to be rethought in order to meet the challenge of providing rigorous, evidence-based assessment of these new approaches within a fairly small patient population and that experimental therapies need to be assessed at diagnosis in very-high-risk patients rather than in relapsed and refractory patients.
Perspective: Quantum Hamiltonians for optical interactions
NASA Astrophysics Data System (ADS)
Andrews, David L.; Jones, Garth A.; Salam, A.; Woolley, R. Guy
2018-01-01
The multipolar Hamiltonian of quantum electrodynamics is extensively employed in chemical and optical physics to treat rigorously the interaction of electromagnetic fields with matter. It is also widely used to evaluate intermolecular interactions. The multipolar version of the Hamiltonian is commonly obtained by carrying out a unitary transformation of the Coulomb gauge Hamiltonian that goes by the name of Power-Zienau-Woolley (PZW). Not only does the formulation provide excellent agreement with experiment, and versatility in its predictive ability, but also superior physical insight. Recently, the foundations and validity of the PZW Hamiltonian have been questioned, raising a concern over issues of gauge transformation and invariance, and whether observable quantities obtained from unitarily equivalent Hamiltonians are identical. Here, an in-depth analysis of theoretical foundations clarifies the issues and enables misconceptions to be identified. Claims of non-physicality are refuted: the PZW transformation and ensuing Hamiltonian are shown to rest on solid physical principles and secure theoretical ground.
Wakefield, J C
1993-02-01
Spitzer and Endicott (1978) proposed an operational definition of mental disorder that is a more rigorous version of the brief definitions that appeared in the 3rd and revised 3rd editions of the Diagnostic and Statistical Manual of Mental Disorders. The heart of their proposal is a translation of the concept of dysfunction into operational terms. I argue that their definition fails to capture the concept of dysfunction and is subject to many counterexamples. I use my harmful dysfunction account of disorder (Wakefield, 1992a, 1992b), which interprets dysfunction in evolutionary terms, to explain both the appeal and the problems of Spitzer and Endicott's definition and to provide support for the harmful dysfunction view. I conclude that the failure of Spitzer and Endicott's sophisticated attempt at operationalization indicates that nonoperational definitions that use functional concepts must play a role in formulating valid diagnostic criteria.
NASA Astrophysics Data System (ADS)
Bancă, Gheorghe; Ivan, Florian; Iozsa, Daniel; Nisulescu, Valentin
2017-10-01
Currently, the tendency of the car manufacturers is to continue the expansion of the global production of SUVs (Sport Utility Vehicle), while observing the requirements imposed by the new pollution standards by developing new technologies like DHEP (Diesel Hybrid Electric Powertrain). Experience has shown that the transient regimes are the most difficult to control from an economic and ecological perspective. As a result, this paper will highlight the behaviour of such engines that are provided in a middle class SUV (Sport Utility Vehicle), which operates in such states. We selected the transient regimes characteristic to the NMVEG (New Motor Vehicle Emissions Group) cycle. The investigations using the modelling platform AMESim allowed for rigorous interpretations for the 16 acceleration and 18 deceleration states. The results obtained from the simulation will be validated by experiments.
Network-based stochastic semisupervised learning.
Silva, Thiago Christiano; Zhao, Liang
2012-03-01
Semisupervised learning is a machine learning approach that is able to employ both labeled and unlabeled samples in the training process. In this paper, we propose a semisupervised data classification model based on a combined random-preferential walk of particles in a network (graph) constructed from the input dataset. The particles of the same class cooperate among themselves, while the particles of different classes compete with each other to propagate class labels to the whole network. A rigorous model definition is provided via a nonlinear stochastic dynamical system and a mathematical analysis of its behavior is carried out. A numerical validation presented in this paper confirms the theoretical predictions. An interesting feature brought by the competitive-cooperative mechanism is that the proposed model can achieve good classification rates while exhibiting low computational complexity order in comparison to other network-based semisupervised algorithms. Computer simulations conducted on synthetic and real-world datasets reveal the effectiveness of the model.
Howard, Steven J; Ehrich, John F; Walton, Russell
2014-01-01
Plagiarism is a significant area of concern in higher education, given university students' high self-reported rates of plagiarism. However, research remains inconsistent in prevalence estimates and suggested precursors of plagiarism. This may be a function of the unclear psychometric properties of the measurement tools adopted. To investigate this, we modified an existing plagiarism scale (to broaden its scope), established its psychometric properties using traditional (EFA, Cronbach's alpha) and modern (Rasch analysis) survey evaluation approaches, and examined results of well-functioning items. Results indicated that traditional and modern psychometric approaches differed in their recommendations. Further, responses indicated that although most respondents acknowledged the seriousness of plagiarism, these attitudes were neither unanimous nor consistent across the range of issues assessed. This study thus provides rigorous psychometric testing of a plagiarism attitude scale and baseline data from which to begin a discussion of contextual, personal, and external factors that influence students' plagiarism attitudes.
Finite-key analysis for measurement-device-independent quantum key distribution.
Curty, Marcos; Xu, Feihu; Cui, Wei; Lim, Charles Ci Wen; Tamaki, Kiyoshi; Lo, Hoi-Kwong
2014-04-29
Quantum key distribution promises unconditionally secure communications. However, as practical devices tend to deviate from their specifications, the security of some practical systems is no longer valid. In particular, an adversary can exploit imperfect detectors to learn a large part of the secret key, even though the security proof claims otherwise. Recently, a practical approach--measurement-device-independent quantum key distribution--has been proposed to solve this problem. However, so far its security has only been fully proven under the assumption that the legitimate users of the system have unlimited resources. Here we fill this gap and provide a rigorous security proof against general attacks in the finite-key regime. This is obtained by applying large deviation theory, specifically the Chernoff bound, to perform parameter estimation. For the first time we demonstrate the feasibility of long-distance implementations of measurement-device-independent quantum key distribution within a reasonable time frame of signal transmission.
Buckling of a stiff thin film on an elastic graded compliant substrate
NASA Astrophysics Data System (ADS)
Chen, Zhou; Chen, Weiqiu; Song, Jizhou
2017-12-01
The buckling of a stiff film on a compliant substrate has attracted much attention due to its wide applications such as thin-film metrology, surface patterning and stretchable electronics. An analytical model is established for the buckling of a stiff thin film on a semi-infinite elastic graded compliant substrate subjected to in-plane compression. The critical compressive strain and buckling wavelength for the sinusoidal mode are obtained analytically for the case with the substrate modulus decaying exponentially. The rigorous finite element analysis (FEA) is performed to validate the analytical model and investigate the postbuckling behaviour of the system. The critical buckling strain for the period-doubling mode is obtained numerically. The influences of various material parameters on the results are investigated. These results are helpful to provide physical insights on the buckling of elastic graded substrate-supported thin film.
Marine Corps Values-Based Ethics Training: A Recipe to Reduce Misconduct
2013-03-01
military personnel) own hearts and minds against the rigors of contemporary warfare...”39 10 Another report states, “Military sociologists and...future is likely to be not only on occasion soldier, policeman, ‘ hearts and minds ’ ambassador or general diplomat, but sometimes all of them...collection of information if it does not display a currently valid OMB control number. PLEASE DO NOT RETURN YOUR FORM TO THE ABOVE ADDRESS. 1
Control of Acoustics and Store Separation in a Cavity in Supersonic Flow
2005-02-01
laser -based flow visualization experiments on the FSU cavity for different microjet pressures. The details of the experiments are given in Zhuang, et. al...developed that rigorously explains the role of leading edge microjets in cavity noise suppression and predicts the magnitude of noise reduction for a...given control input (that is the steady pressure at which the microjets are fired). The model is validated through comparison of its noise reduction
TAMDAR Sensor Validation in 2003 AIRS II
NASA Technical Reports Server (NTRS)
Daniels, Taumi S.; Murray, John J.; Anderson, Mark V.; Mulally, Daniel J.; Jensen, Kristopher R.; Grainger, Cedric A.; Delene, David J.
2005-01-01
This study entails an assessment of TAMDAR in situ temperature, relative humidity and winds sensor data from seven flights of the UND Citation II. These data are undergoing rigorous assessment to determine their viability to significantly augment domestic Meteorological Data Communications Reporting System (MDCRS) and the international Aircraft Meteorological Data Reporting (AMDAR) system observational databases to improve the performance of regional and global numerical weather prediction models. NASA Langley Research Center participated in the Second Alliance Icing Research Study from November 17 to December 17, 2003. TAMDAR data taken during this period is compared with validation data from the UND Citation. The data indicate acceptable performance of the TAMDAR sensor when compared to measurements from the UND Citation research instruments.
Translating the short version of the Perinatal Grief Scale: process and challenges.
Capitulo, K L; Cornelio, M A; Lenz, E R
2001-08-01
Non-English-speaking populations may be excluded from rigorous clinical research because of the lack of reliable and valid instrumentation to measure psychosocial variables. The purpose of this article is to describe the process and challenges when translating a research instrument. The process will be illustrated in the project of translating into Spanish the Short Version of the Perinatal Grief Scale, extensively studied in English-speaking, primarily Caucasian populations. Translation methods, errors, and tips are included. Tools cannot be used in transcultural research and practice without careful and accurate translation and subsequent psychometric evaluation, which are essential to generate credible and valid findings. Copyright 2001 by W.B. Saunders Company
Pictogram Evaluation and Authoring Collaboration Environment
Kim, Hyeoneui; Tamayo, Dorothy; Muhkin, Michael; Kim, Jaemin; Lam, Julius; Ohno-Machado, Lucila; Aronoff-Spencer, Eliah
2012-01-01
Studies showed benefits of using pictograms in health communication such as improved recall and comprehension of health instructions. Pictograms are culturally sensitive thus need to be rigorously validated to ensure they convey the intended meaning correctly to the targeted population. The infeasibility of manually creating pictograms and the lack of robust means to store and validate pictograms are potential barriers to the wider adoption of pictograms in health communication. To address these challenges, we created an open access web-based tool, PEACE (Pictogram Evaluation and Authoring Collaboration Environment) as a part of SHINE (Sustainable Health Informatics and Networking Environment) initiatives. We report the development process and the preliminary evaluation results of PEACE in this paper. PMID:24199088
Rotation and anisotropy of galaxies revisited
NASA Astrophysics Data System (ADS)
Binney, James
2005-11-01
The use of the tensor virial theorem (TVT) as a diagnostic of anisotropic velocity distributions in galaxies is revisited. The TVT provides a rigorous global link between velocity anisotropy, rotation and shape, but the quantities appearing in it are not easily estimated observationally. Traditionally, use has been made of a centrally averaged velocity dispersion and the peak rotation velocity. Although this procedure cannot be rigorously justified, tests on model galaxies show that it works surprisingly well. With the advent of integral-field spectroscopy it is now possible to establish a rigorous connection between the TVT and observations. The TVT is reformulated in terms of sky-averages, and the new formulation is tested on model galaxies.
Validation and Spatiotemporal Analysis of CERES Surface Net Radiation Product
Jia, Aolin; Jiang, Bo; Liang, Shunlin; ...
2016-01-23
The Clouds and the Earth’s Radiant Energy System (CERES) generates one of the few global satellite radiation products. The CERES ARM Validation Experiment (CAVE) has been providing long-term in situ observations for the validation of the CERES products. However, the number of these sites is low and their distribution is globally sparse, and particularly the surface net radiation product has not been rigorously validated yet. Therefore, additional validation efforts are highly required to determine the accuracy of the CERES radiation products. In this study, global land surface measurements were comprehensively collected for use in the validation of the CERES netmore » radiation (R n) product on a daily (340 sites) and a monthly (260 sites) basis, respectively. The validation results demonstrated that the CERES R n product was, overall, highly accurate. The daily validations had a Mean Bias Error (MBE) of 3.43 W·m −2, Root Mean Square Error (RMSE) of 33.56 W·m −2, and R 2 of 0.79, and the monthly validations had an MBE of 3.40 W·m −2, RMSE of 25.57 W·m −2, and R 2 of 0.84. The accuracy was slightly lower for the high latitudes. Following the validation, the monthly CERES R n product, from March 2000 to July 2014, was used for a further analysis. We analysed the global spatiotemporal variation of the R n, which occurred during the measurement period. In addition, two hot spot regions, the southern Great Plains and south-central Africa, were then selected for use in determining the driving factors or attribution of the R n variation. We determined that R n over the southern Great Plains decreased by −0.33 W·m −2 per year, which was mainly driven by changes in surface green vegetation and precipitation. In south-central Africa, R n decreased at a rate of −0.63 W·m −2 per year, the major driving factor of which was surface green vegetation.« less
Validation and Spatiotemporal Analysis of CERES Surface Net Radiation Product
DOE Office of Scientific and Technical Information (OSTI.GOV)
Jia, Aolin; Jiang, Bo; Liang, Shunlin
The Clouds and the Earth’s Radiant Energy System (CERES) generates one of the few global satellite radiation products. The CERES ARM Validation Experiment (CAVE) has been providing long-term in situ observations for the validation of the CERES products. However, the number of these sites is low and their distribution is globally sparse, and particularly the surface net radiation product has not been rigorously validated yet. Therefore, additional validation efforts are highly required to determine the accuracy of the CERES radiation products. In this study, global land surface measurements were comprehensively collected for use in the validation of the CERES netmore » radiation (R n) product on a daily (340 sites) and a monthly (260 sites) basis, respectively. The validation results demonstrated that the CERES R n product was, overall, highly accurate. The daily validations had a Mean Bias Error (MBE) of 3.43 W·m −2, Root Mean Square Error (RMSE) of 33.56 W·m −2, and R 2 of 0.79, and the monthly validations had an MBE of 3.40 W·m −2, RMSE of 25.57 W·m −2, and R 2 of 0.84. The accuracy was slightly lower for the high latitudes. Following the validation, the monthly CERES R n product, from March 2000 to July 2014, was used for a further analysis. We analysed the global spatiotemporal variation of the R n, which occurred during the measurement period. In addition, two hot spot regions, the southern Great Plains and south-central Africa, were then selected for use in determining the driving factors or attribution of the R n variation. We determined that R n over the southern Great Plains decreased by −0.33 W·m −2 per year, which was mainly driven by changes in surface green vegetation and precipitation. In south-central Africa, R n decreased at a rate of −0.63 W·m −2 per year, the major driving factor of which was surface green vegetation.« less
The KP Approximation Under a Weak Coriolis Forcing
NASA Astrophysics Data System (ADS)
Melinand, Benjamin
2018-02-01
In this paper, we study the asymptotic behavior of weakly transverse water-waves under a weak Coriolis forcing in the long wave regime. We derive the Boussinesq-Coriolis equations in this setting and we provide a rigorous justification of this model. Then, from these equations, we derive two other asymptotic models. When the Coriolis forcing is weak, we fully justify the rotation-modified Kadomtsev-Petviashvili equation (also called Grimshaw-Melville equation). When the Coriolis forcing is very weak, we rigorously justify the Kadomtsev-Petviashvili equation. This work provides the first mathematical justification of the KP approximation under a Coriolis forcing.
Comparison of the GHSSmooth and the Rayleigh-Rice surface scatter theories
NASA Astrophysics Data System (ADS)
Harvey, James E.; Pfisterer, Richard N.
2016-09-01
The scalar-based GHSSmooth surface scatter theory results in an expression for the BRDF in terms of the surface PSD that is very similar to that provided by the rigorous Rayleigh-Rice (RR) vector perturbation theory. However it contains correction factors for two extreme situations not shared by the RR theory: (i) large incident or scattered angles that result in some portion of the scattered radiance distribution falling outside of the unit circle in direction cosine space, and (ii) the situation where the relevant rms surface roughness, σrel, is less than the total intrinsic rms roughness of the scattering surface. Also, the RR obliquity factor has been discovered to be an approximation of the more general GHSSmooth obliquity factor due to a little-known (or long-forgotten) implicit assumption in the RR theory that the surface autocovariance length is longer than the wavelength of the scattered radiation. This assumption allowed retaining only quadratic terms and lower in the series expansion for the cosine function, and results in reducing the validity of RR predictions for scattering angles greater than 60°. This inaccurate obliquity factor in the RR theory is also the cause of a complementary unrealistic "hook" at the high spatial frequency end of the predicted surface PSD when performing the inverse scattering problem. Furthermore, if we empirically substitute the polarization reflectance, Q, from the RR expression for the scalar reflectance, R, in the GHSSmooth expression, it inherits all of the polarization capabilities of the rigorous RR vector perturbation theory.
Park, Yoon Soo; Hyderi, Abbas; Heine, Nancy; May, Win; Nevins, Andrew; Lee, Ming; Bordage, Georges; Yudkowsky, Rachel
2017-11-01
To examine validity evidence of local graduation competency examination scores from seven medical schools using shared cases and to provide rater training protocols and guidelines for scoring patient notes (PNs). Between May and August 2016, clinical cases were developed, shared, and administered across seven medical schools (990 students participated). Raters were calibrated using training protocols, and guidelines were developed collaboratively across sites to standardize scoring. Data included scores from standardized patient encounters for history taking, physical examination, and PNs. Descriptive statistics were used to examine scores from the different assessment components. Generalizability studies (G-studies) using variance components were conducted to estimate reliability for composite scores. Validity evidence was collected for response process (rater perception), internal structure (variance components, reliability), relations to other variables (interassessment correlations), and consequences (composite score). Student performance varied by case and task. In the PNs, justification of differential diagnosis was the most discriminating task. G-studies showed that schools accounted for less than 1% of total variance; however, for the PNs, there were differences in scores for varying cases and tasks across schools, indicating a school effect. Composite score reliability was maximized when the PN was weighted between 30% and 40%. Raters preferred using case-specific scoring guidelines with clear point-scoring systems. This multisite study presents validity evidence for PN scores based on scoring rubric and case-specific scoring guidelines that offer rigor and feedback for learners. Variability in PN scores across participating sites may signal different approaches to teaching clinical reasoning among medical schools.
Shao, Jing; Tang, Leiwen; Ye, Zhihong
For the stabilization of the nursing profession in mainland China, a valid and reliable nursing work environment instrument, grounded in China's context, should be developed to better provide rigorous evidence for policy makers. The purpose of the current research was to further develop a scale that could capture the characteristics of the nursing work environment in mainland China. A convenience sample of 542 nurses employed in a tertiary hospital of mainland China completed the 108-item Chinese Nursing Work Environment (C-NWE) Scale (1st ed.). Items that did not differentiate between respondents with the highest and lowest 27% of total scores and those that did not meet criteria for factor loadings were set aside. Exploratory factor analysis based on the maximum likelihood method was used to identify the structure of the scale. The chi-square test was used to evaluate model fit, and expert review was conducted to test content validity. Reliability was estimated using Cronbach's alpha coefficient. The revised C-NWE Scale, which consists of nine subscales and 47 items, is a simplified version of the C-NWE (1st ed.). Using exploratory maximum likelihood factor analysis, the normed chi-square fit index for a nine-factor solution was 1.97. The content validity index for the total scale was 0.93; Cronbach's alpha was .94. Initial evidence of the psychometric properties of C-NWE scores was presented. Further studies could be conducted in various settings to identify the C-NWE Scale's validity and reliability.
Hohashi, Naohiro; Honda, Junko
2012-01-01
Hohashi's Concentric Sphere Family Environment Model (CSFEM; Hohashi & Honda, 2011) is a newly proposed family nursing theory for holistically understanding the family environment that acts on family well-being. The purpose of this article is to develop and psychometrically test the Japanese version of the Survey of Family Environment (SFE-J), grounded in the CSFEM, for measuring family's perceived family functioning and family's perceived needs for family support. The SFE-J is a 30-item self-administered instrument that assesses five domains (suprasystem, macrosystem, microsystem, family internal environment system, and chronosystem) and has been subjected to rigorous reliability and validity investigations among paired partners in child-rearing families (N of family = 1,990). Internal consistency reliability was high as measured by Cronbach's alpha coefficients. Temporal stability over a 2-week interval was supported by high (substantial or perfect) and significant intraclass correlation coefficients. The total score for the SFE-J was significantly correlated with the Japanese version of the Feetham Family Functioning Survey (FFFS-J), indicating an acceptable concurrent validity. Construct validity was supported by a confirmatory factor analysis that evaluated the five-factor structure to measure the concept of CSFEM. Results also demonstrate that the SFE-J family functioning scores show no significant differences between paired partners. The SFE-J is a reliable and valid instrument to assess not only intrafamily functioning but also interfamily functioning and, by identifying items/domains with high requirements for family support, serves to facilitate the providing of appropriate support to families.
Low validity of Google Trends for behavioral forecasting of national suicide rates.
Tran, Ulrich S; Andel, Rita; Niederkrotenthaler, Thomas; Till, Benedikt; Ajdacic-Gross, Vladeta; Voracek, Martin
2017-01-01
Recent research suggests that search volumes of the most popular search engine worldwide, Google, provided via Google Trends, could be associated with national suicide rates in the USA, UK, and some Asian countries. However, search volumes have mostly been studied in an ad hoc fashion, without controls for spurious associations. This study evaluated the validity and utility of Google Trends search volumes for behavioral forecasting of suicide rates in the USA, Germany, Austria, and Switzerland. Suicide-related search terms were systematically collected and respective Google Trends search volumes evaluated for availability. Time spans covered 2004 to 2010 (USA, Switzerland) and 2004 to 2012 (Germany, Austria). Temporal associations of search volumes and suicide rates were investigated with time-series analyses that rigorously controlled for spurious associations. The number and reliability of analyzable search volume data increased with country size. Search volumes showed various temporal associations with suicide rates. However, associations differed both across and within countries and mostly followed no discernable patterns. The total number of significant associations roughly matched the number of expected Type I errors. These results suggest that the validity of Google Trends search volumes for behavioral forecasting of national suicide rates is low. The utility and validity of search volumes for the forecasting of suicide rates depend on two key assumptions ("the population that conducts searches consists mostly of individuals with suicidal ideation", "suicide-related search behavior is strongly linked with suicidal behavior"). We discuss strands of evidence that these two assumptions are likely not met. Implications for future research with Google Trends in the context of suicide research are also discussed.
Jones, J.W.; Jarnagin, T.
2009-01-01
Given the relatively high cost of mapping impervious surfaces at regional scales, substantial effort is being expended in the development of moderate-resolution, satellite-based methods for estimating impervious surface area (ISA). To rigorously assess the accuracy of these data products high quality, independently derived validation data are needed. High-resolution data were collected across a gradient of development within the Mid-Atlantic region to assess the accuracy of National Land Cover Data (NLCD) Landsat-based ISA estimates. Absolute error (satellite predicted area - "reference area") and relative error [satellite (predicted area - "reference area")/ "reference area"] were calculated for each of 240 sample regions that are each more than 15 Landsat pixels on a side. The ability to compile and examine ancillary data in a geographic information system environment provided for evaluation of both validation and NLCD data and afforded efficient exploration of observed errors. In a minority of cases, errors could be explained by temporal discontinuities between the date of satellite image capture and validation source data in rapidly changing places. In others, errors were created by vegetation cover over impervious surfaces and by other factors that bias the satellite processing algorithms. On average in the Mid-Atlantic region, the NLCD product underestimates ISA by approximately 5%. While the error range varies between 2 and 8%, this underestimation occurs regardless of development intensity. Through such analyses the errors, strengths, and weaknesses of particular satellite products can be explored to suggest appropriate uses for regional, satellite-based data in rapidly developing areas of environmental significance. ?? 2009 ASCE.
The SF36 health survey questionnaire: an outcome measure suitable for routine use within the NHS?
Garratt, A M; Ruta, D A; Abdalla, M I; Buckingham, J K; Russell, I T
1993-01-01
OBJECTIVE--To assess the validity, reliability, and acceptability of the short form 36 (SF 36) health survey questionnaire (a shortened version of a battery of 149 health status questions) as a measure of patient outcome in a broad sample of patients suffering from four common clinical conditions. DESIGN--Postal questionnaire, followed up by two reminders at two week intervals. SETTING--Clinics and four training practices in north east Scotland. SUBJECTS--Over 1700 patients aged 16-86 with one of four conditions--low back pain, menorrhagia, suspected peptic ulcer, or varicose veins--and a comparison sample of 900 members of the general population. MAIN OUTCOME MEASURES--The eight scales within the SF36 health profile. RESULTS--The response rate exceeded 75% in the patient population (1310 respondents). The SF36 satisfied rigorous psychometric criteria for validity and internal consistency. Clinical validity was shown by the distinctive profiles generated for each condition, each of which differed from that in the general population in a predictable manner. Furthermore, SF36 scores were lower in referred patients than in patients not referred and were closely related to general practitioners' perceptions of severity. CONCLUSIONS--These results provide support for the SF36 as a potential measure of patient outcome within the NHS. The SF36 seems acceptable to patients, internally consistent, and a valid measure of the health status of a wide range of patients. Before it can be used in the new health service, however, its sensitivity to changes in health status over time must also be tested. PMID:8518640
PASTIS: Bayesian extrasolar planet validation - I. General framework, models, and performance
NASA Astrophysics Data System (ADS)
Díaz, R. F.; Almenara, J. M.; Santerne, A.; Moutou, C.; Lethuillier, A.; Deleuil, M.
2014-06-01
A large fraction of the smallest transiting planet candidates discovered by the Kepler and CoRoT space missions cannot be confirmed by a dynamical measurement of the mass using currently available observing facilities. To establish their planetary nature, the concept of planet validation has been advanced. This technique compares the probability of the planetary hypothesis against that of all reasonably conceivable alternative false positive (FP) hypotheses. The candidate is considered as validated if the posterior probability of the planetary hypothesis is sufficiently larger than the sum of the probabilities of all FP scenarios. In this paper, we present PASTIS, the Planet Analysis and Small Transit Investigation Software, a tool designed to perform a rigorous model comparison of the hypotheses involved in the problem of planet validation, and to fully exploit the information available in the candidate light curves. PASTIS self-consistently models the transit light curves and follow-up observations. Its object-oriented structure offers a large flexibility for defining the scenarios to be compared. The performance is explored using artificial transit light curves of planets and FPs with a realistic error distribution obtained from a Kepler light curve. We find that data support the correct hypothesis strongly only when the signal is high enough (transit signal-to-noise ratio above 50 for the planet case) and remain inconclusive otherwise. PLAnetary Transits and Oscillations of stars (PLATO) shall provide transits with high enough signal-to-noise ratio, but to establish the true nature of the vast majority of Kepler and CoRoT transit candidates additional data or strong reliance on hypotheses priors is needed.
Austin, J Matthew; Demski, Renee; Callender, Tiffany; Lee, K H Ken; Hoffman, Ann; Allen, Lisa; Radke, Deborah A; Kim, Yungjin; Werthman, Ronald J; Peterson, Ronald R; Pronovost, Peter J
2017-04-01
As the health care system in the United States places greater emphasis on the public reporting of quality and safety data and its use to determine payment, provider organizations must implement structures that ensure discipline and rigor regarding these data. An academic health system, as part of a performance management system, applied four key components of a financial reporting structure to support the goal of top-to-bottom accountability for improving quality and safety. The four components implemented by Johns Hopkins Medicine were governance, accountability, reporting of consolidated quality performance statements, and auditing. Governance is provided by the health system's Patient Safety and Quality Board Committee, which reviews goals and strategy for patient safety and quality, reviews quarterly performance for each entity, and holds organizational leaders accountable for performance. An accountability plan includes escalating levels of review corresponding to the number of months an entity misses the defined performance target for a measure. A consolidated quality statement helps inform the Patient Safety and Quality Board Committee and leadership on key quality and safety issues. An audit evaluates the efficiency and effectiveness of processes for data collection, validation, and storage, as to ensure the accuracy and completeness of quality measure reporting. If hospitals and health systems truly want to prioritize improvements in safety and quality, they will need to create a performance management system that ensures data validity and supports performance accountability. Without valid data, it is difficult to know whether a performance gap is due to data quality or clinical quality. Copyright © 2017 The Joint Commission. Published by Elsevier Inc. All rights reserved.
A Standardized Reference Data Set for Vertebrate Taxon Name Resolution
Zermoglio, Paula F.; Guralnick, Robert P.; Wieczorek, John R.
2016-01-01
Taxonomic names associated with digitized biocollections labels have flooded into repositories such as GBIF, iDigBio and VertNet. The names on these labels are often misspelled, out of date, or present other problems, as they were often captured only once during accessioning of specimens, or have a history of label changes without clear provenance. Before records are reliably usable in research, it is critical that these issues be addressed. However, still missing is an assessment of the scope of the problem, the effort needed to solve it, and a way to improve effectiveness of tools developed to aid the process. We present a carefully human-vetted analysis of 1000 verbatim scientific names taken at random from those published via the data aggregator VertNet, providing the first rigorously reviewed, reference validation data set. In addition to characterizing formatting problems, human vetting focused on detecting misspelling, synonymy, and the incorrect use of Darwin Core. Our results reveal a sobering view of the challenge ahead, as less than 47% of name strings were found to be currently valid. More optimistically, nearly 97% of name combinations could be resolved to a currently valid name, suggesting that computer-aided approaches may provide feasible means to improve digitized content. Finally, we associated names back to biocollections records and fit logistic models to test potential drivers of issues. A set of candidate variables (geographic region, year collected, higher-level clade, and the institutional digitally accessible data volume) and their 2-way interactions all predict the probability of records having taxon name issues, based on model selection approaches. We strongly encourage further experiments to use this reference data set as a means to compare automated or computer-aided taxon name tools for their ability to resolve and improve the existing wealth of legacy data. PMID:26760296
2016-01-01
Reliably estimating wildlife abundance is fundamental to effective management. Aerial surveys are one of the only spatially robust tools for estimating large mammal populations, but statistical sampling methods are required to address detection biases that affect accuracy and precision of the estimates. Although various methods for correcting aerial survey bias are employed on large mammal species around the world, these have rarely been rigorously validated. Several populations of feral horses (Equus caballus) in the western United States have been intensively studied, resulting in identification of all unique individuals. This provided a rare opportunity to test aerial survey bias correction on populations of known abundance. We hypothesized that a hybrid method combining simultaneous double-observer and sightability bias correction techniques would accurately estimate abundance. We validated this integrated technique on populations of known size and also on a pair of surveys before and after a known number was removed. Our analysis identified several covariates across the surveys that explained and corrected biases in the estimates. All six tests on known populations produced estimates with deviations from the known value ranging from -8.5% to +13.7% and <0.7 standard errors. Precision varied widely, from 6.1% CV to 25.0% CV. In contrast, the pair of surveys conducted around a known management removal produced an estimated change in population between the surveys that was significantly larger than the known reduction. Although the deviation between was only 9.1%, the precision estimate (CV = 1.6%) may have been artificially low. It was apparent that use of a helicopter in those surveys perturbed the horses, introducing detection error and heterogeneity in a manner that could not be corrected by our statistical models. Our results validate the hybrid method, highlight its potentially broad applicability, identify some limitations, and provide insight and guidance for improving survey designs. PMID:27139732
Lubow, Bruce C; Ransom, Jason I
2016-01-01
Reliably estimating wildlife abundance is fundamental to effective management. Aerial surveys are one of the only spatially robust tools for estimating large mammal populations, but statistical sampling methods are required to address detection biases that affect accuracy and precision of the estimates. Although various methods for correcting aerial survey bias are employed on large mammal species around the world, these have rarely been rigorously validated. Several populations of feral horses (Equus caballus) in the western United States have been intensively studied, resulting in identification of all unique individuals. This provided a rare opportunity to test aerial survey bias correction on populations of known abundance. We hypothesized that a hybrid method combining simultaneous double-observer and sightability bias correction techniques would accurately estimate abundance. We validated this integrated technique on populations of known size and also on a pair of surveys before and after a known number was removed. Our analysis identified several covariates across the surveys that explained and corrected biases in the estimates. All six tests on known populations produced estimates with deviations from the known value ranging from -8.5% to +13.7% and <0.7 standard errors. Precision varied widely, from 6.1% CV to 25.0% CV. In contrast, the pair of surveys conducted around a known management removal produced an estimated change in population between the surveys that was significantly larger than the known reduction. Although the deviation between was only 9.1%, the precision estimate (CV = 1.6%) may have been artificially low. It was apparent that use of a helicopter in those surveys perturbed the horses, introducing detection error and heterogeneity in a manner that could not be corrected by our statistical models. Our results validate the hybrid method, highlight its potentially broad applicability, identify some limitations, and provide insight and guidance for improving survey designs.
NASA Astrophysics Data System (ADS)
Peresan, Antonella; Kossobokov, Vladimir; Romashkova, Leontina; Panza, Giuliano F.
2017-04-01
Predicting earthquakes and related ground shaking is widely recognized among the most challenging scientific problems, both for societal relevance and intrinsic complexity of the problem. The development of reliable forecasting tools requires their rigorous formalization and testing, first in retrospect, and then in an experimental real-time mode, which imply a careful application of statistics to data sets of limited size and different accuracy. Accordingly, the operational issues of prospective validation and use of time-dependent neo-deterministic seismic hazard scenarios are discussed, reviewing the results in their application in Italy and surroundings. Long-term practice and results obtained for the Italian territory in about two decades of rigorous prospective testing, support the feasibility of earthquake forecasting based on the analysis of seismicity patterns at the intermediate-term middle-range scale. Italy is the only country worldwide where two independent, globally tested, algorithms are simultaneously applied, namely CN and M8S, which permit to deal with multiple sets of seismic precursors to allow for a diagnosis of the intervals of time when a strong event is likely to occur inside a given region. Based on routinely updated space-time information provided by CN and M8S forecasts, an integrated procedure has been developed that allows for the definition of time-dependent seismic hazard scenarios, through the realistic modeling of ground motion by the neo-deterministic approach (NDSHA). This scenario-based methodology permits to construct, both at regional and local scale, scenarios of ground motion for the time interval when a strong event is likely to occur within the alerted areas. CN and M8S predictions, as well as the related time-dependent ground motion scenarios associated with the alarmed areas, are routinely updated since 2006. The issues and results from real-time testing of the integrated NDSHA scenarios are illustrated, with special emphasis on the sequence of destructive earthquakes that struck Central Italy starting on August 2016. The results obtained so far evidence the validity of the proposed methodology in anticipating ground shaking from approaching strong earthquakes and prove that the information provided by time-dependent NDSHA can be useful in assigning priorities for timely and effective mitigation actions.
Validation of Land Surface Temperature from Sentinel-3
NASA Astrophysics Data System (ADS)
Ghent, D.
2017-12-01
One of the main objectives of the Sentinel-3 mission is to measure sea- and land-surface temperature with high-end accuracy and reliability in support of environmental and climate monitoring in an operational context. Calibration and validation are thus key criteria for operationalization within the framework of the Sentinel-3 Mission Performance Centre (S3MPC). Land surface temperature (LST) has a long heritage of satellite observations which have facilitated our understanding of land surface and climate change processes, such as desertification, urbanization, deforestation and land/atmosphere coupling. These observations have been acquired from a variety of satellite instruments on platforms in both low-earth orbit and in geostationary orbit. Retrieval accuracy can be a challenge though; surface emissivities can be highly variable owing to the heterogeneity of the land, and atmospheric effects caused by the presence of aerosols and by water vapour absorption can give a bias to the underlying LST. As such, a rigorous validation is critical in order to assess the quality of the data and the associated uncertainties. Validation of the level-2 SL_2_LST product, which became freely available on an operational basis from 5th July 2017 builds on an established validation protocol for satellite-based LST. This set of guidelines provides a standardized framework for structuring LST validation activities. The protocol introduces a four-pronged approach which can be summarised thus: i) in situ validation where ground-based observations are available; ii) radiance-based validation over sites that are homogeneous in emissivity; iii) intercomparison with retrievals from other satellite sensors; iv) time-series analysis to identify artefacts on an interannual time-scale. This multi-dimensional approach is a necessary requirement for assessing the performance of the LST algorithm for the Sea and Land Surface Temperature Radiometer (SLSTR) which is designed around biome-based coefficients, thus emphasizing the importance of non-traditional forms of validation such as radiance-based techniques. Here we present examples of the ongoing routine application of the protocol to operational Sentinel-3 LST data.
Edvardsson, David; Fetherstonhaugh, Deirdre; Nay, Rhonda
2011-10-01
To construct and evaluate an intervention tool for increasing the person-centredness of care in residential aged care services. Providing care that is person-centred and evidence-based is increasingly being regarded as synonymous with best quality aged care. However, consensus about how person-centred care should be defined, operationalised and implemented has not yet been reached. Literature reviews, expert consultation (n = 22) and stakeholder interviews (n = 67) were undertaken to develop the Tool for Understanding Residents' Needs as Individual Persons (TURNIP). Statistical estimates of validity and reliability were employed to evaluate the tool in an Australian convenience sample of aged care staff (n = 220). The 39 item TURNIP conceptualised person-centred care into five dimensions: (1) the care environment, (2) staff members' attitudes towards dementia, (3) staff members' knowledge about dementia, (4) the care organisation and (5) the content of care provided. Psychometric testing indicated satisfactory validity and reliability, as shown for example in a total Cronbach's alpha of 0·89. The TURNIP adds to current literature on person-centred care by presenting a rigorously developed intervention tool based on an explicit conceptual structure that can inform the design, employment and communication of clinical interventions aiming to promote person-centred care. The TURNIP contains clinically relevant items that are ready to be applied in clinical aged care. The tool can be used as a base for clinical interventions applying discussions in aged care organisations about the quality of current care and how to increase person-centredness of the care provided. © 2011 Blackwell Publishing Ltd.
Oliveira, Thaís D; Costa, Danielle de S; Albuquerque, Maicon R; Malloy-Diniz, Leandro F; Miranda, Débora M; de Paula, Jonas J
2018-06-11
The Parenting Styles and Dimensions Questionnaire (PSDQ) is used worldwide to assess three styles (authoritative, authoritarian, and permissive) and seven dimensions of parenting. In this study, we adapted the short version of the PSDQ for use in Brazil and investigated its validity and reliability. Participants were 451 mothers of children aged 3 to 18 years, though sample size varied with analyses. The translation and adaptation of the PSDQ followed a rigorous methodological approach. Then, we investigated the content, criterion, and construct validity of the adapted instrument. The scale content validity index (S-CVI) was considered adequate (0.97). There was evidence of internal validity, with the PSDQ dimensions showing strong correlations with their higher-order parenting styles. Confirmatory factor analysis endorsed the three-factor, second-order solution (i.e., three styles consisting of seven dimensions). The PSDQ showed convergent validity with the validated Brazilian version of the Parenting Styles Inventory (Inventário de Estilos Parentais - IEP), as well as external validity, as it was associated with several instruments measuring sociodemographic and behavioral/emotional-problem variables. The PSDQ is an effective and reliable psychometric instrument to assess childrearing strategies according to Baumrind's model of parenting styles.
Reicks, Marla; Trofholz, Amanda C.; Stang, Jamie S; Laska, Melissa N.
2014-01-01
Objective Cooking programs are growing in popularity; however an extensive review has not examined overall impact. Therefore, this study reviewed previous research on cooking/home food preparation interventions and diet and health-related outcomes among adults and identified implications for practice and research. Design Literature review and descriptive summative method. Main outcome measures Dietary intake, knowledge/skills, cooking attitudes and self-efficacy/confidence, health outcomes. Analysis Articles evaluating effectiveness of interventions that included cooking/home food preparation as the primary aim (January 1980 through December 2011) were identified via OVID MEDLINE, Agricola and Web of Science databases. Studies grouped according to design and outcomes were reviewed for validity using an established coding system. Results were summarized for several outcome categories. Results Of 28 studies identified, 12 included a control group with six as non-randomized and six as randomized controlled trials. Evaluation was done post-intervention for five studies, pre- and post-intervention for 23 and beyond post-intervention for 15. Qualitative and quantitative measures suggested a positive influence on main outcomes. However, non-rigorous study designs, varying study populations, and use of non-validated assessment tools limited stronger conclusions. Conclusions and Implications Well-designed studies are needed that rigorously evaluate long-term impact on cooking behavior, dietary intake, obesity and other health outcomes. PMID:24703245
Hypothetical Case and Scenario Description for International Transportation of Spent Nuclear Fuel.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Williams, Adam David; Osborn, Douglas; Jones, Katherine A.
To support more rigorous analysis on global security issues at Sandia National Laboratories (SNL), there is a need to develop realistic data sets without using "real" data or identifying "real" vulnerabilities, hazards or geopolitically embarrassing shortcomings. In response, an interdisciplinary team led by subject matter experts in SNL's Center for Global Security and Cooperation (CGSC) developed a hypothetical case description. This hypothetical case description assigns various attributes related to international SNF transportation that are representative, illustrative and indicative of "real" characteristics of "real" countries. There is no intent to identify any particular country and any similarity with specific real-world eventsmore » is purely coincidental. To support the goal of this report to provide a case description (and set of scenarios of concern) for international SNF transportation inclusive of as much "real-world" complexity as possible -- without crossing over into politically sensitive or classified information -- this SAND report provides a subject matter expert-validated (and detailed) description of both technical and political influences on the international transportation of spent nuclear fuel. [PAGE INTENTIONALLY LEFT BLANK]« less
Blank, Melissa D.; Breland, Alison B.; Cobb, Caroline O.; Spindle, Tory; Ramôa, Carolina; Eissenberg, Thomas
2017-01-01
Objective Evaluating electronic cigarettes (ECIGs) in the clinical laboratory is critical to understanding their effects. However, laboratory evaluation of ECIGs can be challenging, as they are a novel, varied, and evolving class of products. The objective of this paper is to describe some methodological challenges to the clinical laboratory evaluation of ECIGs. Methods The authors gathered information about challenges involved in the laboratory evaluation of ECIGs. Challenges were categorized and solutions provided when possible. Results Methods used to study combustible cigarettes may need to be adapted to account for ECIG novelty and differences within the class. Challenges to ECIG evaluation can include issues related to 1) identification of ECIG devices and liquids, 2) determination of short -term ECIG abstinence, 3) measurement of use behavior, and 4) assessment of dependence. These challenges are discussed, and some suggestions to inform ECIG evaluation using clinical laboratory methods are provided. Conclusions Awareness of challenges and developing, validating, and reporting methods used to address them aids interpretation of results and replication efforts, thus enhancing the rigor of science used to protect public health through appropriate, empirically-based, ECIG regulation. PMID:28819633
Blank, Melissa D; Breland, Alison B; Cobb, Caroline O; Spindle, Tory; Ramôa, Carolina; Eissenberg, Thomas
2016-10-01
Evaluating electronic cigarettes (ECIGs) in the clinical laboratory is critical to understanding their effects. However, laboratory evaluation of ECIGs can be challenging, as they are a novel, varied, and evolving class of products. The objective of this paper is to describe some methodological challenges to the clinical laboratory evaluation of ECIGs. The authors gathered information about challenges involved in the laboratory evaluation of ECIGs. Challenges were categorized and solutions provided when possible. Methods used to study combustible cigarettes may need to be adapted to account for ECIG novelty and differences within the class. Challenges to ECIG evaluation can include issues related to 1) identification of ECIG devices and liquids, 2) determination of short -term ECIG abstinence, 3) measurement of use behavior, and 4) assessment of dependence. These challenges are discussed, and some suggestions to inform ECIG evaluation using clinical laboratory methods are provided. Awareness of challenges and developing, validating, and reporting methods used to address them aids interpretation of results and replication efforts, thus enhancing the rigor of science used to protect public health through appropriate, empirically-based, ECIG regulation.
A new method to evaluate human-robot system performance
NASA Technical Reports Server (NTRS)
Rodriguez, G.; Weisbin, C. R.
2003-01-01
One of the key issues in space exploration is that of deciding what space tasks are best done with humans, with robots, or a suitable combination of each. In general, human and robot skills are complementary. Humans provide as yet unmatched capabilities to perceive, think, and act when faced with anomalies and unforeseen events, but there can be huge potential risks to human safety in getting these benefits. Robots provide complementary skills in being able to work in extremely risky environments, but their ability to perceive, think, and act by themselves is currently not error-free, although these capabilities are continually improving with the emergence of new technologies. Substantial past experience validates these generally qualitative notions. However, there is a need for more rigorously systematic evaluation of human and robot roles, in order to optimize the design and performance of human-robot system architectures using well-defined performance evaluation metrics. This article summarizes a new analytical method to conduct such quantitative evaluations. While the article focuses on evaluating human-robot systems, the method is generally applicable to a much broader class of systems whose performance needs to be evaluated.
Brosseau, Lucie; Laroche, Chantal; Sutton, Anne; Guitard, Paulette; King, Judy; Poitras, Stéphane; Casimiro, Lynn; Tremblay, Manon; Cardinal, Dominique; Cavallo, Sabrina; Laferrière, Lucie; Grisé, Isabelle; Marshall, Lisa; Smith, Jacky R; Lagacé, Josée; Pharand, Denyse; Galipeau, Roseline; Toupin-April, Karine; Loew, Laurianne; Demers, Catrine; Sauvé-Schenk, Katrine; Paquet, Nicole; Savard, Jacinthe; Tourigny, Jocelyne; Vaillancourt, Véronique
2015-08-01
To prepare a Canadian French translation of the PEDro Scale under the proposed name l'Échelle PEDro, and to examine the validity of its content. A modified approach of Vallerand's cross-cultural validation methodology was used, beginning with a parallel back-translation of the PEDro scale by both professional translators and clinical researchers. These versions were reviewed by an initial panel of experts (P1), who then created the first experimental version of l'Échelle PEDro. This version was evaluated by a second panel of experts (P2). Finally, 32 clinical researchers evaluated the second experimental version of l'Échelle PEDro, using a 5-point clarity scale, and suggested final modifications. The various items on the final version of l'Échelle PEDro show a high degree of clarity (from 4.0 to 4.7 on the 5-point scale). The four rigorous steps of the translation process have produced a valid Canadian French version of the PEDro scale.
Interoperable Data Sharing for Diverse Scientific Disciplines
NASA Astrophysics Data System (ADS)
Hughes, John S.; Crichton, Daniel; Martinez, Santa; Law, Emily; Hardman, Sean
2016-04-01
For diverse scientific disciplines to interoperate they must be able to exchange information based on a shared understanding. To capture this shared understanding, we have developed a knowledge representation framework using ontologies and ISO level archive and metadata registry reference models. This framework provides multi-level governance, evolves independent of implementation technologies, and promotes agile development, namely adaptive planning, evolutionary development, early delivery, continuous improvement, and rapid and flexible response to change. The knowledge representation framework is populated through knowledge acquisition from discipline experts. It is also extended to meet specific discipline requirements. The result is a formalized and rigorous knowledge base that addresses data representation, integrity, provenance, context, quantity, and their relationships within the community. The contents of the knowledge base is translated and written to files in appropriate formats to configure system software and services, provide user documentation, validate ingested data, and support data analytics. This presentation will provide an overview of the framework, present the Planetary Data System's PDS4 as a use case that has been adopted by the international planetary science community, describe how the framework is being applied to other disciplines, and share some important lessons learned.
Virtual reality and cognitive rehabilitation: a review of current outcome research.
Larson, Eric B; Feigon, Maia; Gagliardo, Pablo; Dvorkin, Assaf Y
2014-01-01
Recent advancement in the technology of virtual reality (VR) has allowed improved applications for cognitive rehabilitation. The aim of this review is to facilitate comparisons of therapeutic efficacy of different VR interventions. A systematic approach for the review of VR cognitive rehabilitation outcome research addressed the nature of each sample, treatment apparatus, experimental treatment protocol, control treatment protocol, statistical analysis and results. Using this approach, studies that provide valid evidence of efficacy of VR applications are summarized. Applications that have not yet undergone controlled outcome study but which have promise are introduced. Seventeen studies conducted over the past eight years are reviewed. The few randomized controlled trials that have been completed show that some applications are effective in treating cognitive deficits in people with neurological diagnoses although further study is needed. Innovations requiring further study include the use of enriched virtual environments that provide haptic sensory input in addition to visual and auditory inputs and the use of commercially available gaming systems to provide tele-rehabilitation services. Recommendations are offered to improve efficacy of rehabilitation, to improve scientific rigor of rehabilitation research and to broaden access to the evidence-based treatments that this research has identified.
Evaluating abundance estimate precision and the assumptions of a count-based index for small mammals
Wiewel, A.S.; Adams, A.A.Y.; Rodda, G.H.
2009-01-01
Conservation and management of small mammals requires reliable knowledge of population size. We investigated precision of markrecapture and removal abundance estimates generated from live-trapping and snap-trapping data collected at sites on Guam (n 7), Rota (n 4), Saipan (n 5), and Tinian (n 3), in the Mariana Islands. We also evaluated a common index, captures per unit effort (CPUE), as a predictor of abundance. In addition, we evaluated cost and time associated with implementing live-trapping and snap-trapping and compared species-specific capture rates of selected live- and snap-traps. For all species, markrecapture estimates were consistently more precise than removal estimates based on coefficients of variation and 95 confidence intervals. The predictive utility of CPUE was poor but improved with increasing sampling duration. Nonetheless, modeling of sampling data revealed that underlying assumptions critical to application of an index of abundance, such as constant capture probability across space, time, and individuals, were not met. Although snap-trapping was cheaper and faster than live-trapping, the time difference was negligible when site preparation time was considered. Rattus diardii spp. captures were greatest in Haguruma live-traps (Standard Trading Co., Honolulu, HI) and Victor snap-traps (Woodstream Corporation, Lititz, PA), whereas Suncus murinus and Mus musculus captures were greatest in Sherman live-traps (H. B. Sherman Traps, Inc., Tallahassee, FL) and Museum Special snap-traps (Woodstream Corporation). Although snap-trapping and CPUE may have utility after validation against more rigorous methods, validation should occur across the full range of study conditions. Resources required for this level of validation would likely be better allocated towards implementing rigorous and robust methods.
Tomizawa, Ryoko; Yamano, Mayumi; Osako, Mitue; Hirabayashi, Naotugu; Oshima, Nobuo; Sigeta, Masahiro; Reeves, Scott
2017-12-01
Few scales currently exist to assess the quality of interprofessional teamwork through team members' perceptions of working together in mental health settings. The purpose of this study was to revise and validate an interprofessional scale to assess the quality of teamwork in inpatient psychiatric units and to use it multi-nationally. A literature review was undertaken to identify evaluative teamwork tools and develop an additional 12 items to ensure a broad global focus. Focus group discussions considered adaptation to different care systems using subjective judgements from 11 participants in a pre-test of items. Data quality, construct validity, reproducibility, and internal consistency were investigated in the survey using an international comparative design. Exploratory factor analysis yielded five factors with 21 items: 'patient/community centred care', 'collaborative communication', 'interprofessional conflict', 'role clarification', and 'environment'. High overall internal consistency, reproducibility, adequate face validity, and reasonable construct validity were shown in the USA and Japan. The revised Collaborative Practice Assessment Tool (CPAT) is a valid measure to assess the quality of interprofessional teamwork in psychiatry and identifies the best strategies to improve team performance. Furthermore, the revised scale will generate more rigorous evidence for collaborative practice in psychiatry internationally.
Veldhoen, Nik; Propper, Catherine R; Helbing, Caren C
2014-03-01
Studies performed across diverse frog species have made substantial contributions to our understanding of basic vertebrate development and the natural or anthropogenic environmental factors impacting sensitive life stages. Because, anurans are developmental models, provide ecosystems services, and act as sentinels for the identification of environmental chemical contaminants that interfere with thyroid hormone (TH) action during postembryonic development, there is demand for flexible assessment techniques that can be applied to multiple species. As part of the "thyroid assays across indicator and sentinel species" (TAXISS) initiative, we have designed and validated a series of cross-species real time quantitative PCR (qPCR) primer sets that provide information on transcriptome components in evolutionarily distant anurans. Validation for fifteen gene transcripts involved a rigorous three-tiered quality control within tissue/development-specific contexts. Assay performance was confirmed on multiple tissues (tail fin, liver, brain, and intestine) of Rana catesbeiana and Xenopus laevis tadpoles enabling comparisons between tissues and generation of response profiles to exogenous TH. This revealed notable differences in TH-responsive gene transcripts including thra, thrb, thibz, klf9, col1a2, fn1, plp1, mmp2, timm50, otc, and dio2, suggesting differential regulation and susceptibility to contaminant effects. Evidence for the applicability of the TAXISS anuran qPCR assay across seven other species is also provided with five frog families represented and its utility in defining genome structure was demonstrated. This novel validated approach will enable meaningful comparative studies between frog species and aid in extending knowledge of developmental regulatory pathways and the impact of environmental factors on TH signaling in frog species for which little or no genetic information is currently available. Copyright © 2014 Elsevier B.V. All rights reserved.
NASA Technical Reports Server (NTRS)
Brodnick, Jacob; Richardson, Brian; Ramachandran, Narayanan
2015-01-01
The Low Profile Diffuser (LPD) project originated as an award from the Marshall Space Flight Center (MSFC) Advanced Development (ADO) office to the Main Propulsion Systems Branch (ER22). The task was created to develop and test an LPD concept that could produce comparable performance to a larger, traditionally designed, ullage gas diffuser while occupying a smaller volume envelope. Historically, ullage gas diffusers have been large, bulky devices that occupy a significant portion of the propellant tank, decreasing the tank volume available for propellant. Ullage pressurization of spacecraft propellant tanks is required to prevent boil-off of cryogenic propellants and to provide a positive pressure for propellant extraction. To achieve this, ullage gas diffusers must slow hot, high-pressure gas entering a propellant tank from supersonic speeds to only a few meters per second. Decreasing the incoming gas velocity is typically accomplished through expansion to larger areas within the diffuser which has traditionally led to large diffuser lengths. The Fluid Dynamics Branch (ER42) developed and applied advanced Computational Fluid Dynamics (CFD) analysis methods in order to mature the LPD design from and initial concept to an optimized test prototype and to provide extremely accurate pre-test predictions of diffuser performance. Additionally, the diffuser concept for the Core Stage of the Space Launch System (SLS) was analyzed in a short amount of time to guide test data collection efforts of the qualification of the device. CFD analysis of the SLS diffuser design provided new insights into the functioning of the device and was qualitatively validated against hot wire anemometry of the exterior flow field. Rigorous data analysis of the measurements was performed on static and dynamic pressure data, data from two microphones, accelerometers and hot wire anemometry with automated traverse. Feasibility of the LPD concept and validation of the computational model were demonstrated by the test data.
Mechanical properties of frog skeletal muscles in iodoacetic acid rigor.
Mulvany, M J
1975-01-01
1. Methods have been developed for describing the length: tension characteristics of frog skeletal muscles which go into rigor at 4 degrees C following iodoacetic acid poisoning either in the presence of Ca2+ (Ca-rigor) or its absence (Ca-free-rigor). 2. Such rigor muscles showed less resistance to slow stretch (slow rigor resistance) that to fast stretch (fast rigor resistance). The slow and fast rigor resistances of Ca-free-rigor muscles were much lower than those of Ca-rigor muscles. 3. The slow rigor resistance of Ca-rigor muscles was proportional to the amount of overlap between the contractile filaments present when the muscles were put into rigor. 4. Withdrawing Ca2+ from Ca-rigor muscles (induced-Ca-free rigor) reduced their slow and fast rigor resistances. Readdition of Ca2+ (but not Mg2+, Mn2+ or Sr2+) reversed the effect. 5. The slow and fast rigor resistances of Ca-rigor muscles (but not of Ca-free-rigor muscles) decreased with time. 6.The sarcomere structure of Ca-rigor and induced-Ca-free rigor muscles stretched by 0.2lo was destroyed in proportion to the amount of stretch, but the lengths of the remaining intact sarcomeres were essentially unchanged. This suggests that there had been a successive yielding of the weakeast sarcomeres. 7. The difference between the slow and fast rigor resistance and the effect of calcium on these resistances are discussed in relation to possible variations in the strength of crossbridges between the thick and thin filaments. Images Plate 1 Plate 2 PMID:1082023
Establishing Reliability and Validity of the Criterion Referenced Exam of GeoloGy Standards EGGS
NASA Astrophysics Data System (ADS)
Guffey, S. K.; Slater, S. J.; Slater, T. F.; Schleigh, S.; Burrows, A. C.
2016-12-01
Discipline-based geoscience education researchers have considerable need for a criterion-referenced, easy-to-administer and -score conceptual diagnostic survey for undergraduates taking introductory science survey courses in order for faculty to better be able to monitor the learning impacts of various interactive teaching approaches. To support ongoing education research across the geosciences, we are continuing to rigorously and systematically work to firmly establish the reliability and validity of the recently released Exam of GeoloGy Standards, EGGS. In educational testing, reliability refers to the consistency or stability of test scores whereas validity refers to the accuracy of the inferences or interpretations one makes from test scores. There are several types of reliability measures being applied to the iterative refinement of the EGGS survey, including test-retest, alternate form, split-half, internal consistency, and interrater reliability measures. EGGS rates strongly on most measures of reliability. For one, Cronbach's alpha provides a quantitative index indicating the extent to which if students are answering items consistently throughout the test and measures inter-item correlations. Traditional item analysis methods further establish the degree to which a particular item is reliably assessing students is actually quantifiable, including item difficulty and item discrimination. Validity, on the other hand, is perhaps best described by the word accuracy. For example, content validity is the to extent to which a measurement reflects the specific intended domain of the content, stemming from judgments of people who are either experts in the testing of that particular content area or are content experts. Perhaps more importantly, face validity is a judgement of how representative an instrument is reflective of the science "at face value" and refers to the extent to which a test appears to measure a the targeted scientific domain as viewed by laypersons, examinees, test users, the public, and other invested stakeholders.
NASA Astrophysics Data System (ADS)
Dufaux, Frederic
2011-06-01
The issue of privacy in video surveillance has drawn a lot of interest lately. However, thorough performance analysis and validation is still lacking, especially regarding the fulfillment of privacy-related requirements. In this paper, we first review recent Privacy Enabling Technologies (PET). Next, we discuss pertinent evaluation criteria for effective privacy protection. We then put forward a framework to assess the capacity of PET solutions to hide distinguishing facial information and to conceal identity. We conduct comprehensive and rigorous experiments to evaluate the performance of face recognition algorithms applied to images altered by PET. Results show the ineffectiveness of naïve PET such as pixelization and blur. Conversely, they demonstrate the effectiveness of more sophisticated scrambling techniques to foil face recognition.
Cognitive representations of AIDS: a phenomenological study.
Anderson, Elizabeth H; Spencer, Margaret Hull
2002-12-01
Cognitive representations of illness determine behavior. How persons living with AIDS image their disease might be key to understanding medication adherence and other health behaviors. The authors' purpose was to describe AIDS patients' cognitive representations of their illness. A purposive sample of 58 men and women with AIDS were interviewed. Using Colaizzi's (1978) phenomenological method, rigor was established through application of verification, validation, and validity. From 175 significant statements, 11 themes emerged. Cognitive representations included imaging AIDS as death, bodily destruction, and just a disease. Coping focused on wiping AIDS out of the mind, hoping for the right drug, and caring for oneself. Inquiring about a patient's image of AIDS might help nurses assess coping processes and enhance nurse-patient relationships.
Dynamics of a Chlorophyll Dimer in Collective and Local Thermal Environments
Merkli, M.; Berman, Gennady Petrovich; Sayre, Richard Thomas; ...
2016-01-30
Here we present a theoretical analysis of exciton transfer and decoherence effects in a photosynthetic dimer interacting with collective (correlated) and local (uncorrelated) protein-solvent environments. Our approach is based on the framework of the spin-boson model. We derive explicitly the thermal relaxation and decoherence rates of the exciton transfer process, valid for arbitrary temperatures and for arbitrary (in particular, large) interaction constants between the dimer and the environments. We establish a generalization of the Marcus formula, giving reaction rates for dimer levels possibly individually and asymmetrically coupled to environments. We identify rigorously parameter regimes for the validity of the generalizedmore » Marcus formula. The existence of long living quantum coherences at ambient temperatures emerges naturally from our approach.« less
Measurement uncertainty analysis techniques applied to PV performance measurements
NASA Astrophysics Data System (ADS)
Wells, C.
1992-10-01
The purpose of this presentation is to provide a brief introduction to measurement uncertainty analysis, outline how it is done, and illustrate uncertainty analysis with examples drawn from the PV field, with particular emphasis toward its use in PV performance measurements. The uncertainty information we know and state concerning a PV performance measurement or a module test result determines, to a significant extent, the value and quality of that result. What is measurement uncertainty analysis? It is an outgrowth of what has commonly been called error analysis. But uncertainty analysis, a more recent development, gives greater insight into measurement processes and tests, experiments, or calibration results. Uncertainty analysis gives us an estimate of the interval about a measured value or an experiment's final result within which we believe the true value of that quantity will lie. Why should we take the time to perform an uncertainty analysis? A rigorous measurement uncertainty analysis: Increases the credibility and value of research results; allows comparisons of results from different labs; helps improve experiment design and identifies where changes are needed to achieve stated objectives (through use of the pre-test analysis); plays a significant role in validating measurements and experimental results, and in demonstrating (through the post-test analysis) that valid data have been acquired; reduces the risk of making erroneous decisions; demonstrates quality assurance and quality control measures have been accomplished; define Valid Data as data having known and documented paths of: Origin, including theory; measurements; traceability to measurement standards; computations; uncertainty analysis of results.
Psychometric Assessment of the Mindful Attention Awareness Scale (MAAS) Among Chinese Adolescents
Black, David S.; Sussman, Steve; Johnson, C. Anderson; Milam, Joel
2013-01-01
The Mindful Attention Awareness Scale (MAAS) has the longest empirical track record as a valid measure of trait mindfulness. Most of what is understood about trait mindfulness comes from administering the MAAS to relatively homogenous samples of Caucasian adults. This study rigorously evaluates the psychometric properties of the MAAS among Chinese adolescents attending high school in Chengdu, China. Classrooms from 24 schools were randomly selected to participate in the study. Three waves of longitudinal data (N = 5,287 students) were analyzed. MAAS construct, nomological, and incremental validity were evaluated as well as its measurement invariance across gender using latent factor analyses. Participants’ mean age was 16.2 years (SD = 0.7), and 51% were male. The 15-item MAAS had adequate fit to the one-dimensional factor structure at Wave 1, and this factor structure was replicated at Wave 2. A 6-item short scale of the MAAS fit well to the data at Wave 3. The MAAS maintained reliability (Cronbach’s α = .89–.93; test–restest r = .35–.52), convergent/discriminant validity, and explained additional variance in mental health measures beyond other psychosocial constructs. Both the 15- and 6-item MAAS scales displayed at least partial factorial invariance across gender. The findings suggest that the MAAS is a sound measure of trait mindfulness among Chinese adolescents. To reduce respondent burden, the MAAS 6-item short-scale provides an option to measure trait mindfulness. PMID:21816857
Harris, Sarah Parker; Gould, Robert; Fujiura, Glenn
2015-01-01
There is increasing theoretical consideration about the use of systematic and scoping reviews of evidence in informing disability and rehabilitation research and practice. Indicative of this trend, this journal published a piece by Rumrill, Fitzgerald and Merchant in 2010 explaining the utility and process for conducting reviews of intervention-based research. There is still need to consider how to apply such rigor when conducting more exploratory reviews of heterogeneous research. This article explores the challenges, benefits, and procedures for conducting rigorous exploratory scoping reviews of diverse evidence. The article expands upon Rumrill, Fitzgerald and Merchant's framework and considers its application to more heterogeneous evidence on the impact of social policy. A worked example of a scoping review of the Americans with Disabilities Act is provided with a procedural framework for conducting scoping reviews on the effects of a social policy. The need for more nuanced techniques for enhancing rigor became apparent during the review process. There are multiple methodological steps that can enhance the utility of exploratory scoping reviews. The potential of systematic consideration during the exploratory review process is shown as a viable method to enhance the rigor in reviewing diverse bodies of evidence.
Loescher, Lois J; Rains, Stephen A; Kramer, Sandra S; Akers, Chelsie; Moussa, Renee
2018-05-01
To systematically review healthy lifestyle interventions targeted to adolescents and delivered using text messaging (TM). PubMed, Embase, CINAHL, PsycINFO, and Web of Science databases. Study Inclusion Criteria: Research articles published during 2011 to 2014; analyses focused on intervention targeting adolescents (10-19 years), with healthy lifestyle behaviors as main variables, delivered via mobile phone-based TM. The authors extracted data from 27 of 281 articles using the Preferred Reporting Items for Systematic Reviews and Meta-Analyses method. Adolescent and setting characteristics, study design and rigor, intervention effectiveness, challenges, and risk of bias. Across studies, 16 (59.3%) of 27 included non-Caucasians. The gender was split for 22 (81.5%) of 27 studies. Thirteen studies were randomized controlled trials. There was heterogeneity among targeted conditions, rigor of methods, and intervention effects. Interventions for monitoring/adherence (n = 8) reported more positive results than those for health behavior change (n = 19). Studies that only included message delivered via TM (n = 14) reported more positive effects than studies integrating multiple intervention components. Interventions delivered using TM presented minimal challenges, but selection and performance bias were observed across studies. Interventions delivered using TM have the potential, under certain conditions, to improve healthy lifestyle behaviors in adolescents. However, the rigor of studies varies, and established theory and validated measures have been inconsistently incorporated.
Multi-template polymerase chain reaction.
Kalle, Elena; Kubista, Mikael; Rensing, Christopher
2014-12-01
PCR is a formidable and potent technology that serves as an indispensable tool in a wide range of biological disciplines. However, due to the ease of use and often lack of rigorous standards many PCR applications can lead to highly variable, inaccurate, and ultimately meaningless results. Thus, rigorous method validation must precede its broad adoption to any new application. Multi-template samples possess particular features, which make their PCR analysis prone to artifacts and biases: multiple homologous templates present in copy numbers that vary within several orders of magnitude. Such conditions are a breeding ground for chimeras and heteroduplexes. Differences in template amplification efficiencies and template competition for reaction compounds undermine correct preservation of the original template ratio. In addition, the presence of inhibitors aggravates all of the above-mentioned problems. Inhibitors might also have ambivalent effects on the different templates within the same sample. Yet, no standard approaches exist for monitoring inhibitory effects in multitemplate PCR, which is crucial for establishing compatibility between samples.
NASA Technical Reports Server (NTRS)
Glover, R. M.; Weinhold, F.
1977-01-01
Variational functionals of Braunn and Rebane (1972) for the imagery-frequency polarizability (IFP) have been generalized by the method of Gramian inequalities to give rigorous upper and lower bounds, valid even when the true (but unknown) unperturbed wavefunction must be represented by a variational approximation. Using these formulas in conjunction with flexible variational trial functions, tight error bounds are computed for the IFP and the associated two- and three-body van der Waals interaction constants of the ground 1(1S) and metastable 2(1,3S) states of He and Li(+). These bounds generally establish the ground-state properties to within a fraction of a per cent and metastable properties to within a few per cent, permitting a comparative assessment of competing theoretical methods at this level of accuracy. Unlike previous 'error bounds' for these properties, the present results have a completely a priori theoretical character, with no empirical input data.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Bhardwaj, Shubhendu; Sensale-Rodriguez, Berardi; Xing, Huili Grace
A rigorous theoretical and computational model is developed for the plasma-wave propagation in high electron mobility transistor structures with electron injection from a resonant tunneling diode at the gate. We discuss the conditions in which low-loss and sustainable plasmon modes can be supported in such structures. The developed analytical model is used to derive the dispersion relation for these plasmon-modes. A non-linear full-wave-hydrodynamic numerical solver is also developed using a finite difference time domain algorithm. The developed analytical solutions are validated via the numerical solution. We also verify previous observations that were based on a simplified transmission line model. Itmore » is shown that at high levels of negative differential conductance, plasmon amplification is indeed possible. The proposed rigorous models can enable accurate design and optimization of practical resonant tunnel diode-based plasma-wave devices for terahertz sources, mixers, and detectors, by allowing a precise representation of their coupling when integrated with other electromagnetic structures.« less
A protocol for validating Land Surface Temperature from Sentinel-3
NASA Astrophysics Data System (ADS)
Ghent, D.
2015-12-01
One of the main objectives of the Sentinel-3 mission is to measure sea- and land-surface temperature with high-end accuracy and reliability in support of environmental and climate monitoring in an operational context. Calibration and validation are thus key criteria for operationalization within the framework of the Sentinel-3 Mission Performance Centre (S3MPC).Land surface temperature (LST) has a long heritage of satellite observations which have facilitated our understanding of land surface and climate change processes, such as desertification, urbanization, deforestation and land/atmosphere coupling. These observations have been acquired from a variety of satellite instruments on platforms in both low-earth orbit and in geostationary orbit. Retrieval accuracy can be a challenge though; surface emissivities can be highly variable owing to the heterogeneity of the land, and atmospheric effects caused by the presence of aerosols and by water vapour absorption can give a bias to the underlying LST. As such, a rigorous validation is critical in order to assess the quality of the data and the associated uncertainties. The Sentinel-3 Cal-Val Plan for evaluating the level-2 SL_2_LST product builds on an established validation protocol for satellite-based LST. This set of guidelines provides a standardized framework for structuring LST validation activities, and is rapidly gaining international recognition. The protocol introduces a four-pronged approach which can be summarised thus: i) in situ validation where ground-based observations are available; ii) radiance-based validation over sites that are homogeneous in emissivity; iii) intercomparison with retrievals from other satellite sensors; iv) time-series analysis to identify artefacts on an interannual time-scale. This multi-dimensional approach is a necessary requirement for assessing the performance of the LST algorithm for SLSTR which is designed around biome-based coefficients, thus emphasizing the importance of non-traditional forms of validation such as radiance-based techniques. Here we present examples of the application of the protocol to data produced within the ESA DUE GlobTemperature Project. The lessons learnt here are helping to fine-tune the methodology in preparation for Sentinel-3 commissioning.
Interaction of Theory and Practice to Assess External Validity.
Leviton, Laura C; Trujillo, Mathew D
2016-01-18
Variations in local context bedevil the assessment of external validity: the ability to generalize about effects of treatments. For evaluation, the challenges of assessing external validity are intimately tied to the translation and spread of evidence-based interventions. This makes external validity a question for decision makers, who need to determine whether to endorse, fund, or adopt interventions that were found to be effective and how to ensure high quality once they spread. To present the rationale for using theory to assess external validity and the value of more systematic interaction of theory and practice. We review advances in external validity, program theory, practitioner expertise, and local adaptation. Examples are provided for program theory, its adaptation to diverse contexts, and generalizing to contexts that have not yet been studied. The often critical role of practitioner experience is illustrated in these examples. Work is described that the Robert Wood Johnson Foundation is supporting to study treatment variation and context more systematically. Researchers and developers generally see a limited range of contexts in which the intervention is implemented. Individual practitioners see a different and often a wider range of contexts, albeit not a systematic sample. Organized and taken together, however, practitioner experiences can inform external validity by challenging the developers and researchers to consider a wider range of contexts. Researchers have developed a variety of ways to adapt interventions in light of such challenges. In systematic programs of inquiry, as opposed to individual studies, the problems of context can be better addressed. Evaluators have advocated an interaction of theory and practice for many years, but the process can be made more systematic and useful. Systematic interaction can set priorities for assessment of external validity by examining the prevalence and importance of context features and treatment variations. Practitioner interaction with researchers and developers can assist in sharpening program theory, reducing uncertainty about treatment variations that are consistent or inconsistent with the theory, inductively ruling out the ones that are harmful or irrelevant, and helping set priorities for more rigorous study of context and treatment variation. © The Author(s) 2016.
2013-09-30
data from the IABP ); 2.) Forecasting weather and sea ice conditions; 3.) Forcing, assimilation and validation of global weather and climate models ...International Arctic Buoy Programme ( IABP ) A US Interagency Arctic Buoy Programme (USIABP) contribution to the IABP Dr. Ignatius G. Rigor Polar...ice motion. These observations are assimilated into Numerical Weather Prediction (NWP) models that are used to forecast weather on synoptic time
Advanced topics in evidence-based urologic oncology: surrogate endpoints.
Lavallée, Luke T; Montori, Victor M; Canfield, Stephen E; Breau, Rodney H
2011-01-01
Clinical trials often report surrogate endpoint data. A surrogate endpoint is a biological marker or clinical sign that can be substituted for a patient-important outcome. Using surrogate endpoints correctly may facilitate and expedite clinical trials and may improve medical decisions. However, rigorous criteria must be met for an endpoint to be considered a valid surrogate. The purpose of this article is to review the topic of surrogate endpoints in the context of a urologic encounter. Copyright © 2011 Elsevier Inc. All rights reserved.
[When should a patient with musculoskeletal trauma be referred to emergency ward?].
Feiner, Adam-Scott; Duruz, Henri
2010-08-25
Standardized clinical examination can obviate the need for osteoarticular radiographs for trauma. This paper summarizes a number of decision rules that allow clinical exclusion of significant fracture of the cervical spine, elbow, knee or ankle, making radiographs unnecessary. These criteria were all derived from large cohort studies (Nexus, Ottawa, CCS, etc..., and have been prospectively validated. The rigorous use of these criteria in daily practice improves treatment times and costs with no adverse effect on treatment quality.
Development of a Thermal Wave Interferometry System for Thin-Film Characterisation
2012-10-01
describes a condition where the properties of the interface influence the phase and amplitude of the temperature oscillations at the surface. In the...The measured phase profiles are shown in Figure 13. Overall, the phase variation across the sample is significant, with a strong discontinuity in phase ...d eg ) 91.2 Hz 30 Hz 7 Hz 4 Hz Figure 13: Phase measurements across the coated sample. L1 L2 L3 L4 L5 A more rigorous validation of the
Kawakami, Takao; Nagasaka, Keiko; Takami, Sachiko; Wada, Kazuya; Tu, Hsiao-Kun; Otsuji, Makiko; Kyono, Yutaka; Dobashi, Tae; Komatsu, Yasuhiko; Kihara, Makoto; Akimoto, Shingo; Peers, Ian S.; South, Marie C.; Higenbottam, Tim; Fukuoka, Masahiro; Nakata, Koichiro; Ohe, Yuichiro; Kudoh, Shoji; Clausen, Ib Groth; Nishimura, Toshihide; Marko-Varga, György; Kato, Harubumi
2011-01-01
Interstitial lung disease (ILD) events have been reported in Japanese non-small-cell lung cancer (NSCLC) patients receiving EGFR tyrosine kinase inhibitors. We investigated proteomic biomarkers for mechanistic insights and improved prediction of ILD. Blood plasma was collected from 43 gefitinib-treated NSCLC patients developing acute ILD (confirmed by blinded diagnostic review) and 123 randomly selected controls in a nested case-control study within a pharmacoepidemiological cohort study in Japan. We generated ∼7 million tandem mass spectrometry (MS/MS) measurements with extensive quality control and validation, producing one of the largest proteomic lung cancer datasets to date, incorporating rigorous study design, phenotype definition, and evaluation of sample processing. After alignment, scaling, and measurement batch adjustment, we identified 41 peptide peaks representing 29 proteins best predicting ILD. Multivariate peptide, protein, and pathway modeling achieved ILD prediction comparable to previously identified clinical variables; combining the two provided some improvement. The acute phase response pathway was strongly represented (17 of 29 proteins, p = 1.0×10−25), suggesting a key role with potential utility as a marker for increased risk of acute ILD events. Validation by Western blotting showed correlation for identified proteins, confirming that robust results can be generated from an MS/MS platform implementing strict quality control. PMID:21799770
Bor, Jacob; Geldsetzer, Pascal; Venkataramani, Atheendar; Bärnighausen, Till
2015-01-01
Purpose of review Randomized, population-representative trials of clinical interventions are rare. Quasi-experiments have been used successfully to generate causal evidence on the cascade of HIV care in a broad range of real-world settings. Recent findings Quasi-experiments exploit exogenous, or quasi-random, variation occurring naturally in the world or because of an administrative rule or policy change to estimate causal effects. Well designed quasi-experiments have greater internal validity than typical observational research designs. At the same time, quasi-experiments may also have potential for greater external validity than experiments and can be implemented when randomized clinical trials are infeasible or unethical. Quasi-experimental studies have established the causal effects of HIV testing and initiation of antiretroviral therapy on health, economic outcomes and sexual behaviors, as well as indirect effects on other community members. Recent quasi-experiments have evaluated specific interventions to improve patient performance in the cascade of care, providing causal evidence to optimize clinical management of HIV. Summary Quasi-experiments have generated important data on the real-world impacts of HIV testing and treatment and on interventions to improve the cascade of care. With the growth in large-scale clinical and administrative data, quasi-experiments enable rigorous evaluation of policies implemented in real-world settings. PMID:26371463
Bor, Jacob; Geldsetzer, Pascal; Venkataramani, Atheendar; Bärnighausen, Till
2015-11-01
Randomized, population-representative trials of clinical interventions are rare. Quasi-experiments have been used successfully to generate causal evidence on the cascade of HIV care in a broad range of real-world settings. Quasi-experiments exploit exogenous, or quasi-random, variation occurring naturally in the world or because of an administrative rule or policy change to estimate causal effects. Well designed quasi-experiments have greater internal validity than typical observational research designs. At the same time, quasi-experiments may also have potential for greater external validity than experiments and can be implemented when randomized clinical trials are infeasible or unethical. Quasi-experimental studies have established the causal effects of HIV testing and initiation of antiretroviral therapy on health, economic outcomes and sexual behaviors, as well as indirect effects on other community members. Recent quasi-experiments have evaluated specific interventions to improve patient performance in the cascade of care, providing causal evidence to optimize clinical management of HIV. Quasi-experiments have generated important data on the real-world impacts of HIV testing and treatment and on interventions to improve the cascade of care. With the growth in large-scale clinical and administrative data, quasi-experiments enable rigorous evaluation of policies implemented in real-world settings.
Adapting clinical practice guidelines for diabetic retinopathy in Kenya: process and outputs.
Mwangi, Nyawira; Gachago, Muchai; Gichangi, Michael; Gichuhi, Stephen; Githeko, Kibata; Jalango, Atieno; Karimurio, Jefitha; Kibachio, Joseph; Muthami, Lawrence; Ngugi, Nancy; Nduri, Carmichael; Nyaga, Patrick; Nyamori, Joseph; Zindamoyen, Alain Nazaire Mbongo; Bascaran, Covadonga; Foster, Allen
2018-06-15
The use of clinical practice guidelines envisages augmenting quality and best practice in clinical outcomes. Generic guidelines that are not adapted for local use often fail to produce these outcomes. Adaptation is a systematic and rigorous process that should maintain the quality and validity of the guideline, while making it more usable by the targeted users. Diverse skills are required for the task of adaptation. Although adapting a guideline is not a guarantee that it will be implemented, adaptation may improve acceptance and adherence to its recommendations. We describe the process used to adapt clinical guidelines for diabetic retinopathy in Kenya, using validated tools and manuals. A technical working group consisting of volunteers provided leadership. The process was intensive and required more time than anticipated. Flexibility in the process and concurrent health system activities contributed to the success of the adaptation. The outputs from the adaptation include the guidelines in different formats, point of care instruments, as well as tools for training, monitoring, quality assurance and patient education. Guideline adaptation is applicable and feasible at the national level in Kenya. However, it is labor- and time -intensive. It presents a valuable opportunity to develop several additional outputs that are useful at the point of care.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Ascough, II, James Clifford
1992-05-01
The capability to objectively evaluate design performance of shallow landfill burial (SLB) systems is of great interest to diverse scientific disciplines, including hydrologists, engineers, environmental scientists, and SLB regulators. The goal of this work was to develop and validate a procedure for the nonsubjective evaluation of SLB designs under actual or simulated environmental conditions. A multiobjective decision module (MDM) based on scoring functions (Wymore, 1988) was implemented to evaluate SLB design performance. Input values to the MDM are provided by hydrologic models. The MDM assigns a total score to each SLB design alternative, thereby allowing for rapid and repeatable designmore » performance evaluation. The MDM was validated for a wide range of SLB designs under different climatic conditions. Rigorous assessment of SLB performance also requires incorporation of hydrologic probabilistic analysis and hydrologic risk into the overall design. This was accomplished through the development of a frequency analysis module. The frequency analysis module allows SLB design event magnitudes to be calculated based on the hydrologic return period. The multiobjective decision and freqeuncy anslysis modules were integrated in a decision support system (DSS) framework, SLEUTH (Shallow Landfill Evaluation Using Transport and Hydrology). SLEUTH is a Microsoft Windows {trademark} application, and is written in the Knowledge Pro Windows (Knowledge Garden, Inc., 1991) development language.« less
A Robust and Effective Smart-Card-Based Remote User Authentication Mechanism Using Hash Function
Odelu, Vanga; Goswami, Adrijit
2014-01-01
In a remote user authentication scheme, a remote server verifies whether a login user is genuine and trustworthy, and also for mutual authentication purpose a login user validates whether the remote server is genuine and trustworthy. Several remote user authentication schemes using the password, the biometrics, and the smart card have been proposed in the literature. However, most schemes proposed in the literature are either computationally expensive or insecure against several known attacks. In this paper, we aim to propose a new robust and effective password-based remote user authentication scheme using smart card. Our scheme is efficient, because our scheme uses only efficient one-way hash function and bitwise XOR operations. Through the rigorous informal and formal security analysis, we show that our scheme is secure against possible known attacks. We perform the simulation for the formal security analysis using the widely accepted AVISPA (Automated Validation Internet Security Protocols and Applications) tool to ensure that our scheme is secure against passive and active attacks. Furthermore, our scheme supports efficiently the password change phase always locally without contacting the remote server and correctly. In addition, our scheme performs significantly better than other existing schemes in terms of communication, computational overheads, security, and features provided by our scheme. PMID:24892078
A robust and effective smart-card-based remote user authentication mechanism using hash function.
Das, Ashok Kumar; Odelu, Vanga; Goswami, Adrijit
2014-01-01
In a remote user authentication scheme, a remote server verifies whether a login user is genuine and trustworthy, and also for mutual authentication purpose a login user validates whether the remote server is genuine and trustworthy. Several remote user authentication schemes using the password, the biometrics, and the smart card have been proposed in the literature. However, most schemes proposed in the literature are either computationally expensive or insecure against several known attacks. In this paper, we aim to propose a new robust and effective password-based remote user authentication scheme using smart card. Our scheme is efficient, because our scheme uses only efficient one-way hash function and bitwise XOR operations. Through the rigorous informal and formal security analysis, we show that our scheme is secure against possible known attacks. We perform the simulation for the formal security analysis using the widely accepted AVISPA (Automated Validation Internet Security Protocols and Applications) tool to ensure that our scheme is secure against passive and active attacks. Furthermore, our scheme supports efficiently the password change phase always locally without contacting the remote server and correctly. In addition, our scheme performs significantly better than other existing schemes in terms of communication, computational overheads, security, and features provided by our scheme.
Low validity of Google Trends for behavioral forecasting of national suicide rates
Niederkrotenthaler, Thomas; Till, Benedikt; Ajdacic-Gross, Vladeta; Voracek, Martin
2017-01-01
Recent research suggests that search volumes of the most popular search engine worldwide, Google, provided via Google Trends, could be associated with national suicide rates in the USA, UK, and some Asian countries. However, search volumes have mostly been studied in an ad hoc fashion, without controls for spurious associations. This study evaluated the validity and utility of Google Trends search volumes for behavioral forecasting of suicide rates in the USA, Germany, Austria, and Switzerland. Suicide-related search terms were systematically collected and respective Google Trends search volumes evaluated for availability. Time spans covered 2004 to 2010 (USA, Switzerland) and 2004 to 2012 (Germany, Austria). Temporal associations of search volumes and suicide rates were investigated with time-series analyses that rigorously controlled for spurious associations. The number and reliability of analyzable search volume data increased with country size. Search volumes showed various temporal associations with suicide rates. However, associations differed both across and within countries and mostly followed no discernable patterns. The total number of significant associations roughly matched the number of expected Type I errors. These results suggest that the validity of Google Trends search volumes for behavioral forecasting of national suicide rates is low. The utility and validity of search volumes for the forecasting of suicide rates depend on two key assumptions (“the population that conducts searches consists mostly of individuals with suicidal ideation”, “suicide-related search behavior is strongly linked with suicidal behavior”). We discuss strands of evidence that these two assumptions are likely not met. Implications for future research with Google Trends in the context of suicide research are also discussed. PMID:28813490
Cross-cultural adaptation of instruments assessing breastfeeding determinants: a multi-step approach
2014-01-01
Background Cross-cultural adaptation is a necessary process to effectively use existing instruments in other cultural and language settings. The process of cross-culturally adapting, including translation, of existing instruments is considered a critical set to establishing a meaningful instrument for use in another setting. Using a multi-step approach is considered best practice in achieving cultural and semantic equivalence of the adapted version. We aimed to ensure the content validity of our instruments in the cultural context of KwaZulu-Natal, South Africa. Methods The Iowa Infant Feeding Attitudes Scale, Breastfeeding Self-Efficacy Scale-Short Form and additional items comprise our consolidated instrument, which was cross-culturally adapted utilizing a multi-step approach during August 2012. Cross-cultural adaptation was achieved through steps to maintain content validity and attain semantic equivalence in the target version. Specifically, Lynn’s recommendation to apply an item-level content validity index score was followed. The revised instrument was translated and back-translated. To ensure semantic equivalence, Brislin’s back-translation approach was utilized followed by the committee review to address any discrepancies that emerged from translation. Results Our consolidated instrument was adapted to be culturally relevant and translated to yield more reliable and valid results for use in our larger research study to measure infant feeding determinants effectively in our target cultural context. Conclusions Undertaking rigorous steps to effectively ensure cross-cultural adaptation increases our confidence that the conclusions we make based on our self-report instrument(s) will be stronger. In this way, our aim to achieve strong cross-cultural adaptation of our consolidated instruments was achieved while also providing a clear framework for other researchers choosing to utilize existing instruments for work in other cultural, geographic and population settings. PMID:25285151
Recognizing that novel potential biomarkers are continually being identified and will need to be validated in a rapid, efficient, and scientifically rigorous manner, the NCI has made an enormous commitment to the development of a network that will facilitate biomarker development and validation in multiple organ sites. As part of the National Cancer Institute-funded Early Detection Research Network (EDRN), the Great Lakes-New England Clinical Epidemiological Center (GLNE CEC) proposes a research program that provides the structure for validating and discovering potential surrogate endpoint biomarkers (“biomarkers”). Although examples of such biomarkers are currently in clinical use (i.e. CEA, CA-125), there are limitations to all of them. Our consortium focuses specifically on gastrointestinal neoplasia. There are three goals for this phase of the proposed research. 1. Establish the feasibility of measuring the biomarkers in a multi-center clinical trial. 2. Estimate the variance of the biomarkers in cohorts defined by sex, race, age and histologic diagnosis (non-Barrett’s controls, Barrett’s intestinal metaplasia, Barrett’s intestinal dysplasia [low and high-grade] and adenocarcinoma). 3. Determine if the distributions of the biomarkers differ significantly among patients with different histologic diagnoses. In this protocol, biological samples will consist of serum, plasma, urine, and biopsies from Barrett’s esophagus (metaplasia, low and high-grade dysplasia) patients, from patients with esophageal adenocarcinoma, and from non-Barrett’s controls. Samples will be assayed for villin, p53, Hsp27, cyclooxygenase-2, and Cyclin D1. Samples will also be used for two biomarker discovery projects, one exploring genetic expression using genomic microarrays and a second using two-dimensional gene arrays to discover and characterize amplified proteins associated with esophageal carcinogenesis. Fifty subjects will
Evaluation of Metals Release from Oxidation of Fly Ash during Dredging of the Emory River, TN
2011-08-01
from an oil -free source (trickle flow, 2-5 bubbles per second) to provide some turbulent flow and to maintain dissolved oxygen levels. More rigorous...larval and (b) juvenile Pimephales promelas. ERDC/EL TR-11-9 79 five juvenile fish and was rigorously aerated from an oil -free source to...epithelial width. In contrast, juvenile pike from a reference lake had significantly thicker gill filaments compared to those exposed to Key Lake uranium
Nielsen, Joseph; Tokuhiro, Akira; Hiromoto, Robert; ...
2015-11-13
Evaluation of the impacts of uncertainty and sensitivity in modeling presents a significant set of challenges in particular to high fidelity modeling. Computational costs and validation of models creates a need for cost effective decision making with regards to experiment design. Experiments designed to validate computation models can be used to reduce uncertainty in the physical model. In some cases, large uncertainty in a particular aspect of the model may or may not have a large impact on the final results. For example, modeling of a relief valve may result in large uncertainty, however, the actual effects on final peakmore » clad temperature in a reactor transient may be small and the large uncertainty with respect to valve modeling may be considered acceptable. Additionally, the ability to determine the adequacy of a model and the validation supporting it should be considered within a risk informed framework. Low fidelity modeling with large uncertainty may be considered adequate if the uncertainty is considered acceptable with respect to risk. In other words, models that are used to evaluate the probability of failure should be evaluated more rigorously with the intent of increasing safety margin. Probabilistic risk assessment (PRA) techniques have traditionally been used to identify accident conditions and transients. Traditional classical event tree methods utilize analysts’ knowledge and experience to identify the important timing of events in coordination with thermal-hydraulic modeling. These methods lack the capability to evaluate complex dynamic systems. In these systems, time and energy scales associated with transient events may vary as a function of transition times and energies to arrive at a different physical state. Dynamic PRA (DPRA) methods provide a more rigorous analysis of complex dynamic systems. Unfortunately DPRA methods introduce issues associated with combinatorial explosion of states. This study presents a methodology to address combinatorial explosion using a Branch-and-Bound algorithm applied to Dynamic Event Trees (DET), which utilize LENDIT (L – Length, E – Energy, N – Number, D – Distribution, I – Information, and T – Time) as well as a set theory to describe system, state, resource, and response (S2R2) sets to create bounding functions for the DET. The optimization of the DET in identifying high probability failure branches is extended to create a Phenomenological Identification and Ranking Table (PIRT) methodology to evaluate modeling parameters important to safety of those failure branches that have a high probability of failure. The PIRT can then be used as a tool to identify and evaluate the need for experimental validation of models that have the potential to reduce risk. Finally, in order to demonstrate this methodology, a Boiling Water Reactor (BWR) Station Blackout (SBO) case study is presented.« less
Pueyo Bellafont, Noèlia; Bagus, Paul S; Illas, Francesc
2015-06-07
A systematic study of the N(1s) core level binding energies (BE's) in a broad series of molecules is presented employing Hartree-Fock (HF) and the B3LYP, PBE0, and LC-BPBE density functional theory (DFT) based methods with a near HF basis set. The results show that all these methods give reasonably accurate BE's with B3LYP being slightly better than HF but with both PBE0 and LCBPBE being poorer than HF. A rigorous and general decomposition of core level binding energy values into initial and final state contributions to the BE's is proposed that can be used within either HF or DFT methods. The results show that Koopmans' theorem does not hold for the Kohn-Sham eigenvalues. Consequently, Kohn-Sham orbital energies of core orbitals do not provide estimates of the initial state contribution to core level BE's; hence, they cannot be used to decompose initial and final state contributions to BE's. However, when the initial state contribution to DFT BE's is properly defined, the decompositions of initial and final state contributions given by DFT, with several different functionals, are very similar to those obtained with HF. Furthermore, it is shown that the differences of Kohn-Sham orbital energies taken with respect to a common reference do follow the trend of the properly calculated initial state contributions. These conclusions are especially important for condensed phase systems where our results validate the use of band structure calculations to determine initial state contributions to BE shifts.
Real, Jordi; Forné, Carles; Roso-Llorach, Albert; Martínez-Sánchez, Jose M
2016-05-01
Controlling for confounders is a crucial step in analytical observational studies, and multivariable models are widely used as statistical adjustment techniques. However, the validation of the assumptions of the multivariable regression models (MRMs) should be made clear in scientific reporting. The objective of this study is to review the quality of statistical reporting of the most commonly used MRMs (logistic, linear, and Cox regression) that were applied in analytical observational studies published between 2003 and 2014 by journals indexed in MEDLINE.Review of a representative sample of articles indexed in MEDLINE (n = 428) with observational design and use of MRMs (logistic, linear, and Cox regression). We assessed the quality of reporting about: model assumptions and goodness-of-fit, interactions, sensitivity analysis, crude and adjusted effect estimate, and specification of more than 1 adjusted model.The tests of underlying assumptions or goodness-of-fit of the MRMs used were described in 26.2% (95% CI: 22.0-30.3) of the articles and 18.5% (95% CI: 14.8-22.1) reported the interaction analysis. Reporting of all items assessed was higher in articles published in journals with a higher impact factor.A low percentage of articles indexed in MEDLINE that used multivariable techniques provided information demonstrating rigorous application of the model selected as an adjustment method. Given the importance of these methods to the final results and conclusions of observational studies, greater rigor is required in reporting the use of MRMs in the scientific literature.
Woodruff, Tracey J; Sutton, Patrice
2014-10-01
Synthesizing what is known about the environmental drivers of health is instrumental to taking prevention-oriented action. Methods of research synthesis commonly used in environmental health lag behind systematic review methods developed in the clinical sciences over the past 20 years. We sought to develop a proof of concept of the "Navigation Guide," a systematic and transparent method of research synthesis in environmental health. The Navigation Guide methodology builds on best practices in research synthesis in evidence-based medicine and environmental health. Key points of departure from current methods of expert-based narrative review prevalent in environmental health include a prespecified protocol, standardized and transparent documentation including expert judgment, a comprehensive search strategy, assessment of "risk of bias," and separation of the science from values and preferences. Key points of departure from evidence-based medicine include assigning a "moderate" quality rating to human observational studies and combining diverse evidence streams. The Navigation Guide methodology is a systematic and rigorous approach to research synthesis that has been developed to reduce bias and maximize transparency in the evaluation of environmental health information. Although novel aspects of the method will require further development and validation, our findings demonstrated that improved methods of research synthesis under development at the National Toxicology Program and under consideration by the U.S. Environmental Protection Agency are fully achievable. The institutionalization of robust methods of systematic and transparent review would provide a concrete mechanism for linking science to timely action to prevent harm.
[Comparative analysis of quality labels of health websites].
Padilla-Garrido, N; Aguado-Correa, F; Huelva-López, L; Ortega-Moreno, M
2016-01-01
The search for health related information on the Internet is a growing phenomenon, buts its main drawback is the lack of reliability of information consulted. The aim of this study was to analyse and compare existing quality labels of health websites. A cross-sectional study was performed by searching Medline, IBECS, Google, and Yahoo, in both English and Spanish, between 8 and 9 March, 2015. Different keywords were used depending on whether the search was conducted in medical databases or generic search engines. The quality labels were classified according to their origin, analysing their character, year of implementation, the existence of the accreditation process, number of categories, criteria and standards, possibility of self-assessment, number of levels of certification, certification scope, validity, analytical quality of content, fee, results of the accreditation process, application and number of websites granted the seal, and quality labels obtained by the accrediting organisation. Seven quality labels, five of Spanish origin (WMA, PAWS, WIS, SEAFORMEC and M21) and two international ones (HONcode and Health Web Site Accreditation), were analysed. There was disparity in carrying out the accreditation process, with some not detailing key aspects of the process, or providing incomplete, outdated, or even inaccurate information. The most rigorous guaranteed the level of confidence that the websites had in relation to the content of information, but none checked the quality of them. Although rigorous quality labels may become useful, the deficiencies in some of them cast doubt on their current usefulness. Copyright © 2015 SECA. Publicado por Elsevier España, S.L.U. All rights reserved.
Staff Acceptance of Tele-ICU Coverage
Chan, Paul S.; Cram, Peter
2011-01-01
Background: Remote coverage of ICUs is increasing, but staff acceptance of this new technology is incompletely characterized. We conducted a systematic review to summarize existing research on acceptance of tele-ICU coverage among ICU staff. Methods: We searched for published articles pertaining to critical care telemedicine systems (aka, tele-ICU) between January 1950 and March 2010 using PubMed, Cumulative Index to Nursing and Allied Health Literature, Global Health, Web of Science, and the Cochrane Library and abstracts and presentations delivered at national conferences. Studies were included if they provided original qualitative or quantitative data on staff perceptions of tele-ICU coverage. Studies were imported into content analysis software and coded by tele-ICU configuration, methodology, participants, and findings (eg, positive and negative staff evaluations). Results: Review of 3,086 citations yielded 23 eligible studies. Findings were grouped into four categories of staff evaluation: overall acceptance level of tele-ICU coverage (measured in 70% of studies), impact on patient care (measured in 96%), impact on staff (measured in 100%), and organizational impact (measured in 48%). Overall acceptance was high, despite initial ambivalence. Favorable impact on patient care was perceived by > 82% of participants. Staff impact referenced enhanced collaboration, autonomy, and training, although scrutiny, malfunctions, and contradictory advice were cited as potential barriers. Staff perceived the organizational impact to vary. An important limitation of available studies was a lack of rigorous methodology and validated survey instruments in many studies. Conclusions: Initial reports suggest high levels of staff acceptance of tele-ICU coverage, but more rigorous methodologic study is required. PMID:21051386
A Theoretical Framework for Lagrangian Descriptors
NASA Astrophysics Data System (ADS)
Lopesino, C.; Balibrea-Iniesta, F.; García-Garrido, V. J.; Wiggins, S.; Mancho, A. M.
This paper provides a theoretical background for Lagrangian Descriptors (LDs). The goal of achieving rigorous proofs that justify the ability of LDs to detect invariant manifolds is simplified by introducing an alternative definition for LDs. The definition is stated for n-dimensional systems with general time dependence, however we rigorously prove that this method reveals the stable and unstable manifolds of hyperbolic points in four particular 2D cases: a hyperbolic saddle point for linear autonomous systems, a hyperbolic saddle point for nonlinear autonomous systems, a hyperbolic saddle point for linear nonautonomous systems and a hyperbolic saddle point for nonlinear nonautonomous systems. We also discuss further rigorous results which show the ability of LDs to highlight additional invariants sets, such as n-tori. These results are just a simple extension of the ergodic partition theory which we illustrate by applying this methodology to well-known examples, such as the planar field of the harmonic oscillator and the 3D ABC flow. Finally, we provide a thorough discussion on the requirement of the objectivity (frame-invariance) property for tools designed to reveal phase space structures and their implications for Lagrangian descriptors.
NASA Astrophysics Data System (ADS)
Farrell, Kathryn; Oden, J. Tinsley
2014-07-01
Coarse-grained models of atomic systems, created by aggregating groups of atoms into molecules to reduce the number of degrees of freedom, have been used for decades in important scientific and technological applications. In recent years, interest in developing a more rigorous theory for coarse graining and in assessing the predictivity of coarse-grained models has arisen. In this work, Bayesian methods for the calibration and validation of coarse-grained models of atomistic systems in thermodynamic equilibrium are developed. For specificity, only configurational models of systems in canonical ensembles are considered. Among major challenges in validating coarse-grained models are (1) the development of validation processes that lead to information essential in establishing confidence in the model's ability predict key quantities of interest and (2), above all, the determination of the coarse-grained model itself; that is, the characterization of the molecular architecture, the choice of interaction potentials and thus parameters, which best fit available data. The all-atom model is treated as the "ground truth," and it provides the basis with respect to which properties of the coarse-grained model are compared. This base all-atom model is characterized by an appropriate statistical mechanics framework in this work by canonical ensembles involving only configurational energies. The all-atom model thus supplies data for Bayesian calibration and validation methods for the molecular model. To address the first challenge, we develop priors based on the maximum entropy principle and likelihood functions based on Gaussian approximations of the uncertainties in the parameter-to-observation error. To address challenge (2), we introduce the notion of model plausibilities as a means for model selection. This methodology provides a powerful approach toward constructing coarse-grained models which are most plausible for given all-atom data. We demonstrate the theory and methods through applications to representative atomic structures and we discuss extensions to the validation process for molecular models of polymer structures encountered in certain semiconductor nanomanufacturing processes. The powerful method of model plausibility as a means for selecting interaction potentials for coarse-grained models is discussed in connection with a coarse-grained hexane molecule. Discussions of how all-atom information is used to construct priors are contained in an appendix.
The MIXED framework: A novel approach to evaluating mixed-methods rigor.
Eckhardt, Ann L; DeVon, Holli A
2017-10-01
Evaluation of rigor in mixed-methods (MM) research is a persistent challenge due to the combination of inconsistent philosophical paradigms, the use of multiple research methods which require different skill sets, and the need to combine research at different points in the research process. Researchers have proposed a variety of ways to thoroughly evaluate MM research, but each method fails to provide a framework that is useful for the consumer of research. In contrast, the MIXED framework is meant to bridge the gap between an academic exercise and practical assessment of a published work. The MIXED framework (methods, inference, expertise, evaluation, and design) borrows from previously published frameworks to create a useful tool for the evaluation of a published study. The MIXED framework uses an experimental eight-item scale that allows for comprehensive integrated assessment of MM rigor in published manuscripts. Mixed methods are becoming increasingly prevalent in nursing and healthcare research requiring researchers and consumers to address issues unique to MM such as evaluation of rigor. © 2017 John Wiley & Sons Ltd.
The CUAHSI Water Data Center: Empowering scientists to discover, use, store, and share water data
NASA Astrophysics Data System (ADS)
Couch, A. L.; Hooper, R. P.; Arrigo, J. S.
2012-12-01
The proposed CUAHSI Water Data Center (WDC) will provide production-quality water data resources based upon the successful large-scale data services prototype developed by the CUAHSI Hydrologic Information System (HIS) project. The WDC, using the HIS technology, concentrates on providing time series data collected at fixed points or on moving platforms from sensors primarily (but not exclusively) in the medium of water. The WDC's missions include providing simple and effective data discovery tools useful to researchers in a variety of water-related disciplines, and providing simple and cost-effective data publication mechanisms for projects that do not desire to run their own data servers. The WDC's activities will include: 1. Rigorous curation of the water data catalog already assembled during the CUAHSI HIS project, to ensure accuracy of records and existence of declared sources. 2. Data backup and failover services for "at risk" data sources. 3. Creation and support for ubiquitously accessible data discovery and access, web-based search and smartphone applications. 4. Partnerships with researchers to extend the state of the art in water data use. 5. Partnerships with industry to create plug-and-play data publishing from sensors, and to create domain-specific tools. The WDC will serve as a knowledge resource for researchers of water-related issues, and will interface with other data centers to make their data more accessible to water researchers. The WDC will serve as a vehicle for addressing some of the grand challenges of accessing and using water data, including: a. Cross-domain data discovery: different scientific domains refer to the same kind of water data using different terminologies, making discovery of data difficult for researchers outside the data provider's domain. b. Cross-validation of data sources: much water data comes from sources lacking rigorous quality control procedures; such sources can be compared against others with rigorous quality control. The WDC enables this by making both kinds of sources available in the same search interface. c. Data provenance: the appropriateness of data for use in a specific model or analysis often depends upon the exact details of how data was gathered and processed. The WDC will aid this by curating standards for metadata that are as descriptive as practical of the collection procedures. "Plug and play" sensor interfaces will fill in metadata appropriate to each sensor without human intervention. d. Contextual search: discovering data based upon geological (e.g. aquifer) or geographic (e.g., location in a stream network) features external to metadata. e. Data-driven search: discovering data that exhibit quality factors that are not described by the metadata. The WDC will partner with researchers desiring contextual and data driven search, and make results available to all. Many major data providers (e.g. federal agencies) are not mandated to provide access to data other than those they collect. The HIS project assembled data from over 90 different sources, thus demonstrating the promise of this approach. Meeting the grand challenges listed above will greatly enhance scientists' ability to discover, interpret, access, and analyze water data from across domains and sources to test Earth system hypotheses.
Evaluation of tools used to measure calcium and/or dairy consumption in adults.
Magarey, Anthea; Baulderstone, Lauren; Yaxley, Alison; Markow, Kylie; Miller, Michelle
2015-05-01
To identify and critique tools for the assessment of Ca and/or dairy intake in adults, in order to ascertain the most accurate and reliable tools available. A systematic review of the literature was conducted using defined inclusion and exclusion criteria. Articles reporting on originally developed tools or testing the reliability or validity of existing tools that measure Ca and/or dairy intake in adults were included. Author-defined criteria for reporting reliability and validity properties were applied. Studies conducted in Western countries. Adults. Thirty papers, utilising thirty-six tools assessing intake of dairy, Ca or both, were identified. Reliability testing was conducted on only two dairy and five Ca tools, with results indicating that only one dairy and two Ca tools were reliable. Validity testing was conducted for all but four Ca-only tools. There was high reliance in validity testing on lower-order tests such as correlation and failure to differentiate between statistical and clinically meaningful differences. Results of the validity testing suggest one dairy and five Ca tools are valid. Thus one tool was considered both reliable and valid for the assessment of dairy intake and only two tools proved reliable and valid for the assessment of Ca intake. While several tools are reliable and valid, their application across adult populations is limited by the populations in which they were tested. These results indicate a need for tools that assess Ca and/or dairy intake in adults to be rigorously tested for reliability and validity.
2018-01-01
Artificial intelligence (AI) is projected to substantially influence clinical practice in the foreseeable future. However, despite the excitement around the technologies, it is yet rare to see examples of robust clinical validation of the technologies and, as a result, very few are currently in clinical use. A thorough, systematic validation of AI technologies using adequately designed clinical research studies before their integration into clinical practice is critical to ensure patient benefit and safety while avoiding any inadvertent harms. We would like to suggest several specific points regarding the role that peer-reviewed medical journals can play, in terms of study design, registration, and reporting, to help achieve proper and meaningful clinical validation of AI technologies designed to make medical diagnosis and prediction, focusing on the evaluation of diagnostic accuracy efficacy. Peer-reviewed medical journals can encourage investigators who wish to validate the performance of AI systems for medical diagnosis and prediction to pay closer attention to the factors listed in this article by emphasizing their importance. Thereby, peer-reviewed medical journals can ultimately facilitate translating the technological innovations into real-world practice while securing patient safety and benefit. PMID:29805337
HIV-Risk Index: Development and Validation of a Brief Risk Index for Hispanic Young People.
Ballester-Arnal, Rafael; Gil-Llario, María Dolores; Castro-Calvo, Jesús; Giménez-García, Cristina
2016-08-01
The prevalence of HIV risk behaviors among young people facilitates the spread of HIV, in particular regarding unsafe sex behavior, although this trend is different within this population. For this reason, identifying the riskier young population is required to prevent HIV infection. The main purpose of this study was to develop and validate a risk index to assess the different sexual HIV risk exposure among Hispanic Young people. For this purpose, 9861 Spanish young people were randomly distributed into two groups (derivation and validation group). According to the results, the factor analyses grouped the nine items of the HIV- risk index into two factors (factor 1, direct sexual risk indicators and factor 2, indirect sexual risk indicators) with an equal structure for men and women by a multi-group confirmatory factor analysis. The variance explained was 54.26 %. Moreover, the Cronbach's alpha coefficient revealed high internal reliability (α = .79) and the convergent validity supported its evidence based on different HIV risk indexes. Therefore, the HIV-risk index seem to be a rigorous and valid measure to estimate HIV risk exposure among young people.
Nguyen, Van N B; Forbes, Helen; Mohebbi, Mohammadreza; Duke, Maxine
2017-12-01
Teaching nursing in clinical environments is considered complex and multi-faceted. Little is known about the role of the clinical nurse educator, specifically the challenges related to transition from clinician, or in some cases, from newly-graduated nurse to that of clinical nurse educator, as occurs in developing countries. Confidence in the clinical educator role has been associated with successful transition and the development of role competence. There is currently no valid and reliable instrument to measure clinical nurse educator confidence. This study was conducted to develop and psychometrically test an instrument to measure perceived confidence among clinical nurse educators. A multi-phase, multi-setting survey design was used. A total of 468 surveys were distributed, and 363 were returned. Data were analyzed using exploratory and confirmatory factor analyses. The instrument was successfully tested and modified in phase 1, and factorial validity was subsequently confirmed in phase 2. There was strong evidence of internal consistency, reliability, content, and convergent validity of the Clinical Nurse Educator Skill Acquisition Assessment instrument. The resulting instrument is applicable in similar contexts due to its rigorous development and validation process. © 2017 The Authors. Nursing & Health Sciences published by John Wiley & Sons Australia, Ltd.
Methodology, Methods, and Metrics for Testing and Evaluating Augmented Cognition Systems
DOE Office of Scientific and Technical Information (OSTI.GOV)
Greitzer, Frank L.
The augmented cognition research community seeks cognitive neuroscience-based solutions to improve warfighter performance by applying and managing mitigation strategies to reduce workload and improve the throughput and quality of decisions. The focus of augmented cognition mitigation research is to define, demonstrate, and exploit neuroscience and behavioral measures that support inferences about the warfighter’s cognitive state that prescribe the nature and timing of mitigation. A research challenge is to develop valid evaluation methodologies, metrics and measures to assess the impact of augmented cognition mitigations. Two considerations are external validity, which is the extent to which the results apply to operational contexts;more » and internal validity, which reflects the reliability of performance measures and the conclusions based on analysis of results. The scientific rigor of the research methodology employed in conducting empirical investigations largely affects the validity of the findings. External validity requirements also compel us to demonstrate operational significance of mitigations. Thus it is important to demonstrate effectiveness of mitigations under specific conditions. This chapter reviews some cognitive science and methodological considerations in designing augmented cognition research studies and associated human performance metrics and analysis methods to assess the impact of augmented cognition mitigations.« less
Park, Seong Ho; Kressel, Herbert Y
2018-05-28
Artificial intelligence (AI) is projected to substantially influence clinical practice in the foreseeable future. However, despite the excitement around the technologies, it is yet rare to see examples of robust clinical validation of the technologies and, as a result, very few are currently in clinical use. A thorough, systematic validation of AI technologies using adequately designed clinical research studies before their integration into clinical practice is critical to ensure patient benefit and safety while avoiding any inadvertent harms. We would like to suggest several specific points regarding the role that peer-reviewed medical journals can play, in terms of study design, registration, and reporting, to help achieve proper and meaningful clinical validation of AI technologies designed to make medical diagnosis and prediction, focusing on the evaluation of diagnostic accuracy efficacy. Peer-reviewed medical journals can encourage investigators who wish to validate the performance of AI systems for medical diagnosis and prediction to pay closer attention to the factors listed in this article by emphasizing their importance. Thereby, peer-reviewed medical journals can ultimately facilitate translating the technological innovations into real-world practice while securing patient safety and benefit.
Evaluation of nucleus segmentation in digital pathology images through large scale image synthesis
NASA Astrophysics Data System (ADS)
Zhou, Naiyun; Yu, Xiaxia; Zhao, Tianhao; Wen, Si; Wang, Fusheng; Zhu, Wei; Kurc, Tahsin; Tannenbaum, Allen; Saltz, Joel; Gao, Yi
2017-03-01
Digital histopathology images with more than 1 Gigapixel are drawing more and more attention in clinical, biomedical research, and computer vision fields. Among the multiple observable features spanning multiple scales in the pathology images, the nuclear morphology is one of the central criteria for diagnosis and grading. As a result it is also the mostly studied target in image computing. Large amount of research papers have devoted to the problem of extracting nuclei from digital pathology images, which is the foundation of any further correlation study. However, the validation and evaluation of nucleus extraction have yet been formulated rigorously and systematically. Some researches report a human verified segmentation with thousands of nuclei, whereas a single whole slide image may contain up to million. The main obstacle lies in the difficulty of obtaining such a large number of validated nuclei, which is essentially an impossible task for pathologist. We propose a systematic validation and evaluation approach based on large scale image synthesis. This could facilitate a more quantitatively validated study for current and future histopathology image analysis field.
Increased scientific rigor will improve reliability of research and effectiveness of management
Sells, Sarah N.; Bassing, Sarah B.; Barker, Kristin J.; Forshee, Shannon C.; Keever, Allison; Goerz, James W.; Mitchell, Michael S.
2018-01-01
Rigorous science that produces reliable knowledge is critical to wildlife management because it increases accurate understanding of the natural world and informs management decisions effectively. Application of a rigorous scientific method based on hypothesis testing minimizes unreliable knowledge produced by research. To evaluate the prevalence of scientific rigor in wildlife research, we examined 24 issues of the Journal of Wildlife Management from August 2013 through July 2016. We found 43.9% of studies did not state or imply a priori hypotheses, which are necessary to produce reliable knowledge. We posit that this is due, at least in part, to a lack of common understanding of what rigorous science entails, how it produces more reliable knowledge than other forms of interpreting observations, and how research should be designed to maximize inferential strength and usefulness of application. Current primary literature does not provide succinct explanations of the logic behind a rigorous scientific method or readily applicable guidance for employing it, particularly in wildlife biology; we therefore synthesized an overview of the history, philosophy, and logic that define scientific rigor for biological studies. A rigorous scientific method includes 1) generating a research question from theory and prior observations, 2) developing hypotheses (i.e., plausible biological answers to the question), 3) formulating predictions (i.e., facts that must be true if the hypothesis is true), 4) designing and implementing research to collect data potentially consistent with predictions, 5) evaluating whether predictions are consistent with collected data, and 6) drawing inferences based on the evaluation. Explicitly testing a priori hypotheses reduces overall uncertainty by reducing the number of plausible biological explanations to only those that are logically well supported. Such research also draws inferences that are robust to idiosyncratic observations and unavoidable human biases. Offering only post hoc interpretations of statistical patterns (i.e., a posteriorihypotheses) adds to uncertainty because it increases the number of plausible biological explanations without determining which have the greatest support. Further, post hocinterpretations are strongly subject to human biases. Testing hypotheses maximizes the credibility of research findings, makes the strongest contributions to theory and management, and improves reproducibility of research. Management decisions based on rigorous research are most likely to result in effective conservation of wildlife resources.
Forgatch, Marion S.; Patterson, Gerald R.; DeGarmo, David S.
2006-01-01
When efficacious interventions are implemented in real-world conditions, it is important to evaluate whether or not the programs are practiced as intended. This article presents the Fidelity of Implementation Rating System (FIMP), an observation-based measure assessing competent adherence to the Oregon model of Parent Management Training (PMTO). FIMP evaluates 5 dimensions of competent adherence to PMTO (i.e., knowledge, structure, teaching skill, clinical skill, and overall effectiveness) specified in the intervention model. Predictive validity for FIMP was evaluated with a subsample of stepfamilies participating in a preventive PMTO intervention. As hypothesized, high FIMP ratings predicted change in observed parenting practices from baseline to 12 months. The rigor and scope of adherence measures are discussed. PMID:16718302
How Confident can we be in Flood Risk Assessments?
NASA Astrophysics Data System (ADS)
Merz, B.
2017-12-01
Flood risk management should be based on risk analyses quantifying the risk and its reduction for different risk reduction strategies. However, validating risk estimates by comparing model simulations with past observations is hardly possible, since the assessment typically encompasses extreme events and their impacts that have not been observed before. Hence, risk analyses are strongly based on assumptions and expert judgement. This situation opens the door for cognitive biases, such as `illusion of certainty', `overconfidence' or `recency bias'. Such biases operate specifically in complex situations with many factors involved, when uncertainty is high and events are probabilistic, or when close learning feedback loops are missing - aspects that all apply to risk analyses. This contribution discusses how confident we can be in flood risk assessments, and reflects about more rigorous approaches towards their validation.
Goodman, L A; Corcoran, C; Turner, K; Yuan, N; Green, B L
1998-07-01
This article reviews the psychometric properties of the Stressful Life Events Screening Questionnaire (SLESQ), a recently developed trauma history screening measure, and discusses the complexities involved in assessing trauma exposure. There are relatively few general measures of exposure to a variety of types of traumatic events, and most of those that exist have not been subjected to rigorous psychometric evaluation. The SLESQ showed good test-retest reliability, with a median kappa of .73, adequate convergent validity (with a lengthier interview) with a median kappa of .64, and good discrimination between Criterion A and non-Criterion A events. The discussion addresses some of the challenges of assessing traumatic event exposure along the dimensions of defining traumatic events, assessment methodologies, reporting consistency, and incident validation.
Hsu, Ya-Chuan
2011-09-01
: Diverse social and recreational activities in elder care institutions have been provided to enrich a person's mental well-being amidst what is a relatively monotonous life. However, few instruments that measure the social activities of long-term care residents are available. : This study was designed to develop a culturally sensitive instrument (Socially Supportive Activity Inventory, SSAI) to assess quantity and quality of social activities for long-term care institutions and validate the instrument's psychometric properties. : The SSAI was developed on the basis of the social support theory, a synthesis of literature, and Taiwanese cultural mores. The instrument was rigorously subjected to a two-stage process to evaluate its reliability and validity. In Stage 1, six experts from diverse backgrounds were recruited to evaluate instrument items and estimate the content validity of the instrument using a content validity questionnaire. Items were modified and refined on the basis of the responses of the expert panel and a set of criteria. After obtaining approval from a university institutional review board, in the second stage of evaluating test-retest reliability, a convenience sample of 10 Taiwanese institutionalized elders in a pilot study, recruited from a nursing home, completed the revised instrument at two separate times over 2 weeks. : Results showed a content validity of .96. Test-retest reliability from a sample of 10 participants yielded stability coefficients of .76-1.00. The stability coefficient was 1.00 for the component of frequency, .76-1.00 for the component of meaningfulness, and .78-1.00 for the component of enjoyment. : The SSAI is a highly relevant and reliable culturally based instrument that can measure social activity in long-term care facilities. Because of the pilot nature of this study, future directions include further exploration of the SSAI instrument's psychometric properties. This should be done by enlarging the sample size to include more long-term care facilities and individual participants. Future studies can utilize diverse measures of social activity for comparison and validation of the SSAI.
Hagelstein, V; Ortland, I; Wilmer, A; Mitchell, S A; Jaehde, U
2016-12-01
Integrating the patient's perspective has become an increasingly important component of adverse event reporting. The National Cancer Institute has developed a Patient-Reported Outcomes version of the Common Terminology Criteria for Adverse Events (PRO-CTCAE™). This instrument has been translated into German and linguistically validated; however, its quantitative measurement properties have not been evaluated. A German language survey that included 31 PRO-CTCAE items, as well as the EORTC QLQ-C30 and the Oral Mucositis Daily Questionnaire (OMDQ), was distributed at 10 cancer treatment settings in Germany and Austria. Item quality was assessed by analysis of acceptability and comprehensibility. Reliability was evaluated by using Cronbach's' alpha and validity by principal components analysis (PCA), multitrait-multimethod matrix (MTMM) and known groups validity techniques. Of 660 surveys distributed to the study centres, 271 were returned (return rate 41%), and data from 262 were available for analysis. Participants' median age was 59.7 years, and 69.5% of the patients were female. Analysis of item quality supported the comprehensibility of the 31 PRO-CTCAE items. Reliability was very good; Cronbach's' alpha correlation coefficients were >0.9 for almost all item clusters. Construct validity of the PRO-CTCAE core item set was shown by identifying 10 conceptually meaningful item clusters via PCA. Moreover, construct validity was confirmed by the MTMM: monotrait-heteromethod comparison showed 100% high correlation, whereas heterotrait-monomethod comparison indicated 0% high correlation. Known groups validity was supported; PRO-CTCAE scores were significantly lower for those with impaired versus preserved health-related quality of life. A set of 31 items drawn from the German PRO-CTCAE item library demonstrated favourable measurement properties. These findings add to the body of evidence that PRO-CTCAE provides a rigorous method to capture patient self-reports of symptomatic toxicity for use in cancer clinical trials. © The Author 2016. Published by Oxford University Press on behalf of the European Society for Medical Oncology. All rights reserved. For permissions, please email: journals.permissions@oup.com.
Hazardous Asteroids: Cloaking STEM Skills Training within an Attention-Grabbing Science/Math Course
NASA Astrophysics Data System (ADS)
Ryan, Eileen V.; Ryan, William H.
2015-11-01
A graduate-level course was designed and taught during the summer months from 2009 - 2015 in order to contribute to the training and professional development of K-12 teachers residing in the Southwest. The teachers were seeking Master’s degrees via the New Mexico Institute of Mining and Technology’s (NMT’s) Masters of Science Teaching (MST) program, and the course satisfied a science or math requirement. The MST program provides opportunities for in-service teachers to enhance their content backgrounds in science, mathematics, engineering, and technology (SMET). The ultimate goal is to assist teachers in gaining knowledge that has direct application in the classroom.The engaging topic area of near-Earth object (NEO) characterization studies was used to create a fun and exciting framework for mastering basic skills and concepts in physics and astronomy. The objective was to offer a class that had the appropriate science rigor (with an emphasis on mathematics) within a non-threatening format. The course, entitled “Hazardous Asteroids”, incorporates a basic planetary physics curriculum, with challenging laboratories that include a heavy emphasis on math and technology. Since the authors run a NASA-funded NEO research and follow-up program, also folded into the course is the use of the Magdalena Ridge Observatory’s 2.4-meter telescope so participants can take and reduce their own data on a near-Earth asteroid.In exit assessments, the participants have given the course excellent ratings for design and implementation, and the overall degree of satisfaction was high. This validates that a well-constructed (and rigorous) course can be effective in receptively reaching teachers in need of basic skills refreshment. Many of the teachers taking the course were employed in school districts serving at-risk or under-prepared students, and the course helped provide them with the confidence vital to developing new strategies for successful teaching.
On the use of log-transformation vs. nonlinear regression for analyzing biological power laws.
Xiao, Xiao; White, Ethan P; Hooten, Mevin B; Durham, Susan L
2011-10-01
Power-law relationships are among the most well-studied functional relationships in biology. Recently the common practice of fitting power laws using linear regression (LR) on log-transformed data has been criticized, calling into question the conclusions of hundreds of studies. It has been suggested that nonlinear regression (NLR) is preferable, but no rigorous comparison of these two methods has been conducted. Using Monte Carlo simulations, we demonstrate that the error distribution determines which method performs better, with NLR better characterizing data with additive, homoscedastic, normal error and LR better characterizing data with multiplicative, heteroscedastic, lognormal error. Analysis of 471 biological power laws shows that both forms of error occur in nature. While previous analyses based on log-transformation appear to be generally valid, future analyses should choose methods based on a combination of biological plausibility and analysis of the error distribution. We provide detailed guidelines and associated computer code for doing so, including a model averaging approach for cases where the error structure is uncertain.
NASA Technical Reports Server (NTRS)
Grindle, Laurie; Hackenberg, Davis L.
2016-01-01
UAS Integration in the NAS Project has: a) Developed Technical Challenges that are crucial to UAS integration, aligned with NASA's Strategic Plan and Thrusts, and support FAA standards development. b) Demonstrated rigorous project management processes through the execution of previous phases. c) Defined Partnership Plans. d) Established path to KDP-C. Request approval of Technical Challenges, execution of partnerships and plans, and execution of near-term FY17 activities. There is an increasing need to fly UAS in the NAS to perform missions of vital importance to National Security and Defense, Emergency Management, and Science. There is also an emerging need to enable commercial applications such as cargo transport (e.g. FedEx). Unencumbered NAS Access for Civil/Commercial UAS. Provide research findings, utilizing simulation and flight tests, to support the development and validation of DAA and C2 technologies necessary for integrating Unmanned Aircraft Systems into the National Airspace System.
NASA Astrophysics Data System (ADS)
Cashmore, Matthew; Bond, Alan; Cobb, Dick
2007-09-01
It has long been suggested that environmental assessment has the potential to contribute to sustainable development through mechanisms above and beyond informing design and consent decisions, and while theories have been proposed to explain how this might occur, few have been subjected to rigorous empirical validation. This research advances the theoretical debate by building a rich empirical understanding of environmental assessment’s practical outcomes, from which its potential to contribute to sustainable development can be gauged. Three case study environmental assessment processes in England were investigated using a combination of data generated from content analysis, in-depth interviews, and a questionnaire survey. Four categories of outcomes are delineated based on the research data: learning outcomes; governance outcomes; attitudinal and value changes; and developmental outcomes. The data provide a robust critique of mainstream theory, with its focus on design and consent decisions. The article concludes with an examination of the consequences of the context-specific nature of environmental assessment practices in terms of developing theory and focusing future research.
Kumar, Nagi; Crocker, Theresa; Smith, Tiffany; Connors, Shahnjayla; Pow-Sang, Julio; Spiess, Philippe E; Egan, Kathleen; Quinn, Gwen; Schell, Michael; Sebti, Said; Kazi, Aslam; Chuang, Tian; Salup, Raoul; Helal, Mohamed; Zagaja, Gregory; Trabulsi, Edouard; McLarty, Jerry; Fazili, Tajammul; Williams, Christopher R; Schreiber, Fred; Anderson, Kyle
2012-01-21
In spite of the large number of nutrient-derived agents demonstrating promise as potential chemopreventive agents, most have failed to prove effectiveness in clinical trials. Critical requirements for moving nutrient-derived agents to recommendation for clinical use include adopting a systematic, molecular-mechanism based approach and utilizing the same ethical and rigorous methods such as are used to evaluate other pharmacological agents. Preliminary data on a mechanistic rationale for chemoprevention activity as observed from epidemiological, in vitro and preclinical studies, phase I data of safety in suitable cohorts, duration of intervention based on time to progression of preneoplastic disease to cancer and the use of a valid panel of biomarkers representing the hypothesized carcinogenesis pathway for measuring efficacy must inform the design of phase II clinical trials. The goal of this paper is to provide a model for evaluating a well characterized agent- Polyphenon E- in a phase II clinical trial of prostate cancer chemoprevention.
Kumar, Nagi; Crocker, Theresa; Smith, Tiffany; Connors, Shahnjayla; Pow-Sang, Julio; Spiess, Philippe E.; Egan, Kathleen; Quinn, Gwen; Schell, Michael; Sebti, Said; Kazi, Aslam; Chuang, Tian; Salup, Raoul; Helal, Mohamed; Zagaja, Gregory; Trabulsi, Edouard; McLarty, Jerry; Fazili, Tajammul; Williams, Christopher R.; Schreiber, Fred; Anderson, Kyle
2014-01-01
In spite of the large number of nutrient-derived agents demonstrating promise as potential chemopreventive agents, most have failed to prove effectiveness in clinical trials. Critical requirements for moving nutrient-derived agents to recommendation for clinical use include adopting a systematic, molecular-mechanism based approach and utilizing the same ethical and rigorous methods such as are used to evaluate other pharmacological agents. Preliminary data on a mechanistic rationale for chemoprevention activity as observed from epidemiological, in vitro and preclinical studies, phase I data of safety in suitable cohorts, duration of intervention based on time to progression of preneoplastic disease to cancer and the use of a valid panel of biomarkers representing the hypothesized carcinogenesis pathway for measuring efficacy must inform the design of phase II clinical trials. The goal of this paper is to provide a model for evaluating a well characterized agent- Polyphenon E- in a phase II clinical trial of prostate cancer chemoprevention. PMID:24533253
Simulation of Watts Bar Unit 1 Initial Startup Tests with Continuous Energy Monte Carlo Methods
DOE Office of Scientific and Technical Information (OSTI.GOV)
Godfrey, Andrew T; Gehin, Jess C; Bekar, Kursat B
2014-01-01
The Consortium for Advanced Simulation of Light Water Reactors* is developing a collection of methods and software products known as VERA, the Virtual Environment for Reactor Applications. One component of the testing and validation plan for VERA is comparison of neutronics results to a set of continuous energy Monte Carlo solutions for a range of pressurized water reactor geometries using the SCALE component KENO-VI developed by Oak Ridge National Laboratory. Recent improvements in data, methods, and parallelism have enabled KENO, previously utilized predominately as a criticality safety code, to demonstrate excellent capability and performance for reactor physics applications. The highlymore » detailed and rigorous KENO solutions provide a reliable nu-meric reference for VERAneutronics and also demonstrate the most accurate predictions achievable by modeling and simulations tools for comparison to operating plant data. This paper demonstrates the performance of KENO-VI for the Watts Bar Unit 1 Cycle 1 zero power physics tests, including reactor criticality, control rod worths, and isothermal temperature coefficients.« less
Robust approximation-free prescribed performance control for nonlinear systems and its application
NASA Astrophysics Data System (ADS)
Sun, Ruisheng; Na, Jing; Zhu, Bin
2018-02-01
This paper presents a robust prescribed performance control approach and its application to nonlinear tail-controlled missile systems with unknown dynamics and uncertainties. The idea of prescribed performance function (PPF) is incorporated into the control design, such that both the steady-state and transient control performance can be strictly guaranteed. Unlike conventional PPF-based control methods, we further tailor a recently proposed systematic control design procedure (i.e. approximation-free control) using the transformed tracking error dynamics, which provides a proportional-like control action. Hence, the function approximators (e.g. neural networks, fuzzy systems) that are widely used to address the unknown nonlinearities in the nonlinear control designs are not needed. The proposed control design leads to a robust yet simplified function approximation-free control for nonlinear systems. The closed-loop system stability and the control error convergence are all rigorously proved. Finally, comparative simulations are conducted based on nonlinear missile systems to validate the improved response and the robustness of the proposed control method.
Wang, Boyi; Tan, Hua-Wei; Fang, Wanping; Meinhardt, Lyndel W; Mischke, Sue; Matsumoto, Tracie; Zhang, Dapeng
2015-01-01
Longan (Dimocarpus longan Lour.) is an important tropical fruit tree crop. Accurate varietal identification is essential for germplasm management and breeding. Using longan transcriptome sequences from public databases, we developed single nucleotide polymorphism (SNP) markers; validated 60 SNPs in 50 longan germplasm accessions, including cultivated varieties and wild germplasm; and designated 25 SNP markers that unambiguously identified all tested longan varieties with high statistical rigor (P<0.0001). Multiple trees from the same clone were verified and off-type trees were identified. Diversity analysis revealed genetic relationships among analyzed accessions. Cultivated varieties differed significantly from wild populations (Fst=0.300; P<0.001), demonstrating untapped genetic diversity for germplasm conservation and utilization. Within cultivated varieties, apparent differences between varieties from China and those from Thailand and Hawaii indicated geographic patterns of genetic differentiation. These SNP markers provide a powerful tool to manage longan genetic resources and breeding, with accurate and efficient genotype identification. PMID:26504559
Unmanned Aircraft Systems in the National Airspace System: A Formal Methods Perspective
NASA Technical Reports Server (NTRS)
Munoz, Cesar A.; Dutle, Aaron; Narkawicz, Anthony; Upchurch, Jason
2016-01-01
As the technological and operational capabilities of unmanned aircraft systems (UAS) have grown, so too have international efforts to integrate UAS into civil airspace. However, one of the major concerns that must be addressed in realizing this integration is that of safety. For example, UAS lack an on-board pilot to comply with the legal requirement that pilots see and avoid other aircraft. This requirement has motivated the development of a detect and avoid (DAA) capability for UAS that provides situational awareness and maneuver guidance to UAS operators to aid them in avoiding and remaining well clear of other aircraft in the airspace. The NASA Langley Research Center Formal Methods group has played a fundamental role in the development of this capability. This article gives a selected survey of the formal methods work conducted in support of the development of a DAA concept for UAS. This work includes specification of low-level and high-level functional requirements, formal verification of algorithms, and rigorous validation of software implementations.
NASA Astrophysics Data System (ADS)
Hirst, Jonathan D.; King, Ross D.; Sternberg, Michael J. E.
1994-08-01
Neural networks and inductive logic programming (ILP) have been compared to linear regression for modelling the QSAR of the inhibition of E. coli dihydrofolate reductase (DHFR) by 2,4-diamino-5-(substitured benzyl)pyrimidines, and, in the subsequent paper [Hirst, J.D., King, R.D. and Sternberg, M.J.E., J. Comput.-Aided Mol. Design, 8 (1994) 421], the inhibition of rodent DHFR by 2,4-diamino-6,6-dimethyl-5-phenyl-dihydrotriazines. Cross-validation trials provide a statistically rigorous assessment of the predictive capabilities of the methods, with training and testing data selected randomly and all the methods developed using identical training data. For the ILP analysis, molecules are represented by attributes other than Hansch parameters. Neural networks and ILP perform better than linear regression using the attribute representation, but the difference is not statistically significant. The major benefit from the ILP analysis is the formulation of understandable rules relating the activity of the inhibitors to their chemical structure.
Unified quantitative characterization of epithelial tissue development
Guirao, Boris; Rigaud, Stéphane U; Bosveld, Floris; Bailles, Anaïs; López-Gay, Jesús; Ishihara, Shuji; Sugimura, Kaoru
2015-01-01
Understanding the mechanisms regulating development requires a quantitative characterization of cell divisions, rearrangements, cell size and shape changes, and apoptoses. We developed a multiscale formalism that relates the characterizations of each cell process to tissue growth and morphogenesis. Having validated the formalism on computer simulations, we quantified separately all morphogenetic events in the Drosophila dorsal thorax and wing pupal epithelia to obtain comprehensive statistical maps linking cell and tissue scale dynamics. While globally cell shape changes, rearrangements and divisions all significantly participate in tissue morphogenesis, locally, their relative participations display major variations in space and time. By blocking division we analyzed the impact of division on rearrangements, cell shape changes and tissue morphogenesis. Finally, by combining the formalism with mechanical stress measurement, we evidenced unexpected interplays between patterns of tissue elongation, cell division and stress. Our formalism provides a novel and rigorous approach to uncover mechanisms governing tissue development. DOI: http://dx.doi.org/10.7554/eLife.08519.001 PMID:26653285
A VLF-based technique in applications to digital control of nonlinear hybrid multirate systems
NASA Astrophysics Data System (ADS)
Vassilyev, Stanislav; Ulyanov, Sergey; Maksimkin, Nikolay
2017-01-01
In this paper, a technique for rigorous analysis and design of nonlinear multirate digital control systems on the basis of the reduction method and sublinear vector Lyapunov functions is proposed. The control system model under consideration incorporates continuous-time dynamics of the plant and discrete-time dynamics of the controller and takes into account uncertainties of the plant, bounded disturbances, nonlinear characteristics of sensors and actuators. We consider a class of multirate systems where the control update rate is slower than the measurement sampling rates and periodic non-uniform sampling is admitted. The proposed technique does not use the preliminary discretization of the system, and, hence, allows one to eliminate the errors associated with the discretization and improve the accuracy of analysis. The technique is applied to synthesis of digital controller for a flexible spacecraft in the fine stabilization mode and decentralized controller for a formation of autonomous underwater vehicles. Simulation results are provided to validate the good performance of the designed controllers.
Aeronomy of Ice in the Mesosphere (AIM)
NASA Technical Reports Server (NTRS)
2003-01-01
The overall goal of the Aeronomy of Ice in the Mesosphere (AIM) experiment is to resolve why Polar Mesospheric Clouds form and why they vary. By measuring PMCs and the thermal, chemical and dynamical environment in which they form, we will quanti@ the connection between these clouds and the meteorology of the polar mesosphere. In the end, this will provide the basis for study of long-term variability in the mesospheric climate and its relationship to global change. The results of AIM will be a rigorous validation of predictive models that can reliably use past PMC changes and present trends as indicators of global change. The AIM goal will be achieved by measuring PMC extinction, brightness, spatial distribution, particle size distributions, gravity wave activity, dust influx to the atmosphere and precise, vertical profile measurements of temperature, H20, C&, 0 3 , C02, NO. and aerosols. These data can only be obtained by a complement of instruments on an orbiting spacecraft (S/C).
NASA Astrophysics Data System (ADS)
Ren, Jie
2017-12-01
The process by which a kinesin motor couples its ATPase activity with concerted mechanical hand-over-hand steps is a foremost topic of molecular motor physics. Two major routes toward elucidating kinesin mechanisms are the motility performance characterization of velocity and run length, and single-molecular state detection experiments. However, these two sets of experimental approaches are largely uncoupled to date. Here, we introduce an integrative motility state analysis based on a theorized kinetic graph theory for kinesin, which, on one hand, is validated by a wealth of accumulated motility data, and, on the other hand, allows for rigorous quantification of state occurrences and chemomechanical cycling probabilities. An interesting linear scaling for kinesin motility performance across species is discussed as well. An integrative kinetic graph theory analysis provides a powerful tool to bridge motility and state characterization experiments, so as to forge a unified effort for the elucidation of the working mechanisms of molecular motors.
One-electron reduced density matrices of strongly correlated harmonium atoms.
Cioslowski, Jerzy
2015-03-21
Explicit asymptotic expressions are derived for the reduced one-electron density matrices (the 1-matrices) of strongly correlated two- and three-electron harmonium atoms in the ground and first excited states. These expressions, which are valid at the limit of small confinement strength ω, yield electron densities and kinetic energies in agreement with the published values. In addition, they reveal the ω(5/6) asymptotic scaling of the exchange components of the electron-electron repulsion energies that differs from the ω(2/3) scaling of their Coulomb and correlation counterparts. The natural orbitals of the totally symmetric ground state of the two-electron harmonium atom are found to possess collective occupancies that follow a mixed power/Gaussian dependence on the angular momentum in variance with the simple power-law prediction of Hill's asymptotics. Providing rigorous constraints on energies as functionals of 1-matrices, these results are expected to facilitate development of approximate implementations of the density matrix functional theory and ensure their proper description of strongly correlated systems.
Change rates and prevalence of a dichotomous variable: simulations and applications.
Brinks, Ralph; Landwehr, Sandra
2015-01-01
A common modelling approach in public health and epidemiology divides the population under study into compartments containing persons that share the same status. Here we consider a three-state model with the compartments: A, B and Dead. States A and B may be the states of any dichotomous variable, for example, Healthy and Ill, respectively. The transitions between the states are described by change rates, which depend on calendar time and on age. So far, a rigorous mathematical calculation of the prevalence of property B has been difficult, which has limited the use of the model in epidemiology and public health. We develop a partial differential equation (PDE) that simplifies the use of the three-state model. To demonstrate the validity of the PDE, it is applied to two simulation studies, one about a hypothetical chronic disease and one about dementia in Germany. In two further applications, the PDE may provide insights into smoking behaviour of males in Germany and the knowledge about the ovulatory cycle in Egyptian women.
Data Assimilation - Advances and Applications
DOE Office of Scientific and Technical Information (OSTI.GOV)
Williams, Brian J.
2014-07-30
This presentation provides an overview of data assimilation (model calibration) for complex computer experiments. Calibration refers to the process of probabilistically constraining uncertain physics/engineering model inputs to be consistent with observed experimental data. An initial probability distribution for these parameters is updated using the experimental information. Utilization of surrogate models and empirical adjustment for model form error in code calibration form the basis for the statistical methodology considered. The role of probabilistic code calibration in supporting code validation is discussed. Incorporation of model form uncertainty in rigorous uncertainty quantification (UQ) analyses is also addressed. Design criteria used within a batchmore » sequential design algorithm are introduced for efficiently achieving predictive maturity and improved code calibration. Predictive maturity refers to obtaining stable predictive inference with calibrated computer codes. These approaches allow for augmentation of initial experiment designs for collecting new physical data. A standard framework for data assimilation is presented and techniques for updating the posterior distribution of the state variables based on particle filtering and the ensemble Kalman filter are introduced.« less
Weber, Erica; Blackstone, Kaitlin; Woods, Steven Paul
2013-01-01
Despite significant advances in the virologic management of HIV infection over the last two decades, effective treatments for HIV-associated neurocognitive disorders (HAND) remain elusive. While pharmacological interventions have yielded some success in improving neurocognitive outcomes in HIV, there is a dearth of rigorous studies examining the efficacy of cognitive rehabilitation for remediating HIV-associated neurocognitive impairment. This qualitative review summarizes and critiques the emerging literature on cognitive and behavioral treatments for HAND, which provides many reasons for optimism, but also has major limitations that underscore the scope of the work that lies ahead. Considering the notable real-world consequences of HAND, the development, validation, and clinical deployment of cognitive neurorehabilitation interventions tailored to the needs of persons living with HIV infection is a priority for clinical neuroAIDS investigators. In describing potential future directions for this endeavor, particular attention was paid to the application of cognitive neuropsychological principles in developing theory-driven approaches to managing HAND, improving everyday functioning, and enhancing HIV health outcomes. PMID:23417497
Development of rigor mortis is not affected by muscle volume.
Kobayashi, M; Ikegaya, H; Takase, I; Hatanaka, K; Sakurada, K; Iwase, H
2001-04-01
There is a hypothesis suggesting that rigor mortis progresses more rapidly in small muscles than in large muscles. We measured rigor mortis as tension determined isometrically in rat musculus erector spinae that had been cut into muscle bundles of various volumes. The muscle volume did not influence either the progress or the resolution of rigor mortis, which contradicts the hypothesis. Differences in pre-rigor load on the muscles influenced the onset and resolution of rigor mortis in a few pairs of samples, but did not influence the time taken for rigor mortis to reach its full extent after death. Moreover, the progress of rigor mortis in this muscle was biphasic; this may reflect the early rigor of red muscle fibres and the late rigor of white muscle fibres.
Plutonium Critical Mass Curve Comparison to Mass at Upper Subcritical Limit (USL) Using Whisper
DOE Office of Scientific and Technical Information (OSTI.GOV)
Alwin, Jennifer Louise; Zhang, Ning
Whisper is computational software designed to assist the nuclear criticality safety analyst with validation studies with the MCNP ® Monte Carlo radiation transport package. Standard approaches to validation rely on the selection of benchmarks based upon expert judgment. Whisper uses sensitivity/uncertainty (S/U) methods to select relevant benchmarks to a particular application or set of applications being analyzed. Using these benchmarks, Whisper computes a calculational margin. Whisper attempts to quantify the margin of subcriticality (MOS) from errors in software and uncertainties in nuclear data. The combination of the Whisper-derived calculational margin and MOS comprise the baseline upper subcritical limit (USL), tomore » which an additional margin may be applied by the nuclear criticality safety analyst as appropriate to ensure subcriticality. A series of critical mass curves for plutonium, similar to those found in Figure 31 of LA-10860-MS, have been generated using MCNP6.1.1 and the iterative parameter study software, WORM_Solver. The baseline USL for each of the data points of the curves was then computed using Whisper 1.1. The USL was then used to determine the equivalent mass for plutonium metal-water system. ANSI/ANS-8.1 states that it is acceptable to use handbook data, such as the data directly from the LA-10860-MS, as it is already considered validated (Section 4.3 4) “Use of subcritical limit data provided in ANSI/ANS standards or accepted reference publications does not require further validation.”). This paper attempts to take a novel approach to visualize traditional critical mass curves and allows comparison with the amount of mass for which the k eff is equal to the USL (calculational margin + margin of subcriticality). However, the intent is to plot the critical mass data along with USL, not to suggest that already accepted handbook data should have new and more rigorous requirements for validation.« less
Fundamentals of endoscopic surgery: creation and validation of the hands-on test.
Vassiliou, Melina C; Dunkin, Brian J; Fried, Gerald M; Mellinger, John D; Trus, Thadeus; Kaneva, Pepa; Lyons, Calvin; Korndorffer, James R; Ujiki, Michael; Velanovich, Vic; Kochman, Michael L; Tsuda, Shawn; Martinez, Jose; Scott, Daniel J; Korus, Gary; Park, Adrian; Marks, Jeffrey M
2014-03-01
The Fundamentals of Endoscopic Surgery™ (FES) program consists of online materials and didactic and skills-based tests. All components were designed to measure the skills and knowledge required to perform safe flexible endoscopy. The purpose of this multicenter study was to evaluate the reliability and validity of the hands-on component of the FES examination, and to establish the pass score. Expert endoscopists identified the critical skill set required for flexible endoscopy. They were then modeled in a virtual reality simulator (GI Mentor™ II, Simbionix™ Ltd., Airport City, Israel) to create five tasks and metrics. Scores were designed to measure both speed and precision. Validity evidence was assessed by correlating performance with self-reported endoscopic experience (surgeons and gastroenterologists [GIs]). Internal consistency of each test task was assessed using Cronbach's alpha. Test-retest reliability was determined by having the same participant perform the test a second time and comparing their scores. Passing scores were determined by a contrasting groups methodology and use of receiver operating characteristic curves. A total of 160 participants (17 % GIs) performed the simulator test. Scores on the five tasks showed good internal consistency reliability and all had significant correlations with endoscopic experience. Total FES scores correlated 0.73, with participants' level of endoscopic experience providing evidence of their validity, and their internal consistency reliability (Cronbach's alpha) was 0.82. Test-retest reliability was assessed in 11 participants, and the intraclass correlation was 0.85. The passing score was determined and is estimated to have a sensitivity (true positive rate) of 0.81 and a 1-specificity (false positive rate) of 0.21. The FES hands-on skills test examines the basic procedural components required to perform safe flexible endoscopy. It meets rigorous standards of reliability and validity required for high-stakes examinations, and, together with the knowledge component, may help contribute to the definition and determination of competence in endoscopy.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Valerio, Luis G., E-mail: luis.valerio@fda.hhs.gov; Cross, Kevin P.
Control and minimization of human exposure to potential genotoxic impurities found in drug substances and products is an important part of preclinical safety assessments of new drug products. The FDA's 2008 draft guidance on genotoxic and carcinogenic impurities in drug substances and products allows use of computational quantitative structure–activity relationships (QSAR) to identify structural alerts for known and expected impurities present at levels below qualified thresholds. This study provides the information necessary to establish the practical use of a new in silico toxicology model for predicting Salmonella t. mutagenicity (Ames assay outcome) of drug impurities and other chemicals. We describemore » the model's chemical content and toxicity fingerprint in terms of compound space, molecular and structural toxicophores, and have rigorously tested its predictive power using both cross-validation and external validation experiments, as well as case studies. Consistent with desired regulatory use, the model performs with high sensitivity (81%) and high negative predictivity (81%) based on external validation with 2368 compounds foreign to the model and having known mutagenicity. A database of drug impurities was created from proprietary FDA submissions and the public literature which found significant overlap between the structural features of drug impurities and training set chemicals in the QSAR model. Overall, the model's predictive performance was found to be acceptable for screening drug impurities for Salmonella mutagenicity. -- Highlights: ► We characterize a new in silico model to predict mutagenicity of drug impurities. ► The model predicts Salmonella mutagenicity and will be useful for safety assessment. ► We examine toxicity fingerprints and toxicophores of this Ames assay model. ► We compare these attributes to those found in drug impurities known to FDA/CDER. ► We validate the model and find it has a desired predictive performance.« less
In defense of the classical height system
NASA Astrophysics Data System (ADS)
Foroughi, Ismael; Vaníček, Petr; Sheng, Michael; Kingdon, Robert William; Santos, Marcelo C.
2017-11-01
In many European countries, normal heights referred to the quasi-geoid as introduced by Molodenskij in the mid-20th century are preferred to the classical height system that consists of orthometric heights and the geoid as a reference surface for these heights. The rationale for this choice is supposed to be that in the classical height system, neither the geoid, nor the orthometric height can be ever known with centimetre level accuracy because one would need to know the topographical mass density to a level that can never be achieved. The aim of this paper is to question the validity of this rationale. The common way of assessing the congruency of a local geoid model and the orthometric heights is to compare the geoid heights with the difference between orthometric heights provided by leveling and geodetic heights provided by GNSS. On the other hand, testing the congruency of a quasi-geoidal model with normal height a similar procedure is used, except that instead of orthometric heights, normal heights are employed. For the area of Auvergne, France, which is now a more or less standard choice for precise geoid or quasi-geoid testing, only the normal heights are supplied by the Institute Geographic National, the provider of the data. This is clearly the consequence of the European preference for the Molodenskij system. The quality of the height system is to be judged by the congruency of the difference of the geoid/quasi-geoid heights subtracted from the geodetic heights and orthometric/normal heights. To assess the congruency of the classical height system, the Helmert approximation of orthometric heights is typically used as the transformation between normal and Helmert's heights is easily done. However, the evaluation of the differences between Helmert's and the rigorous orthometric heights is somewhat more involved as will be seen from the review in this paper. For the area of interest, the differences between normal and Helmert's heights at the control leveling points range between - 9.5 and 0 cm, differences between Helmert's and the rigorous orthometric heights vary between - 3.6 and 1.1 cm. The local gravimetric geoid model of Auvergne, computed by the Stokes-Helmert technique, is used here to illustrate the accuracy of the classical height system. Results show a very reasonable standard deviation (STD) of 3.2 cm of the differences between geoid values, derived from control leveling points, and gravimetric geoid heights when Helmert's heights are employed and even a smaller STD of 2.9 cm when rigorous orthometric heights are used. A corresponding comparison of a quasi-geoid model, computed by Least-Squares Modification of Stokes method, with normal heights show an STD of 3.4 cm.
Hughes, Brianna H; Greenberg, Neil J; Yang, Tom C; Skonberg, Denise I
2015-01-01
High-pressure processing (HPP) is used to increase meat safety and shelf-life, with conflicting quality effects depending on rigor status during HPP. In the seafood industry, HPP is used to shuck and pasteurize oysters, but its use on abalones has only been minimally evaluated and the effect of rigor status during HPP on abalone quality has not been reported. Farm-raised abalones (Haliotis rufescens) were divided into 12 HPP treatments and 1 unprocessed control treatment. Treatments were processed pre-rigor or post-rigor at 2 pressures (100 and 300 MPa) and 3 processing times (1, 3, and 5 min). The control was analyzed post-rigor. Uniform plugs were cut from adductor and foot meat for texture profile analysis, shear force, and color analysis. Subsamples were used for scanning electron microscopy of muscle ultrastructure. Texture profile analysis revealed that post-rigor processed abalone was significantly (P < 0.05) less firm and chewy than pre-rigor processed irrespective of muscle type, processing time, or pressure. L values increased with pressure to 68.9 at 300 MPa for pre-rigor processed foot, 73.8 for post-rigor processed foot, 90.9 for pre-rigor processed adductor, and 89.0 for post-rigor processed adductor. Scanning electron microscopy images showed fraying of collagen fibers in processed adductor, but did not show pressure-induced compaction of the foot myofibrils. Post-rigor processed abalone meat was more tender than pre-rigor processed meat, and post-rigor processed foot meat was lighter in color than pre-rigor processed foot meat, suggesting that waiting for rigor to resolve prior to processing abalones may improve consumer perceptions of quality and market value. © 2014 Institute of Food Technologists®
NASA Technical Reports Server (NTRS)
Priestley, K. J.; Matthews, G.; Thomas, S.
2006-01-01
The CERES Flight Models 1 through 4 instruments were launched aboard NASA's Earth Observing System (EOS) Terra and Aqua Spacecraft into 705 Km sun-synchronous orbits with 10:30 a.m. and 1:30 p.m. equatorial crossing times. These instruments supplement measurements made by the CERES Proto Flight Model (PFM) instrument launched aboard NASA's Tropical Rainfall Measuring Mission (TRMM) into a 350 Km, 38-degree mid-inclined orbit. CERES Climate Data Records consist of geolocated and calibrated instantaneous filtered and unfiltered radiances through temporally and spatially averaged TOA, Surface and Atmospheric fluxes. CERES filtered radiance measurements cover three spectral bands including shortwave (0.3 to 5 microns), total (0.3 to 100 microns) and an atmospheric window channel (8 to 12 microns). The CERES Earth Radiation Budget measurements represent a new era in radiation climate data, realizing a factor of 2 to 4 improvement in calibration accuracy and stability over the previous ERBE climate records, while striving for the next goal of 0.3-percent per decade absolute stability. The current improvement is derived from two sources: the incorporation of lessons learned from the ERBE mission in the design of the CERES instruments and the development of a rigorous and comprehensive radiometric validation protocol consisting of individual studies covering different spatial, spectral and temporal time scales on data collected both pre and post launch. Once this ensemble of individual perspectives is collected and organized, a cohesive and highly rigorous picture of the overall end-to-end performance of the CERES instrument's and data processing algorithms may be clearly established. This approach has resulted in unprecedented levels of accuracy for radiation budget instruments and data products with calibration stability of better than 0.2-percent and calibration traceability from ground to flight of 0.25-percent. The current work summarizes the development, philosophy and implementation of the protocol designed to rigorously quantify the quality of the data products as well as the level of agreement between the CERES TRMM, Terra and Aqua climate data records.
A practical guide to surveys and questionnaires.
Slattery, Eric L; Voelker, Courtney C J; Nussenbaum, Brian; Rich, Jason T; Paniello, Randal C; Neely, J Gail
2011-06-01
Surveys with questionnaires play a vital role in decision and policy making in society. Within medicine, including otolaryngology, surveys with questionnaires may be the only method for gathering data on rare or unusual events. In addition, questionnaires can be developed and validated to be used as outcome measures in clinical trials and other clinical research architecture. Consequently, it is fundamentally important that such tools be properly developed and validated. Just asking questions that have not gone through rigorous design and development may be misleading and unfair at best; at worst, they can result in under- or overtreatment and unnecessary expense. Furthermore, it is important that consumers of the data produced by these instruments understand the principles of questionnaire design to interpret results in an optimal and meaningful way. This article presents a practical guide for understanding the methodologies of survey and questionnaire design, including the concepts of validity and reliability, how surveys are administered and implemented, and, finally, biases and pitfalls of surveys.
NASA Astrophysics Data System (ADS)
Zimmermann, Judith; von Davier, Alina A.; Buhmann, Joachim M.; Heinimann, Hans R.
2018-01-01
Graduate admission has become a critical process in tertiary education, whereby selecting valid admissions instruments is key. This study assessed the validity of Graduate Record Examination (GRE) General Test scores for admission to Master's programmes at a technical university in Europe. We investigated the indicative value of GRE scores for the Master's programme grade point average (GGPA) with and without the addition of the undergraduate GPA (UGPA) and the TOEFL score, and of GRE scores for study completion and Master's thesis performance. GRE scores explained 20% of the variation in the GGPA, while additional 7% were explained by the TOEFL score and 3% by the UGPA. Contrary to common belief, the GRE quantitative reasoning score showed only little explanatory power. GRE scores were also weakly related to study progress but not to thesis performance. Nevertheless, GRE and TOEFL scores were found to be sensible admissions instruments. Rigorous methodology was used to obtain highly reliable results.
Kim, Jung-Hee; Shin, Sujin; Park, Jin-Hwa
2015-04-01
The purpose of this study was to evaluate the methodological quality of nursing studies using structural equation modeling in Korea. Databases of KISS, DBPIA, and National Assembly Library up to March 2014 were searched using the MeSH terms 'nursing', 'structure', 'model'. A total of 152 studies were screened. After removal of duplicates and non-relevant titles, 61 papers were read in full. Of the sixty-one articles retrieved, 14 studies were published between 1992 and 2000, 27, between 2001 and 2010, and 20, between 2011 and March 2014. The methodological quality of the review examined varied considerably. The findings of this study suggest that more rigorous research is necessary to address theoretical identification, two indicator rule, distribution of sample, treatment of missing values, mediator effect, discriminant validity, convergent validity, post hoc model modification, equivalent models issues, and alternative models issues should be undergone. Further research with robust consistent methodological study designs from model identification to model respecification is needed to improve the validity of the research.
An atomic model of brome mosaic virus using direct electron detection and real-space optimization.
Wang, Zhao; Hryc, Corey F; Bammes, Benjamin; Afonine, Pavel V; Jakana, Joanita; Chen, Dong-Hua; Liu, Xiangan; Baker, Matthew L; Kao, Cheng; Ludtke, Steven J; Schmid, Michael F; Adams, Paul D; Chiu, Wah
2014-09-04
Advances in electron cryo-microscopy have enabled structure determination of macromolecules at near-atomic resolution. However, structure determination, even using de novo methods, remains susceptible to model bias and overfitting. Here we describe a complete workflow for data acquisition, image processing, all-atom modelling and validation of brome mosaic virus, an RNA virus. Data were collected with a direct electron detector in integrating mode and an exposure beyond the traditional radiation damage limit. The final density map has a resolution of 3.8 Å as assessed by two independent data sets and maps. We used the map to derive an all-atom model with a newly implemented real-space optimization protocol. The validity of the model was verified by its match with the density map and a previous model from X-ray crystallography, as well as the internal consistency of models from independent maps. This study demonstrates a practical approach to obtain a rigorously validated atomic resolution electron cryo-microscopy structure.
Using the Depression Anxiety Stress Scale 21 (DASS-21) across cultures.
Oei, Tian P S; Sawang, Sukanlaya; Goh, Yong Wah; Mukhtar, Firdaus
2013-01-01
The DASS-21 is a well-established instrument for measuring depression, anxiety, and stress with good reliability and validity reported from Hispanic American, British, and Australian adults. However, the lack of appropriate validation among Asian populations continues to pose concerns over the use of DASS-21 in Asian samples. Cultural variation may influence the individual's experience and emotional expression. Thus, when researchers and practitioners employ Western-based assessments with Asian populations by directly translating them without an appropriate validation, the process can be challenging. We conducted a series of rigorous statistical tests and minimized any potential confounds from the demographic information. Following factor analyses, we performed multigroup analysis across six nations to demonstrate consistency of our findings. The advantages of this revised DASS-18 stress scale are twofold. First, it possesses fewer items, which results in a cleaner factorial structure. Second, it has a smaller interfactor correlation. With these justifications, the revised DASS-18 stress scale is potentially more suitable for Asian populations. Nonetheless, given limitations, findings should be considered preliminary.
NASA Occupant Protection Standards Development
NASA Technical Reports Server (NTRS)
Somers, Jeffrey; Gernhardt, Michael; Lawrence, Charles
2012-01-01
Historically, spacecraft landing systems have been tested with human volunteers, because analytical methods for estimating injury risk were insufficient. These tests were conducted with flight-like suits and seats to verify the safety of the landing systems. Currently, NASA uses the Brinkley Dynamic Response Index to estimate injury risk, although applying it to the NASA environment has drawbacks: (1) Does not indicate severity or anatomical location of injury (2) Unclear if model applies to NASA applications. Because of these limitations, a new validated, analytical approach was desired. Leveraging off of the current state of the art in automotive safety and racing, a new approach was developed. The approach has several aspects: (1) Define the acceptable level of injury risk by injury severity (2) Determine the appropriate human surrogate for testing and modeling (3) Mine existing human injury data to determine appropriate Injury Assessment Reference Values (IARV). (4) Rigorously Validate the IARVs with sub-injurious human testing (5) Use validated IARVs to update standards and vehicle requirement
Student peer assessment in evidence-based medicine (EBM) searching skills training: an experiment
Eldredge, Jonathan D.; Bear, David G.; Wayne, Sharon J.; Perea, Paul P.
2013-01-01
Background: Student peer assessment (SPA) has been used intermittently in medical education for more than four decades, particularly in connection with skills training. SPA generally has not been rigorously tested, so medical educators have limited evidence about SPA effectiveness. Methods: Experimental design: Seventy-one first-year medical students were stratified by previous test scores into problem-based learning tutorial groups, and then these assigned groups were randomized further into intervention and control groups. All students received evidence-based medicine (EBM) training. Only the intervention group members received SPA training, practice with assessment rubrics, and then application of anonymous SPA to assignments submitted by other members of the intervention group. Results: Students in the intervention group had higher mean scores on the formative test with a potential maximum score of 49 points than did students in the control group, 45.7 and 43.5, respectively (P = 0.06). Conclusions: SPA training and the application of these skills by the intervention group resulted in higher scores on formative tests compared to those in the control group, a difference approaching statistical significance. The extra effort expended by librarians, other personnel, and medical students must be factored into the decision to use SPA in any specific educational context. Implications: SPA has not been rigorously tested, particularly in medical education. Future, similarly rigorous studies could further validate use of SPA so that librarians can optimally make use of limited contact time for information skills training in medical school curricula. PMID:24163593
Do we need methodological theory to do qualitative research?
Avis, Mark
2003-09-01
Positivism is frequently used to stand for the epistemological assumption that empirical science based on principles of verificationism, objectivity, and reproducibility is the foundation of all genuine knowledge. Qualitative researchers sometimes feel obliged to provide methodological alternatives to positivism that recognize their different ethical, ontological, and epistemological commitments and have provided three theories: phenomenology, grounded theory, and ethnography. The author argues that positivism was a doomed attempt to define empirical foundations for knowledge through a rigorous separation of theory and evidence; offers a pragmatic, coherent view of knowledge; and suggests that rigorous, rational empirical investigation does not need methodological theory. Therefore, qualitative methodological theory is unnecessary and counterproductive because it hinders critical reflection on the relation between methodological theory and empirical evidence.
NASA Astrophysics Data System (ADS)
Hincks, Ian; Granade, Christopher; Cory, David G.
2018-01-01
The analysis of photon count data from the standard nitrogen vacancy (NV) measurement process is treated as a statistical inference problem. This has applications toward gaining better and more rigorous error bars for tasks such as parameter estimation (e.g. magnetometry), tomography, and randomized benchmarking. We start by providing a summary of the standard phenomenological model of the NV optical process in terms of Lindblad jump operators. This model is used to derive random variables describing emitted photons during measurement, to which finite visibility, dark counts, and imperfect state preparation are added. NV spin-state measurement is then stated as an abstract statistical inference problem consisting of an underlying biased coin obstructed by three Poisson rates. Relevant frequentist and Bayesian estimators are provided, discussed, and quantitatively compared. We show numerically that the risk of the maximum likelihood estimator is well approximated by the Cramér-Rao bound, for which we provide a simple formula. Of the estimators, we in particular promote the Bayes estimator, owing to its slightly better risk performance, and straightforward error propagation into more complex experiments. This is illustrated on experimental data, where quantum Hamiltonian learning is performed and cross-validated in a fully Bayesian setting, and compared to a more traditional weighted least squares fit.
The diabetes online community: Older adults supporting self-care through peer health.
Litchman, Michelle L; Rothwell, Erin; Edelman, Linda S
2018-03-01
The use of the diabetes online community (DOC) is growing across all age groups. The aim of this exploratory study was to describe why older adults participated in the DOC, and how DOC users interacted with their healthcare providers. Telephone interviews (N=20) were conducted with older adult DOC users (born between 1946 and 1964) living in the United States. Interviews were analyzed using qualitative content analysis adhering to rigor and reproducibility standards. Themes that emerged from the data related to DOC participation included: information to improve self-care, emotional support, belonging to a community, validation of information, cause for concern and interaction with healthcare providers. Participants used the DOC for day to day diabetes management advice and healthcare providers for medical information and care. Participants highly valued the DOC and regarded their participation as a way to increase knowledge to improve self-care and reciprocate emotional support with others for diabetes management. The DOC filled gaps in knowledge and support participants were not able to get elsewhere. The DOC serves as an important source of information and support for individuals with diabetes and may be a cost-effective strategy to augment standard diabetes care. Copyright © 2017 Elsevier B.V. All rights reserved.
Hall, Kelli Stidham; Manu, Abubakar; Morhe, Emmanuel; Harris, Lisa H; Loll, Dana; Ela, Elizabeth; Kolenic, Giselle; Dozier, Jessica L; Challa, Sneha; Zochowski, Melissa K; Boakye, Andrew; Adanu, Richard; Dalton, Vanessa K
2018-01-01
Young women's experiences with sexual and reproductive health (SRH) stigma may contribute to unintended pregnancy. Thus, stigma interventions and rigorous measures to assess their impact are needed. Based on formative work, we generated a pool of 51 items on perceived stigma around different dimensions of adolescent SRH and family planning (sex, contraception, pregnancy, childbearing, abortion). We tested items in a survey study of 1,080 women ages 15 to 24 recruited from schools, health facilities, and universities in Ghana. Confirmatory factor analysis (CFA) identified the most conceptually and statistically relevant scale, and multivariable regression established construct validity via associations between stigma and contraceptive use. CFA provided strong support for our hypothesized Adolescent SRH Stigma Scale (chi-square p value < 0.001; root mean square error of approximation [RMSEA] = 0.07; standardized root mean square residual [SRMR] = 0.06). The final 20-item scale included three subscales: internalized stigma (six items), enacted stigma (seven items), and stigmatizing lay attitudes (seven items). The scale demonstrated good internal consistency (α = 0.74) and strong subscale correlations (α = 0.82 to 0.93). Higher SRH stigma scores were inversely associated with ever having used modern contraception (adjusted odds ratio [AOR] = 0.96, confidence interval [CI] = 0.94 to 0.99, p value = 0.006). A valid, reliable instrument for assessing SRH stigma and its impact on family planning, the Adolescent SRH Stigma Scale can inform and evaluate interventions to reduce/manage stigma and foster resilience among young women in Africa and beyond.
Hall, Kelli Stidham; Manu, Abubakar; Morhe, Emmanuel; Harris, Lisa H.; Loll, Dana; Ela, Elizabeth; Kolenic, Giselle; Dozier, Jessica L.; Challa, Sneha; Zochowski, Melissa K.; Boakye, Andrew; Adanu, Richard; Dalton, Vanessa K.
2018-01-01
Young women’s experiences with sexual and reproductive health (SRH) stigma may contribute to unintended pregnancy. Thus, stigma interventions and rigorous measures to assess their impact are needed. Based on formative work, we generated a pool of 51 items on perceived stigma around different dimensions of adolescent SRH and family planning (sex, contraception, pregnancy, child-bearing, abortion). We tested items in a survey study of 1,080 women ages 15 to 24 recruited from schools, health facilities, and universities in Ghana. Confirmatory factor analysis (CFA) identified the most conceptually and statistically relevant scale, and multivariable regression established construct validity via associations between stigma and contraceptive use. CFA provided strong support for our hypothesized Adolescent SRH Stigma Scale (chi-square p value < 0.001; root mean square error of approximation [RMSEA] = 0.07; standardized root mean square residual [SRMR] = 0.06). The final 20-item scale included three subscales: internalized stigma (six items), enacted stigma (seven items), and stigmatizing lay attitudes (seven items). The scale demonstrated good internal consistency (α = 0.74) and strong subscale correlations (α = 0.82 to 0.93). Higher SRH stigma scores were inversely associated with ever having used modern contraception (adjusted odds ratio [AOR] = 0.96, confidence interval [CI] = 0.94 to 0.99, p value = 0.006). A valid, reliable instrument for assessing SRH stigma and its impact on family planning, the Adolescent SRH Stigma Scale can inform and evaluate interventions to reduce/manage stigma and foster resilience among young women in Africa and beyond. PMID:28266874
In silico modeling to predict drug-induced phospholipidosis
DOE Office of Scientific and Technical Information (OSTI.GOV)
Choi, Sydney S.; Kim, Jae S.; Valerio, Luis G., E-mail: luis.valerio@fda.hhs.gov
2013-06-01
Drug-induced phospholipidosis (DIPL) is a preclinical finding during pharmaceutical drug development that has implications on the course of drug development and regulatory safety review. A principal characteristic of drugs inducing DIPL is known to be a cationic amphiphilic structure. This provides evidence for a structure-based explanation and opportunity to analyze properties and structures of drugs with the histopathologic findings for DIPL. In previous work from the FDA, in silico quantitative structure–activity relationship (QSAR) modeling using machine learning approaches has shown promise with a large dataset of drugs but included unconfirmed data as well. In this study, we report the constructionmore » and validation of a battery of complementary in silico QSAR models using the FDA's updated database on phospholipidosis, new algorithms and predictive technologies, and in particular, we address high performance with a high-confidence dataset. The results of our modeling for DIPL include rigorous external validation tests showing 80–81% concordance. Furthermore, the predictive performance characteristics include models with high sensitivity and specificity, in most cases above ≥ 80% leading to desired high negative and positive predictivity. These models are intended to be utilized for regulatory toxicology applied science needs in screening new drugs for DIPL. - Highlights: • New in silico models for predicting drug-induced phospholipidosis (DIPL) are described. • The training set data in the models is derived from the FDA's phospholipidosis database. • We find excellent predictivity values of the models based on external validation. • The models can support drug screening and regulatory decision-making on DIPL.« less
NASA Astrophysics Data System (ADS)
Kahveci, Ajda; Kahveci, Murat; Mansour, Nasser; Alarfaj, Maher Mohammed
2017-06-01
Teachers play a key role in moving reform-based science education practices into the classroom. Based on research that emphasizes the importance of teachers' affective states, this study aimed to explore the constructs pedagogical discontentment, science teaching self-efficacy, intentions to reform, and their correlations. Also, it aimed to provide empirical evidence in light of a previously proposed theoretical model while focusing on an entirely new context in Middle East. Data were collected in Saudi Arabia with a total of randomly selected 994 science teachers, 656 of whom were females and 338 were males. To collect the data, the Arabic versions of the Science Teachers' Pedagogical Discontentment scale, the Science Teaching Efficacy Beliefs Instrument and the Intentions to Reform Science Teaching scale were developed. For assuring the validity of the instruments in a non-Western context, rigorous cross-cultural validations procedures were followed. Factor analyses were conducted for construct validation and descriptive statistical analyses were performed including frequency distributions and normality checks. Univariate analyses of variance were run to explore statistically significant differences between groups of teachers. Cross-tabulation and correlation analyses were conducted to explore relationships. The findings suggest effect of teacher characteristics such as age and professional development program attendance on the affective states. The results demonstrate that teachers who attended a relatively higher number of programs had lower level of intentions to reform raising issues regarding the conduct and outcomes of professional development. Some of the findings concerning interrelationships among the three constructs challenge and serve to expand the previously proposed theoretical model.
Carney, Patricia A; Palmer, Ryan T; Fuqua Miller, Marissa; Thayer, Erin K; Estroff, Sue E; Litzelman, Debra K; Biagioli, Frances E; Teal, Cayla R; Lambros, Ann; Hatt, William J; Satterfield, Jason M
2016-05-01
Behavioral and social science (BSS) competencies are needed to provide quality health care, but psychometrically validated measures to assess these competencies are difficult to find. Moreover, they have not been mapped to existing frameworks, like those from the Liaison Committee on Medical Education (LCME) and Accreditation Council for Graduate Medical Education (ACGME). This systematic review aimed to identify and evaluate the quality of assessment tools used to measure BSS competencies. The authors searched the literature published between January 2002 and March 2014 for articles reporting psychometric or other validity/reliability testing, using OVID, CINAHL, PubMed, ERIC, Research and Development Resource Base, SOCIOFILE, and PsycINFO. They reviewed 5,104 potentially relevant titles and abstracts. To guide their review, they mapped BSS competencies to existing LCME and ACGME frameworks. The final included articles fell into three categories: instrument development, which were of the highest quality; educational research, which were of the second highest quality; and curriculum evaluation, which were of lower quality. Of the 114 included articles, 33 (29%) yielded strong evidence supporting tools to assess communication skills, cultural competence, empathy/compassion, behavioral health counseling, professionalism, and teamwork. Sixty-two (54%) articles yielded moderate evidence and 19 (17%) weak evidence. Articles mapped to all LCME standards and ACGME core competencies; the most common was communication skills. These findings serve as a valuable resource for medical educators and researchers. More rigorous measurement validation and testing and more robust study designs are needed to understand how educational strategies contribute to BSS competency development.
Tools to Assess Behavioral and Social Science Competencies in Medical Education: A Systematic Review
Carney, Patricia A.; Palmer, Ryan T.; Miller, Marissa Fuqua; Thayer, Erin K.; Estroff, Sue E.; Litzelman, Debra K.; Biagioli, Frances E.; Teal, Cayla R.; Lambros, Ann; Hatt, William J.; Satterfield, Jason M.
2015-01-01
Purpose Behavioral and social science (BSS) competencies are needed to provide quality health care, but psychometrically validated measures to assess these competencies are difficult to find. Moreover, they have not been mapped to existing frameworks, like those from the Liaison Committee on Medical Education (LCME) and Accreditation Council for Graduate Medical Education (ACGME). This systematic review aimed to identify and evaluate the quality of assessment tools used to measure BSS competencies. Method The authors searched the literature published between January 2002 and March 2014 for articles reporting psychometric or other validity/reliability testing, using OVID, CINAHL, PubMed, ERIC, Research and Development Resource Base, SOCIOFILE, and PsycINFO. They reviewed 5,104 potentially relevant titles and abstracts. To guide their review, they mapped BSS competencies to existing LCME and ACGME frameworks. The final, included articles fell into three categories: instrument development, which were of the highest quality; educational research, which were of the second highest quality; and curriculum evaluation, which were of lower quality. Results Of the 114 included articles, 33 (29%) yielded strong evidence supporting tools to assess communication skills, cultural competence, empathy/compassion, behavioral health counseling, professionalism, and teamwork. Sixty-two (54%) articles yielded moderate evidence and 19 (17%) weak evidence. Articles mapped to all LCME standards and ACGME core competencies; the most common was communication skills. Conclusions These findings serve as a valuable resource for medical educators and researchers. More rigorous measurement validation and testing and more robust study designs are needed to understand how educational strategies contribute to BSS competency development. PMID:26796091
Modelling dynamic changes in blood flow and volume in the cerebral vasculature.
Payne, S J; El-Bouri, W K
2018-08-01
The cerebral microvasculature plays a key role in the transport of blood and the delivery of nutrients to the cells that perform brain function. Although recent advances in experimental imaging techniques mean that its structure and function can be interrogated to very small length scales, allowing individual vessels to be mapped to a fraction of 1 μm, these techniques currently remain confined to animal models. In-vivo human data can only be obtained at a much coarser length scale, of order 1 mm, meaning that mathematical models of the microvasculature play a key role in interpreting flow and metabolism data. However, there are close to 10,000 vessels even within a single voxel of size 1 mm 3 . Given the number of vessels present within a typical voxel and the complexity of the governing equations for flow and volume changes, it is computationally challenging to solve these in full, particularly when considering dynamic changes, such as those found in response to neural activation. We thus consider here the governing equations and some of the simplifications that have been proposed in order more rigorously to justify in what generations of blood vessels these approximations are valid. We show that two approximations (neglecting the advection term and assuming a quasi-steady state solution for blood volume) can be applied throughout the cerebral vasculature and that two further approximations (a simple first order differential relationship between inlet and outlet flows and inlet and outlet pressures, and matching of static pressure at nodes) can be applied in vessels smaller than approximately 1 mm in diameter. We then show how these results can be applied in solving flow fields within cerebral vascular networks providing a simplified yet rigorous approach to solving dynamic flow fields and compare the results to those obtained with alternative approaches. We thus provide a framework to model cerebral blood flow and volume within the cerebral vasculature that can be used, particularly at sub human imaging length scales, to provide greater insight into the behaviour of blood flow and volume in the cerebral vasculature. Copyright © 2018 Elsevier Inc. All rights reserved.
Marr-Lyon, Lisa R; Gupchup, Gireesh V; Anderson, Joe R
2012-01-01
The Purdue Pharmacist Directive Guidance (PPDG) Scale was developed to assess patients' perceptions of the level of pharmacist-provided (1) instruction and (2) feedback and goal-setting-2 aspects of pharmaceutical care. Calculations of its psychometric properties stemming from SPSS and R were similar, but distinct differences were apparent. Using SPSS and R software packages, researchers aimed to examine the construct validity of the PPDG using a higher order factoring procedure; in tandem, McDonald's omega and Cronbach's alpha were calculated as means of reliability analyses. Ninety-nine patients with either type I or type II diabetes, aged 18 years or older, able to read and write English, and who could provide written-informed consent participated in the study. Data were collected in 8 community pharmacies in New Mexico. Using R, (1) a principal axis factor analysis with promax (oblique) rotation was conducted, (2) a Schmid-Leiman transformation was attained, and (3) McDonald's omega and Cronbach's alpha were computed. Using SPSS, subscale findings were validated by conducting a principal axis factor analysis with promax rotation; strict parallels and Cronbach's alpha reliabilities were calculated. McDonald's omega and Cronbach's alpha were robust, with coefficients greater than 0.90; principal axis factor analysis with promax rotation revealed construct similarities with an overall general factor emerging from R. Further subjecting the PPDG to rigorous psychometric testing revealed stronger quantitative support of the overall general factor of directive guidance and subscales of instruction and feedback and goal-setting. Copyright © 2012 Elsevier Inc. All rights reserved.
Smith, Madison B; Macieira, Tamara G R; Bumbach, Michael D; Garbutt, Susan J; Citty, Sandra W; Stephen, Anita; Ansell, Margaret; Glover, Toni L; Keenan, Gail
2018-01-01
To present the findings of a systematic review on the use of simulation-based learning experiences (SBLEs) to teach communication skills to nursing students and clinicians who provide palliative and end-of-life care to patients and their families. Palliative care communication skills are fundamental to providing holistic patient care. Since nurses have the greatest amount of direct exposure to patients, building such communication competencies is essential. However, exposure to patients and families receiving palliative and end-of-life care is often limited, resulting in few opportunities to learn these skills in the clinical setting. Simulation-based learning experiences can be used to supplement didactic teaching and clinical experiences to build the requisite communication skills. Searches of CINAHL, MEDLINE, PsychINFO, ERIC, and Web of Science electronic databases and Grey Literature returned 442 unique records. Thirty articles met the established criteria, including the SBLE must contain a nursing role. Simulation-based learning experience are being used to teach palliative and end-of-life communication skills to nursing students and clinicians. Lack of standardization, poor evaluation methods, and limited exposure to the entire interprofessional team makes it difficult to identify and disseminate validated best practices. While the need for further research is acknowledged, we recommend this evidence be augmented by training programs that utilize SBLEs through (1) applying standards, (2) clearly specifying goals and objectives, (3) integrating externally validated scenarios, and (4) employing rigorous evaluation methods and measures that link the SBLE to the training objectives and desired clinician practice behaviors and patient outcomes.
NASA Astrophysics Data System (ADS)
Parilla, Philip A.; Gross, Karl; Hurst, Katherine; Gennett, Thomas
2016-03-01
The ultimate goal of the hydrogen economy is the development of hydrogen storage systems that meet or exceed the US DOE's goals for onboard storage in hydrogen-powered vehicles. In order to develop new materials to meet these goals, it is extremely critical to accurately, uniformly and precisely measure materials' properties relevant to the specific goals. Without this assurance, such measurements are not reliable and, therefore, do not provide a benefit toward the work at hand. In particular, capacity measurements for hydrogen storage materials must be based on valid and accurate results to ensure proper identification of promising materials for further development. Volumetric capacity determinations are becoming increasingly important for identifying promising materials, yet there exists controversy on how such determinations are made and whether such determinations are valid due to differing methodologies to count the hydrogen content. These issues are discussed herein, and we show mathematically that capacity determinations can be made rigorously and unambiguously if the constituent volumes are well defined and measurable in practice. It is widely accepted that this occurs for excess capacity determinations and we show here that this can happen for the total capacity determination. Because the adsorption volume is undefined, the absolute capacity determination remains imprecise. Furthermore, we show that there is a direct relationship between determining the respective capacities and the calibration constants used for the manometric and gravimetric techniques. Several suggested volumetric capacity figure-of-merits are defined, discussed and reporting requirements recommended. Finally, an example is provided to illustrate these protocols and concepts.
Implications of chronic kidney disease for dietary treatment in cardiovascular disease.
Packard, Diane P; Milton, Joan E; Shuler, Lynn A; Short, Robert A; Tuttle, Katherine R
2006-07-01
Chronic kidney disease (CKD) often accompanies cardiovascular disease (CVD). Trends foretelling a greater burden of CKD and CVD are largely a result of increasing frequencies of obesity, hypertension, and diabetes. Nutritional therapy occupies a critical role in reducing risk factors and preventing progressive damage to the kidneys and heart. Nutritional assessment and treatment should take into account both health concerns. This review examines several diet components and eating styles for efficacy in the treatment of these conditions. A variety of dietary regimens claim to provide health benefits, but rigorous scientific validation of long-term efficacy is frequently lacking. An urgent need exists for eating styles that reduce risk of chronic diseases and that are acceptable and achievable in free-living populations. We describe our ongoing study, a randomized controlled trial comparing the American Heart Association Step II diet and a Mediterranean diet, in survivors of a first myocardial infarction. The primary end point is a composite of mortality and major CVD events. Because many in this population have CKD, indicators of kidney damage and function are prespecified secondary end points. Results of this trial should provide insight into optimal dietary interventions for persons with both CVD and CKD.