ERIC Educational Resources Information Center
Applied Management Sciences, Inc., Silver Spring, MD.
A Pre-Award Validation Analysis was conducted in 1978-1979 to provide the federal government information about the accuracy of data provided by applicants for Basic Educational Opportunity Grants. New procedures involved: validation of selected applications by college financial aid officers using documentation such as Federal Income Tax forms;…
Multinational Validation of the Spanish Bracken Basic Concept Scale for Cross-Cultural Assessments.
ERIC Educational Resources Information Center
Bracken, Bruce A.; And Others
1990-01-01
Investigated construct validity of the Spanish translation of the Bracken Basic Concept Scale (BBCS) in Latino children (n=293) including monolingual Spanish-speaking children from Puerto Rico and Venezuela and Spanish-dominant bilingual Latino children from Texas. Results provided support for construct validity of the Spanish version of the…
Kawaguchi, Koji; Egi, Hiroyuki; Hattori, Minoru; Sawada, Hiroyuki; Suzuki, Takahisa; Ohdan, Hideki
2014-10-01
Virtual reality surgical simulators are becoming popular as a means of providing trainees with an opportunity to practice laparoscopic skills. The Lap-X (Epona Medical, Rotterdam, the Netherlands) is a novel VR simulator for training basic skills in laparoscopic surgery. The objective of this study was to validate the LAP-X laparoscopic virtual reality simulator by assessing the face and construct validity in order to determine whether the simulator is adequate for basic skills training. The face and content validity were evaluated using a structured questionnaire. To assess the construct validity, the participants, nine expert surgeons (median age: 40 (32-45)) (>100 laparoscopic procedures) and 11 novices performed three basic laparoscopic tasks using the Lap-X. The participants reported a high level of content validity. No significant differences were found between the expert surgeons and the novices (Ps > 0.246). The performance of the expert surgeons on the three tasks was significantly better than that of the novices in all parameters (Ps < 0.05). This study demonstrated the face, content and construct validity of the Lap-X. The Lap-X holds real potential as a home and hospital training device.
ERIC Educational Resources Information Center
Walters, Pamela Barnhouse; And Others
The validity of information provided in applications to the Basic Educational Opportunity Grant (BEOG) program was studied, along with procedures used in recovering overpayments. The objective was to investigate misreporting and misuse of the BEOG program. A 1977-1978 study reviewed cases referred to the U.S. Office of Education by educational…
Validation of the Basic Psychological Needs in Exercise Scale in a Portuguese sample.
Moutão, João Miguel Raimundo Peres; Serra, Luis Filipe Cid; Alves, José Augusto Marinho; Leitão, José Carlos; Vlachopoulos, Symeon P
2012-03-01
In line with self-determination theory (SDT: Deci & Ryan, 1985, 2002) the satisfaction of the basic psychological needs for autonomy, competence, and relatedness has been identified as an important predictor of behavior and optimal functioning in various contexts including exercise. The lack of a valid and reliable instrument to assess the extent to which these needs are fulfilled among Portuguese exercise participants limits the evaluation of causal links proposed by SDT in the Portuguese exercise context. The aim of the present study was to translate into Portuguese and validate the Basic Psychological Needs in Exercise Scale (BPNES: Vlachopoulos & Michailidou, 2006). Using data from 522 exercise participants the findings provided evidence of strong internal consistency of the translated BPNES subscales while confirmatory factor analysis supported a good fit of the correlated 3-factor model to the data. The present findings support the use of the translated into Portuguese BPNES to assess the extent of basic psychological need fulfilment among Portuguese exercise participants.
The new Guidance in AQUATOX Setup and Application provides a quick start guide to introduce major model features, as well as being a type of cookbook to guide basic model setup, calibration, and validation.
Validation study and routine control monitoring of moist heat sterilization procedures.
Shintani, Hideharu
2012-06-01
The proposed approach to validation of steam sterilization in autoclaves follows the basic life cycle concepts applicable to all validation programs. Understand the function of sterilization process, develop and understand the cycles to carry out the process, and define a suitable test or series of tests to confirm that the function of the process is suitably ensured by the structure provided. Sterilization of product and components and parts that come in direct contact with sterilized product is the most critical of pharmaceutical processes. Consequently, this process requires a most rigorous and detailed approach to validation. An understanding of the process requires a basic understanding of microbial death, the parameters that facilitate that death, the accepted definition of sterility, and the relationship between the definition and sterilization parameters. Autoclaves and support systems need to be designed, installed, and qualified in a manner that ensures their continued reliability. Lastly, the test program must be complete and definitive. In this paper, in addition to validation study, documentation of IQ, OQ and PQ concretely were described.
ERIC Educational Resources Information Center
Johnson, Teresa R.; Khalil, Mohammed K.; Peppler, Richard D.; Davey, Diane D.; Kibble, Jonathan D.
2014-01-01
In the present study, we describe the innovative use of the National Board of Medical Examiners (NBME) Comprehensive Basic Science Examination (CBSE) as a progress test during the preclerkship medical curriculum. The main aim of this study was to provide external validation of internally developed multiple-choice assessments in a new medical…
Clinical prognostic rules for severe acute respiratory syndrome in low- and high-resource settings.
Cowling, Benjamin J; Muller, Matthew P; Wong, Irene O L; Ho, Lai-Ming; Lo, Su-Vui; Tsang, Thomas; Lam, Tai Hing; Louie, Marie; Leung, Gabriel M
2006-07-24
An accurate prognostic model for patients with severe acute respiratory syndrome (SARS) could provide a practical clinical decision aid. We developed and validated prognostic rules for both high- and low-resource settings based on data available at the time of admission. We analyzed data on all 1755 and 291 patients with SARS in Hong Kong (derivation cohort) and Toronto (validation cohort), respectively, using a multivariable logistic scoring method with internal and external validation. Scores were assigned on the basis of patient history in a basic model, and a full model additionally incorporated radiological and laboratory results. The main outcome measure was death. Predictors for mortality in the basic model included older age, male sex, and the presence of comorbid conditions. Additional predictors in the full model included haziness or infiltrates on chest radiography, less than 95% oxygen saturation on room air, high lactate dehydrogenase level, and high neutrophil and low platelet counts. The basic model had an area under the receiver operating characteristic (ROC) curve of 0.860 in the derivation cohort, which was maintained on external validation with an area under the ROC curve of 0.882. The full model improved discrimination with areas under the ROC curve of 0.877 and 0.892 in the derivation and validation cohorts, respectively. The model performs well and could be useful in assessing prognosis for patients who are infected with re-emergent SARS.
Fransson, Boel A; Chen, Chi-Ya; Noyes, Julie A; Ragle, Claude A
2016-11-01
To determine the construct and concurrent validity of instrument motion metrics for laparoscopic skills assessment in virtual reality and augmented reality simulators. Evaluation study. Veterinarian students (novice, n = 14) and veterinarians (experienced, n = 11) with no or variable laparoscopic experience. Participants' minimally invasive surgery (MIS) experience was determined by hospital records of MIS procedures performed in the Teaching Hospital. Basic laparoscopic skills were assessed by 5 tasks using a physical box trainer. Each participant completed 2 tasks for assessments in each type of simulator (virtual reality: bowel handling and cutting; augmented reality: object positioning and a pericardial window model). Motion metrics such as instrument path length, angle or drift, and economy of motion of each simulator were recorded. None of the motion metrics in a virtual reality simulator showed correlation with experience, or to the basic laparoscopic skills score. All metrics in augmented reality were significantly correlated with experience (time, instrument path, and economy of movement), except for the hand dominance metric. The basic laparoscopic skills score was correlated to all performance metrics in augmented reality. The augmented reality motion metrics differed between American College of Veterinary Surgeons diplomates and residents, whereas basic laparoscopic skills score and virtual reality metrics did not. Our results provide construct validity and concurrent validity for motion analysis metrics for an augmented reality system, whereas a virtual reality system was validated only for the time score. © Copyright 2016 by The American College of Veterinary Surgeons.
Observing the sick child: part 2b. Respiratory palpation.
Aylott, Marion
2007-02-01
Assessment is a major nursing role, and expanding assessment techniques, traditionally seen as the remit of the medical profession, can assist nursing assessment and the provision of appropriate care. This is the third article in a series of five articles. Parts 1 and 2a provided a practical critical review of the validity and reliability of basic respiratory assessment focusing on measurement of respiratory rate, rhythm and depth. This article provides a practical step-by-step introduction to the theory and practice of advanced respiratory assessment using palpation. Next month we will build on these skills and provide a practical step-by-step introduction to using auscultation in order to augment basic respiratory assessment skills.
ERIC Educational Resources Information Center
Smith, Karen; And Others
Procedures for validating data reported by students and parents on an application for Basic Educational Opportunity Grants were developed in 1978 for the U.S. Office of Education (OE). Validation activities include: validation of flagged Student Eligibility Reports (SERs) for students whose schools are part of the Alternate Disbursement System;…
Development and Validation of a Project Package for Junior Secondary School Basic Science
ERIC Educational Resources Information Center
Udofia, Nsikak-Abasi
2014-01-01
This was a Research and Developmental study designed to develop and validate projects for Junior Secondary School Basic Science instruction and evaluation. The projects were developed using the project blueprint and sent for validation by experts in science education and measurement and evaluation; using a project validation scale. They were to…
Ferrer, Rebecca A.; McDonald, Paige Green; Barrett, Lisa Feldman
2015-01-01
Cancer control research involves the conduct of basic and applied behavioral and social sciences to reduce cancer incidence, morbidity, and mortality, and improve quality of life. Given the importance of behavior in cancer control, fundamental research is necessary to identify psychological mechanisms underlying cancer risk, prevention, and management behaviors. Cancer prevention, diagnosis, and treatment are often emotionally-laden. As such, affective science research to elucidate questions related to basic phenomenological nature of emotion, stress, and mood is necessary to understand how cancer control can be hindered or facilitated by emotional experiences. To date, the intersection of basic affective science research and cancer control remains largely unexplored. The goal of this paper is to outline key questions in the cancer control research domain that provide an ecologically valid context for new affective science discoveries. We also provide examples of ways in which basic affective discoveries could inform future cancer prevention and control research. These examples are not meant to be exhaustive or prescriptive, but instead are offered to generate creative thought about the promise of a cancer research context for answering basic affective science questions. Together, these examples provide a compelling argument for fostering collaborations between affective and cancer control scientists. PMID:25987511
Czerniecki, Joseph M; Turner, Aaron P; Williams, Rhonda M; Thompson, Mary Lou; Landry, Greg; Hakimi, Kevin; Speckman, Rebecca; Norvell, Daniel C
2017-01-01
The objective of this study was the development of AMPREDICT-Mobility, a tool to predict the probability of independence in either basic or advanced (iBASIC or iADVANCED) mobility 1 year after dysvascular major lower extremity amputation. Two prospective cohort studies during consecutive 4-year periods (2005-2009 and 2010-2014) were conducted at seven medical centers. Multiple demographic and biopsychosocial predictors were collected in the periamputation period among individuals undergoing their first major amputation because of complications of peripheral arterial disease or diabetes. The primary outcomes were iBASIC and iADVANCED mobility, as measured by the Locomotor Capabilities Index. Combined data from both studies were used for model development and internal validation. Backwards stepwise logistic regression was used to develop the final prediction models. The discrimination and calibration of each model were assessed. Internal validity of each model was assessed with bootstrap sampling. Twelve-month follow-up was reached by 157 of 200 (79%) participants. Among these, 54 (34%) did not achieve iBASIC mobility, 103 (66%) achieved at least iBASIC mobility, and 51 (32%) also achieved iADVANCED mobility. Predictive factors associated with reduced odds of achieving iBASIC mobility were increasing age, chronic obstructive pulmonary disease, dialysis, diabetes, prior history of treatment for depression or anxiety, and very poor to fair self-rated health. Those who were white, were married, and had at least a high-school degree had a higher probability of achieving iBASIC mobility. The odds of achieving iBASIC mobility increased with increasing body mass index up to 30 kg/m 2 and decreased with increasing body mass index thereafter. The prediction model of iADVANCED mobility included the same predictors with the exception of diabetes, chronic obstructive pulmonary disease, and education level. Both models showed strong discrimination with C statistics of 0.85 and 0.82, respectively. The mean difference in predicted probabilities for those who did and did not achieve iBASIC and iADVANCED mobility was 33% and 29%, respectively. Tests for calibration and observed vs predicted plots suggested good fit for both models; however, the precision of the estimates of the predicted probabilities was modest. Internal validation through bootstrapping demonstrated some overoptimism of the original model development, with the optimism-adjusted C statistic for iBASIC and iADVANCED mobility being 0.74 and 0.71, respectively, and the discrimination slope 19% and 16%, respectively. AMPREDICT-Mobility is a user-friendly prediction tool that can inform the patient undergoing a dysvascular amputation and the patient's provider about the probability of independence in either basic or advanced mobility at each major lower extremity amputation level. Copyright © 2016 Society for Vascular Surgery. All rights reserved.
Standardized Radiation Shield Design Methods: 2005 HZETRN
NASA Technical Reports Server (NTRS)
Wilson, John W.; Tripathi, Ram K.; Badavi, Francis F.; Cucinotta, Francis A.
2006-01-01
Research committed by the Langley Research Center through 1995 resulting in the HZETRN code provides the current basis for shield design methods according to NASA STD-3000 (2005). With this new prominence, the database, basic numerical procedures, and algorithms are being re-examined with new methods of verification and validation being implemented to capture a well defined algorithm for engineering design processes to be used in this early development phase of the Bush initiative. This process provides the methodology to transform the 1995 HZETRN research code into the 2005 HZETRN engineering code to be available for these early design processes. In this paper, we will review the basic derivations including new corrections to the codes to insure improved numerical stability and provide benchmarks for code verification.
NASA Astrophysics Data System (ADS)
Hidayati, A.; Rahmi, A.; Yohandri; Ratnawulan
2018-04-01
The importance of teaching materials in accordance with the characteristics of students became the main reason for the development of basic electronics I module integrated character values based on conceptual change teaching model. The module development in this research follows the development procedure of Plomp which includes preliminary research, prototyping phase and assessment phase. In the first year of this research, the module is validated. Content validity is seen from the conformity of the module with the development theory in accordance with the demands of learning model characteristics. The validity of the construct is seen from the linkage and consistency of each module component developed with the characteristic of the integrated learning model of character values obtained through validator assessment. The average validation value assessed by the validator belongs to a very valid category. Based on the validator assessment then revised the basic electronics I module integrated character values based on conceptual change teaching model.
ERIC Educational Resources Information Center
Applied Management Sciences, Inc., Silver Spring, MD.
The 1978-1979 pre-award institution validation process for the Basic Educational Opportunity Grant (BEOG) program was studied, based on applicant and grant recipient files as of the end of February 1979. The objective was to assess the impact of the validation process on the proper award of BEOGs, and to determine whether the criteria for…
Schmetz, Emilie; Rousselle, Laurence; Ballaz, Cécile; Detraux, Jean-Jacques; Barisnikov, Koviljka
2017-06-20
This study aims to examine the different levels of visual perceptual object recognition (early, intermediate, and late) defined in Humphreys and Riddoch's model as well as basic visual spatial processing in children using a new test battery (BEVPS). It focuses on the age sensitivity, internal coherence, theoretical validity, and convergent validity of this battery. French-speaking, typically developing children (n = 179; 5 to 14 years) were assessed using 15 new computerized subtests. After selecting the most age-sensitive tasks though ceiling effect and correlation analyses, an exploratory factorial analysis was run with the 12 remaining subtests to examine the BEVPS' theoretical validity. Three separate factors were identified for the assessment of the stimuli's basic features (F1, four subtests), view-dependent and -independent object representations (F2, six subtests), and basic visual spatial processing (F3, two subtests). Convergent validity analyses revealed positive correlations between F1 and F2 and the Beery-VMI visual perception subtest, while no such correlations were found for F3. Children's performances progressed until the age of 9-10 years in F1 and in view-independent representations (F2), and until 11-12 years in view-dependent representations (F2). However, no progression with age was observed in F3. Moreover, the selected subtests, present good-to-excellent internal consistency, which indicates that they provide reliable measures for the assessment of visual perceptual processing abilities in children.
Corrections Analysis Report: Phase II.
ERIC Educational Resources Information Center
Applied Management Sciences, Inc., Silver Spring, MD.
The validity of data provided through rejection edits of students applying for Basic Educational Opportunity Grants (BEOGs) was assessed. The question of whether some edits were associated with applicants' failure to reapply was also considered. The analysis used the 1978-1979 BEOG applicant population. A total of 3,823,008 applicants had applied…
Vocational Education and Training in Denmark: Short Description. CEDEFOP Panorama Series.
ERIC Educational Resources Information Center
Cort, Pia
Denmark has a uniform, nationwide vocational education and training (VET) system that provides qualifications that are valid throughout the country and recognized by employers and trade unions. Initial VET (IVET) includes the following components: VET, including commercial and technical training; basic social and health care training; agricultural…
Ferrer, Rebecca A; Green, Paige A; Barrett, Lisa Feldman
2015-05-01
Cancer control research involves the conduct of basic and applied behavioral and social sciences to reduce cancer incidence, morbidity, and mortality and improve quality of life. Given the importance of behavior in cancer control, fundamental research is necessary to identify psychological mechanisms underlying cancer risk, prevention, and management behaviors. Cancer prevention, diagnosis, and treatment are often emotionally laden. As such, affective science research to elucidate questions related to the basic phenomenological nature of emotion, stress, and mood is necessary to understand how cancer control can be hindered or facilitated by emotional experiences. To date, the intersection of basic affective science research and cancer control remains largely unexplored. The goal of this article is to outline key questions in the cancer control research domain that provide an ecologically valid context for new affective science discoveries. We also provide examples of ways in which basic affective discoveries could inform future cancer prevention and control research. These examples are not meant to be exhaustive or prescriptive but instead are offered to generate creative thought about the promise of a cancer research context for answering basic affective science questions. Together, these examples provide a compelling argument for fostering collaborations between affective and cancer control scientists. © The Author(s) 2015.
Validation, Edits, and Application Processing System Report: Phase I.
ERIC Educational Resources Information Center
Gray, Susan; And Others
Findings of phase 1 of a study of the 1979-1980 Basic Educational Opportunity Grants validation, edits, and application processing system are presented. The study was designed to: assess the impact of the validation effort and processing system edits on the correct award of Basic Grants; and assess the characteristics of students most likely to…
Johannesen, Jason K; Lurie, Jessica B; Fiszdon, Joanna M; Bell, Morris D
2013-01-01
The Social Attribution Task-Multiple Choice (SAT-MC) uses a 64-second video of geometric shapes set in motion to portray themes of social relatedness and intentions. Considered a test of "Theory of Mind," the SAT-MC assesses implicit social attribution formation while reducing verbal and basic cognitive demands required of other common measures. We present a comparability analysis of the SAT-MC and the new SAT-MC-II, an alternate form created for repeat testing, in a university sample (n = 92). Score distributions and patterns of association with external validation measures were nearly identical between the two forms, with convergent and discriminant validity supported by association with affect recognition ability and lack of association with basic visual reasoning. Internal consistency of the SAT-MC-II was superior (alpha = .81) to the SAT-MC (alpha = .56). Results support the use of SAT-MC and new SAT-MC-II as equivalent test forms. Demonstrating relatively higher association to social cognitive than basic cognitive abilities, the SAT-MC may provide enhanced sensitivity as an outcome measure of social cognitive intervention trials.
Bidell, Markus P
2017-01-01
These three studies provide initial evidence for the development, factor structure, reliability, and validity of the Lesbian, Gay, Bisexual, and Transgender Development of Clinical Skills Scale (LGBT-DOCSS), a new interdisciplinary LGBT clinical self-assessment for health and mental health providers. Research participants were voluntarily recruited in the United States and United Kingdom and included trainees, clinicians, and educators from applied psychology, counseling, psychotherapy, and primary care medicine. Study 1 (N = 602) used exploratory and confirmatory factor analytic techniques, revealing an 18-item three-factor structure (Clinical Preparedness, Attitudinal Awareness, and Basic Knowledge). Study 2 established internal consistency for the overall LGBT-DOCSS (α = .86) and for each of the three subscales (Clinical Preparedness = .88, Attitudinal Awareness = .80, and Basic Knowledge = .83) and 2-week test-retest reliability (.87). In study 3 (N = 564), participant criteria (sexual orientation and education level) and four established scales that measured LGBT prejudice, assessment skills, and social desirability were used to support initial content and discriminant validity. Psychometric properties, limitations, and recommendations are discussed.
Moreno-Murcia, Juan A; Martínez-Galindo, Celestina; Moreno-Pérez, Víctor; Marcos, Pablo J.; Borges, Fernanda
2012-01-01
This study aimed to cross-validate the psychometric properties of the Basic Psychological Needs in Exercise Scale (BPNES) by Vlachopoulos and Michailidou, 2006 in a Spanish context. Two studies were conducted. Confirmatory factor analysis results confirmed the hypothesized three-factor solution In addition, we documented evidence of reliability, analysed as internal consistency and temporal stability. Future studies should analyse the scale's validity and reliability with different populations and check their experimental effect. Key pointsThe Basic Psychological Needs in Exercise Scale (BPNES) is valid and reliable for measuring basic psychological needs in healthy physical exercise in the Spanish context.The factor structure of three correlated factors has shown minimal invariance across gender. PMID:24149130
Rebuilding health systems in post-conflict countries: estimating the costs of basic services.
Newbrander, William; Yoder, Richard; Debevoise, Anne Bilby
2007-01-01
After the fall of the Taliban in 2001, the Afghan transitional government and international donors found the health system near collapse. Afghanistan had some of the worst health indicators ever recorded. To begin activities that would quickly improve the health situation, the Ministry of Health (MOH) needed both a national package of health services and reliable data on the costs of providing those services. This study details the process of determining national health priorities, creating a basic package of services, and estimating per capita and unit costs for providing those services, with an emphasis on the costing exercise. Strategies for obtaining a rapid yet reasonably accurate estimate of health service costs nationwide are discussed. In 2002 this costing exercise indicated that the basic package of services could be provided for US dollars 4.55 per person. In 2006, the findings were validated: the four major donors who contracted with non-governmental organizations (NGOs) to provide basic health services for nearly 80% of the population found per capita costs ranging from dollars 4.30 to dollars 5.12. This study is relevant for other post-conflict countries that are re-establishing health services and seeking to develop cost-effective and equitable health systems. Copyright (c) 2007 John Wiley & Sons, Ltd.
Climatological Processing and Product Development for the TRMM Ground Validation Program
NASA Technical Reports Server (NTRS)
Marks, D. A.; Kulie, M. S.; Robinson, M.; Silberstein, D. S.; Wolff, D. B.; Ferrier, B. S.; Amitai, E.; Fisher, B.; Wang, J.; Augustine, D.;
2000-01-01
The Tropical Rainfall Measuring Mission (TRMM) satellite was successfully launched in November 1997.The main purpose of TRMM is to sample tropical rainfall using the first active spaceborne precipitation radar. To validate TRMM satellite observations, a comprehensive Ground Validation (GV) Program has been implemented. The primary goal of TRMM GV is to provide basic validation of satellite-derived precipitation measurements over monthly climatologies for the following primary sites: Melbourne, FL; Houston, TX; Darwin, Australia- and Kwajalein Atoll, RMI As part of the TRMM GV effort, research analysts at NASA Goddard Space Flight Center (GSFC) generate standardized rainfall products using quality-controlled ground-based radar data from the four primary GV sites. This presentation will provide an overview of TRMM GV climatological processing and product generation. A description of the data flow between the primary GV sites, NASA GSFC, and the TRMM Science and Data Information System (TSDIS) will be presented. The radar quality control algorithm, which features eight adjustable height and reflectivity parameters, and its effect on monthly rainfall maps, will be described. The methodology used to create monthly, gauge-adjusted rainfall products for each primary site will also be summarized. The standardized monthly rainfall products are developed in discrete, modular steps with distinct intermediate products. A summary of recently reprocessed official GV rainfall products available for TRMM science users will be presented. Updated basic standardized product results involving monthly accumulation, Z-R relationship, and gauge statistics for each primary GV site will also be displayed.
Pedrosa, Linda Délia C O; Sarinho, Silvia W; Ordonha, Manoelina R
2007-10-01
Analysis of the quality of information on basic causes of neonatal deaths in Brazil is crucially important, since it allows one to estimate how many deaths are avoidable and provide support for policies to decrease neonatal mortality. The current study aimed to evaluate the reliability and validity of the Mortality Information System (MIS) for discriminating between basic causes of neonatal deaths and defining percentages of reducible causes. The basic causes of early neonatal deaths in hospitals in Maceió, Alagoas State, were analyzed, and the causes recorded in medical records were compared to the MIS data in order to measure reliability and validity. The modified SEADE Foundation and Wigglesworth classifications were compared to analyze the capacity for reduction of neonatal mortality. Maternal causes predominated in the medical records, as compared to respiratory disorders on the death certificates and in the MIS. The percentage of avoidable deaths may be much higher than observed from the MIS, due to imprecision in completing death certificates. Based on the MIS, the greatest problems are in early diagnosis and treatment of neonatal causes. However, the results show that the most pressing problems relate to failures in prenatal care and lack of control of diseases.
Indian Manpower Resources in the Southwest: A Pilot Study.
ERIC Educational Resources Information Center
Taylor, Benjamin J.; O'Connor, Dennis J.
The Indian Manpower Resource Study (IMRS) was undertaken to provide basic information essential to the planning and development of effective programs to make use of the manpower resource of reservation Indians in Arizona and New Mexico. In the past, decision making has largely been undertaken with a minimum of valid information. The reservations…
Relationship of Elementary and Secondary School Achievement Test Scores to Later Academic Success.
ERIC Educational Resources Information Center
Loyd, Brenda H.; And Others
1980-01-01
This study investigated the relationship between achievement test scores on the Iowa Tests of Basic Skills (ITBS) and Iowa Tests of Educational Development (ITED), and high school and college grade point average. Support for the predictive validity of the ITBS and ITED achievement test batteries is provided. (Author/GK)
An experimental study of the validity of the heat-field concept for sonic-boom alleviation
NASA Technical Reports Server (NTRS)
Swigart, R. J.
1974-01-01
An experimental program was carried out in the NASA-Langley 4 ft x 4 ft supersonic pressure tunnel to investigate the validity of the heat-field concept for sonic boom alleviation. The concept involves heating the flow about a supersonic aircraft in such a manner as to obtain an increase in effective aircraft length and yield an effective aircraft shape that will result in a shock-free pressure signature on the ground. First, a basic body-of-revolution representing an SST configuration with its lift equivalence in volume was tested to provide a baseline pressure signature. Second, a model having a 5/2-power area distribution which, according to theory, should yield a linear pressure rise with no front shock wave was tested. Third, the concept of providing the 5/2-power area distribution by using an off-axis slender fin below the basic body was investigated. Then a substantial portion (approximately 40 percent) of the solid fin was replaced by a heat field generated by passing heated nitrogen through the rear of the fin.
The Basic Psychological Needs at Work Scale: Measurement Invariance between Canada and France.
Brien, Maryse; Forest, Jacques; Mageau, Geneviève A; Boudrias, Jean-Sébastien; Desrumaux, Pascale; Brunet, Luc; Morin, Estelle M
2012-07-01
The purpose of this study is to develop and validate the Basic Psychological Needs at Work Scale (BPNWS) in French, but items are also provided in English in the article. The BPNWS is a work-related self-report instrument designed to measure the degree to which the needs for autonomy, competence, and relatedness, as identified by Self-Determination Theory (Deci & Ryan, 2000), are satisfied at work. Using exploratory and confirmatory factor analysis, the first study examines the structure of the BPNWS in a group of 271 workers. The second study tests the measurement invariance of the scale in a group of 851 teachers from two different cultures, Canada and France. Results support the three-factor structure and show adequate internal consistency, as well as nomological validity across samples. © 2012 The Authors. Applied Psychology: Health and Well-Being © 2012 The International Association of Applied Psychology.
NASA Technical Reports Server (NTRS)
Noll, Thomas E.; Perry, Boyd, III; Tiffany, Sherwood H.; Cole, Stanley R.; Buttrill, Carey S.; Adams, William M., Jr.; Houck, Jacob A.; Srinathkumar, S.; Mukhopadhyay, Vivek; Pototzky, Anthony S.
1989-01-01
The status of the joint NASA/Rockwell Active Flexible Wing Wind-Tunnel Test Program is described. The objectives are to develop and validate the analysis, design, and test methodologies required to apply multifunction active control technology for improving aircraft performance and stability. Major tasks include designing digital multi-input/multi-output flutter-suppression and rolling-maneuver-load alleviation concepts for a flexible full-span wind-tunnel model, obtaining an experimental data base for the basic model and each control concept and providing comparisons between experimental and analytical results to validate the methodologies. The opportunity is provided to improve real-time simulation techniques and to gain practical experience with digital control law implementation procedures.
Pol-Pons Rn Ibclc, Anna; Aubanell-Serra ScB, Mercè; Vidal Rn, Mireia; Ojeda-Ciurana Rn, Imma; Martí-Lluch PhD, Ruth; Ponjoan PhD, Anna
2016-11-01
validated and reliable tools measuring the level of competence about breastfeeding amongst health professionals who attend mothers in primary care are scarce. to develop and validate a self-administered online questionnaire to assess basic competence about breastfeeding in the full range of health professionals attending lactating mothers in primary care. methodological study. the study was conducted in 33 primary care centers administered by the Catalan Health Institute (Institut Català de la Salut) in the Girona Region (northeast Spain). the questionnaire was sent to 398 health professionals who were likely to be in contact with breast feeding mothers. We included midwives, general practitioners, pediatric nurses (who provide care for children from birth to 14 years old), non-pediatric nurses (who provide care for patients older than 14 year), pediatricians, and gynecologists. The random sampling was stratified by discipline to ensure representativeness. the development and validation of the self-administered online questionnaire had five phases: (1) literature review, preparation of a draft for evaluation by an expert panel and pilot study; (2) design of a questionnaire based on observations from Phase 1; (3) questionnaire administration to a random sample of 398 health professionals, stratified by discipline; (4) re-testing after four weeks; and (5) assessment of construct validity using factor analysis and hypothesis-testing, comparing scores between professional groups and estimating effect size. Internal consistency was assessed by Cronbach alpha and reproducibility by the intraclass correlation coefficient. The response rate was 69.3%. The final version of the questionnaire has 24 items. The uni-dimensionality of the questionnaire was confirmed by the factor analysis. The score differed significantly between professional groups (F=12.904; p<0.001), among whom midwives achieved the highest score (129±10 points) and non-pediatric nurses the lowest (111±14 points).The effect sized values ranged from 0.6 and 1.5. Cronbach alpha was 0.870 (95% CI: 0.847, 0.891) and intraclass correlation coefficient was 0.856 (95%CI: 0.817, 0.888). this is a valid and reliable questionnaire to assess a basic level of competence about breast feeding among primary care professionals involved in supporting breast-feeding mothers. This questionnaire may help to identify health professionals who require support to improve their capability to manage women who are breast feeding. Copyright © 2016 The Authors. Published by Elsevier Ltd.. All rights reserved.
Grants4Targets - an innovative approach to translate ideas from basic research into novel drugs.
Lessl, Monika; Schoepe, Stefanie; Sommer, Anette; Schneider, Martin; Asadullah, Khusru
2011-04-01
Collaborations between industry and academia are steadily gaining importance. To combine expertises Bayer Healthcare has set up a novel open innovation approach called Grants4Targets. Ideas on novel drug targets can easily be submitted to http://www.grants4targets.com. After a review process, grants are provided to perform focused experiments to further validate the proposed targets. In addition to financial support specific know-how on target validation and drug discovery is provided. Experienced scientists are nominated as project partners and, depending on the project, tools or specific models are provided. Around 280 applications have been received and 41 projects granted. According to our experience, this type of bridging fund combined with joint efforts provides a valuable tool to foster drug discovery collaborations. Copyright © 2010 Elsevier Ltd. All rights reserved.
Design for validation: An approach to systems validation
NASA Technical Reports Server (NTRS)
Carter, William C.; Dunham, Janet R.; Laprie, Jean-Claude; Williams, Thomas; Howden, William; Smith, Brian; Lewis, Carl M. (Editor)
1989-01-01
Every complex system built is validated in some manner. Computer validation begins with review of the system design. As systems became too complicated for one person to review, validation began to rely on the application of adhoc methods by many individuals. As the cost of the changes mounted and the expense of failure increased, more organized procedures became essential. Attempts at devising and carrying out those procedures showed that validation is indeed a difficult technical problem. The successful transformation of the validation process into a systematic series of formally sound, integrated steps is necessary if the liability inherent in the future digita-system-based avionic and space systems is to be minimized. A suggested framework and timetable for the transformtion are presented. Basic working definitions of two pivotal ideas (validation and system life-cyle) are provided and show how the two concepts interact. Many examples are given of past and present validation activities by NASA and others. A conceptual framework is presented for the validation process. Finally, important areas are listed for ongoing development of the validation process at NASA Langley Research Center.
Openness to Experience as a Basic Dimension of Personality.
ERIC Educational Resources Information Center
McCrae, Robert R.
This paper opens by describing research since 1975 (McCrae and Costa) on a set of related traits that identified as aspects of Openness to Experience. The historic roots of the concept of Openness to Experience are traced. Data are provided on the convergent and discriminant validity of the six Revised NEO-Personality Inventory facets of Fantasy,…
Expert system validation in prolog
NASA Technical Reports Server (NTRS)
Stock, Todd; Stachowitz, Rolf; Chang, Chin-Liang; Combs, Jacqueline
1988-01-01
An overview of the Expert System Validation Assistant (EVA) is being implemented in Prolog at the Lockheed AI Center. Prolog was chosen to facilitate rapid prototyping of the structure and logic checkers and since February 1987, we have implemented code to check for irrelevance, subsumption, duplication, deadends, unreachability, and cycles. The architecture chosen is extremely flexible and expansible, yet concise and complementary with the normal interactive style of Prolog. The foundation of the system is in the connection graph representation. Rules and facts are modeled as nodes in the graph and arcs indicate common patterns between rules. The basic activity of the validation system is then a traversal of the connection graph, searching for various patterns the system recognizes as erroneous. To aid in specifying these patterns, a metalanguage is developed, providing the user with the basic facilities required to reason about the expert system. Using the metalanguage, the user can, for example, give the Prolog inference engine the goal of finding inconsistent conclusions among the rules, and Prolog will search the graph intantiations which can match the definition of inconsistency. Examples of code for some of the checkers are provided and the algorithms explained. Technical highlights include automatic construction of a connection graph, demonstration of the use of metalanguage, the A* algorithm modified to detect all unique cycles, general-purpose stacks in Prolog, and a general-purpose database browser with pattern completion.
Herbort, Maike C.; Iseev, Jenny; Stolz, Christopher; Roeser, Benedict; Großkopf, Nora; Wüstenberg, Torsten; Hellweg, Rainer; Walter, Henrik; Dziobek, Isabel; Schott, Björn H.
2016-01-01
We present the ToMenovela, a stimulus set that has been developed to provide a set of normatively rated socio-emotional stimuli showing varying amount of characters in emotionally laden interactions for experimental investigations of (i) cognitive and (ii) affective Theory of Mind (ToM), (iii) emotional reactivity, and (iv) complex emotion judgment with respect to Ekman’s basic emotions (happiness, anger, disgust, fear, sadness, surprise, Ekman and Friesen, 1975). Stimuli were generated with focus on ecological validity and consist of 190 scenes depicting daily-life situations. Two or more of eight main characters with distinct biographies and personalities are depicted on each scene picture. To obtain an initial evaluation of the stimulus set and to pave the way for future studies in clinical populations, normative data on each stimulus of the set was obtained from a sample of 61 neurologically and psychiatrically healthy participants (31 female, 30 male; mean age 26.74 ± 5.84), including a visual analog scale rating of Ekman’s basic emotions (happiness, anger, disgust, fear, sadness, surprise) and free-text descriptions of the content of each scene. The ToMenovela is being developed to provide standardized material of social scenes that are available to researchers in the study of social cognition. It should facilitate experimental control while keeping ecological validity high. PMID:27994562
Johannesen, Jason K.; Lurie, Jessica B.; Fiszdon, Joanna M.; Bell, Morris D.
2013-01-01
The Social Attribution Task-Multiple Choice (SAT-MC) uses a 64-second video of geometric shapes set in motion to portray themes of social relatedness and intentions. Considered a test of “Theory of Mind,” the SAT-MC assesses implicit social attribution formation while reducing verbal and basic cognitive demands required of other common measures. We present a comparability analysis of the SAT-MC and the new SAT-MC-II, an alternate form created for repeat testing, in a university sample (n = 92). Score distributions and patterns of association with external validation measures were nearly identical between the two forms, with convergent and discriminant validity supported by association with affect recognition ability and lack of association with basic visual reasoning. Internal consistency of the SAT-MC-II was superior (alpha = .81) to the SAT-MC (alpha = .56). Results support the use of SAT-MC and new SAT-MC-II as equivalent test forms. Demonstrating relatively higher association to social cognitive than basic cognitive abilities, the SAT-MC may provide enhanced sensitivity as an outcome measure of social cognitive intervention trials. PMID:23864984
Verhagen, H; Aruoma, O I; van Delft, J H M; Dragsted, L O; Ferguson, L R; Knasmüller, S; Pool-Zobel, B L; Poulsen, H E; Williamson, G; Yannai, S
2003-05-01
There is increasing evidence that chemicals/test substances cannot only have adverse effects, but that there are many substances that can (also) have a beneficial effect on health. As this journal regularly publishes papers in this area and has every intention in continuing to do so in the near future, it has become essential that studies reported in this journal reflect an adequate level of scientific scrutiny. Therefore a set of essential characteristics of studies has been defined. These basic requirements are default properties rather than non-negotiables: deviations are possible and useful, provided they can be justified on scientific grounds. The 10 basic requirements for a scientific paper reporting antioxidant, antimutagenic or anticarcinogenic potential of test substances in in vitro experiments and animal studies in vivo concern the following areas: (1) Hypothesis-driven study design; (2) The nature of the test substance; (3) Valid and invalid test systems; (4) The selection of dose levels and gender; (5) Reversal of the effects induced by oxidants, carcinogens and mutagens; (6) Route of administration; (7) Number and validity of test variables; (8) Repeatability and reproducibility; (9) Statistics; and (10) Quality Assurance.
Tzonev, Svilen
2018-01-01
Current commercially available digital PCR (dPCR) systems and assays are capable of detecting individual target molecules with considerable reliability. As tests are developed and validated for use on clinical samples, the need to understand and develop robust statistical analysis routines increases. This chapter covers the fundamental processes and limitations of detecting and reporting on single molecule detection. We cover the basics of quantification of targets and sources of imprecision. We describe the basic test concepts: sensitivity, specificity, limit of blank, limit of detection, and limit of quantification in the context of dPCR. We provide basic guidelines how to determine those, how to choose and interpret the operating point, and what factors may influence overall test performance in practice.
Validation of the Military Entrance Physical Strength Capacity Test. Technical Report 610.
ERIC Educational Resources Information Center
Myers, David C.; And Others
A battery of physical ability tests was validated using a predictive, criterion-related strategy. The battery was given to 1,003 female soldiers and 980 male soldiers before they had begun Army Basic Training. Criterion measures which represented physical competency in Basic Training (physical proficiency tests, sick call, profiles, and separation…
Spanish Translation and Validation of the Bracken Basic Concept Scale.
ERIC Educational Resources Information Center
Bracken, Bruce A; Fouad, Nadya
1987-01-01
The Bracken Basic Concept Scale (BBCS) was translated into Spanish, and 32 preschool and primary age bilingual children were assessed in a counter-balanced format with the English and newly translated Spanish forms to assess the adequacy of the translation. Preliminary content validity of the Spanish BBCS was demonstrated. (Author/JAZ)
Fracture mechanics life analytical methods verification testing
NASA Technical Reports Server (NTRS)
Favenesi, J. A.; Clemmons, T. G.; Lambert, T. J.
1994-01-01
Verification and validation of the basic information capabilities in NASCRAC has been completed. The basic information includes computation of K versus a, J versus a, and crack opening area versus a. These quantities represent building blocks which NASCRAC uses in its other computations such as fatigue crack life and tearing instability. Several methods were used to verify and validate the basic information capabilities. The simple configurations such as the compact tension specimen and a crack in a finite plate were verified and validated versus handbook solutions for simple loads. For general loads using weight functions, offline integration using standard FORTRAN routines was performed. For more complicated configurations such as corner cracks and semielliptical cracks, NASCRAC solutions were verified and validated versus published results and finite element analyses. A few minor problems were identified in the basic information capabilities of the simple configurations. In the more complicated configurations, significant differences between NASCRAC and reference solutions were observed because NASCRAC calculates its solutions as averaged values across the entire crack front whereas the reference solutions were computed for a single point.
NASA Astrophysics Data System (ADS)
Gurova, E. G.
2016-08-01
During the researches the mathematical description of the traction characteristics of the stiffness compensators of the vibration isolation devices, relatively of the each axis, has been done. Representation of the compensators properties considers the variable load, thereby provide the wide enough spectrum of the action of the suggested vibration isolators. The derived expressions are valid for all three axes of space at the different stiffnesses, i.e. basic basic and two compensating. The research was supported by the scholarships of Russian Federation President for young scientists №184 from 10th of March 2015.
A Validity and Reliability Study of the Basic Electronics Skills Self-Efficacy Scale (BESS)
ERIC Educational Resources Information Center
Korkmaz, Ö.; Korkmaz, M. K.
2016-01-01
The aim of this study is to improve a measurement tool to evaluate the self-efficacy of Electrical-Electronics Engineering students through their basic electronics skills. The sample group is composed of 124 Electrical-Electronics engineering students. The validity of the scale is analyzed with two different methods through factor analysis and…
Confirmatory Factor Analysis of the TerraNova Comprehensive Tests of Basic Skills/5
ERIC Educational Resources Information Center
Stevens, Joseph J.; Zvoch, Keith
2007-01-01
Confirmatory factor analysis was used to explore the internal validity of scores on the TerraNova Comprehensive Tests of Basic Skills/5 using samples from a southwestern school district and standardization samples reported by the publisher. One of the strengths claimed for battery-type achievement tests is provision of reliable and valid samples…
Basic School Skills Inventory-3: Validity and Reliability Study
ERIC Educational Resources Information Center
Yildiz, F. Ülkü; Çagdas, Aysel; Kayili, Gökhan
2017-01-01
The purpose of this study is to perform the validity-reliability analysis of the three subtests of Basic School Skills Inventory 3--Mathematics, Classroom Behavior and Daily Life skills--and do its adaptation for four to six year-old Turkish children. The sample of the study included 595 four to six year-old Turkish children attending public and…
IMAGESEER - IMAGEs for Education and Research
NASA Technical Reports Server (NTRS)
Le Moigne, Jacqueline; Grubb, Thomas; Milner, Barbara
2012-01-01
IMAGESEER is a new Web portal that brings easy access to NASA image data for non-NASA researchers, educators, and students. The IMAGESEER Web site and database are specifically designed to be utilized by the university community, to enable teaching image processing (IP) techniques on NASA data, as well as to provide reference benchmark data to validate new IP algorithms. Along with the data and a Web user interface front-end, basic knowledge of the application domains, benchmark information, and specific NASA IP challenges (or case studies) are provided.
A Supersonic/Hypersonic Aerodynamic Investigation of the Saturn 1B/Apollo Upper Stage
NASA Technical Reports Server (NTRS)
1966-01-01
The static stability and axial force characteristics of an upper stage Saturn 1B/Apollo model were investigated in a series of wind tunnel tests at Mach numbers between 1. 93 and 8. 05. This report presents and analyzes the results of these small-scale studies. Test procedures and models are described, and consideration of boundary layer characteristics and violated modeling rules yields an assessment of the validity of the data. Four basic models provided evaluation of the basic launch vehicle characteristics, as well as the effects of two modes of mission abort and jettison of the launch escape system. The launch configuration experienced sudden, strong changes in stability derivatives and axial force coefficients near Mach 5. No similar changes were observed on the other configurations. Effects of Reynolds number, Mach number and boundary layer trips were obtained for each shape. The faired curves are believed to be valid for full-scale flight of the Saturn 1B/Apollo upper stage.
Applications of the pipeline environment for visual informatics and genomics computations
2011-01-01
Background Contemporary informatics and genomics research require efficient, flexible and robust management of large heterogeneous data, advanced computational tools, powerful visualization, reliable hardware infrastructure, interoperability of computational resources, and detailed data and analysis-protocol provenance. The Pipeline is a client-server distributed computational environment that facilitates the visual graphical construction, execution, monitoring, validation and dissemination of advanced data analysis protocols. Results This paper reports on the applications of the LONI Pipeline environment to address two informatics challenges - graphical management of diverse genomics tools, and the interoperability of informatics software. Specifically, this manuscript presents the concrete details of deploying general informatics suites and individual software tools to new hardware infrastructures, the design, validation and execution of new visual analysis protocols via the Pipeline graphical interface, and integration of diverse informatics tools via the Pipeline eXtensible Markup Language syntax. We demonstrate each of these processes using several established informatics packages (e.g., miBLAST, EMBOSS, mrFAST, GWASS, MAQ, SAMtools, Bowtie) for basic local sequence alignment and search, molecular biology data analysis, and genome-wide association studies. These examples demonstrate the power of the Pipeline graphical workflow environment to enable integration of bioinformatics resources which provide a well-defined syntax for dynamic specification of the input/output parameters and the run-time execution controls. Conclusions The LONI Pipeline environment http://pipeline.loni.ucla.edu provides a flexible graphical infrastructure for efficient biomedical computing and distributed informatics research. The interactive Pipeline resource manager enables the utilization and interoperability of diverse types of informatics resources. The Pipeline client-server model provides computational power to a broad spectrum of informatics investigators - experienced developers and novice users, user with or without access to advanced computational-resources (e.g., Grid, data), as well as basic and translational scientists. The open development, validation and dissemination of computational networks (pipeline workflows) facilitates the sharing of knowledge, tools, protocols and best practices, and enables the unbiased validation and replication of scientific findings by the entire community. PMID:21791102
Personality and organizational influences on aerospace human performance
NASA Technical Reports Server (NTRS)
Helmreich, Robert L.
1989-01-01
Individual and organizational influences on performance in aerospace environments are discussed. A model of personality with demonstrated validity is described along with reasons why personality's effects on performance have been underestimated. Organizational forces including intergroup conflict and coercive pressures are also described. It is suggested that basic and applied research in analog situations is needed to provide necessary guidance for planning future space missions.
ERIC Educational Resources Information Center
Teeter, Phyllis Anne; Smith, Philip L.
The final report of the 2-year project describes the development and validation of microcomputer software to help assess reading disabled elementary grade children and to provide basic reading instruction. Accomplishments of the first year included: design of the STAR Neuro-Cognitive Assessment Program which includes a reproduction of…
ERIC Educational Resources Information Center
Lethbridge Catholic Separate School District #9 (Alberta).
The objectives of the 2-year Thinking Skills Project were to provide teachers and students with a set of thinking skills, to develop and validate a model of cognition for teachers, to devise a Measure of Questioning Skills, and to establish a normative base for this instrument. The model of essential thinking skills covers the basic processes: (1)…
ERIC Educational Resources Information Center
Murphy, Colleen; Martin, Garry L.; Yu, C. T.
2014-01-01
The Assessment of Basic Learning Abilities (ABLA) is an empirically validated clinical tool for assessing the learning ability of persons with intellectual disabilities and children with autism. An ABLA tester uses standardized prompting and reinforcement procedures to attempt to teach, individually, each of six tasks, called levels, to a testee,…
ERIC Educational Resources Information Center
Spivey, Bruce E.; And Others
A systems approach has been used to develop and validate instructional materials in medical education. Seven self-instructional units, currently in various stages of completion, form the basis of a basic curriculum in opthamology. Performance objectives for the units were derived from the results of a questionnaire responded to by 1,600…
Betschart, Patrick; Zumstein, Valentin; Bentivoglio, Maico; Engeler, Daniel; Schmid, Hans-Peter; Abt, Dominik
2017-12-01
To assess the readability of the web-based patient education material provided by the European Association of Urology. English patient education materials (PEM) as available in May 2017 were obtained from the EAU website. Each topic was analyzed separately using six well-established readability assessment tools, including Flesch-Kincaid Grade Level (FKGL), SMOG Grade Level (SMOG), Coleman-Liau Index (CLI), Gunning Fog Index (GFI), Flesch Reading Ease Formula (FRE) and Fry Readability Graph (FRG). A total of 17 main topics were identified of which separate basic and in-depth information is provided for 14 topics. Calculation of grade levels (FKGL, SMOG, CLI, GFI) showed readability scores of 7th-13th grade for basic information, 8th-15th grade for in-depth information and 7th-15th grade for single PEM. Median FRE score was 54 points (range 45-65) for basic information and 56 points (41-64) for in-depth information. The FRG as a graphical assessment revealed only 13 valid results with an approximate 8th-17th grade level. The EAU provides carefully worked out PEM for 17 urological topics. Although improved readability compared to similar analyses was found, a simplification of certain chapters might be helpful to facilitate better patient understanding.
Siedlecki, Sandra L; Albert, Nancy M
This article will describe how to assess interrater reliability and validity of risk assessment tools, using easy-to-follow formulas, and to provide calculations that demonstrate principles discussed. Clinical nurse specialists should be able to identify risk assessment tools that provide high-quality interrater reliability and the highest validity for predicting true events of importance to clinical settings. Making best practice recommendations for assessment tool use is critical to high-quality patient care and safe practices that impact patient outcomes and nursing resources. Optimal risk assessment tool selection requires knowledge about interrater reliability and tool validity. The clinical nurse specialist will understand the reliability and validity issues associated with risk assessment tools, and be able to evaluate tools using basic calculations. Risk assessment tools are developed to objectively predict quality and safety events and ultimately reduce the risk of event occurrence through preventive interventions. To ensure high-quality tool use, clinical nurse specialists must critically assess tool properties. The better the tool's ability to predict adverse events, the more likely that event risk is mediated. Interrater reliability and validity assessment is relatively an easy skill to master and will result in better decisions when selecting or making recommendations for risk assessment tool use.
Show and tell: disclosure and data sharing in experimental pathology.
Schofield, Paul N; Ward, Jerrold M; Sundberg, John P
2016-06-01
Reproducibility of data from experimental investigations using animal models is increasingly under scrutiny because of the potentially negative impact of poor reproducibility on the translation of basic research. Histopathology is a key tool in biomedical research, in particular for the phenotyping of animal models to provide insights into the pathobiology of diseases. Failure to disclose and share crucial histopathological experimental details compromises the validity of the review process and reliability of the conclusions. We discuss factors that affect the interpretation and validation of histopathology data in publications and the importance of making these data accessible to promote replicability in research. © 2016. Published by The Company of Biologists Ltd.
ERIC Educational Resources Information Center
Hintze, John M.; Ryan, Amanda L.; Stoner, Gary
2003-01-01
The purpose of this study was to (a) examine the concurrent validity of the Dynamic Indicators of Basic Early Literacy Skills (DIBELS) with the Comprehensive Test of Phonological Processing (CTOPP), and (b) explore the diagnostic accuracy of the DIBELS in predicting CTOPP performance using suggested and alternative cut-scores. Eighty-six students…
ERIC Educational Resources Information Center
Vlachopoulos, Symeon P.; Michailidou, Sotiria
2006-01-01
The development process and initial validation of the Basic Psychological Needs in Exercise Scale (BPNES) are presented in this study. The BPNES is a domain-specific self-report instrument designed to assess perceptions of the extent to which the innate needs for autonomy, competence, and relatedness (Deci & Ryan, 2000) are satisfied in…
ERIC Educational Resources Information Center
Struyf, Elke; Adriaensens, Stefanie; Meynen, Karen
2011-01-01
Society has become more complex in recent decades, and this has increased the demands placed on the educational system and the teaching profession. This study focuses on the development and validation of an instrument that measures the basic skills of beginning teachers. The instrument was developed according to scientific knowledge on teacher…
ERIC Educational Resources Information Center
Abós Catalán, Ángel; Sevil Serrano, Javier; Julián Clemente, José Antonio; Martín-Albo Lucas, José; García-González, Luis
2018-01-01
The present study aimed to validate a Spanish-version of the Basic Psychological Needs at Work Scale (BPNWS-Sp) and to examine the associations between needs satisfaction and engagement and burnout in secondary education teachers. Using a sample of 584 secondary education teachers, the results supported the three-factor model, composite…
ERIC Educational Resources Information Center
Maherally, Uzma Nooreen
2014-01-01
The purpose of this study was to develop and validate a science assessment tool termed the Life Sciences Assessment (LSA) in order to assess preschool children's conceptions of basic life sciences. The hypothesis was that the four sub-constructs, each of which can be measured through a series of questions on the LSA, will make a significant…
Basic statistics (the fundamental concepts).
Lim, Eric
2014-12-01
An appreciation and understanding of statistics is import to all practising clinicians, not simply researchers. This is because mathematics is the fundamental basis to which we base clinical decisions, usually with reference to the benefit in relation to risk. Unless a clinician has a basic understanding of statistics, he or she will never be in a position to question healthcare management decisions that have been handed down from generation to generation, will not be able to conduct research effectively nor evaluate the validity of published evidence (usually making an assumption that most published work is either all good or all bad). This article provides a brief introduction to basic statistical methods and illustrates its use in common clinical scenarios. In addition, pitfalls of incorrect usage have been highlighted. However, it is not meant to be a substitute for formal training or consultation with a qualified and experienced medical statistician prior to starting any research project.
Vachon, David D; Lynam, Donald R; Widiger, Thomas A; Miller, Joshua D; McCrae, Robert R; Costa, Paul T
2013-05-01
Personality disorders (PDs) may be better understood in terms of dimensions of general personality functioning rather than as discrete categorical conditions. Personality-trait descriptions of PDs are robust across methods and settings, and PD assessments based on trait measures show good construct validity. The study reported here extends research showing that basic traits (e.g., impulsiveness, warmth, straightforwardness, modesty, and deliberation) can re-create the epidemiological characteristics associated with PDs. Specifically, we used normative changes in absolute trait levels to simulate age-related differences in the prevalence of psychopathy in a forensic setting. Results demonstrated that trait information predicts the rate of decline for psychopathy over the life span; discriminates the decline of psychopathy from that of a similar disorder, antisocial PD; and accurately predicts the differential decline of subfactors of psychopathy. These findings suggest that basic traits provide a parsimonious account of PD prevalence across the life span.
Jalink, M B; Goris, J; Heineman, E; Pierie, J P E N; ten Cate Hoedemaker, H O
2014-02-01
Virtual reality (VR) laparoscopic simulators have been around for more than 10 years and have proven to be cost- and time-effective in laparoscopic skills training. However, most simulators are, in our experience, considered less interesting by residents and are often poorly accessible. Consequently, these devices are rarely used in actual training. In an effort to make a low-cost and more attractive simulator, a custom-made Nintendo Wii game was developed. This game could ultimately be used to train the same basic skills as VR laparoscopic simulators ought to. Before such a video game can be implemented into a surgical training program, it has to be validated according to international standards. The main goal of this study was to test construct and concurrent validity of the controls of a prototype of the game. In this study, the basic laparoscopic skills of experts (surgeons, urologists, and gynecologists, n = 15) were compared to those of complete novices (internists, n = 15) using the Wii Laparoscopy (construct validity). Scores were also compared to the Fundamentals of Laparoscopy (FLS) Peg Transfer test, an already established assessment method for measuring basic laparoscopic skills (concurrent validity). Results showed that experts were 111 % faster (P = 0.001) on the Wii Laparoscopy task than novices. Also, scores of the FLS Peg Transfer test and the Wii Laparoscopy showed a significant, high correlation (r = 0.812, P < 0.001). The prototype setup of the Wii Laparoscopy possesses solid construct and concurrent validity.
Validation of the procedures. [integrated multidisciplinary optimization of rotorcraft
NASA Technical Reports Server (NTRS)
Mantay, Wayne R.
1989-01-01
Validation strategies are described for procedures aimed at improving the rotor blade design process through a multidisciplinary optimization approach. Validation of the basic rotor environment prediction tools and the overall rotor design are discussed.
ERIC Educational Resources Information Center
Postert, Christian; Averbeck-Holocher, Marlies; Beyer, Thomas; Muller, Jorg; Furniss, Tilman
2009-01-01
"DSM-IV" and "ICD-10" have limitations in the diagnostic classification of psychiatric disorders at preschool age (0-5 years). The publication of the "Diagnostic Classification 0-3 (DC:0-3)" in 1994, its basically revised second edition ("DC:0-3R") in 2005 and the "Research Diagnostic Criteria-Preschool Age (RDC-PA)" in 2004 have provided several…
Technical skills assessment toolbox: a review using the unitary framework of validity.
Ghaderi, Iman; Manji, Farouq; Park, Yoon Soo; Juul, Dorthea; Ott, Michael; Harris, Ilene; Farrell, Timothy M
2015-02-01
The purpose of this study was to create a technical skills assessment toolbox for 35 basic and advanced skills/procedures that comprise the American College of Surgeons (ACS)/Association of Program Directors in Surgery (APDS) surgical skills curriculum and to provide a critical appraisal of the included tools, using contemporary framework of validity. Competency-based training has become the predominant model in surgical education and assessment of performance is an essential component. Assessment methods must produce valid results to accurately determine the level of competency. A search was performed, using PubMed and Google Scholar, to identify tools that have been developed for assessment of the targeted technical skills. A total of 23 assessment tools for the 35 ACS/APDS skills modules were identified. Some tools, such as Operative Performance Rating System (OSATS) and Objective Structured Assessment of Technical Skill (OPRS), have been tested for more than 1 procedure. Therefore, 30 modules had at least 1 assessment tool, with some common surgical procedures being addressed by several tools. Five modules had none. Only 3 studies used Messick's framework to design their validity studies. The remaining studies used an outdated framework on the basis of "types of validity." When analyzed using the contemporary framework, few of these studies demonstrated validity for content, internal structure, and relationship to other variables. This study provides an assessment toolbox for common surgical skills/procedures. Our review shows that few authors have used the contemporary unitary concept of validity for development of their assessment tools. As we progress toward competency-based training, future studies should provide evidence for various sources of validity using the contemporary framework.
The Basic Empathy Scale: A Chinese Validation of a Measure of Empathy in Adolescents
ERIC Educational Resources Information Center
Geng, Yaoguo; Xia, Dan; Qin, Beibei
2012-01-01
The purpose of this study was to evaluate the reliability and validity of the Chinese version of the Basic Empathy Scale (BES). The Chinese version of BES was administered to a sample (n = 1,524) aged 9-18 and 65 males with conduct disorder aged 13-18. The result of confirmatory factor analysis showed a two-factor structure with four items deleted…
ERIC Educational Resources Information Center
Vogel, Ronald J.
Errors made on applications to the Basic Educational Opportunity Grant (BEOG) program were studied to determine the scope and nature of misreporting and misuse of the BEOG program. Characteristics of applicants who misreport personal and/or family financial circumstances were identified, and procedures were established to regularly screen and…
Virtual reality-based assessment of basic laparoscopic skills using the Leap Motion controller.
Lahanas, Vasileios; Loukas, Constantinos; Georgiou, Konstantinos; Lababidi, Hani; Al-Jaroudi, Dania
2017-12-01
The majority of the current surgical simulators employ specialized sensory equipment for instrument tracking. The Leap Motion controller is a new device able to track linear objects with sub-millimeter accuracy. The aim of this study was to investigate the potential of a virtual reality (VR) simulator for assessment of basic laparoscopic skills, based on the low-cost Leap Motion controller. A simple interface was constructed to simulate the insertion point of the instruments into the abdominal cavity. The controller provided information about the position and orientation of the instruments. Custom tools were constructed to simulate the laparoscopic setup. Three basic VR tasks were developed: camera navigation (CN), instrument navigation (IN), and bimanual operation (BO). The experiments were carried out in two simulation centers: MPLSC (Athens, Greece) and CRESENT (Riyadh, Kingdom of Saudi Arabia). Two groups of surgeons (28 experts and 21 novices) participated in the study by performing the VR tasks. Skills assessment metrics included time, pathlength, and two task-specific errors. The face validity of the training scenarios was also investigated via a questionnaire completed by the participants. Expert surgeons significantly outperformed novices in all assessment metrics for IN and BO (p < 0.05). For CN, a significant difference was found in one error metric (p < 0.05). The greatest difference between the performances of the two groups occurred for BO. Qualitative analysis of the instrument trajectory revealed that experts performed more delicate movements compared to novices. Subjects' ratings on the feedback questionnaire highlighted the training value of the system. This study provides evidence regarding the potential use of the Leap Motion controller for assessment of basic laparoscopic skills. The proposed system allowed the evaluation of dexterity of the hand movements. Future work will involve comparison studies with validated simulators and development of advanced training scenarios on current Leap Motion controller.
Consensus methods: review of original methods and their main alternatives used in public health
Bourrée, Fanny; Michel, Philippe; Salmi, Louis Rachid
2008-01-01
Summary Background Consensus-based studies are increasingly used as decision-making methods, for they have lower production cost than other methods (observation, experimentation, modelling) and provide results more rapidly. The objective of this paper is to describe the principles and methods of the four main methods, Delphi, nominal group, consensus development conference and RAND/UCLA, their use as it appears in peer-reviewed publications and validation studies published in the healthcare literature. Methods A bibliographic search was performed in Pubmed/MEDLINE, Banque de Données Santé Publique (BDSP), The Cochrane Library, Pascal and Francis. Keywords, headings and qualifiers corresponding to a list of terms and expressions related to the consensus methods were searched in the thesauri, and used in the literature search. A search with the same terms and expressions was performed on Internet using the website Google Scholar. Results All methods, precisely described in the literature, are based on common basic principles such as definition of subject, selection of experts, and direct or remote interaction processes. They sometimes use quantitative assessment for ranking items. Numerous variants of these methods have been described. Few validation studies have been implemented. Not implementing these basic principles and failing to describe the methods used to reach the consensus were both frequent reasons contributing to raise suspicion regarding the validity of consensus methods. Conclusion When it is applied to a new domain with important consequences in terms of decision making, a consensus method should be first validated. PMID:19013039
Cannon, Joanna E; Hubley, Anita M; Millhoff, Courtney; Mazlouman, Shahla
2016-01-01
The aim of the current study was to gather validation evidence for the Comprehension of Written Grammar (CWG; Easterbrooks, 2010) receptive test of 26 grammatical structures of English print for use with children who are deaf and hard of hearing (DHH). Reliability and validity data were collected for 98 participants (49 DHH and 49 hearing) in Grades 2-6. The objectives were to: (a) examine 4-week test-retest reliability data; and (b) provide evidence of known-groups validity by examining expected differences between the groups on the CWG vocabulary pretest and main test, as well as selected structures. Results indicated excellent test-retest reliability estimates for CWG test scores. DHH participants performed statistically significantly lower on the CWG vocabulary pretest and main test than the hearing participants. Significantly lower performance by DHH participants on most expected grammatical structures (e.g., basic sentence patterns, auxiliary "be" singular/plural forms, tense, comparatives, and complementation) also provided known groups evidence. Overall, the findings of this study showed strong evidence of the reliability of scores and known group-based validity of inferences made from the CWG. © The Author 2015. Published by Oxford University Press. All rights reserved. For Permissions, please email: journals.permissions@oup.com.
Latifi, Rifat; Ziemba, Michelle; Leppäniemi, Ari; Dasho, Erion; Dogjani, Agron; Shatri, Zhaneta; Kociraj, Agim; Oldashi, Fatos; Shosha, Lida
2014-08-01
Trauma continues to be a major health problem worldwide, particularly in the developing world, with high mortality and morbidity. Yet most developing countries lack an organized trauma system. Furthermore, developing countries do not have in place any accreditation process for trauma centers; thus, no accepted standard assessment tools exist to evaluate their trauma services. The aims of this study were to evaluate the trauma system in Albania, using the basic trauma criteria of the American College of Surgeons/Committee on Trauma (ACS/COT) as assessment tools, and to provide the Government with a situational analysis relative to these criteria. We used the ACS/COT basic criteria as assessment tools to evaluate the trauma system in Albania. We conducted a series of semi-structured interviews, unstructured interviews, and focus groups with all stakeholders at the Ministry of Health, at the University Trauma Hospital (UTH) based in Tirana (the capital city), and at ten regional hospitals across the country. Albania has a dedicated national trauma center that serves as the only tertiary center, plus ten regional hospitals that provide some trauma care. However, overall, its trauma system is in need of major reforms involving all essential elements in order to meet the basic requirements of a structured trauma system. The ACS/COT basic criteria can be used as assessment tools to evaluate trauma care in developing countries. Further studies are needed in other developing countries to validate the applicability of these criteria.
NASA Astrophysics Data System (ADS)
Adams, Elizabeth L.; Carrier, Sarah J.; Minogue, James; Porter, Stephen R.; McEachin, Andrew; Walkowiak, Temple A.; Zulli, Rebecca A.
2017-02-01
The Instructional Practices Log in Science (IPL-S) is a daily teacher log developed for K-5 teachers to self-report their science instruction. The items on the IPL-S are grouped into scales measuring five dimensions of science instruction: Low-level Sense-making, High-level Sense-making, Communication, Integrated Practices, and Basic Practices. As part of the current validation study, 206 elementary teachers completed 4137 daily log entries. The purpose of this paper is to provide evidence of validity for the IPL-S's scales, including (a) support for the theoretical framework; (b) cognitive interviews with logging teachers; (c) item descriptive statistics; (d) comparisons of 28 pairs of teacher and rater logs; and (e) an examination of the internal structure of the IPL-S. We present evidence to describe the extent to which the items and the scales are completed accurately by teachers and differentiate various types of science instructional strategies employed by teachers. Finally, we point to several practical implications of our work and potential uses for the IPL-S. Overall, results provide neutral to positive support for the validity of the groupings of items or scales.
Passive haptics in a knee arthroscopy simulator: is it valid for core skills training?
McCarthy, Avril D; Moody, Louise; Waterworth, Alan R; Bickerstaff, Derek R
2006-01-01
Previous investigation of a cost-effective virtual reality arthroscopic training system, the Sheffield Knee Arthroscopy Training System (SKATS), indicated the desirability of including haptic feedback. A formal task analysis confirmed the importance of knee positioning as a core skill for trainees learning to navigate the knee arthroscopically. The system cost and existing limb interface, which permits knee positioning, would be compromised by the addition of commercial active haptic devices available currently. The validation results obtained when passive haptic feedback (resistance provided by physical structures) is provided indicate that SKATS has construct, predictive and face validity for navigation and triangulation training. When tested using SKATS, experienced surgeons (n = 11) performed significantly faster, located significantly more pathologies, and showed significantly shorter arthroscope path lengths than a less experienced surgeon cohort (n = 12). After SKATS training sessions, novices (n = 3) showed significant improvements in: task completion time, shorter arthroscope path lengths, shorter probe path lengths, and fewer arthroscope tip contacts. Main improvements occurred after the first two practice sessions, indicating rapid familiarization and a training effect. Feedback from questionnaires completed by orthopaedic surgeons indicates that the system has face validity for its remit of basic arthroscopic training.
ERIC Educational Resources Information Center
Hildebrand, Myrene; Hoover, H. D.
This study compared the reliability and validity of two different measures of reading ability, the Degrees of Reading Power (DRP) and the Iowa Tests of Basic Skills (ITBS) Reading test and the ITBS Vocabulary test. The data consisted of scores of 377 grade 5 and grade 6 students on these tests, along with their assigned reading levels in the…
Clow, Gary D.
2014-01-01
System (GTOS). The data will also be useful for refining our basic understanding of the physical conditions in permafrost in Arctic Alaska, as well as providing important information for validating predictive models used for climate impact assessments. The processed data are available from the Advanced Cooperative Arctic Data and Information Service (ACADIS) repository at doi:10.5065/D6N014HK.
Eastwood, Sophie V; Tillin, Therese; Wright, Andrew; Heasman, John; Willis, Joseph; Godsland, Ian F; Forouhi, Nita; Whincup, Peter; Hughes, Alun D; Chaturvedi, Nishi
2013-01-01
South Asians and African Caribbeans experience more cardiometabolic disease than Europeans. Risk factors include visceral (VAT) and subcutaneous abdominal (SAT) adipose tissue, which vary with ethnicity and are difficult to quantify using anthropometry. We developed and cross-validated ethnicity and gender-specific equations using anthropometrics to predict VAT and SAT. 669 Europeans, 514 South Asians and 227 African Caribbeans (70 ± 7 years) underwent anthropometric measurement and abdominal CT scanning. South Asian and African Caribbean participants were first-generation migrants living in London. Prediction equations were derived for CT-measured VAT and SAT using stepwise regression, then cross-validated by comparing actual and predicted means. South Asians had more and African Caribbeans less VAT than Europeans. For basic VAT prediction equations (age and waist circumference), model fit was better in men (R(2) range 0.59-0.71) than women (range 0.35-0.59). Expanded equations (+ weight, height, hip and thigh circumference) improved fit for South Asian and African Caribbean women (R(2) 0.35 to 0.55, and 0.43 to 0.56 respectively). For basic SAT equations, R(2) was 0.69-0.77, and for expanded equations it was 0.72-0.86. Cross-validation showed differences between actual and estimated VAT of <7%, and SAT of <8% in all groups, apart from VAT in South Asian women which disagreed by 16%. We provide ethnicity- and gender-specific VAT and SAT prediction equations, derived from a large tri-ethnic sample. Model fit was reasonable for SAT and VAT in men, while basic VAT models should be used cautiously in South Asian and African Caribbean women. These equations will aid studies of mechanisms of cardiometabolic disease in later life, where imaging data are not available.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Nallasivam, Ulaganathan; Shah, Vishesh H.; Shenvi, Anirudh A.
We present a general Global Minimization Algorithm (GMA) to identify basic or thermally coupled distillation configurations that require the least vapor duty under minimum reflux conditions for separating any ideal or near-ideal multicomponent mixture into a desired number of product streams. In this algorithm, global optimality is guaranteed by modeling the system using Underwood equations and reformulating the resulting constraints to bilinear inequalities. The speed of convergence to the globally optimal solution is increased by using appropriate feasibility and optimality based variable-range reduction techniques and by developing valid inequalities. As a result, the GMA can be coupled with already developedmore » techniques that enumerate basic and thermally coupled distillation configurations, to provide for the first time, a global optimization based rank-list of distillation configurations.« less
Bone grafts, bone substitutes and orthobiologics
Roberts, Timothy T.; Rosenbaum, Andrew J.
2012-01-01
The biology of fracture healing is better understood than ever before, with advancements such as the locking screw leading to more predictable and less eventful osseous healing. However, at times one’s intrinsic biological response, and even concurrent surgical stabilization, is inadequate. In hopes of facilitating osseous union, bone grafts, bone substitutes and orthobiologics are being relied on more than ever before. The osteoinductive, osteoconductive and osteogenic properties of these substrates have been elucidated in the basic science literature and validated in clinical orthopaedic practice. Furthermore, an industry built around these items is more successful and in demand than ever before. This review provides a comprehensive overview of the basic science, clinical utility and economics of bone grafts, bone substitutes and orthobiologics. PMID:23247591
NASA Technical Reports Server (NTRS)
Eisfeld, Bernhard; Rumsey, Chris; Togiti, Vamshi
2015-01-01
The implementation of the SSG/LRR-omega differential Reynolds stress model into the NASA flow solvers CFL3D and FUN3D and the DLR flow solver TAU is verified by studying the grid convergence of the solution of three different test cases from the Turbulence Modeling Resource Website. The model's predictive capabilities are assessed based on four basic and four extended validation cases also provided on this website, involving attached and separated boundary layer flows, effects of streamline curvature and secondary flow. Simulation results are compared against experimental data and predictions by the eddy-viscosity models of Spalart-Allmaras (SA) and Menter's Shear Stress Transport (SST).
Ketcham, Jonathan D; Kuminoff, Nicolai V; Powers, Christopher A
2016-12-01
Consumers' enrollment decisions in Medicare Part D can be explained by Abaluck and Gruber’s (2011) model of utility maximization with psychological biases or by a neoclassical version of their model that precludes such biases. We evaluate these competing hypotheses by applying nonparametric tests of utility maximization and model validation tests to administrative data. We find that 79 percent of enrollment decisions from 2006 to 2010 satisfied basic axioms of consumer theory under the assumption of full information. The validation tests provide evidence against widespread psychological biases. In particular, we find that precluding psychological biases improves the structural model's out-of-sample predictions for consumer behavior.
Diagnosis and Management of Acute Concussion.
McCrea, Michael A; Nelson, Lindsay D; Guskiewicz, Kevin
2017-05-01
Over the past 2 decades, there have been major advances in the basic and clinical science of concussion and mild traumatic brain injury. These advances now provide a more evidence-informed approach to the definition, diagnosis, assessment, and management of acute concussion. Standardized clinical tools have been developed and validated for assessment of acute concussion across injury settings (eg, civilian, sport, military). Consensus guidelines now provide guidance regarding injury management and approaches to ensure safe return to activity after acute concussion. This article provides a brief, high-level overview of approaches to best practice in diagnosis, assessment, and management of acute concussion. Copyright © 2017 Elsevier Inc. All rights reserved.
Ethical Implications of Validity-vs.-Reliability Trade-Offs in Educational Research
ERIC Educational Resources Information Center
Fendler, Lynn
2016-01-01
In educational research that calls itself empirical, the relationship between validity and reliability is that of trade-off: the stronger the bases for validity, the weaker the bases for reliability (and vice versa). Validity and reliability are widely regarded as basic criteria for evaluating research; however, there are ethical implications of…
Commentary on "Validating the Interpretations and Uses of Test Scores"
ERIC Educational Resources Information Center
Brennan, Robert L.
2013-01-01
Kane's paper "Validating the Interpretations and Uses of Test Scores" is the most complete and clearest discussion yet available of the argument-based approach to validation. At its most basic level, validation as formulated by Kane is fundamentally a simply-stated two-step enterprise: (1) specify the claims inherent in a particular interpretation…
Mayer, B; Muche, R
2013-01-01
Animal studies are highly relevant for basic medical research, although their usage is discussed controversially in public. Thus, an optimal sample size for these projects should be aimed at from a biometrical point of view. Statistical sample size calculation is usually the appropriate methodology in planning medical research projects. However, required information is often not valid or only available during the course of an animal experiment. This article critically discusses the validity of formal sample size calculation for animal studies. Within the discussion, some requirements are formulated to fundamentally regulate the process of sample size determination for animal experiments.
Emerging Drugs for the Treatment of Anxiety
Murrough, James W.; Yaqubi, Sahab; Sayed, Sehrish; Charney, Dennis S.
2016-01-01
Introduction Anxiety disorders are among the most prevalent and disabling psychiatric disorders in the United States and worldwide. Basic research has provided critical insights into the mechanism regulating fear behavior in animals and a host of animal models have been developed in order to screen compounds for anxiolytic properties. Despite this progress, no mechanistically novel agents for the treatment of anxiety have come to market in more than two decades. Areas covered The current review will provide a critical summary of current pharmacological approaches to the treatment of anxiety and will examine the pharmacotherapeutic pipeline for treatments in development. Anxiety and related disorders considered herein include panic disorder, social anxiety disorder, generalized anxiety disorder and posttraumatic stress disorder. The glutamate, neuropeptide and endocannabinoid systems show particular promise as future targets for novel drug development. Expert opinion In the face of an ever-growing understanding of fear related behavior, the field awaits the translation of this research into mechanistically novel treatments. Obstacles will be overcome through close collaboration between basic and clinical researchers with the goal of aligning valid endophenotypes of human anxiety disorders with improved animal models. Novel approaches are needed to move basic discoveries into new, more effective treatments for our patients. PMID:26012843
Assessment of MARMOT. A Mesoscale Fuel Performance Code
DOE Office of Scientific and Technical Information (OSTI.GOV)
Tonks, M. R.; Schwen, D.; Zhang, Y.
2015-04-01
MARMOT is the mesoscale fuel performance code under development as part of the US DOE Nuclear Energy Advanced Modeling and Simulation Program. In this report, we provide a high level summary of MARMOT, its capabilities, and its current state of validation. The purpose of MARMOT is to predict the coevolution of microstructure and material properties of nuclear fuel and cladding. It accomplished this using the phase field method coupled to solid mechanics and heat conduction. MARMOT is based on the Multiphysics Object-Oriented Simulation Environment (MOOSE), and much of its basic capability in the areas of the phase field method, mechanics,more » and heat conduction come directly from MOOSE modules. However, additional capability specific to fuel and cladding is available in MARMOT. While some validation of MARMOT has been completed in the areas of fission gas behavior and grain growth, much more validation needs to be conducted. However, new mesoscale data needs to be obtained in order to complete this validation.« less
Summary of BISON Development and Validation Activities - NEAMS FY16 Report
DOE Office of Scientific and Technical Information (OSTI.GOV)
Williamson, R. L.; Pastore, G.; Gamble, K. A.
This summary report contains an overview of work performed under the work package en- titled “FY2016 NEAMS INL-Engineering Scale Fuel Performance (BISON)” A first chapter identifies the specific FY-16 milestones, providing a basic description of the associated work and references to related detailed documentation. Where applicable, a representative technical result is provided. A second chapter summarizes major additional accomplishments, which in- clude: 1) publication of a journal article on solution verification and validation of BISON for LWR fuel, 2) publication of a journal article on 3D Missing Pellet Surface (MPS) analysis of BWR fuel, 3) use of BISON to designmore » a unique 3D MPS validation experiment for future in- stallation in the Halden research reactor, 4) participation in an OECD benchmark on Pellet Clad Mechanical Interaction (PCMI), 5) participation in an OECD benchmark on Reactivity Insertion Accident (RIA) analysis, 6) participation in an OECD activity on uncertainity quantification and sensitivity analysis in nuclear fuel modeling and 7) major improvements to BISON’s fission gas behavior models. A final chapter outlines FY-17 future work.« less
van Dongen, Koen W; Ahlberg, Gunnar; Bonavina, Luigi; Carter, Fiona J; Grantcharov, Teodor P; Hyltander, Anders; Schijven, Marlies P; Stefani, Alessandro; van der Zee, David C; Broeders, Ivo A M J
2011-01-01
Virtual reality (VR) simulators have been demonstrated to improve basic psychomotor skills in endoscopic surgery. The exercise configuration settings used for validation in studies published so far are default settings or are based on the personal choice of the tutors. The purpose of this study was to establish consensus on exercise configurations and on a validated training program for a virtual reality simulator, based on the experience of international experts to set criterion levels to construct a proficiency-based training program. A consensus meeting was held with eight European teams, all extensively experienced in using the VR simulator. Construct validity of the training program was tested by 20 experts and 60 novices. The data were analyzed by using the t test for equality of means. Consensus was achieved on training designs, exercise configuration, and examination. Almost all exercises (7/8) showed construct validity. In total, 50 of 94 parameters (53%) showed significant difference. A European, multicenter, validated, training program was constructed according to the general consensus of a large international team with extended experience in virtual reality simulation. Therefore, a proficiency-based training program can be offered to training centers that use this simulator for training in basic psychomotor skills in endoscopic surgery.
Climatological Processing of Radar Data for the TRMM Ground Validation Program
NASA Technical Reports Server (NTRS)
Kulie, Mark; Marks, David; Robinson, Michael; Silberstein, David; Wolff, David; Ferrier, Brad; Amitai, Eyal; Fisher, Brad; Wang, Jian-Xin; Augustine, David;
2000-01-01
The Tropical Rainfall Measuring Mission (TRMM) satellite was successfully launched in November, 1997. The main purpose of TRMM is to sample tropical rainfall using the first active spaceborne precipitation radar. To validate TRMM satellite observations, a comprehensive Ground Validation (GV) Program has been implemented. The primary goal of TRMM GV is to provide basic validation of satellite-derived precipitation measurements over monthly climatologies for the following primary sites: Melbourne, FL; Houston, TX; Darwin, Australia; and Kwajalein Atoll, RMI. As part of the TRMM GV effort, research analysts at NASA Goddard Space Flight Center (GSFC) generate standardized TRMM GV products using quality-controlled ground-based radar data from the four primary GV sites as input. This presentation will provide an overview of the TRMM GV climatological processing system. A description of the data flow between the primary GV sites, NASA GSFC, and the TRMM Science and Data Information System (TSDIS) will be presented. The radar quality control algorithm, which features eight adjustable height and reflectivity parameters, and its effect on monthly rainfall maps will be described. The methodology used to create monthly, gauge-adjusted rainfall products for each primary site will also be summarized. The standardized monthly rainfall products are developed in discrete, modular steps with distinct intermediate products. These developmental steps include: (1) extracting radar data over the locations of rain gauges, (2) merging rain gauge and radar data in time and space with user-defined options, (3) automated quality control of radar and gauge merged data by tracking accumulations from each instrument, and (4) deriving Z-R relationships from the quality-controlled merged data over monthly time scales. A summary of recently reprocessed official GV rainfall products available for TRMM science users will be presented. Updated basic standardized product results and trends involving monthly accumulation, Z-R relationship, and gauge statistics for each primary GV site will be also displayed.
Flight Validation of On-Demand Operations: The Deep Space One Beacon Monitor Operations Experiment
NASA Technical Reports Server (NTRS)
Wyatt, Jay; Sherwood, Rob; Sue, Miles; Szijjarto, John
2000-01-01
After a brief overview of the operational concept, this paper will provide a detailed description of the _as-flown_ flight software components, the DS1 experiment plan, and experiment results to date. Special emphasis will be given to experiment results and lessons learned since the basic system design has been previously reported. Mission scenarios where beacon operations is highly applicable will be described. Detailed cost savings estimates for a sample science mission will be provided as will cumulative savings that are possible over the next fifteen years of NASA missions.
Geelen, Sven Jacobus Gertruda; Valkenet, Karin; Veenhof, Cindy
2018-05-12
To evaluate the construct validity and the inter-rater reliability of the Dutch Activity Measure for Post-Acute Care "6-clicks" Basic Mobility short form measuring the patient's mobility in Dutch hospital care. First, the "6-clicks" was translated by using a forward-backward translation protocol. Next, 64 patients were assessed by the physiotherapist to determine the validity while being admitted to the Internal Medicine wards of a university medical center. Six hypotheses were tested regarding the construct "mobility" which showed that: Better "6-clicks" scores were related to less restrictive pre-admission living situations (p = 0.011), less restrictive discharge locations (p = 0.001), more independence in activities of daily living (p = 0.001) and less physiotherapy visits (p < 0.001). A correlation was found between the "6-clicks" and length of stay (r= -0.408, p = 0.001), but not between the "6-clicks" and age (r= -0.180, p = 0.528). To determine the inter-rater reliability, an additional 50 patients were assessed by pairs of physiotherapists who independently scored the patients. Intraclass Correlation Coefficients of 0.920 (95%CI: 0.828-0.964) were found. The Kappa Coefficients for the individual items ranged from 0.649 (walking stairs) to 0.841 (sit-to-stand). The Dutch "6-clicks" shows a good construct validity and moderate-to-excellent inter-rater reliability when used to assess the mobility of hospitalized patients. Implications for Rehabilitation Even though various measurement tools have been developed, it appears the majority of physiotherapists working in a hospital currently do not use these tools as a standard part of their care. The Activity Measure for Post-Acute Care "6-clicks" Basic Mobility is the only tool which is designed to be short, easy to use within usual care and has been validated in the entire hospital population. This study shows that the Dutch version of the Activity Measure for Post-Acute Care "6-clicks" Basic Mobility form is a valid, easy to use, quick tool to assess the basic mobility of Dutch hospitalized patients.
Auditing radiation sterilization facilities
NASA Astrophysics Data System (ADS)
Beck, Jeffrey A.
The diversity of radiation sterilization systems available today places renewed emphasis on the need for thorough Quality Assurance audits of these facilities. Evaluating compliance with Good Manufacturing Practices is an obvious requirement, but an effective audit must also evaluate installation and performance qualification programs (validation_, and process control and monitoring procedures in detail. The present paper describes general standards that radiation sterilization operations should meet in each of these key areas, and provides basic guidance for conducting QA audits of these facilities.
ERIC Educational Resources Information Center
Binder, Katherine S.; Snyder, Melissa A.; Ardoin, Scott P.; Morris, Robin K.
2011-01-01
This study examined the reliability and validity of administering Dynamic Indicators of Basic Early Literacy Skills (DIBELS) to adult basic education (ABE) students. Ninety ABE participants were administered DIBELS measures, the Woodcock-Johnson III Broad Reading (WJ III BR) measures, and four orthographic ability tests. Since ABE students are a…
Validation of a partial coherence interferometry method for estimating retinal shape
Verkicharla, Pavan K.; Suheimat, Marwan; Pope, James M.; Sepehrband, Farshid; Mathur, Ankit; Schmid, Katrina L.; Atchison, David A.
2015-01-01
To validate a simple partial coherence interferometry (PCI) based retinal shape method, estimates of retinal shape were determined in 60 young adults using off-axis PCI, with three stages of modeling using variants of the Le Grand model eye, and magnetic resonance imaging (MRI). Stage 1 and 2 involved a basic model eye without and with surface ray deviation, respectively and Stage 3 used model with individual ocular biometry and ray deviation at surfaces. Considering the theoretical uncertainty of MRI (12-14%), the results of the study indicate good agreement between MRI and all three stages of PCI modeling with <4% and <7% differences in retinal shapes along horizontal and vertical meridians, respectively. Stage 2 and Stage 3 gave slightly different retinal co-ordinates than Stage 1 and we recommend the intermediate Stage 2 as providing a simple and valid method of determining retinal shape from PCI data. PMID:26417496
Validation of a partial coherence interferometry method for estimating retinal shape.
Verkicharla, Pavan K; Suheimat, Marwan; Pope, James M; Sepehrband, Farshid; Mathur, Ankit; Schmid, Katrina L; Atchison, David A
2015-09-01
To validate a simple partial coherence interferometry (PCI) based retinal shape method, estimates of retinal shape were determined in 60 young adults using off-axis PCI, with three stages of modeling using variants of the Le Grand model eye, and magnetic resonance imaging (MRI). Stage 1 and 2 involved a basic model eye without and with surface ray deviation, respectively and Stage 3 used model with individual ocular biometry and ray deviation at surfaces. Considering the theoretical uncertainty of MRI (12-14%), the results of the study indicate good agreement between MRI and all three stages of PCI modeling with <4% and <7% differences in retinal shapes along horizontal and vertical meridians, respectively. Stage 2 and Stage 3 gave slightly different retinal co-ordinates than Stage 1 and we recommend the intermediate Stage 2 as providing a simple and valid method of determining retinal shape from PCI data.
[Nurse's competence indicators: linguistic and cultural validation of the Nurse Competence Scale].
Finotto, Stefano; Cantarelli, William
2009-01-01
For some years, the clinical performance of new-graduate nurses, has been a leading topic in international scientific literature. In Italy there are many criticisms to basic education; ever since the basic education moved from the regional schools to the university, the main question that the teachers, the clinical nurses and the nursing managers are asking is whether the level of competence of new-graduates is appropriate to the demands of the world of work. Many criticisms have been addressed to the gap between theory and practice and between education and clinic. In Italy this has stimulated a debate towards a shared definition of competence and especially towards defining indicators that can assess/measure this phenomenon. The purposes of this study are: translating the indicators of Nurse Competence Scale (NCS) in the Italian language and test its validity and reliability; provide a tool for evaluating competence in Italian in order to use it in the context of our country. after a research on the Medline and Cinhal electronic data base, the NCS was identified and submitted to a process of linguistic translation (English-Italian-English) and to a process of validation using the test-retest methodology (test of Wilcoxon), the Intraclass Correlation Coefficient (ICC) and Cronbach's alpha. the evaluation given by nurses in the first administration does not differ significantly with those of the second one. For all sections of the NCS the ICC reports values greater than 0.85. the Nurse Competence Scale appears valid in its Italian version and it might be used to measure the competences of Italian nurses.
Rotthoff, Thomas; Ostapczuk, Martin S; Kröncke, Klaus D; Zimmerhofer, Alexander; Decking, Ulrich; Schneider, Matthias; Ritz-Timme, Stefanie
2014-01-01
Core competencies have progressively gained importance in medical education. In other contexts, especially personnel selection and development, assessment centers (ACs) are used to assess competencies, but there is only a limited number of studies on competency-based ACs in medical education. To the best of our knowledge, the present study provides the first data on the criterion-related validity of a competency-based AC in medical education. We developed an AC tailored to measure core competencies relevant to medical education (social-ethical, communicative, self, and teaching) and tested its validity in n=30 first-year medical students using 3- to 4-year follow-up measures such as (a) objective structured clinical examinations (OSCE) on basic clinical skills (n=26), (b) OSCE on communication skills (n=21), and (c) peer feedback (n=18). The AC contained three elements: interview, group discussion, and role play. Additionally, a self-report questionnaire was provided as a basis for the interview. Baseline AC average score and teaching competency correlated moderately with the communication OSCE average score (r=0.41, p=0.03, and r=0.38, p=0.04, respectively). Social-ethical competency in the AC showed a very strong convergent association with the communication OSCE average score (r=0.60, p<0.01). The AC total score also showed a moderate correlation with the overall peer feedback score provided in Year 4 (r=0.38, p=0.06). In addition, communicative competency correlated strongly with the overall peer feedback (r=0.50, p=0.02). We found predominantly low and insignificant correlations between the AC and the OSCE on basic clinical skills (r=-0.33 to 0.30, all p's>0.05). The results showed that competency-based ACs can be used at a very early stage of medical training to successfully predict future performance in core competencies.
Rotthoff, Thomas; Ostapczuk, Martin S; Kröncke, Klaus D; Zimmerhofer, Alexander; Decking, Ulrich; Schneider, Matthias; Ritz-Timme, Stefanie
2014-01-01
Introduction Core competencies have progressively gained importance in medical education. In other contexts, especially personnel selection and development, assessment centers (ACs) are used to assess competencies, but there is only a limited number of studies on competency-based ACs in medical education. To the best of our knowledge, the present study provides the first data on the criterion-related validity of a competency-based AC in medical education. Methods We developed an AC tailored to measure core competencies relevant to medical education (social-ethical, communicative, self, and teaching) and tested its validity in n=30 first-year medical students using 3- to 4-year follow-up measures such as (a) objective structured clinical examinations (OSCE) on basic clinical skills (n=26), (b) OSCE on communication skills (n=21), and (c) peer feedback (n=18). The AC contained three elements: interview, group discussion, and role play. Additionally, a self-report questionnaire was provided as a basis for the interview. Results Baseline AC average score and teaching competency correlated moderately with the communication OSCE average score (r=0.41, p=0.03, and r=0.38, p=0.04, respectively). Social-ethical competency in the AC showed a very strong convergent association with the communication OSCE average score (r=0.60, p<0.01). The AC total score also showed a moderate correlation with the overall peer feedback score provided in Year 4 (r=0.38, p=0.06). In addition, communicative competency correlated strongly with the overall peer feedback (r=0.50, p=0.02). We found predominantly low and insignificant correlations between the AC and the OSCE on basic clinical skills (r=-0.33 to 0.30, all p's>0.05). Conclusion The results showed that competency-based ACs can be used at a very early stage of medical training to successfully predict future performance in core competencies.
Cognitive simulators for medical education and training.
Kahol, Kanav; Vankipuram, Mithra; Smith, Marshall L
2009-08-01
Simulators for honing procedural skills (such as surgical skills and central venous catheter placement) have proven to be valuable tools for medical educators and students. While such simulations represent an effective paradigm in surgical education, there is an opportunity to add a layer of cognitive exercises to these basic simulations that can facilitate robust skill learning in residents. This paper describes a controlled methodology, inspired by neuropsychological assessment tasks and embodied cognition, to develop cognitive simulators for laparoscopic surgery. These simulators provide psychomotor skill training and offer the additional challenge of accomplishing cognitive tasks in realistic environments. A generic framework for design, development and evaluation of such simulators is described. The presented framework is generalizable and can be applied to different task domains. It is independent of the types of sensors, simulation environment and feedback mechanisms that the simulators use. A proof of concept of the framework is provided through developing a simulator that includes cognitive variations to a basic psychomotor task. The results of two pilot studies are presented that show the validity of the methodology in providing an effective evaluation and learning environments for surgeons.
An elementary tutorial on formal specification and verification using PVS
NASA Technical Reports Server (NTRS)
Butler, Ricky W.
1993-01-01
A tutorial on the development of a formal specification and its verification using the Prototype Verification System (PVS) is presented. The tutorial presents the formal specification and verification techniques by way of specific example - an airline reservation system. The airline reservation system is modeled as a simple state machine with two basic operations. These operations are shown to preserve a state invariant using the theorem proving capabilities of PVS. The technique of validating a specification via 'putative theorem proving' is also discussed and illustrated in detail. This paper is intended for the novice and assumes only some of the basic concepts of logic. A complete description of user inputs and the PVS output is provided and thus it can be effectively used while one is sitting at a computer terminal.
Nallasivam, Ulaganathan; Shah, Vishesh H.; Shenvi, Anirudh A.; ...
2016-02-10
We present a general Global Minimization Algorithm (GMA) to identify basic or thermally coupled distillation configurations that require the least vapor duty under minimum reflux conditions for separating any ideal or near-ideal multicomponent mixture into a desired number of product streams. In this algorithm, global optimality is guaranteed by modeling the system using Underwood equations and reformulating the resulting constraints to bilinear inequalities. The speed of convergence to the globally optimal solution is increased by using appropriate feasibility and optimality based variable-range reduction techniques and by developing valid inequalities. As a result, the GMA can be coupled with already developedmore » techniques that enumerate basic and thermally coupled distillation configurations, to provide for the first time, a global optimization based rank-list of distillation configurations.« less
Face validity of a Wii U video game for training basic laparoscopic skills.
Jalink, Maarten B; Goris, Jetse; Heineman, Erik; Pierie, Jean-Pierre E N; Ten Cate Hoedemaker, Henk O
2015-06-01
Although the positive effects of playing video games on basic laparoscopic skills have been studied for several years, no games are actually used in surgical training. This article discusses the face validity of the first video game and custom-made hardware, which takes advantage of these effects. Participants were recruited at the Chirurgendagen 2013 and the Society of American Gastrointestinal and Endoscopic Surgeons 2014 annual meeting. In total, 72 laparoscopic surgeons completed a demo of the game and filled in a questionnaire. On a 1-to-10 scale, the mean score for hardware realism was 7.2 and the mean score for usefulness as a training tool was 8.4. Participants did not mind the fact that the workspace does not look like an abdominal cavity, but do have some trouble with the absence of tactile feedback. We obtained face validity for both the hardware and the usefulness of Underground, a video game made for training basic laparoscopic skills. Copyright © 2015 Elsevier Inc. All rights reserved.
Albiero, Paolo; Matricardi, Giada; Speltri, Daniela; Toso, Diana
2009-04-01
The present study examined the validity of the Basic Empathy Scale (BES) [Jolliffe, D., & Farrington, D. P. (2006a). Development and validation of the Basic Empathy Scale. Journal of Adolescence, 29, 589-611; Jolliffe, D., & Farrington, D. P. (2006b). Examining the relationship between low empathy and bullying. Aggressive Behavior, 32(6), 540-550.] and found further evidence for the scale's good psychometric properties. The BES was administered to a sample of 655 Italian adolescents to examine the generalizability and reliability of its factor structure. The results of the confirmatory factor analysis showed a reasonable data fit with the two hypothesized BES domains of Cognitive Empathy and Affective Empathy. Scale reliability was also satisfactory, showing good internal consistency. Lastly, BES scores were significantly associated with other empathy questionnaire scores (the Interpersonal Reactivity Index and Balanced Emotional Empathy Scale) and with a scale measuring prosocial behavior. Research and practice implications are discussed.
Bridges, Mindy Sittner; Catts, Hugh W.
2013-01-01
This study examined the usefulness and predictive validity of a dynamic screening of phonological awareness in two samples of kindergarten children. In one sample (n = 90), the predictive validity of the dynamic assessment was compared to a static version of the same screening measure. In the second sample (n = 96), the dynamic screening measure was compared to a commonly used screening tool, Dynamic Indicators of Basic Early Literacy Skills Initial Sound Fluency. Results showed that the dynamic screening measure uniquely predicted end-of-year reading achievement and outcomes in both samples. These results provide preliminary support for the usefulness of a dynamic screening measure of phonological awareness for kindergarten students. PMID:21571700
NASA Astrophysics Data System (ADS)
Endryansyah; Wanarti Rusimamto, Puput; Ridianto, Adam; Sugiarto, Hariyadi
2018-04-01
In the Department of Electrical Engineering FT Unesa, there are 3 majors: S1 Electrical Engineering Education, S1 Electrical Engineering, and D3 Electrical Engineering. Courses the Basic System Settings go to in the curriculum of the three programs. Team lecturer college of basic system settings seek learning innovation, focused on the development of trainer to student practicum at the laboratory of systems control. Trainer developed is a servo motor along with the lab module that contains a wide variety of theories about the servo motor and guide the practicum. This research type is development research using methods Research & development (R & D). In which the steps are applied in this study is as follows: pay attention to the potential and existing problems, gather information and study the literature, design the product, validate the design, revise the design, a limited trial. The results of the validation of learning device in the form of modules and trainer obtained as follows: score validation of learning device is 3,64; score validation lab module Servo Motor is 3,47; and questionnaire responses of students is 3,73. The result of the whole validation value is located in the interval >of 3.25 s/d 4 with the category of “Very Valid”, so it can be concluded that all instruments have a level of validity “Very Valid” and worthy of use for further learning.
Good validity of the international spinal cord injury quality of life basic data set.
Post, M W M; Adriaansen, J J E; Charlifue, S; Biering-Sørensen, F; van Asbeck, F W A
2016-04-01
Cross-sectional validation study. To examine the construct and concurrent validity of the International Spinal Cord Injury (SCI) Quality of Life (QoL) Basic Data Set. Dutch community. People 28-65 years of age, who obtained their SCI between 18 and 35 years of age, were at least 10 years post SCI and were wheelchair users in daily life. MEASURE(S): The International SCI QoL Basic Data Set consists of three single items on satisfaction with life as a whole, physical health and psychological health (0=complete dissatisfaction; 10=complete satisfaction). Reference measures were the Mental Health Inventory-5 and three items of the World Health Organization Quality of Life measure. Data of 261 participants were available. Mean time after SCI was 24.1 years (s.d. 9.1); 90.4% had a traumatic SCI, 81.5% a motor complete SCI and 40% had tetraplegia. Mean age was 47.9 years (s.d. 8.8) and 73.2% were male. Mean scores were 6.9 (s.d. 1.9) for general QoL, 5.8 (s.d. 2.2) for physical health and 7.1 (s.d. 1.9) for psychological health. No floor or ceiling effects were found. Strong inter-correlations (0.48-0.71) were found between the items, and Cronbach's alpha of the scale was good (0.81). Correlations with the reference measures showed the strongest correlations between the WHOQOL general satisfaction item and general QoL (0.64), the WHOQOL health and daily activities items and physical health (0.69 and 0.60) and the Mental Health Inventory-5 and psychological health (0.70). This first validity study of the International SCI QoL Basic Data Set shows that it appears valid for persons with SCI.
PCA as a practical indicator of OPLS-DA model reliability.
Worley, Bradley; Powers, Robert
Principal Component Analysis (PCA) and Orthogonal Projections to Latent Structures Discriminant Analysis (OPLS-DA) are powerful statistical modeling tools that provide insights into separations between experimental groups based on high-dimensional spectral measurements from NMR, MS or other analytical instrumentation. However, when used without validation, these tools may lead investigators to statistically unreliable conclusions. This danger is especially real for Partial Least Squares (PLS) and OPLS, which aggressively force separations between experimental groups. As a result, OPLS-DA is often used as an alternative method when PCA fails to expose group separation, but this practice is highly dangerous. Without rigorous validation, OPLS-DA can easily yield statistically unreliable group separation. A Monte Carlo analysis of PCA group separations and OPLS-DA cross-validation metrics was performed on NMR datasets with statistically significant separations in scores-space. A linearly increasing amount of Gaussian noise was added to each data matrix followed by the construction and validation of PCA and OPLS-DA models. With increasing added noise, the PCA scores-space distance between groups rapidly decreased and the OPLS-DA cross-validation statistics simultaneously deteriorated. A decrease in correlation between the estimated loadings (added noise) and the true (original) loadings was also observed. While the validity of the OPLS-DA model diminished with increasing added noise, the group separation in scores-space remained basically unaffected. Supported by the results of Monte Carlo analyses of PCA group separations and OPLS-DA cross-validation metrics, we provide practical guidelines and cross-validatory recommendations for reliable inference from PCA and OPLS-DA models.
ERIC Educational Resources Information Center
Karl, Luis C.
The adult basic education (ABE) program at Nicolet Area Technical College (NATC) evaluated its training and development (T&D) process for new basic education instructors. The study gathered monitoring and screening criteria that addressed valuable components for use in an instrument for validating effectiveness of the ABE program (T&D)…
2012-07-01
SCI-228) Executive Summary This AGARDograph presents an introduction to fiber optic systems and is intended to provide a basic understanding of...22 / SCI-228) Synthèse Cette AGARDograph présente une introduction aux systèmes à fibres optiques et a pour objet d’expliquer l’utilisation de ces...l’instrumentation des essais en vol dans l’ensemble des pays membres de l’OTAN, entraînant la disparition progressive des jauges extensométriques et des
2015-11-09
and intelligence warfighting support. AFSPC operates sensors that provide direct attack warning and assessment to U.S. Strategic Command and North...combinations. AFRL conducted low-speed wind tunnel tests of 9%-scale model completed at NASA Langley Research Center (LaRC); data validated analytical...by $2M across JTAC platforms and expanding mobile device operation usage by 95 hours. The BATMAN-II team also delivered a new wireless mobile
2015-11-09
missile warning, weather and intelligence warfighting support. AFSPC operates sensors that provide direct attack warning and assessment to U.S...toughness combinations. AFRL conducted low-speed wind tunnel tests of 9%-scale model completed at NASA Langley Research Center (LaRC); data validated... wireless mobile monitoring capability designed for dismounted Pararescue Jumpers (PJ) called United States Air Force 89 Battlefield Airmen Trauma
1997-09-01
factor values are identified. For SASET, revised cost estimating relationships are provided ( Apgar et al., 1991). A 1991 AFIT thesis by Gerald Ourada...description of the model is a paragraph directly quoted from the user’s manual . This is not to imply that a lack of a thorough analysis indicates...constraints imposed by the system. The effective technology rating is computed from the basic technology rating by the following equation ( Apgar et al., 1991
BASIC BIOCHEMICAL AND CLINICAL ASPECTS OF NONINVASIVE TESTS HELIC.
Dmitrienko, M A; Dmitrienko, V S; Kornienko, E A; Parolova, N I; Colomina, E O; Aronov, E B
Biochemical process that lay in the core of non-invasive detection of Helico ho cter pylod with the help of HELIC Ammonia breath test, manufactured by AMA Co Ltd., St.Petersburg, is shown. Patents from various countries, describing ammonia as H.pyiori diagnostic marker, are reviewed. Approaches for evaluation of efficacy of the test-system are analyzed, validation and verification data is provided. High diagnostic characteristics are confirmed by the results of comparative studies on patients of different age groups, reaching 97% sensitivity and 96% specificity.
Tyl, Rochelle W.
2009-01-01
Background Myers et al. [Environ Health Perspect 117:309–315 (2009)] argued that Good Laboratory Practices (GLPs) cannot be used as a criterion for selecting data for risk assessment, using bisphenol A (BPA) as a case study. They did not discuss the role(s) of guideline-compliant studies versus basic/exploratory research studies, and they criticized both GLPs and guideline-compliant studies and their roles in formal hazard evaluation and risk assessment. They also specifically criticized our published guideline-compliant dietary studies on BPA in rats and mice and 17β-estradiol (E2) in mice. Objectives As the study director/first author of the criticized E2 and BPA studies, I discuss the uses of basic research versus guideline-compliant studies, how testing guidelines are developed and revised, how new end points are validated, and the role of GLPs. I also provide an overview of the BPA guideline-compliant and exploratory research animal studies and describe BPA pharmacokinetics in rats and humans. I present responses to specific criticisms by Myers et al. Discussion and conclusions Weight-of-evidence evaluations have consistently concluded that low-level BPA oral exposures do not adversely affect human developmental or reproductive health, and I encourage increased validation efforts for “new” end points for inclusion in guideline studies, as well as performance of robust long-term studies to follow early effects (observed in small exploratory studies) to any adverse consequences. PMID:20049112
Lauffer, A; Solé, L; Bernstein, S; Lopes, M H; Francisconi, C F
2013-01-01
The development and validation of questionnaires for evaluating quality of life (QoL) has become an important area of research. However, there is a proliferation of non-validated measuring instruments in the health setting that do not contribute to advances in scientific knowledge. To present, through the analysis of available validated questionnaires, a checklist of the practical aspects of how to carry out the cross-cultural adaptation of QoL questionnaires (generic, or disease-specific) so that no step is overlooked in the evaluation process, and thus help prevent the elaboration of insufficient or incomplete validations. We have consulted basic textbooks and Pubmed databases using the following keywords quality of life, questionnaires, and gastroenterology, confined to «validation studies» in English, Spanish, and Portuguese, and with no time limit, for the purpose of analyzing the translation and validation of the questionnaires available through the Mapi Institute and PROQOLID websites. A checklist is presented to aid in the planning and carrying out of the cross-cultural adaptation of QoL questionnaires, in conjunction with a glossary of key terms in the area of knowledge. The acronym DSTAC was used, which refers to each of the 5 stages involved in the recommended procedure. In addition, we provide a table of the QoL instruments that have been validated into Spanish. This article provides information on how to adapt QoL questionnaires from a cross-cultural perspective, as well as to minimize common errors. Copyright © 2012 Asociación Mexicana de Gastroenterología. Published by Masson Doyma México S.A. All rights reserved.
Evolving the future: Toward a science of intentional change
Wilson, David Sloan; Hayes, Steven C.; Biglan, Anthony; Embry, Dennis D.
2015-01-01
Humans possess great capacity for behavioral and cultural change, but our ability to manage change is still limited. This article has two major objectives: first, to sketch a basic science of intentional change centered on evolution; second, to provide examples of intentional behavioral and cultural change from the applied behavioral sciences, which are largely unknown to the basic sciences community. All species have evolved mechanisms of phenotypic plasticity that enable them to respond adaptively to their environments. Some mechanisms of phenotypic plasticity count as evolutionary processes in their own right. The human capacity for symbolic thought provides an inheritance system having the same kind of combinatorial diversity as does genetic recombination and antibody formation. Taking these propositions seriously allows an integration of major traditions within the basic behavioral sciences, such as behaviorism, social constructivism, social psychology, cognitive psychology, and evolutionary psychology, which are often isolated and even conceptualized as opposed to one another. The applied behavioral sciences include well-validated examples of successfully managing behavioral and cultural change at scales ranging from individuals to small groups to large populations. However, these examples are largely unknown beyond their disciplinary boundaries, for lack of a unifying theoretical framework. Viewed from an evolutionary perspective, they are examples of managing evolved mechanisms of phenotypic plasticity, including open-ended processes of variation and selection. Once the many branches of the basic and applied behavioral sciences become conceptually unified, we are closer to a science of intentional change than one might think. PMID:24826907
Cid, Luis; Lettnin, Carla; Stobäus, Claus; Monteiro, Diogo; Davoglio, Tárcia; Moutão, João
2016-02-19
The main propose of this study is the cross-cultural validation of the Basic Psychological Needs in Physical Education Scale among Portugal and Brazil samples, through the study of measurement model invariance. Participated in this study, two independent samples, one with 616 Portuguese students, of the 2nd and 3rd basic grade of public school, aged between 9 and 18 years old and another with 450 Brazilian students, from 1st, 2nd and 3rd middle grade of public and private school, aged between 14 and 18 years old. The results support the suitability of the model (three factors, each one with four items), showing an adequate fit to the data in each sample (Portugal: χ2 = 203.8, p = .001, df = 51, SRMR = .062, NNFI = .926, CFI = .943, RMSEA = .070, RMSEA 90% IC = .060-.080; Brazil: χ2 = 173.7, p = .001, df = 51, SRMR = .052, NNFI = .940, CFI = .954, RMSEA = .073, RMSEA 90% IC = .061-.085), as well valid cross-cultural invariance among Portuguese e Brazilian samples (∆CFI ≤ .01). Those findings allow us to conclude that scale can be used to measure the basic psychological needs in physical education in Portugal and in Brazil.
NASA Astrophysics Data System (ADS)
Mandigo Anggana Raras, Gustav
2018-04-01
This research aims to produce a product in the form of flash based interactive learning media on a basic electronic engineering subject that reliable to be used and to know students’ responses about the media. The target of this research is X-TEI 1 class at SMK Negeri 1 Driyorejo – Gresik. The method used in this study is R&D that has been limited into seven stages only (1) potential and problems, (2) data collection, (3) product design, (4) product validation, (5) product revision, (6) field test, and (7) analysis and writing. The obtained result is interactive learning media named MELDASH. Validation process used to produce a valid interactive learning media. The result of media validation state that the interactive learning media has a 90.83% rating. Students’ responses to this interactive learning media is really good with 88.89% rating.
AGENT-BASED MODELS IN EMPIRICAL SOCIAL RESEARCH*
Bruch, Elizabeth; Atwell, Jon
2014-01-01
Agent-based modeling has become increasingly popular in recent years, but there is still no codified set of recommendations or practices for how to use these models within a program of empirical research. This article provides ideas and practical guidelines drawn from sociology, biology, computer science, epidemiology, and statistics. We first discuss the motivations for using agent-based models in both basic science and policy-oriented social research. Next, we provide an overview of methods and strategies for incorporating data on behavior and populations into agent-based models, and review techniques for validating and testing the sensitivity of agent-based models. We close with suggested directions for future research. PMID:25983351
Advanced Technology Training System on Motor-Operated Valves
NASA Technical Reports Server (NTRS)
Wiederholt, Bradley J.; Widjaja, T. Kiki; Yasutake, Joseph Y.; Isoda, Hachiro
1993-01-01
This paper describes how features from the field of Intelligent Tutoring Systems are applied to the Motor-Operated Valve (MOV) Advanced Technology Training System (ATTS). The MOV ATTS is a training system developed at Galaxy Scientific Corporation for the Central Research Institute of Electric Power Industry in Japan and the Electric Power Research Institute in the United States. The MOV ATTS combines traditional computer-based training approaches with system simulation, integrated expert systems, and student and expert modeling. The primary goal of the MOV ATTS is to reduce human errors that occur during MOV overhaul and repair. The MOV ATTS addresses this goal by providing basic operational information of the MOV, simulating MOV operation, providing troubleshooting practice of MOV failures, and tailoring this training to the needs of each individual student. The MOV ATTS integrates multiple expert models (functional and procedural) to provide advice and feedback to students. The integration also provides expert model validation support to developers. Student modeling is supported by two separate student models: one model registers and updates the student's current knowledge of basic MOV information, while another model logs the student's actions and errors during troubleshooting exercises. These two models are used to provide tailored feedback to the student during the MOV course.
Assessment of construct validity of a virtual reality laparoscopy simulator.
Rosenthal, Rachel; Gantert, Walter A; Hamel, Christian; Hahnloser, Dieter; Metzger, Juerg; Kocher, Thomas; Vogelbach, Peter; Scheidegger, Daniel; Oertli, Daniel; Clavien, Pierre-Alain
2007-08-01
The aim of this study was to assess whether virtual reality (VR) can discriminate between the skills of novices and intermediate-level laparoscopic surgical trainees (construct validity), and whether the simulator assessment correlates with an expert's evaluation of performance. Three hundred and seven (307) participants of the 19th-22nd Davos International Gastrointestinal Surgery Workshops performed the clip-and-cut task on the Xitact LS 500 VR simulator (Xitact S.A., Morges, Switzerland). According to their previous experience in laparoscopic surgery, participants were assigned to the basic course (BC) or the intermediate course (IC). Objective performance parameters recorded by the simulator were compared to the standardized assessment by the course instructors during laparoscopic pelvitrainer and conventional surgery exercises. IC participants performed significantly better on the VR simulator than BC participants for the task completion time as well as the economy of movement of the right instrument, not the left instrument. Participants with maximum scores in the pelvitrainer cholecystectomy task performed the VR trial significantly faster, compared to those who scored less. In the conventional surgery task, a significant difference between those who scored the maximum and those who scored less was found not only for task completion time, but also for economy of movement of the right instrument. VR simulation provides a valid assessment of psychomotor skills and some basic aspects of spatial skills in laparoscopic surgery. Furthermore, VR allows discrimination between trainees with different levels of experience in laparoscopic surgery establishing construct validity for the Xitact LS 500 clip-and-cut task. Virtual reality may become the gold standard to assess and monitor surgical skills in laparoscopic surgery.
A novel augmented reality simulator for skills assessment in minimal invasive surgery.
Lahanas, Vasileios; Loukas, Constantinos; Smailis, Nikolaos; Georgiou, Evangelos
2015-08-01
Over the past decade, simulation-based training has come to the foreground as an efficient method for training and assessment of surgical skills in minimal invasive surgery. Box-trainers and virtual reality (VR) simulators have been introduced in the teaching curricula and have substituted to some extent the traditional model of training based on animals or cadavers. Augmented reality (AR) is a new technology that allows blending of VR elements and real objects within a real-world scene. In this paper, we present a novel AR simulator for assessment of basic laparoscopic skills. The components of the proposed system include: a box-trainer, a camera and a set of laparoscopic tools equipped with custom-made sensors that allow interaction with VR training elements. Three AR tasks were developed, focusing on basic skills such as perception of depth of field, hand-eye coordination and bimanual operation. The construct validity of the system was evaluated via a comparison between two experience groups: novices with no experience in laparoscopic surgery and experienced surgeons. The observed metrics included task execution time, tool pathlength and two task-specific errors. The study also included a feedback questionnaire requiring participants to evaluate the face-validity of the system. Between-group comparison demonstrated highly significant differences (<0.01) in all performance metrics and tasks denoting the simulator's construct validity. Qualitative analysis on the instruments' trajectories highlighted differences between novices and experts regarding smoothness and economy of motion. Subjects' ratings on the feedback questionnaire highlighted the face-validity of the training system. The results highlight the potential of the proposed simulator to discriminate groups with different expertise providing a proof of concept for the potential use of AR as a core technology for laparoscopic simulation training.
Knowledge representation in fuzzy logic
NASA Technical Reports Server (NTRS)
Zadeh, Lotfi A.
1989-01-01
The author presents a summary of the basic concepts and techniques underlying the application of fuzzy logic to knowledge representation. He then describes a number of examples relating to its use as a computational system for dealing with uncertainty and imprecision in the context of knowledge, meaning, and inference. It is noted that one of the basic aims of fuzzy logic is to provide a computational framework for knowledge representation and inference in an environment of uncertainty and imprecision. In such environments, fuzzy logic is effective when the solutions need not be precise and/or it is acceptable for a conclusion to have a dispositional rather than categorical validity. The importance of fuzzy logic derives from the fact that there are many real-world applications which fit these conditions, especially in the realm of knowledge-based systems for decision-making and control.
Miller, Joan G
2002-01-01
D. Oyserman, H. M. Coon, and M. Kemmelmeir (2002) offered a comprehensive literature review on individualism and collectivism that forwards valuable suggestions for ways to enhance future research conducted within this framework. The author argues that although their criticisms of much contemporary social psychological research on individualism and collectivism are valid, even more fundamental problems need to be recognized as characterizing work within this tradition, such as the insufficiently subtle nature of the views held of culture, the limited attention given to meanings, and the downplaying of contextual variation. The author suggests adopting more nuanced and process-oriented conceptions of culture and more contextually grounded views of its impact on psychological functioning as a way of realizing the promise of cultural psychology to broaden and provide insight into basic psychological theory.
Basic biostatistics for post-graduate students
Dakhale, Ganesh N.; Hiware, Sachin K.; Shinde, Abhijit T.; Mahatme, Mohini S.
2012-01-01
Statistical methods are important to draw valid conclusions from the obtained data. This article provides background information related to fundamental methods and techniques in biostatistics for the use of postgraduate students. Main focus is given to types of data, measurement of central variations and basic tests, which are useful for analysis of different types of observations. Few parameters like normal distribution, calculation of sample size, level of significance, null hypothesis, indices of variability, and different test are explained in detail by giving suitable examples. Using these guidelines, we are confident enough that postgraduate students will be able to classify distribution of data along with application of proper test. Information is also given regarding various free software programs and websites useful for calculations of statistics. Thus, postgraduate students will be benefitted in both ways whether they opt for academics or for industry. PMID:23087501
The use of quizStar application for online examination in basic physics course
NASA Astrophysics Data System (ADS)
Kustijono, R.; Budiningarti, H.
2018-03-01
The purpose of the study is to produce an online Basic Physics exam system using the QuizStar application. This is a research and development with ADDIE model. The steps are: 1) analysis; 2) design; 3) development; 4) implementation; 5) evaluation. System feasibility is reviewed for its validity, practicality, and effectiveness. The subjects of research are 60 Physics Department students of Universitas Negeri Surabaya. The data analysis used is a descriptive statistic. The validity, practicality, and effectiveness scores are measured using a Likert scale. Criteria feasible if the total score of all aspects obtained is ≥ 61%. The results obtained from the online test system by using QuizStar developed are 1) conceptually feasible to use; 2) the system can be implemented in the Basic Physics assessment process, and the existing constraints can be overcome; 3) student's response to system usage is in a good category. The results conclude that QuizStar application is eligible to be used for online Basic Physics exam system.
Current advances in esophageal cancer proteomics.
Uemura, Norihisa; Kondo, Tadashi
2015-06-01
We review the current status of proteomics for esophageal cancer (EC) from a clinician's viewpoint. The ultimate goal of cancer proteomics is the improvement of clinical outcome. The proteome as a functional translation of the genome is a straightforward representation of genomic mechanisms that trigger carcinogenesis. Cancer proteomics has identified the mechanisms of carcinogenesis and tumor progression, detected biomarker candidates for early diagnosis, and provided novel therapeutic targets for personalized treatments. Our review focuses on three major topics in EC proteomics: diagnostics, treatment, and molecular mechanisms. We discuss the major histological differences between EC types, i.e., esophageal squamous cell carcinoma and adenocarcinoma, and evaluate the clinical significance of published proteomics studies, including promising diagnostic biomarkers and novel therapeutic targets, which should be further validated prior to launching clinical trials. Multi-disciplinary collaborations between basic scientists, clinicians, and pathologists should be established for inter-institutional validation. In conclusion, EC proteomics has provided significant results, which after thorough validation, should lead to the development of novel clinical tools and improvement of the clinical outcome for esophageal cancer patients. This article is part of a Special Issue entitled: Medical Proteomics. Copyright © 2014 Elsevier B.V. All rights reserved.
Clinical decision rules for termination of resuscitation in out-of-hospital cardiac arrest.
Sherbino, Jonathan; Keim, Samuel M; Davis, Daniel P
2010-01-01
Out-of-hospital cardiac arrest (OHCA) has a low probability of survival to hospital discharge. Four clinical decision rules (CDRs) have been validated to identify patients with no probability of survival. Three of these rules focus on exclusive prehospital basic life support care for OHCA, and two of these rules focus on prehospital advanced life support care for OHCA. Can a CDR for the termination of resuscitation identify a patient with no probability of survival in the setting of OHCA? Six validation studies were selected from a PubMed search. A structured review of each of the studies is presented. In OHCA receiving basic life support care, the BLS-TOR (basic life support termination of resuscitation) rule has a positive predictive value for death of 99.5% (95% confidence interval 98.9-99.8%), and decreases the transportation of all patients by 62.6%. This rule has been appropriately validated for widespread use. In OHCA receiving advanced life support care, no current rule has been appropriately validated for widespread use. The BLS-TOR rule is a simple rule that identifies patients who will not survive OHCA. Further research is required to identify similarly robust CDRs for patients receiving advanced life support care in the setting of OHCA. Copyright 2010 Elsevier Inc. All rights reserved.
ERIC Educational Resources Information Center
College of the Canyons, Valencia, CA. Office of Institutional Development.
California's College of the Canyons has used the College Board Assessment and Placement Services (APS) test to assess students' abilities in basic and college English since spring 1993. These two reports summarize data from a May 1994 study of the predictive validity of the APS writing and reading tests and a June 1994 effort to validate the cut…
Methodology Series Module 9: Designing Questionnaires and Clinical Record Forms - Part II.
Setia, Maninder Singh
2017-01-01
This article is a continuation of the previous module on designing questionnaires and clinical record form in which we have discussed some basic points about designing the questionnaire and clinical record forms. In this section, we will discuss the reliability and validity of questionnaires. The different types of validity are face validity, content validity, criterion validity, and construct validity. The different types of reliability are test-retest reliability, inter-rater reliability, and intra-rater reliability. Some of these parameters are assessed by subject area experts. However, statistical tests should be used for evaluation of other parameters. Once the questionnaire has been designed, the researcher should pilot test the questionnaire. The items in the questionnaire should be changed based on the feedback from the pilot study participants and the researcher's experience. After the basic structure of the questionnaire has been finalized, the researcher should assess the validity and reliability of the questionnaire or the scale. If an existing standard questionnaire is translated in the local language, the researcher should assess the reliability and validity of the translated questionnaire, and these values should be presented in the manuscript. The decision to use a self- or interviewer-administered, paper- or computer-based questionnaire depends on the nature of the questions, literacy levels of the target population, and resources.
Methodology Series Module 9: Designing Questionnaires and Clinical Record Forms – Part II
Setia, Maninder Singh
2017-01-01
This article is a continuation of the previous module on designing questionnaires and clinical record form in which we have discussed some basic points about designing the questionnaire and clinical record forms. In this section, we will discuss the reliability and validity of questionnaires. The different types of validity are face validity, content validity, criterion validity, and construct validity. The different types of reliability are test-retest reliability, inter-rater reliability, and intra-rater reliability. Some of these parameters are assessed by subject area experts. However, statistical tests should be used for evaluation of other parameters. Once the questionnaire has been designed, the researcher should pilot test the questionnaire. The items in the questionnaire should be changed based on the feedback from the pilot study participants and the researcher's experience. After the basic structure of the questionnaire has been finalized, the researcher should assess the validity and reliability of the questionnaire or the scale. If an existing standard questionnaire is translated in the local language, the researcher should assess the reliability and validity of the translated questionnaire, and these values should be presented in the manuscript. The decision to use a self- or interviewer-administered, paper- or computer-based questionnaire depends on the nature of the questions, literacy levels of the target population, and resources. PMID:28584367
Eastwood, Sophie V.; Tillin, Therese; Wright, Andrew; Heasman, John; Willis, Joseph; Godsland, Ian F.; Forouhi, Nita; Whincup, Peter; Hughes, Alun D.; Chaturvedi, Nishi
2013-01-01
Background South Asians and African Caribbeans experience more cardiometabolic disease than Europeans. Risk factors include visceral (VAT) and subcutaneous abdominal (SAT) adipose tissue, which vary with ethnicity and are difficult to quantify using anthropometry. Objective We developed and cross-validated ethnicity and gender-specific equations using anthropometrics to predict VAT and SAT. Design 669 Europeans, 514 South Asians and 227 African Caribbeans (70±7 years) underwent anthropometric measurement and abdominal CT scanning. South Asian and African Caribbean participants were first-generation migrants living in London. Prediction equations were derived for CT-measured VAT and SAT using stepwise regression, then cross-validated by comparing actual and predicted means. Results South Asians had more and African Caribbeans less VAT than Europeans. For basic VAT prediction equations (age and waist circumference), model fit was better in men (R2 range 0.59-0.71) than women (range 0.35-0.59). Expanded equations (+ weight, height, hip and thigh circumference) improved fit for South Asian and African Caribbean women (R2 0.35 to 0.55, and 0.43 to 0.56 respectively). For basic SAT equations, R2 was 0.69-0.77, and for expanded equations it was 0.72-0.86. Cross-validation showed differences between actual and estimated VAT of <7%, and SAT of <8% in all groups, apart from VAT in South Asian women which disagreed by 16%. Conclusion We provide ethnicity- and gender-specific VAT and SAT prediction equations, derived from a large tri-ethnic sample. Model fit was reasonable for SAT and VAT in men, while basic VAT models should be used cautiously in South Asian and African Caribbean women. These equations will aid studies of mechanisms of cardiometabolic disease in later life, where imaging data are not available. PMID:24069381
Gill, Michael J.; Andreychik, Michael R.
2014-01-01
Why is he poor? Why is she failing academically? Why is he so generous? Why is she so conscientious? Answers to such everyday questions—social explanations—have powerful effects on relationships at the interpersonal and societal levels. How do people select an explanation in particular cases? We suggest that, often, explanations are selected based on the individual's pre-existing general theories of social causality. More specifically, we suggest that over time individuals develop general beliefs regarding the causes of social events. We refer to these beliefs as social explanatory styles. Our goal in the present article is to offer and validate a measure of individual differences in social explanatory styles. Accordingly, we offer the Social Explanatory Styles Questionnaire (SESQ), which measures three independent dimensions of social explanatory style: Dispositionism, historicism, and controllability. Studies 1–3 examine basic psychometric properties of the SESQ and provide positive evidence regarding internal consistency, factor structure, and both convergent and divergent validity. Studies 4–6 examine predictive validity for each subscale: Does each explanatory dimension moderate an important phenomenon of social cognition? Results suggest that they do. In Study 4, we show that SESQ dispositionism moderates the tendency to make spontaneous trait inferences. In Study 5, we show that SESQ historicism moderates the tendency to commit the Fundamental Attribution Error. Finally, in Study 6 we show that SESQ controllability predicts polarization of moral blame judgments: Heightened blaming toward controllable stigmas (assimilation), and attenuated blaming toward uncontrollable stigmas (contrast). Decades of research suggest that explanatory style regarding the self is a powerful predictor of self-functioning. We think it is likely that social explanatory styles—perhaps comprising interactive combinations of the basic dimensions tapped by the SESQ—will be similarly potent predictors of social functioning. We hope the SESQ will be a useful tool for exploring that possibility. PMID:25007152
Boubouchairopoulou, N; Kollias, A; Chiu, B; Chen, B; Lagou, S; Anestis, P; Stergiou, G S
2017-07-01
A pocket-size cuffless electronic device for self-measurement of blood pressure (BP) has been developed (Freescan, Maisense Inc., Zhubei, Taiwan). The device estimates BP within 10 s using three embedded electrodes and one force sensor that is applied over the radial pulse to evaluate the pulse wave. Before use, basic anthropometric characteristics are recorded on the device, and individualized initial calibration is required based on a standard BP measurement performed using an upper-arm BP monitor. The device performance in providing valid BP readings was evaluated in 313 normotensive and hypertensive adults in three study phases during which the device sensor was upgraded. A formal validation study of a prototype device against mercury sphygmomanometer was performed according to the American National Standards Institute/Association for the Advancement of Medical Instrumentation/International Organization for Standardization (ANSI/AAMI/ISO) 2013 protocol. The test device succeeded in obtaining a valid BP measurement (three successful readings within up to five attempts) in 55-72% of the participants, which reached 87% with device sensor upgrade. For the validation study, 125 adults were recruited and 85 met the protocol requirements for inclusion. The mean device-observers BP difference was 3.2±6.7 (s.d.) mm Hg for systolic and 2.6±4.6 mm Hg for diastolic BP (criterion 1). The estimated s.d. (inter-subject variability) were 5.83 and 4.17 mm Hg respectively (criterion 2). These data suggest that this prototype cuffless BP monitor provides valid self-measurements in the vast majority of adults, and satisfies the BP measurement accuracy criteria of the ANSI/AAMI/ISO 2013 validation protocol.
Validation of software for calculating the likelihood ratio for parentage and kinship.
Drábek, J
2009-03-01
Although the likelihood ratio is a well-known statistical technique, commercial off-the-shelf (COTS) software products for its calculation are not sufficiently validated to suit general requirements for the competence of testing and calibration laboratories (EN/ISO/IEC 17025:2005 norm) per se. The software in question can be considered critical as it directly weighs the forensic evidence allowing judges to decide on guilt or innocence or to identify person or kin (i.e.: in mass fatalities). For these reasons, accredited laboratories shall validate likelihood ratio software in accordance with the above norm. To validate software for calculating the likelihood ratio in parentage/kinship scenarios I assessed available vendors, chose two programs (Paternity Index and familias) for testing, and finally validated them using tests derived from elaboration of the available guidelines for the field of forensics, biomedicine, and software engineering. MS Excel calculation using known likelihood ratio formulas or peer-reviewed results of difficult paternity cases were used as a reference. Using seven testing cases, it was found that both programs satisfied the requirements for basic paternity cases. However, only a combination of two software programs fulfills the criteria needed for our purpose in the whole spectrum of functions under validation with the exceptions of providing algebraic formulas in cases of mutation and/or silent allele.
Domínguez, Evelia; Martín, Patricia; Martín-Albo, José; Núñez, Juan L; León, Jaime
2010-11-01
The aim of the present research was to translate and to analyze the psychometric properties of the Spanish version of the Satisfaction of Psychological Needs Scale, using a sample of 284 athletes (204 male and 78 female). Results of the confirmatory factor analysis confirmed the correlated three-factor structure of the scale. Furthermore, the results showed evidence of convergence validity with the Basic Psychological Needs in Exercise Scale. The predictive validity was tested using a structural equation model in which task orientation climate predicted the three basic psychological needs and these, in turn, intrinsic motivation. Likewise, we documented evidence of reliability, analyzed as internal consistency and temporal stability. Results partially support the use of the Spanish version of the scale in sports.
NIST TXR Validation of S-HIS radiances and a UW-SSEC Blackbody
NASA Astrophysics Data System (ADS)
Taylor, J. K.; O'Connell, J.; Rice, J. P.; Revercomb, H. E.; Best, F. A.; Tobin, D. C.; Knuteson, R. O.; Adler, D. P.; Ciganovich, N. C.; Dutcher, S. T.; Laporte, D. D.; Ellington, S. D.; Werner, M. W.; Garcia, R. K.
2007-12-01
The ability to accurately validate infrared spectral radiances measured from space by direct comparison with airborne spectrometer radiances was first demonstrated using the Scanning High-resolution Interferometer Sounder (S-HIS) aircraft instrument flown under the AIRS on the NASA Aqua spacecraft in 2002 with subsequent successful comparisons in 2004 and 2006. The comparisons span a range of conditions, including arctic and tropical atmospheres, daytime and nighttime, and ocean and land surfaces. Similar comprehensive and successful comparisons have also been conducted with S-HIS for the MODIS sensors, the Tropospheric Emission Spectrometer (TES), and most recently the MetOp Infrared Atmospheric Sounding Interferometer (IASI). These comparisons are part of a larger picture that already shows great progress toward transforming our ability to make, and verify, highly accurate spectral radiance observations from space. A key challenge, especially for climate, is to carefully define the absolute accuracy of satellite radiances. Our vision of the near-term future of spectrally resolved infrared radiance observation includes a new space-borne mission that provides benchmark observations of the emission spectrum for climate. This concept, referred to as the CLimate Absolute Radiance and REfractivity Observatory (CLARREO) in the recent NRC Decadal Survey provides more complete spectral and time-of-day coverage and would fly basic physical standards to eliminate the need to assume on-board reference stability. Therefore, the spectral radiances from this mission will also serve as benchmarks to propagate a highly accurate calibration to other space-borne IR instruments. For the current approach of calibrating infrared flight sensors, in which thermal vacuum tests are conducted before launch and stability is assumed after launch, in-flight calibration validation is essential for highly accurate applications. At present, airborne observations provide the only source of direct radiance validation with resulting traceable uncertainties approaching the level required for remote sensing and climate applications (0.1 K 3- sigma). For the calibration validation process to be accurate, repeatable, and meaningful, the reference instrument must be extremely well characterized and understood, carefully maintained, and accurately calibrated, with the calibration accuracy of the reference instrument tied to absolute standards. Tests of the S-HIS absolute calibration have been conducted using the NIST transfer radiometer (TXR). The TXR provides a more direct connection to the Blackbody reference sources maintained by NIST than the normal traceability of blackbody temperature scales and paint emissivity measurements. Two basic tests were conducted: (1) comparison of radiances measured by the S-HIS to those from the TXR, and (2) measuring the reflectivity of a UW-SSEC blackbody by using the TXR as a stable detector. Preliminary results from both tests are very promising for confirming and refining the expected absolute accuracy of the S-HIS.
Mathematical foundations of biomechanics.
Niederer, Peter F
2010-01-01
The aim of biomechanics is the analysis of the structure and function of humans, animals, and plants by means of the methods of mechanics. Its foundations are in particular embedded in mathematics, physics, and informatics. Due to the inherent multidisciplinary character deriving from its aim, biomechanics has numerous connections and overlapping areas with biology, biochemistry, physiology, and pathophysiology, along with clinical medicine, so its range is enormously wide. This treatise is mainly meant to serve as an introduction and overview for readers and students who intend to acquire a basic understanding of the mathematical principles and mechanics that constitute the foundation of biomechanics; accordingly, its contents are limited to basic theoretical principles of general validity and long-range significance. Selected examples are included that are representative for the problems treated in biomechanics. Although ultimate mathematical generality is not in the foreground, an attempt is made to derive the theory from basic principles. A concise and systematic formulation is thereby intended with the aim that the reader is provided with a working knowledge. It is assumed that he or she is familiar with the principles of calculus, vector analysis, and linear algebra.
The basic aerodynamics of floatation
DOE Office of Scientific and Technical Information (OSTI.GOV)
Davies, M.J.; Wood, D.H.
1983-09-01
The original derivation of the basic theory governing the aerodynamics of both hovercraft and modern floatation ovens, requires the validity of some extremely crude assumptions. However, the basic theory is surprisingly accurate. It is shown that this accuracy occurs because the final expression of the basic theory can be derived by approximating the full Navier-Stokes equations in a manner that clearly shows the limitations of the theory. These limitations are used in discussing the relatively small discrepancies between the theory and experiment, which may not be significant for practical purposes.
Digital Fly-By-Wire Flight Control Validation Experience
NASA Technical Reports Server (NTRS)
Szalai, K. J.; Jarvis, C. R.; Krier, G. E.; Megna, V. A.; Brock, L. D.; Odonnell, R. N.
1978-01-01
The experience gained in digital fly-by-wire technology through a flight test program being conducted by the NASA Dryden Flight Research Center in an F-8C aircraft is described. The system requirements are outlined, along with the requirements for flight qualification. The system is described, including the hardware components, the aircraft installation, and the system operation. The flight qualification experience is emphasized. The qualification process included the theoretical validation of the basic design, laboratory testing of the hardware and software elements, systems level testing, and flight testing. The most productive testing was performed on an iron bird aircraft, which used the actual electronic and hydraulic hardware and a simulation of the F-8 characteristics to provide the flight environment. The iron bird was used for sensor and system redundancy management testing, failure modes and effects testing, and stress testing in many cases with the pilot in the loop. The flight test program confirmed the quality of the validation process by achieving 50 flights without a known undetected failure and with no false alarms.
Preliminary development of the adolescent students' basic psychological needs at school scale.
Tian, Lili; Han, Mengmeng; Huebner, E Scott
2014-04-01
The aim of the present study was to develop and provide evidence for the validity of a new measure of adolescent students' psychological need satisfaction at school, using a sample of Chinese students. We conducted four studies with four independent samples (total n = 1872). The first study aimed to develop items for the new instrument and to ascertain its factorial structure using exploratory factor analysis procedures. The second study aimed to examine the instrument's factorial structure using confirmatory factor analysis procedures as well as to assess its internal consistency reliability, convergent and divergent validity. The third study aimed to assess its measurement invariance across gender and age. The fourth study aimed to test its test-retest reliability over time and predictive validity. These preliminary results showed that the new instrument has promising psychometric properties. The potential contributions of the new instrument for future research and educational practices were discussed. Copyright © 2014 The Foundation for Professionals in Services for Adolescents. Published by Elsevier Ltd. All rights reserved.
Latzman, Robert D.; Drislane, Laura E.; Hecht, Lisa K.; Brislin, Sarah J.; Patrick, Christopher J.; Lilienfeld, Scott O.; Freeman, Hani J.; Schapiro, Steven J.; Hopkins, William D.
2015-01-01
The current work sought to operationalize constructs of the triarchic model of psychopathy in chimpanzees (Pan troglodytes), a species well-suited for investigations of basic biobehavioral dispositions relevant to psychopathology. Across three studies, we generated validity evidence for scale measures of the triarchic model constructs in a large sample (N=238) of socially-housed chimpanzees. Using a consensus-based rating approach, we first identified candidate items for the chimpanzee triarchic (CHMP-Tri) scales from an existing primate personality instrument and refined these into scales. In Study 2, we collected data for these scales from human informants (N=301), and examined their convergent and divergent relations with scales from another triarchic inventory developed for human use. In Study 3, we undertook validation work examining associations between CHMP-Tri scales and task measures of approach-avoidance behavior (N=73) and ability to delay gratification (N=55). Current findings provide support for a chimpanzee model of core dispositions relevant to psychopathy and other forms of psychopathology. PMID:26779396
LaBudde, Robert A; Harnly, James M
2012-01-01
A qualitative botanical identification method (BIM) is an analytical procedure that returns a binary result (1 = Identified, 0 = Not Identified). A BIM may be used by a buyer, manufacturer, or regulator to determine whether a botanical material being tested is the same as the target (desired) material, or whether it contains excessive nontarget (undesirable) material. The report describes the development and validation of studies for a BIM based on the proportion of replicates identified, or probability of identification (POI), as the basic observed statistic. The statistical procedures proposed for data analysis follow closely those of the probability of detection, and harmonize the statistical concepts and parameters between quantitative and qualitative method validation. Use of POI statistics also harmonizes statistical concepts for botanical, microbiological, toxin, and other analyte identification methods that produce binary results. The POI statistical model provides a tool for graphical representation of response curves for qualitative methods, reporting of descriptive statistics, and application of performance requirements. Single collaborator and multicollaborative study examples are given.
Takamura, Ayari; Watanabe, Ken; Akutsu, Tomoko
2017-07-01
Identification of human semen is indispensable for the investigation of sexual assaults. Fluorescence staining methods using commercial kits, such as the series of SPERM HY-LITER™ kits, have been useful to detect human sperm via strong fluorescence. These kits have been examined from various forensic aspects. However, because of a lack of evaluation methods, these studies did not provide objective, or quantitative, descriptions of the results nor clear criteria for the decisions reached. In addition, the variety of validations was considerably limited. In this study, we conducted more advanced validations of SPERM HY-LITER™ Express using our established image analysis method. Use of this method enabled objective and specific identification of fluorescent sperm's spots and quantitative comparisons of the sperm detection performance under complex experimental conditions. For body fluid mixtures, we examined interference with the fluorescence staining from other body fluid components. Effects of sample decomposition were simulated in high humidity and high temperature conditions. Semen with quite low sperm concentrations, such as azoospermia and oligospermia samples, represented the most challenging cases in application of the kit. Finally, the tolerance of the kit against various acidic and basic environments was analyzed. The validations herein provide useful information for the practical applications of the SPERM HY-LITER™ Express kit, which were previously unobtainable. Moreover, the versatility of our image analysis method toward various complex cases was demonstrated.
Zhao, Jin-yao; Liu, Chun-qing; Zhao, He-nan; Ding, Yan-Fang; Bi, Tie; Wang, Bo; Lin, Xing-chi; Guo, Gordon; Cui, Shi-ying
2012-10-01
After discovering new miRNAs, it is often difficult to determine their targets and effects on downstream protein expression. In situ hybridization (ISH) and immunohistochemistry (IHC) are two commonly used methods for clinical diagnosis and basic research. We used an optimized technique that simultaneously detects miRNAs, their binding targets and corresponding proteins on transferred serial formalin fixed paraffin embedded (FFPE) sections from patients. Combined with bioinformatics, this method was used to validate the reciprocal expression of specific miRNAs and targets that were detected by ISH, as well as the expression of downstream proteins that were detected by IHC. A complete analysis was performed using a limited number of transferred serial FFPE sections that had been stored for 1-4 years at room temperature. Some sections had even been previously stained with H&E. We identified a miRNA that regulates epithelial ovarian cancer, along with its candidate target and related downstream protein. These findings were directly validated using sub-cellular components obtained from the same patient sample. In addition, the expression of Nephrin (a podocyte marker) and Stmn1 (a recently identified marker related to glomerular development) were confirmed in transferred FFPE sections of mouse kidney. This procedure may be adapted for clinical diagnosis and basic research, providing a qualitative and efficient method to dissect the detailed spatial expression patterns of miRNA pathways in FFPE tissue, especially in cases where only a small biopsy sample can be obtained. Copyright © 2012 Elsevier Inc. All rights reserved.
Developments in Sensitivity Methodologies and the Validation of Reactor Physics Calculations
Palmiotti, Giuseppe; Salvatores, Massimo
2012-01-01
The sensitivity methodologies have been a remarkable story when adopted in the reactor physics field. Sensitivity coefficients can be used for different objectives like uncertainty estimates, design optimization, determination of target accuracy requirements, adjustment of input parameters, and evaluations of the representativity of an experiment with respect to a reference design configuration. A review of the methods used is provided, and several examples illustrate the success of the methodology in reactor physics. A new application as the improvement of nuclear basic parameters using integral experiments is also described.
Experimental investigation of hypersonic aerodynamics
NASA Technical Reports Server (NTRS)
Heinemann, K.; Intrieri, Peter F.
1987-01-01
An extensive series of ballistic range tests are currently being conducted at the Ames Research Center. These tests are intended to investigate the hypersonic aerodynamic characteristics of two basic configurations, which are: the blunt-cone Galileo probe which is scheduled to be launched in late 1989 and will enter the atmosphere of Jupiter in 1994, and a generic slender cone configuration to provide experimental aerodynamic data including good flow-field definition which computational aerodynamicists could use to validate their computer codes. Some of the results obtained thus far are presented and work for the near future is discussed.
Violation of the 2nd Law of Thermodynamics in the Quantum Microworld
NASA Astrophysics Data System (ADS)
Čápek, V.; Frege, O.
2002-05-01
For one open quantum system recently reported to work as a perpetuum mobile of the second kind, basic equations providing basis for discussion of physics beyond the system activity are rederived in an appreciably simpler manner. The equations become exact in one specific scaling limit corresponding to the physical regime where internal processes (relaxations) in the system are commensurable or even slower than relaxation processes induced by bath. In the high-temperature (i.e. classical) limit, the system ceases to work, i.e., validity of the second law is reestablished.
Baseline mathematics and geodetics for tracking operations
NASA Technical Reports Server (NTRS)
James, R.
1981-01-01
Various geodetic and mapping algorithms are analyzed as they apply to radar tracking systems and tested in extended BASIC computer language for real time computer applications. Closed-form approaches to the solution of converting Earth centered coordinates to latitude, longitude, and altitude are compared with classical approximations. A simplified approach to atmospheric refractivity called gradient refraction is compared with conventional ray tracing processes. An extremely detailed set of documentation which provides the theory, derivations, and application of algorithms used in the programs is included. Validation methods are also presented for testing the accuracy of the algorithms.
Pozos-Guillén, Amaury; Ruiz-Rodríguez, Socorro; Garrocho-Rangel, Arturo
The main purpose of the second part of this series was to provide the reader with some basic aspects of the most common biostatistical methods employed in health sciences, in order to better understand the validity, significance and reliability of the results from any article on Pediatric Dentistry. Currently, as mentioned in the first paper, Pediatric Dentists need basic biostatistical knowledge to be able to apply it when critically appraise a dental article during the Evidence-based Dentistry (EBD) process, or when participating in the development of a clinical study with dental pediatric patients. The EBD process provides a systematic approach of collecting, review and analyze current and relevant published evidence about oral health care in order to answer a particular clinical question; then this evidence should be applied in everyday practice. This second report describes the most commonly used statistical methods for analyzing and interpret collected data, and the methodological criteria to be considered when choosing the most appropriate tests for a specific study. These are available to Pediatric Dentistry practicants interested in reading or designing original clinical or epidemiological studies.
Liu, Nehemiah T; Holcomb, John B; Wade, Charles E; Batchinsky, Andriy I; Cancio, Leopoldo C; Darrah, Mark I; Salinas, José
2014-02-01
Accurate and effective diagnosis of actual injury severity can be problematic in trauma patients. Inherent physiologic compensatory mechanisms may prevent accurate diagnosis and mask true severity in many circumstances. The objective of this project was the development and validation of a multiparameter machine learning algorithm and system capable of predicting the need for life-saving interventions (LSIs) in trauma patients. Statistics based on means, slopes, and maxima of various vital sign measurements corresponding to 79 trauma patient records generated over 110,000 feature sets, which were used to develop, train, and implement the system. Comparisons among several machine learning models proved that a multilayer perceptron would best implement the algorithm in a hybrid system consisting of a machine learning component and basic detection rules. Additionally, 295,994 feature sets from 82 h of trauma patient data showed that the system can obtain 89.8 % accuracy within 5 min of recorded LSIs. Use of machine learning technologies combined with basic detection rules provides a potential approach for accurately assessing the need for LSIs in trauma patients. The performance of this system demonstrates that machine learning technology can be implemented in a real-time fashion and potentially used in a critical care environment.
[A functional analysis of healthcare auditors' skills in Venezuela, 2008].
Chirinos-Muñoz, Mónica S
2010-10-01
Using functional analysis for identifying the basic, working, specific and generic skills and values which a health service auditor must have. Implementing the functional analysis technique with 10 experts, identifying specific, basic, generic skills and values by means of deductive logic. A functional map was obtained which started by establishing a key purpose based on improving healthcare and service quality from which three key functions emerged. The main functions and skills' units were then broken down into the competitive elements defining what a health service auditor is able to do. This functional map (following functional analysis methodology) shows in detail the simple and complex tasks which a healthcare auditor should apply in the workplace, adopting a forward management approach for improving healthcare and health service quality. This methodology, expressing logical-deductive awareness raising, provides expert consensual information validating each element regarding overall skills.
Semi-supervised Learning for Phenotyping Tasks.
Dligach, Dmitriy; Miller, Timothy; Savova, Guergana K
2015-01-01
Supervised learning is the dominant approach to automatic electronic health records-based phenotyping, but it is expensive due to the cost of manual chart review. Semi-supervised learning takes advantage of both scarce labeled and plentiful unlabeled data. In this work, we study a family of semi-supervised learning algorithms based on Expectation Maximization (EM) in the context of several phenotyping tasks. We first experiment with the basic EM algorithm. When the modeling assumptions are violated, basic EM leads to inaccurate parameter estimation. Augmented EM attenuates this shortcoming by introducing a weighting factor that downweights the unlabeled data. Cross-validation does not always lead to the best setting of the weighting factor and other heuristic methods may be preferred. We show that accurate phenotyping models can be trained with only a few hundred labeled (and a large number of unlabeled) examples, potentially providing substantial savings in the amount of the required manual chart review.
Dementia diagnoses from clinical and neuropsychological data compared: the Cache County study.
Tschanz, J T; Welsh-Bohmer, K A; Skoog, I; West, N; Norton, M C; Wyse, B W; Nickles, R; Breitner, J C
2000-03-28
To validate a neuropsychological algorithm for dementia diagnosis. We developed a neuropsychological algorithm in a sample of 1,023 elderly residents of Cache County, UT. We compared algorithmic and clinical dementia diagnoses both based on DSM-III-R criteria. The algorithm diagnosed dementia when there was impairment in memory and at least one other cognitive domain. We also tested a variant of the algorithm that incorporated functional measures that were based on structured informant reports. Of 1,023 participants, 87% could be classified by the basic algorithm, 94% when functional measures were considered. There was good concordance between basic psychometric and clinical diagnoses (79% agreement, kappa = 0.57). This improved after incorporating functional measures (90% agreement, kappa = 0.76). Neuropsychological algorithms may reasonably classify individuals on dementia status across a range of severity levels and ages and may provide a useful adjunct to clinical diagnoses in population studies.
Cloud and Thermodynamic Parameters Retrieved from Satellite Ultraspectral Infrared Measurements
NASA Technical Reports Server (NTRS)
Zhou, Daniel K.; Smith, William L.; Larar, Allen M.; Liu, Xu; Taylor, Jonathan P.; Schluessel, Peter; Strow, L. Larrabee; Mango, Stephen A.
2008-01-01
Atmospheric-thermodynamic parameters and surface properties are basic meteorological parameters for weather forecasting. A physical geophysical parameter retrieval scheme dealing with cloudy and cloud-free radiance observed with satellite ultraspectral infrared sounders has been developed and applied to the Infrared Atmospheric Sounding Interferometer (IASI) and the Atmospheric InfraRed Sounder (AIRS). The retrieved parameters presented herein are from radiance data gathered during the Joint Airborne IASI Validation Experiment (JAIVEx). JAIVEx provided intensive aircraft observations obtained from airborne Fourier Transform Spectrometer (FTS) systems, in-situ measurements, and dedicated dropsonde and radiosonde measurements for the validation of the IASI products. Here, IASI atmospheric profile retrievals are compared with those obtained from dedicated dropsondes, radiosondes, and the airborne FTS system. The IASI examples presented here demonstrate the ability to retrieve fine-scale horizontal features with high vertical resolution from satellite ultraspectral sounder radiance spectra.
Synthetic Jet Flow Field Database for CFD Validation
NASA Technical Reports Server (NTRS)
Yao, Chung-Sheng; Chen, Fang Jenq; Neuhart, Dan; Harris, Jerome
2004-01-01
An oscillatory zero net mass flow jet was generated by a cavity-pumping device, namely a synthetic jet actuator. This basic oscillating jet flow field was selected as the first of the three test cases for the Langley workshop on CFD Validation of Synthetic Jets and Turbulent Separation Control. The purpose of this workshop was to assess the current CFD capabilities to predict unsteady flow fields of synthetic jets and separation control. This paper describes the characteristics and flow field database of a synthetic jet in a quiescent fluid. In this experiment, Particle Image Velocimetry (PIV), Laser Doppler Velocimetry (LDV), and hot-wire anemometry were used to measure the jet velocity field. In addition, the actuator operating parameters including diaphragm displacement, internal cavity pressure, and internal cavity temperature were also documented to provide boundary conditions for CFD modeling.
Using Model Replication to Improve the Reliability of Agent-Based Models
NASA Astrophysics Data System (ADS)
Zhong, Wei; Kim, Yushim
The basic presupposition of model replication activities for a computational model such as an agent-based model (ABM) is that, as a robust and reliable tool, it must be replicable in other computing settings. This assumption has recently gained attention in the community of artificial society and simulation due to the challenges of model verification and validation. Illustrating the replication of an ABM representing fraudulent behavior in a public service delivery system originally developed in the Java-based MASON toolkit for NetLogo by a different author, this paper exemplifies how model replication exercises provide unique opportunities for model verification and validation process. At the same time, it helps accumulate best practices and patterns of model replication and contributes to the agenda of developing a standard methodological protocol for agent-based social simulation.
Transport Phenomena During Equiaxed Solidification of Alloys
NASA Technical Reports Server (NTRS)
Beckermann, C.; deGroh, H. C., III
1997-01-01
Recent progress in modeling of transport phenomena during dendritic alloy solidification is reviewed. Starting from the basic theorems of volume averaging, a general multiphase modeling framework is outlined. This framework allows for the incorporation of a variety of microscale phenomena in the macroscopic transport equations. For the case of diffusion dominated solidification, a simplified set of model equations is examined in detail and validated through comparisons with numerous experimental data for both columnar and equiaxed dendritic growth. This provides a critical assessment of the various model assumptions. Models that include melt flow and solid phase transport are also discussed, although their validation is still at an early stage. Several numerical results are presented that illustrate some of the profound effects of convective transport on the final compositional and structural characteristics of a solidified part. Important issues that deserve continuing attention are identified.
Establishing the Content Validity of a Basic Computer Literacy Course.
ERIC Educational Resources Information Center
Clements, James; Carifio, James
1995-01-01
Content analysis of 13 textbooks and 2 Department of Education documents was conducted to ascertain common word processing, database, and spreadsheet software skills in order to determine which specific skills should be taught in a high school computer literacy course. Aspects of a basic computer course, created from this analysis, are described.…
ERIC Educational Resources Information Center
Scheuer, Claude; Bund, Andreas; Becker, Werner; Herrmann, Christian
2017-01-01
Basic motor competencies (in German: Motorische Basiskompetenzen; MOBAK) are motor performance dispositions formulated as minimum standards that empower children to participate in the culture of human movement. In opposition to movement-specific and process-oriented fundamental movement skills assessing the quality of movement execution, basic…
Reading Component Skills of Learners in Adult Basic Education
ERIC Educational Resources Information Center
MacArthur, Charles A.; Konold, Timothy R.; Glutting, Joseph J.; Alamprese, Judith A.
2010-01-01
The purposes of this study were to investigate the reliability and construct validity of measures of reading component skills with a sample of adult basic education (ABE) learners, including both native and nonnative English speakers, and to describe the performance of those learners on the measures. Investigation of measures of reading components…
Development of a Child Abuse Checklist to Evaluate Prehospital Provider Performance.
Alphonso, Aimee; Auerbach, Marc; Bechtel, Kirsten; Bilodeau, Kyle; Gawel, Marcie; Koziel, Jeannette; Whitfill, Travis; Tiyyagura, Gunjan Kamdar
2017-01-01
To develop and provide validity evidence for a performance checklist to evaluate the child abuse screening behaviors of prehospital providers. Checklist Development: We developed the first iteration of the checklist after review of the relevant literature and on the basis of the authors' clinical experience. Next, a panel of six content experts participated in three rounds of Delphi review to reach consensus on the final checklist items. Checklist Validation: Twenty-eight emergency medical services (EMS) providers (16 EMT-Basics, 12 EMT-Paramedics) participated in a standardized simulated case of physical child abuse to an infant followed by one-on-one semi-structured qualitative interviews. Three reviewers scored the videotaped performance using the final checklist. Light's kappa and Cronbach's alpha were calculated to assess inter-rater reliability (IRR) and internal consistency, respectively. The correlation of successful child abuse screening with checklist task completion and with participant characteristics were compared using Pearson's chi squared test to gather evidence for construct validity. The Delphi review process resulted in a final checklist that included 24 items classified with trichotomous scoring (done, not done, or not applicable). The overall IRR of the three raters was 0.70 using Light's kappa, indicating substantial agreement. Internal consistency of the checklist was low, with an overall Cronbach's alpha of 0.61. Of 28 participants, only 14 (50%) successfully screened for child abuse in simulation. Participants who successfully screened for child abuse did not differ significantly from those who failed to screen in terms of training level, past experience with child abuse reporting, or self-reported confidence in detecting child abuse (all p > 0.30). Of all 24 tasks, only the task of exposing the infant significantly correlated with successful detection of child abuse (p < 0.05). We developed a child abuse checklist that demonstrated strong content validity and substantial inter-rater reliability, but successful item completion did not correlate with other markers of provider experience. The validated instrument has important potential for training, continuing education, and research for prehospital providers at all levels of training.
Development and Validation of Coaches' Interpersonal Style Questionnaire
ERIC Educational Resources Information Center
Pulido, Juan J.; Sánchez-Oliva, David; Leo, Francisco M.; Sánchez-Cano, Jorge; García-Calvo, Tomás
2018-01-01
Purpose: The objectives were to develop and validate the Coaches' Interpersonal Style Questionnaire. The Coaches' Interpersonal Style Questionnaire analyzes the interpersonal style adopted by coaches when implementing their strategy of supporting or thwarting athletes' basic psychological needs. Method: In Study 1, an exploratory factor analysis…
Application of a framework to assess the usefulness of alternative sepsis criteria
Seymour, Christopher W.; Coopersmith, Craig M.; Deutschman, Clifford S; Gesten, Foster; Klompas, Michael; Levy, Mitchell; Martin, Gregory S.; Osborn, Tiffany M.; Rhee, Chanu; Warren, David; Watson, R. Scott; Angus, Derek C.
2016-01-01
The current definition for sepsis is life-threatening, acute organ dysfunction secondary to a dysregulated host response to infection. Criteria to operationalize this definition can be judged by 6 domains of usefulness (reliability; content, construct and criterion validity, measurement burden, and timeliness). The relative importance of these 6 domains depends on the intended purpose for the criteria (clinical care, basic and clinical research, surveillance, or quality improvement (QI) and audit). For example, criteria for clinical care should have high content and construct validity, timeliness, and low measurement burden to facilitate prompt care. Criteria for surveillance or QI/audit place greater emphasis on reliability across individuals and sites and lower emphasis on timeliness. Criteria for clinical trials require timeliness to ensure prompt enrollment and reasonable reliability but can tolerate high measurement burden. Basic research also tolerates high measurement burden and may not need stability over time. In an illustrative case study, we compared examples of criteria designed for clinical care, surveillance and QI/audit among 396,241 patients admitted to 12 academic and community hospitals in an integrated health system. Case rates differed 4-fold and mortality 3-fold. Predictably, clinical care criteria, which emphasized timeliness and low burden and therefore used vital signs and routine laboratory tests, had the highest case identification with lowest mortality. QI /audit criteria, which emphasized reliability and criterion validity, used discharge information and had the lowest case identification with highest mortality. Using this framework to identify the purpose and apply domains of usefulness can help with the evaluation of existing sepsis diagnostic criteria and provide a roadmap for future work. PMID:26901560
Application of a Framework to Assess the Usefulness of Alternative Sepsis Criteria.
Seymour, Christopher W; Coopersmith, Craig M; Deutschman, Clifford S; Gesten, Foster; Klompas, Michael; Levy, Mitchell; Martin, Gregory S; Osborn, Tiffany M; Rhee, Chanu; Warren, David K; Watson, R Scott; Angus, Derek C
2016-03-01
The current definition of sepsis is life-threatening, acute organ dysfunction secondary to a dysregulated host response to infection. Criteria to operationalize this definition can be judged by six domains of usefulness (reliability, content, construct and criterion validity, measurement burden, and timeliness). The relative importance of these six domains depends on the intended purpose for the criteria (clinical care, basic and clinical research, surveillance, or quality improvement [QI] and audit). For example, criteria for clinical care should have high content and construct validity, timeliness, and low measurement burden to facilitate prompt care. Criteria for surveillance or QI/audit place greater emphasis on reliability across individuals and sites and lower emphasis on timeliness. Criteria for clinical trials require timeliness to ensure prompt enrollment and reasonable reliability but can tolerate high measurement burden. Basic research also tolerates high measurement burden and may not need stability over time. In an illustrative case study, we compared examples of criteria designed for clinical care, surveillance and QI/audit among 396,241 patients admitted to 12 academic and community hospitals in an integrated health system. Case rates differed four-fold and mortality three-fold. Predictably, clinical care criteria, which emphasized timeliness and low burden and therefore used vital signs and routine laboratory tests, had the greater case identification with lowest mortality. QI/audit criteria, which emphasized reliability and criterion validity, used discharge information and had the lowest case identification with highest mortality. Using this framework to identify the purpose and apply domains of usefulness can help with the evaluation of existing sepsis diagnostic criteria and provide a roadmap for future work.
Wilson-Sands, Cathy; Brahn, Pamela; Graves, Kristal
2015-01-01
Validating participants' ability to correctly perform cardiopulmonary resuscitation (CPR) skills during basic life support courses can be a challenge for nursing professional development specialists. This study compares two methods of basic life support training, instructor-led and computer-based learning with voice-activated manikins, to identify if one method is more effective for performance of CPR skills. The findings suggest that a computer-based learning course with voice-activated manikins is a more effective method of training for improved CPR performance.
Emergency medical services capacity for prehospital stroke care in North Carolina.
Patel, Mehul D; Brice, Jane H; Evenson, Kelly R; Rose, Kathryn M; Suchindran, Chirayath M; Rosamond, Wayne D
2013-09-05
Prior assessments of emergency medical services (EMS) stroke capacity found deficiencies in education and training, use of protocols and screening tools, and planning for the transport of patients. A 2001 survey of North Carolina EMS providers found many EMS systems lacked basic stroke services. Recent statewide efforts have sought to standardize and improve prehospital stroke care. The objective of this study was to assess EMS stroke care capacity in North Carolina and evaluate statewide changes since 2001. In June 2012, we conducted a web-based survey on stroke education and training and stroke care practices and policies among all EMS systems in North Carolina. We used the McNemar test to assess changes from 2001 to 2012. Of 100 EMS systems in North Carolina, 98 responded to our survey. Most systems reported providing stroke education and training (95%) to EMS personnel, using a validated stroke scale or screening tool (96%), and having a hospital prenotification policy (98%). Many were suboptimal in covering basic stroke educational topics (71%), always communicating stroke screen results to the destination hospital (46%), and always using a written destination plan (49%). Among 70 EMS systems for which we had data for 2001 and 2012, we observed significant improvements in education on stroke scales or screening tools (61% to 93%, P < .001) and use of validated stroke scales or screening tools (23% to 96%, P < .001). Major improvements in EMS stroke care, especially in prehospital stroke screening, have occurred in North Carolina in the past decade, whereas other practices and policies, including use of destination plans, remain in need of improvement.
The rationale, development, and standardization of a basic word vocabulary test.
Dupuy, H J
1974-04-01
The results of the studies to date indicate that the Basic Word Vocabulary Test provides a range of items in terms of item difficulty levels useful in printed form from about the third grade to the highest educational levels. Since pictorial and orally given vocabulary tests are used from about ages 2 to 8 years, further work should be done to extend the scale downward so that a single comprehensive vocabulary scale ranging from age 2 years to the highest level of verbal development is available for general use. Validation studies should also be conducted with other well-known intelligence tests so that scores can be compared. Alternate forms need to be developed to allow for longitudinal studies of growth and development. The use of a single standard of measurement of vocabulary development, suitable over a wide range of age and ability levels, by different investigators should materially aid in comparing results across studies and samples and lead to more consistent findings, advances in knowledge, and wider application of findings in practical circumstances, The findings presented in this report indicate that the Basic Word Vocabulary Test adequately measures basic word knowledge acquisition and development. The BWVT is suitable for evaluation of individuals and for use in making group comparisons in levels of basic word knowledge attainment, growth, and development. All material appearing in this report is in the public domain and may be reproduced or copied without permission; citation as to source, however, is appreciated.
User-Friendly Interface Developed for a Web-Based Service for SpaceCAL Emulations
NASA Technical Reports Server (NTRS)
Liszka, Kathy J.; Holtz, Allen P.
2004-01-01
A team at the NASA Glenn Research Center is developing a Space Communications Architecture Laboratory (SpaceCAL) for protocol development activities for coordinated satellite missions. SpaceCAL will provide a multiuser, distributed system to emulate space-based Internet architectures, backbone networks, formation clusters, and constellations. As part of a new effort in 2003, building blocks are being defined for an open distributed system to make the satellite emulation test bed accessible through an Internet connection. The first step in creating a Web-based service to control the emulation remotely is providing a user-friendly interface for encoding the data into a well-formed and complete Extensible Markup Language (XML) document. XML provides coding that allows data to be transferred between dissimilar systems. Scenario specifications include control parameters, network routes, interface bandwidths, delay, and bit error rate. Specifications for all satellite, instruments, and ground stations in a given scenario are also included in the XML document. For the SpaceCAL emulation, the XML document can be created using XForms, a Webbased forms language for data collection. Contrary to older forms technology, the interactive user interface makes the science prevalent, not the data representation. Required versus optional input fields, default values, automatic calculations, data validation, and reuse will help researchers quickly and accurately define missions. XForms can apply any XML schema defined for the test mission to validate data before forwarding it to the emulation facility. New instrument definitions, facilities, and mission types can be added to the existing schema. The first prototype user interface incorporates components for interactive input and form processing. Internet address, data rate, and the location of the facility are implemented with basic form controls with default values provided for convenience and efficiency using basic XForms operations. Because different emulation scenarios will vary widely in their component structure, more complex operations are used to add and delete facilities.
Validation of Satellite Aerosol Retrievals from AERONET Ground-Based Measurements
NASA Technical Reports Server (NTRS)
Holben, Brent; Remer, Lorraine; Torres, Omar; Zhao, Tom; Smith, David E. (Technical Monitor)
2001-01-01
Accurate and comprehensive assessment of the parameters that control key atmospheric and biospheric processes including assessment of anthropogenic effects on climate change is a fundamental measurement objective of NASA's EOS program (King and Greenstone, 1999). Satellite assessment programs and associated global climate models require validation and additional parameterization with frequent reliable ground-based observations. A critical and highly uncertain element of the measurement program is characterization of tropospheric aerosols requiring basic observations of aerosols optical and microphysical properties. Unfortunately as yet we do not know the aerosol burden man is contributing to the atmosphere and thus we will have no definitive measure of change for the future. This lack of aerosol assessment is the impetus for some of the EOS measurement activities (Kaufman et al., 1997; King et al., 1999) and the formation of the AERONET program (Holben et al., 1998). The goals of the AERONET program are to develop long term monitoring at globally distributed sites providing critical data for multiannual trend changes in aerosol loading and optical properties with the specific goal of providing a data base for validation of satellite derived aerosol optical properties. The AERONET program has evolved into an international federated network of approximately 100 ground-based remote sensing monitoring stations to characterize the optical and microphysical properties of aerosols.
NASA Astrophysics Data System (ADS)
Vosk, Ted
2011-10-01
The principles, methods and technologies of physics can provide a powerful tool for the discovery of truth in the criminal justice system. Accordingly, physics based forensic evidence is relied upon in criminal prosecutions around the country every day. Infrared spectroscopy for the determination of the alcohol concentration of an individual's breath, force, momentum and multi-body dynamics for purposes of accident reconstruction and the basic application of sound metrological (measurement) practices constitute but a few examples. In many cases, a jury's determination of guilt or innocence, upon which the liberty of a Citizen rests, may in fact be determined by such evidence. Society may well place a high degree of confidence in the integrity of verdicts so obtained when ``the physics'' has been applied in a valid manner. Unfortunately, as concluded by the National Academy of Sciences, ``The law's greatest dilemma in its heavy reliance on forensic evidence--concerns the question of whether---and to what extent-- -there is science in any given `forensic science' discipline.'' Even where valid physical principles are relied upon, their improper application by forensic practitioners who have little physics training, background and/or understanding calls into question the validity of results or conclusions obtained. This presentation provides examples of the application of physics in the courtroom, where problems have been discovered and how they can be addressed by the physics community.
Demurie, Ellen; Roeyers, Herbert; Baeyens, Dieter; Sonuga-Barke, Edmund
2012-12-01
The current study compared the effects of reward anticipation on task performance in children and adolescents (8-16 years old) using monetary and various social rewards. Eighty-five typically developing children undertook the Monetary Incentive Delay (MID) task. Of these 44 also undertook the Social Incentive Delay (SID-basic) task where social reward was operationalized as a smiling face and spoken compliments. Forty-one children participated in the SID-plus where points were added to a pictogram with written compliments. In a preparatory validation study participants were asked howmuch they liked the SID-basic rewards.Results showed that there was an effect of reward size on accuracy and RT in both the MID task and SID-plus, but not SID-basic. Subjective value of the SID-basic rewards was rated higher with hypothesized increasing reward intensity. In conclusion, although the social rewards in SID-basic were liked by children andadolescents in the validation study, they had no effect on the behaviour. Only when points were added (SID-plus), anticipated social reward affected task performance. Thus our results highlight (i) the difference between likeability andreinforcing quality and (ii) the need for a quantifiable element to rewards for themto be reinforcing for children. Copyright © 2012 John Wiley & Sons, Ltd.
ASTM Validates Air Pollution Test Methods
ERIC Educational Resources Information Center
Chemical and Engineering News, 1973
1973-01-01
The American Society for Testing and Materials (ASTM) has validated six basic methods for measuring pollutants in ambient air as the first part of its Project Threshold. Aim of the project is to establish nationwide consistency in measuring pollutants; determining precision, accuracy and reproducibility of 35 standard measuring methods. (BL)
Student Ratings: The Validity of Use.
ERIC Educational Resources Information Center
McKeachie, Wilbert J.
1997-01-01
Concludes that there is concurrence on the validity of student ratings but that contextual variables affect the level of ratings. However, there is disagreement on the use of statistical corrections for such bias. The basic problem lies in the lack of sophistication of personnel committees who use the ratings. (MMU)
The Validity of the Interpersonal Behaviors Questionnaire (IBQ) in Sport
ERIC Educational Resources Information Center
Rocchi, Meredith; Pelletier, Luc; Desmarais, Philippe
2017-01-01
According to Self-Determination Theory (SDT), basic psychological needs will be influenced by other individuals' interpersonal behaviors. The objective of the present research is to extend the validity of the Interpersonal Behaviors Questionnaire (IBQ and IBQ-Self) to the sport context. The measure was designed to assess perceptions of…
A Validation Study of the Existential Anxiety Scale.
ERIC Educational Resources Information Center
Hullett, Michael A.
Logotherapy is a meaning-centered psychotherapy which focuses on both the meaning of human existence and the personal search for meaning. If the will to search for meaning is frustrated, "existential frustration" may result. This study validates the Existential Anxiety Scale (EAS) developed by Good and Good (1974). Basic principles of…
Reaction Time and Self-Report Psychopathological Assessment: Convergent and Discriminant Validity.
ERIC Educational Resources Information Center
Holden, Ronald R.; Fekken, G. Cynthia
The processing of incoming psychological information along the network, or schemata, of self-knowledge was studied to determine the convergent and discriminant validity of the patterns of schemata-specific response latencies. Fifty-three female and 52 male university students completed the Basic Personality Inventory (BPI). BPI scales assess…
Basic Navigator Battery: An Experimental Selection Composite for Undergraduate Navigator Training.
ERIC Educational Resources Information Center
Shanahan, Frank M.; Kantor, Jeffrey E.
High rates of attrition among students in Undergraduate Navigator Training (UNT) is a major concern for Air Training Command. The main objective of this research was to evaluate the Basic Navigator Battery (BNB), a multi-test experimental selection instrument, for its potential to increase the validity of the Air Force Officer Qualifying Test…
New regulations have emerged in recent years giving guidance on how best to conduct skin sensitization testing on new chemical entities. For example, recently The Interagency Coordinating Committee on the Validation of Alternative Methods (ICCVAM) in the U.S. and the European Ce...
A Multicomponent Measure of Writing Motivation with Basic College Writers
ERIC Educational Resources Information Center
MacArthur, Charles A.; Philippakos, Zoi A.; Graham, Steve
2016-01-01
The purpose of the current study was to develop and validate a measure of motivation for use with basic college writers that would measure self-efficacy, achievement goals, beliefs, and affect. As part of a design research project on curriculum for community college developmental writing classes, 133 students in 11 classes completed the motivation…
Validation of Student and Parent Reported Data on the Basic Grant Application Form. Project Summary.
ERIC Educational Resources Information Center
Applied Management Sciences, Inc., Silver Spring, MD.
Results of studies to assess accuracy of information reported by applicants to the Basic Educational Opportunity Grant (BEOG) program are summarized. Attention is also focused on applicant characteristics and corrective actions taken as a result of the studies. Overall, the studies found that the majority of BEOG applicants reported income…
Learning and Motivation in Thailand: A Comparative Regional Study on Basic Education Ninth Graders
ERIC Educational Resources Information Center
Loima, Jyrki; Vibulphol, Jutarat
2016-01-01
This qualitative research studied regional motivation and learning of the basic education 9th graders in Thailand. Second topic was the school size and its possible effect on motivation. Furthermore, the data gave an opportunity to discuss, whether international research on motivation and learning was valid in Thai classrooms. The informants were…
Sokoliess, Torsten; Köller, Gerhard
2005-06-01
A chiral capillary electrophoresis system allowing the determination of the enantiomeric purity of an investigational new drug was developed using a generic method development approach for basic analytes. The method was optimized in terms of type and concentration of both cyclodextrin (CD) and electrolyte, buffer pH, temperature, voltage, and rinsing procedure. Optimal chiral separation of the analyte was obtained using an electrolyte with 2.5% carboxymethyl-beta-CD in 25 mM NaH2PO4 (pH 4.0). Interchanging the inlet and outlet vials after each run improved the method's precision. To assure the method's suitability for the control of enantiomeric impurities in pharmaceutical quality control, its specificity, linearity, precision, accuracy, and robustness were validated according to the requirements of the International Conference on Harmonization. The usefulness of our generic method development approach for the validation of robustness was demonstrated.
Wu, Huey-Min; Li, Cheng-Hsaun; Kuo, Bor-Chen; Yang, Yu-Mao; Lin, Chin-Kai; Wan, Wei-Hsiang
2017-08-01
Morphological awareness is the foundation for the important developmental skills involved with vocabulary, as well as understanding the meaning of words, orthographic knowledge, reading, and writing. Visual perception of space and radicals in two-dimensional positions of Chinese characters' morphology is very important in identifying Chinese characters. The important predictive variables of special and visual perception in Chinese characters identification were investigated in the growth model in this research. The assessment tool is the "Computerized Visual Perception Assessment Tool for Chinese Characters Structures" developed by this study. There are two constructs, basic stroke and character structure. In the basic stroke, there are three subtests of one, two, and more than three strokes. In the character structure, there are three subtests of single-component character, horizontal-compound character, and vertical-compound character. This study used purposive sampling. In the first year, 551 children 4-6 years old participated in the study and were monitored for one year. In the second year, 388 children remained in the study and the successful follow-up rate was 70.4%. This study used a two-wave cross-lagged panel design to validate the growth model of the basic stroke and the character structure. There was significant correlation of the basic stroke and the character structure at different time points. The abilities in the basic stroke and in the character structure steadily developed over time for preschool children. Children's knowledge of the basic stroke effectively predicted their knowledge of the basic stroke and the character structure. Copyright © 2017 Elsevier Ltd. All rights reserved.
Collins, Anne; Ross, Janine
2017-01-01
We performed a systematic review to identify all original publications describing the asymmetric inheritance of cellular organelles in normal animal eukaryotic cells and to critique the validity and imprecision of the evidence. Searches were performed in Embase, MEDLINE and Pubmed up to November 2015. Screening of titles, abstracts and full papers was performed by two independent reviewers. Data extraction and validity were performed by one reviewer and checked by a second reviewer. Study quality was assessed using the SYRCLE risk of bias tool, for animal studies and by developing validity tools for the experimental model, organelle markers and imprecision. A narrative data synthesis was performed. We identified 31 studies (34 publications) of the asymmetric inheritance of organelles after mitotic or meiotic division. Studies for the asymmetric inheritance of centrosomes (n = 9); endosomes (n = 6), P granules (n = 4), the midbody (n = 3), mitochondria (n = 3), proteosomes (n = 2), spectrosomes (n = 2), cilia (n = 2) and endoplasmic reticulum (n = 2) were identified. Asymmetry was defined and quantified by variable methods. Assessment of the statistical reliability of the results indicated only two studies (7%) were judged to have low concern, the majority of studies (77%) were 'unclear' and five (16%) were judged to have 'high concerns'; the main reasons were low technical repeats (<10). Assessment of model validity indicated that the majority of studies (61%) were judged to be valid, ten studies (32%) were unclear and two studies (7%) were judged to have 'high concerns'; both described 'stem cells' without providing experimental evidence to confirm this (pluripotency and self-renewal). Assessment of marker validity indicated that no studies had low concern, most studies were unclear (96.5%), indicating there were insufficient details to judge if the markers were appropriate. One study had high concern for marker validity due to the contradictory results of two markers for the same organelle. For most studies the validity and imprecision of results could not be confirmed. In particular, data were limited due to a lack of reporting of interassay variability, sample size calculations, controls and functional validation of organelle markers. An evaluation of 16 systematic reviews containing cell assays found that only 50% reported adherence to PRISMA or ARRIVE reporting guidelines and 38% reported a formal risk of bias assessment. 44% of the reviews did not consider how relevant or valid the models were to the research question. 75% reviews did not consider how valid the markers were. 69% of reviews did not consider the impact of the statistical reliability of the results. Future systematic reviews in basic or preclinical research should ensure the rigorous reporting of the statistical reliability of the results in addition to the validity of the methods. Increased awareness of the importance of reporting guidelines and validation tools is needed for the scientific community. PMID:28562636
Relevance of motion-related assessment metrics in laparoscopic surgery.
Oropesa, Ignacio; Chmarra, Magdalena K; Sánchez-González, Patricia; Lamata, Pablo; Rodrigues, Sharon P; Enciso, Silvia; Sánchez-Margallo, Francisco M; Jansen, Frank-Willem; Dankelman, Jenny; Gómez, Enrique J
2013-06-01
Motion metrics have become an important source of information when addressing the assessment of surgical expertise. However, their direct relationship with the different surgical skills has not been fully explored. The purpose of this study is to investigate the relevance of motion-related metrics in the evaluation processes of basic psychomotor laparoscopic skills and their correlation with the different abilities sought to measure. A framework for task definition and metric analysis is proposed. An explorative survey was first conducted with a board of experts to identify metrics to assess basic psychomotor skills. Based on the output of that survey, 3 novel tasks for surgical assessment were designed. Face and construct validation was performed, with focus on motion-related metrics. Tasks were performed by 42 participants (16 novices, 22 residents, and 4 experts). Movements of the laparoscopic instruments were registered with the TrEndo tracking system and analyzed. Time, path length, and depth showed construct validity for all 3 tasks. Motion smoothness and idle time also showed validity for tasks involving bimanual coordination and tasks requiring a more tactical approach, respectively. Additionally, motion smoothness and average speed showed a high internal consistency, proving them to be the most task-independent of all the metrics analyzed. Motion metrics are complementary and valid for assessing basic psychomotor skills, and their relevance depends on the skill being evaluated. A larger clinical implementation, combined with quality performance information, will give more insight on the relevance of the results shown in this study.
Space human factors publications: 1980-1990
NASA Technical Reports Server (NTRS)
Dickson, Katherine J.
1991-01-01
A 10 year cummulative bibliography of publications resulting from research supported by the NASA Space Human Factors Program of the Life Science Division is provided. The goal of this program is to understand the basic mechanisms underlying behavioral adaptation to space and to develop and validate system design requirements, protocols, and countermeasures to ensure the psychological well-being, safety, and productivity of crewmembers. Subjects encompassed by this bibliography include selection and training, group dynamics, psychophysiological interactions, habitability issues, human-machine interactions, psychological support measures, and anthropometric data. Principal Investigators whose research tasks resulted in publication are identified by asterisk.
Tropospheric ozone as a fungal elicitor.
Zuccarini, Paolo
2009-03-01
Tropospheric ozone has been proven to trigger biochemical plant responses that are similar to the ones induced by an attack of fungal pathogens,i.e. it resembles fungal elicitors.This suggests that ozone can represent a valid tool for the study of stress responses and induction of resistance to pathogens. This review provides an overview of the implications of such a phenomenon for basic and applied research. After an introduction about the environmental implications of tropospheric ozone and plant responses to biotic stresses, the biochemistry of ozone stress is analysed, pointing out its similarities with plant responses to pathogens and its possible applications.
On the transferability of RegCM4: Europe, Africa and Asia
NASA Astrophysics Data System (ADS)
Belda, Michal; Halenka, Tomas
2013-04-01
Simulations driven by ERA-interim reanalysis for CORDEX domains covering Europe, Africa and Asia have been performed using RegCM4 at 50 km resolution. The same settings are used in basic simulations and preliminary evaluation of model performance for individual regions will be presented. Several settings of different options is tested and sensitivity of selected ones will be shown in individual regions. Secant Mercator projection is introduced for Africa providing more efficient model geometry setting, the impact of proper emissivity inclusion is compared especially for Africa and Asia desserts. CRU data are used for the validation.
BASiCS: Bayesian Analysis of Single-Cell Sequencing Data
Vallejos, Catalina A.; Marioni, John C.; Richardson, Sylvia
2015-01-01
Single-cell mRNA sequencing can uncover novel cell-to-cell heterogeneity in gene expression levels in seemingly homogeneous populations of cells. However, these experiments are prone to high levels of unexplained technical noise, creating new challenges for identifying genes that show genuine heterogeneous expression within the population of cells under study. BASiCS (Bayesian Analysis of Single-Cell Sequencing data) is an integrated Bayesian hierarchical model where: (i) cell-specific normalisation constants are estimated as part of the model parameters, (ii) technical variability is quantified based on spike-in genes that are artificially introduced to each analysed cell’s lysate and (iii) the total variability of the expression counts is decomposed into technical and biological components. BASiCS also provides an intuitive detection criterion for highly (or lowly) variable genes within the population of cells under study. This is formalised by means of tail posterior probabilities associated to high (or low) biological cell-to-cell variance contributions, quantities that can be easily interpreted by users. We demonstrate our method using gene expression measurements from mouse Embryonic Stem Cells. Cross-validation and meaningful enrichment of gene ontology categories within genes classified as highly (or lowly) variable supports the efficacy of our approach. PMID:26107944
BASiCS: Bayesian Analysis of Single-Cell Sequencing Data.
Vallejos, Catalina A; Marioni, John C; Richardson, Sylvia
2015-06-01
Single-cell mRNA sequencing can uncover novel cell-to-cell heterogeneity in gene expression levels in seemingly homogeneous populations of cells. However, these experiments are prone to high levels of unexplained technical noise, creating new challenges for identifying genes that show genuine heterogeneous expression within the population of cells under study. BASiCS (Bayesian Analysis of Single-Cell Sequencing data) is an integrated Bayesian hierarchical model where: (i) cell-specific normalisation constants are estimated as part of the model parameters, (ii) technical variability is quantified based on spike-in genes that are artificially introduced to each analysed cell's lysate and (iii) the total variability of the expression counts is decomposed into technical and biological components. BASiCS also provides an intuitive detection criterion for highly (or lowly) variable genes within the population of cells under study. This is formalised by means of tail posterior probabilities associated to high (or low) biological cell-to-cell variance contributions, quantities that can be easily interpreted by users. We demonstrate our method using gene expression measurements from mouse Embryonic Stem Cells. Cross-validation and meaningful enrichment of gene ontology categories within genes classified as highly (or lowly) variable supports the efficacy of our approach.
Application of omics data in regulatory toxicology: report of an international BfR expert workshop.
Marx-Stoelting, P; Braeuning, A; Buhrke, T; Lampen, A; Niemann, L; Oelgeschlaeger, M; Rieke, S; Schmidt, F; Heise, T; Pfeil, R; Solecki, R
2015-11-01
Advances in omics techniques and molecular toxicology are necessary to provide new perspectives for regulatory toxicology. By the application of modern molecular techniques, more mechanistic information should be gained to support standard toxicity studies and to contribute to a reduction and refinement of animal experiments required for certain regulatory purposes. The relevance and applicability of data obtained by omics methods to regulatory purposes such as grouping of chemicals, mode of action analysis or classification and labelling needs further improvement, defined validation and cautious expert judgment. Based on the results of an international expert workshop organized 2014 by the Federal Institute for Risk Assessment in Berlin, this paper is aimed to provide a critical overview of the regulatory relevance and reliability of omics methods, basic requirements on data quality and validation, as well as regulatory criteria to decide which effects observed by omics methods should be considered adverse or non-adverse. As a way forward, it was concluded that the inclusion of omics data can facilitate a more flexible approach for regulatory risk assessment and may help to reduce or refine animal testing.
The human role in space: Technology, economics and optimization
NASA Technical Reports Server (NTRS)
Hall, S. B. (Editor)
1985-01-01
Man-machine interactions in space are explored in detail. The role and the degree of direct involvement of humans that will be required in future space missions are investigated. An attempt is made to establish valid criteria for allocating functional activities between humans and machines and to provide insight into the technological requirements, economics, and benefits of the human presence in space. Six basic categories of man-machine interactions are considered: manual, supported, augmented, teleoperated, supervised, and independent. Appendices are included which provide human capability data, project analyses, activity timeline profiles and data sheets for 37 generic activities, support equipment and human capabilities required in these activities, and cumulative costs as a function of activity for seven man-machine modes.
Using Literacy Techniques to Teach Astronomy to Non-Science Majors
NASA Astrophysics Data System (ADS)
Garland, C. A.; Ratay, D. L.
We discuss an introductory-level college astronomy class that significantly relied on reading and writing assignments to deliver basic content knowledge and provide a basis for deeper analysis of the material. As opposed to the traditional problem-set method of homework, students were required to read popular articles from magazines and newspapers related to the content presented in class, and then prepare responses. These responses ranged from methodological analyzes to using the readings to create original science journalism. Additional forms of assessment indicated that students benefited from this type of course design. We propose that given the background of students in this type of course, our course design is better suited to engage students in the material and provides a valid alternative method of assessment.
DNA vaccination for prostate cancer, from preclinical to clinical trials - where we stand?
2012-01-01
Development of various vaccines for prostate cancer (PCa) is becoming an active research area. PCa vaccines are perceived to have less toxicity compared with the available cytotoxic agents. While various immune-based strategies can elicit anti-tumour responses, DNA vaccines present increased efficacy, inducing both humoural and cellular immunity. This immune activation has been proven effective in animal models and initial clinical trials are encouraging. However, to validate the role of DNA vaccination in currently available PCa management paradigms, strong clinical evidence is still lacking. This article provides an overview of the basic principles of DNA vaccines and aims to provide a summary of preclinical and clinical trials outlining the benefits of this immunotherapy in the management of PCa. PMID:23046944
NASA Astrophysics Data System (ADS)
Drob, D. P.; Huba, J.; Kordella, L.; Earle, G. D.; Ridley, A. J.
2017-12-01
The great American solar eclipse of August 21, 2017 provides a unique opportunity to study the basic physics of the upper atmosphere and ionosphere. While the effects of solar eclipses on the upper atmosphere and ionosphere have been studied since the 1930s, and later matured in the last several decades, recent advances in first principles numerical models and multi-instrument observational capabilities continue to provide new insights. Upper atmospheric eclipse phenomena such as ionospheric conjugate effects and the generation of a thermospheric bow wave that propagates into the nightside are simulated with high-resolution first principles upper atmospheric models and compared with observations to validate this understanding.
Shedden, Kerby; Taylor, Jeremy M.G.; Enkemann, Steve A.; Tsao, Ming S.; Yeatman, Timothy J.; Gerald, William L.; Eschrich, Steve; Jurisica, Igor; Venkatraman, Seshan E.; Meyerson, Matthew; Kuick, Rork; Dobbin, Kevin K.; Lively, Tracy; Jacobson, James W.; Beer, David G.; Giordano, Thomas J.; Misek, David E.; Chang, Andrew C.; Zhu, Chang Qi; Strumpf, Dan; Hanash, Samir; Shepherd, Francis A.; Ding, Kuyue; Seymour, Lesley; Naoki, Katsuhiko; Pennell, Nathan; Weir, Barbara; Verhaak, Roel; Ladd-Acosta, Christine; Golub, Todd; Gruidl, Mike; Szoke, Janos; Zakowski, Maureen; Rusch, Valerie; Kris, Mark; Viale, Agnes; Motoi, Noriko; Travis, William; Sharma, Anupama
2009-01-01
Although prognostic gene expression signatures for survival in early stage lung cancer have been proposed, for clinical application it is critical to establish their performance across different subject populations and in different laboratories. Here we report a large, training-testing, multi-site blinded validation study to characterize the performance of several prognostic models based on gene expression for 442 lung adenocarcinomas. The hypotheses proposed examined whether microarray measurements of gene expression either alone or combined with basic clinical covariates (stage, age, sex) can be used to predict overall survival in lung cancer subjects. Several models examined produced risk scores that substantially correlated with actual subject outcome. Most methods performed better with clinical data, supporting the combined use of clinical and molecular information when building prognostic models for early stage lung cancer. This study also provides the largest available set of microarray data with extensive pathological and clinical annotation for lung adenocarcinomas. PMID:18641660
SINERGIA laparoscopic virtual reality simulator: didactic design and technical development.
Lamata, Pablo; Gómez, Enrique J; Sánchez-Margallo, Francisco M; López, Oscar; Monserrat, Carlos; García, Verónica; Alberola, Carlos; Florido, Miguel Angel Rodríguez; Ruiz, Juan; Usón, Jesús
2007-03-01
VR laparoscopic simulators have demonstrated its validity in recent studies, and research should be directed towards a high training effectiveness and efficacy. In this direction, an insight into simulators' didactic design and technical development is provided, by describing the methodology followed in the building of the SINERGIA simulator. It departs from a clear analysis of training needs driven by a surgical training curriculum. Existing solutions and validation studies are an important reference for the definition of specifications, which are described with a suitable use of simulation technologies. Five new didactic exercises are proposed to train some of the basic laparoscopic skills. Simulator construction has required existing algorithms and the development of a particle-based biomechanical model, called PARSYS, and a collision handling solution based in a multi-point strategy. The resulting VR laparoscopic simulator includes new exercises and enhanced simulation technologies, and is finding a very good acceptance among surgeons.
Xie, Yuan; Bergström, Tobias; Jiang, Yiwen; Johansson, Patrik; Marinescu, Voichita Dana; Lindberg, Nanna; Segerman, Anna; Wicher, Grzegorz; Niklasson, Mia; Baskaran, Sathishkumar; Sreedharan, Smitha; Everlien, Isabelle; Kastemar, Marianne; Hermansson, Annika; Elfineh, Lioudmila; Libard, Sylwia; Holland, Eric Charles; Hesselager, Göran; Alafuzoff, Irina; Westermark, Bengt; Nelander, Sven; Forsberg-Nilsson, Karin; Uhrbom, Lene
2015-10-01
Glioblastoma (GBM) is the most frequent and malignant form of primary brain tumor. GBM is essentially incurable and its resistance to therapy is attributed to a subpopulation of cells called glioma stem cells (GSCs). To meet the present shortage of relevant GBM cell (GC) lines we developed a library of annotated and validated cell lines derived from surgical samples of GBM patients, maintained under conditions to preserve GSC characteristics. This collection, which we call the Human Glioblastoma Cell Culture (HGCC) resource, consists of a biobank of 48 GC lines and an associated database containing high-resolution molecular data. We demonstrate that the HGCC lines are tumorigenic, harbor genomic lesions characteristic of GBMs, and represent all four transcriptional subtypes. The HGCC panel provides an open resource for in vitro and in vivo modeling of a large part of GBM diversity useful to both basic and translational GBM research.
NASA Astrophysics Data System (ADS)
Fu, Yao; Song, Jeong-Hoon
2014-08-01
Hardy stress definition has been restricted to pair potentials and embedded-atom method potentials due to the basic assumptions in the derivation of a symmetric microscopic stress tensor. Force decomposition required in the Hardy stress expression becomes obscure for multi-body potentials. In this work, we demonstrate the invariance of the Hardy stress expression for a polymer system modeled with multi-body interatomic potentials including up to four atoms interaction, by applying central force decomposition of the atomic force. The balance of momentum has been demonstrated to be valid theoretically and tested under various numerical simulation conditions. The validity of momentum conservation justifies the extension of Hardy stress expression to multi-body potential systems. Computed Hardy stress has been observed to converge to the virial stress of the system with increasing spatial averaging volume. This work provides a feasible and reliable linkage between the atomistic and continuum scales for multi-body potential systems.
King, Marika R.; Binger, Cathy; Kent-Walsh, Jennifer
2015-01-01
The developmental readiness of four 5-year-old children to produce basic sentences using graphic symbols on an augmentative and alternative communication (AAC) device during a dynamic assessment (DA) task was examined. Additionally, the ability of the DA task to predict performance on a subsequent experimental task was evaluated. A graduated prompting framework was used during DA. Measures included amount of support required to produce the targets, modifiability (change in participant performance) within a DA session, and predictive validity of DA. Participants accurately produced target structures with varying amounts of support. Modifiability within DA sessions was evident for some participants, and partial support was provided for the measures of predictive validity. These initial results indicate that DA may be a viable way to measure young children’s developmental readiness to learn how to sequence simple, rule-based messages via aided AAC. PMID:25621928
An Examination of Construct Validity for the EARLI Numeracy Skill Measures
ERIC Educational Resources Information Center
Cheng, Weiyi; Lei, Pui-Wa; DiPerna, James C.
2017-01-01
The purpose of the current study was to examine dimensionality and concurrent validity evidence of the EARLI numeracy measures (DiPerna, Morgan, & Lei, 2007), which were developed to assess key skills such as number identification, counting, and basic arithmetic. Two methods (NOHARM with approximate chi-square test and DIMTEST with DETECT…
ERIC Educational Resources Information Center
Han, Jeanie Nam; Vanderwood, Michael L.; Lee, Catherine Y.
2015-01-01
This study examined the predictive validity of early literacy measures with first-grade Korean English language learners (ELLs) in the United States at varying levels of English proficiency. Participants were screened using Dynamic Indicators of Basic Early Literacy Skills (DIBELS) Phoneme Segmentation Fluency (PSF), DIBELS Nonsense Word Fluency…
Oja, M; Maran, U
2015-01-01
Absorption in gastrointestinal tract compartments varies and is largely influenced by pH. Therefore, considering pH in studies and analyses of membrane permeability provides an opportunity to gain a better understanding of the behaviour of compounds and to obtain good permeability estimates for prediction purposes. This study concentrates on relationships between the chemical structure and membrane permeability of acidic and basic drugs and drug-like compounds. The membrane permeability of 36 acidic and 61 basic compounds was measured using the parallel artificial membrane permeability assay (PAMPA) at pH 3, 5, 7.4 and 9. Descriptive and/or predictive single-parameter quantitative structure-permeability relationships were derived for all pH values. For acidic compounds, membrane permeability is mainly influenced by hydrogen bond donor properties, as revealed by models with r(2) > 0.8 for pH 3 and pH 5. For basic compounds, the best (r(2) > 0.7) structure-permeability relationships are obtained with the octanol-water distribution coefficient for pH 7.4 and pH 9, indicating the importance of partition properties. In addition to the validation set, the prediction quality of the developed models was tested with folic acid and astemizole, showing good matches between experimental and calculated membrane permeabilities at key pHs. Selected QSAR models are available at the QsarDB repository ( http://dx.doi.org/10.15152/QDB.166 ).
Benej, Martin; Bendlova, Bela; Vaclavikova, Eliska; Poturnajova, Martina
2011-10-06
Reliable and effective primary screening of mutation carriers is the key condition for common diagnostic use. The objective of this study is to validate the method high resolution melting (HRM) analysis for routine primary mutation screening and accomplish its optimization, evaluation and validation. Due to their heterozygous nature, germline point mutations of c-RET proto-oncogene, associated to multiple endocrine neoplasia type 2 (MEN2), are suitable for HRM analysis. Early identification of mutation carriers has a major impact on patients' survival due to early onset of medullary thyroid carcinoma (MTC) and resistance to conventional therapy. The authors performed a series of validation assays according to International Conference on Harmonization of Technical Requirements for Registration of Pharmaceuticals for Human Use (ICH) guidelines for validation of analytical procedures, along with appropriate design and optimization experiments. After validated evaluation of HRM, the method was utilized for primary screening of 28 pathogenic c-RET mutations distributed among nine exons of c-RET gene. Validation experiments confirm the repeatability, robustness, accuracy and reproducibility of HRM. All c-RET gene pathogenic variants were detected with no occurrence of false-positive/false-negative results. The data provide basic information about design, establishment and validation of HRM for primary screening of genetic variants in order to distinguish heterozygous point mutation carriers among the wild-type sequence carriers. HRM analysis is a powerful and reliable tool for rapid and cost-effective primary screening, e.g., of c-RET gene germline and/or sporadic mutations and can be used as a first line potential diagnostic tool.
NASA Astrophysics Data System (ADS)
de Greef, Maurice; Segers, Mien; Nijhuis, Jan; Lam, Jo Fond; van Groenestijn, Mieke; van Hoek, Frans; van Deursen, Alexander J. A. M.; Bohnenn, Ella; Tubbing, Marga
2015-10-01
Besides work-oriented training, most Dutch adult learning courses of formal and non-formal education focus on three basic skills: literacy, numeracy and problem solving in technology-rich environments. In the Netherlands, the Ministry of Education, Culture and Science recently initiated the development of a new adult education framework concerning literacy, numeracy and digital skills. In order to monitor the progress of literacy, numeracy and digital competencies, it is necessary to develop and validate testing materials for specific competencies. This study validates the testing materials which were developed to assess learners' proficiency in literacy (reading and writing), numeracy and digital skills based on the new Dutch framework. The outcome is that the materials proved valid and can be used in different courses referring to basic skills and adult learning, though there are still some limitations. Besides adult education professionals (such teachers and trainers), policy makers can also use the results of these tests in order to describe and monitor the impact of adult education on the lives of adult learners.
PRA (Probabilistic Risk Assessments) Participation versus Validation
NASA Technical Reports Server (NTRS)
DeMott, Diana; Banke, Richard
2013-01-01
Probabilistic Risk Assessments (PRAs) are performed for projects or programs where the consequences of failure are highly undesirable. PRAs primarily address the level of risk those projects or programs posed during operations. PRAs are often developed after the design has been completed. Design and operational details used to develop models include approved and accepted design information regarding equipment, components, systems and failure data. This methodology basically validates the risk parameters of the project or system design. For high risk or high dollar projects, using PRA methodologies during the design process provides new opportunities to influence the design early in the project life cycle to identify, eliminate or mitigate potential risks. Identifying risk drivers before the design has been set allows the design engineers to understand the inherent risk of their current design and consider potential risk mitigation changes. This can become an iterative process where the PRA model can be used to determine if the mitigation technique is effective in reducing risk. This can result in more efficient and cost effective design changes. PRA methodology can be used to assess the risk of design alternatives and can demonstrate how major design changes or program modifications impact the overall program or project risk. PRA has been used for the last two decades to validate risk predictions and acceptability. Providing risk information which can positively influence final system and equipment design the PRA tool can also participate in design development, providing a safe and cost effective product.
ERIC Educational Resources Information Center
Munger, Kristen A.; LoFaro, Stephen A.; Kawryga, Erin A.; Sovocool, Elizabeth A.; Medina, Siani Y.
2014-01-01
This study involved examination of the validity evidence of the Dynamic Indicators of Basic Early Literacy Skills-Next Edition (DIBELS Next) for a sample of 85 third-and fifth-grade students, in reference to the "simple view" of reading. Tests administered included DIBELS Next, Peabody Picture Vocabulary Test-IV (PPVT-IV), Group Reading…
ERIC Educational Resources Information Center
Petry, Katja; Maes, Bea; Vlaskamp, Carla
2005-01-01
Background: This study considered the general validity of the basic domains of quality of life that appear in theoretical models, in relation to people with profound multiple disabilities. The authors examined how parents and direct support staff operationalized these basic domains for people with profound multiple disabilities. They investigated…
ERIC Educational Resources Information Center
Jenkins-Guarnieri, Michael A.; Vaughan, Angela L.; Wright, Stephen L.
2015-01-01
We adapted a work self-determination measure to create the Basic Needs Satisfaction at College Scale. Confirmatory factor analysis and item response theory analyses with data from 525 adults supported a 3-factor model with 13 items most sensitive for lower to middle range levels of the autonomy, competence, and relatedness constructs.
ERIC Educational Resources Information Center
Perkins, Kyle; Duncan, Ann
An assessment analysis was performed to determine whether sets of items designed to measure three different subskills of reading comprehension of the Iowa Tests of Basic Skills (ITBSs) did, in fact, distinguish among these subskills. The three major skills objectives were: (1) facts; (2) generalizations; and (3) inferences. Data from…
ERIC Educational Resources Information Center
Lynam, Donald R.; Gaughan, Eric T.; Miller, Joshua D.; Miller, Drew J.; Mullins-Sweatt, Stephanie; Widiger, Thomas A.
2011-01-01
A new self-report assessment of the basic traits of psychopathy was developed with a general trait model of personality (five-factor model [FFM]) as a framework. Scales were written to assess maladaptive variants of the 18 FFM traits that are robustly related to psychopathy across a variety of perspectives including empirical correlations, expert…
ERIC Educational Resources Information Center
Vogel, Ronald J.
A study was conducted in 1976 of applicants who submitted corrections or amendments to their Student Eligibility Reports (SERs) for the Basic Educational Opportunity Grant (BEOG) Program. The objective was to review the applications corrections process and to determine factors linked to applicants' use of correction procedures. Attention was…
ERIC Educational Resources Information Center
Pennsylvania Blue Shield, Camp Hill.
A project developed a model curriculum to be delivered by computer-based instruction to teach the required literacy skills for entry workers in the health insurance industry. Literacy task analyses were performed for the targeted jobs and then validated with focus groups. The job tasks and related basic skills were divided into modules. The job…
The behavior of the Higgs field in the new inflationary universe
NASA Technical Reports Server (NTRS)
Guth, Alan H.; Pi, So-Young
1986-01-01
Answers are provided to questions about the standard model of the new inflationary universe (NIU) which have raised concerns about the model's validity. A baby toy problem which consists of the study of a single particle moving in one dimension under the influence of a potential with the form of an upside-down harmonic oscillator is studied, showing that the quantum mechanical wave function at large times is accurately described by classical physics. Then, an exactly soluble toy model for the behavior of the Higgs field in the NIU is described which should provide a reasonable approximation to the behavior of the Higgs field in the NIU. The dynamics of the toy model is described, and calculative results are reviewed which, the authors claim, provide strong evidence that the basic features of the standard picture are correct.
Pollak, Andrew N; Ficke, Col James R
2010-01-01
The fourth annual Extremity War Injuries (EWI) Symposium addressed ongoing challenges and opportunities in the management of combat-related musculoskeletal injury. The symposium, which also examined host-nation care and disaster preparedness and response, defined opportunities for synergy between several organizations with similar missions and goals. Within the Department of Defense, the Orthopaedic Extremity Trauma Research Program (OETRP) has funded basic research related to a series of protocols first identified and validated at prior EWI symposia. A well-funded clinical research arm of OETRP has been developed to help translate and validate research advances from each of the protocols. The Armed Forces Institute for Regenerative Medicine, a consortium of academic research institutions, employs a tissue-engineering approach to EWI challenges, particularly with regard to tissue loss. Programs within the National Institute of Arthritis and Musculoskeletal and Skin Diseases and throughout the National Institutes of Health have also expanded tissue-engineering efforts by emphasizing robust mechanistic basic science programs. Much of the clinical care delivered by US military medical personnel and nongovernmental agencies has been to host-nation populations; coordinating delivery to maximize the number of injured who receive care requires understanding of the breadth and scope of resources available within the war zone. Similarly, providing the most comprehensive care to the greatest number of injured in the context of domestic mass casualty requires discussion and planning by all groups involved.
Comino, María Rosa Salvador; Garcia, Victor Regife; López, Maria Auxiliadora Fernández; Feddersen, Berend; Roselló, María Luisa Martin; Sanftenberg, Linda; Schelling, Jörg
2017-07-11
Background Palliative medicine is an essential component of the health care system. Basic palliative care should be provided by primary care services (family physician and home nursing) with palliative-medical basic qualification. Often it is very difficult to identify patients that would profit from a specialized palliative care team. For the evaluation of the case complexity of a palliative patient, we present a Spanish diagnostic tool IDC-Pal, which tries to specify when, why and where a palliative patient should be referred. The aims of this study were the translation and cultural adaptation of the diagnostic tool for complexity in palliative care IDC-Pal to the German language, and the measurement of its feasibility and face validity. Methods During the first phase, a forward-backward translation with linguistic and cultural adaptation of the tool IDC-Pal as well as the validation of its content by a review committee was performed. During the second phase, the preliminary version of the tool was tested by 38 family physicians that were asked for a qualitative assessment using a 10-points Likert scale (1 = "strongly disagree" and 10 = "totally agree"). Finally, a definitive version was developed. Results The translation and adaptation were achieved without major problems. Both feasibility and apparent validity of the tool IDC-Pal were rated as high. The mean response in the Likert scale was 7.79, with a SD of 0.36. Participants strongly agreed on the apparent validity of the tool with a mean of 7.82 and a SD of 0.26 and on its feasibility with a mean of 7.79, and a SD of 0.39. Conclusions A conceptually, culturally and linguistically equivalent version of the original instrument IDC-Pal was obtained. German family physicians agreed on the usability of IDC-Pal as a tool for rating the case complexity of palliative patients. These results indicate that physicians in Bavaria and eventually in Germany could benefit of the full validation of IDC-Pal. © Georg Thieme Verlag KG Stuttgart · New York.
Health Literacy Impact on National Healthcare Utilization and Expenditure.
Rasu, Rafia S; Bawa, Walter Agbor; Suminski, Richard; Snella, Kathleen; Warady, Bradley
2015-08-17
Health literacy presents an enormous challenge in the delivery of effective healthcare and quality outcomes. We evaluated the impact of low health literacy (LHL) on healthcare utilization and healthcare expenditure. Database analysis used Medical Expenditure Panel Survey (MEPS) from 2005-2008 which provides nationally representative estimates of healthcare utilization and expenditure. Health literacy scores (HLSs) were calculated based on a validated, predictive model and were scored according to the National Assessment of Adult Literacy (NAAL). HLS ranged from 0-500. Health literacy level (HLL) and categorized in 2 groups: Below basic or basic (HLS <226) and above basic (HLS ≥226). Healthcare utilization expressed as a physician, nonphysician, or emergency room (ER) visits and healthcare spending. Expenditures were adjusted to 2010 rates using the Consumer Price Index (CPI). A P value of 0.05 or less was the criterion for statistical significance in all analyses. Multivariate regression models assessed the impact of the predicted HLLs on outpatient healthcare utilization and expenditures. All analyses were performed with SAS and STATA® 11.0 statistical software. The study evaluated 22 599 samples representing 503 374 648 weighted individuals nationally from 2005-2008. The cohort had an average age of 49 years and included more females (57%). Caucasian were the predominant racial ethnic group (83%) and 37% of the cohort were from the South region of the United States of America. The proportion of the cohort with basic or below basic health literacy was 22.4%. Annual predicted values of physician visits, nonphysician visits, and ER visits were 6.6, 4.8, and 0.2, respectively, for basic or below basic compared to 4.4, 2.6, and 0.1 for above basic. Predicted values of office and ER visits expenditures were $1284 and $151, respectively, for basic or below basic and $719 and $100 for above basic (P < .05). The extrapolated national estimates show that the annual costs for prescription alone for adults with LHL possibly associated with basic and below basic health literacy could potentially reach about $172 billion. Health literacy is inversely associated with healthcare utilization and expenditure. Individuals with below basic or basic HLL have greater healthcare utilization and expendituresspending more on prescriptions compared to individuals with above basic HLL. Public health strategies promoting appropriate education among individuals with LHL may help to improve health outcomes and reduce unnecessary healthcare visits and costs. © 2015 by Kerman University of Medical Sciences.
Nguyen, Khoa; Ben Khallouq, Bertha; Schuster, Amanda; Beevers, Christopher; Dil, Nyla; Kay, Denise; Kibble, Jonathan D; Harris, David M
2017-12-01
Most assessments of physiology in medical school use multiple choice tests that may not provide information about a student's critical thinking (CT) process. There are limited performance assessments, but high-fidelity patient simulations (HFPS) may be a feasible platform. The purpose of this pilot study was to determine whether a group's CT process could be observed over a series of HFPS. An instrument [Critical Thinking Skills Rating Instrument CTSRI)] was designed with the IDEAS framework. Fifteen groups of students participated in three HFPS that consisted of a basic knowledge quiz and introduction, HFPS session, and debriefing. HFPS were video recorded, and two raters reviewed and scored all HFPS encounters with the CTSRI independently. Interrater analysis suggested good reliability. There was a correlation between basic knowledge scores and three of the six observations on the CTSRI providing support for construct validity. The median CT ratings significantly increased for all observations between the groups' first and last simulation. However, there were still large percentages of video ratings that indicated students needed substantial prompting during the HFPS. The data from this pilot study suggest that it is feasible to observe CT skills in HFPS using the CTSRI. Based on the findings from this study, we strongly recommend that first-year medical students be competent in basic knowledge of the relevant physiology of the HFPS before participating, to minimize the risk of a poor learning experience. Copyright © 2017 the American Physiological Society.
Tavassoli, Teresa; Hoekstra, Rosa A; Baron-Cohen, Simon
2014-01-01
Questionnaire-based studies suggest atypical sensory perception in over 90% of individuals with autism spectrum conditions (ASC). Sensory questionnaire-based studies in ASC mainly record parental reports of their child's sensory experience; less is known about sensory reactivity in adults with ASC. Given the DSM-5 criteria for ASC now include sensory reactivity, there is a need for an adult questionnaire investigating basic sensory functioning. We aimed to develop and validate the Sensory Perception Quotient (SPQ), which assesses basic sensory hyper- and hyposensitivity across all five modalities. A total of 359 adults with (n = 196) and without (n = 163) ASC were asked to fill in the SPQ, the Sensory Over-Responsivity Inventory (SensOR) and the Autism-Spectrum Quotient (AQ) online. Adults with ASC reported more sensory hypersensitivity on the SPQ compared to controls (P < .001). SPQ scores were correlated with AQ scores both across groups (r = .-38) and within the ASC (r = -.18) and control groups (r = -.15). Principal component analyses conducted separately in both groups indicated that one factor comprising 35 items consistently assesses sensory hypersensitivity. The SPQ showed high internal consistency for both the total SPQ (Cronbach's alpha = .92) and the reduced 35-item version (alpha = .93). The SPQ was significantly correlated with the SensOR across groups (r = -.46) and within the ASC (r = -.49) and control group (r = -.21). The SPQ shows good internal consistency and concurrent validity and differentiates between adults with and without ASC. Adults with ASC report more sensitivity to sensory stimuli on the SPQ. Finally, greater sensory sensitivity is associated with more autistic traits. The SPQ provides a new tool to measure individual differences on this dimension.
The MCNP6 Analytic Criticality Benchmark Suite
DOE Office of Scientific and Technical Information (OSTI.GOV)
Brown, Forrest B.
2016-06-16
Analytical benchmarks provide an invaluable tool for verifying computer codes used to simulate neutron transport. Several collections of analytical benchmark problems [1-4] are used routinely in the verification of production Monte Carlo codes such as MCNP® [5,6]. Verification of a computer code is a necessary prerequisite to the more complex validation process. The verification process confirms that a code performs its intended functions correctly. The validation process involves determining the absolute accuracy of code results vs. nature. In typical validations, results are computed for a set of benchmark experiments using a particular methodology (code, cross-section data with uncertainties, and modeling)more » and compared to the measured results from the set of benchmark experiments. The validation process determines bias, bias uncertainty, and possibly additional margins. Verification is generally performed by the code developers, while validation is generally performed by code users for a particular application space. The VERIFICATION_KEFF suite of criticality problems [1,2] was originally a set of 75 criticality problems found in the literature for which exact analytical solutions are available. Even though the spatial and energy detail is necessarily limited in analytical benchmarks, typically to a few regions or energy groups, the exact solutions obtained can be used to verify that the basic algorithms, mathematics, and methods used in complex production codes perform correctly. The present work has focused on revisiting this benchmark suite. A thorough review of the problems resulted in discarding some of them as not suitable for MCNP benchmarking. For the remaining problems, many of them were reformulated to permit execution in either multigroup mode or in the normal continuous-energy mode for MCNP. Execution of the benchmarks in continuous-energy mode provides a significant advance to MCNP verification methods.« less
Emergency Medical Services Capacity for Prehospital Stroke Care in North Carolina
Brice, Jane H.; Evenson, Kelly R.; Rose, Kathryn M.; Suchindran, Chirayath M.; Rosamond, Wayne D.
2013-01-01
Introduction Prior assessments of emergency medical services (EMS) stroke capacity found deficiencies in education and training, use of protocols and screening tools, and planning for the transport of patients. A 2001 survey of North Carolina EMS providers found many EMS systems lacked basic stroke services. Recent statewide efforts have sought to standardize and improve prehospital stroke care. The objective of this study was to assess EMS stroke care capacity in North Carolina and evaluate statewide changes since 2001. Methods In June 2012, we conducted a web-based survey on stroke education and training and stroke care practices and policies among all EMS systems in North Carolina. We used the McNemar test to assess changes from 2001 to 2012. Results Of 100 EMS systems in North Carolina, 98 responded to our survey. Most systems reported providing stroke education and training (95%) to EMS personnel, using a validated stroke scale or screening tool (96%), and having a hospital prenotification policy (98%). Many were suboptimal in covering basic stroke educational topics (71%), always communicating stroke screen results to the destination hospital (46%), and always using a written destination plan (49%). Among 70 EMS systems for which we had data for 2001 and 2012, we observed significant improvements in education on stroke scales or screening tools (61% to 93%, P < .001) and use of validated stroke scales or screening tools (23% to 96%, P < .001). Conclusion Major improvements in EMS stroke care, especially in prehospital stroke screening, have occurred in North Carolina in the past decade, whereas other practices and policies, including use of destination plans, remain in need of improvement. PMID:24007677
DOE Office of Scientific and Technical Information (OSTI.GOV)
Ricci, P., E-mail: paolo.ricci@epfl.ch; Riva, F.; Theiler, C.
In the present work, a Verification and Validation procedure is presented and applied showing, through a practical example, how it can contribute to advancing our physics understanding of plasma turbulence. Bridging the gap between plasma physics and other scientific domains, in particular, the computational fluid dynamics community, a rigorous methodology for the verification of a plasma simulation code is presented, based on the method of manufactured solutions. This methodology assesses that the model equations are correctly solved, within the order of accuracy of the numerical scheme. The technique to carry out a solution verification is described to provide a rigorousmore » estimate of the uncertainty affecting the numerical results. A methodology for plasma turbulence code validation is also discussed, focusing on quantitative assessment of the agreement between experiments and simulations. The Verification and Validation methodology is then applied to the study of plasma turbulence in the basic plasma physics experiment TORPEX [Fasoli et al., Phys. Plasmas 13, 055902 (2006)], considering both two-dimensional and three-dimensional simulations carried out with the GBS code [Ricci et al., Plasma Phys. Controlled Fusion 54, 124047 (2012)]. The validation procedure allows progress in the understanding of the turbulent dynamics in TORPEX, by pinpointing the presence of a turbulent regime transition, due to the competition between the resistive and ideal interchange instabilities.« less
Development and Validation of a Diagnostic Grammar Test for Japanese Learners of English
ERIC Educational Resources Information Center
Koizumi, Rie; Sakai, Hideki; Ido, Takahiro; Ota, Hiroshi; Hayama, Megumi; Sato, Masatoshi; Nemoto, Akiko
2011-01-01
This article reports on the development and validation of the English Diagnostic Test of Grammar (EDiT Grammar) for Japanese learners of English. From among the many aspects of grammar, this test focuses on the knowledge of basic English noun phrases (NPs), especially their internal structures, because previous research has indicated the…
Coping with Health Problems: Developing a Reliable and Valid Multidimensional Measure.
ERIC Educational Resources Information Center
Endler, Norman S.; Parker, James D. A.; Summerfeldt, Laura J.
1998-01-01
A self-report measure, the Coping with Health Injuries and Problems Scale (CHIP), was developed to identify basic coping dimensions for responding to health problems. The CHIP factor structure, established with samples of 532 adults and 598 adults in Canada, is cross-validated with 390 general medical patients and 286 chronic back pain patients.…
easyCBM® Reading Criterion Related Validity Evidence: Grades K-1. Technical Report #1309
ERIC Educational Resources Information Center
Lai, Cheng-Fei; Alonzo, Julie; Tindal, Gerald
2013-01-01
In this technical report, we present the results of a study to gather criterion-related evidence for Grade K-1 easyCBM® reading measures. We used correlations to examine the relation between the easyCBM® measures and other published measures with known reliability and validity evidence, including the Dynamic Indicators of Basic Early Literacy…
Examining the Validity of Self-Reports on Scales Measuring Students' Strategic Processing
ERIC Educational Resources Information Center
Samuelstuen, Marit S.; Braten, Ivar
2007-01-01
Background: Self-report inventories trying to measure strategic processing at a global level have been much used in both basic and applied research. However, the validity of global strategy scores is open to question because such inventories assess strategy perceptions outside the context of specific task performance. Aims: The primary aim was to…
Hsiao, Pei-Chi; Yu, Wan-Hui; Lee, Shih-Chieh; Chen, Mei-Hsiang; Hsieh, Ching-Lin
2018-06-14
The responsiveness and predictive validity of the Tablet-based Symbol Digit Modalities Test (T-SDMT) are unknown, which limits the utility of the T-SDMT in both clinical and research settings. The purpose of this study was to examine the responsiveness and predictive validity of the T-SDMT in inpatients with stroke. A follow-up, repeated-assessments design. One rehabilitation unit at a local medical center. A total of 50 inpatients receiving rehabilitation completed T-SDMT assessments at admission to and discharge from a rehabilitation ward. The median follow-up period was 14 days. The Barthel index (BI) was assessed at discharge and was used as the criterion of the predictive validity. The mean changes in the T-SDMT scores between admission and discharge were statistically significant (paired t-test = 3.46, p = 0.001). The T-SDMT scores showed a nearly moderate standardized response mean (0.49). A moderate association (Pearson's r = 0.47) was found between the scores of the T-SDMT at admission and those of the BI at discharge, indicating good predictive validity of the T-SDMT. Our results support the responsiveness and predictive validity of the T-SDMT in patients with stroke receiving rehabilitation in hospitals. This study provides empirical evidence supporting the use of the T-SDMT as an outcome measure for assessing processingspeed in inpatients with stroke. The scores of the T-SDMT could be used to predict basic activities of daily living function in inpatients with stroke.
ERIC Educational Resources Information Center
Salem, Ashraf Atta M. S.
2013-01-01
The present study aimed at investigating the effects of using a program based on the writing workshop approach on developing basic writing skills of prospective teachers of English in Hurgada faculty of Education. For that purpose, the researcher constructed and validated a teaching program based on the writing workshop approach, checklist of the…
A Manual Control Test for the Detection and Deterrence of Impaired Drivers
NASA Technical Reports Server (NTRS)
Stein, A. C.; Allen, R. W.; Jex, H. R.
1984-01-01
A brief manual control test and a decision strategy were developed, laboratory tested, and field validated which provide a means for detecting human operator impairment from alcohol or other drugs. The test requires the operator to stabilize progressively unstable controlled element dynamics. Control theory and experimental data verify that the human operator's control ability on this task is constrained by basic cybernetic characteristics, and that task performance is reliably affected by impairment effects on these characteristics. Assessment of human operator control ability is determined by a statistically based decision strategy. The operator is allowed several chances to exceed a preset pass criterion. Procedures are described for setting the pass criterion based on individual ability and a desired unimpaired failure rate. These procedures were field tested with apparatus installed in automobiles that were designed to discourage drunk drivers from operating their vehicles. This test program demonstrated that the control task and detection strategy could be applied in a practical setting to screen human operators for impairment in their basic cybernetic skills.
Developmental Neurotoxicology: History and Outline of ...
The present work provides a brief review of basic concepts in developmental neurotoxicology, as well as current representative testing guidelines for evaluating developmental neurotoxicity (DNT) of xenobiotics. Historically, DNT was initially recognized as a “functional” teratogenicity: the main concern was that prenatal and/or early postnatal exposures to chemicals during critical periods of central nervous system (CNS) development would cause later functional abnormalities of the brain. Current internationally harmonized DNT study guidelines are thus intended to predict adverse effects of test compounds on the developing CNS by observing such postnatal parameters as motor activity, startle response, and learning and memory, as well as neropathological alterations. The reliability of current DNT study guidelines and sensitivity of testing methodologies recommended in these guidelines have been confirmed by retrospective evaluations of the many international and domestic collaborative validation studies in developed nations including Japan. Invited review with brief review of basic concepts in developmental neurotoxicology, as well as current representative testing guidelines for evaluating developmental neurotoxicity (DNT) of xenobiotics.
Experimental Evaluation of a Planning Language Suitable for Formal Verification
NASA Technical Reports Server (NTRS)
Butler, Rick W.; Munoz, Cesar A.; Siminiceanu, Radu I.
2008-01-01
The marriage of model checking and planning faces two seemingly diverging alternatives: the need for a planning language expressive enough to capture the complexity of real-life applications, as opposed to a language simple, yet robust enough to be amenable to exhaustive verification and validation techniques. In an attempt to reconcile these differences, we have designed an abstract plan description language, ANMLite, inspired from the Action Notation Modeling Language (ANML) [17]. We present the basic concepts of the ANMLite language as well as an automatic translator from ANMLite to the model checker SAL (Symbolic Analysis Laboratory) [7]. We discuss various aspects of specifying a plan in terms of constraints and explore the implications of choosing a robust logic behind the specification of constraints, rather than simply propose a new planning language. Additionally, we provide an initial assessment of the efficiency of model checking to search for solutions of planning problems. To this end, we design a basic test benchmark and study the scalability of the generated SAL models in terms of plan complexity.
Sugand, Kapil; Wescott, Robert A; Carrington, Richard; Hart, Alister; Van Duren, Bernard H
2018-05-10
Background and purpose - Simulation is an adjunct to surgical education. However, nothing can accurately simulate fluoroscopic procedures in orthopedic trauma. Current options for training with fluoroscopy are either intraoperative, which risks radiation, or use of expensive and unrealistic virtual reality simulators. We introduce FluoroSim, an inexpensive digital fluoroscopy simulator without the need for radiation. Patients and methods - This was a multicenter study with 26 surgeons in which everyone completed 1 attempt at inserting a guide-wire into a femoral dry bone using surgical equipment and FluoroSim. 5 objective performance metrics were recorded in real-time to assess construct validity. The surgeons were categorized based on the number of dynamic hip screws (DHS) performed: novices (< 10), intermediates (10-39) and experts (≥ 40). A 7-point Likert scale questionnaire assessed the face and content validity of FluoroSim. Results - Construct validity was present for 2 clinically validated metrics in DHS surgery. Experts and intermediates statistically significantly outperformed novices for tip-apex distance and for cut-out rate. Novices took the least number of radiographs. Face and content validity were also observed. Interpretation - FluoroSim discriminated between novice and intermediate or expert surgeons based on tip-apex distance and cut-out rate while demonstrating face and content validity. FluoroSim provides a useful adjunct to orthopedic training. Our findings concur with results from studies using other simulation modalities. FluoroSim can be implemented for education easily and cheaply away from theater in a safe and controlled environment.
Havemann, Maria Cecilie; Dalsgaard, Torur; Sørensen, Jette Led; Røssaak, Kristin; Brisling, Steffen; Mosgaard, Berit Jul; Høgdall, Claus; Bjerrum, Flemming
2018-05-14
Increasing focus on patient safety makes it important to ensure surgical competency among surgeons before operating on patients. The objective was to gather validity evidence for a virtual-reality simulator test for robotic surgical skills and evaluate its potential as a training tool. Surgeons with varying experience in robotic surgery were recruited: novices (zero procedures), intermediates (1-50), experienced (> 50). Five experienced surgeons rated five exercises on the da Vinci Skills Simulator. Participants were tested using the five exercises. Participants were invited back 3 times and completed a total of 10 attempts per exercise. The outcome was the average simulator performance score for the 5 exercises. 32 participants from 5 surgical specialties were included. 38 participants completed all 4 sessions. A moderate correlation between the average total score and robotic experience was identified for the first attempt (Spearman r = 0.58; p = 0.0004). A difference in average total score was observed between novices and intermediates [median score 61% (IQR 52-66) vs. 83% (IQR 75-91), adjusted p < 0.0001], as well as novices and experienced [median score 61% (IQR 52-66) vs. 80 (IQR 69-85), adjusted p = 0.002]. All three groups improved their performance between the 1st and 10th attempts (p < 0.00). This study describes validity evidence for a virtual-reality simulator for basic robotic surgical skills, which can be used for assessment of basic competency and as a training tool. However, more validity evidence is needed before it can be used for certification or high-stakes assessment.
Real-Time Sensor Validation System Developed for Reusable Launch Vehicle Testbed
NASA Technical Reports Server (NTRS)
Jankovsky, Amy L.
1997-01-01
A real-time system for validating sensor health has been developed for the reusable launch vehicle (RLV) program. This system, which is part of the propulsion checkout and control system (PCCS), was designed for use in an integrated propulsion technology demonstrator testbed built by Rockwell International and located at the NASA Marshall Space Flight Center. Work on the sensor health validation system, a result of an industry-NASA partnership, was completed at the NASA Lewis Research Center, then delivered to Marshall for integration and testing. The sensor validation software performs three basic functions: it identifies failed sensors, it provides reconstructed signals for failed sensors, and it identifies off-nominal system transient behavior that cannot be attributed to a failed sensor. The code is initiated by host software before the start of a propulsion system test, and it is called by the host program every control cycle. The output is posted to global memory for use by other PCCS modules. Output includes a list indicating the status of each sensor (i.e., failed, healthy, or reconstructed) and a list of features that are not due to a sensor failure. If a sensor failure is found, the system modifies that sensor's data array by substituting a reconstructed signal, when possible, for use by other PCCS modules.
Role of natural analogs in performance assessment of nuclear waste repositories
DOE Office of Scientific and Technical Information (OSTI.GOV)
Sagar, B.; Wittmeyer, G.W.
1995-09-01
Mathematical models of the flow of water and transport of radionuclides in porous media will be used to assess the ability of deep geologic repositories to safely contain nuclear waste. These models must, in some sense, be validated to ensure that they adequately describe the physical processes occurring within the repository and its geologic setting. Inasmuch as the spatial and temporal scales over which these models must be applied in performance assessment are very large, validation of these models against laboratory and small-scale field experiments may be considered inadequate. Natural analogs may provide validation data that are representative of physico-chemicalmore » processes that occur over spatial and temporal scales as large or larger than those relevant to repository design. The authors discuss the manner in which natural analog data may be used to increase confidence in performance assessment models and conclude that, while these data may be suitable for testing the basic laws governing flow and transport, there is insufficient control of boundary and initial conditions and forcing functions to permit quantitative validation of complex, spatially distributed flow and transport models. The authors also express their opinion that, for collecting adequate data from natural analogs, resources will have to be devoted to them that are much larger than are devoted to them at present.« less
Pitchford, Nicola J; Outhwaite, Laura A
2016-01-01
Assessment of cognitive and motor functions is fundamental for developmental and neuropsychological profiling. Assessments are usually conducted on an individual basis, with a trained examiner, using standardized paper and pencil tests, and can take up to an hour or more to complete, depending on the nature of the test. This makes traditional standardized assessments of child development largely unsuitable for use in low-income countries. Touch screen tablets afford the opportunity to assess cognitive functions in groups of participants, with untrained administrators, with precision recording of responses, thus automating the assessment process. In turn, this enables cognitive profiling to be conducted in contexts where access to qualified examiners and standardized assessments are rarely available. As such, touch screen assessments could provide a means of assessing child development in both low- and high-income countries, which would afford cross-cultural comparisons to be made with the same assessment tool. However, before touch screen tablet assessments can be used for cognitive profiling in low-to-high-income countries they need to be shown to provide reliable and valid measures of performance. We report the development of a new touch screen tablet assessment of basic cognitive and motor functions for use with early years primary school children in low- and high-income countries. Measures of spatial intelligence, visual attention, short-term memory, working memory, manual processing speed, and manual coordination are included as well as mathematical knowledge. To investigate if this new touch screen assessment tool can be used for cross-cultural comparisons we administered it to a sample of children ( N = 283) spanning standards 1-3 in a low-income country, Malawi, and a smaller sample of children ( N = 70) from first year of formal schooling from a high-income country, the UK. Split-half reliability, test-retest reliability, face validity, convergent construct validity, predictive criterion validity, and concurrent criterion validity were investigated. Results demonstrate "proof of concept" that touch screen tablet technology can provide reliable and valid psychometric measures of performance in the early years, highlighting its potential to be used in cross-cultural comparisons and research.
Peraman, Ramalingam; Nayakanti, Devanna; Dugga, Hari Hara Theja; Kodikonda, Sudhakara
2013-01-01
A validated stability-indicating RP-HPLC method for etofenamate (ETF) was developed by separating its degradation products on a C18 (250 mm × 4.6 mm 5 μm) Qualisil BDS column using a phosphate buffer (pH-adjusted to 6.0 with orthophosphoric acid) and methanol in the ratio of 20:80 % v/v as the mobile phase at a flow rate of 1.0 mL/min. The column effluents were monitored by a photodiode array detector set at 286 nm. The method was validated in terms of specificity, linearity, accuracy, precision, detection limit, quantification limit, and robustness. Forced degradation of etofenamate was carried out under acidic, basic, thermal, photo, and peroxide conditions and the major degradation products of acidic and basic degradation were isolated and characterized by 1H-NMR, 13C-NMR, and mass spectral studies. The mass balance of the method varied between 92–99%. PMID:24482770
Methodology for testing and validating knowledge bases
NASA Technical Reports Server (NTRS)
Krishnamurthy, C.; Padalkar, S.; Sztipanovits, J.; Purves, B. R.
1987-01-01
A test and validation toolset developed for artificial intelligence programs is described. The basic premises of this method are: (1) knowledge bases have a strongly declarative character and represent mostly structural information about different domains, (2) the conditions for integrity, consistency, and correctness can be transformed into structural properties of knowledge bases, and (3) structural information and structural properties can be uniformly represented by graphs and checked by graph algorithms. The interactive test and validation environment have been implemented on a SUN workstation.
Castrejon, I; Carmona, L; Agrinier, N; Andres, M; Briot, K; Caron, M; Christensen, R; Consolaro, A; Curbelo, R; Ferrer, Montserrat; Foltz, Violaine; Gonzalez, C; Guillemin, F; Machado, P M; Prodinger, Birgit; Ravelli, A; Scholte-Voshaar, M; Uhlig, T; van Tuyl, L H D; Zink, A; Gossec, L
2015-01-01
Patient reported outcomes (PROs) are relevant in rheumatology. Variable accessibility and validity of commonly used PROs are obstacles to homogeneity in evidence synthesis. The objective of this project was to provide a comprehensive library of "validated PROs". A launch meeting with rheumatologists, PROs methodological experts, and patients, was held to define the library's aims and scope, and basic requirements. To feed the library we performed systematic reviews on selected diseases and domains. Relevant information on PROs was collected using standardised data collection forms based on the COSMIN checklist. The EULAR Outcomes Measures Library (OML), whose aims are to provide and to advise on PROs on a user-friendly manner albeit based on scientific grounds, has been launched and made accessible to all. PROs currently included cover any domain and, are generic or specifically target to the following diseases: rheumatoid arthritis, osteoarthritis, spondyloarthritis, low back pain, systemic lupus erythematosus, gout, osteoporosis, juvenile idiopathic arthritis, and fibromyalgia. Up to 236 instruments (106 generic and 130 specific) have been identified, evaluated, and included. The systematic review for SLE, which yielded 10 specific instruments, is presented here as an example. The OML website includes, for each PRO, information on the construct being measured and the extent of validation, recommendations for use, and available versions; it also contains a glossary on common validation terms. The OML is an in progress library led by rheumatologists, related professionals and patients, that will help to better understand and apply PROs in rheumatic and musculoskeletal diseases.
Minimum-complexity helicopter simulation math model
NASA Technical Reports Server (NTRS)
Heffley, Robert K.; Mnich, Marc A.
1988-01-01
An example of a minimal complexity simulation helicopter math model is presented. Motivating factors are the computational delays, cost, and inflexibility of the very sophisticated math models now in common use. A helicopter model form is given which addresses each of these factors and provides better engineering understanding of the specific handling qualities features which are apparent to the simulator pilot. The technical approach begins with specification of features which are to be modeled, followed by a build up of individual vehicle components and definition of equations. Model matching and estimation procedures are given which enable the modeling of specific helicopters from basic data sources such as flight manuals. Checkout procedures are given which provide for total model validation. A number of possible model extensions and refinement are discussed. Math model computer programs are defined and listed.
American Alcohol Photo Stimuli (AAPS): A standardized set of alcohol and matched non-alcohol images.
Stauffer, Christopher S; Dobberteen, Lily; Woolley, Joshua D
2017-11-01
Photographic stimuli are commonly used to assess cue reactivity in the research and treatment of alcohol use disorder. The stimuli used are often non-standardized, not properly validated, and poorly controlled. There are no previously published, validated, American-relevant sets of alcohol images created in a standardized fashion. We aimed to: 1) make available a standardized, matched set of photographic alcohol and non-alcohol beverage stimuli, 2) establish face validity, the extent to which the stimuli are subjectively viewed as what they are purported to be, and 3) establish construct validity, the degree to which a test measures what it claims to be measuring. We produced a standardized set of 36 images consisting of American alcohol and non-alcohol beverages matched for basic color, form, and complexity. A total of 178 participants (95 male, 82 female, 1 genderqueer) rated each image for appetitiveness. An arrow-probe task, in which matched pairs were categorized after being presented for 200 ms, assessed face validity. Criteria for construct validity were met if variation in AUDIT scores were associated with variation in performance on tasks during alcohol image presentation. Overall, images were categorized with >90% accuracy. Participants' AUDIT scores correlated significantly with alcohol "want" and "like" ratings [r(176) = 0.27, p = <0.001; r(176) = 0.36, p = <0.001] and arrow-probe latency [r(176) = -0.22, p = 0.004], but not with non-alcohol outcomes. Furthermore, appetitive ratings and arrow-probe latency for alcohol, but not non-alcohol, differed significantly for heavy versus light drinkers. Our image set provides valid and reliable alcohol stimuli for both explicit and implicit tests of cue reactivity. The use of standardized, validated, reliable image sets may improve consistency across research and treatment paradigms.
Risk for suicide and risk for violence: a case for separating the current violence diagnoses.
Smith, J E; Early, J A; Green, P T; Lauck, D L; Oblaczynski, C; Smochek, M R; Wright, G
1997-01-01
To identify accurate descriptive terms for risk for violence and risk for suicide and to provide operational definitions for these terms The Delphi technique, with two rounds, was used to differentiate the operational definitions that represent risk for suicide from those that represent risk for violence. The expert panel consisted of 23 healthcare professionals with a minimum of a master's degree. In addition to the expert panel, a control group (N = 11) participated to assess content validity. Thirty-six definitions were agreed upon for suicide, 39 for violence. These definitions represent the basic distinctions between the two behavioral manifestations.
NASA Astrophysics Data System (ADS)
Imai, Takashi; Ota, Kaiichiro; Aoyagi, Toshio
2017-02-01
Phase reduction has been extensively used to study rhythmic phenomena. As a result of phase reduction, the rhythm dynamics of a given system can be described using the phase response curve. Measuring this characteristic curve is an important step toward understanding a system's behavior. Recently, a basic idea for a new measurement method (called the multicycle weighted spike-triggered average method) was proposed. This paper confirms the validity of this method by providing an analytical proof and demonstrates its effectiveness in actual experimental systems by applying the method to an oscillating electric circuit. Some practical tips to use the method are also presented.
ERIC Educational Resources Information Center
Centeno, Vera G.; Kauko, Jaakko; Candido, Helena H. D.
2018-01-01
In our present research we address the question of whether it is valid to apply the Quality Assurance and Evaluation (QAE) umbrella concept, which was formulated to explain new phenomena in European educational governance, to similar developments in Brazilian basic education. This led us to reflect on the possible pitfalls and potential strengths…
ERIC Educational Resources Information Center
Milenkovic, Dusica D.; Hrin, Tamara N.; Segedinac, Mirjana D.; Horvat, Sasa
2016-01-01
This study describes the development and application of a three-tier test as a valid and reliable tool in diagnosing students' misconceptions regarding some basic concepts about carbohydrates. The test was administrated to students of the Pharmacy Department at the University of Bijeljina (Serb Republic). The results denoted construct and content…
Validation of the Impact of Event Scale for Psychological Sequelae of Combat.
ERIC Educational Resources Information Center
Schwarzwald, Joseph; And Others
1987-01-01
Attempted to validate the Impact of Event Scale (IES) with a large sample of males who had been exposed to combat trauma. The IES was shown to be sensitive to differences between and within the three subject groups (combat stress reaction group, combat control group, noncombat control group). Four basic patterns of intrusion and avoidance were…
Predictive Validity of a Multiple-Choice Test for Placement in a Community College
ERIC Educational Resources Information Center
Verbout, Mary F.
2013-01-01
Multiple-choice tests of punctuation and usage are used throughout the United States to assess the writing skills of new community college students in order to place them in either a basic writing course or first-year composition. To determine whether using the COMPASS Writing Test (CWT) is a valid placement at a community college, student test…
Update: Validation, Edits, and Application Processing. Phase II and Error-Prone Model Report.
ERIC Educational Resources Information Center
Gray, Susan; And Others
An update to the Validation, Edits, and Application Processing and Error-Prone Model Report (Section 1, July 3, 1980) is presented. The objective is to present the most current data obtained from the June 1980 Basic Educational Opportunity Grant applicant and recipient files and to determine whether the findings reported in Section 1 of the July…
The Error Prone Model and the Basic Grants Validation Selection System. Draft Final Report.
ERIC Educational Resources Information Center
System Development Corp., Falls Church, VA.
An evaluation of existing and proposed mechanisms to ensure data accuracy for the Pell Grant program is reported, and recommendations for efficient detection of fraud and error in the program are offered. One study objective was to examine the existing system of pre-established criteria (PEC), which are validation criteria that select students on…
Commissioning and validation of COMPASS system for VMAT patient specific quality assurance
NASA Astrophysics Data System (ADS)
Pimthong, J.; Kakanaporn, C.; Tuntipumiamorn, L.; Laojunun, P.; Iampongpaiboon, P.
2016-03-01
Pre-treatment patient specific quality assurance (QA) of advanced treatment techniques such as volumetric modulated arc therapy (VMAT) is one of important QA in radiotherapy. The fast and reliable dosimetric device is required. The objective of this study is to commission and validate the performance of COMPASS system for dose verification of VMAT technique. The COMPASS system is composed of an array of ionization detectors (MatriXX) mounted to the gantry using a custom holder and software for the analysis and visualization of QA results. We validated the COMPASS software for basic and advanced clinical application. For the basic clinical study, the simple open field in various field sizes were validated in homogeneous phantom. And the advanced clinical application, the fifteen prostate and fifteen nasopharyngeal cancers VMAT plans were chosen to study. The treatment plans were measured by the MatriXX. The doses and dose-volume histograms (DVHs) reconstructed from the fluence measurements were compared to the TPS calculated plans. And also, the doses and DVHs computed using collapsed cone convolution (CCC) Algorithm were compared with Eclipse TPS calculated plans using Analytical Anisotropic Algorithm (AAA) that according to dose specified in ICRU 83 for PTV.
Current status and standards for establishment of hemodialysis units in Korea
Lee, Young-Ki; Kim, Kiwon
2013-01-01
The number of hemodialysis patients and dialysis facilities is increasing each year, but there are no surveillance programs validating that the services and equipment of each hemodialysis unit meet specified safety and quality standards. There is a concern that excessive competition and illegal activities committed by some dialysis facilities may violate patients' right to health. Contrastingly, developed countries often have their own survey program to provide initial certification and monitoring to ensure that these clinics continue to meet basic requirements. Because hemodialysis units provide renal replacement therapy to critical patients suffering from severe chronic renal failure, appropriate legal regulation is important for the provision of initial certification and maintenance of facility, equipment, and human resource quality. Therefore, several standards providing minimum requirements for the area of hemodialysis unit, equipment for emergency care, physician and nurse staffs, water purification and quality management are urgently needed. PMID:23682218
Best practices for evaluating single nucleotide variant calling methods for microbial genomics
Olson, Nathan D.; Lund, Steven P.; Colman, Rebecca E.; Foster, Jeffrey T.; Sahl, Jason W.; Schupp, James M.; Keim, Paul; Morrow, Jayne B.; Salit, Marc L.; Zook, Justin M.
2015-01-01
Innovations in sequencing technologies have allowed biologists to make incredible advances in understanding biological systems. As experience grows, researchers increasingly recognize that analyzing the wealth of data provided by these new sequencing platforms requires careful attention to detail for robust results. Thus far, much of the scientific Communit’s focus for use in bacterial genomics has been on evaluating genome assembly algorithms and rigorously validating assembly program performance. Missing, however, is a focus on critical evaluation of variant callers for these genomes. Variant calling is essential for comparative genomics as it yields insights into nucleotide-level organismal differences. Variant calling is a multistep process with a host of potential error sources that may lead to incorrect variant calls. Identifying and resolving these incorrect calls is critical for bacterial genomics to advance. The goal of this review is to provide guidance on validating algorithms and pipelines used in variant calling for bacterial genomics. First, we will provide an overview of the variant calling procedures and the potential sources of error associated with the methods. We will then identify appropriate datasets for use in evaluating algorithms and describe statistical methods for evaluating algorithm performance. As variant calling moves from basic research to the applied setting, standardized methods for performance evaluation and reporting are required; it is our hope that this review provides the groundwork for the development of these standards. PMID:26217378
ERIC Educational Resources Information Center
Mark, Jorie Lester
A questionnaire was distributed to 1,305 companies to study the basic skills training provided. Of 62 responses, 41 companies had basic skills training programs. Respondents represented these types of companies: communications and utilities, finance and insurance, manufacturing, wholesalers, retailers, health and hospitals, and mining, and had…
Contreras Ortiz, Oscar; Rizk, Diaa Ee; Falconi, Gabriele; Schreiner, Lucas; Gorbea Chávez, Viridiana
2017-02-01
For four decades, the training for fellows in Urogynecology has been defined by taking into account the proposals of the relevant international societies. Primary health care providers and general OB/GYN practitioners could not find validated guidelines for the integration of knowledge in pelvic floor dysfunctions. The FIGO Working Group (FWG) in Pelvic Floor Medicine and Reconstructive Surgery has looked for the consensus of international opinion leaders in order to develop a set of minimal requirements of knowledge and skills in this area. This manuscript is divided into three categories of knowledge and skills, these are: to know, to understand, and to perform in order to offer the patients a more holistic health care in this area. The FWG reached consensus on the minimal requirements of knowledge and skills regarding each of the enabling objectives identified for postgraduate obstetrics and gynecology physicians and for residents in obstetrics and gynecology. Our goal is to propose and validate the basic objectives of minimal knowledge in pelvic floor medicine and reconstructive surgery. Neurourol. Urodynam. 36:514-517, 2017. © 2015 Wiley Periodicals, Inc. © 2015 Wiley Periodicals, Inc.
Nogami, Kentaro; Taniguchi, Shogo; Ichiyama, Tomoko
2016-01-01
The aim of this study was to investigate the correlation between basic life support skills in dentists who had completed the American Heart Association's Basic Life Support (BLS) Healthcare Provider qualification and time since course completion. Thirty-six dentists who had completed the 2005 BLS Healthcare Provider course participated in the study. We asked participants to perform 2 cycles of cardiopulmonary resuscitation on a mannequin and evaluated basic life support skills. Dentists who had previously completed the BLS Healthcare Provider course displayed both prolonged reaction times, and the quality of their basic life support skills deteriorated rapidly. There were no correlations between basic life support skills and time since course completion. Our results suggest that basic life support skills deteriorate rapidly for dentists who have completed the BLS Healthcare Provider. Newer guidelines stressing chest compressions over ventilation may help improve performance over time, allowing better cardiopulmonary resuscitation in dental office emergencies. Moreover, it may be effective to provide a more specialized version of the life support course to train the dentists, stressing issues that may be more likely to occur in the dental office.
Modeling techniques for quantum cascade lasers
NASA Astrophysics Data System (ADS)
Jirauschek, Christian; Kubis, Tillmann
2014-03-01
Quantum cascade lasers are unipolar semiconductor lasers covering a wide range of the infrared and terahertz spectrum. Lasing action is achieved by using optical intersubband transitions between quantized states in specifically designed multiple-quantum-well heterostructures. A systematic improvement of quantum cascade lasers with respect to operating temperature, efficiency, and spectral range requires detailed modeling of the underlying physical processes in these structures. Moreover, the quantum cascade laser constitutes a versatile model device for the development and improvement of simulation techniques in nano- and optoelectronics. This review provides a comprehensive survey and discussion of the modeling techniques used for the simulation of quantum cascade lasers. The main focus is on the modeling of carrier transport in the nanostructured gain medium, while the simulation of the optical cavity is covered at a more basic level. Specifically, the transfer matrix and finite difference methods for solving the one-dimensional Schrödinger equation and Schrödinger-Poisson system are discussed, providing the quantized states in the multiple-quantum-well active region. The modeling of the optical cavity is covered with a focus on basic waveguide resonator structures. Furthermore, various carrier transport simulation methods are discussed, ranging from basic empirical approaches to advanced self-consistent techniques. The methods include empirical rate equation and related Maxwell-Bloch equation approaches, self-consistent rate equation and ensemble Monte Carlo methods, as well as quantum transport approaches, in particular the density matrix and non-equilibrium Green's function formalism. The derived scattering rates and self-energies are generally valid for n-type devices based on one-dimensional quantum confinement, such as quantum well structures.
Modeling techniques for quantum cascade lasers
DOE Office of Scientific and Technical Information (OSTI.GOV)
Jirauschek, Christian; Kubis, Tillmann
2014-03-15
Quantum cascade lasers are unipolar semiconductor lasers covering a wide range of the infrared and terahertz spectrum. Lasing action is achieved by using optical intersubband transitions between quantized states in specifically designed multiple-quantum-well heterostructures. A systematic improvement of quantum cascade lasers with respect to operating temperature, efficiency, and spectral range requires detailed modeling of the underlying physical processes in these structures. Moreover, the quantum cascade laser constitutes a versatile model device for the development and improvement of simulation techniques in nano- and optoelectronics. This review provides a comprehensive survey and discussion of the modeling techniques used for the simulation ofmore » quantum cascade lasers. The main focus is on the modeling of carrier transport in the nanostructured gain medium, while the simulation of the optical cavity is covered at a more basic level. Specifically, the transfer matrix and finite difference methods for solving the one-dimensional Schrödinger equation and Schrödinger-Poisson system are discussed, providing the quantized states in the multiple-quantum-well active region. The modeling of the optical cavity is covered with a focus on basic waveguide resonator structures. Furthermore, various carrier transport simulation methods are discussed, ranging from basic empirical approaches to advanced self-consistent techniques. The methods include empirical rate equation and related Maxwell-Bloch equation approaches, self-consistent rate equation and ensemble Monte Carlo methods, as well as quantum transport approaches, in particular the density matrix and non-equilibrium Green's function formalism. The derived scattering rates and self-energies are generally valid for n-type devices based on one-dimensional quantum confinement, such as quantum well structures.« less
Nelissen, Ellen; Ersdal, Hege; Ostergaard, Doris; Mduma, Estomih; Broerse, Jacqueline; Evjen-Olsen, Bjørg; van Roosmalen, Jos; Stekelenburg, Jelle
2014-03-01
To evaluate "Helping Mothers Survive Bleeding After Birth" (HMS BAB) simulation-based training in a low-resource setting. Educational intervention study. Rural referral hospital in Northern Tanzania. Clinicians, nurse-midwives, medical attendants, and ambulance drivers involved in maternity care. In March 2012, health care workers were trained in HMS BAB, a half-day simulation-based training, using a train-the-trainer model. The training focused on basic delivery care, active management of third stage of labor, and treatment of postpartum hemorrhage, including bimanual uterine compression. Evaluation questionnaires provided information on course perception. Knowledge, skills, and confidence of facilitators and learners were tested before and after training. Four master trainers trained eight local facilitators, who subsequently trained 89 learners. After training, all facilitators passed the knowledge test, but pass rates for the skills test were low (29% pass rate for basic delivery and 0% pass rate for management of postpartum hemorrhage). Evaluation revealed that HMS BAB training was considered acceptable and feasible, although more time should be allocated for training, and teaching materials should be translated into the local language. Knowledge, skills, and confidence of learners increased significantly immediately after training. However, overall pass rates for skills tests of learners after training were low (3% pass rate for basic delivery and management of postpartum hemorrhage). The HMS BAB simulation-based training has potential to contribute to education of health care providers. We recommend a full day of training and validation of the facilitators to improve the training. © 2013 Nordic Federation of Societies of Obstetrics and Gynecology.
A Diagnostic Assessment for Introductory Molecular and Cell Biology
Wood, William B.; Martin, Jennifer M.; Guild, Nancy A.; Vicens, Quentin; Knight, Jennifer K.
2010-01-01
We have developed and validated a tool for assessing understanding of a selection of fundamental concepts and basic knowledge in undergraduate introductory molecular and cell biology, focusing on areas in which students often have misconceptions. This multiple-choice Introductory Molecular and Cell Biology Assessment (IMCA) instrument is designed for use as a pre- and posttest to measure student learning gains. To develop the assessment, we first worked with faculty to create a set of learning goals that targeted important concepts in the field and seemed likely to be emphasized by most instructors teaching these subjects. We interviewed students using open-ended questions to identify commonly held misconceptions, formulated multiple-choice questions that included these ideas as distracters, and reinterviewed students to establish validity of the instrument. The assessment was then evaluated by 25 biology experts and modified based on their suggestions. The complete revised assessment was administered to more than 1300 students at three institutions. Analysis of statistical parameters including item difficulty, item discrimination, and reliability provides evidence that the IMCA is a valid and reliable instrument with several potential uses in gauging student learning of key concepts in molecular and cell biology. PMID:21123692
Evolving Improvements to TRMM Ground Validation Rainfall Estimates
NASA Technical Reports Server (NTRS)
Robinson, M.; Kulie, M. S.; Marks, D. A.; Wolff, D. B.; Ferrier, B. S.; Amitai, E.; Silberstein, D. S.; Fisher, B. L.; Wang, J.; Einaudi, Franco (Technical Monitor)
2000-01-01
The primary function of the TRMM Ground Validation (GV) Program is to create GV rainfall products that provide basic validation of satellite-derived precipitation measurements for select primary sites. Since the successful 1997 launch of the TRMM satellite, GV rainfall estimates have demonstrated systematic improvements directly related to improved radar and rain gauge data, modified science techniques, and software revisions. Improved rainfall estimates have resulted in higher quality GV rainfall products and subsequently, much improved evaluation products for the satellite-based precipitation estimates from TRMM. This presentation will demonstrate how TRMM GV rainfall products created in a semi-automated, operational environment have evolved and improved through successive generations. Monthly rainfall maps and rainfall accumulation statistics for each primary site will be presented for each stage of GV product development. Contributions from individual product modifications involving radar reflectivity (Ze)-rain rate (R) relationship refinements, improvements in rain gauge bulk-adjustment and data quality control processes, and improved radar and gauge data will be discussed. Finally, it will be demonstrated that as GV rainfall products have improved, rainfall estimation comparisons between GV and satellite have converged, lending confidence to the satellite-derived precipitation measurements from TRMM.
Basic psychometric properties of the transfer assessment instrument (version 3.0).
Tsai, Chung-Ying; Rice, Laura A; Hoelmer, Claire; Boninger, Michael L; Koontz, Alicia M
2013-12-01
To refine the Transfer Assessment Instrument (TAI 2.0), develop a training program for the TAI, and analyze the basic psychometric properties of the TAI 3.0, including reliability, standard error of measurement (SEM), minimal detectable change (MDC), and construct validity. Repeated measures. A winter sports clinic for disabled veterans. Wheelchair users (N=41) who perform sitting-pivot or standing-pivot transfers. Not applicable. TAI version 3.0, intraclass correlation coefficients, SEMs, and MDCs for reliable measurement of raters' responses. Spearman correlation coefficient, 1-way analysis of variance, and independent t tests to evaluate construct validity. TAI 3.0 had acceptable to high levels of reliability (range, .74-.88). The SEMs for part 1, part 2, and final scores ranged from .45 to .75. The MDC was 1.5 points on the 10-point scale for the final score. There were weak correlations (ρ range, -.13 to .25; P>.11) between TAI final scores and subjects' characteristics (eg, sex, body mass index, age, type of disability, length of wheelchair use, grip and elbow strength, sitting balance). With comprehensive training, the refined TAI 3.0 yields high reliability among raters of different clinical backgrounds and experience. TAI 3.0 was unbiased toward certain physical characteristics that may influence transfer. TAI fills a void in the field by providing a quantitative measurement of transfers and a tool that can be used to detect problems and guide transfer training. Copyright © 2013 American Congress of Rehabilitation Medicine. Published by Elsevier Inc. All rights reserved.
ERIC Educational Resources Information Center
Novalis, Carol
The use of estimated income to analyze financial need of applicants to the Basic Educational Opportunity Grant (BEOG) program was investigated. Attention was focused on: how well applicants estimate their income; reasons for errors in estimation, and whether applicants supplying income tax returns supply true versions. For 1,547 eligible BEOG…
ERIC Educational Resources Information Center
Vogel, Ronald J.; And Others
The types and ranges of errors made on applications to the Basic Educational Opportunity Grant (BEOG) program were studied, along with procedures used in recovering overpayments. The objective was to assess the scope and nature of misreporting and misuse of the BEOG program. A 1975-1976 study reviewed cases referred to the U.S. Office of Education…
ERIC Educational Resources Information Center
Rees, Libby
This guide provides information on basic skills needs and programs in the workplace and issues affecting basic skills provision from a British perspective. Section 1 aims to provide a context for workplace basic skills provision. Sections 2-7 provide practical suggestions and advice on the following topics: (1) marketing; (2) contacting employers;…
Postert, Christian; Averbeck-Holocher, Marlies; Beyer, Thomas; Müller, Jörg; Furniss, Tilman
2009-03-01
DSM-IV and ICD-10 have limitations in the diagnostic classification of psychiatric disorders at preschool age (0-5 years). The publication of the Diagnostic Classification 0-3 (DC:0-3) in 1994, its basically revised second edition (DC:0-3R) in 2005 and the Research Diagnostic Criteria-Preschool Age (RDC-PA) in 2004 have provided several modifications of these manuals. Taking into account the growing empirical evidence highlighting the need for a diagnostic classification system for psychiatric disorders in preschool children, the main categorical classification systems in preschool psychiatry will be presented and discussed. The paper will focus on issues of validity, usefulness and reliability in DSM-IV, ICD-10, RDC-PA, DC:0-3, and DC:0-3R. The reasons for including or excluding postulated psychiatric disorder categories for preschool children with variable degrees of empirical evidence into the different diagnostic systems will be discussed.
Molecular beacon-enabled purification of living cells by targeting cell type-specific mRNAs.
Wile, Brian M; Ban, Kiwon; Yoon, Young-Sup; Bao, Gang
2014-10-01
Molecular beacons (MBs) are dual-labeled oligonucleotides that fluoresce only in the presence of complementary mRNA. The use of MBs to target specific mRNAs allows sorting of specific cells from a mixed cell population. In contrast to existing approaches that are limited by available surface markers or selectable metabolic characteristics, the MB-based method enables the isolation of a wide variety of cells. For example, the ability to purify specific cell types derived from pluripotent stem cells (PSCs) is important for basic research and therapeutics. In addition to providing a general protocol for MB design, validation and nucleofection into cells, we describe how to isolate a specific cell population from differentiating PSCs. By using this protocol, we have successfully isolated cardiomyocytes differentiated from mouse or human PSCs (hPSCs) with ∼ 97% purity, as confirmed by electrophysiology and immunocytochemistry. After designing MBs, their ordering and validation requires 2 weeks, and the isolation process requires 3 h.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Fu, Yao, E-mail: fu5@mailbox.sc.edu, E-mail: jhsong@cec.sc.edu; Song, Jeong-Hoon, E-mail: fu5@mailbox.sc.edu, E-mail: jhsong@cec.sc.edu
2014-08-07
Hardy stress definition has been restricted to pair potentials and embedded-atom method potentials due to the basic assumptions in the derivation of a symmetric microscopic stress tensor. Force decomposition required in the Hardy stress expression becomes obscure for multi-body potentials. In this work, we demonstrate the invariance of the Hardy stress expression for a polymer system modeled with multi-body interatomic potentials including up to four atoms interaction, by applying central force decomposition of the atomic force. The balance of momentum has been demonstrated to be valid theoretically and tested under various numerical simulation conditions. The validity of momentum conservation justifiesmore » the extension of Hardy stress expression to multi-body potential systems. Computed Hardy stress has been observed to converge to the virial stress of the system with increasing spatial averaging volume. This work provides a feasible and reliable linkage between the atomistic and continuum scales for multi-body potential systems.« less
Measuring Constructs in Family Science: How Can Item Response Theory Improve Precision and Validity?
Gordon, Rachel A.
2014-01-01
This article provides family scientists with an understanding of contemporary measurement perspectives and the ways in which item response theory (IRT) can be used to develop measures with desired evidence of precision and validity for research uses. The article offers a nontechnical introduction to some key features of IRT, including its orientation toward locating items along an underlying dimension and toward estimating precision of measurement for persons with different levels of that same construct. It also offers a didactic example of how the approach can be used to refine conceptualization and operationalization of constructs in the family sciences, using data from the National Longitudinal Survey of Youth 1979 (n = 2,732). Three basic models are considered: (a) the Rasch and (b) two-parameter logistic models for dichotomous items and (c) the Rating Scale Model for multicategory items. Throughout, the author highlights the potential for researchers to elevate measurement to a level on par with theorizing and testing about relationships among constructs. PMID:25663714
Swart, Marcel; Bickelhaupt, F Matthias
2006-03-01
We have carried out an extensive exploration of the gas-phase basicity of archetypal anionic bases across the periodic system using the generalized gradient approximation of density functional theory (DFT) at BP86/QZ4P//BP86/TZ2P. First, we validate DFT as a reliable tool for computing proton affinities and related thermochemical quantities: BP86/QZ4P//BP86/TZ2P is shown to yield a mean absolute deviation of 1.6 kcal/mol for the proton affinity at 0 K with respect to high-level ab initio benchmark data. The main purpose of this work is to provide the proton affinities (and corresponding entropies) at 298 K of the anionic conjugate bases of all main-group-element hydrides of groups 14-17 and periods 2-6. We have also studied the effect of stepwise methylation of the protophilic center of the second- and third-period bases.
Advanced online control mode selection for gas turbine aircraft engines
NASA Astrophysics Data System (ADS)
Wiseman, Matthew William
The modern gas turbine aircraft engine is a complex, highly nonlinear system the operates in a widely varying environment. Traditional engine control techniques based on the hydro mechanical control concepts of early turbojet engines are unable to deliver the performance required from today's advanced engine designs. A new type of advanced control utilizing multiple control modes and an online mode selector is investigated, and various strategies for improving the baseline mode selection architecture are introduced. The ability to five-tune actuator command outputs is added to the basic mode selection and blending process, and mode selection designs that we valid for the entire flight envelope are presented. Methods for optimizing the mode selector to improve overall engine performance are also discussed. Finally, using flight test data from a GE F110-powered F16 aircraft, the full-envelope mode selector designs are validated and shown to provide significant performance benefits. Specifically, thrust command tracking is enhanced while critical engine limits are protected, with very little impact on engine efficiency.
Nucleus-acoustic shock waves in white dwarfs
NASA Astrophysics Data System (ADS)
Jannat, S.; Mamun, A. A.
2018-04-01
The nucleus-acoustic shock waves (NASWs) propagating in a white dwarf plasma system, which contain non-relativistically or ultrarelativistically degenerate electrons, non-relativistically degenerate, viscous fluid of light nuclei, and immobile nuclei of heavy elements, have been theoretically investigated. We have used the reductive perturbation method, which is valid for small but finite-amplitude NASWs to derive the Burgers equation. The NASWs are, in fact, associated with the nucleus-acoustic (NA) waves in which the inertia is provided by the light nuclei, and restoring force is provided by the degenerate pressure of electrons. On the other hand, the stationary heavy nuclei participate only in maintaining the background charge neutrality condition at equilibrium. It is found that the viscous force acting in the fluid of light nuclei is a source of dissipation, and is responsible for the formation of NASWs. It is also observed that the basic features (polarity, amplitude, width, etc.) of the NASWs are significantly modified by the presence of heavy nuclei, and that NASWs are formed with either positive or negative potential depending on the values of the charge density of the heavy nuclei. The basic properties are also found to be significantly modified by the effects of ultrarelativistically degenerate electrons. The implications of our results in white dwarfs are briefly discussed.
Koo, Laura W.; Horowitz, Alice M.; Radice, Sarah D.; Wang, Min Q.; Kleinman, Dushanka V.
2016-01-01
Objectives We examined nurse practitioners’ use and opinions of recommended communication techniques for the promotion of oral health as part of a Maryland state-wide oral health literacy assessment. Use of recommended health-literate and patient-centered communication techniques have demonstrated improved health outcomes. Methods A 27-item self-report survey, containing 17 communication technique items, across 5 domains, was mailed to 1,410 licensed nurse practitioners (NPs) in Maryland in 2010. Use of communication techniques and opinions about their effectiveness were analyzed using descriptive statistics. General linear models explored provider and practice characteristics to predict differences in the total number and the mean number of communication techniques routinely used in a week. Results More than 80% of NPs (N = 194) routinely used 3 of the 7 basic communication techniques: simple language, limiting teaching to 2–3 concepts, and speaking slowly. More than 75% of respondents believed that 6 of the 7 basic communication techniques are effective. Sociodemographic provider characteristics and practice characteristics were not significant predictors of the mean number or the total number of communication techniques routinely used by NPs in a week. Potential predictors for using more of the 7 basic communication techniques, demonstrating significance in one general linear model each, were: assessing the office for user-friendliness and ever taking a communication course in addition to nursing school. Conclusions NPs in Maryland self-reported routinely using some recommended health-literate communication techniques, with belief in their effectiveness. Our findings suggest that NPs who had assessed the office for patient-friendliness or who had taken a communication course beyond their initial education may be predictors for using more of the 7 basic communication techniques. These self-reported findings should be validated with observational studies. Graduate and continuing education for NPs should increase emphasis on health-literate and patient-centered communication techniques to increase patient understanding of dental caries prevention. Non-dental healthcare providers, such as NPs, are uniquely positioned to contribute to preventing early childhood dental caries through health-literate and patient-centered communication. PMID:26766557
Koo, Laura W; Horowitz, Alice M; Radice, Sarah D; Wang, Min Q; Kleinman, Dushanka V
2016-01-01
We examined nurse practitioners' use and opinions of recommended communication techniques for the promotion of oral health as part of a Maryland state-wide oral health literacy assessment. Use of recommended health-literate and patient-centered communication techniques have demonstrated improved health outcomes. A 27-item self-report survey, containing 17 communication technique items, across 5 domains, was mailed to 1,410 licensed nurse practitioners (NPs) in Maryland in 2010. Use of communication techniques and opinions about their effectiveness were analyzed using descriptive statistics. General linear models explored provider and practice characteristics to predict differences in the total number and the mean number of communication techniques routinely used in a week. More than 80% of NPs (N = 194) routinely used 3 of the 7 basic communication techniques: simple language, limiting teaching to 2-3 concepts, and speaking slowly. More than 75% of respondents believed that 6 of the 7 basic communication techniques are effective. Sociodemographic provider characteristics and practice characteristics were not significant predictors of the mean number or the total number of communication techniques routinely used by NPs in a week. Potential predictors for using more of the 7 basic communication techniques, demonstrating significance in one general linear model each, were: assessing the office for user-friendliness and ever taking a communication course in addition to nursing school. NPs in Maryland self-reported routinely using some recommended health-literate communication techniques, with belief in their effectiveness. Our findings suggest that NPs who had assessed the office for patient-friendliness or who had taken a communication course beyond their initial education may be predictors for using more of the 7 basic communication techniques. These self-reported findings should be validated with observational studies. Graduate and continuing education for NPs should increase emphasis on health-literate and patient-centered communication techniques to increase patient understanding of dental caries prevention. Non-dental healthcare providers, such as NPs, are uniquely positioned to contribute to preventing early childhood dental caries through health-literate and patient-centered communication.
Using 'big data' to validate claims made in the pharmaceutical approval process.
Wasser, Thomas; Haynes, Kevin; Barron, John; Cziraky, Mark
2015-01-01
Big Data in the healthcare setting refers to the storage, assimilation, and analysis of large quantities of information regarding patient care. These data can be collected and stored in a wide variety of ways including electronic medical records collected at the patient bedside, or through medical records that are coded and passed to insurance companies for reimbursement. When these data are processed it is possible to validate claims as a part of the regulatory review process regarding the anticipated performance of medications and devices. In order to analyze properly claims by manufacturers and others, there is a need to express claims in terms that are testable in a timeframe that is useful and meaningful to formulary committees. Claims for the comparative benefits and costs, including budget impact, of products and devices need to be expressed in measurable terms, ideally in the context of submission or validation protocols. Claims should be either consistent with accessible Big Data or able to support observational studies where Big Data identifies target populations. Protocols should identify, in disaggregated terms, key variables that would lead to direct or proxy validation. Once these variables are identified, Big Data can be used to query massive quantities of data in the validation process. Research can be passive or active in nature. Passive, where the data are collected retrospectively; active where the researcher is prospectively looking for indicators of co-morbid conditions, side-effects or adverse events, testing these indicators to determine if claims are within desired ranges set forth by the manufacturer. Additionally, Big Data can be used to assess the effectiveness of therapy through health insurance records. This, for example, could indicate that disease or co-morbid conditions cease to be treated. Understanding the basic strengths and weaknesses of Big Data in the claim validation process provides a glimpse of the value that this research can provide to industry. Big Data can support a research agenda that focuses on the process of claims validation to support formulary submissions as well as inputs to ongoing disease area and therapeutic class reviews.
Schwartz, John; Macomber, Christopher
2017-01-01
Translational research for new drugs, medical devices, and diagnostics encompasses aspects of both basic science and clinical research, requiring multidisciplinary skills and resources that are not all readily available in either a basic laboratory or clinical setting alone. We propose that, to be successful, “translational” research ought to be understood as a defined process from basic science through manufacturing, regulatory, clinical testing all the way to market. The authors outline a process which has worked well for them to identify and commercialize academic innovation. The academic environment places a high value on novelty and less value on whether, among other things, data are reproducible, scalable, reimbursable, or have commercial freedom to operate. In other words, when investors, strategic companies, or other later stage stakeholders evaluate academic efforts at translational research the relative lack of attention to clinical, regulatory, reimbursement, and manufacturing and intellectual property freedom to operate almost universally results in more questions and doubts about the potential of the proposed product, thereby inhibiting further interest. This contrasts with industry-based R&D, which often emphasizes manufacturing, regulatory and commercial factors. Academics do not so much choose to ignore those necessary and standard elements of translation development, but rather, they are not built into the culture or incentive structure of the university environment. Acknowledging and addressing this mismatch of approach and lack of common language in a systematic way facilitates a more effective “translation” handoffs of academic project concepts into meaningful clinical solutions help translational researchers more efficiently develop and progress new and better medical products which address validated needs. The authors provide an overview and framework for academic researchers to use which will help them define the elements of a market-driven translational program (1) problem identification and validation; (2) defining the conceptual model of disease; and (3) risk evaluation and mitigation strategies. PMID:28952508
Statistical primer: how to deal with missing data in scientific research?
Papageorgiou, Grigorios; Grant, Stuart W; Takkenberg, Johanna J M; Mokhles, Mostafa M
2018-05-10
Missing data are a common challenge encountered in research which can compromise the results of statistical inference when not handled appropriately. This paper aims to introduce basic concepts of missing data to a non-statistical audience, list and compare some of the most popular approaches for handling missing data in practice and provide guidelines and recommendations for dealing with and reporting missing data in scientific research. Complete case analysis and single imputation are simple approaches for handling missing data and are popular in practice, however, in most cases they are not guaranteed to provide valid inferences. Multiple imputation is a robust and general alternative which is appropriate for data missing at random, surpassing the disadvantages of the simpler approaches, but should always be conducted with care. The aforementioned approaches are illustrated and compared in an example application using Cox regression.
Estimating job runtime for CMS analysis jobs
NASA Astrophysics Data System (ADS)
Sfiligoi, I.
2014-06-01
The basic premise of pilot systems is to create an overlay scheduling system on top of leased resources. And by definition, leases have a limited lifetime, so any job that is scheduled on such resources must finish before the lease is over, or it will be killed and all the computation is wasted. In order to effectively schedule jobs to resources, the pilot system thus requires the expected runtime of the users' jobs. Past studies have shown that relying on user provided estimates is not a valid strategy, so the system should try to make an estimate by itself. This paper provides a study of the historical data obtained from the Compact Muon Solenoid (CMS) experiment's Analysis Operations submission system. Clear patterns are observed, suggesting that making prediction of an expected job lifetime range is achievable with high confidence level in this environment.
Hands-Free Transcranial Color Doppler Probe
NASA Technical Reports Server (NTRS)
Chin, Robert; Madala, Srihdar; Sattler, Graham
2012-01-01
Current transcranial color Doppler (TCD) transducer probes are bulky and difficult to move in tiny increments to search and optimize TCD signals. This invention provides miniature motions of a TCD transducer probe to optimize TCD signals. The mechanical probe uses spherical bearing in guiding and locating the tilting crystal face. The lateral motion of the crystal face as it tilts across the full range of motion was achieved by minimizing the distance between the pivot location and the crystal face. The smallest commonly available metal spherical bearing was used with an outer diameter of 12 mm, a 3-mm tall retaining ring, and 5-mm overall height. Small geared motors were used that would provide sufficient power in a very compact package. After confirming the validity of the basic positioning concept, optimization design loops were completed to yield the final design.
How well do basic models describe the turbidity currents coming down Monterey and Congo Canyon?
NASA Astrophysics Data System (ADS)
Cartigny, M.; Simmons, S.; Heerema, C.; Xu, J. P.; Azpiroz, M.; Clare, M. A.; Cooper, C.; Gales, J. A.; Maier, K. L.; Parsons, D. R.; Paull, C. K.; Sumner, E. J.; Talling, P.
2017-12-01
Turbidity currents rival rivers in their global capacity to transport sediment and organic carbon. Furthermore, turbidity currents break submarine cables that now transport >95% of our global data traffic. Accurate turbidity current models are thus needed to quantify their transport capacity and to predict the forces exerted on seafloor structures. Despite this need, existing numerical models are typically only calibrated with scaled-down laboratory measurements due to the paucity of direct measurements of field-scale turbidity currents. This lack of calibration thus leaves much uncertainty in the validity of existing models. Here we use the most detailed observations of turbidity currents yet acquired to validate one of the most fundamental models proposed for turbidity currents, the modified Chézy model. Direct measurements on which the validation is based come from two sites that feature distinctly different flow modes and grain sizes. The first are from the multi-institution Coordinated Canyon Experiment (CCE) in Monterey Canyon, California. An array of six moorings along the canyon axis captured at least 15 flow events that lasted up to hours. The second is the deep-sea Congo Canyon, where 10 finer grained flows were measured by a single mooring, each lasting several days. Moorings captured depth-resolved velocity and suspended sediment concentration at high resolution (<30 second) for each of the 25 events. We use both datasets to test the most basic model available for turbidity currents; the modified Chézy model. This basic model has been very useful for river studies over the past 200 years, as it provides a rapid estimate of how flow velocity varies with changes in river level and energy slope. Chézy-type models assume that the gravitational force of the flow equals the friction of the river-bed. Modified Chézy models have been proposed for turbidity currents. However, the absence of detailed measurements of friction and sediment concentration within full-scale turbidity currents has forced modellers to make rough assumptions for these parameters. Here we use mooring data to deduce observation-based relations that can replace the previous assumptions. This improvement will significantly enhance the model predictions and allow us to better constrain the behaviour of turbidity currents.
NASA Astrophysics Data System (ADS)
Gatlin, P. N.; Conover, H.; Berendes, T.; Maskey, M.; Naeger, A. R.; Wingo, S. M.
2017-12-01
A key component of NASA's Earth observation system is its field experiments, for intensive observation of particular weather phenomena, or for ground validation of satellite observations. These experiments collect data from a wide variety of airborne and ground-based instruments, on different spatial and temporal scales, often in unique formats. The field data are often used with high volume satellite observations that have very different spatial and temporal coverage. The challenges inherent in working with such diverse datasets make it difficult for scientists to rapidly collect and analyze the data for physical process studies and validation of satellite algorithms. The newly-funded VISAGE project will address these issues by combining and extending nascent efforts to provide on-line data fusion, exploration, analysis and delivery capabilities. A key building block is the Field Campaign Explorer (FCX), which allows users to examine data collected during field campaigns and simplifies data acquisition for event-based research. VISAGE will extend FCX's capabilities beyond interactive visualization and exploration of coincident datasets, to provide interrogation of data values and basic analyses such as ratios and differences between data fields. The project will also incorporate new, higher level fused and aggregated analysis products from the System for Integrating Multi-platform data to Build the Atmospheric column (SIMBA), which combines satellite and ground-based observations into a common gridded atmospheric column data product; and the Validation Network (VN), which compiles a nationwide database of coincident ground- and satellite-based radar measurements of precipitation for larger scale scientific analysis. The VISAGE proof-of-concept will target "golden cases" from Global Precipitation Measurement Ground Validation campaigns. This presentation will introduce the VISAGE project, initial accomplishments and near term plans.
Validation of SMAP data using Cosmic-ray Neutron Probes during the SMAPVEX16-IA Campaign
NASA Astrophysics Data System (ADS)
Russell, M. V.
2016-12-01
Global trends in consumptive water-use indicate a growing and unsustainable reliance on water resources. Each year it is estimated that 60 percent of water used for agriculture is wasted through inadequate water conservation, losses in distribution, and inappropriate times and rates of irrigation. Satellite remote sensing offers a variety of water balance datasets (precipitation, evapotranspiration, soil moisture, groundwater storage) to increase the water use efficiency in agricultural systems. In this work, we aim to validate the Soil Moisture Active Passive (SMAP) soil moisture product using the ground based cosmic-ray neutron probe (CRNP) for estimating field scale soil moisture at intermediate spatial scales as part of SMAPVEX16-IA experiment. Typical SMAP calibration and validation has been done using a combination of direct gravimetric sampling and in-situ soil moisture point observations. Although these measurements provide accurate data, it is time consuming and labor intensive to collect data over a 36 by 36 km SMAP pixel. Through a joint effort with rovers provided by the US Army Corps of Engineers and University of Nebraska-Lincoln, we are able to cover the domain in 7 hours. Data from both rovers was combined in order to produce a 1, 3, 9 and 36 km resolution product on the day of 12 SMAP overpasses in May and August 2016. Here we will describe basic QAQC procedures for estimating soil moisture from the dual rover experiment. This will include discussion about calibration, validation, and accounting for conditions such as variable road type and growing vegetation. Lastly, we will compare the calibrated rover and SMAP products. If the products are highly correlated the ground based rovers offer a strategy for collecting finer resolution products that may be used in future downscaling efforts in support of high resolution Land Surface Modeling.
Physical Behavior in Older Persons during Daily Life: Insights from Instrumented Shoes.
Moufawad El Achkar, Christopher; Lenoble-Hoskovec, Constanze; Paraschiv-Ionescu, Anisoara; Major, Kristof; Büla, Christophe; Aminian, Kamiar
2016-08-03
Activity level and gait parameters during daily life are important indicators for clinicians because they can provide critical insights into modifications of mobility and function over time. Wearable activity monitoring has been gaining momentum in daily life health assessment. Consequently, this study seeks to validate an algorithm for the classification of daily life activities and to provide a detailed gait analysis in older adults. A system consisting of an inertial sensor combined with a pressure sensing insole has been developed. Using an algorithm that we previously validated during a semi structured protocol, activities in 10 healthy elderly participants were recorded and compared to a wearable reference system over a 4 h recording period at home. Detailed gait parameters were calculated from inertial sensors. Dynamics of physical behavior were characterized using barcodes that express the measure of behavioral complexity. Activity classification based on the algorithm led to a 93% accuracy in classifying basic activities of daily life, i.e., sitting, standing, and walking. Gait analysis emphasizes the importance of metrics such as foot clearance in daily life assessment. Results also underline that measures of physical behavior and gait performance are complementary, especially since gait parameters were not correlated to complexity. Participants gave positive feedback regarding the use of the instrumented shoes. These results extend previous observations in showing the concurrent validity of the instrumented shoes compared to a body-worn reference system for daily-life physical behavior monitoring in older adults.
Benedict, Ralph HB; DeLuca, John; Phillips, Glenn; LaRocca, Nicholas; Hudson, Lynn D; Rudick, Richard
2017-01-01
Cognitive and motor performance measures are commonly employed in multiple sclerosis (MS) research, particularly when the purpose is to determine the efficacy of treatment. The increasing focus of new therapies on slowing progression or reversing neurological disability makes the utilization of sensitive, reproducible, and valid measures essential. Processing speed is a basic elemental cognitive function that likely influences downstream processes such as memory. The Multiple Sclerosis Outcome Assessments Consortium (MSOAC) includes representatives from advocacy organizations, Food and Drug Administration (FDA), European Medicines Agency (EMA), National Institute of Neurological Disorders and Stroke (NINDS), academic institutions, and industry partners along with persons living with MS. Among the MSOAC goals is acceptance and qualification by regulators of performance outcomes that are highly reliable and valid, practical, cost-effective, and meaningful to persons with MS. A critical step for these neuroperformance metrics is elucidation of clinically relevant benchmarks, well-defined degrees of disability, and gradients of change that are deemed clinically meaningful. This topical review provides an overview of research on one particular cognitive measure, the Symbol Digit Modalities Test (SDMT), recognized as being particularly sensitive to slowed processing of information that is commonly seen in MS. The research in MS clearly supports the reliability and validity of this test and recently has supported a responder definition of SDMT change approximating 4 points or 10% in magnitude. PMID:28206827
Benedict, Ralph Hb; DeLuca, John; Phillips, Glenn; LaRocca, Nicholas; Hudson, Lynn D; Rudick, Richard
2017-04-01
Cognitive and motor performance measures are commonly employed in multiple sclerosis (MS) research, particularly when the purpose is to determine the efficacy of treatment. The increasing focus of new therapies on slowing progression or reversing neurological disability makes the utilization of sensitive, reproducible, and valid measures essential. Processing speed is a basic elemental cognitive function that likely influences downstream processes such as memory. The Multiple Sclerosis Outcome Assessments Consortium (MSOAC) includes representatives from advocacy organizations, Food and Drug Administration (FDA), European Medicines Agency (EMA), National Institute of Neurological Disorders and Stroke (NINDS), academic institutions, and industry partners along with persons living with MS. Among the MSOAC goals is acceptance and qualification by regulators of performance outcomes that are highly reliable and valid, practical, cost-effective, and meaningful to persons with MS. A critical step for these neuroperformance metrics is elucidation of clinically relevant benchmarks, well-defined degrees of disability, and gradients of change that are deemed clinically meaningful. This topical review provides an overview of research on one particular cognitive measure, the Symbol Digit Modalities Test (SDMT), recognized as being particularly sensitive to slowed processing of information that is commonly seen in MS. The research in MS clearly supports the reliability and validity of this test and recently has supported a responder definition of SDMT change approximating 4 points or 10% in magnitude.
Structured learning for robotic surgery utilizing a proficiency score: a pilot study.
Hung, Andrew J; Bottyan, Thomas; Clifford, Thomas G; Serang, Sarfaraz; Nakhoda, Zein K; Shah, Swar H; Yokoi, Hana; Aron, Monish; Gill, Inderbir S
2017-01-01
We evaluated feasibility and benefit of implementing structured learning in a robotics program. Furthermore, we assessed validity of a proficiency assessment tool for stepwise graduation. Teaching cases included robotic radical prostatectomy and partial nephrectomy. Procedure steps were categorized: basic, intermediate, and advanced. An assessment tool ["proficiency score" (PS)] was developed to evaluate ability to safely and autonomously complete a step. Graduation required a passing PS (PS ≥ 3) on three consecutive attempts. PS and validated global evaluative assessment of robotic skills (GEARS) were evaluated for completed steps. Linear regression was utilized to determine postgraduate year/PS relationship (construct validity). Spearman's rank correlation coefficient measured correlation between PS and GEARS evaluations (concurrent validity). Intraclass correlation (ICC) evaluated PS agreement between evaluator classes. Twenty-one robotic trainees participated within the pilot program, completing a median of 14 (2-69) cases each. Twenty-three study evaluators scored 14 (1-60) cases. Over 4 months, 229/294 (78 %) cases were designated "teaching" cases. Residents completed 91 % of possible evaluations; faculty completed 78 %. Verbal and quantitative feedback received by trainees increased significantly (p = 0.002, p < 0.001, respectively). Average PS increased with PGY (post-graduate year) for basic and intermediate steps (regression slopes: 0.402 (p < 0.0001), 0.323 (p < 0.0001), respectively) (construct validation). Overall, PS correlated highly with GEARS (ρ = 0.81, p < 0.0001) (concurrent validity). ICC was 0.77 (95 % CI 0.61-0.88) for resident evaluations. Structured learning can be implemented in an academic robotic program with high levels of trainee and evaluator participation, encouraging both quantitative and verbal feedback. A proficiency assessment tool developed for step-specific proficiency has construct and concurrent validity.
NASA Astrophysics Data System (ADS)
Yusmaita, E.; Nasra, Edi
2018-04-01
This research aims to produce instrument for measuring chemical literacy assessment in basic chemistry courses with solubility topic. The construction of this measuring instrument is adapted to the PISA (Programme for International Student Assessment) problem’s characteristics and the Syllaby of Basic Chemistry in KKNI-IndonesianNational Qualification Framework. The PISA is a cross-country study conducted periodically to monitor the outcomes of learners' achievement in each participating country. So far, studies conducted by PISA include reading literacy, mathematic literacy and scientific literacy. Refered to the scientific competence of the PISA study on science literacy, an assessment designed to measure the chemical literacy of the chemistry department’s students in UNP. The research model used is MER (Model of Educational Reconstruction). The validity and reliability values of discourse questions is measured using the software ANATES. Based on the acquisition of these values is obtained a valid and reliable chemical literacy questions.There are seven question items limited response on the topic of solubility with valid category, the acquisition value of test reliability is 0,86, and has a difficulty index and distinguishing good
ERIC Educational Resources Information Center
Tang, Yang; Cook, Thomas D.; Kisbu-Sakarya, Yasemin
2015-01-01
Regression discontinuity design (RD) has been widely used to produce reliable causal estimates. Researchers have validated the accuracy of RD design using within study comparisons (Cook, Shadish & Wong, 2008; Cook & Steiner, 2010; Shadish et al, 2011). Within study comparisons examines the validity of a quasi-experiment by comparing its…
Blinov, N N
2000-01-01
Specifications for the main element of a modern X-ray diagnostic device an X-ray feeder are formulated. There is evidence for choosing its parameters. The new rational routine of X-ray study and the layout of a X-ray room are proposed. Information on the up-to-date commercially manufactured basic medium-frequency general-purpose X-ray feeder "URP-30 SCh Amico" is given.
Basic research on radiant burners. Semi-annual report, through July 1991
DOE Office of Scientific and Technical Information (OSTI.GOV)
Sullivan, J.D.
1991-10-01
Basic performance characteristics of radiant burners are explored in the broad-based study combining theoretical modeling and experimental validation of predictions. The work included fabrication of catalyzed substrates and fibers; incorporation of the catalysts into burners; testing of catalysts; and investigation of new catalyst sources. The progress of the study is detailed and further plans are outlined. A report on the preparation of palladium catalysts by Andre Blaise Kooh is included in the appendix.
Tannure, Meire Chucre; Salgado, Patrícia de Oliveira; Chianca, Tânia Couto Machado
2014-01-01
This descriptive study aimed at elaborating nursing diagnostic labels according to ICNP®; conducting a cross-mapping between the diagnostic formulations and the diagnostic labels of NANDA-I; identifying the diagnostic labels thus obtained that were also listed in the NANDA-I; and mapping them according to Basic Human Needs. The workshop technique was applied to 32 intensive care nurses, the cross-mapping and validation based on agreement with experts. The workshop produced 1665 diagnostic labels which were further refined into 120 labels. They were then submitted to a cross-mapping process with both NANDA-I diagnostic labels and the Basic Human Needs. The mapping results underwent content validation by two expert nurses leading to concordance rates of 92% and 100%. It was found that 63 labels were listed in NANDA-I and 47 were not.
Van Iddekinge, Chad H; Roth, Philip L; Putka, Dan J; Lanivich, Stephen E
2011-11-01
A common belief among researchers is that vocational interests have limited value for personnel selection. However, no comprehensive quantitative summaries of interests validity research have been conducted to substantiate claims for or against the use of interests. To help address this gap, we conducted a meta-analysis of relations between interests and employee performance and turnover using data from 74 studies and 141 independent samples. Overall validity estimates (corrected for measurement error in the criterion but not for range restriction) for single interest scales were .14 for job performance, .26 for training performance, -.19 for turnover intentions, and -.15 for actual turnover. Several factors appeared to moderate interest-criterion relations. For example, validity estimates were larger when interests were theoretically relevant to the work performed in the target job. The type of interest scale also moderated validity, such that corrected validities were larger for scales designed to assess interests relevant to a particular job or vocation (e.g., .23 for job performance) than for scales designed to assess a single, job-relevant realistic, investigative, artistic, social, enterprising, or conventional (i.e., RIASEC) interest (.10) or a basic interest (.11). Finally, validity estimates were largest when studies used multiple interests for prediction, either by using a single job or vocation focused scale (which tend to tap multiple interests) or by using a regression-weighted composite of several RIASEC or basic interest scales. Overall, the results suggest that vocational interests may hold more promise for predicting employee performance and turnover than researchers may have thought. (c) 2011 APA, all rights reserved.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Not Available
This report was prepared at the request of the Lawrence Livermore Laboratory (LLL) to provide background information for analyzing soil-structure interaction by the frequency-independent impedance function approach. LLL is conducting such analyses as part of its seismic review of selected operating plants under the Systematic Evaluation Program for the US Nuclear Regulatory Commission. The analytical background and basic assumptionsof the impedance function theory are briefly reviewed, and the role of radiation damping in soil-structure interaction analysis is discussed. The validity of modeling soil-structure interaction by using frequency-independent functions is evaluated based on data from several field tests. Finally, the recommendedmore » procedures for performing soil-structure interaction analyses are discussed with emphasis on the modal superposition method.« less
Performance Investigation of Millimeter Wave Generation Reliant on Stimulated Brillouin Scattering
NASA Astrophysics Data System (ADS)
Tickoo, Sheetal; Gupta, Amit
2018-04-01
In this work, photonic method of generating the millimeter waves has been done based on Brillouin scattering effect in optical fiber. Here different approaches are proposed to get maximum frequency shift in mm-wave region using only pumps, radio signals with Mach-Zehnder modulator. Moreover for generated signal validation, signals modulated and send to both wired and wireless medium in optical domain. It is observed that maximum shift of 300 GHz is realized using 60 GHz input sine wave. Basically a frequency doubler is proposed which double shift of input frequency and provide better SNR. For the future generation network system, the generation of millimeter waves makes them well reliable for the transmission of the data.
A psychoanalyst-liaison psychiatrist's overview of DSM III.
Grossman, S
1982-12-01
There has been a groundswell of reaction, mostly favorable, to the most recent edition of the Diagnostic and Statistical Manual. In this paper the author attempts to evaluate the Manual, which purports to provide an atheoretical and descriptive diagnostic model on its own ground, as well as from a psychodynamic and ego-psychological point of view. More specifically, the Manual is evaluated in terms of its usefulness in the diagnosis and management of the Somatoform Disorders, as well as other typical general hospital problems. This assessment raises many questions about the basic tenets of DSM III in general. Concomitantly, suggestions are made to enhance the reliability, validity, and clinical relevance of the Manual.
NASA Technical Reports Server (NTRS)
Klopfer, Goetz H.
1993-01-01
The work performed during the past year on this cooperative agreement covered two major areas and two lesser ones. The two major items included further development and validation of the Compressible Navier-Stokes Finite Volume (CNSFV) code and providing computational support for the Laminar Flow Supersonic Wind Tunnel (LFSWT). The two lesser items involve a Navier-Stokes simulation of an oscillating control surface at transonic speeds and improving the basic algorithm used in the CNSFV code for faster convergence rates and more robustness. The work done in all four areas is in support of the High Speed Research Program at NASA Ames Research Center.
NASA Astrophysics Data System (ADS)
Schmitt, Rainer M.; Scott, W. Guy; Irving, Richard D.; Arnold, Joe; Bardons, Charles; Halpert, Daniel; Parker, Lawrence
2004-09-01
A new type of fingerprint sensor is presented. The sensor maps the acoustic impedance of the fingerprint pattern by estimating the electrical impedance of its sensor elements. The sensor substrate, made of 1-3 piezo-ceramic, which is fabricated inexpensively at large scales, can provide a resolution up to 50 μm over an area of 20 x 25 mm2. Using FE modeling the paper presents the numerical validation of the basic principle. It evaluates an optimized pillar aspect ratio, estimates spatial resolution and the point spread function for a 100 μm and 50 μm pitch model. In addition, first fingerprints obtained with the prototype sensor are presented.
Using Space Weather for Enhanced, Extreme Terrestrial Weather Predictions.
NASA Astrophysics Data System (ADS)
McKenna, M. H.; Lee, T. A., III
2017-12-01
Considering the complexities of the Sun-Earth system, the impacts of space weather to weather here on Earth are not fully understood. This study attempts to analyze this interrelationship by providing a theoretical framework for studying the varied modalities of solar inclination and explores the extent to which they contribute, both in formation and intensity, to extreme terrestrial weather. Using basic topologic and ontology engineering concepts (TOEC), the transdisciplinary syntaxes of space physics, geophysics, and meteorology are analyzed as a seamless interrelated system. This paper reports this investigation's initial findings and examines the validity of the question "Does space weather contribute to extreme weather on Earth, and if so, to what degree?"
Vector 33: A reduce program for vector algebra and calculus in orthogonal curvilinear coordinates
NASA Astrophysics Data System (ADS)
Harper, David
1989-06-01
This paper describes a package with enables REDUCE 3.3 to perform algebra and calculus operations upon vectors. Basic algebraic operations between vectors and between scalars and vectors are provided, including scalar (dot) product and vector (cross) product. The vector differential operators curl, divergence, gradient and Laplacian are also defined, and are valid in any orthogonal curvilinear coordinate system. The package is written in RLISP to allow algebra and calculus to be performed using notation identical to that for operations. Scalars and vectors can be mixed quite freely in the same expression. The package will be of interest to mathematicians, engineers and scientists who need to perform vector calculations in orthogonal curvilinear coordinates.
NASA Astrophysics Data System (ADS)
Salon, Stefano; Cossarini, Gianpiero; Bolzon, Giorgio; Teruzzi, Anna
2017-04-01
The Mediterranean Monitoring and Forecasting Centre (Med-MFC) is one of the regional production centres of the EU Copernicus Marine Environment Monitoring Service (CMEMS). Med-MFC manages a suite of numerical model systems for the operational delivery of the CMEMS products, providing continuous monitoring and forecasting of the Mediterranean marine environment. The CMEMS products of fundamental biogeochemical variables (chlorophyll, nitrate, phosphate, oxygen, phytoplankton biomass, primary productivity, pH, pCO2) are organised as gridded datasets and are available at the marine.copernicus.eu web portal. Quantitative estimates of CMEMS products accuracy are prerequisites to release reliable information to intermediate users, end users and to other downstream services. In particular, validation activities aim to deliver accuracy information of the model products and to serve as a long term monitoring of the performance of the modelling systems. The quality assessment of model output is implemented using a multiple-stages approach, basically inspired to the classic "GODAE 4 Classes" metrics and criteria (consistency, quality, performance and benefit). Firstly, pre-operational runs qualify the operational model system against historical data, also providing a verification of the improvements of the new model system release with respect to the previous version. Then, the near real time (NRT) validation aims at delivering a sustained on-line skill assessment of the model analysis and forecast, relying on the NRT available relevant observations (e.g. in situ, Bio Argo and satellite observations). NRT validation results are operated on weekly basis and published on the MEDEAF web portal (www.medeaf.inogs.it). On a quarterly basis, the integration of the NRT validation activities delivers a comprehensive view of the accuracy of model forecast through the official CMEMS validation webpage. Multi-Year production (e.g. reanalysis runs) follows a similar procedure, and the validation is achieved using the same metrics on available historical observations (e.g. the World Ocean Atlas 2013 dataset). Results of the validation activities show that the comparison of the different variables of the CMEMS products with experimental data is feasible at different levels (i.e. either as skill assessment of the short-term forecast and as model consistency through different system versions) and at different spatial and temporal scales. In particular, the accuracy of some variables (chlorophyll, nitrate, oxygen) can be provided at weekly scale and sub-mesoscale, others (carbonate system, phosphate) at quarterly/annual and sub-basin scale, and others (phytoplankton biomass, primary production) only at the level of consistency of model functioning (e.g. literature- or climatology-based). In spite of a wide literature on model validation has been produced so far, maintaining a validation framework in the biogeochemical operational contest that fulfils GODAE criteria is still a challenge. Recent results of the validation activities and new potential validation framework at the Med-MFC will be presented in our contribution.
Griesinger, Claudius; Desprez, Bertrand; Coecke, Sandra; Casey, Warren; Zuang, Valérie
This chapter explores the concepts, processes, tools and challenges relating to the validation of alternative methods for toxicity and safety testing. In general terms, validation is the process of assessing the appropriateness and usefulness of a tool for its intended purpose. Validation is routinely used in various contexts in science, technology, the manufacturing and services sectors. It serves to assess the fitness-for-purpose of devices, systems, software up to entire methodologies. In the area of toxicity testing, validation plays an indispensable role: "alternative approaches" are increasingly replacing animal models as predictive tools and it needs to be demonstrated that these novel methods are fit for purpose. Alternative approaches include in vitro test methods, non-testing approaches such as predictive computer models up to entire testing and assessment strategies composed of method suites, data sources and decision-aiding tools. Data generated with alternative approaches are ultimately used for decision-making on public health and the protection of the environment. It is therefore essential that the underlying methods and methodologies are thoroughly characterised, assessed and transparently documented through validation studies involving impartial actors. Importantly, validation serves as a filter to ensure that only test methods able to produce data that help to address legislative requirements (e.g. EU's REACH legislation) are accepted as official testing tools and, owing to the globalisation of markets, recognised on international level (e.g. through inclusion in OECD test guidelines). Since validation creates a credible and transparent evidence base on test methods, it provides a quality stamp, supporting companies developing and marketing alternative methods and creating considerable business opportunities. Validation of alternative methods is conducted through scientific studies assessing two key hypotheses, reliability and relevance of the test method for a given purpose. Relevance encapsulates the scientific basis of the test method, its capacity to predict adverse effects in the "target system" (i.e. human health or the environment) as well as its applicability for the intended purpose. In this chapter we focus on the validation of non-animal in vitro alternative testing methods and review the concepts, challenges, processes and tools fundamental to the validation of in vitro methods intended for hazard testing of chemicals. We explore major challenges and peculiarities of validation in this area. Based on the notion that validation per se is a scientific endeavour that needs to adhere to key scientific principles, namely objectivity and appropriate choice of methodology, we examine basic aspects of study design and management, and provide illustrations of statistical approaches to describe predictive performance of validated test methods as well as their reliability.
NASA Astrophysics Data System (ADS)
Karakoti, Indira; Kesarwani, Kapil; Mehta, Manish; Dobhal, D. P.
2016-10-01
Two enhanced temperature-index (T-index) models are proposed by incorporating meteorological parameters viz. relative humidity, wind speed and net radiation. The models are an attempt to explore different climatic variables other than temperature affecting glacier surface melting. Weather data were recorded at Chorabari Glacier using an automatic weather station during the summers of 2010 (July 10 to September 10) and 2012 (June 10 to October 25). The modelled surface melt is validated against the measured point surface melting at the snout. Performance of the developed models is evaluated by comparing with basic temperature-index model and is quantified through different efficiency criteria. The results suggest that proposed models yield considerable improvement in surface melt simulation . Consequently, the study reveals that glacier surface melt depends not only on temperature but also on weather parameters viz. relative humidity, wind speed and net radiation play a significant role in glacier surface melting. This approach provides a major improvement on basic temperature-index method and offers an alternative to energy balance model.
Basic Brackets of a 2D Model for the Hodge Theory Without its Canonical Conjugate Momenta
NASA Astrophysics Data System (ADS)
Kumar, R.; Gupta, S.; Malik, R. P.
2016-06-01
We deduce the canonical brackets for a two (1+1)-dimensional (2D) free Abelian 1-form gauge theory by exploiting the beauty and strength of the continuous symmetries of a Becchi-Rouet-Stora-Tyutin (BRST) invariant Lagrangian density that respects, in totality, six continuous symmetries. These symmetries entail upon this model to become a field theoretic example of Hodge theory. Taken together, these symmetries enforce the existence of exactly the same canonical brackets amongst the creation and annihilation operators that are found to exist within the standard canonical quantization scheme. These creation and annihilation operators appear in the normal mode expansion of the basic fields of this theory. In other words, we provide an alternative to the canonical method of quantization for our present model of Hodge theory where the continuous internal symmetries play a decisive role. We conjecture that our method of quantization is valid for a class of field theories that are tractable physical examples for the Hodge theory. This statement is true in any arbitrary dimension of spacetime.
Real-time PCR (qPCR) primer design using free online software.
Thornton, Brenda; Basu, Chhandak
2011-01-01
Real-time PCR (quantitative PCR or qPCR) has become the preferred method for validating results obtained from assays which measure gene expression profiles. The process uses reverse transcription polymerase chain reaction (RT-PCR), coupled with fluorescent chemistry, to measure variations in transcriptome levels between samples. The four most commonly used fluorescent chemistries are SYBR® Green dyes and TaqMan®, Molecular Beacon or Scorpion probes. SYBR® Green is very simple to use and cost efficient. As SYBR® Green dye binds to any double-stranded DNA product, its success depends greatly on proper primer design. Many types of online primer design software are available, which can be used free of charge to design desirable SYBR® Green-based qPCR primers. This laboratory exercise is intended for those who have a fundamental background in PCR. It addresses the basic fluorescent chemistries of real-time PCR, the basic rules and pitfalls of primer design, and provides a step-by-step protocol for designing SYBR® Green-based primers with free, online software. Copyright © 2010 Wiley Periodicals, Inc.
Onan, Arif; Simsek, Nurettin
2017-04-01
Automated external defibrillators are pervasive computing devices designed for the treatment and management of acute sudden cardiac arrest. This study aims to explain users' actual use behavior in teams formed by different professions taken after a short time span of interaction with automated external defibrillator. Before the intervention, all the participants were certified with the American Heart Association Basic Life Support for healthcare providers. A statistically significant difference was revealed in mean individual automated external defibrillator technical skills between uniprofessional and interprofessional groups. The technical automated external defibrillator team scores were greater for groups with interprofessional than for those with uniprofessional education. The nontechnical automated external defibrillator skills of interprofessional and uniprofessional teams revealed differences in advantage of interprofessional teams. Students positively accept automated external defibrillators if well-defined and validated training opportunities to use them expertly are available. Uniprofessional teams were successfully supported by their members and, thereby, used automated external defibrillator effectively. Furthermore, the interprofessional approach resulted in as much effective teamwork as the uniprofessional approach.
A basic plasma test for gyrokinetics: GDC turbulence in LAPD
NASA Astrophysics Data System (ADS)
Pueschel, M. J.; Rossi, G.; Told, D.; Terry, P. W.; Jenko, F.; Carter, T. A.
2017-02-01
Providing an important step towards validating gyrokinetics under comparatively little-explored conditions, simulations of pressure-gradient-driven plasma turbulence in the Large Plasma Device (LAPD) are compared with experimental observations. The corresponding signatures confirm the existence of a novel regime of turbulence, based on the recently-discovered gradient-driven drift coupling (GDC) instability, which is thus confirmed as a candidate mechanism for turbulence in basic, space and astrophysical plasmas. Despite the limitations of flux-tube gyrokinetics for this scenario, when accounting for box size scaling by applying a scalar factor η =6, agreement between simulations and experiment improves to within a factor of two for key observables: compressional magnetic, density, and temperature fluctuations, both in amplitude and structure. Thus, a first, strong indication is presented that the GDC instability seen in gyrokinetics appears to operate in the experiment and that the essential instability physics is present in the numerical model. Overall, the gyrokinetic framework and its numerical implementation in the Gene code therefore perform well for LAPD plasmas very different from their brethren in fusion experiments.
A Mobile Sensor Network System for Monitoring of Unfriendly Environments.
Song, Guangming; Zhou, Yaoxin; Ding, Fei; Song, Aiguo
2008-11-14
Observing microclimate changes is one of the most popular applications of wireless sensor networks. However, some target environments are often too dangerous or inaccessible to humans or large robots and there are many challenges for deploying and maintaining wireless sensor networks in those unfriendly environments. This paper presents a mobile sensor network system for solving this problem. The system architecture, the mobile node design, the basic behaviors and advanced network capabilities have been investigated respectively. A wheel-based robotic node architecture is proposed here that can add controlled mobility to wireless sensor networks. A testbed including some prototype nodes has also been created for validating the basic functions of the proposed mobile sensor network system. Motion performance tests have been done to get the positioning errors and power consumption model of the mobile nodes. Results of the autonomous deployment experiment show that the mobile nodes can be distributed evenly into the previously unknown environments. It provides powerful support for network deployment and maintenance and can ensure that the sensor network will work properly in unfriendly environments.
Near Zone: Basic scattering code user's manual with space station applications
NASA Technical Reports Server (NTRS)
Marhefka, R. J.; Silvestro, J. W.
1989-01-01
The Electromagnetic Code - Basic Scattering Code, Version 3, is a user oriented computer code to analyze near and far zone patterns of antennas in the presence of scattering structures, to provide coupling between antennas in a complex environment, and to determine radiation hazard calculations at UHF and above. The analysis is based on uniform asymptotic techniques formulated in terms of the Uniform Geometrical Theory of Diffraction (UTD). Complicated structures can be simulated by arbitrarily oriented flat plates and an infinite ground plane that can be perfectly conducting or dielectric. Also, perfectly conducting finite elliptic cylinder, elliptic cone frustum sections, and finite composite ellipsoids can be used to model the superstructure of a ship, the body of a truck, and airplane, a satellite, etc. This manual gives special consideration to space station modeling applications. This is a user manual designed to give an overall view of the operation of the computer code, to instruct a user in how to model structures, and to show the validity of the code by comparing various computed results against measured and alternative calculations such as method of moments whenever available.
Torlakovic, Emina E; Cheung, Carol C; D'Arrigo, Corrado; Dietel, Manfred; Francis, Glenn D; Gilks, C Blake; Hall, Jacqueline A; Hornick, Jason L; Ibrahim, Merdol; Marchetti, Antonio; Miller, Keith; van Krieken, J Han; Nielsen, Soren; Swanson, Paul E; Vyberg, Mogens; Zhou, Xiaoge; Taylor, Clive R
2017-03-01
Validation of immunohistochemistry (IHC) assays is a subject that is of great importance to clinical practice as well as basic research and clinical trials. When applied to clinical practice and focused on patient safety, validation of IHC assays creates objective evidence that IHC assays used for patient care are "fit-for-purpose." Validation of IHC assays needs to be properly informed by and modeled to assess the purpose of the IHC assay, which will further determine what sphere of validation is required, as well as the scope, type, and tier of technical validation. These concepts will be defined in this review, part 3 of the 4-part series "Evolution of Quality Assurance for Clinical Immunohistochemistry in the Era of Precision Medicine."
Camilleri, Michael; Chang, Lin
2008-01-01
Recent advances in our understanding of basic neuroenteric mechanisms and the role of effectors and transmitters in the brain-gut axis have provided opportunities to develop new therapeutic agents for irritable bowel syndrome (IBS). Furthermore, human pharmacodynamic studies utilizing transit, colonic or rectal sensitivity, and brain imaging have been useful in determining therapeutic efficacy (particularly for drugs that act on motor function). This review provides an overview of medications that have not yet been approved for treatment of patients with IBS, yet have shown promise in phase IIB trials. These include drugs that act on the serotonin receptor and transporter system, antidepressants, norepinephrine reuptake inhibitors, opioids, cholecystokinin antagonists, neurokinin-antagonists, chloride channel activators, guanylate cyclase C agonists, atypical benzodiazepines, probiotics and antibiotics. The changing landscape in the regulatory approval process has impacted the development of IBS drugs. Guidance documents from regulatory agencies in Europe and the United States have focused on patients’ reported outcomes and associated quality of life. After a decade of experience with different endpoints that have generated some data on psychometric validation and unprecedented information about responsiveness of the binary or global endpoints to drug therapy, it is necessary to pursue further validation studies before or during pivotal phase IIB or III trials. The hope of providing relief to patients should galvanize all parties to achieve these goals. PMID:18848833
GGFC Special Bureau for Loading: current status and plans
NASA Astrophysics Data System (ADS)
van Dam, T.; Plag, H.-P.; Francis, O.; Gegout, P.
The Earth's surface is perpetually being displaced due to temporally varying atmospheric, oceanic and continental water mass surface loads. These non-geodynamic signals are of substantial magnitude that they contribute significantly to the scatter in geodetic observations of crustal motion. In February, 2002, the International Earth Rotation Service (IERS) established a Special Bureau of Loading (SBL) whose primary charge is to provide consistent and valid estimates of surface mass loading effects to the IERS community for the purpose of correcting geodetic time series. Here we outline the primary principles involved in modelling the surface displacements and gravity changes induced by surface mass loading including the basic theory, the Earth model and the surface load data. We then identify a list of operational issues, including product validation, that need to be addressed by the SBL before products can be provided to the community. Finally, we outline areas for future research to further improve the loading estimates. We conclude by formulating a recommendation on the best procedure for including loading corrections into geodetic data. Success of the SBL will depend on our ability to efficiently provide consistent and reliable estimates of surface mass loading effects. It is imperative that we work closely with the existing Global Geophysical Fluids Center (GGFC) Special Bureaus and with the community to as much as possible to verify the products.
Feasibility study for early removal of HEU from CPP-651-Phase II
DOE Office of Scientific and Technical Information (OSTI.GOV)
Smith, C.V.; Henry, R.; Milligan, C.
1997-09-01
A two-phase feasibility study was initiated in late 1996 to identify a way to expedite the removal of SNM from the CPP-651 vault. The first phase of this study provided preliminary information that appeared promising, but needed additional detailed planning and evaluate to validate the concepts and conclusions. The focus of Phase 2 was to provide the validation via resource-loaded schedules and more detailed cost estimates. Section 1 describes the purpose and objectives of the Phase 2 tasks and the programmatic drivers that influence related CPP-651 high-enriched uranium (HEU) management issues. Section 2 identifies the evaluation criteria and methodology andmore » the transfer issues and barriers preventing shipment. Section 3 provides site-specific background information for the CPP-651 facility and the Idaho National Engineering and Environmental Laboratory (INEEL) and describes the development of the basic material removal schedule, the proposed base case plan for removal of SNM, and the proposed HEU material management/shipping issues and strategies. Section 4 identifies the proposed options for accelerated removal of SNM and how they were evaluated via detailed scheduling, resource histograms, and cost analysis. Section 5 summarizes principal tasks for implementing this plan and other related HEU CPP-651 management issues that require continued planning efforts to assure successful implementation of this proposed early removal strategy.« less
Auditory fitness for duty: a review.
Tufts, Jennifer B; Vasil, Kristin A; Briggs, Sarah
2009-10-01
Auditory fitness for duty (AFFD) refers to the possession of hearing abilities sufficient for safe and effective job performance. In jobs such as law enforcement and piloting, where the ability to hear is critical to job performance and safety, hearing loss can decrease performance, even to the point of being hazardous to self and others. Tests of AFFD should provide an employer with a valid assessment of an employee's ability to perform the job safely, without discriminating against the employee purely on the basis of hearing loss. The purpose of this review is to provide a basic description of the functional hearing abilities required in hearing-critical occupations, and a summary of current practices in AFFD evaluation. In addition, we suggest directions for research and standardization to ensure best practices in the evaluation of AFFD in the future. We conducted a systematic review of the English-language peer-reviewed literature in AFFD. "Popular" search engines were consulted for governmental regulations and trade journal articles. We also contacted professionals with expertise in AFFD regarding research projects, unpublished material, and current standards. The literature review provided information regarding the functional hearing abilities required to perform hearing-critical tasks, the development of and characteristics of AFFD protocols, and the current implementation of AFFD protocols. This review paper provides evidence of the need to institute job-specific AFFD protocols, move beyond the pure-tone audiogram, and establish the validity of test protocols. These needs are arguably greater now than in times past.
77 FR 22707 - Electronic Reporting Under the Toxic Substances Control Act
Federal Register 2010, 2011, 2012, 2013, 2014
2012-04-17
... completes metadata information, the web-based tool validates the submission by performing a basic error... uploading PDF attachments or other file types, such as XML, and completing metadata information would be...
2014-01-01
Background Questionnaire-based studies suggest atypical sensory perception in over 90% of individuals with autism spectrum conditions (ASC). Sensory questionnaire-based studies in ASC mainly record parental reports of their child’s sensory experience; less is known about sensory reactivity in adults with ASC. Given the DSM-5 criteria for ASC now include sensory reactivity, there is a need for an adult questionnaire investigating basic sensory functioning. We aimed to develop and validate the Sensory Perception Quotient (SPQ), which assesses basic sensory hyper- and hyposensitivity across all five modalities. Methods A total of 359 adults with (n = 196) and without (n = 163) ASC were asked to fill in the SPQ, the Sensory Over-Responsivity Inventory (SensOR) and the Autism-Spectrum Quotient (AQ) online. Results Adults with ASC reported more sensory hypersensitivity on the SPQ compared to controls (P < .001). SPQ scores were correlated with AQ scores both across groups (r = .-38) and within the ASC (r = -.18) and control groups (r = -.15). Principal component analyses conducted separately in both groups indicated that one factor comprising 35 items consistently assesses sensory hypersensitivity. The SPQ showed high internal consistency for both the total SPQ (Cronbach’s alpha = .92) and the reduced 35-item version (alpha = .93). The SPQ was significantly correlated with the SensOR across groups (r = -.46) and within the ASC (r = -.49) and control group (r = -.21). Conclusions The SPQ shows good internal consistency and concurrent validity and differentiates between adults with and without ASC. Adults with ASC report more sensitivity to sensory stimuli on the SPQ. Finally, greater sensory sensitivity is associated with more autistic traits. The SPQ provides a new tool to measure individual differences on this dimension. PMID:24791196
Sherman, Sean P; Pyle, April D
2013-01-01
Differentiated cells from human embryonic stem cells (hESCs) provide an unlimited source of cells for use in regenerative medicine. The recent derivation of human induced pluripotent cells (hiPSCs) provides a potential supply of pluripotent cells that avoid immune rejection and could provide patient-tailored therapy. In addition, the use of pluripotent cells for drug screening could enable routine toxicity testing and evaluation of underlying disease mechanisms. However, prior to establishment of patient specific cells for cell therapy it is important to understand the basic regulation of cell fate decisions in hESCs. One critical issue that hinders the use of these cells is the fact that hESCs survive poorly upon dissociation, which limits genetic manipulation because of poor cloning efficiency of individual hESCs, and hampers production of large-scale culture of hESCs. To address the problems associated with poor growth in culture and our lack of understanding of what regulates hESC signaling, we successfully developed a screening platform that allows for large scale screening for small molecules that regulate survival. In this work we developed the first large scale platform for hESC screening using laser scanning cytometry and were able to validate this platform by identifying the pro-survival molecule HA-1077. These small molecules provide targets for both improving our basic understanding of hESC survival as well as a tool to improve our ability to expand and genetically manipulate hESCs for use in regenerative applications.
Hernandez, Sasha; Oliveira, Jessica Bastos; Shirazian, Taraneh
2017-01-01
In low- and middle-income countries (LMICs), where the rates of maternal mortality continue to be inappropriately high, there has been recognition of the importance of training traditional birth attendants (TBAs) to help improve outcomes during pregnancy and childbirth. In Guatemala, there is no national comprehensive training program in place despite the fact that the majority of women rely on TBAs during pregnancy and childbirth. This community case study presents a unique education program led by TBAs for TBAs in rural Guatemala. Discussion of this training program focuses on programming implementation, curriculum development, sustainable methodology, and how an educational partnership with the current national health-care system can increase access to health care for women in LMICs. Recent modifications to this training model are also discussed including how a change in the clinical curriculum is further integrating TBAs into the national health infrastructure. The training program has demonstrated that Guatemalan TBAs are able to improve their basic obstetrical knowledge, are capable of identifying and referring early complications of pregnancy and labor, and can deliver basic prenatal care that would otherwise not be provided. This training model is helping transform the role of the TBA from a sole cultural practitioner to a validated health-care provider within the health-care infrastructure of Guatemala and has the potential to do the same in other LMICs.
Causality and headache triggers
Turner, Dana P.; Smitherman, Todd A.; Martin, Vincent T.; Penzien, Donald B.; Houle, Timothy T.
2013-01-01
Objective The objective of this study was to explore the conditions necessary to assign causal status to headache triggers. Background The term “headache trigger” is commonly used to label any stimulus that is assumed to cause headaches. However, the assumptions required for determining if a given stimulus in fact has a causal-type relationship in eliciting headaches have not been explicated. Methods A synthesis and application of Rubin’s Causal Model is applied to the context of headache causes. From this application the conditions necessary to infer that one event (trigger) causes another (headache) are outlined using basic assumptions and examples from relevant literature. Results Although many conditions must be satisfied for a causal attribution, three basic assumptions are identified for determining causality in headache triggers: 1) constancy of the sufferer; 2) constancy of the trigger effect; and 3) constancy of the trigger presentation. A valid evaluation of a potential trigger’s effect can only be undertaken once these three basic assumptions are satisfied during formal or informal studies of headache triggers. Conclusions Evaluating these assumptions is extremely difficult or infeasible in clinical practice, and satisfying them during natural experimentation is unlikely. Researchers, practitioners, and headache sufferers are encouraged to avoid natural experimentation to determine the causal effects of headache triggers. Instead, formal experimental designs or retrospective diary studies using advanced statistical modeling techniques provide the best approaches to satisfy the required assumptions and inform causal statements about headache triggers. PMID:23534872
Do codependent traits involve more than basic dimensions of personality and psychopathology?
Gotham, H J; Sher, K J
1996-01-01
Despite widespread use of the term codependency, empirical evidence regarding its construct validity is generally lacking. This study analyzed the construct validity of codependency as measured by Potter-Efron and Potter-Efron's Codependency Assessment Questionnaire (CAQ). It attempted to determine the CAQ's factor structure and whether there are any unique relations between symptoms of codependency and parental alcoholism after controlling for basic dimensions of personality and psychopathology. Participants were 467 (246 male, 221 female) young adult children of alcoholics and controls who contributed complete questionnaire data at the fourth wave of a longitudinal study of factors related to alcohol use and abuse. The CAQ showed reliability and basically a one dimensional structure, and CAQ scores were significantly related to family history. Although much of this relation between family history and codependency was accounted for by neuroticism and symptoms of general psychopathology, a small, but significant, association between family history and codependency remained even after statistically controlling for personality and psychopathology. We conclude that, although there may be unique aspects of the purported codependency syndrome that are related to a family history of alcoholism, most of the relation between codependency and family history appears to be "explained" by general negative affectivity.
Removal of basic nitrogen compounds from hydrocarbon liquids
Givens, Edwin N.; Hoover, David S.
1985-01-01
A method is provided for reducing the concentration of basic nitrogen compounds in hydrocarbonaceous feedstock fluids used in the refining industry by providing a solid particulate carbonaceous adsorbent/fuel material such as coal having active basic nitrogen complexing sites on the surface thereof and the coal with a hydrocarbonaceous feedstock containing basic nitrogen compounds to facilitate attraction of the basic nitrogen compounds to the complexing sites and the formation of complexes thereof on the surface of the coal. The adsorbent coal material and the complexes formed thereon are from the feedstock fluid to provide a hydrocarbonaceous fluid of reduced basic nitrogen compound concentration. The coal can then be used as fuel for boilers and the like.
Rivard, Justin D; Vergis, Ashley S; Unger, Bertram J; Hardy, Krista M; Andrew, Chris G; Gillman, Lawrence M; Park, Jason
2014-06-01
Computer-based surgical simulators capture a multitude of metrics based on different aspects of performance, such as speed, accuracy, and movement efficiency. However, without rigorous assessment, it may be unclear whether all, some, or none of these metrics actually reflect technical skill, which can compromise educational efforts on these simulators. We assessed the construct validity of individual performance metrics on the LapVR simulator (Immersion Medical, San Jose, CA, USA) and used these data to create task-specific summary metrics. Medical students with no prior laparoscopic experience (novices, N = 12), junior surgical residents with some laparoscopic experience (intermediates, N = 12), and experienced surgeons (experts, N = 11) all completed three repetitions of four LapVR simulator tasks. The tasks included three basic skills (peg transfer, cutting, clipping) and one procedural skill (adhesiolysis). We selected 36 individual metrics on the four tasks that assessed six different aspects of performance, including speed, motion path length, respect for tissue, accuracy, task-specific errors, and successful task completion. Four of seven individual metrics assessed for peg transfer, six of ten metrics for cutting, four of nine metrics for clipping, and three of ten metrics for adhesiolysis discriminated between experience levels. Time and motion path length were significant on all four tasks. We used the validated individual metrics to create summary equations for each task, which successfully distinguished between the different experience levels. Educators should maintain some skepticism when reviewing the plethora of metrics captured by computer-based simulators, as some but not all are valid. We showed the construct validity of a limited number of individual metrics and developed summary metrics for the LapVR. The summary metrics provide a succinct way of assessing skill with a single metric for each task, but require further validation.
NASA Astrophysics Data System (ADS)
Debernardi, N.; Dunias, P.; van El, B.; Statham, A. E.
2014-03-01
A novel methodology is presented to mimic diffuse reflectance spectra of arbitrary biological tissues in the visible and near-infrared ranges. The prerequisite for this method is that the spectral information of basic components is sufficient to mimic an arbitrary tissue. Using a sterile disposable fiber optic probe the diffuse reflectance spectrum of a tissue (either in vivo or ex vivo) is measured, which forms the target spectrum. With the same type of fiber probe, a wide variety of basic components (ingredients) has been previously measured and all together forms a spectral database. A "recipe" for the optimal mixture of ingredients can then be derived using an algorithm that fits the absorption and scattering behavior of the target spectrum using the spectra of the basic components in the database. The spectral mimicking accuracy refines by adding more ingredients to the database. The validity of the principle is demonstrated by mimicking an arbitrary mixture of components. The method can be applied with different kinds of materials, e.g. gelatins, waxes and silicones, thus providing the possibility of mimicking the mechanical properties of target tissues as well. The algorithm can be extended from single point contact spectral measurement to contactless multi- and hyper-spectral camera acquisition. It can be applied to produce portable and durable tissue-like phantoms that provides consistent results over time for calibration, demonstration, comparison of instruments or other such tasks. They are also more readily available than living tissue or a cadaver and are not so limited by ease of handling and legislation; hence they are highly useful when developing new devices.
The Use of Artificial Neural Networks for Forecasting the Electric Demand of Stand-Alone Consumers
NASA Astrophysics Data System (ADS)
Ivanin, O. A.; Direktor, L. B.
2018-05-01
The problem of short-term forecasting of electric power demand of stand-alone consumers (small inhabited localities) situated outside centralized power supply areas is considered. The basic approaches to modeling the electric power demand depending on the forecasting time frame and the problems set, as well as the specific features of such modeling, are described. The advantages and disadvantages of the methods used for the short-term forecast of the electric demand are indicated, and difficulties involved in the solution of the problem are outlined. The basic principles of arranging artificial neural networks are set forth; it is also shown that the proposed method is preferable when the input information necessary for prediction is lacking or incomplete. The selection of the parameters that should be included into the list of the input data for modeling the electric power demand of residential areas using artificial neural networks is validated. The structure of a neural network is proposed for solving the problem of modeling the electric power demand of residential areas. The specific features of generation of the training dataset are outlined. The results of test modeling of daily electric demand curves for some settlements of Kamchatka and Yakutia based on known actual electric demand curves are provided. The reliability of the test modeling has been validated. A high value of the deviation of the modeled curve from the reference curve obtained in one of the four reference calculations is explained. The input data and the predicted power demand curves for the rural settlement of Kuokuiskii Nasleg are provided. The power demand curves were modeled for four characteristic days of the year, and they can be used in the future for designing a power supply system for the settlement. To enhance the accuracy of the method, a series of measures based on specific features of a neural network's functioning are proposed.
Evaluation of Sintering Behaviors of Saprolitic Nickeliferous Laterite Based on Quaternary Basicity
NASA Astrophysics Data System (ADS)
Luo, Jun; Li, Guanghui; Rao, Mingjun; Zhang, Yuanbo; Peng, Zhiwei; Zhi, Qian; Jiang, Tao
2015-09-01
The sintering behaviors of saprolitic nickeliferous laterite with various quaternary basicities [(CaO + MgO)/(SiO2 + Al2O3) mass ratio] in a reductive atmosphere are investigated by simulative sintering and validated by sintering pot tests. The simulative sintering results show that the generation of diopside (CaMgSi2O6) with low melting point is the key reason for the decrease in characteristic fusion temperatures when the quaternary basicity increases from 0.5 to 0.8-1.0. Continuous increase of basicity leads to transformation of diopside (CaMgSi2O6) into akermanite (Ca2MgSi2O7), which adversely increases the characteristic fusion temperatures. These findings are confirmed by the sinter pot tests, which demonstrate that the sintering indexes including vertical sintering velocity (VSV), yield ( Y), and productivity ( P), can be improved by optimizing quaternary basicity. At basicity of 1.0, the VSV, Y, P, and ISO tumbling index reach 49.2 mm/min, 80.5%, 1.0 t/(h m2), and 66.5%, respectively.
Health literacy and 30-day hospital readmission after acute myocardial infarction
Bailey, Stacy Cooper; Fang, Gang; Annis, Izabela E; O'Conor, Rachel; Paasche-Orlow, Michael K; Wolf, Michael S
2015-01-01
Objective To assess the validity of a predictive model of health literacy, and to examine the relationship between derived health literacy estimates and 30-day hospital readmissions for acute myocardial infarction (AMI). Design Retrospective cohort study. Setting and participants A National Institute of Aging (NIA) study cohort of 696 adult, English-speaking primary care patients, aged 55–74 years, was used to assess the validity of derived health literacy estimates. Claims from 7733 Medicare beneficiaries hospitalised for AMI in 2008 in North Carolina and Illinois were used to investigate the association between health literacy estimates and 30-day hospital readmissions. Measures The NIA cohort was administered 3 common health literacy assessments (Newest Vital Sign, Test of Functional Health Literacy in Adults, and Rapid Estimate of Adult Literacy in Medicine). Health literacy estimates at the census block group level were derived via a predictive model. 30-day readmissions were measured from Medicare claims data using a validated algorithm. Results Fair agreement was found between derived estimates and in-person literacy assessments (Pearson Correlation coefficients: 0.38–0.51; κ scores: 0.38–0.40). Medicare enrollees with above basic literacy according to derived health literacy estimates had an 18% lower risk of a 30-day readmission (RR=0.82, 95% CI 0.73 to 0.92) and 21% lower incidence rate of 30-day readmission (IRR=0.79, 95% CI 0.68 to 0.87) than patients with basic or below basic literacy. After adjusting for demographic and clinical characteristics, the risk of 30-day readmission was 12% lower (p=0.03), and the incidence rate 16% lower (p<0.01) for patients with above basic literacy. Conclusions Health literacy, as measured by a predictive model, was found to be a significant, independent predictor of 30-day readmissions. As a modifiable risk factor with evidence-based solutions, health literacy should be considered in readmission reduction efforts. PMID:26068508
Xiao, Dongjuan; Jakimowicz, Jack J; Albayrak, Armagan; Buzink, Sonja N; Botden, Sanne M B I; Goossens, Richard H M
2014-01-01
Laparoscopic skills can be improved effectively through laparoscopic simulation. The purpose of this study was to verify the face and content validity of a new portable Ergonomic Laparoscopic Skills simulator (Ergo-Lap simulator) and assess the construct validity of the Ergo-Lap simulator in 4 basic skills tasks. Four tasks were evaluated: 2 different translocation exercises (a basic bimanual exercise and a challenging single-handed exercise), an exercise involving tissue manipulation under tension, and a needle-handling exercise. Task performance was analyzed according to speed and accuracy. The participants rated the usability and didactic value of each task and the Ergo-Lap simulator along a 5-point Likert scale. Institutional academic medical center with its affiliated general surgery residency. Forty-six participants were allotted into 2 groups: a Novice group (n = 26, <10 clinical laparoscopic procedures) and an Experienced group (n = 20, >50 clinical laparoscopic procedures). The Experienced group completed all tasks in less time than the Novice group did (p < 0.001, Mann-Whitney U test). The Experienced group also completed tasks 1, 2, and 4 with fewer errors than the Novice group did (p < 0.05). Of the Novice participants, 96% considered that the present Ergo-Lap simulator could encourage more frequent practice of laparoscopic skills. In addition, 92% would like to purchase this simulator. All of the experienced participants confirmed that the Ergo-Lap simulator was easy to use and useful for practicing basic laparoscopic skills in an ergonomic manner. Most (95%) of these respondents would recommend this simulator to other surgical trainees. This Ergo-Lap simulator with multiple tasks was rated as a useful training tool that can distinguish between various levels of laparoscopic expertise. The Ergo-Lap simulator is also an inexpensive alternative, which surgical trainees could use to update their skills in the skills laboratory, at home, or in the office. Copyright © 2014 Association of Program Directors in Surgery. Published by Elsevier Inc. All rights reserved.
Mötteli, S; Barbey, J; Keller, C; Bucher, T; Siegrist, M
2016-04-01
As a high-quality diet is associated with a lower risk for several diseases and all-cause mortality, current nutrition education tools provide people with information regarding how to build a healthy and a balanced meal. To assess this basic nutrition knowledge, the research aim was to develop and validate a brief scale to measure the Practical Knowledge about Balanced meals (PKB-7). A pool of 25 items was pretested with experts and laypeople before being tested on a random sample in Switzerland (n=517). For item selection, a Rasch model analysis was applied. The validity and reliability of the new scale were assessed by three additional studies including laypeople (n=597; n=145) and nutrition experts (n=59). The final scale consists of seven multiple-choice items, which met the assumptions of the Rasch model. The validity of the new scale was shown by several aspects: the Rasch model was replicated in a second study, and nutrition experts achieved significantly higher scores than laypeople (t(148)=20.27, P<0.001, d=1.78). In addition, the PKB-7 scale was correlated with other nutrition-related constructs and associated with reported vegetable consumption. Test-retest reliability (r=0.68, P<0.001) was acceptable. The PKB-7 scale is a reliable and a valid Rasch-based instrument in Swiss citizens aged between 18 and 80 years for measuring the practical knowledge about balanced meals based on current dietary guidelines. This brief and easy-to-use scale is intended for application in both research and practice.
Barton, K L; Wrieden, W L; Anderson, A S
2011-12-01
Food skills programmes are widely used as a means to improve confidence in food preparation, the use of basic food skills and food selections amongst low income communities. However, the impact of such interventions are rarely evaluated as a result of a lack of validated assessment tools appropriate for use within this target group. A two-page questionnaire utilising a closed-question format was designed based on key domains known to be influenced by cooking skills programmes. Content validity was assessed by a panel of public health experts and face validity by individuals, typical of those who may attend cooking skills classes. Internal and repeat reliability were assessed with groups of adults attending community-based classes. The feasibility of using the tool in community settings was also assessed. The draft questionnaire was amended as appropriate subsequent to content and face validity testing. Cronbach's alpha for confidence and knowledge sections was 0.86 and 0.84, respectively, indicating good internal consistency. Spearman correlation coefficients for repeat reliability testing between time 1 and time 2 for each item were in the range 0.46-0.91 (all significant at P < 0.001), indicating that the questionnaire elicited stable responses for repeated use. Feasibility testing highlighted the need for detailed instructions for course tutors on how to distribute and check questionnaires for completion. This tool provides a standardised method of evaluating cooking skills interventions that could be utilised in the development and evaluation of multicentre cooking skills interventions. © 2011 The Authors. Journal of Human Nutrition and Dietetics © 2011 The British Dietetic Association Ltd.
Allelic variants of hereditary prions: The bimodularity principle.
Tikhodeyev, Oleg N; Tarasov, Oleg V; Bondarev, Stanislav A
2017-01-02
Modern biology requires modern genetic concepts equally valid for all discovered mechanisms of inheritance, either "canonical" (mediated by DNA sequences) or epigenetic. Applying basic genetic terms such as "gene" and "allele" to protein hereditary factors is one of the necessary steps toward these concepts. The basic idea that different variants of the same prion protein can be considered as alleles has been previously proposed by Chernoff and Tuite. In this paper, the notion of prion allele is further developed. We propose the idea that any prion allele is a bimodular hereditary system that depends on a certain DNA sequence (DNA determinant) and a certain epigenetic mark (epigenetic determinant). Alteration of any of these 2 determinants may lead to establishment of a new prion allele. The bimodularity principle is valid not only for hereditary prions; it seems to be universal for any epigenetic hereditary factor.
Allelic variants of hereditary prions: The bimodularity principle
Tikhodeyev, Oleg N.; Tarasov, Oleg V.; Bondarev, Stanislav A.
2017-01-01
ABSTRACT Modern biology requires modern genetic concepts equally valid for all discovered mechanisms of inheritance, either “canonical” (mediated by DNA sequences) or epigenetic. Applying basic genetic terms such as “gene” and “allele” to protein hereditary factors is one of the necessary steps toward these concepts. The basic idea that different variants of the same prion protein can be considered as alleles has been previously proposed by Chernoff and Tuite. In this paper, the notion of prion allele is further developed. We propose the idea that any prion allele is a bimodular hereditary system that depends on a certain DNA sequence (DNA determinant) and a certain epigenetic mark (epigenetic determinant). Alteration of any of these 2 determinants may lead to establishment of a new prion allele. The bimodularity principle is valid not only for hereditary prions; it seems to be universal for any epigenetic hereditary factor. PMID:28281926
Validity in work-based assessment: expanding our horizons.
Govaerts, Marjan; van der Vleuten, Cees P M
2013-12-01
Although work-based assessments (WBA) may come closest to assessing habitual performance, their use for summative purposes is not undisputed. Most criticism of WBA stems from approaches to validity consistent with the quantitative psychometric framework. However, there is increasing research evidence that indicates that the assumptions underlying the predictive, deterministic framework of psychometrics may no longer hold. In this discussion paper we argue that meaningfulness and appropriateness of current validity evidence can be called into question and that we need alternative strategies to assessment and validity inquiry that build on current theories of learning and performance in complex and dynamic workplace settings. Drawing from research in various professional fields we outline key issues within the mechanisms of learning, competence and performance in the context of complex social environments and illustrate their relevance to WBA. In reviewing recent socio-cultural learning theory and research on performance and performance interpretations in work settings, we demonstrate that learning, competence (as inferred from performance) as well as performance interpretations are to be seen as inherently contextualised, and can only be under-stood 'in situ'. Assessment in the context of work settings may, therefore, be more usefully viewed as a socially situated interpretive act. We propose constructivist-interpretivist approaches towards WBA in order to capture and understand contextualised learning and performance in work settings. Theoretical assumptions underlying interpretivist assessment approaches call for a validity theory that provides the theoretical framework and conceptual tools to guide the validation process in the qualitative assessment inquiry. Basic principles of rigour specific to qualitative research have been established, and they can and should be used to determine validity in interpretivist assessment approaches. If used properly, these strategies generate trustworthy evidence that is needed to develop the validity argument in WBA, allowing for in-depth and meaningful information about professional competence. © 2013 John Wiley & Sons Ltd.
Software reliability through fault-avoidance and fault-tolerance
NASA Technical Reports Server (NTRS)
Vouk, Mladen A.; Mcallister, David F.
1993-01-01
Strategies and tools for the testing, risk assessment and risk control of dependable software-based systems were developed. Part of this project consists of studies to enable the transfer of technology to industry, for example the risk management techniques for safety-concious systems. Theoretical investigations of Boolean and Relational Operator (BRO) testing strategy were conducted for condition-based testing. The Basic Graph Generation and Analysis tool (BGG) was extended to fully incorporate several variants of the BRO metric. Single- and multi-phase risk, coverage and time-based models are being developed to provide additional theoretical and empirical basis for estimation of the reliability and availability of large, highly dependable software. A model for software process and risk management was developed. The use of cause-effect graphing for software specification and validation was investigated. Lastly, advanced software fault-tolerance models were studied to provide alternatives and improvements in situations where simple software fault-tolerance strategies break down.
Instrumental variable methods in comparative safety and effectiveness research.
Brookhart, M Alan; Rassen, Jeremy A; Schneeweiss, Sebastian
2010-06-01
Instrumental variable (IV) methods have been proposed as a potential approach to the common problem of uncontrolled confounding in comparative studies of medical interventions, but IV methods are unfamiliar to many researchers. The goal of this article is to provide a non-technical, practical introduction to IV methods for comparative safety and effectiveness research. We outline the principles and basic assumptions necessary for valid IV estimation, discuss how to interpret the results of an IV study, provide a review of instruments that have been used in comparative effectiveness research, and suggest some minimal reporting standards for an IV analysis. Finally, we offer our perspective of the role of IV estimation vis-à-vis more traditional approaches based on statistical modeling of the exposure or outcome. We anticipate that IV methods will be often underpowered for drug safety studies of very rare outcomes, but may be potentially useful in studies of intended effects where uncontrolled confounding may be substantial.
A methodology for long-range prediction of air transportation
NASA Technical Reports Server (NTRS)
Ayati, M. B.; English, J. M.
1980-01-01
A framework and methodology for long term projection of demand for aviation fuels is presented. The approach taken includes two basic components. The first was a new technique for establishing the socio-economic environment within which the future aviation industry is embedded. The concept utilized was a definition of an overall societal objective for the very long run future. Within a framework so defined, a set of scenarios by which the future will unfold are then written. These scenarios provide the determinants of the air transport industry operations and accordingly provide an assessment of future fuel requirements. The second part was the modeling of the industry in terms of an abstracted set of variables to represent the overall industry performance on a macro scale. The model was validated by testing the desired output variables from the model with historical data over the past decades.
DRS: Derivational Reasoning System
NASA Technical Reports Server (NTRS)
Bose, Bhaskar
1995-01-01
The high reliability requirements for airborne systems requires fault-tolerant architectures to address failures in the presence of physical faults, and the elimination of design flaws during the specification and validation phase of the design cycle. Although much progress has been made in developing methods to address physical faults, design flaws remain a serious problem. Formal methods provides a mathematical basis for removing design flaws from digital systems. DRS (Derivational Reasoning System) is a formal design tool based on advanced research in mathematical modeling and formal synthesis. The system implements a basic design algebra for synthesizing digital circuit descriptions from high level functional specifications. DRS incorporates an executable specification language, a set of correctness preserving transformations, verification interface, and a logic synthesis interface, making it a powerful tool for realizing hardware from abstract specifications. DRS integrates recent advances in transformational reasoning, automated theorem proving and high-level CAD synthesis systems in order to provide enhanced reliability in designs with reduced time and cost.
Cohn, Daniel Wagner Hamada; Kinoshita, Denise; Palermo-Neto, João
2012-07-01
In spite of the high prevalence and negative impact of depression, little is known about its pathophysiology. Basic research on depression needs new animal models in order to increase knowledge of the disease and search for new therapies. The work presented here aims to provide a neurobiologically validated model for investigating the relationships among sickness behavior, antidepressants treatment, and social dominance behavior. For this purpose, dominant individuals from dyads of male Swiss mice were treated with the bacterial endotoxin lipopolysaccharide (LPS) to induce social hierarchy destabilization. Two groups were treated with the antidepressants imipramine and fluoxetine prior to LPS administration. In these groups, antidepressant treatment prevented the occurrence of social destabilization. These results indicate that this model could be useful in providing new insights into the understanding of the brain systems involved in depression. © 2012 New York Academy of Sciences.
Control of Flexible Structures (COFS) Flight Experiment Background and Description
NASA Technical Reports Server (NTRS)
Hanks, B. R.
1985-01-01
A fundamental problem in designing and delivering large space structures to orbit is to provide sufficient structural stiffness and static configuration precision to meet performance requirements. These requirements are directly related to control requirements and the degree of control system sophistication available to supplement the as-built structure. Background and rationale are presented for a research study in structures, structural dynamics, and controls using a relatively large, flexible beam as a focus. This experiment would address fundamental problems applicable to large, flexible space structures in general and would involve a combination of ground tests, flight behavior prediction, and instrumented orbital tests. Intended to be multidisciplinary but basic within each discipline, the experiment should provide improved understanding and confidence in making design trades between structural conservatism and control system sophistication for meeting static shape and dynamic response/stability requirements. Quantitative results should be obtained for use in improving the validity of ground tests for verifying flight performance analyses.
Salas, Daniela; Borrull, Francesc; Fontanals, Núria; Marcé, Rosa Maria
2018-01-01
The aim of the present study is to broaden the applications of mixed-mode ion-exchange solid-phase extraction sorbents to extract both basic and acidic compounds simultaneously by combining the sorbents in a single cartridge and developing a simplified extraction procedure. Four different cartridges containing negative and positive charges in the same configuration were evaluated and compared to extract a group of basic, neutral, and acidic pharmaceuticals selected as model compounds. After a thorough optimization of the extraction conditions, the four different cartridges showed to be capable of retaining basic and acidic pharmaceuticals simultaneously through ionic interactions, allowing the introduction of a washing step with 15 mL methanol to eliminate interferences retained by hydrophobic interactions. Using the best combined cartridge, a method was developed, validated, and further applied to environmental waters to demonstrate that the method is promising for the extraction of basic and acidic compounds from very complex samples.
NASA Astrophysics Data System (ADS)
Witarsyah Jacob, Deden; Fudzee, Mohd Farhan Md; Aizi Salamat, Mohamad; Kasim, Shahreen; Mahdin, Hairulnizam; Azhar Ramli, Azizul
2017-08-01
Many governments around the world increasingly use internet technologies such as electronic government to provide public services. These services range from providing the most basic informational website to deploying sophisticated tools for managing interactions between government agencies and beyond government. Electronic government (e-government) aims to provide a more accurate, easily accessible, cost-effective and time saving for the community. In this study, we develop a new model of e-government adoption service by extending the Unified Theory of Acceptance and Use of Technology (UTAUT) through the incorporation of some variables such as System Quality, Information Quality and Trust. The model is then tested using a large-scale, multi-site survey research of 237 Indonesian citizens. This model will be validated by using Structural Equation Modeling (SEM). The result indicates that System Quality, Information Quality and Trust variables proven to effect user behavior. This study extends the current understanding on the influence of System Quality, Information Quality and Trust factors to researchers, practitioners, and policy makers.
NASA Astrophysics Data System (ADS)
Antognini, Luca M.; Assenza, Salvatore; Speziale, Chiara; Mezzenga, Raffaele
2016-08-01
Lyotropic Liquid Crystals (LLCs) are a class of lipid-based membranes with a strong potential for drug-delivery employment. The characterization and control of their transport properties is a central issue in this regard, and has recently prompted a notable volume of research on the topic. A promising experimental approach is provided by the so-called diffusion setup, where the drug molecules diffuse from a feeding chamber filled with water to a receiving one passing through a LLC. In the present work we provide a theoretical framework for the proper description of this setup, and validate it by means of targeted experiments. Due to the inhomogeneity of the system, a rich palette of different diffusion dynamics emerges from the interplay of the different time- and lengthscales thereby present. Our work paves the way to the employment of diffusion experiments to quantitatively characterize the transport properties of LLCs, and provides the basic tools for device diffusion setups with controlled kinetic properties.
Fitzpatrick, Megan J; Mathewson, Paul D; Porter, Warren P
2015-01-01
Mechanistic models provide a powerful, minimally invasive tool for gaining a deeper understanding of the ecology of animals across geographic space and time. In this paper, we modified and validated the accuracy of the mechanistic model Niche Mapper for simulating heat exchanges of animals with counter-current heat exchange mechanisms in their legs and animals that wade in water. We then used Niche Mapper to explore the effects of wading and counter-current heat exchange on the energy expenditures of Whooping Cranes, a long-legged wading bird. We validated model accuracy against the energy expenditure of two captive Whooping Cranes measured using the doubly-labeled water method and time energy budgets. Energy expenditure values modeled by Niche Mapper were similar to values measured by the doubly-labeled water method and values estimated from time-energy budgets. Future studies will be able to use Niche Mapper as a non-invasive tool to explore energy-based limits to the fundamental niche of Whooping Cranes and apply this knowledge to management decisions. Basic questions about the importance of counter-current exchange and wading to animal physiological tolerances can also now be explored with the model.
Fitzpatrick, Megan J.; Mathewson, Paul D.; Porter, Warren P.
2015-01-01
Mechanistic models provide a powerful, minimally invasive tool for gaining a deeper understanding of the ecology of animals across geographic space and time. In this paper, we modified and validated the accuracy of the mechanistic model Niche Mapper for simulating heat exchanges of animals with counter-current heat exchange mechanisms in their legs and animals that wade in water. We then used Niche Mapper to explore the effects of wading and counter-current heat exchange on the energy expenditures of Whooping Cranes, a long-legged wading bird. We validated model accuracy against the energy expenditure of two captive Whooping Cranes measured using the doubly-labeled water method and time energy budgets. Energy expenditure values modeled by Niche Mapper were similar to values measured by the doubly-labeled water method and values estimated from time-energy budgets. Future studies will be able to use Niche Mapper as a non-invasive tool to explore energy-based limits to the fundamental niche of Whooping Cranes and apply this knowledge to management decisions. Basic questions about the importance of counter-current exchange and wading to animal physiological tolerances can also now be explored with the model. PMID:26308207
Pérez-Del-Olmo, A; Montero, F E; Fernández, M; Barrett, J; Raga, J A; Kostadinova, A
2010-10-01
We address the effect of spatial scale and temporal variation on model generality when forming predictive models for fish assignment using a new data mining approach, Random Forests (RF), to variable biological markers (parasite community data). Models were implemented for a fish host-parasite system sampled along the Mediterranean and Atlantic coasts of Spain and were validated using independent datasets. We considered 2 basic classification problems in evaluating the importance of variations in parasite infracommunities for assignment of individual fish to their populations of origin: multiclass (2-5 population models, using 2 seasonal replicates from each of the populations) and 2-class task (using 4 seasonal replicates from 1 Atlantic and 1 Mediterranean population each). The main results are that (i) RF are well suited for multiclass population assignment using parasite communities in non-migratory fish; (ii) RF provide an efficient means for model cross-validation on the baseline data and this allows sample size limitations in parasite tag studies to be tackled effectively; (iii) the performance of RF is dependent on the complexity and spatial extent/configuration of the problem; and (iv) the development of predictive models is strongly influenced by seasonal change and this stresses the importance of both temporal replication and model validation in parasite tagging studies.
The Development and Validation of a Rapid Assessment Tool of Primary Care in China
Mei, Jie; Liang, Yuan; Shi, LeiYu; Zhao, JingGe; Wang, YuTan; Kuang, Li
2016-01-01
Introduction. With Chinese health care reform increasingly emphasizing the importance of primary care, the need for a tool to evaluate primary care performance and service delivery is clear. This study presents a methodology for a rapid assessment of primary care organizations and service delivery in China. Methods. The study translated and adapted the Primary Care Assessment Tool-Adult Edition (PCAT-AE) into a Chinese version to measure core dimensions of primary care, namely, first contact, continuity, comprehensiveness, and coordination. A cross-sectional survey was conducted to assess the validity and reliability of the Chinese Rapid Primary Care Assessment Tool (CR-PCAT). Eight community health centers in Guangdong province have been selected to participate in the survey. Results. A total of 1465 effective samples were included for data analysis. Eight items were eliminated following principal component analysis and reliability testing. The principal component analysis extracted five multiple-item scales (first contact utilization, first contact accessibility, ongoing care, comprehensiveness, and coordination). The tests of scaling assumptions were basically met. Conclusion. The standard psychometric evaluation indicates that the scales have achieved relatively good reliability and validity. The CR-PCAT provides a rapid and reliable measure of four core dimensions of primary care, which could be applied in various scenarios. PMID:26885509
Brinkmann, Christian; Fritz, Mathias; Pankratius, Ulrich; Bahde, Ralf; Neumann, Philipp; Schlueter, Steffen; Senninger, Norbert; Rijcken, Emile
Simulation training improves laparoscopic performance. Laparoscopic basic skills can be learned in simulators as box- or virtual-reality (VR) trainers. However, there is no clear recommendation for either box or VR trainers as the most appropriate tool for the transfer of acquired laparoscopic basic skills into a surgical procedure. Both training tools were compared, using validated and well-established curricula in the acquirement of basic skills, in a prospective randomized trial in a 5-day structured laparoscopic training course. Participants completed either a box- or VR-trainer curriculum and then applied the learned skills performing an ex situ laparoscopic cholecystectomy on a pig liver. The performance was recorded on video and evaluated offline by 4 blinded observers using the Global Operative Assessment of Laparoscopic Skills (GOALS) score. Learning curves of the various exercises included in the training course were compared and the improvement in each exercise was analyzed. Surgical Skills Lab of the Department of General and Visceral Surgery, University Hospital Muenster. Surgical novices without prior surgical experience (medical students, n = 36). Posttraining evaluation showed significant improvement compared with baseline in both groups, indicating acquisition of laparoscopic basic skills. Learning curves showed almost the same progression with no significant differences. In simulated laparoscopic cholecystectomy, total GOALS score was significantly higher for the box-trained group than the VR-trained group (box: 15.31 ± 3.61 vs. VR: 12.92 ± 3.06; p = 0.039; Hedge׳s g* = 0.699), indicating higher technical skill levels. Despite both systems having advantages and disadvantages, they can both be used for simulation training for laparoscopic skills. In the setting with 2 structured, validated and almost identical curricula, the box-trained group appears to be superior in the better transfer of basic skills into an experimental but structured surgical procedure. Copyright © 2017 Association of Program Directors in Surgery. Published by Elsevier Inc. All rights reserved.
Rosser, James C; Liu, Xinwei; Jacobs, Charles; Choi, Katherine Mia; Jalink, Maarten B; Ten Cate Hoedemaker, Henk O
2017-04-01
This abstract profiles the comparison of correlations between previously validated Super Monkey Ball (SMB) and recently introduced Underground (U) video game on the Nintendo Wii U to multiple validated tasks used for developing basic and advanced laparoscopic skills. Sixty-eight participants, 53 residents and 15 attending surgeons, performed the Top Gun Pea Drop, FLS Peg Pass, intracorporeal suturing, and two video games (SMB and U). SMB is an over-the-counter game, and U was formulated for laparoscopic skill training. Spearman's rank correlations were performed looking at performance comparing the three validated laparoscopic training tasks, and SMB/U. The SMB score had a moderate correlation with intracorporeal suturing (ρ = 0.39, p < 0.01), and the final score involving all three tasks (ρ = 0.39, p < 0.01), but low correlations with Pea Drop Drill and FLS Peg Transfer (ρ = 0.11, 0.18, p < 0.01). The U score had a small correlation with intracorporeal suturing and final score (ρ = 0.09, 0.13, p < 0.01). However, there were correlations between U score and Pea Drop Drill, and FLS Peg Transfer (ρ = 0.24, 0.27, p < 0.01, respectively). In this study, SMB had a very significant correlation with intracorporeal suturing. U demonstrated more of a correlation with basic skills. At this point, our conclusion would be that both are effective for laparoscopic skill training, and they should be used in tandem rather than alone.
ASK ME--A Merit Promotion System
ERIC Educational Resources Information Center
Johnson, Michael L.; McCloskey, Kenneth R.
1978-01-01
Describes a procedure developed at the Red River Army Depot, Texarkana, Texas, for selecting and evaluating personnel called AKS ME (Ability, Skill and Knowledge Merit Evaluation), consisting of four basic steps: job analysis, examination plan, candidate evaluation, and validation. (MF)
PA.NET International Quality Certification Protocol for blood pressure monitors.
Omboni, Stefano; Costantini, Carlo; Pini, Claudio; Bulegato, Roberto; Manfellotto, Dario; Rizzoni, Damiano; Palatini, Paolo; O'brien, Eoin; Parati, Gianfranco
2008-10-01
Although standard validation protocols provide assurance of the accuracy of blood pressure monitors (BPMs), there is no guidance for the consumer as to the overall quality of a device. The PA.NET International Quality Certification Protocol, developed by the Association for Research and Development of Biomedical Technologies and for Continuing Medical Education (ARSMED), a nonprofit organization, with the support of the Italian Society of Hypertension-Italian Hypertension League, and the dabl Educational Trust denotes additional criteria of quality for BPMs that fulfilled basic validation criteria, published in full in peer-reviewed medical journals. The certification is characterized by three phases: (i) to determine that the device fulfilled standard validation criteria; (ii) to determine the technical and functional characteristics of the device (e.g. operativity, display dimension, accessory functions, memory availability, etc.) and (iii) to determine the commercial characteristics (e.g. price-quality ratio, after-sale service, guarantee, etc.). At the end of the certification process, ARSMED attributes a quality index to the device, based on a scale ranging from 1 to 100, and a quality seal with four different grades (bronze, silver, gold and diamond) according to the achieved score. The seal is identified by a unique alphanumeric code. The quality seal may be used on the packaging of the appliance or in advertising. A quality certification is released to the manufacturer and published on www.pressionearteriosa.net and www.dableducational.org. The PA.NET International Quality Certification Protocol represents the first attempt to provide health care personnel and consumers with an independent and objective assessment of BPMs based on their quality.
Gençöz, Tülin; Öcül, Öznur
2012-01-01
The aim of the present study was to test the cross-cultural validity of the five-factor nature of personality. For this aim, an indigenous, psychometrically strong instrument measuring the basic personality dimensions within Turkish culture and language was developed through three consecutive studies. The first study aimed to reveal the adjectives that have been most frequently used to define people in the Turkish culture. In the second study, factor analysis of these personality characteristics revealed big five personality factors, along with the sixth factor, which had been called as the Negative Valence factor. The adjectives that most strongly represented and differentiated each factor constituted 45-item "Basic Personality Traits Inventory". Finally, in the third study, psychometric characteristics of the Basic Personality Traits Inventory were examined. Factor structure and psychometric properties of this instrument confirmed that five-factor nature of personality may not hold true in every culture.
Blais, Marie-Claude; Maunsell, Elizabeth; Grenier, Sophie; Lauzier, Sophie; Dorval, Michel; Pelletier, Sylvie; Guay, Stéphane; Robidoux, André; Provencher, Louise
2014-09-01
The study objectives were to identify key information components that would be the basic content of a brief informational intervention, developed from a population perspective, to empower individual couple members facing breast cancer and to validate the relevance and acceptability of these components. A review of information relevant to couples facing cancer presented in internet sites and documents of national cancer organizations was made to identify information components to include in a brief informational intervention. These information components were framed as messages, that is, very brief sentences or tips. To validate the relevance and acceptability of these messages, six focus groups were conducted in Quebec City and Montreal among women who had had breast cancer and their spouses. Reactions to the messages were synthesized by analyzing the verbatim transcripts. A total of 70 individuals (35 women with an average of 14 months since diagnosis and 35 spouses) participated in the groups. The content of almost all messages was seen as relevant, although nuances and improvements were discussed. However, the message format provoked irritation and will need improvement. Findings provide validation of the relevance, from the viewpoint of women and their spouses, of the message content to be part of a brief informational intervention intended to empower couples as they cope with breast cancer. Couples approved of the idea of being guided in their adjustment to breast cancer. However, the message format requires adaptation and further testing.
Jacqmin, Dustin J; Bredfeldt, Jeremy S; Frigo, Sean P; Smilowitz, Jennifer B
2017-01-01
The AAPM Medical Physics Practice Guideline (MPPG) 5.a provides concise guidance on the commissioning and QA of beam modeling and dose calculation in radiotherapy treatment planning systems. This work discusses the implementation of the validation testing recommended in MPPG 5.a at two institutions. The two institutions worked collaboratively to create a common set of treatment fields and analysis tools to deliver and analyze the validation tests. This included the development of a novel, open-source software tool to compare scanning water tank measurements to 3D DICOM-RT Dose distributions. Dose calculation algorithms in both Pinnacle and Eclipse were tested with MPPG 5.a to validate the modeling of Varian TrueBeam linear accelerators. The validation process resulted in more than 200 water tank scans and more than 50 point measurements per institution, each of which was compared to a dose calculation from the institution's treatment planning system (TPS). Overall, the validation testing recommended in MPPG 5.a took approximately 79 person-hours for a machine with four photon and five electron energies for a single TPS. Of the 79 person-hours, 26 person-hours required time on the machine, and the remainder involved preparation and analysis. The basic photon, electron, and heterogeneity correction tests were evaluated with the tolerances in MPPG 5.a, and the tolerances were met for all tests. The MPPG 5.a evaluation criteria were used to assess the small field and IMRT/VMAT validation tests. Both institutions found the use of MPPG 5.a to be a valuable resource during the commissioning process. The validation testing in MPPG 5.a showed the strengths and limitations of the TPS models. In addition, the data collected during the validation testing is useful for routine QA of the TPS, validation of software upgrades, and commissioning of new algorithms. © 2016 The Authors. Journal of Applied Clinical Medical Physics published by Wiley Periodicals, Inc. on behalf of American Association of Physicists in Medicine.
Revisiting sample size: are big trials the answer?
Lurati Buse, Giovanna A L; Botto, Fernando; Devereaux, P J
2012-07-18
The superiority of the evidence generated in randomized controlled trials over observational data is not only conditional to randomization. Randomized controlled trials require proper design and implementation to provide a reliable effect estimate. Adequate random sequence generation, allocation implementation, analyses based on the intention-to-treat principle, and sufficient power are crucial to the quality of a randomized controlled trial. Power, or the probability of the trial to detect a difference when a real difference between treatments exists, strongly depends on sample size. The quality of orthopaedic randomized controlled trials is frequently threatened by a limited sample size. This paper reviews basic concepts and pitfalls in sample-size estimation and focuses on the importance of large trials in the generation of valid evidence.
Collapse of a Liquid Column: Numerical Simulation and Experimental Validation
NASA Astrophysics Data System (ADS)
Cruchaga, Marcela A.; Celentano, Diego J.; Tezduyar, Tayfun E.
2007-03-01
This paper is focused on the numerical and experimental analyses of the collapse of a liquid column. The measurements of the interface position in a set of experiments carried out with shampoo and water for two different initial column aspect ratios are presented together with the corresponding numerical predictions. The experimental procedure was found to provide acceptable recurrence in the observation of the interface evolution. Basic models describing some of the relevant physical aspects, e.g. wall friction and turbulence, are included in the simulations. Numerical experiments are conducted to evaluate the influence of the parameters involved in the modeling by comparing the results with the data from the measurements. The numerical predictions reasonably describe the physical trends.
Standards of care for obsessive–compulsive disorder centres
Menchón, José M.; van Ameringen, Michael; Dell’Osso, Bernardo; Denys, Damiaan; Figee, Martijn; Grant, Jon E.; Hollander, Eric; Marazziti, Donatella; Nicolini, Humberto; Pallanti, Stefano; Ruck, Christian; Shavitt, Roseli; Stein, Dan J.; Andersson, Erik; Bipeta, Rajshekhar; Cath, Danielle C.; Drummond, Lynne; Feusner, Jamie; Geller, Daniel A.; Hranov, Georgi; Lochner, Christine; Matsunaga, Hisato; McCabe, Randy E.; Mpavaenda, Davis; Nakamae, Takashi; O'Kearney, Richard; Pasquini, Massimo; Pérez Rivera, Ricardo; Poyurovsky, Michael; Real, Eva; do Rosário, Maria Conceição; Soreni, Noam; Swinson, Richard P.; Vulink, Nienke; Zohar, Joseph; Fineberg, Naomi
2016-01-01
Abstract In recent years, many assessment and care units for obsessive–compulsive disorder (OCD) have been set up in order to detect, diagnose and to properly manage this complex disorder, but there is no consensus regarding the key functions that these units should perform. The International College of Obsessive-Compulsive Spectrum Disorders (ICOCS) together with the Obsessive Compulsive and Related Disorders Network (OCRN) of the European College of Neuropsychopharmacology (ECNP) and the Anxiety and Obsessive Compulsive Disorders Section of the World Psychiaric Association (WPA) has developed a standards of care programme for OCD centres. The goals of this collaborative initiative are promoting basic standards, improving the quality of clinical care and enhance the validity and reliability of research results provided by different facilities and countries. PMID:27359333
Slininger, P J; Dien, B S; Lomont, J M; Bothast, R J; Ladisch, M R; Okos, M R
2014-08-01
Scheffersomyces (formerly Pichia) stipitis is a potential biocatalyst for converting lignocelluloses to ethanol because the yeast natively ferments xylose. An unstructured kinetic model based upon a system of linear differential equations has been formulated that describes growth and ethanol production as functions of ethanol, oxygen, and xylose concentrations for both growth and fermentation stages. The model was validated for various growth conditions including batch, cell recycle, batch with in situ ethanol removal and fed-batch. The model provides a summary of basic physiological yeast properties and is an important tool for simulating and optimizing various culture conditions and evaluating various bioreactor designs for ethanol production. © 2014 Wiley Periodicals, Inc.
Li, Qiuping; Lin, Yi; Hu, Caiping; Xu, Yinghua; Zhou, Huiya; Yang, Liping; Xu, Yongyong
2016-12-01
The Hospital Anxiety and Depression Scale (HADS) acts as one of the most frequently used self-reported measures in cancer practice. The evidence for construct validity of HADS, however, remains inconclusive. The objective of this study is to evaluate the psychometric properties of the Chinese version HADS (C-HADS) in terms of construct validity, internal consistency reliability, and concurrent validity in dyads of Chinese cancer patients and their family caregivers. This was a cross-sectional study, conducted in multiple centers: one hospital in each of the seven different administrative regions in China from October 2014 to May 2015. A total of 641 dyads, consisting of cancer patients and family caregivers, completed a survey assessing their demographic and background information, anxiety and depression using C-HADS, and quality of life (QOL) using Chinese version SF-12. Data analysis methods included descriptive statistics, confirmatory factor analysis (CFA), and Pearson correlations. Both the two-factor and one-factor models offered the best and adequate fit to the data in cancer patients and family caregivers respectively. The comparison of the two-factor and single-factor models supports the basic assumption of two-factor construct of C-HADS. The overall and two subscales of C-HADS in both cancer patients and family caregivers had good internal consistency and acceptable concurrent validity. The Chinese version of the HADS may be a reliable and valid screening tool, as indicated by its original two-factor structure. The finding supports the basic assumption of two-factor construct of HADS. Copyright © 2016 Elsevier Ltd. All rights reserved.
Hybrid and conventional hydrogen engine vehicles that meet EZEV emissions
DOE Office of Scientific and Technical Information (OSTI.GOV)
Aceves, S.M.; Smith, J.R.
In this paper, a time-dependent engine model is used for predicting hydrogen engine efficiency and emissions. The model uses basic thermodynamic equations for the compression and expansion processes, along with an empirical correlation for heat transfer, to predict engine indicated efficiency. A friction correlation and a supercharger/turbocharger model are then used to calculate brake thermal efficiency. The model is validated with many experimental points obtained in a recent evaluation of a hydrogen research engine. A The validated engine model is then used to calculate fuel economy and emissions for three hydrogen-fueled vehicles: a conventional, a parallel hybrid, and a seriesmore » hybrid. All vehicles use liquid hydrogen as a fuel. The hybrid vehicles use a flywheel for energy storage. Comparable ultra capacitor or battery energy storage performance would give similar results. This paper analyzes the engine and flywheel sizing requirements for obtaining a desired level of performance. The results indicate that hydrogen lean-burn spark-ignited engines can provide a high fuel economy and Equivalent Zero Emission Vehicle (EZEV) levels in the three vehicle configurations being analyzed.« less
CaveMan Enterprise version 1.0 Software Validation and Verification.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Hart, David
The U.S. Department of Energy Strategic Petroleum Reserve stores crude oil in caverns solution-mined in salt domes along the Gulf Coast of Louisiana and Texas. The CaveMan software program has been used since the late 1990s as one tool to analyze pressure mea- surements monitored at each cavern. The purpose of this monitoring is to catch potential cavern integrity issues as soon as possible. The CaveMan software was written in Microsoft Visual Basic, and embedded in a Microsoft Excel workbook; this method of running the CaveMan software is no longer sustainable. As such, a new version called CaveMan Enter- prisemore » has been developed. CaveMan Enterprise version 1.0 does not have any changes to the CaveMan numerical models. CaveMan Enterprise represents, instead, a change from desktop-managed work- books to an enterprise framework, moving data management into coordinated databases and porting the numerical modeling codes into the Python programming language. This document provides a report of the code validation and verification testing.« less
Swart, Marcel; Rösler, Ernst; Bickelhaupt, F Matthias
2006-10-01
We have carried out an extensive exploration of the gas-phase basicity of archetypal neutral bases across the periodic system using the generalized gradient approximation (GGA) of the density functional theory (DFT) at BP86/QZ4P//BP86/TZ2P. First, we validate DFT as a reliable tool for computing proton affinities and related thermochemical quantities: BP86/QZ4P//BP86/TZ2P is shown to yield a mean absolute deviation of 2.0 kcal/mol for the proton affinity at 298 K with respect to experiment, and 1.2 kcal/mol with high-level ab initio benchmark data. The main purpose of this work is to provide the proton affinities (and corresponding entropies) at 298 K of the neutral bases constituted by all maingroup-element hydrides of groups 15-17 and the noble gases, that is, group 18, and periods 1-6. We have also studied the effect of step-wise methylation of the protophilic center of the second- and third-period bases. Copyright 2006 Wiley Periodicals, Inc.
Prediction of protein subcellular locations by GO-FunD-PseAA predictor.
Chou, Kuo-Chen; Cai, Yu-Dong
2004-08-06
The localization of a protein in a cell is closely correlated with its biological function. With the explosion of protein sequences entering into DataBanks, it is highly desired to develop an automated method that can fast identify their subcellular location. This will expedite the annotation process, providing timely useful information for both basic research and industrial application. In view of this, a powerful predictor has been developed by hybridizing the gene ontology approach [Nat. Genet. 25 (2000) 25], functional domain composition approach [J. Biol. Chem. 277 (2002) 45765], and the pseudo-amino acid composition approach [Proteins Struct. Funct. Genet. 43 (2001) 246; Erratum: ibid. 44 (2001) 60]. As a showcase, the recently constructed dataset [Bioinformatics 19 (2003) 1656] was used for demonstration. The dataset contains 7589 proteins classified into 12 subcellular locations: chloroplast, cytoplasmic, cytoskeleton, endoplasmic reticulum, extracellular, Golgi apparatus, lysosomal, mitochondrial, nuclear, peroxisomal, plasma membrane, and vacuolar. The overall success rate of prediction obtained by the jackknife cross-validation was 92%. This is so far the highest success rate performed on this dataset by following an objective and rigorous cross-validation procedure.
Simultaneous acquisition of EEG and NIRS during cognitive tasks for an open access dataset.
Shin, Jaeyoung; von Lühmann, Alexander; Kim, Do-Won; Mehnert, Jan; Hwang, Han-Jeong; Müller, Klaus-Robert
2018-02-13
We provide an open access multimodal brain-imaging dataset of simultaneous electroencephalography (EEG) and near-infrared spectroscopy (NIRS) recordings. Twenty-six healthy participants performed three cognitive tasks: 1) n-back (0-, 2- and 3-back), 2) discrimination/selection response task (DSR) and 3) word generation (WG) tasks. The data provided includes: 1) measured data, 2) demographic data, and 3) basic analysis results. For n-back (dataset A) and DSR tasks (dataset B), event-related potential (ERP) analysis was performed, and spatiotemporal characteristics and classification results for 'target' versus 'non-target' (dataset A) and symbol 'O' versus symbol 'X' (dataset B) are provided. Time-frequency analysis was performed to show the EEG spectral power to differentiate the task-relevant activations. Spatiotemporal characteristics of hemodynamic responses are also shown. For the WG task (dataset C), the EEG spectral power and spatiotemporal characteristics of hemodynamic responses are analyzed, and the potential merit of hybrid EEG-NIRS BCIs was validated with respect to classification accuracy. We expect that the dataset provided will facilitate performance evaluation and comparison of many neuroimaging analysis techniques.
Simultaneous acquisition of EEG and NIRS during cognitive tasks for an open access dataset
Shin, Jaeyoung; von Lühmann, Alexander; Kim, Do-Won; Mehnert, Jan; Hwang, Han-Jeong; Müller, Klaus-Robert
2018-01-01
We provide an open access multimodal brain-imaging dataset of simultaneous electroencephalography (EEG) and near-infrared spectroscopy (NIRS) recordings. Twenty-six healthy participants performed three cognitive tasks: 1) n-back (0-, 2- and 3-back), 2) discrimination/selection response task (DSR) and 3) word generation (WG) tasks. The data provided includes: 1) measured data, 2) demographic data, and 3) basic analysis results. For n-back (dataset A) and DSR tasks (dataset B), event-related potential (ERP) analysis was performed, and spatiotemporal characteristics and classification results for ‘target’ versus ‘non-target’ (dataset A) and symbol ‘O’ versus symbol ‘X’ (dataset B) are provided. Time-frequency analysis was performed to show the EEG spectral power to differentiate the task-relevant activations. Spatiotemporal characteristics of hemodynamic responses are also shown. For the WG task (dataset C), the EEG spectral power and spatiotemporal characteristics of hemodynamic responses are analyzed, and the potential merit of hybrid EEG-NIRS BCIs was validated with respect to classification accuracy. We expect that the dataset provided will facilitate performance evaluation and comparison of many neuroimaging analysis techniques. PMID:29437166
Real-time recording and classification of eye movements in an immersive virtual environment.
Diaz, Gabriel; Cooper, Joseph; Kit, Dmitry; Hayhoe, Mary
2013-10-10
Despite the growing popularity of virtual reality environments, few laboratories are equipped to investigate eye movements within these environments. This primer is intended to reduce the time and effort required to incorporate eye-tracking equipment into a virtual reality environment. We discuss issues related to the initial startup and provide algorithms necessary for basic analysis. Algorithms are provided for the calculation of gaze angle within a virtual world using a monocular eye-tracker in a three-dimensional environment. In addition, we provide algorithms for the calculation of the angular distance between the gaze and a relevant virtual object and for the identification of fixations, saccades, and pursuit eye movements. Finally, we provide tools that temporally synchronize gaze data and the visual stimulus and enable real-time assembly of a video-based record of the experiment using the Quicktime MOV format, available at http://sourceforge.net/p/utdvrlibraries/. This record contains the visual stimulus, the gaze cursor, and associated numerical data and can be used for data exportation, visual inspection, and validation of calculated gaze movements.
Real-time recording and classification of eye movements in an immersive virtual environment
Diaz, Gabriel; Cooper, Joseph; Kit, Dmitry; Hayhoe, Mary
2013-01-01
Despite the growing popularity of virtual reality environments, few laboratories are equipped to investigate eye movements within these environments. This primer is intended to reduce the time and effort required to incorporate eye-tracking equipment into a virtual reality environment. We discuss issues related to the initial startup and provide algorithms necessary for basic analysis. Algorithms are provided for the calculation of gaze angle within a virtual world using a monocular eye-tracker in a three-dimensional environment. In addition, we provide algorithms for the calculation of the angular distance between the gaze and a relevant virtual object and for the identification of fixations, saccades, and pursuit eye movements. Finally, we provide tools that temporally synchronize gaze data and the visual stimulus and enable real-time assembly of a video-based record of the experiment using the Quicktime MOV format, available at http://sourceforge.net/p/utdvrlibraries/. This record contains the visual stimulus, the gaze cursor, and associated numerical data and can be used for data exportation, visual inspection, and validation of calculated gaze movements. PMID:24113087
Nationwide evaluation of X-Ray trends (NEXT): eight years of data (1974-1981)
DOE Office of Scientific and Technical Information (OSTI.GOV)
Not Available
Production of radiologic images of acceptable diagnostic quality obtained with minimum radiation exposure to patients is a basic goal of the Center for Devices and Radiological Health (CDRH). The Office of Training and Assistance (OTA) of the CDRH conducts a number of educational programs to meet this goal. These programs have provided guidance to practices which promote the safe and effective use of radiation in the ordering, conduct, and interpretation of diagnostic radiology examinations. NEXT has provided much useful information in the past and is now undergoing revision to increase the specificity of the individual facility data as well asmore » to improve the statistical validity of the cumulative results in order to provide more accurate national indices of patient exposure and dose. In view of the coming changes in NEXT, it seemed particularly appropriate that a summary of past activities and findings should be provided to the state agencies whose activities have been essential to the success of the program and to other interested parties. This publication is intended to serve that purpose.« less
Validation of LOC-I interventions
DOT National Transportation Integrated Search
2012-08-13
The basic tenet of this paper is that todays national airspace systems, at least in advanced industrial countries, qualify as so-called Highly Reliable Systems (HRS). In an HRS, even the type of accident that causes the most fatalities is a rare e...
Development, Validation and Dissemination of Instructional Materials: An Analysis
ERIC Educational Resources Information Center
Creech, Jane C.; Arsham, G. M.
1976-01-01
The activities of SIMO (Self-Instructional Materials in Ophthalmology), a two-year project to develop four slide/tape units and one videocassette in basic ophthalmology for medical students, are described. Instructional effectiveness was documented using various measures. (Author/LBH)
1978-12-01
Poisson processes . The method is valid for Poisson processes with any given intensity function. The basic thinning algorithm is modified to exploit several refinements which reduce computer execution time by approximately one-third. The basic and modified thinning programs are compared with the Poisson decomposition and gap-statistics algorithm, which is easily implemented for Poisson processes with intensity functions of the form exp(a sub 0 + a sub 1t + a sub 2 t-squared. The thinning programs are competitive in both execution
Discussion of solutions to ethical issues in real-world study.
Wang, Sicheng; Liu, Baoyan; Xiong, Ningning; Xie, Qi; Zhang, Runshun; Zhou, Xuezhong; Qiao, Jie
2014-09-01
In recent years, the paradigm of real-world study (RWS) has been at the forefront of clinical research worldwide, particularly in the field of traditional Chinese medicine. In this paper, basic features and nature of real-world clinical studies are discussed, and ethical issues in different stages of RWS are raised and reviewed. Moreover, some preliminary solutions to these issues, such as protecting subjects during the process of RWS and performing ethical review, are presented based on recent practices and basic ethical rules to improve the scientific validity and ethical level of RWS.
Preliminary Face and Construct Validation Study of a Virtual Basic Laparoscopic Skill Trainer
Sankaranarayanan, Ganesh; Lin, Henry; Arikatla, Venkata S.; Mulcare, Maureen; Zhang, Likun; Derevianko, Alexandre; Lim, Robert; Fobert, David; Cao, Caroline; Schwaitzberg, Steven D.; Jones, Daniel B.
2010-01-01
Abstract Background The Virtual Basic Laparoscopic Skill Trainer (VBLaST™) is a developing virtual-reality–based surgical skill training system that incorporates several of the tasks of the Fundamentals of Laparoscopic Surgery (FLS) training system. This study aimed to evaluate the face and construct validity of the VBLaST™ system. Materials and Methods Thirty-nine subjects were voluntarily recruited at the Beth Israel Deaconess Medical Center (Boston, MA) and classified into two groups: experts (PGY 5, fellow and practicing surgeons) and novice (PGY 1–4). They were then asked to perform three FLS tasks, consisting of peg transfer, pattern cutting, and endoloop, on both the VBLaST and FLS systems. The VBLaST performance scores were automatically computed, while the FLS scores were rated by a trained evaluator. Face validity was assessed using a 5-point Likert scale, varying from not realistic/useful (1) to very realistic/useful (5). Results Face-validity scores showed that the VBLaST system was significantly realistic in portraying the three FLS tasks (3.95 ± 0.909), as well as the reality in trocar placement and tool movements (3.67 ± 0.874). Construct-validity results show that VBLaST was able to differentiate between the expert and novice group (P = 0.015). However, of the two tasks used for evaluating VBLaST, only the peg-transfer task showed a significant difference between the expert and novice groups (P = 0.003). Spearman correlation coefficient analysis between the two scores showed significant correlation for the peg-transfer task (Spearman coefficient 0.364; P = 0.023). Conclusions VBLaST demonstrated significant face and construct validity. A further set of studies, involving improvement to the current VBLaST system, is needed to thoroughly demonstrate face and construct validity for all the tasks. PMID:20201683
1988-05-01
ifforiable manpower investement. On the basis of our current experience it seems that the basic design principles are valid. The system developed will... system is operational on various computer networks, and in both industrial and in research environments. The design pri,lciples for the construction of...to a useful numerical simulation and design system for very complex configurations and flows. 7. REFERENCES 1. Bartlett G. W. , "An experimental
General Aviation Pilot Advisory and Training System (GAPATS)
NASA Technical Reports Server (NTRS)
Painter, John; Ward, Donald T.; Kelly, Wallace; Crump, John W.; Phillips, Ron; Trang, Jeff; Lee, Kris; Branham, Paul A.; Krishnamurthy, Karthik; Alcorn, William P., Jr.;
1997-01-01
The goal of this project is to achieve a validated General Aviation Pilot Advisor and Training System (GAPATS) engineering prototype, implemented according to commercial software standards and Federal Aviation Administration (FAA) issues of certification. Phase 2 builds on progress during Phase 1, which exceeded proposed objectives. The basic technology has been transferred from previous NASA research (1989 to 1994). We anticipate a commercially licensable prototype, validated by pilots in a flight simulator and in a light twin-engine research aircraft for FAA certification, by January 1998.
2012-01-01
numerical oil spill model validation showing the need for improvedmodel param- eterizations of basic oil spill processes (Cheng et al., 2010). 3.1.2...2004). Modelling the bidirectional reflectance distribution function ( BRDF ) of seawater polluted by an oil film. Optics Express, 12, 1671–1676. Pilon...display a currently valid OMB control number. 1. REPORT DATE 2012 2. REPORT TYPE 3. DATES COVERED 00-00-2012 to 00-00-2012 4. TITLE AND
Assessing the significance of pedobarographic signals using random field theory.
Pataky, Todd C
2008-08-07
Traditional pedobarographic statistical analyses are conducted over discrete regions. Recent studies have demonstrated that regionalization can corrupt pedobarographic field data through conflation when arbitrary dividing lines inappropriately delineate smooth field processes. An alternative is to register images such that homologous structures optimally overlap and then conduct statistical tests at each pixel to generate statistical parametric maps (SPMs). The significance of SPM processes may be assessed within the framework of random field theory (RFT). RFT is ideally suited to pedobarographic image analysis because its fundamental data unit is a lattice sampling of a smooth and continuous spatial field. To correct for the vast number of multiple comparisons inherent in such data, recent pedobarographic studies have employed a Bonferroni correction to retain a constant family-wise error rate. This approach unfortunately neglects the spatial correlation of neighbouring pixels, so provides an overly conservative (albeit valid) statistical threshold. RFT generally relaxes the threshold depending on field smoothness and on the geometry of the search area, but it also provides a framework for assigning p values to suprathreshold clusters based on their spatial extent. The current paper provides an overview of basic RFT concepts and uses simulated and experimental data to validate both RFT-relevant field smoothness estimations and RFT predictions regarding the topological characteristics of random pedobarographic fields. Finally, previously published experimental data are re-analysed using RFT inference procedures to demonstrate how RFT yields easily understandable statistical results that may be incorporated into routine clinical and laboratory analyses.
NASA Astrophysics Data System (ADS)
Anderson, T.
2015-12-01
The Northeast Fisheries Science Center's (NEFSC) Student Drifters Program is providing education opportunities for students of all ages. Using GPS-tracked ocean drifters, various educational institutions can provide students with hands-on experience in physical oceanography, engineering, and computer science. In building drifters many high school and undergraduate students may focus on drifter construction, sometimes designing their own drifter or attempting to improve current NEFSC models. While learning basic oceanography younger students can build drifters with the help of an educator and directions available on the studentdrifters.org website. Once drifters are deployed, often by a local mariner or oceanographic partner, drifter tracks can be visualised on maps provided at http://nefsc.noaa.gov/drifter. With the lesson plans available for those interested in computer science, students may download, process, and plot the drifter position data with basic Python code provided. Drifter tracks help students to visualize ocean currents, and also allow them to understand real particle tracking applications such as in search and rescue, oil spill dispersion, larval transport, and the movement of injured sea animals. Additionally, ocean circulation modelers can use student drifter paths to validate their models. The Student Drifters Program has worked with over 100 schools, several of them having deployed drifters on the West Coast. Funding for the program often comes from individual schools and small grants but in the future will preferably come from larger government grants. NSF, Sea-Grant, NOAA, and EPA are all possible sources of funding, especially with the support of multiple schools and large marine education associations. The Student Drifters Program is a unique resource for educators, students, and scientists alike.
Finite Element Analysis of Reverberation Chambers
NASA Technical Reports Server (NTRS)
Bunting, Charles F.; Nguyen, Duc T.
2000-01-01
The primary motivating factor behind the initiation of this work was to provide a deterministic means of establishing the validity of the statistical methods that are recommended for the determination of fields that interact in -an avionics system. The application of finite element analysis to reverberation chambers is the initial step required to establish a reasonable course of inquiry in this particularly data-intensive study. The use of computational electromagnetics provides a high degree of control of the "experimental" parameters that can be utilized in a simulation of reverberating structures. As the work evolved there were four primary focus areas they are: 1. The eigenvalue problem for the source free problem. 2. The development of a complex efficient eigensolver. 3. The application of a source for the TE and TM fields for statistical characterization. 4. The examination of shielding effectiveness in a reverberating environment. One early purpose of this work was to establish the utility of finite element techniques in the development of an extended low frequency statistical model for reverberation phenomena. By employing finite element techniques, structures of arbitrary complexity can be analyzed due to the use of triangular shape functions in the spatial discretization. The effects of both frequency stirring and mechanical stirring are presented. It is suggested that for the low frequency operation the typical tuner size is inadequate to provide a sufficiently random field and that frequency stirring should be used. The results of the finite element analysis of the reverberation chamber illustrate io-W the potential utility of a 2D representation for enhancing the basic statistical characteristics of the chamber when operating in a low frequency regime. The basic field statistics are verified for frequency stirring over a wide range of frequencies. Mechanical stirring is shown to provide an effective frequency deviation.
Lampis, Valentina; Maziade, Michel; Battaglia, Marco
2011-05-01
We are witnessing a tremendous expansion of strategies and techniques that derive from basic and preclinical science to study the fine genetic, epigenetic, and proteomic regulation of behavior in the laboratory animal. In this endeavor, animal models of psychiatric illness are becoming the almost exclusive domain of basic researchers, with lesser involvement of clinician researchers in their conceptual design, and transfer into practice of new paradigms. From the side of human behavioral research, the growing interest in gene-environment interplay and the fostering of valid endophenotypes are among the few substantial innovations in the effort of linking common mental disorders to cutting-edge clinical research questions. We argue that it is time for cross-fertilization between these camps. In this article, we a) observe that the "translational divide" can-and should-be crossed by having investigators from both the basic and the clinical sides cowork on simpler, valid "endophenotypes" of neurodevelopmental relevance; b) emphasize the importance of unambiguous physiological readouts, more than behavioral equivalents of human symptoms/syndromes, for animal research; c) indicate and discuss how this could be fostered and implemented in a developmental framework of reference for some common anxiety disorders and ultimately lead to better animal models of human mental disorders.
Basic symptoms, temperament and character in adolescent psychiatric disorders.
Poustka, L; Parzer, P; Brunner, R; Resch, F
2007-01-01
Basic symptoms are early subtle changes in thinking, feeling and perception that are subjectively experienced and precede the onset of a psychotic illness. In adult samples, high basic symptom scores are regarded as specific risk factors for the development of schizophrenia. The aim of this study was to explore the relevance of basic symptoms of psychiatric disorders in adolescent patients with special focus on early detection of psychosis. Furthermore, the association between basic symptoms and personality traits has been investigated. From 89 adolescents, who were consecutive inpatients with different psychiatric disorders in 1995 and 1997, 54 were followed up 4.7 years later. Patients were examined with the Bonn Scale for the Assessment of Basic Symptoms at the time of the first presentation and follow-up. Additionally, personality traits were assessed at follow-up using the Junior Temperament and Character Inventory. None of the subjects made the transition to schizophrenia, despite high baseline scores of basic symptoms at the initial assessment. Indirect minus symptoms were found to be the most valid predictor of a persisting psychiatric diagnosis. In addition, this specific category was strongly associated with the personality traits harm avoidance and self-directedness. In adolescents, basic symptoms in association with personality traits present as a nonspecific indicator of psychopathology rather than as an indicator of vulnerability to schizophrenia. Copyright (c) 2007 S. Karger AG, Basel.
Dewitt, James; Capistrant, Benjamin; Kohli, Nidhi; Mitteldorf, Darryl; Merengwa, Enyinnaya; West, William
2018-01-01
Background While deduplication and cross-validation protocols have been recommended for large Web-based studies, protocols for survey response validation of smaller studies have not been published. Objective This paper reports the challenges of survey validation inherent in a small Web-based health survey research. Methods The subject population was North American, gay and bisexual, prostate cancer survivors, who represent an under-researched, hidden, difficult-to-recruit, minority-within-a-minority population. In 2015-2016, advertising on a large Web-based cancer survivor support network, using email and social media, yielded 478 completed surveys. Results Our manual deduplication and cross-validation protocol identified 289 survey submissions (289/478, 60.4%) as likely spam, most stemming from advertising on social media. The basic components of this deduplication and validation protocol are detailed. An unexpected challenge encountered was invalid survey responses evolving across the study period. This necessitated the static detection protocol be augmented with a dynamic one. Conclusions Five recommendations for validation of Web-based samples, especially with smaller difficult-to-recruit populations, are detailed. PMID:29691203
Communication Competence among Teachers: The Ohio Solution.
ERIC Educational Resources Information Center
DeWine, Sue; Pearson, Judy C.
1989-01-01
Describes a summer instructional program for elementary and secondary teachers involving a set of workshops on basic communication concepts, nonverbal communication, organizational communication, listening, and small group communication. Discusses issues of marketing the program, defending its academic validity, and decreasing the gap between the…
DOE Office of Scientific and Technical Information (OSTI.GOV)
Ricci, Paolo; Theiler, C.; Fasoli, A.
A methodology for plasma turbulence code validation is discussed, focusing on quantitative assessment of the agreement between experiments and simulations. The present work extends the analysis carried out in a previous paper [P. Ricci et al., Phys. Plasmas 16, 055703 (2009)] where the validation observables were introduced. Here, it is discussed how to quantify the agreement between experiments and simulations with respect to each observable, how to define a metric to evaluate this agreement globally, and - finally - how to assess the quality of a validation procedure. The methodology is then applied to the simulation of the basic plasmamore » physics experiment TORPEX [A. Fasoli et al., Phys. Plasmas 13, 055902 (2006)], considering both two-dimensional and three-dimensional simulation models.« less
On the validation of a code and a turbulence model appropriate to circulation control airfoils
NASA Technical Reports Server (NTRS)
Viegas, J. R.; Rubesin, M. W.; Maccormack, R. W.
1988-01-01
A computer code for calculating flow about a circulation control airfoil within a wind tunnel test section has been developed. This code is being validated for eventual use as an aid to design such airfoils. The concept of code validation being used is explained. The initial stages of the process have been accomplished. The present code has been applied to a low-subsonic, 2-D flow about a circulation control airfoil for which extensive data exist. Two basic turbulence models and variants thereof have been successfully introduced into the algorithm, the Baldwin-Lomax algebraic and the Jones-Launder two-equation models of turbulence. The variants include adding a history of the jet development for the algebraic model and adding streamwise curvature effects for both models. Numerical difficulties and difficulties in the validation process are discussed. Turbulence model and code improvements to proceed with the validation process are also discussed.
Reinventing Biostatistics Education for Basic Scientists
Weissgerber, Tracey L.; Garovic, Vesna D.; Milin-Lazovic, Jelena S.; Winham, Stacey J.; Obradovic, Zoran; Trzeciakowski, Jerome P.; Milic, Natasa M.
2016-01-01
Numerous studies demonstrating that statistical errors are common in basic science publications have led to calls to improve statistical training for basic scientists. In this article, we sought to evaluate statistical requirements for PhD training and to identify opportunities for improving biostatistics education in the basic sciences. We provide recommendations for improving statistics training for basic biomedical scientists, including: 1. Encouraging departments to require statistics training, 2. Tailoring coursework to the students’ fields of research, and 3. Developing tools and strategies to promote education and dissemination of statistical knowledge. We also provide a list of statistical considerations that should be addressed in statistics education for basic scientists. PMID:27058055
Extended score interval in the assessment of basic surgical skills.
Acosta, Stefan; Sevonius, Dan; Beckman, Anders
2015-01-01
The Basic Surgical Skills course uses an assessment score interval of 0-3. An extended score interval, 1-6, was proposed by the Swedish steering committee of the course. The aim of this study was to analyze the trainee scores in the current 0-3 scored version compared to a proposed 1-6 scored version. Sixteen participants, seven females and nine males, were evaluated in the current and proposed assessment forms by instructors, observers, and learners themselves during the first and second day. In each assessment form, 17 tasks were assessed. The inter-rater reliability between the current and the proposed score sheets were evaluated with intraclass correlation (ICC) with 95% confidence intervals (CI). The distribution of scores for 'knot tying' at the last time point and 'bowel anastomosis side to side' given by the instructors in the current assessment form showed that the highest score was given in 31 and 62%, respectively. No ceiling effects were found in the proposed assessment form. The overall ICC between the current and proposed score sheets after assessment by the instructors increased from 0.38 (95% CI 0.77-0.78) on Day 1 to 0.83 (95% CI 0.51-0.94) on Day 2. A clear ceiling effect of scores was demonstrated in the current assessment form, questioning its validity. The proposed score sheet provides more accurate scores and seems to be a better feedback instrument for learning technical surgical skills in the Basic Surgical Skills course.
Extended score interval in the assessment of basic surgical skills.
Acosta, Stefan; Sevonius, Dan; Beckman, Anders
2015-01-01
Introduction The Basic Surgical Skills course uses an assessment score interval of 0-3. An extended score interval, 1-6, was proposed by the Swedish steering committee of the course. The aim of this study was to analyze the trainee scores in the current 0-3 scored version compared to a proposed 1-6 scored version. Methods Sixteen participants, seven females and nine males, were evaluated in the current and proposed assessment forms by instructors, observers, and learners themselves during the first and second day. In each assessment form, 17 tasks were assessed. The inter-rater reliability between the current and the proposed score sheets were evaluated with intraclass correlation (ICC) with 95% confidence intervals (CI). Results The distribution of scores for 'knot tying' at the last time point and 'bowel anastomosis side to side' given by the instructors in the current assessment form showed that the highest score was given in 31 and 62%, respectively. No ceiling effects were found in the proposed assessment form. The overall ICC between the current and proposed score sheets after assessment by the instructors increased from 0.38 (95% CI 0.77-0.78) on Day 1 to 0.83 (95% CI 0.51-0.94) on Day 2. Discussion A clear ceiling effect of scores was demonstrated in the current assessment form, questioning its validity. The proposed score sheet provides more accurate scores and seems to be a better feedback instrument for learning technical surgical skills in the Basic Surgical Skills course.
Sun, Xiaojun; Guo, Zhimou; Yu, Mengqi; Lin, Chao; Sheng, Anran; Wang, Zhiyu; Linhardt, Robert J; Chi, Lianli
2017-01-06
Low molecular weight heparins (LMWHs) are important anticoagulant drugs that are prepared through depolymerization of unfractionated heparin. Based on the types of processing reactions and the structures of the products, LMWHs can be divided into different classifications. Enoxaparin is prepared by benzyl esterification and alkaline depolymerization, while dalteparin and nadroparin are prepared through nitrous acid depolymerization followed by borohydride reduction. Compositional analysis of their basic building blocks is an effective way to provide structural information on heparin and LMWHs. However, most current compositional analysis methods have been limited to heparin and enoxaparin. A sensitive and comprehensive approach is needed for detailed investigation of the structure of LMWHs prepared through nitrous acid depolymerization, especially their characteristic saturated non-reducing end (NRE) and 2,5-anhydro-d-mannitol reducing end (RE). A maltose modified hydrophilic interaction column offers improved separation of complicated mixtures of acidic disaccharides and oligosaccharides. A total of 36 basic building blocks were unambiguously identified by high-resolution tandem mass spectrometry (MS). Multiple reaction monitoring (MRM) MS/MS quantification was developed and validated in the analysis of dalteparin and nadroparin samples. Each group of building blocks revealed different aspects of the properties of LMWHs, such as functional motifs required for anticoagulant activity, the structure of heparin starting materials, cleavage sites in the depolymerization reaction, and undesired structural modifications resulting from side reactions. Copyright © 2016 Elsevier B.V. All rights reserved.
A novel simulation theory and model system for multi-field coupling pipe-flow system
NASA Astrophysics Data System (ADS)
Chen, Yang; Jiang, Fan; Cai, Guobiao; Xu, Xu
2017-09-01
Due to the lack of a theoretical basis for multi-field coupling in many system-level models, a novel set of system-level basic equations for flow/heat transfer/combustion coupling is put forward. Then a finite volume model of quasi-1D transient flow field for multi-species compressible variable-cross-section pipe flow is established by discretising the basic equations on spatially staggered grids. Combining with the 2D axisymmetric model for pipe-wall temperature field and specific chemical reaction mechanisms, a finite volume model system is established; a set of specific calculation methods suitable for multi-field coupling system-level research is structured for various parameters in this model; specific modularisation simulation models can be further derived in accordance with specific structures of various typical components in a liquid propulsion system. This novel system can also be used to derive two sub-systems: a flow/heat transfer two-field coupling pipe-flow model system without chemical reaction and species diffusion; and a chemical equilibrium thermodynamic calculation-based multi-field coupling system. The applicability and accuracy of two sub-systems have been verified through a series of dynamic modelling and simulations in earlier studies. The validity of this system is verified in an air-hydrogen combustion sample system. The basic equations and the model system provide a unified universal theory and numerical system for modelling and simulation and even virtual testing of various pipeline systems.
Contribution of NIH funding to new drug approvals 2010–2016
Beierlein, Jennifer M.; Khanuja, Navleen Surjit; McNamee, Laura M.; Ledley, Fred D.
2018-01-01
This work examines the contribution of NIH funding to published research associated with 210 new molecular entities (NMEs) approved by the Food and Drug Administration from 2010–2016. We identified >2 million publications in PubMed related to the 210 NMEs (n = 131,092) or their 151 known biological targets (n = 1,966,281). Of these, >600,000 (29%) were associated with NIH-funded projects in RePORTER. This funding included >200,000 fiscal years of NIH project support (1985–2016) and project costs >$100 billion (2000–2016), representing ∼20% of the NIH budget over this period. NIH funding contributed to every one of the NMEs approved from 2010–2016 and was focused primarily on the drug targets rather than on the NMEs themselves. There were 84 first-in-class products approved in this interval, associated with >$64 billion of NIH-funded projects. The percentage of fiscal years of project funding identified through target searches, but not drug searches, was greater for NMEs discovered through targeted screening than through phenotypic methods (95% versus 82%). For targeted NMEs, funding related to targets preceded funding related to the NMEs, consistent with the expectation that basic research provides validated targets for targeted screening. This analysis, which captures basic research on biological targets as well as applied research on NMEs, suggests that the NIH contribution to research associated with new drug approvals is greater than previously appreciated and highlights the risk of reducing federal funding for basic biomedical research. PMID:29440428
NASA Astrophysics Data System (ADS)
Schreiner, L. J.
2017-05-01
For seventeen years a community of basic and clinical scientists and researchers has been meeting bi-annually to promote the clinical advance of techniques to measure radiation dose in three dimensions. The interest in this dosimetry was motivated by its promise as an effective methodology for 3D measurement of the complex conformal dose distributions achieved by modern techniques such as Intensity Modulated and Volumetric Arc Radiation Therapy. Each of the International Conferences on 3D Radiation Dosimetry resulted in the publication of informative proceedings [1-8], the majority openly available on the internet. The proceedings included papers that: i) reviewed the basic science of the radiation sensitive materials used to accumulate the dose information, ii) introduced the science and engineering of the imaging systems required to read the information out, iii) described the work flows and systems required for efficient dosimetry, iv) reported the protocols required for reproducible dosimetry, and v) showed examples of clinical use illustrating advantage and limitations of the dosimetry. This paper is intended to use the framework provided by these proceedings to review the current 3D chemical dosimeters available and to discuss the requirements for their use. The paper describes how 3D dosimetry can complement other dose delivery validation approaches available in the clinic. It closes with some personal reflections of how the motivation for, and practice of, 3D dosimetry have changed (or not) over the years.
Fahey, Marion; Rudd, Anthony; Béjot, Yannick; Wolfe, Charles; Douiri, Abdel
2017-01-01
Introduction Stroke is a leading cause of adult disability and death worldwide. The neurological impairments associated with stroke prevent patients from performing basic daily activities and have enormous impact on families and caregivers. Practical and accurate tools to assist in predicting outcome after stroke at patient level can provide significant aid for patient management. Furthermore, prediction models of this kind can be useful for clinical research, health economics, policymaking and clinical decision support. Methods 2869 patients with first-ever stroke from South London Stroke Register (SLSR) (1995–2004) will be included in the development cohort. We will use information captured after baseline to construct multilevel models and a Cox proportional hazard model to predict cognitive impairment, functional outcome and mortality up to 5 years after stroke. Repeated random subsampling validation (Monte Carlo cross-validation) will be evaluated in model development. Data from participants recruited to the stroke register (2005–2014) will be used for temporal validation of the models. Data from participants recruited to the Dijon Stroke Register (1985–2015) will be used for external validation. Discrimination, calibration and clinical utility of the models will be presented. Ethics Patients, or for patients who cannot consent their relatives, gave written informed consent to participate in stroke-related studies within the SLSR. The SLSR design was approved by the ethics committees of Guy’s and St Thomas’ NHS Foundation Trust, Kings College Hospital, Queens Square and Westminster Hospitals (London). The Dijon Stroke Registry was approved by the Comité National des Registres and the InVS and has authorisation of the Commission Nationale de l’Informatique et des Libertés. PMID:28821511
Sánchez-Pérez, Noelia; Fuentes, Luis J; Jolliffe, Darrick; González-Salinas, Carmen
2014-01-01
The aim of the current research was to study cognitive and affective empathy in children aged 6-12 years old, and their associations with children's family environment and social adjustment. For this purpose, we developed the Spanish version of the Basic Empathy Scale (BES), self- and parent-report forms. Factorial analyses confirmed a two-component model of empathy in both self- and parent-report forms. Concordance between parent-child measures of empathy was low for cognitive and affective factors. Analyses of variance on the cognitive and affective components brought a significant effect of age for self-reported cognitive empathy, with older children scoring higher than younger ones. Gender brought out a significant principal effect for self-reported affective empathy, with girls scoring higher than boys. No other main effects were found for age and gender for the rest of the factors analyzed. Children's empathy was associated with socioeconomic status and other family socialization processes, as well as children' social behaviors. Overall the new measures provided a coherent view of empathy in middle childhood and early adolescence when measured through self and parent reports, and illustrate the similarity of the validity of the BES in a European-Spanish culture.
NASA Astrophysics Data System (ADS)
Lesiuk, Michał; Moszynski, Robert
2014-12-01
In this paper we consider the calculation of two-center exchange integrals over Slater-type orbitals (STOs). We apply the Neumann expansion of the Coulomb interaction potential and consider calculation of all basic quantities which appear in the resulting expression. Analytical closed-form equations for all auxiliary quantities have already been known but they suffer from large digital erosion when some of the parameters are large or small. We derive two differential equations which are obeyed by the most difficult basic integrals. Taking them as a starting point, useful series expansions for small parameter values or asymptotic expansions for large parameter values are systematically derived. The resulting expansions replace the corresponding analytical expressions when the latter introduce significant cancellations. Additionally, we reconsider numerical integration of some necessary quantities and present a new way to calculate the integrand with a controlled precision. All proposed methods are combined to lead to a general, stable algorithm. We perform extensive numerical tests of the introduced expressions to verify their validity and usefulness. Advances reported here provide methodology to compute two-electron exchange integrals over STOs for a broad range of the nonlinear parameters and large angular momenta.
A Navier-Stokes-Based Approach for Mean Flow Perturbation Analysis
NASA Astrophysics Data System (ADS)
Bhaumik, Swagata; Gaitonde, Datta; Waindim, Mbu; The Ohio State University Team
2014-11-01
The manner in which a basic state, obtained from a time-averaged unsteady flowfield, processes perturbations can provide significant insight into the cause and evolution of instabilities. A widely used approach is based on Parabolized Stability Equations (PSE), which limits streamwise mean flow variation and is often applied to 2-D base flows. To avoid some of these issues, we advance a Navier-Stokes-based method, which can address non-trivial three-dimensional fields. The method stems from that employed by Touber and Sandham (Theor. Comput. Fluid. Dyn., 23, 79, 2009) to analyze global modes in nominally 2-D shock-wave turbulent-boundary layer interactions (STBLI). We first develop its theoretical underpinnings by examining conditions under which it degenerates to traditional methods. We then illustrate the application by considering perturbations to an entropy layer at Mach 6, a turbulent supersonic jet at Mach 1.3 and STBLI at Mach 2.3. For the entropy layer and jet cases, known linear stability and PSE results are successfully reproduced, while global modes are obtained for STBLI. The results not only validate the proposed technique, but also demonstrate its suitability in analyzing instabilities for any general 3D basic state, including impulse response. Sponsored by AFOSR.
Bridging the Gap Between Research and Practice: Implementation Science.
Olswang, Lesley B; Prelock, Patricia A
2015-12-01
This article introduces implementation science, which focuses on research methods that promote the systematic application of research findings to practice. The narrative defines implementation science and highlights the importance of moving research along the pipeline from basic science to practice as one way to facilitate evidence-based service delivery. This review identifies challenges in developing and testing interventions in order to achieve widespread adoption in practice settings. A framework for conceptualizing implementation research is provided, including an example to illustrate the application of principles in speech-language pathology. Last, the authors reflect on the status of implementation research in the discipline of communication sciences and disorders. The extant literature highlights the value of implementation science for reducing the gap between research and practice in our discipline. While having unique principles guiding implementation research, many of the challenges and questions are similar to those facing any investigators who are attempting to design valid and reliable studies. This article is intended to invigorate interest in the uniqueness of implementation science among those pursuing both basic and applied research. In this way, it should help ensure the discipline's knowledge base is realized in practice and policy that affects the lives of individuals with communication disorders.
Basic Skills Support in Business and Industry.
ERIC Educational Resources Information Center
Byatt, Janet; Davies, Karen
This guide is designed as a tool for English and Welsh businesses wanting to provide basic skills training for their employees. It provides practical solutions to the problems of identifying employees' basic skills needs and selecting the best model of training delivery to address identified training needs. The introductory section discusses basic…
Kinematic Characterization of Left Ventricular Chamber Stiffness and Relaxation
NASA Astrophysics Data System (ADS)
Mossahebi, Sina
Heart failure is the most common cause of hospitalization today, and diastolic heart failure accounts for 40-50% of cases. Therefore, it is critical to identify diastolic dysfunction at a subclinical stage so that appropriate therapy can be administered before ventricular function is further, and perhaps irreversibly impaired. Basic concepts in physics such as kinematic modeling provide a unique method with which to characterize cardiovascular physiology, specifically diastolic function (DF). The advantage of an approach that is standard in physics, such as the kinematic modeling is its causal formulation that functions in contrast to correlative approaches traditionally utilized in the life sciences. Our research group has pioneered theoretical and experimental quantitative analysis of DF in humans, using both non-invasive (echocardiography, cardiac MRI) and invasive (simultaneous catheterization-echocardiography) methods. Our group developed and validated the Parametrized Diastolic Filling (PDF) formalism which is motivated by basic physiologic principles (LV is a mechanical suction pump at the mitral valve opening) that obey Newton's Laws. PDF formalism is a kinematic model of filling employing an equation of motion, the solution of which accurately predicts all E-wave contours in accordance with the rules of damped harmonic oscillatory motion. The equation's lumped parameters---ventricular stiffness, ventricular viscoelasticity/relaxation and ventricular load---are obtained by solving the 'inverse problem'. The parameters' physiologic significance and clinical utility have been repeatedly demonstrated in multiple clinical settings. In this work we apply our kinematic modeling approach to better understand how the heart works as it fills in order to advance the relationship between physiology and mathematical modeling. Through the use of this modeling, we thereby define and validate novel, causal indexes of diastolic function such as early rapid filling energy, diastatic stiffness, and relaxation and stiffness components of E-wave deceleration time.
Gillespie-Bennett, Julie; Keall, Michael; Howden-Chapman, Philippa; Baker, Michael G
2013-08-02
Substandard housing is a problem in New Zealand. Historically there has been little recognition of the important aspects of housing quality that affect people's health and safety. In this viewpoint article we outline the importance of assessing these factors as an essential step to improving the health and safety of New Zealanders and household energy efficiency. A practical risk assessment tool adapted to New Zealand conditions, the Healthy Housing Index (HHI), measures the physical characteristics of houses that affect the health and safety of the occupants. This instrument is also the only tool that has been validated against health and safety outcomes and reported in the international peer-reviewed literature. The HHI provides a framework on which a housing warrant of fitness (WOF) can be based. The HHI inspection takes about one hour to conduct and is performed by a trained building inspector. To maximise the effectiveness of this housing quality assessment we envisage the output having two parts. The first would be a pass/fail WOF assessment showing whether or not the house meets basic health, safety and energy efficiency standards. The second component would rate each main assessment area (health, safety and energy efficiency), potentially on a five-point scale. This WOF system would establish a good minimum standard for rental accommodation as well encouraging improved housing performance over time. In this article we argue that the HHI is an important, validated, housing assessment tool that will improve housing quality, leading to better health of the occupants, reduced home injuries, and greater energy efficiency. If required, this tool could be extended to also cover resilience to natural hazards, broader aspects of sustainability, and the suitability of the dwelling for occupants with particular needs.
New, P W; Currie, K E
2016-08-01
Questionnaire development, validation and completion. Develop comprehensive survey of sexuality issues including validated self-report versions of the International Spinal Cord Injury male sexual function and female sexual and reproductive function basic data sets (SR-iSCI-sexual function). People with spinal cord damage (SCD) living in the community, Australia from August 2013 to June 2014. An iterative process involving rehabilitation medicine clinicians, a nurse specialising in sexuality issues in SCD and people with SCD who developed a comprehensive survey that included the SR-iSCI-sexual function. Participants recruitment through spinal rehabilitation review clinic and community organisations that support people with SCD. Surveys completed by 154 people. Most were male (n=101, 65.6%). Respondents' median age was 50 years (interquartile range (IQR) 38-58), and they were a median of 10 years (IQR 4-20) after the onset of SCD. Sexual problems unrelated to SCD were reported by 12 (8%) respondents, and 114 (n=75.5%) reported sexual problems because of SCD. Orgasms were much less likely (χ(2)=13.1, P=0.006) to be normal in males (n=5, 5%) compared with females (n=11, 22%). Males had significantly worse (χ(2)=26.0, P=0.001) psychogenic genital functioning (normal n=9, 9%) than females (normal n=13, 26%) and worse (χ(2)=10.8, P=0.013) reflex genital functioning. Normal ejaculation was reported in only three (3%) men. Most (n=26, 52%) women reported reduced or absent menstruation pattern since SCD. The SR-iSCI-sexual function provides a useful tool for researchers and clinicians to collect information regarding patient-reported sexual functioning after SCD and to facilitate comparative studies.
An update on SCARLET hardware development and flight programs
DOE Office of Scientific and Technical Information (OSTI.GOV)
Jones, P.A.; Murphy, D.M.; Piszczor, M.F.
1995-10-01
Solar Concentrator Array with Refractive Linear Element Technology (SCARLET) is one of the first practical photovoltaic concentrator array technologies that offers a number of benefits for space applications (i.e. high array efficiency, protection from space radiation effects, a relatively light weight system, minimized plasma interactions, etc.) The line-focus concentrator concept, however, also offers two very important advantages: (1) low-cost mass production potential of the lens material; and (2) relaxation of precise array tracking requirements to only a single axis. These benefits offer unique capabilities to both commercial and government spacecraft users, specifically those interested in high radiation missions, such asmore » MEO orbits, and electric-powered propulsion LEO-to-GEO orbit raising applications. SCARLET is an aggressive hardware development and flight validation program sponsored by the Ballistic Missile Defense Organization (BMDO) and NASA Lewis Research Center. Its intent is to bring technology to the level of performance and validation necessary for use by various government and commercial programs. The first phase of the SCARLET program culminated with the design, development and fabrication of a small concentrator array for flight on the METEOR satellite. This hardware will be the first in-space demonstration of concentrator technology at the `array level` and will provide valuable in-orbit performance measurements. The METEOR satellite is currently planned for a September/October 1995 launch. The next phase of the program is the development of large array for use by one of the NASA New Millenium Program missions. This hardware will incorporate a number of the significant improvements over the basic METEOR design. This presentation will address the basic SCARLET technology, examine its benefits to users, and describe the expected improvements for future missions.« less
Solar Pumped Solid State Lasers for Space Solar Power: Experimental Path
NASA Technical Reports Server (NTRS)
Fork, Richard L.; Carrington, Connie K.; Walker, Wesley W.; Cole, Spencer T.; Green, Jason J. A.; Laycock, Rustin L.
2003-01-01
We outline an experimentally based strategy designed to lead to solar pumped solid state laser oscillators useful for space solar power. Our method involves solar pumping a novel solid state gain element specifically designed to provide efficient conversion of sunlight in space to coherent laser light. Kilowatt and higher average power is sought from each gain element. Multiple such modular gain elements can be used to accumulate total average power of interest for power beaming in space, e.g., 100 kilowatts and more. Where desirable the high average power can also be produced as a train of pulses having high peak power (e.g., greater than 10(exp 10 watts). The modular nature of the basic gain element supports an experimental strategy in which the core technology can be validated by experiments on a single gain element. We propose to do this experimental validation both in terrestrial locations and also on a smaller scale in space. We describe a terrestrial experiment that includes diagnostics and the option of locating the laser beam path in vacuum environment. We describe a space based experiment designed to be compatible with the Japanese Experimental Module (JEM) on the International Space Station (ISS). We anticipate the gain elements will be based on low temperature (approx. 100 degrees Kelvin) operation of high thermal conductivity (k approx. 100 W/cm-K) diamond and sapphire (k approx. 4 W/cm-K). The basic gain element will be formed by sequences of thin alternating layers of diamond and Ti:sapphire with special attention given to the material interfaces. We anticipate this strategy will lead to a particularly simple, robust, and easily maintained low mass modelocked multi-element laser oscillator useful for space solar power.
Dimitriadis, Alexis; Palmer, Antony L; Thomas, Russell A S; Nisbet, Andrew; Clark, Catharine H
2017-06-01
To adapt and validate an anthropomorphic head phantom for use in a cranial radiosurgery audit. Two bespoke inserts were produced for the phantom: one for providing the target and organ at risk for delineation and the other for performing dose measurements. The inserts were tested to assess their positional accuracy. A basic treatment plan dose verification with an ionization chamber was performed to establish a baseline accuracy for the phantom and beam model. The phantom and inserts were then used to perform dose verification measurements of a radiosurgery plan. The dose was measured with alanine pellets, EBT extended dose film and a plastic scintillation detector (PSD). Both inserts showed reproducible positioning (±0.5 mm) and good positional agreement between them (±0.6 mm). The basic treatment plan measurements showed agreement to the treatment planning system (TPS) within 0.5%. Repeated film measurements showed consistent gamma passing rates with good agreement to the TPS. For 2%-2 mm global gamma, the mean passing rate was 96.7% and the variation in passing rates did not exceed 2.1%. The alanine pellets and PSD showed good agreement with the TPS (-0.1% and 0.3% dose difference in the target) and good agreement with each other (within 1%). The adaptations to the phantom showed acceptable accuracies. The presence of alanine and PSD do not affect film measurements significantly, enabling simultaneous measurements by all three detectors. Advances in knowledge: A novel method for thorough end-to-end test of radiosurgery, with capability to incorporate all steps of the clinical pathway in a time-efficient and reproducible manner, suitable for a national audit.
ERIC Educational Resources Information Center
Ballard, W.L.
1968-01-01
The article discusses models of synchronic and diachronic phonology and suggests changes in them. The basic generative model of phonology is outlined with the author's reinterpretations. The systematic phonemic level is questioned in terms of its unreality with respect to linguistic performance and its lack of validity with respect to historical…
ERIC Educational Resources Information Center
Szymanski, Theodore
1999-01-01
Discusses a common misunderstanding demonstrated by many students in basic mathematics courses: not knowing how to properly "cancel" factors in simplifying mathematical equations. Asserts that "crossing-out" or "canceling" is not a valid mathematical operation, and that instructors should be wary about using these terms because of the ease with…
NASA Astrophysics Data System (ADS)
Brustein, R.
I review some basic facts about entropy bounds in general and about cosmological entropy bounds. Then I review the causal entropy bound, the conditions for its validity and its application to the study of cosmological singularities. This article is based on joint work with Gabriele Veneziano and subsequent related research.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Curceanu, C.; Bragadireanu, M.; Sirghi, D.
The Pauli Exclusion Principle (PEP) is one of the basic principles of modern physics and, even if there are no compelling reasons to doubt its validity, it is still debated today because an intuitive, elementary explanation is still missing, and because of its unique stand among the basic symmetries of physics. We present an experimental test of the validity of the Pauli Exclusion Principle for electrons based on a straightforward idea put forward a few years ago by Ramberg and Snow (E. Ramberg and G. A. Snow 1990 Phys. Lett. B 238 438). We performed a very accurate search ofmore » X-rays from the Pauli-forbidden atomic transitions of electrons in the already filled 1S shells of copper atoms. Although the experiment has a very simple structure, it poses deep conceptual and interpretational problems. Here we describe the experimental method and recent experimental results interpreted as an upper limit for the probability to violate the Pauli Exclusion Principle. We also present future plans to upgrade the experimental apparatus.« less
Verloock, Leen; Joseph, Wout; Gati, Azeddine; Varsier, Nadège; Flach, Björn; Wiart, Joe; Martens, Luc
2013-06-01
An experimental validation of a low-cost method for extrapolation and estimation of the maximal electromagnetic-field exposure from long-term evolution (LTE) radio base station installations are presented. No knowledge on downlink band occupation or service characteristics is required for the low-cost method. The method is applicable in situ. It only requires a basic spectrum analyser with appropriate field probes without the need of expensive dedicated LTE decoders. The method is validated both in laboratory and in situ, for a single-input single-output antenna LTE system and a 2×2 multiple-input multiple-output system, with low deviations in comparison with signals measured using dedicated LTE decoders.
NASA Technical Reports Server (NTRS)
Price J. M.; Ortega, R.
1998-01-01
Probabilistic method is not a universally accepted approach for the design and analysis of aerospace structures. The validity of this approach must be demonstrated to encourage its acceptance as it viable design and analysis tool to estimate structural reliability. The objective of this Study is to develop a well characterized finite population of similar aerospace structures that can be used to (1) validate probabilistic codes, (2) demonstrate the basic principles behind probabilistic methods, (3) formulate general guidelines for characterization of material drivers (such as elastic modulus) when limited data is available, and (4) investigate how the drivers affect the results of sensitivity analysis at the component/failure mode level.
Validity and reliability of the Japanese version of the Newest Vital Sign: a preliminary study.
Kogure, Takamichi; Sumitani, Masahiko; Suka, Machi; Ishikawa, Hirono; Odajima, Takeshi; Igarashi, Ataru; Kusama, Makiko; Okamoto, Masako; Sugimori, Hiroki; Kawahara, Kazuo
2014-01-01
Health literacy (HL) refers to the ability to obtain, process, and understand basic health information and services, and is thus needed to make appropriate health decisions. The Newest Vital Sign (NVS) is comprised of 6 questions about an ice cream nutrition label and assesses HL numeracy skills. We developed a Japanese version of the NVS (NVS-J) and evaluated the validity and reliability of the NVS-J in patients with chronic pain. The translation of the original NVS into Japanese was achieved as per the published guidelines. An observational study was subsequently performed to evaluate the validity and reliability of the NVS-J in 43 Japanese patients suffering from chronic pain. Factor analysis with promax rotation, using the Kaiser criterion (eigenvalues ≥1.0), and a scree plot revealed that the main component of the NVS-J consists of three determinative factors, and each factor consists of two NVS-J items. The criterion-related validity of the total NVS-J score was significantly correlated with the total score of Ishikawa et al.'s self-rated HL Questionnaire, the clinical global assessment of comprehensive HL level, cognitive function, and the Brinkman index. In addition, Cronbach's coefficient for the total score of the NVS-J was adequate (alpha = 0.72). This study demonstrated that the NVS-J has good validity and reliability. Further, the NVS-J consists of three determinative factors: "basic numeracy ability," "complex numeracy ability," and "serious-minded ability." These three HL abilities comprise a 3-step hierarchical structure. Adequate HL should be promoted in chronic pain patients to enable coping, improve functioning, and increase activities of daily living (ADLs) and quality of life (QOL).
NASA Technical Reports Server (NTRS)
Hilsenrath, E.; Schoeberl, M.; Douglass, A.; Anderson, J.; Bhartia, P. K. (Technical Monitor)
2002-01-01
The EOS-Aura Mission is designed to answer three basic questions concerning the Earth's atmosphere: 1) Is ozone recovering as predicted, 2) is air quality getting worse, and 3) how is climate changing? Aura's four instruments work synergistically and are dedicated to answering these questions. These questions relate to NASA Earth Science Enterprise's overall strategic questions, which seek to understand the consequences of climate change for human civilization and determine if these changes can be predicted. NASA supports an ongoing research and analysis program, which is conducted independently and in support of satellite missions. The research program conducts several on-going field campaigns employing aircraft, balloons, and ground based systems. These campaigns have focused on exploring processes in the tropics, high latitudes, and continental outflow to explain the chemistry and transport in the troposphere and stratosphere and how these regions interact. NASA is now studying how the Aura mission and requirements of the research and analysis program might be merged to achieve its strategic goals related to global atmospheric chemistry changes. In addition, NASA field campaign resources will be folded into Aura's validation requirements. Aura validation requires correlative measurements throughout the troposphere and stratosphere under a range of observing and geophysical conditions. Because of the recent launches of Envisat and other smaller international chemistry satellites, the NASA program plans to collaborate with European space agencies in developing a series of campaigns that will provide continuity between those satellites missions and Aura.
Understanding personality from Ayurvedic perspective for psychological assessment: A case
Shilpa, S; Venkatesha Murthy, C. G.
2011-01-01
The study of personality has fascinated psychologists since a long time. Personality as an area of study has grown so much that a wealth of literature is available. On the other hand, the socio-ethnic dynamics of human race in the emerging global village context provoking psychologists to develop a personality theory which can treat certain basic components of personality as invariants, so that irrespective of culture, race, and nativity could still be able to study personality which will have universal applicability and relevance, is still far away. In the above emerging backdrop, “Ayurveda” has perhaps an important role to play as it can enable providing a theoretical and empirical base of personality traits and types. These Ayurvedic concepts are applicable to all human beings irrespective of caste, color, sex or race. Therefore, it is interesting to build on the Ayurvedic knowledge which has already given us so much since time immemorial, and validate some of these issues related to personality from psychological perspective. There are already certain efforts in understanding Sattva, Rajas, and Tamas Gunas. It can be furthered so that a comprehensive personality picture can be generated, which can have implications for health, career, education and many other dimensions of life. The present paper is a theoretical attempt in developing such a personality proposition which can be validated. Thus, the present paper only builds a theoretical framework for their possible empirical validity. PMID:22131752
2010-01-01
Background The modular approach to analysis of genetically modified organisms (GMOs) relies on the independence of the modules combined (i.e. DNA extraction and GM quantification). The validity of this assumption has to be proved on the basis of specific performance criteria. Results An experiment was conducted using, as a reference, the validated quantitative real-time polymerase chain reaction (PCR) module for detection of glyphosate-tolerant Roundup Ready® GM soybean (RRS). Different DNA extraction modules (CTAB, Wizard and Dellaporta), were used to extract DNA from different food/feed matrices (feed, biscuit and certified reference material [CRM 1%]) containing the target of the real-time PCR module used for validation. Purity and structural integrity (absence of inhibition) were used as basic criteria that a DNA extraction module must satisfy in order to provide suitable template DNA for quantitative real-time (RT) PCR-based GMO analysis. When performance criteria were applied (removal of non-compliant DNA extracts), the independence of GMO quantification from the extraction method and matrix was statistically proved, except in the case of Wizard applied to biscuit. A fuzzy logic-based procedure also confirmed the relatively poor performance of the Wizard/biscuit combination. Conclusions For RRS, this study recognises that modularity can be generally accepted, with the limitation of avoiding combining highly processed material (i.e. biscuit) with a magnetic-beads system (i.e. Wizard). PMID:20687918
[Benchmarking and other functions of ROM: back to basics].
Barendregt, M
2015-01-01
Since 2011 outcome data in the Dutch mental health care have been collected on a national scale. This has led to confusion about the position of benchmarking in the system known as routine outcome monitoring (rom). To provide insight into the various objectives and uses of aggregated outcome data. A qualitative review was performed and the findings were analysed. Benchmarking is a strategy for finding best practices and for improving efficacy and it belongs to the domain of quality management. Benchmarking involves comparing outcome data by means of instrumentation and is relatively tolerant with regard to the validity of the data. Although benchmarking is a function of rom, it must be differentiated form other functions from rom. Clinical management, public accountability, research, payment for performance and information for patients are all functions of rom which require different ways of data feedback and which make different demands on the validity of the underlying data. Benchmarking is often wrongly regarded as being simply a synonym for 'comparing institutions'. It is, however, a method which includes many more factors; it can be used to improve quality and has a more flexible approach to the validity of outcome data and is less concerned than other rom functions about funding and the amount of information given to patients. Benchmarking can make good use of currently available outcome data.
Health literacy and 30-day hospital readmission after acute myocardial infarction.
Bailey, Stacy Cooper; Fang, Gang; Annis, Izabela E; O'Conor, Rachel; Paasche-Orlow, Michael K; Wolf, Michael S
2015-06-11
To assess the validity of a predictive model of health literacy, and to examine the relationship between derived health literacy estimates and 30-day hospital readmissions for acute myocardial infarction (AMI). Retrospective cohort study. A National Institute of Aging (NIA) study cohort of 696 adult, English-speaking primary care patients, aged 55-74 years, was used to assess the validity of derived health literacy estimates. Claims from 7733 Medicare beneficiaries hospitalised for AMI in 2008 in North Carolina and Illinois were used to investigate the association between health literacy estimates and 30-day hospital readmissions. The NIA cohort was administered 3 common health literacy assessments (Newest Vital Sign, Test of Functional Health Literacy in Adults, and Rapid Estimate of Adult Literacy in Medicine). Health literacy estimates at the census block group level were derived via a predictive model. 30-day readmissions were measured from Medicare claims data using a validated algorithm. Fair agreement was found between derived estimates and in-person literacy assessments (Pearson Correlation coefficients: 0.38-0.51; κ scores: 0.38-0.40). Medicare enrollees with above basic literacy according to derived health literacy estimates had an 18% lower risk of a 30-day readmission (RR=0.82, 95% CI 0.73 to 0.92) and 21% lower incidence rate of 30-day readmission (IRR=0.79, 95% CI 0.68 to 0.87) than patients with basic or below basic literacy. After adjusting for demographic and clinical characteristics, the risk of 30-day readmission was 12% lower (p=0.03), and the incidence rate 16% lower (p<0.01) for patients with above basic literacy. Health literacy, as measured by a predictive model, was found to be a significant, independent predictor of 30-day readmissions. As a modifiable risk factor with evidence-based solutions, health literacy should be considered in readmission reduction efforts. Published by the BMJ Publishing Group Limited. For permission to use (where not already granted under a licence) please go to http://group.bmj.com/group/rights-licensing/permissions.
Pengra, Bruce; Long, Jordan; Dahal, Devendra; Stehman, Stephen V.; Loveland, Thomas R.
2015-01-01
The methodology for selection, creation, and application of a global remote sensing validation dataset using high resolution commercial satellite data is presented. High resolution data are obtained for a stratified random sample of 500 primary sampling units (5 km × 5 km sample blocks), where the stratification based on Köppen climate classes is used to distribute the sample globally among biomes. The high resolution data are classified to categorical land cover maps using an analyst mediated classification workflow. Our initial application of these data is to evaluate a global 30 m Landsat-derived, continuous field tree cover product. For this application, the categorical reference classification produced at 2 m resolution is converted to percent tree cover per 30 m pixel (secondary sampling unit)for comparison to Landsat-derived estimates of tree cover. We provide example results (based on a subsample of 25 sample blocks in South America) illustrating basic analyses of agreement that can be produced from these reference data. Commercial high resolution data availability and data quality are shown to provide a viable means of validating continuous field tree cover. When completed, the reference classifications for the full sample of 500 blocks will be released for public use.
NASA Astrophysics Data System (ADS)
O'Brien, Enda; McKinstry, Alastair; Ralph, Adam
2015-04-01
Building on previous work presented at EGU 2013 (http://www.sciencedirect.com/science/article/pii/S1876610213016068 ), more results are available now from a different wind-farm in complex terrain in southwest Ireland. The basic approach is to interpolate wind-speed forecasts from an operational weather forecast model (i.e., HARMONIE in the case of Ireland) to the precise location of each wind-turbine, and then use Bayes Model Averaging (BMA; with statistical information collected from a prior training-period of e.g., 25 days) to remove systematic biases. Bias-corrected wind-speed forecasts (and associated power-generation forecasts) are then provided twice daily (at 5am and 5pm) out to 30 hours, with each forecast validation fed back to BMA for future learning. 30-hr forecasts from the operational Met Éireann HARMONIE model at 2.5km resolution have been validated against turbine SCADA observations since Jan. 2014. An extra high-resolution (0.5km grid-spacing) HARMONIE configuration has been run since Nov. 2014 as an extra member of the forecast "ensemble". A new version of HARMONIE with extra filters designed to stabilize high-resolution configurations has been run since Jan. 2015. Measures of forecast skill and forecast errors will be provided, and the contributions made by the various physical and computational enhancements to HARMONIE will be quantified.
Validation of a DICE Simulation Against a Discrete Event Simulation Implemented Entirely in Code.
Möller, Jörgen; Davis, Sarah; Stevenson, Matt; Caro, J Jaime
2017-10-01
Modeling is an essential tool for health technology assessment, and various techniques for conceptualizing and implementing such models have been described. Recently, a new method has been proposed-the discretely integrated condition event or DICE simulation-that enables frequently employed approaches to be specified using a common, simple structure that can be entirely contained and executed within widely available spreadsheet software. To assess if a DICE simulation provides equivalent results to an existing discrete event simulation, a comparison was undertaken. A model of osteoporosis and its management programmed entirely in Visual Basic for Applications and made public by the National Institute for Health and Care Excellence (NICE) Decision Support Unit was downloaded and used to guide construction of its DICE version in Microsoft Excel ® . The DICE model was then run using the same inputs and settings, and the results were compared. The DICE version produced results that are nearly identical to the original ones, with differences that would not affect the decision direction of the incremental cost-effectiveness ratios (<1% discrepancy), despite the stochastic nature of the models. The main limitation of the simple DICE version is its slow execution speed. DICE simulation did not alter the results and, thus, should provide a valid way to design and implement decision-analytic models without requiring specialized software or custom programming. Additional efforts need to be made to speed up execution.
NASA Astrophysics Data System (ADS)
Blot, R.; Nedelec, P.; Petetin, H.; Thouret, V.; Cohen, Y.
2017-12-01
The In-Service Aircraft for a Global Observing System (IAGOS; http://www.iagos.org) is an European Research Infrastructure that provides cost-effective global atmospheric composition measurements at high resolution using commercial passenger aircraft. It is the continuation of the MOZAIC (1994-2014) and the CARIBIC (since 1997) programs that has provided a unique scientific database using 6 aircraft operated by European airlines over two decades. Thanks to growing interests of several international Airlines to contribute to the academic climate research, the IAGOS aircraft fleet (started in 2011), with the IAGOS-CORE basic instrumentation, has expanded to 9 Airbus A340/A330 aircraft up to now. Here, we present this IAGOS-CORE instrumentation that continuously sample carbon monoxide, ozone, water vapor and cloud droplets. We focus on carbon monoxide and ozone measurements which are performed by optimized, but well known, methods such as UV absorption and IR correlation. We describe the data processing/validation and the data quality control. With already more than 20 and 15 years of continuous ozone and carbon monoxide measurements, respectively, the IAGOS/MOZAIC data are particularly suitable for climatologies and trends. Also, since commercial aircraft are daily operated, the near-real time IAGOS-CORE data are also used to observe pollution plumes and to validate air-quality models as well as satellite products.
Spectral studies related to dissociation of HBr, HCl and BrO
NASA Technical Reports Server (NTRS)
Ginter, M. L.
1986-01-01
Concern over halogen catalyzed decomposition of O3 in the upper atmosphere has generated need for data on the atomic and molecular species X, HX and XO (where X is Cl and Br). Of special importance are Cl produced from freon decomposition and Cl and Br produced from natural processes and from other industrial and agricultural chemicals. Basic spectral data is provided on HCl, HBr, and BrO necessary to detect specific states and energy levels, to enable detailed modeling of the processes involving molecular dissociation, ionization, etc., and to help evaluate field experiments to check the validity of model calculations for these species in the upper atmosphere. Results contained in four published papers and two major spectral compilations are summarized together with other results obtained.
[Biological markers in epidemiology: concepts, applications, perspectives (part I)].
Hoffmann, W; Latza, U; Ahrens, W; Greiser, K H; Kroke, A; Nieters, A; Schulze, M B; Steiner, M; Terschüren, C; Wjst, M
2002-02-01
The inclusion of biomarkers in epidemiological research provides new possibilities for exposure assessment and the study of early structural or functional changes and pre-clinical stages of diseases. At the same time issues of validity, reliability, and quality control as well as logistics require special attention. Usually epidemiological studies become more expensive with regard to time and cost. Interdisciplinary collaboration between epidemiology, basic research, and laboratory research is crucial. A prerequisite for this collaboration are agreements on definitions, methods and procedures. The definition of "biomarker" and a description of previous uses of biomarkers in epidemiological studies are presented in the first part of this paper. The second part addresses genetic markers and markers of individual sensitivity and susceptibility. We will end with a discussion about the possible future of biomarkers in epidemiology.
Functional differentiability in time-dependent quantum mechanics.
Penz, Markus; Ruggenthaler, Michael
2015-03-28
In this work, we investigate the functional differentiability of the time-dependent many-body wave function and of derived quantities with respect to time-dependent potentials. For properly chosen Banach spaces of potentials and wave functions, Fréchet differentiability is proven. From this follows an estimate for the difference of two solutions to the time-dependent Schrödinger equation that evolve under the influence of different potentials. Such results can be applied directly to the one-particle density and to bounded operators, and present a rigorous formulation of non-equilibrium linear-response theory where the usual Lehmann representation of the linear-response kernel is not valid. Further, the Fréchet differentiability of the wave function provides a new route towards proving basic properties of time-dependent density-functional theory.
Software Tools to Support Research on Airport Departure Planning
NASA Technical Reports Server (NTRS)
Carr, Francis; Evans, Antony; Feron, Eric; Clarke, John-Paul
2003-01-01
A simple, portable and useful collection of software tools has been developed for the analysis of airport surface traffic. The tools are based on a flexible and robust traffic-flow model, and include calibration, validation and simulation functionality for this model. Several different interfaces have been developed to help promote usage of these tools, including a portable Matlab(TM) implementation of the basic algorithms; a web-based interface which provides online access to automated analyses of airport traffic based on a database of real-world operations data which covers over 250 U.S. airports over a 5-year period; and an interactive simulation-based tool currently in use as part of a college-level educational module. More advanced applications for airport departure traffic include taxi-time prediction and evaluation of "windowing" congestion control.
Diller, David J; Swanson, Jon; Bayden, Alexander S; Jarosinski, Mark; Audie, Joseph
2015-01-01
Peptides provide promising templates for developing drugs to occupy a middle space between small molecules and antibodies and for targeting 'undruggable' intracellular protein-protein interactions. Importantly, rational or in cerebro design, especially when coupled with validated in silico tools, can be used to efficiently explore chemical space and identify islands of 'drug-like' peptides to satisfy diverse drug discovery program objectives. Here, we consider the underlying principles of and recent advances in rational, computer-enabled peptide drug design. In particular, we consider the impact of basic physicochemical properties, potency and ADME/Tox opportunities and challenges, and recently developed computational tools for enabling rational peptide drug design. Key principles and practices are spotlighted by recent case studies. We close with a hypothetical future case study.
FILTSoft: A computational tool for microstrip planar filter design
NASA Astrophysics Data System (ADS)
Elsayed, M. H.; Abidin, Z. Z.; Dahlan, S. H.; Cholan N., A.; Ngu, Xavier T. I.; Majid, H. A.
2017-09-01
Filters are key component of any communication system to control spectrum and suppress interferences. Designing a filter involves long process as well as good understanding of the basic hardware technology. Hence this paper introduces an automated design tool based on Matlab-GUI, called the FILTSoft (acronym for Filter Design Software) to ease the process. FILTSoft is a user friendly filter design tool to aid, guide and expedite calculations from lumped elements level to microstrip structure. Users just have to provide the required filter specifications as well as the material description. FILTSoft will calculate and display the lumped element details, the planar filter structure, and the expected filter's response. An example of a lowpass filter design was calculated using FILTSoft and the results were validated through prototype measurement for comparison purposes.
Yu, Xiaochu; Jiang, Jingmei; Liu, Changwei; Shen, Keng; Wang, Zixing; Han, Wei; Liu, Xingrong; Lin, Guole; Zhang, Ye; Zhang, Ying; Ma, Yufen; Bo, Haixin; Zhao, Yupei
2017-06-15
Surgical safety has emerged as a crucial global health issue in the past two decades. Although several safety-enhancing tools are available, the pace of large-scale improvement remains slow, especially in developing countries such as China. The present project (Modern Surgery and Anesthesia Safety Management System Construction and Promotion) aims to develop and validate system-based integrated approaches for reducing perioperative deaths and complications using a multicentre, multistage design. The project involves collection of clinical and outcome information for 1 20 000 surgical inpatients at four regionally representative academic/teaching general hospitals in China during three sequential stages: preparation and development, effectiveness validation and improvement of implementation for promotion. These big data will provide the evidence base for the formulation, validation and improvement processes of a system-based stratified safety intervention package covering the entire surgical pathway. Attention will be directed to managing inherent patient risks and regulating medical safety behaviour. Information technology will facilitate data collection and intervention implementation, provide supervision mechanisms and guarantee transfer of key patient safety messages between departments and personnel. Changes in rates of deaths, surgical complications during hospitalisation, length of stay, system adoption and implementation rates will be analysed to evaluate effectiveness and efficiency. This study was approved by the institutional review boards of Peking Union Medical College Hospital, First Hospital of China Medical University, Qinghai Provincial People's Hospital, Xiangya Hospital Central South University and the Institute of Basic Medical Sciences, Chinese Academy of Medical Sciences. Study findings will be disseminated via peer-reviewed journals, conference presentations and patent papers. © Article author(s) (or their employer(s) unless otherwise stated in the text of the article) 2017. All rights reserved. No commercial use is permitted unless otherwise expressly granted.
RUCAM in Drug and Herb Induced Liver Injury: The Update
Danan, Gaby; Teschke, Rolf
2015-01-01
RUCAM (Roussel Uclaf Causality Assessment Method) or its previous synonym CIOMS (Council for International Organizations of Medical Sciences) is a well established tool in common use to quantitatively assess causality in cases of suspected drug induced liver injury (DILI) and herb induced liver injury (HILI). Historical background and the original work confirm the use of RUCAM as single term for future cases, dismissing now the term CIOMS for reasons of simplicity and clarity. RUCAM represents a structured, standardized, validated, and hepatotoxicity specific diagnostic approach that attributes scores to individual key items, providing final quantitative gradings of causality for each suspect drug/herb in a case report. Experts from Europe and the United States had previously established in consensus meetings the first criteria of RUCAM to meet the requirements of clinicians and practitioners in care for their patients with suspected DILI and HILI. RUCAM was completed by additional criteria and validated, assisting to establish the timely diagnosis with a high degree of certainty. In many countries and for more than two decades, physicians, regulatory agencies, case report authors, and pharmaceutical companies successfully applied RUCAM for suspected DILI and HILI. Their practical experience, emerging new data on DILI and HILI characteristics, and few ambiguous questions in domains such alcohol use and exclusions of non-drug causes led to the present update of RUCAM. The aim was to reduce interobserver and intraobserver variability, to provide accurately defined, objective core elements, and to simplify the handling of the items. We now present the update of the well accepted original RUCAM scale and recommend its use for clinical, regulatory, publication, and expert purposes to validly establish causality in cases of suspected DILI and HILI, facilitating a straightforward application and an internationally harmonized approach of causality assessment as a common basic tool. PMID:26712744
Design and validation of an intelligent wheelchair towards a clinically-functional outcome.
Boucher, Patrice; Atrash, Amin; Kelouwani, Sousso; Honoré, Wormser; Nguyen, Hai; Villemure, Julien; Routhier, François; Cohen, Paul; Demers, Louise; Forget, Robert; Pineau, Joelle
2013-06-17
Many people with mobility impairments, who require the use of powered wheelchairs, have difficulty completing basic maneuvering tasks during their activities of daily living (ADL). In order to provide assistance to this population, robotic and intelligent system technologies have been used to design an intelligent powered wheelchair (IPW). This paper provides a comprehensive overview of the design and validation of the IPW. The main contributions of this work are three-fold. First, we present a software architecture for robot navigation and control in constrained spaces. Second, we describe a decision-theoretic approach for achieving robust speech-based control of the intelligent wheelchair. Third, we present an evaluation protocol motivated by a meaningful clinical outcome, in the form of the Robotic Wheelchair Skills Test (RWST). This allows us to perform a thorough characterization of the performance and safety of the system, involving 17 test subjects (8 non-PW users, 9 regular PW users), 32 complete RWST sessions, 25 total hours of testing, and 9 kilometers of total running distance. User tests with the RWST show that the navigation architecture reduced collisions by more than 60% compared to other recent intelligent wheelchair platforms. On the tasks of the RWST, we measured an average decrease of 4% in performance score and 3% in safety score (not statistically significant), compared to the scores obtained with conventional driving model. This analysis was performed with regular users that had over 6 years of wheelchair driving experience, compared to approximately one half-hour of training with the autonomous mode. The platform tested in these experiments is among the most experimentally validated robotic wheelchairs in realistic contexts. The results establish that proficient powered wheelchair users can achieve the same level of performance with the intelligent command mode, as with the conventional command mode.
Experimental evaluation of the certification-trail method
NASA Technical Reports Server (NTRS)
Sullivan, Gregory F.; Wilson, Dwight S.; Masson, Gerald M.; Itoh, Mamoru; Smith, Warren W.; Kay, Jonathan S.
1993-01-01
Certification trails are a recently introduced and promising approach to fault-detection and fault-tolerance. A comprehensive attempt to assess experimentally the performance and overall value of the method is reported. The method is applied to algorithms for the following problems: huffman tree, shortest path, minimum spanning tree, sorting, and convex hull. Our results reveal many cases in which an approach using certification-trails allows for significantly faster overall program execution time than a basic time redundancy-approach. Algorithms for the answer-validation problem for abstract data types were also examined. This kind of problem provides a basis for applying the certification-trail method to wide classes of algorithms. Answer-validation solutions for two types of priority queues were implemented and analyzed. In both cases, the algorithm which performs answer-validation is substantially faster than the original algorithm for computing the answer. Next, a probabilistic model and analysis which enables comparison between the certification-trail method and the time-redundancy approach were presented. The analysis reveals some substantial and sometimes surprising advantages for ther certification-trail method. Finally, the work our group performed on the design and implementation of fault injection testbeds for experimental analysis of the certification trail technique is discussed. This work employs two distinct methodologies, software fault injection (modification of instruction, data, and stack segments of programs on a Sun Sparcstation ELC and on an IBM 386 PC) and hardware fault injection (control, address, and data lines of a Motorola MC68000-based target system pulsed at logical zero/one values). Our results indicate the viability of the certification trail technique. It is also believed that the tools developed provide a solid base for additional exploration.
NASA Astrophysics Data System (ADS)
Ungermann, J.; Blank, J.; Dick, M.; Ebersoldt, A.; Friedl-Vallon, F.; Giez, A.; Guggenmoser, T.; Höpfner, M.; Jurkat, T.; Kaufmann, M.; Kaufmann, S.; Kleinert, A.; Krämer, M.; Latzko, T.; Oelhaf, H.; Olchewski, F.; Preusse, P.; Rolf, C.; Schillings, J.; Suminska-Ebersoldt, O.; Tan, V.; Thomas, N.; Voigt, C.; Zahn, A.; Zöger, M.; Riese, M.
2015-06-01
The Gimballed Limb Observer for Radiance Imaging of the Atmosphere (GLORIA) is an airborne infrared limb imager combining a two-dimensional infrared detector with a Fourier transform spectrometer. It was operated aboard the new German Gulfstream G550 High Altitude LOng Range (HALO) research aircraft during the Transport And Composition in the upper Troposphere/lowermost Stratosphere (TACTS) and Earth System Model Validation (ESMVAL) campaigns in summer 2012. This paper describes the retrieval of temperature and trace gas (H2O, O3, HNO3) volume mixing ratios from GLORIA dynamics mode spectra that are spectrally sampled every 0.625 cm-1. A total of 26 integrated spectral windows are employed in a joint fit to retrieve seven targets using consecutively a fast and an accurate tabulated radiative transfer model. Typical diagnostic quantities are provided including effects of uncertainties in the calibration and horizontal resolution along the line of sight. Simultaneous in situ observations by the Basic Halo Measurement and Sensor System (BAHAMAS), the Fast In-situ Stratospheric Hygrometer (FISH), an ozone detector named Fairo, and the Atmospheric chemical Ionization Mass Spectrometer (AIMS) allow a validation of retrieved values for three flights in the upper troposphere/lowermost stratosphere region spanning polar and sub-tropical latitudes. A high correlation is achieved between the remote sensing and the in situ trace gas data, and discrepancies can to a large extent be attributed to differences in the probed air masses caused by different sampling characteristics of the instruments. This 1-D processing of GLORIA dynamics mode spectra provides the basis for future tomographic inversions from circular and linear flight paths to better understand selected dynamical processes of the upper troposphere and lowermost stratosphere.
NASA Astrophysics Data System (ADS)
Ungermann, J.; Blank, J.; Dick, M.; Ebersoldt, A.; Friedl-Vallon, F.; Giez, A.; Guggenmoser, T.; Höpfner, M.; Jurkat, T.; Kaufmann, M.; Kaufmann, S.; Kleinert, A.; Krämer, M.; Latzko, T.; Oelhaf, H.; Olchewski, F.; Preusse, P.; Rolf, C.; Schillings, J.; Suminska-Ebersoldt, O.; Tan, V.; Thomas, N.; Voigt, C.; Zahn, A.; Zöger, M.; Riese, M.
2014-12-01
The Gimballed Limb Observer for Radiance Imaging of the Atmosphere (GLORIA) is an airborne infrared limb-imager combining a two-dimensional infrared detector with a Fourier transform spectrometer. It was operated aboard the new German Gulfstream G550 research aircraft HALO during the Transport And Composition in the upper Troposphere/lowermost Stratosphere (TACTS) and Earth System Model Validation (ESMVAL) campaigns in summer 2012. This paper describes the retrieval of temperature and trace gas (H2O, O3, HNO3) volume mixing ratios from GLORIA dynamics mode spectra. 26 integrated spectral windows are employed in a joint fit to retrieve seven targets using consecutively a fast and an accurate tabulated radiative transfer model. Typical diagnostic quantities are provided including effects of uncertainties in the calibration and horizontal resolution along the line-of-sight. Simultaneous in-situ observations by the BAsic HALO Measurement And Sensor System (BAHAMAS), the Fast In-Situ Stratospheric Hygrometer (FISH), FAIRO, and the Atmospheric chemical Ionization Mass Spectrometer (AIMS) allow a validation of retrieved values for three flights in the upper troposphere/lowermost stratosphere region spanning polar and sub-tropical latitudes. A high correlation is achieved between the remote sensing and the in-situ trace gas data, and discrepancies can to a large fraction be attributed to differences in the probed air masses caused by different sampling characteristics of the instruments. This 1-D processing of GLORIA dynamics mode spectra provides the basis for future tomographic inversions from circular and linear flight paths to better understand selected dynamical processes of the upper troposphere and lowermost stratosphere.
Yu, Xiaochu; Jiang, Jingmei; Liu, Changwei; Shen, Keng; Wang, Zixing; Han, Wei; Liu, Xingrong; Lin, Guole; Zhang, Ye; Zhang, Ying; Ma, Yufen; Bo, Haixin; Zhao, Yupei
2017-01-01
Introduction Surgical safety has emerged as a crucial global health issue in the past two decades. Although several safety-enhancing tools are available, the pace of large-scale improvement remains slow, especially in developing countries such as China. The present project (Modern Surgery and Anesthesia Safety Management System Construction and Promotion) aims to develop and validate system-based integrated approaches for reducing perioperative deaths and complications using a multicentre, multistage design. Methods and analysis The project involves collection of clinical and outcome information for 1 20 000 surgical inpatients at four regionally representative academic/teaching general hospitals in China during three sequential stages: preparation and development, effectiveness validation and improvement of implementation for promotion. These big data will provide the evidence base for the formulation, validation and improvement processes of a system-based stratified safety intervention package covering the entire surgical pathway. Attention will be directed to managing inherent patient risks and regulating medical safety behaviour. Information technology will facilitate data collection and intervention implementation, provide supervision mechanisms and guarantee transfer of key patient safety messages between departments and personnel. Changes in rates of deaths, surgical complications during hospitalisation, length of stay, system adoption and implementation rates will be analysed to evaluate effectiveness and efficiency. Ethics and dissemination This study was approved by the institutional review boards of Peking Union Medical College Hospital, First Hospital of China Medical University, Qinghai Provincial People’s Hospital, Xiangya Hospital Central South University and the Institute of Basic Medical Sciences, Chinese Academy of Medical Sciences. Study findings will be disseminated via peer-reviewed journals, conference presentations and patent papers. PMID:28619774
CFD Approach To Investigate The Flow Characteristics In Bi-Directional Ventilated Disc Brake
NASA Astrophysics Data System (ADS)
Munisamy, Kannan M.; Yusoff, Mohd. Zamri; Shuaib, Norshah Hafeez; Thangaraju, Savithry K.
2010-06-01
This paper presents experimental and Computational Fluids Dynamics (CFD) investigations of the flow in ventilated brake discs. Development of an experiment rig with basic measuring devices are detailed out and following a validation study, the possible improvement in the brake cooling can be further analyzed using CFD analysis. The mass flow rate is determined from basic flow measurement technique following that the conventional bi-directional passenger car is simulated using commercial CFD software FLUENT™. The CFD simulation is used to investigate the flow characteristics in between blade flow of the bi-directional ventilated disc brake.
NASA Technical Reports Server (NTRS)
Wilson, T. G.; Lee, F. C. Y.; Burns, W. W., III; Owen, H. A., Jr.
1974-01-01
A procedure is developed for classifying dc-to-square-wave two-transistor parallel inverters used in power conditioning applications. The inverters are reduced to equivalent RLC networks and are then grouped with other inverters with the same basic equivalent circuit. Distinction between inverter classes is based on the topology characteristics of the equivalent circuits. Information about one class can then be extended to another class using the basic oscillation theory and the concept of duality. Oscillograms from test circuits confirm the validity of the procedure adopted.
NASA Astrophysics Data System (ADS)
Mohanty, B.; Jena, S.; Panda, R. K.
2016-12-01
The overexploitation of groundwater elicited in abandoning several shallow tube wells in the study Basin in Eastern India. For the sustainability of groundwater resources, basin-scale modelling of groundwater flow is indispensable for the effective planning and management of the water resources. The basic intent of this study is to develop a 3-D groundwater flow model of the study basin using the Visual MODFLOW Flex 2014.2 package and successfully calibrate and validate the model using 17 years of observed data. The sensitivity analysis was carried out to quantify the susceptibility of aquifer system to the river bank seepage, recharge from rainfall and agriculture practices, horizontal and vertical hydraulic conductivities, and specific yield. To quantify the impact of parameter uncertainties, Sequential Uncertainty Fitting Algorithm (SUFI-2) and Markov chain Monte Carlo (McMC) techniques were implemented. Results from the two techniques were compared and the advantages and disadvantages were analysed. Nash-Sutcliffe coefficient (NSE), Coefficient of Determination (R2), Mean Absolute Error (MAE), Mean Percent Deviation (Dv) and Root Mean Squared Error (RMSE) were adopted as criteria of model evaluation during calibration and validation of the developed model. NSE, R2, MAE, Dv and RMSE values for groundwater flow model during calibration and validation were in acceptable range. Also, the McMC technique was able to provide more reasonable results than SUFI-2. The calibrated and validated model will be useful to identify the aquifer properties, analyse the groundwater flow dynamics and the change in groundwater levels in future forecasts.
Federal Register 2010, 2011, 2012, 2013, 2014
2013-07-31
..., and the facilities and controls used for, the manufacture, preproduction design validation (including... part 820 (21 CFR part 820) and sets forth basic CGMP requirements governing the design, manufacture... manufacturers for compliance with QS requirements encompassing design, production, installation, and servicing...
Federal Register 2010, 2011, 2012, 2013, 2014
2010-06-24
... the facilities and controls used for, the manufacture, pre-production design validation (including a... 820 (21 CFR part 820) and sets forth basic CGMP requirements governing the design, manufacture... manufacturers for compliance with QS requirements encompassing design, production, installation, and servicing...
Humanities, Religion, and the Arts Tomorrow.
ERIC Educational Resources Information Center
Hunter, Howard, Ed.
Intended as a basic resource in new primary sources for interdisciplinary studies, this book consists of twelve essays on contemporary culture, religion, and the arts. The authors, specialists in the humanities, are concerned with interdisciplinary investigation, including such issues as determining methods of study, methods of validating claims…
Human Services. Georgia Core Standards for Occupational Clusters.
ERIC Educational Resources Information Center
Georgia Univ., Athens. Dept. of Occupational Studies.
This document lists core standards and occupational knowledge and skills that have been identified and validated by industry as necessary to all Georgia students in secondary-level human services occupations programs. First, foundation skills are grouped as follows: basic skills (reading, writing, arithmetic/mathematics, listening, speaking);…
Technical/Engineering. Georgia Core Standards for Occupational Clusters.
ERIC Educational Resources Information Center
Georgia Univ., Athens. Dept. of Occupational Studies.
This document lists core standards and occupational knowledge and skills that have been identified and validated by industry as necessary to all Georgia students in secondary-level technical/engineering programs. First, foundation skills are grouped as follows: basic skills (reading, writing, arithmetic/mathematics, listening, speaking); thinking…
Health Care. Georgia Core Standards for Occupational Clusters.
ERIC Educational Resources Information Center
Georgia Univ., Athens. Dept. of Occupational Studies.
This document lists core standards and occupational knowledge and skills that have been identified/validated by industry as necessary to all Georgia students in secondary-level health care occupations programs. First, foundation skills are grouped as follows: basic skills (reading, writing, arithmetic/mathematics, listening, speaking); thinking…
Measuring Problem Solving Skills in "Portal 2"
ERIC Educational Resources Information Center
Shute, Valerie J.; Wang, Lubin
2013-01-01
This paper examines possible improvement to problem solving skills as a function of playing the video game "Portal 2." Stealth assessment is used in the game to evaluate students' problem solving abilities--specifically basic and flexible rule application. The stealth assessment measures will be validated against commonly accepted…
working with partners on a global initiative developing the Global Register of Introduced and Invasive Species (GRIIS) (site is under development) which is aimed at developing country-wise validated, verified will feature at least basic information on each of these species. ISSG with partners is developing the
Reviews of Selected System and Software Tools for Strategic Defense Applications
1990-02-01
Interleaf and FrameMaker . IStatic Diagnostics Basic testing includes validating flows, detecting orphan activity, and checking completeness of activities...Publisher, Aldus PageMaker, Unix pic, Apple .pict metafile, Interleaf, Framemaker , or Postscript format. There are no forms for standard documents such as 3
Teacher Competency: A Public Farce!
ERIC Educational Resources Information Center
Weitman, Catheryn J.
The current popularity of teacher testing allows for content, criterion, and construct validity to be assessed, as pertaining to achievement levels on basic knowledge examinations. Teacher competency is a complex issue that is inaccurately confused with or identified as measures derived from academic testing. The problems in addressing the…
What Developmental Educators Should Know About Learning Styles and Cognitive Styles.
ERIC Educational Resources Information Center
Lemire, David
2002-01-01
Considers three serious problems associated with learning styles: confusion in definitions, weaknesses in reliability and validity, and the identification of relevant characteristics in instructional settings, or aptitude-treatment interactions. Discusses four basic groups of learning styles that developmental educators should be aware of.…
Student Climate Survey, Spring 1998.
ERIC Educational Resources Information Center
Truckee Meadows Community Coll., Sparks, NV.
Every three years, Truckee Meadows Community College (Nevada) administers a student climate survey that measures the attitudes, perceptions, and opinions of its student population. The instrument used to survey the student body was designed with three basic objectives in mind: (1) validate the institutional mission; (2) obtain input regarding the…
2012-01-01
Background Ensuring the quality of malaria medicines is crucial in working toward malaria control and eventual elimination. Unlike other validated tests that can assess all critical quality attributes, which is the standard for determining the quality of medicines, basic tests are significantly less expensive, faster, and require less skilled labour; yet, these tests provide reproducible data and information on several critical quality attributes, such as identity, purity, content, and disintegration. Visual and physical inspection also provides valuable information about the manufacturing and the labelling of medicines, and in many cases this inspection is sufficient to detect counterfeit medicines. The Promoting the Quality of Medicines (PQM) programme has provided technical assistance to Amazon Malaria Initiative (AMI) countries to implement the use of basic tests as a key screening mechanism to assess the quality of malaria medicines available to patients in decentralized regions. Methods Trained personnel from the National Malaria Control Programmes (NMCPs), often in collaboration with country’s Official Medicine Control Laboratory (OMCL), developed country- specific protocols that encompassed sampling methods, sample analysis, and data reporting. Sampling sites were selected based on malaria burden, accessibility, and geographical location. Convenience sampling was performed and countries were recommended to store the sampled medicines under conditions that did not compromise their quality. Basic analytical tests, such as disintegration and thin layer chromatography (TLC), were performed utilizing a portable mini-laboratory. Results Results were originally presented at regional meetings in a non-standardized format that lacked relevant medicines information. However, since 2008 information has been submitted utilizing a template specifically developed by PQM for that purpose. From 2005 to 2010, the quality of 1,663 malaria medicines from seven AMI countries was evaluated, mostly collected from the public sector, 1,445/1,663 (86.9%). Results indicate that 193/1,663 (11.6%) were found not to meet quality specifications. Most failures were reported during visual and physical inspection, 142/1663 (8.5%), and most of these were due to expired medicines, 118/142 (83.1%). Samples failing TLC accounted for 27/1,663 (1.6%) and those failing disintegration accounted for 24/1,663 (1.4%). Medicines quality failures decreased significantly during the last two years. Conclusions Basic tests revealed that the quality of medicines in the public sector improved over the years, since the implementation of this type of quality monitoring programme in 2005. However, the lack of consistent confirmatory tests in the quality control (QC) laboratory, utilizing methods that can also evaluate additional quality attributes, could still mask quality issues. In the future, AMI countries should improve coordination with their health authorities and their QC lab consistently, to provide a more complete picture of malaria medicines quality and support the implementation of corrective actions. Facilities in the private and informal sectors also should be included when these sectors constitute an important source of medicines used by malaria patients. PMID:22704680
Pribluda, Victor S; Barojas, Adrian; Añez, Arletta; López, Cecilia G; Figueroa, Ruth; Herrera, Roxana; Nakao, Gladys; Nogueira, Fernando Ha; Pianetti, Gerson A; Povoa, Marinete M; Viana, Giselle Mr; Gomes, Margarete S Mendonça; Escobar, Jose P; Sierra, Olga L Muñoz; Norena, Susana P Rendon; Veloz, Raúl; Bravo, Marcy Silva; Aldás, Martha R; Hindssemple, Alison; Collins, Marilyn; Ceron, Nicolas; Krishnalall, Karanchand; Adhin, Malti; Bretas, Gustavo; Hernandez, Nelly; Mendoza, Marjorie; Smine, Abdelkrim; Chibwe, Kennedy; Lukulay, Patrick; Evans, Lawrence
2012-06-15
Ensuring the quality of malaria medicines is crucial in working toward malaria control and eventual elimination. Unlike other validated tests that can assess all critical quality attributes, which is the standard for determining the quality of medicines, basic tests are significantly less expensive, faster, and require less skilled labour; yet, these tests provide reproducible data and information on several critical quality attributes, such as identity, purity, content, and disintegration. Visual and physical inspection also provides valuable information about the manufacturing and the labelling of medicines, and in many cases this inspection is sufficient to detect counterfeit medicines. The Promoting the Quality of Medicines (PQM) programme has provided technical assistance to Amazon Malaria Initiative (AMI) countries to implement the use of basic tests as a key screening mechanism to assess the quality of malaria medicines available to patients in decentralized regions. Trained personnel from the National Malaria Control Programmes (NMCPs), often in collaboration with country's Official Medicine Control Laboratory (OMCL), developed country- specific protocols that encompassed sampling methods, sample analysis, and data reporting. Sampling sites were selected based on malaria burden, accessibility, and geographical location. Convenience sampling was performed and countries were recommended to store the sampled medicines under conditions that did not compromise their quality. Basic analytical tests, such as disintegration and thin layer chromatography (TLC), were performed utilizing a portable mini-laboratory. Results were originally presented at regional meetings in a non-standardized format that lacked relevant medicines information. However, since 2008 information has been submitted utilizing a template specifically developed by PQM for that purpose. From 2005 to 2010, the quality of 1,663 malaria medicines from seven AMI countries was evaluated, mostly collected from the public sector, 1,445/1,663 (86.9%). Results indicate that 193/1,663 (11.6%) were found not to meet quality specifications. Most failures were reported during visual and physical inspection, 142/1663 (8.5%), and most of these were due to expired medicines, 118/142 (83.1%). Samples failing TLC accounted for 27/1,663 (1.6%) and those failing disintegration accounted for 24/1,663 (1.4%). Medicines quality failures decreased significantly during the last two years. Basic tests revealed that the quality of medicines in the public sector improved over the years, since the implementation of this type of quality monitoring programme in 2005. However, the lack of consistent confirmatory tests in the quality control (QC) laboratory, utilizing methods that can also evaluate additional quality attributes, could still mask quality issues. In the future, AMI countries should improve coordination with their health authorities and their QC lab consistently, to provide a more complete picture of malaria medicines quality and support the implementation of corrective actions. Facilities in the private and informal sectors also should be included when these sectors constitute an important source of medicines used by malaria patients.
A simple risk scoring system for prediction of relapse after inpatient alcohol treatment.
Pedersen, Mads Uffe; Hesse, Morten
2009-01-01
Predicting relapse after alcoholism treatment can be useful in targeting patients for aftercare services. However, a valid and practical instrument for predicting relapse risk does not exist. Based on a prospective study of alcoholism treatment, we developed the Risk of Alcoholic Relapse Scale (RARS) using items taken from the Addiction Severity Index and some basic demographic information. The RARS was cross-validated using two non-overlapping samples, and tested for its ability to predict relapse across different models of treatment. The RARS predicted relapse to drinking within 6 months after alcoholism treatment in both the original and the validation sample, and in a second validation sample it predicted admission to new treatment 3 years after treatment. The RARS can identify patients at high risk of relapse who need extra aftercare and support after treatment.
Veldhoen, Nik; Propper, Catherine R; Helbing, Caren C
2014-03-01
Studies performed across diverse frog species have made substantial contributions to our understanding of basic vertebrate development and the natural or anthropogenic environmental factors impacting sensitive life stages. Because, anurans are developmental models, provide ecosystems services, and act as sentinels for the identification of environmental chemical contaminants that interfere with thyroid hormone (TH) action during postembryonic development, there is demand for flexible assessment techniques that can be applied to multiple species. As part of the "thyroid assays across indicator and sentinel species" (TAXISS) initiative, we have designed and validated a series of cross-species real time quantitative PCR (qPCR) primer sets that provide information on transcriptome components in evolutionarily distant anurans. Validation for fifteen gene transcripts involved a rigorous three-tiered quality control within tissue/development-specific contexts. Assay performance was confirmed on multiple tissues (tail fin, liver, brain, and intestine) of Rana catesbeiana and Xenopus laevis tadpoles enabling comparisons between tissues and generation of response profiles to exogenous TH. This revealed notable differences in TH-responsive gene transcripts including thra, thrb, thibz, klf9, col1a2, fn1, plp1, mmp2, timm50, otc, and dio2, suggesting differential regulation and susceptibility to contaminant effects. Evidence for the applicability of the TAXISS anuran qPCR assay across seven other species is also provided with five frog families represented and its utility in defining genome structure was demonstrated. This novel validated approach will enable meaningful comparative studies between frog species and aid in extending knowledge of developmental regulatory pathways and the impact of environmental factors on TH signaling in frog species for which little or no genetic information is currently available. Copyright © 2014 Elsevier B.V. All rights reserved.
Ortiz, Glorimar; Schacht, Lucille
2012-01-01
Measurement of consumers' satisfaction in psychiatric settings is important because it has been correlated with improved clinical outcomes and administrative measures of high-quality care. These consumer satisfaction measurements are actively used as performance measures required by the accreditation process and for quality improvement activities. Our objectives were (i) to re-evaluate, through exploratory factor analysis (EFA) and confirmatory factor analysis (CFA), the structure of an instrument intended to measure consumers' satisfaction with care in psychiatric settings and (ii) to examine and publish the psychometric characteristics, validity and reliability, of the Inpatient Consumer Survey (ICS). To psychometrically test the structure of the ICS, 34 878 survey results, submitted by 90 psychiatric hospitals in 2008, were extracted from the Behavioral Healthcare Performance Measurement System (BHPMS). Basic descriptive item-response and correlation analyses were performed for total surveys. Two datasets were randomly created for analysis. A random sample of 8229 survey results was used for EFA. Another random sample of 8261 consumer survey results was used for CFA. This same sample was used to perform validity and reliability analyses. The item-response analysis showed that the mean range for a disagree/agree five-point scale was 3.10-3.94. Correlation analysis showed a strong relationship between items. Six domains (dignity, rights, environment, empowerment, participation, and outcome) with internal reliabilities between good to moderate (0.87-0.73) were shown to be related to overall care satisfaction. Overall reliability for the instrument was excellent (0.94). Results from CFA provided support for the domains structure of the ICS proposed through EFA. The overall findings from this study provide evidence that the ICS is a reliable measure of consumer satisfaction in psychiatric inpatient settings. The analysis has shown the ICS to provide valid and reliable results and to focus on the specific concerns of consumers of psychiatric inpatient care. Scores by item indicate that opportunity for improvement exists across healthcare organizations.
Development and necessary norms of reasoning
Markovits, Henry
2014-01-01
The question of whether reasoning can, or should, be described by a single normative model is an important one. In the following, I combine epistemological considerations taken from Piaget’s notion of genetic epistemology, a hypothesis about the role of reasoning in communication and developmental data to argue that some basic logical principles are in fact highly normative. I argue here that explicit, analytic human reasoning, in contrast to intuitive reasoning, uniformly relies on a form of validity that allows distinguishing between valid and invalid arguments based on the existence of counterexamples to conclusions. PMID:24904501
Bayesian truthing and experimental validation in homeland security and defense
NASA Astrophysics Data System (ADS)
Jannson, Tomasz; Forrester, Thomas; Wang, Wenjian; Kostrzewski, Andrew; Pradhan, Ranjit
2014-05-01
In this paper we discuss relations between Bayesian Truthing (experimental validation), Bayesian statistics, and Binary Sensing in the context of selected Homeland Security and Intelligence, Surveillance, Reconnaissance (ISR) optical and nonoptical application scenarios. The basic Figure of Merit (FoM) is Positive Predictive Value (PPV), as well as false positives and false negatives. By using these simple binary statistics, we can analyze, classify, and evaluate a broad variety of events including: ISR; natural disasters; QC; and terrorism-related, GIS-related, law enforcement-related, and other C3I events.
Lee, Ciaran M; Zhu, Haibao; Davis, Timothy H; Deshmukh, Harshahardhan; Bao, Gang
2017-01-01
The CRISPR/Cas9 system is a powerful tool for precision genome editing. The ability to accurately modify genomic DNA in situ with single nucleotide precision opens up new possibilities for not only basic research but also biotechnology applications and clinical translation. In this chapter, we outline the procedures for design, screening, and validation of CRISPR/Cas9 systems for targeted modification of coding sequences in the human genome and how to perform genome editing in induced pluripotent stem cells with high efficiency and specificity.
NASA Technical Reports Server (NTRS)
Humphreys, William M., Jr.; Bartram, Scott M.
2001-01-01
A novel multiple-camera system for the recording of digital particle image velocimetry (DPIV) images acquired in a two-dimensional separating/reattaching flow is described. The measurements were performed in the NASA Langley Subsonic Basic Research Tunnel as part of an overall series of experiments involving the simultaneous acquisition of dynamic surface pressures and off-body velocities. The DPIV system utilized two frequency-doubled Nd:YAG lasers to generate two coplanar, orthogonally polarized light sheets directed upstream along the horizontal centerline of the test model. A recording system containing two pairs of matched high resolution, 8-bit cameras was used to separate and capture images of illuminated tracer particles embedded in the flow field. Background image subtraction was used to reduce undesirable flare light emanating from the surface of the model, and custom pixel alignment algorithms were employed to provide accurate registration among the various cameras. Spatial cross correlation analysis with median filter validation was used to determine the instantaneous velocity structure in the separating/reattaching flow region illuminated by the laser light sheets. In operation the DPIV system exhibited a good ability to resolve large-scale separated flow structures with acceptable accuracy over the extended field of view of the cameras. The recording system design provided enhanced performance versus traditional DPIV systems by allowing a variety of standard and non-standard cameras to be easily incorporated into the system.
Harkness, Kate L; Monroe, Scott M
2016-07-01
Life stress is a central factor in the onset and course of a wide range of medical and psychiatric conditions. Determining the precise etiological and pathological consequences of stress, though, has been hindered by weaknesses in prevailing definitional and measurement practices. The purpose of the current paper is to evaluate the primary strategies for defining and measuring major and minor acute life events, chronic stressors, and daily hassles as informed by 3 basic scientific premises. The first premise concerns the manner in which stress is conceptualized and operationally defined, and specifically we assert that stress measures must not conflate the stress exposure with the stress response. The second premise concerns how stress exposures are measured, and we provide guidelines for optimizing standardized and sensitive indicators of life stress. The third premise addresses the consequences of variations in the procedures for life event measurement with regard to the validity of the research designs employed. We show that life stress measures are susceptible to several sources of bias, and if these potential sources of bias are not controlled in the design of the research, spurious findings may result. Our goal is to provide a useful guide for researchers who consider life stress to be an important factor in their theoretical models of disease, wish to incorporate measures of life stress in their research, and seek to avoid the common pitfalls of past measurement practices. (PsycINFO Database Record (c) 2016 APA, all rights reserved).
2017-01-01
Background Palliative care is nowadays essential in nursing care, due to the increasing number of patients who require attention in final stages of their life. Nurses need to acquire specific knowledge and abilities to provide quality palliative care. Palliative Care Quiz for Nurses is a questionnaire that evaluates their basic knowledge about palliative care. The Palliative Care Quiz for Nurses (PCQN) is useful to evaluate basic knowledge about palliative care, but its adaptation into the Spanish language and the analysis of its effectiveness and utility for Spanish culture is lacking. Purpose To report the adaptation into the Spanish language and the psychometric analysis of the Palliative Care Quiz for Nurses. Method The Palliative Care Quiz for Nurses-Spanish Version (PCQN-SV) was obtained from a process including translation, back-translation, comparison with versions in other languages, revision by experts, and pilot study. Content validity and reliability of questionnaire were analyzed. Difficulty and discrimination indexes of each item were also calculated according to Item Response Theory (IRT). Findings Adequate internal consistency was found (S-CVI = 0.83); Cronbach's alpha coefficient of 0.67 and KR-20 test result of 0,72 reflected the reliability of PCQN-SV. The questionnaire had a global difficulty index of 0,55, with six items which could be considered as difficult or very difficult, and five items with could be considered easy or very easy. The discrimination indexes of the 20 items, show us that eight items are good or very good while six items are bad to discriminate between good and bad respondents. Discussion Although in shows internal consistency, reliability and difficulty indexes similar to those obtained by versions of PCQN in other languages, a reformulation of the items with lowest content validity or discrimination indexes and those showing difficulties with their comprehension is an aspect to take into account in order to improve the PCQN-SV. Conclusion The PCQN-SV is a useful Spanish language instrument for measuring Spanish nurses’ knowledge in palliative care and it is adequate to establish international comparisons. PMID:28545037
Chover-Sierra, Elena; Martínez-Sabater, Antonio; Lapeña-Moñux, Yolanda Raquel
2017-01-01
Palliative care is nowadays essential in nursing care, due to the increasing number of patients who require attention in final stages of their life. Nurses need to acquire specific knowledge and abilities to provide quality palliative care. Palliative Care Quiz for Nurses is a questionnaire that evaluates their basic knowledge about palliative care. The Palliative Care Quiz for Nurses (PCQN) is useful to evaluate basic knowledge about palliative care, but its adaptation into the Spanish language and the analysis of its effectiveness and utility for Spanish culture is lacking. To report the adaptation into the Spanish language and the psychometric analysis of the Palliative Care Quiz for Nurses. The Palliative Care Quiz for Nurses-Spanish Version (PCQN-SV) was obtained from a process including translation, back-translation, comparison with versions in other languages, revision by experts, and pilot study. Content validity and reliability of questionnaire were analyzed. Difficulty and discrimination indexes of each item were also calculated according to Item Response Theory (IRT). Adequate internal consistency was found (S-CVI = 0.83); Cronbach's alpha coefficient of 0.67 and KR-20 test result of 0,72 reflected the reliability of PCQN-SV. The questionnaire had a global difficulty index of 0,55, with six items which could be considered as difficult or very difficult, and five items with could be considered easy or very easy. The discrimination indexes of the 20 items, show us that eight items are good or very good while six items are bad to discriminate between good and bad respondents. Although in shows internal consistency, reliability and difficulty indexes similar to those obtained by versions of PCQN in other languages, a reformulation of the items with lowest content validity or discrimination indexes and those showing difficulties with their comprehension is an aspect to take into account in order to improve the PCQN-SV. The PCQN-SV is a useful Spanish language instrument for measuring Spanish nurses' knowledge in palliative care and it is adequate to establish international comparisons.
Allen Gomes, Ana; Ruivo Marques, Daniel; Meia-Via, Ana Maria; Meia-Via, Mariana; Tavares, José; Fernandes da Silva, Carlos; Pinto de Azevedo, Maria Helena
2015-04-01
Based on successive samples totaling more than 5000 higher education students, we scrutinized the reliability, structure, initial validity and normative scores of a brief self-report seven-item scale to screen for the continuum of nighttime insomnia complaints/perceived sleep quality, used by our team for more than a decade, henceforth labeled the Basic Scale on Insomnia complaints and Quality of Sleep (BaSIQS). In study/sample 1 (n = 1654), the items were developed based on part of a larger survey on higher education sleep-wake patterns. The test-retest study was conducted in an independent small group (n = 33) with a 2-8 week gap. In study/sample 2 (n = 360), focused mainly on validity, the BaSIQS was completed together with the Pittsburgh Sleep Quality Index (PSQI). In study 3, a large recent sample of students from universities all over the country (n = 2995) answered the BaSIQS items, based on which normative scores were determined, and an additional question on perceived sleep problems in order to further analyze the scale's validity. Regarding reliability, Cronbach alpha coefficients were systematically higher than 0.7, and the test-retest correlation coefficient was greater than 0.8. Structure analyses revealed consistently satisfactory two-factor and single-factor solutions. Concerning validity analyses, BaSIQS scores were significantly correlated with PSQI component scores and overall score (r = 0.652 corresponding to a large association); mean scores were significantly higher in those students classifying themselves as having sleep problems (p < 0.0001, d = 0.99 corresponding to a large effect size). In conclusion, the BaSIQS is very easy to administer, and appears to be a reliable and valid scale in higher education students. It might be a convenient short tool in research and applied settings to rapidly assess sleep quality or screen for insomnia complaints, and it may be easily used in other populations with minor adaptations.
Westbury, Chris; Keith, Jeff; Briesemeister, Benny B; Hofmann, Markus J; Jacobs, Arthur M
2015-01-01
Ever since Aristotle discussed the issue in Book II of his Rhetoric, humans have attempted to identify a set of "basic emotion labels". In this paper we propose an algorithmic method for evaluating sets of basic emotion labels that relies upon computed co-occurrence distances between words in a 12.7-billion-word corpus of unselected text from USENET discussion groups. Our method uses the relationship between human arousal and valence ratings collected for a large list of words, and the co-occurrence similarity between each word and emotion labels. We assess how well the words in each of 12 emotion label sets-proposed by various researchers over the past 118 years-predict the arousal and valence ratings on a test and validation dataset, each consisting of over 5970 items. We also assess how well these emotion labels predict lexical decision residuals (LDRTs), after co-varying out the effects attributable to basic lexical predictors. We then demonstrate a generalization of our method to determine the most predictive "basic" emotion labels from among all of the putative models of basic emotion that we considered. As well as contributing empirical data towards the development of a more rigorous definition of basic emotions, our method makes it possible to derive principled computational estimates of emotionality-specifically, of arousal and valence-for all words in the language.
Porto, Paolo; Walling, Des E
2012-10-01
Information on rates of soil loss from agricultural land is a key requirement for assessing both on-site soil degradation and potential off-site sediment problems. Many models and prediction procedures have been developed to estimate rates of soil loss and soil redistribution as a function of the local topography, hydrometeorology, soil type and land management, but empirical data remain essential for validating and calibrating such models and prediction procedures. Direct measurements using erosion plots are, however, costly and the results obtained relate to a small enclosed area, which may not be representative of the wider landscape. In recent years, the use of fallout radionuclides and more particularly caesium-137 ((137)Cs) and excess lead-210 ((210)Pb(ex)) has been shown to provide a very effective means of documenting rates of soil loss and soil and sediment redistribution in the landscape. Several of the assumptions associated with the theoretical conversion models used with such measurements remain essentially unvalidated. This contribution describes the results of a measurement programme involving five experimental plots located in southern Italy, aimed at validating several of the basic assumptions commonly associated with the use of mass balance models for estimating rates of soil redistribution on cultivated land from (137)Cs and (210)Pb(ex) measurements. Overall, the results confirm the general validity of these assumptions and the importance of taking account of the fate of fresh fallout. However, further work is required to validate the conversion models employed in using fallout radionuclide measurements to document soil redistribution in the landscape and this could usefully direct attention to different environments and to the validation of the final estimates of soil redistribution rate as well as the assumptions of the models employed. Copyright © 2012 Elsevier Ltd. All rights reserved.
38 CFR 3.812 - Special allowance payable under section 156 of Pub. L. 97-377.
Code of Federal Regulations, 2014 CFR
2014-07-01
... consideration, and death on active duty subsequent to August 12, 1981, is qualifying provided that the death... payment rate—(1) Basic entitlement rate. A basic entitlement rate will be computed for each eligible...-377 using data to be provided by the Social Security Administration. This basic entitlement rate will...
38 CFR 3.812 - Special allowance payable under section 156 of Pub. L. 97-377.
Code of Federal Regulations, 2010 CFR
2010-07-01
... consideration, and death on active duty subsequent to August 12, 1981, is qualifying provided that the death... payment rate—(1) Basic entitlement rate. A basic entitlement rate will be computed for each eligible...-377 using data to be provided by the Social Security Administration. This basic entitlement rate will...
38 CFR 3.812 - Special allowance payable under section 156 of Pub. L. 97-377.
Code of Federal Regulations, 2013 CFR
2013-07-01
... consideration, and death on active duty subsequent to August 12, 1981, is qualifying provided that the death... payment rate—(1) Basic entitlement rate. A basic entitlement rate will be computed for each eligible...-377 using data to be provided by the Social Security Administration. This basic entitlement rate will...
38 CFR 3.812 - Special allowance payable under section 156 of Pub. L. 97-377.
Code of Federal Regulations, 2012 CFR
2012-07-01
... consideration, and death on active duty subsequent to August 12, 1981, is qualifying provided that the death... payment rate—(1) Basic entitlement rate. A basic entitlement rate will be computed for each eligible...-377 using data to be provided by the Social Security Administration. This basic entitlement rate will...
38 CFR 3.812 - Special allowance payable under section 156 of Pub. L. 97-377.
Code of Federal Regulations, 2011 CFR
2011-07-01
... consideration, and death on active duty subsequent to August 12, 1981, is qualifying provided that the death... payment rate—(1) Basic entitlement rate. A basic entitlement rate will be computed for each eligible...-377 using data to be provided by the Social Security Administration. This basic entitlement rate will...
ERIC Educational Resources Information Center
Ohio State Univ., Columbus. National Center for Research in Vocational Education.
This guide provides information and guidelines intended to assist vocational administrators in developing and evaluating programs to improve the basic skills of vocational-technical students. Part one provides background information about basic skills and examines their role in vocational education. Discussed next are various program types,…
Phase 2 STS new user development program. Volume 1: Executive summary
NASA Technical Reports Server (NTRS)
Mcdowell, J. R.
1976-01-01
A methodology for developing new users for STS other than NASA and DoD, thereby maximizing the use of the STS system was developed. The approach to user development, reflected in the implementation plan, and attendant informational material to be used were evaluated by conducting a series of test cases with selected user organizations. These test case organizations were, in effect, used as consultants to evaluate the effectiveness, the needs, the completeness, and the adequacy of the user development approach and informational material. The selection of the test cases provided a variety of potential STS users covering industry, other government agencies, and the educational sector. The test cases covered various use areas and provided a mix of user organization types. A summary of the actual test cases conducted is given. The conduct of the test cases verified the general approach of the implementation plan, the validity of the user development strategy prepared for each test case organization and the effectiveness of the STS basic and user customized informational material.
Instrumental variable methods in comparative safety and effectiveness research†
Brookhart, M. Alan; Rassen, Jeremy A.; Schneeweiss, Sebastian
2010-01-01
Summary Instrumental variable (IV) methods have been proposed as a potential approach to the common problem of uncontrolled confounding in comparative studies of medical interventions, but IV methods are unfamiliar to many researchers. The goal of this article is to provide a non-technical, practical introduction to IV methods for comparative safety and effectiveness research. We outline the principles and basic assumptions necessary for valid IV estimation, discuss how to interpret the results of an IV study, provide a review of instruments that have been used in comparative effectiveness research, and suggest some minimal reporting standards for an IV analysis. Finally, we offer our perspective of the role of IV estimation vis-à-vis more traditional approaches based on statistical modeling of the exposure or outcome. We anticipate that IV methods will be often underpowered for drug safety studies of very rare outcomes, but may be potentially useful in studies of intended effects where uncontrolled confounding may be substantial. PMID:20354968
The Effectiveness of Streaming Video on Medical Student Learning: A Case Study
Bridge, Patrick D.; Jackson, Matt; Robinson, Leah
2009-01-01
Information technology helps meet today's medical students’ needs by providing multiple curriculum delivery methods. Video streaming is an e-learning technology that uses the Internet to deliver curriculum while giving the student control of the content's delivery. There have been few studies conducted on the effectiveness of streaming video in medical schools. A 5-year retrospective study was conducted using three groups of students (n = 1736) to determine if the availability of streaming video in Years 1–2 of the basic science curriculum affected overall Step 1 scores for first-time test-takers. The results demonstrated a positive effect on program outcomes as streaming video became more readily available to students. Based on these findings, streaming video technology seems to be a viable tool to complement in-class delivery methods, to accommodate the needs of medical students, and to provide options for meeting the challenges of delivering the undergraduate medical curriculum. Further studies need to be conducted to continue validating the effectiveness of streaming video technology. PMID:20165525
Operational properties of fluctuation X-ray scattering data
Malmerberg, Erik; Kerfeld, Cheryl A.; Zwart, Petrus H.
2015-03-20
X-ray scattering images collected on timescales shorter than rotation diffusion times using a (partially) coherent beam result in a significant increase in information content in the scattered data. These measurements, named fluctuation X-ray scattering (FXS), are typically performed on an X-ray free-electron laser (XFEL) and can provide fundamental insights into the structure of biological molecules, engineered nanoparticles or energy-related mesoscopic materials beyond what can be obtained with standard X-ray scattering techniques. In order to understand, use and validate experimental FXS data, the availability of basic data characteristics and operational properties is essential, but has been absent up to this point.more » In this communication, an intuitive view of the nature of FXS data and their properties is provided, the effect of FXS data on the derived structural models is highlighted, and generalizations of the Guinier and Porod laws that can ultimately be used to plan experiments and assess the quality of experimental data are presented.« less
The effectiveness of streaming video on medical student learning: a case study.
Bridge, Patrick D; Jackson, Matt; Robinson, Leah
2009-08-19
Information technology helps meet today's medical students' needs by providing multiple curriculum delivery methods. Video streaming is an e-learning technology that uses the Internet to deliver curriculum while giving the student control of the content's delivery. There have been few studies conducted on the effectiveness of streaming video in medical schools. A 5-year retrospective study was conducted using three groups of students (n = 1736) to determine if the availability of streaming video in Years 1-2 of the basic science curriculum affected overall Step 1 scores for first-time test-takers. The results demonstrated a positive effect on program outcomes as streaming video became more readily available to students. Based on these findings, streaming video technology seems to be a viable tool to complement in-class delivery methods, to accommodate the needs of medical students, and to provide options for meeting the challenges of delivering the undergraduate medical curriculum. Further studies need to be conducted to continue validating the effectiveness of streaming video technology.
No. 127-The Evaluation of Stress Incontinence Prior to Primary Surgery.
Farrell, Scott A
2018-02-01
To provide clinical guidelines for the evaluation of women with stress urinary incontinence prior to primary anti-incontinence surgery. The modalities of evaluation range from basic pelvic examination through to the use of adjuncts including ultrasound and urodynamic testing. These guidelines provide a comprehensive approach to the preoperative evaluation of urinary incontinence to ensure that excessive evaluation is avoided without sacrificing diagnostic accuracy. Published opinions of experts, supplemented by evidence from clinical trials, where appropriate. The quality of the evidence is rated using the criteria described by the Canadian Task Force on the Periodic Health Examination. Comprehensive evaluation of women considering surgery to treat urinary incontinence is essential to rule out causes of incontinence that may not be amenable to surgical treatment. Simplifying the evaluation minimizes the discomfort and embarrassment potentially experienced by women. VALIDATION: These guidelines have been approved by the Urogynaecology Committee and the Executive and Council of The Society of Obstetricians and Gynaecologists of Canada. Copyright © 2018. Published by Elsevier Inc.
SEPT9: A Specific Circulating Biomarker for Colorectal Cancer.
Song, Lele; Li, Yuemin
2015-01-01
SEPT9 gene methylation has been implicated as a biomarker for colorectal cancer (CRC) for more than 10 years and has been used clinically for more than 6 years. Studies have proven it to be an accurate, reliable, fast, and convenient method for CRC. In this chapter, we will first provide the background on the role of septin9 protein and the theoretical basis of the SEPT9 gene methylation assay. We will then focus on the performance of SEPT9 gene methylation assay for CRC early detection and screening by analyzing the data obtained in clinical trials and comparing its performance with other methods or markers. Finally, we will discuss the future applications of the assay in monitoring cancer recurrence, evaluating surgery, chemotherapy, and predicting long-term survival. We hope this chapter can provide a full overview of the theoretical basis, development, validation, and clinical applications of the SEPT9 assay for both basic science researchers and clinical practitioners. © 2015 Elsevier Inc. All rights reserved.
Chen, Fanxiu; Zhuang, Qi; Zhang, Huixin
2016-06-20
The mechanical behaviors of granular materials are governed by the grain properties and microstructure of the materials. We conducted experiments to study the force transmission in granular materials using plane strain tests. The large amount of nearly continuous displacement data provided by the advanced noncontact experimental technique of digital image correlation (DIC) has provided a means to quantify local displacements and strains at the particle level. The average strain of each particle could be calculated based on the DIC method, and the average stress could be obtained using Hooke's law. The relationship between the stress and particle force could be obtained based on basic Newtonian mechanics and the balance of linear momentum at the particle level. This methodology is introduced and validated. In the testing procedure, the system is tested in real 2D particle cases, and the contact forces and force chain are obtained and analyzed. The system has great potential for analyzing a real granular system and measuring the contact forces and force chain.
The Definition of Basic Skills in Manufacturing Industries.
ERIC Educational Resources Information Center
Jones, Charles M.
1996-01-01
Survey responses from 186 of 250 trainers/human resource managers and a Delphi panel of 21 rated companies' basic skills training provisions and the importance of basic skills. Although most stated their companies provide basic skills training, results show most are reluctant to hire, train, or retain workers with low basic skills. (SK)
Bax, Leon; Yu, Ly-Mee; Ikeda, Noriaki; Tsuruta, Harukazu; Moons, Karel GM
2006-01-01
Background Meta-analysis has become a well-known method for synthesis of quantitative data from previously conducted research in applied health sciences. So far, meta-analysis has been particularly useful in evaluating and comparing therapies and in assessing causes of disease. Consequently, the number of software packages that can perform meta-analysis has increased over the years. Unfortunately, it can take a substantial amount of time to get acquainted with some of these programs and most contain little or no interactive educational material. We set out to create and validate an easy-to-use and comprehensive meta-analysis package that would be simple enough programming-wise to remain available as a free download. We specifically aimed at students and researchers who are new to meta-analysis, with important parts of the development oriented towards creating internal interactive tutoring tools and designing features that would facilitate usage of the software as a companion to existing books on meta-analysis. Results We took an unconventional approach and created a program that uses Excel as a calculation and programming platform. The main programming language was Visual Basic, as implemented in Visual Basic 6 and Visual Basic for Applications in Excel 2000 and higher. The development took approximately two years and resulted in the 'MIX' program, which can be downloaded from the program's website free of charge. Next, we set out to validate the MIX output with two major software packages as reference standards, namely STATA (metan, metabias, and metatrim) and Comprehensive Meta-Analysis Version 2. Eight meta-analyses that had been published in major journals were used as data sources. All numerical and graphical results from analyses with MIX were identical to their counterparts in STATA and CMA. The MIX program distinguishes itself from most other programs by the extensive graphical output, the click-and-go (Excel) interface, and the educational features. Conclusion The MIX program is a valid tool for performing meta-analysis and may be particularly useful in educational environments. It can be downloaded free of charge via or . PMID:17038197
Bax, Leon; Yu, Ly-Mee; Ikeda, Noriaki; Tsuruta, Harukazu; Moons, Karel G M
2006-10-13
Meta-analysis has become a well-known method for synthesis of quantitative data from previously conducted research in applied health sciences. So far, meta-analysis has been particularly useful in evaluating and comparing therapies and in assessing causes of disease. Consequently, the number of software packages that can perform meta-analysis has increased over the years. Unfortunately, it can take a substantial amount of time to get acquainted with some of these programs and most contain little or no interactive educational material. We set out to create and validate an easy-to-use and comprehensive meta-analysis package that would be simple enough programming-wise to remain available as a free download. We specifically aimed at students and researchers who are new to meta-analysis, with important parts of the development oriented towards creating internal interactive tutoring tools and designing features that would facilitate usage of the software as a companion to existing books on meta-analysis. We took an unconventional approach and created a program that uses Excel as a calculation and programming platform. The main programming language was Visual Basic, as implemented in Visual Basic 6 and Visual Basic for Applications in Excel 2000 and higher. The development took approximately two years and resulted in the 'MIX' program, which can be downloaded from the program's website free of charge. Next, we set out to validate the MIX output with two major software packages as reference standards, namely STATA (metan, metabias, and metatrim) and Comprehensive Meta-Analysis Version 2. Eight meta-analyses that had been published in major journals were used as data sources. All numerical and graphical results from analyses with MIX were identical to their counterparts in STATA and CMA. The MIX program distinguishes itself from most other programs by the extensive graphical output, the click-and-go (Excel) interface, and the educational features. The MIX program is a valid tool for performing meta-analysis and may be particularly useful in educational environments. It can be downloaded free of charge via http://www.mix-for-meta-analysis.info or http://sourceforge.net/projects/meta-analysis.
Mastoidectomy performance assessment of virtual simulation training using final-product analysis.
Andersen, Steven A W; Cayé-Thomasen, Per; Sørensen, Mads S
2015-02-01
The future development of integrated automatic assessment in temporal bone virtual surgical simulators calls for validation against currently established assessment tools. This study aimed to explore the relationship between mastoidectomy final-product performance assessment in virtual simulation and traditional dissection training. Prospective trial with blinding. A total of 34 novice residents performed a mastoidectomy on the Visible Ear Simulator and on a cadaveric temporal bone. Two blinded senior otologists assessed the final-product performance using a modified Welling scale. The simulator gathered basic metrics on time, steps, and volumes in relation to the on-screen tutorial and collisions with vital structures. Substantial inter-rater reliability (kappa = 0.77) for virtual simulation and moderate inter-rater reliability (kappa = 0.59) for dissection final-product assessment was found. The simulation and dissection performance scores had significant correlation (P = .014). None of the basic simulator metrics correlated significantly with the final-product score except for number of steps completed in the simulator. A modified version of a validated final-product performance assessment tool can be used to assess mastoidectomy on virtual temporal bones. Performance assessment of virtual mastoidectomy could potentially save the use of cadaveric temporal bones for more advanced training when a basic level of competency in simulation has been achieved. NA. © 2014 The American Laryngological, Rhinological and Otological Society, Inc.
Lifeline: A Tool for Logistics Professionals
2017-06-01
proof of concept study is designed to provide a basic understanding of the Supply Corps community, provide a comparative analysis of the organizational...concept study is designed to provide a basic understanding of the Supply Corps community, provide a comparative analysis of the organizational...APPLICATION) ......................................................................................63 G. DESIGN
Understanding and Observing Subglacial Friction Using Seismology
NASA Astrophysics Data System (ADS)
Tsai, V. C.
2017-12-01
Glaciology began with a focus on understanding basic mechanical processes and producing physical models that could explain the principal observations. Recently, however, more attention has been paid to the wealth of recent observations, with many modeling efforts relying on data assimilation and empirical scalings, rather than being based on first-principles physics. Notably, ice sheet models commonly assume that subglacial friction is characterized by a "slipperiness" coefficient that is determined by inverting surface velocity observations. Predictions are usually then made by assuming these slipperiness coefficients are spatially and temporally fixed. However, this is only valid if slipperiness is an unchanging material property of the bed and, despite decades of work on subglacial friction, it has remained unclear how to best account for such subglacial physics in ice sheet models. Here, we describe how basic seismological concepts and observations can be used to improve our understanding and determination of subglacial friction. First, we discuss how standard models of granular friction can and should be used in basal friction laws for marine ice sheets, where very low effective pressures exist. We show that under realistic West Antarctic Ice Sheet conditions, standard Coulomb friction should apply in a relatively narrow zone near the grounding line and that this should transition abruptly as one moves inland to a different, perhaps Weertman-style, dependence of subglacial stress on velocity. We show that this subglacial friction law predicts significantly different ice sheet behavior even as compared with other friction laws that include effective pressure. Secondly, we explain how seismological observations of water flow noise and basal icequakes constrain subglacial physics in important ways. Seismically observed water flow noise can provide constraints on water pressures and channel sizes and geometry, leading to important data on subglacial friction. Basal icequake mechanisms also provide unique constraints on subglacial stress state as well as variations in water pressures. Together, the use of standard seismological concepts and new observations thus promises to provide new constraints on subglacial mechanics and focus attention back on the basic physical processes involved.
Interactive basic mathematics web using Wordpress
NASA Astrophysics Data System (ADS)
Septia, Tika; Husna; Cesaria, Anna
2017-12-01
Wordpress is a popular open source tool that can be used for developing learning media. Basic Mathematics is the difficult subject for a physics student. The students need an interactive learning to improve their knowledge. The aims of this study were to develop the interactive media using Wordpress and to know the effectiveness of web as a learning media to improve the ICT Literacy students. This study used ADDIE models. The effectiveness of interactive web can be described as the students’ equipness of ICT literacy. The population is physics students. The findings show that the interactive web is valid for the content, presentation, linguistic, and graphic aspects. The results concluded that basic mathematic interactive web is effective to equip the learners ICT literacy of categories of high, medium, and low with the observations and questionnaires are in very good criteria.
Are I.Q. Tests Valid Measures of Intelligence? An Interview with Howard Gardner.
ERIC Educational Resources Information Center
School Administrator, 1986
1986-01-01
Discuss seven basic intellectual competencies possessed by all human beings: linguistic, logical-mathematical, musical, spatial, bodily-kinesthetic, interpersonal, and intrapersonal. Because I.Q. tests can measure only the first two qualities, this interview suggests evaluating schools by quality of educational experiences and students by project…
DOT National Transportation Integrated Search
1968-09-01
The study concerns the experimental use of several commercially-published aptitude tests and a determination of their validity as predictors of training-performance grades for more than 300 Marines and over 600 Naval students who entered a basic air-...
A Diagnostic Assessment for Introductory Molecular and Cell Biology
ERIC Educational Resources Information Center
Shi, Jia; Wood, William B.; Martin, Jennifer M.; Guild, Nancy A.; Vicens, Quentin; Knight, Jennifer K.
2010-01-01
We have developed and validated a tool for assessing understanding of a selection of fundamental concepts and basic knowledge in undergraduate introductory molecular and cell biology, focusing on areas in which students often have misconceptions. This multiple-choice Introductory Molecular and Cell Biology Assessment (IMCA) instrument is designed…
Keeping in Character: A Time-Tested Solution.
ERIC Educational Resources Information Center
Benninga, Jaques S.; Wynne, Edward A.
1998-01-01
Refutes Alfie Kohn's criticisms of modern character education programs in the February 1997 "Kappan." The basic structure of true "for character" education relies on an approach relevant for students of all ages, has been time-tested over 2,500 years, has broad public support, and has a valid research base. Kohn advocates…
Validation of QTL in commercial-type pigs at USMARC
USDA-ARS?s Scientific Manuscript database
The US Meat Animal Research Center’s commercial swine resource population was a closed population maintained for 10 generations and produced over 40,000 pigs with basic phenotypic measurements. Approximately 3,000 of the most heavily phenotyped pigs as well as all boars used were genotyped with the ...
DOE Office of Scientific and Technical Information (OSTI.GOV)
Shaisultanov, Rashid; Eichler, David
2011-03-15
The dielectric tensor is obtained for a general anisotropic distribution function that is represented as a sum over Legendre polynomials. The result is valid over all of k-space. We obtain growth rates for the Weibel instability for some basic examples of distribution functions.
Environmental and Agricultural Sciences. Georgia Core Standards for Occupational Clusters.
ERIC Educational Resources Information Center
Georgia Univ., Athens. Dept. of Occupational Studies.
This document lists core standards and occupational knowledge amd skills that have been identified/validated by industry as necessary to all Georgia students in secondary-level environmental and agricultural sciences programs. First, foundation skills are grouped as follows: basic skills (reading, writing, arithmetic/mathematics, listening,…
Business, Marketing, and Information Management. Georgia Core Standards for Occupational Clusters.
ERIC Educational Resources Information Center
Georgia Univ., Athens. Dept. of Occupational Studies.
This document lists core standards and occupational knowledge and skills that have been identified and validated by industry as necessary to all Georgia students in business, marketing, and information management programs. First, foundation skills are grouped as follows: basic skills (reading, writing, arithmetic/mathematics, listening, speaking);…
Code of Federal Regulations, 2010 CFR
2010-10-01
... & Validation services), steady state operations, basic and applied research, and routine services (e.g... of the cognizant HCA. (b) EVM is not required, but may be applied with prior written approval of the...(s) under FAR Part 12. (c) When full EVM is required on a prime contract, it applies to subcontracts...
Styles and Style-Stretching: How Are They Related to Successful Learning?
ERIC Educational Resources Information Center
Griffiths, Carol; Inceçay, Görsev
2016-01-01
Although the learning style construct has aroused much interest over the years, questions remain regarding basic issues such as definition, the validity and/or reliability of various measurement instruments, and the relationship between learning style and successful learning. Furthermore, although maintaining stylistic flexibility is recommended…
Describes procedures written based on the assumption that they will be performed by analysts who are formally trained in at least the basic principles of chemical analysis and in the use of the subject technology.
Code of Federal Regulations, 2014 CFR
2014-07-01
... otherwise, to establish the validity and competence of his estimates. He must familiarize himself with basic..., 1971, dictates that written statement of, and summary of the basis for, the amount of the estimate of... Appraisers which sets out market value “as the highest price estimated in terms of money which a property...