Sample records for reliability analysis empirical

  1. A study of fault prediction and reliability assessment in the SEL environment

    NASA Technical Reports Server (NTRS)

    Basili, Victor R.; Patnaik, Debabrata

    1986-01-01

    An empirical study on estimation and prediction of faults, prediction of fault detection and correction effort, and reliability assessment in the Software Engineering Laboratory environment (SEL) is presented. Fault estimation using empirical relationships and fault prediction using curve fitting method are investigated. Relationships between debugging efforts (fault detection and correction effort) in different test phases are provided, in order to make an early estimate of future debugging effort. This study concludes with the fault analysis, application of a reliability model, and analysis of a normalized metric for reliability assessment and reliability monitoring during development of software.

  2. Alternative Methods for Calculating Intercoder Reliability in Content Analysis: Kappa, Weighted Kappa and Agreement Charts Procedures.

    ERIC Educational Resources Information Center

    Kang, Namjun

    If content analysis is to satisfy the requirement of objectivity, measures and procedures must be reliable. Reliability is usually measured by the proportion of agreement of all categories identically coded by different coders. For such data to be empirically meaningful, a high degree of inter-coder reliability must be demonstrated. Researchers in…

  3. Τhe observational and empirical thermospheric CO2 and NO power do not exhibit power-law behavior; an indication of their reliability

    NASA Astrophysics Data System (ADS)

    Varotsos, C. A.; Efstathiou, M. N.

    2018-03-01

    In this paper we investigate the evolution of the energy emitted by CO2 and NO from the Earth's thermosphere on a global scale using both observational and empirically derived data. In the beginning, we analyze the daily power observations of CO2 and NO received from the Sounding of the Atmosphere using Broadband Emission Radiometry (SABER) equipment on the NASA Thermosphere-Ionosphere-Mesosphere Energetics and Dynamics (TIMED) satellite for the entire period 2002-2016. We then perform the same analysis on the empirical daily power emitted by CO2 and NO that were derived recently from the infrared energy budget of the thermosphere during 1947-2016. The tool used for the analysis of the observational and empirical datasets is the detrended fluctuation analysis, in order to investigate whether the power emitted by CO2 and by NO from the thermosphere exhibits power-law behavior. The results obtained from both observational and empirical data do not support the establishment of the power-law behavior. This conclusion reveals that the empirically derived data are characterized by the same intrinsic properties as those of the observational ones, thus enhancing the validity of their reliability.

  4. Crowd-sourcing relative preferences for ecosystem services in the St. Louis River AOC

    EPA Science Inventory

    Analysis of ecosystem service tradeoffs among project scenarios is more reliable when valuation data are available. Empirical valuation data are expensive and difficult to collect. As a possible alternative or supplement to empirical data, we downloaded and classified images from...

  5. Issues in benchmarking human reliability analysis methods : a literature review.

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Lois, Erasmia; Forester, John Alan; Tran, Tuan Q.

    There is a diversity of human reliability analysis (HRA) methods available for use in assessing human performance within probabilistic risk assessment (PRA). Due to the significant differences in the methods, including the scope, approach, and underlying models, there is a need for an empirical comparison investigating the validity and reliability of the methods. To accomplish this empirical comparison, a benchmarking study is currently underway that compares HRA methods with each other and against operator performance in simulator studies. In order to account for as many effects as possible in the construction of this benchmarking study, a literature review was conducted,more » reviewing past benchmarking studies in the areas of psychology and risk assessment. A number of lessons learned through these studies are presented in order to aid in the design of future HRA benchmarking endeavors.« less

  6. Issues in Benchmarking Human Reliability Analysis Methods: A Literature Review

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Ronald L. Boring; Stacey M. L. Hendrickson; John A. Forester

    There is a diversity of human reliability analysis (HRA) methods available for use in assessing human performance within probabilistic risk assessments (PRA). Due to the significant differences in the methods, including the scope, approach, and underlying models, there is a need for an empirical comparison investigating the validity and reliability of the methods. To accomplish this empirical comparison, a benchmarking study comparing and evaluating HRA methods in assessing operator performance in simulator experiments is currently underway. In order to account for as many effects as possible in the construction of this benchmarking study, a literature review was conducted, reviewing pastmore » benchmarking studies in the areas of psychology and risk assessment. A number of lessons learned through these studies are presented in order to aid in the design of future HRA benchmarking endeavors.« less

  7. Development and empirical validation of symmetric component measures of multidimensional constructs: customer and competitor orientation.

    PubMed

    Sørensen, Hans Eibe; Slater, Stanley F

    2008-08-01

    Atheoretical measure purification may lead to construct deficient measures. The purpose of this paper is to provide a theoretically driven procedure for the development and empirical validation of symmetric component measures of multidimensional constructs. Particular emphasis is placed on establishing a formalized three-step procedure for achieving a posteriori content validity. Then the procedure is applied to development and empirical validation of two symmetrical component measures of market orientation, customer orientation and competitor orientation. Analysis suggests that average variance extracted is particularly critical to reliability in the respecification of multi-indicator measures. In relation to this, the results also identify possible deficiencies in using Cronbach alpha for establishing reliable and valid measures.

  8. Reliability, Validity, and Factor Structure of the Imaginative Capability Scale

    ERIC Educational Resources Information Center

    Liang, Chaoyun; Chia, Tsorng-Lin

    2014-01-01

    Three studies were combined to test the reliability, validity, and factor structure of the imaginative capability scale (ICS). The ICS was a new self-report measure, which was developed to be empirically valid and easy to administer. Study 1 consisted in an exploratory factor analysis to determine the most appropriate structure of the ICS in a…

  9. Inter-rater reliability for movement pattern analysis (MPA): measuring patterning of behaviors versus discrete behavior counts as indicators of decision-making style

    PubMed Central

    Connors, Brenda L.; Rende, Richard; Colton, Timothy J.

    2014-01-01

    The unique yield of collecting observational data on human movement has received increasing attention in a number of domains, including the study of decision-making style. As such, interest has grown in the nuances of core methodological issues, including the best ways of assessing inter-rater reliability. In this paper we focus on one key topic – the distinction between establishing reliability for the patterning of behaviors as opposed to the computation of raw counts – and suggest that reliability for each be compared empirically rather than determined a priori. We illustrate by assessing inter-rater reliability for key outcome measures derived from movement pattern analysis (MPA), an observational methodology that records body movements as indicators of decision-making style with demonstrated predictive validity. While reliability ranged from moderate to good for raw counts of behaviors reflecting each of two Overall Factors generated within MPA (Assertion and Perspective), inter-rater reliability for patterning (proportional indicators of each factor) was significantly higher and excellent (ICC = 0.89). Furthermore, patterning, as compared to raw counts, provided better prediction of observable decision-making process assessed in the laboratory. These analyses support the utility of using an empirical approach to inform the consideration of measuring patterning versus discrete behavioral counts of behaviors when determining inter-rater reliability of observable behavior. They also speak to the substantial reliability that may be achieved via application of theoretically grounded observational systems such as MPA that reveal thinking and action motivations via visible movement patterns. PMID:24999336

  10. Inter-rater reliability for movement pattern analysis (MPA): measuring patterning of behaviors versus discrete behavior counts as indicators of decision-making style.

    PubMed

    Connors, Brenda L; Rende, Richard; Colton, Timothy J

    2014-01-01

    The unique yield of collecting observational data on human movement has received increasing attention in a number of domains, including the study of decision-making style. As such, interest has grown in the nuances of core methodological issues, including the best ways of assessing inter-rater reliability. In this paper we focus on one key topic - the distinction between establishing reliability for the patterning of behaviors as opposed to the computation of raw counts - and suggest that reliability for each be compared empirically rather than determined a priori. We illustrate by assessing inter-rater reliability for key outcome measures derived from movement pattern analysis (MPA), an observational methodology that records body movements as indicators of decision-making style with demonstrated predictive validity. While reliability ranged from moderate to good for raw counts of behaviors reflecting each of two Overall Factors generated within MPA (Assertion and Perspective), inter-rater reliability for patterning (proportional indicators of each factor) was significantly higher and excellent (ICC = 0.89). Furthermore, patterning, as compared to raw counts, provided better prediction of observable decision-making process assessed in the laboratory. These analyses support the utility of using an empirical approach to inform the consideration of measuring patterning versus discrete behavioral counts of behaviors when determining inter-rater reliability of observable behavior. They also speak to the substantial reliability that may be achieved via application of theoretically grounded observational systems such as MPA that reveal thinking and action motivations via visible movement patterns.

  11. Assessing the Measurement Properties of the Principal Instructional Management Rating Scale: A Meta-Analysis of Reliability Studies

    ERIC Educational Resources Information Center

    Hallinger, Phillip; Wang, Wen-Chung; Chen, Chia-Wen

    2013-01-01

    Background: In a recent article, Hallinger (2011b) reviewed 135 empirical studies that had employed the Principal Instructional Management Rating Scale (PIMRS) over the prior three decades. The author concluded that the PIMRS appeared to have attained a consistent record of yielding reliable and valid data on principal instructional leadership.…

  12. Ability Self-Estimates and Self-Efficacy: Meaningfully Distinct?

    ERIC Educational Resources Information Center

    Bubany, Shawn T.; Hansen, Jo-Ida C.

    2010-01-01

    Conceptual differences between self-efficacy and ability self-estimate scores, used in vocational psychology and career counseling, were examined with confirmatory factor analysis, discriminate relations, and reliability analysis. Results suggest that empirical differences may be due to measurement error or scale content, rather than due to the…

  13. Optimal Measurement Conditions for Spatiotemporal EEG/MEG Source Analysis.

    ERIC Educational Resources Information Center

    Huizenga, Hilde M.; Heslenfeld, Dirk J.; Molenaar, Peter C. M.

    2002-01-01

    Developed a method to determine the required number and position of sensors for human brain electromagnetic source analysis. Studied the method through a simulation study and an empirical study on visual evoked potentials in one adult male. Results indicate the method is fast and reliable and improves source precision. (SLD)

  14. Validity Issues in Assessing Dispositions: The Confirmatory Factor Analysis of a Teacher Dispositions Form

    ERIC Educational Resources Information Center

    Niu, Chunling; Everson, Kimberlee; Dietrich, Sylvia; Zippay, Cassie

    2017-01-01

    Critics against the inclusion of dispositions as part of the teacher education accreditation focus on the dearth of empirical literature on reliably and validly accessing dispositions (Borko, Liston, & Whitcomb, 2007). In this study, a confirmatory factor analysis (CFA) was performed to test the factorial validity of a teacher dispositions…

  15. An empirical comparison of a dynamic software testability metric to static cyclomatic complexity

    NASA Technical Reports Server (NTRS)

    Voas, Jeffrey M.; Miller, Keith W.; Payne, Jeffrey E.

    1993-01-01

    This paper compares the dynamic testability prediction technique termed 'sensitivity analysis' to the static testability technique termed cyclomatic complexity. The application that we chose in this empirical study is a CASE generated version of a B-737 autoland system. For the B-737 system we analyzed, we isolated those functions that we predict are more prone to hide errors during system/reliability testing. We also analyzed the code with several other well-known static metrics. This paper compares and contrasts the results of sensitivity analysis to the results of the static metrics.

  16. Estimating Ordinal Reliability for Likert-Type and Ordinal Item Response Data: A Conceptual, Empirical, and Practical Guide

    ERIC Educational Resources Information Center

    Gadermann, Anne M.; Guhn, Martin; Zumbo, Bruno D.

    2012-01-01

    This paper provides a conceptual, empirical, and practical guide for estimating ordinal reliability coefficients for ordinal item response data (also referred to as Likert, Likert-type, ordered categorical, or rating scale item responses). Conventionally, reliability coefficients, such as Cronbach's alpha, are calculated using a Pearson…

  17. An empirical Bayes approach for the Poisson life distribution.

    NASA Technical Reports Server (NTRS)

    Canavos, G. C.

    1973-01-01

    A smooth empirical Bayes estimator is derived for the intensity parameter (hazard rate) in the Poisson distribution as used in life testing. The reliability function is also estimated either by using the empirical Bayes estimate of the parameter, or by obtaining the expectation of the reliability function. The behavior of the empirical Bayes procedure is studied through Monte Carlo simulation in which estimates of mean-squared errors of the empirical Bayes estimators are compared with those of conventional estimators such as minimum variance unbiased or maximum likelihood. Results indicate a significant reduction in mean-squared error of the empirical Bayes estimators over the conventional variety.

  18. Detecting intrinsic dynamics of traffic flow with recurrence analysis and empirical mode decomposition

    NASA Astrophysics Data System (ADS)

    Xiong, Hui; Shang, Pengjian; Bian, Songhan

    2017-05-01

    In this paper, we apply the empirical mode decomposition (EMD) method to the recurrence plot (RP) and recurrence quantification analysis (RQA), to evaluate the frequency- and time-evolving dynamics of the traffic flow. Based on the cumulative intrinsic mode functions extracted by the EMD, the frequency-evolving RP regarding different oscillation of modes suggests that apparent dynamics of the data considered are mainly dominated by its components of medium- and low-frequencies while severely affected by fast oscillated noises contained in the signal. Noises are then eliminated to analyze the intrinsic dynamics and consequently, the denoised time-evolving RQA diversely characterizes the properties of the signal and marks crucial points more accurately where white bands in the RP occur, whereas a strongly qualitative agreement exists between all the non-denoised RQA measures. Generally, the EMD combining with the recurrence analysis sheds more reliable, abundant and inherent lights into the traffic flow, which is meaningful to the empirical analysis of complex systems.

  19. Improved estimation of subject-level functional connectivity using full and partial correlation with empirical Bayes shrinkage.

    PubMed

    Mejia, Amanda F; Nebel, Mary Beth; Barber, Anita D; Choe, Ann S; Pekar, James J; Caffo, Brian S; Lindquist, Martin A

    2018-05-15

    Reliability of subject-level resting-state functional connectivity (FC) is determined in part by the statistical techniques employed in its estimation. Methods that pool information across subjects to inform estimation of subject-level effects (e.g., Bayesian approaches) have been shown to enhance reliability of subject-level FC. However, fully Bayesian approaches are computationally demanding, while empirical Bayesian approaches typically rely on using repeated measures to estimate the variance components in the model. Here, we avoid the need for repeated measures by proposing a novel measurement error model for FC describing the different sources of variance and error, which we use to perform empirical Bayes shrinkage of subject-level FC towards the group average. In addition, since the traditional intra-class correlation coefficient (ICC) is inappropriate for biased estimates, we propose a new reliability measure denoted the mean squared error intra-class correlation coefficient (ICC MSE ) to properly assess the reliability of the resulting (biased) estimates. We apply the proposed techniques to test-retest resting-state fMRI data on 461 subjects from the Human Connectome Project to estimate connectivity between 100 regions identified through independent components analysis (ICA). We consider both correlation and partial correlation as the measure of FC and assess the benefit of shrinkage for each measure, as well as the effects of scan duration. We find that shrinkage estimates of subject-level FC exhibit substantially greater reliability than traditional estimates across various scan durations, even for the most reliable connections and regardless of connectivity measure. Additionally, we find partial correlation reliability to be highly sensitive to the choice of penalty term, and to be generally worse than that of full correlations except for certain connections and a narrow range of penalty values. This suggests that the penalty needs to be chosen carefully when using partial correlations. Copyright © 2018. Published by Elsevier Inc.

  20. Empirical test of the performance of an acoustic-phonetic approach to forensic voice comparison under conditions similar to those of a real case.

    PubMed

    Enzinger, Ewald; Morrison, Geoffrey Stewart

    2017-08-01

    In a 2012 case in New South Wales, Australia, the identity of a speaker on several audio recordings was in question. Forensic voice comparison testimony was presented based on an auditory-acoustic-phonetic-spectrographic analysis. No empirical demonstration of the validity and reliability of the analytical methodology was presented. Unlike the admissibility standards in some other jurisdictions (e.g., US Federal Rule of Evidence 702 and the Daubert criteria, or England & Wales Criminal Practice Directions 19A), Australia's Unified Evidence Acts do not require demonstration of the validity and reliability of analytical methods and their implementation before testimony based upon them is presented in court. The present paper reports on empirical tests of the performance of an acoustic-phonetic-statistical forensic voice comparison system which exploited the same features as were the focus of the auditory-acoustic-phonetic-spectrographic analysis in the case, i.e., second-formant (F2) trajectories in /o/ tokens and mean fundamental frequency (f0). The tests were conducted under conditions similar to those in the case. The performance of the acoustic-phonetic-statistical system was very poor compared to that of an automatic system. Copyright © 2017 Elsevier B.V. All rights reserved.

  1. Rollover risk prediction of heavy vehicles by reliability index and empirical modelling

    NASA Astrophysics Data System (ADS)

    Sellami, Yamine; Imine, Hocine; Boubezoul, Abderrahmane; Cadiou, Jean-Charles

    2018-03-01

    This paper focuses on a combination of a reliability-based approach and an empirical modelling approach for rollover risk assessment of heavy vehicles. A reliability-based warning system is developed to alert the driver to a potential rollover before entering into a bend. The idea behind the proposed methodology is to estimate the rollover risk by the probability that the vehicle load transfer ratio (LTR) exceeds a critical threshold. Accordingly, a so-called reliability index may be used as a measure to assess the vehicle safe functioning. In the reliability method, computing the maximum of LTR requires to predict the vehicle dynamics over the bend which can be in some cases an intractable problem or time-consuming. With the aim of improving the reliability computation time, an empirical model is developed to substitute the vehicle dynamics and rollover models. This is done by using the SVM (Support Vector Machines) algorithm. The preliminary obtained results demonstrate the effectiveness of the proposed approach.

  2. The Conceptualisation and Measurement of DSM-5 Internet Gaming Disorder: The Development of the IGD-20 Test

    PubMed Central

    Pontes, Halley M.; Király, Orsolya; Demetrovics, Zsolt; Griffiths, Mark D.

    2014-01-01

    Background Over the last decade, there has been growing concern about ‘gaming addiction’ and its widely documented detrimental impacts on a minority of individuals that play excessively. The latest (fifth) edition of the American Psychiatric Association's Diagnostic and Statistical Manual of Mental Disorders (DSM-5) included nine criteria for the potential diagnosis of Internet Gaming Disorder (IGD) and noted that it was a condition that warranted further empirical study. Aim: The main aim of this study was to develop a valid and reliable standardised psychometrically robust tool in addition to providing empirically supported cut-off points. Methods A sample of 1003 gamers (85.2% males; mean age 26 years) from 57 different countries were recruited via online gaming forums. Validity was assessed by confirmatory factor analysis (CFA), criterion-related validity, and concurrent validity. Latent profile analysis was also carried to distinguish disordered gamers from non-disordered gamers. Sensitivity and specificity analyses were performed to determine an empirical cut-off for the test. Results The CFA confirmed the viability of IGD-20 Test with a six-factor structure (salience, mood modification, tolerance, withdrawal, conflict and relapse) for the assessment of IGD according to the nine criteria from DSM-5. The IGD-20 Test proved to be valid and reliable. According to the latent profile analysis, 5.3% of the total participants were classed as disordered gamers. Additionally, an optimal empirical cut-off of 71 points (out of 100) seemed to be adequate according to the sensitivity and specificity analyses carried. Conclusions The present findings support the viability of the IGD-20 Test as an adequate standardised psychometrically robust tool for assessing internet gaming disorder. Consequently, the new instrument represents the first step towards unification and consensus in the field of gaming studies. PMID:25313515

  3. The conceptualisation and measurement of DSM-5 Internet Gaming Disorder: the development of the IGD-20 Test.

    PubMed

    Pontes, Halley M; Király, Orsolya; Demetrovics, Zsolt; Griffiths, Mark D

    2014-01-01

    Over the last decade, there has been growing concern about 'gaming addiction' and its widely documented detrimental impacts on a minority of individuals that play excessively. The latest (fifth) edition of the American Psychiatric Association's Diagnostic and Statistical Manual of Mental Disorders (DSM-5) included nine criteria for the potential diagnosis of Internet Gaming Disorder (IGD) and noted that it was a condition that warranted further empirical study. The main aim of this study was to develop a valid and reliable standardised psychometrically robust tool in addition to providing empirically supported cut-off points. A sample of 1003 gamers (85.2% males; mean age 26 years) from 57 different countries were recruited via online gaming forums. Validity was assessed by confirmatory factor analysis (CFA), criterion-related validity, and concurrent validity. Latent profile analysis was also carried to distinguish disordered gamers from non-disordered gamers. Sensitivity and specificity analyses were performed to determine an empirical cut-off for the test. The CFA confirmed the viability of IGD-20 Test with a six-factor structure (salience, mood modification, tolerance, withdrawal, conflict and relapse) for the assessment of IGD according to the nine criteria from DSM-5. The IGD-20 Test proved to be valid and reliable. According to the latent profile analysis, 5.3% of the total participants were classed as disordered gamers. Additionally, an optimal empirical cut-off of 71 points (out of 100) seemed to be adequate according to the sensitivity and specificity analyses carried. The present findings support the viability of the IGD-20 Test as an adequate standardised psychometrically robust tool for assessing internet gaming disorder. Consequently, the new instrument represents the first step towards unification and consensus in the field of gaming studies.

  4. Crowd-Sourcing Relative Preferences for Ecosystem Services in The St. Louis River AOC (Poster)

    EPA Science Inventory

    Analysis of ecosystem service tradeoffs among AOC project scenarios is more reliable if valuation data are available. As an alternative to empirical data, we classified images from social media sites (SMS) Panoramio (n= 638), Instagram (n=2085), and Flickr (n=6643) for the St. Lo...

  5. Body surface assessment with 3D laser-based anthropometry: reliability, validation, and improvement of empirical surface formulae.

    PubMed

    Kuehnapfel, Andreas; Ahnert, Peter; Loeffler, Markus; Scholz, Markus

    2017-02-01

    Body surface area is a physiological quantity relevant for many medical applications. In clinical practice, it is determined by empirical formulae. 3D laser-based anthropometry provides an easy and effective way to measure body surface area but is not ubiquitously available. We used data from laser-based anthropometry from a population-based study to assess validity of published and commonly used empirical formulae. We performed a large population-based study on adults collecting classical anthropometric measurements and 3D body surface assessments (N = 1435). We determined reliability of the 3D body surface assessment and validity of 18 different empirical formulae proposed in the literature. The performance of these formulae is studied in subsets of sex and BMI. Finally, improvements of parameter settings of formulae and adjustments for sex and BMI were considered. 3D body surface measurements show excellent intra- and inter-rater reliability of 0.998 (overall concordance correlation coefficient, OCCC was used as measure of agreement). Empirical formulae of Fujimoto and Watanabe, Shuter and Aslani and Sendroy and Cecchini performed best with excellent concordance with OCCC > 0.949 even in subgroups of sex and BMI. Re-parametrization of formulae and adjustment for sex and BMI slightly improved results. In adults, 3D laser-based body surface assessment is a reliable alternative to estimation by empirical formulae. However, there are empirical formulae showing excellent results even in subgroups of sex and BMI with only little room for improvement.

  6. An Empirical Evaluation of Factor Reliability.

    ERIC Educational Resources Information Center

    Jackson, Douglas N.; Morf, Martin E.

    The psychometric reliability of a factor, defined as its generalizability across samples drawn from the same population of tests, is considered as a necessary precondition for the scientific meaningfulness of factor analytic results. A solution to the problem of generalizability is illustrated empirically on data from a set of tests designed to…

  7. Improving reliability of aggregation, numerical simulation and analysis of complex systems by empirical data

    NASA Astrophysics Data System (ADS)

    Dobronets, Boris S.; Popova, Olga A.

    2018-05-01

    The paper considers a new approach of regression modeling that uses aggregated data presented in the form of density functions. Approaches to Improving the reliability of aggregation of empirical data are considered: improving accuracy and estimating errors. We discuss the procedures of data aggregation as a preprocessing stage for subsequent to regression modeling. An important feature of study is demonstration of the way how represent the aggregated data. It is proposed to use piecewise polynomial models, including spline aggregate functions. We show that the proposed approach to data aggregation can be interpreted as the frequency distribution. To study its properties density function concept is used. Various types of mathematical models of data aggregation are discussed. For the construction of regression models, it is proposed to use data representation procedures based on piecewise polynomial models. New approaches to modeling functional dependencies based on spline aggregations are proposed.

  8. An Empirical Analysis of Teacher Spillover Effects in Secondary School

    ERIC Educational Resources Information Center

    Koedel, Cory

    2009-01-01

    This paper examines whether educational production in secondary school involves joint production among teachers across subjects. In doing so, it also provides insights into the reliability of value-added modeling. Teacher value-added to reading test scores is estimated for four different teacher types: English, math, science and social-studies.…

  9. Measuring Work Environment and Performance in Nursing Homes

    PubMed Central

    Temkin-Greener, Helena; Zheng, Nan (Tracy); Katz, Paul; Zhao, Hongwei; Mukamel, Dana B.

    2008-01-01

    Background Qualitative studies of the nursing home work environment have long suggested that such attributes as leadership and communication may be related to nursing home performance, including residents' outcomes. However, empirical studies examining these relationships have been scant. Objectives This study is designed to: develop an instrument for measuring nursing home work environment and perceived work effectiveness; test the reliability and validity of the instrument; and identify individual and facility-level factors associated with better facility performance. Research Design and Methods The analysis was based on survey responses provided by managers (N=308) and direct care workers (N=7,418) employed in 162 facilities throughout New York State. Exploratory factor analysis, Chronbach's alphas, analysis of variance, and regression models were used to assess instrument reliability and validity. Multivariate regression models, with fixed facility effects, were used to examine factors associated with work effectiveness. Results The reliability and the validity of the survey instrument for measuring work environment and perceived work effectiveness has been demonstrated. Several individual (e.g. occupation, race) and facility characteristics (e.g. management style, workplace conditions, staffing) that are significant predictors of perceived work effectiveness were identified. Conclusions The organizational performance model used in this study recognizes the multidimensionality of the work environment in nursing homes. Our findings suggest that efforts at improving work effectiveness must also be multifaceted. Empirical findings from such a line of research may provide insights for improving the quality of the work environment and ultimately the quality of residents' care. PMID:19330892

  10. A Bayesian approach to parameter and reliability estimation in the Poisson distribution.

    NASA Technical Reports Server (NTRS)

    Canavos, G. C.

    1972-01-01

    For life testing procedures, a Bayesian analysis is developed with respect to a random intensity parameter in the Poisson distribution. Bayes estimators are derived for the Poisson parameter and the reliability function based on uniform and gamma prior distributions of that parameter. A Monte Carlo procedure is implemented to make possible an empirical mean-squared error comparison between Bayes and existing minimum variance unbiased, as well as maximum likelihood, estimators. As expected, the Bayes estimators have mean-squared errors that are appreciably smaller than those of the other two.

  11. Design and experimentation of an empirical multistructure framework for accurate, sharp and reliable hydrological ensembles

    NASA Astrophysics Data System (ADS)

    Seiller, G.; Anctil, F.; Roy, R.

    2017-09-01

    This paper outlines the design and experimentation of an Empirical Multistructure Framework (EMF) for lumped conceptual hydrological modeling. This concept is inspired from modular frameworks, empirical model development, and multimodel applications, and encompasses the overproduce and select paradigm. The EMF concept aims to reduce subjectivity in conceptual hydrological modeling practice and includes model selection in the optimisation steps, reducing initial assumptions on the prior perception of the dominant rainfall-runoff transformation processes. EMF generates thousands of new modeling options from, for now, twelve parent models that share their functional components and parameters. Optimisation resorts to ensemble calibration, ranking and selection of individual child time series based on optimal bias and reliability trade-offs, as well as accuracy and sharpness improvement of the ensemble. Results on 37 snow-dominated Canadian catchments and 20 climatically-diversified American catchments reveal the excellent potential of the EMF in generating new individual model alternatives, with high respective performance values, that may be pooled efficiently into ensembles of seven to sixty constitutive members, with low bias and high accuracy, sharpness, and reliability. A group of 1446 new models is highlighted to offer good potential on other catchments or applications, based on their individual and collective interests. An analysis of the preferred functional components reveals the importance of the production and total flow elements. Overall, results from this research confirm the added value of ensemble and flexible approaches for hydrological applications, especially in uncertain contexts, and open up new modeling possibilities.

  12. Rasch Analysis of a New Hierarchical Scoring System for Evaluating Hand Function on the Motor Assessment Scale for Stroke

    PubMed Central

    Sabari, Joyce S.; Woodbury, Michelle; Velozo, Craig A.

    2014-01-01

    Objectives. (1) To develop two independent measurement scales for use as items assessing hand movements and hand activities within the Motor Assessment Scale (MAS), an existing instrument used for clinical assessment of motor performance in stroke survivors; (2) To examine the psychometric properties of these new measurement scales. Design. Scale development, followed by a multicenter observational study. Setting. Inpatient and outpatient occupational therapy programs in eight hospital and rehabilitation facilities in the United States and Canada. Participants. Patients (N = 332) receiving stroke rehabilitation following left (52%) or right (48%) cerebrovascular accident; mean age 64.2 years (sd 15); median 1 month since stroke onset. Intervention. Not applicable. Main Outcome Measures. Data were tested for unidimensionality and reliability, and behavioral criteria were ordered according to difficulty level with Rasch analysis. Results. The new scales assessing hand movements and hand activities met Rasch expectations of unidimensionality and reliability. Conclusion. Following a multistep process of test development, analysis, and refinement, we have redesigned the two scales that comprise the hand function items on the MAS. The hand movement scale contains an empirically validated 10-behavior hierarchy and the hand activities item contains an empirically validated 8-behavior hierarchy. PMID:25177513

  13. Reliability Generalization of the Alcohol Use Disorder Identification Test.

    ERIC Educational Resources Information Center

    Shields, Alan L.; Caruso, John C.

    2002-01-01

    Evaluated the reliability of scores from the Alcohol Use Disorders Identification Test (AUDIT; J. Sounders and others, 1993) in a reliability generalization study based on 17 empirical journal articles. Results show AUDIT scores to be generally reliable for basic assessment. (SLD)

  14. An empirical comparative study on biological age estimation algorithms with an application of Work Ability Index (WAI).

    PubMed

    Cho, Il Haeng; Park, Kyung S; Lim, Chang Joo

    2010-02-01

    In this study, we described the characteristics of five different biological age (BA) estimation algorithms, including (i) multiple linear regression, (ii) principal component analysis, and somewhat unique methods developed by (iii) Hochschild, (iv) Klemera and Doubal, and (v) a variant of Klemera and Doubal's method. The objective of this study is to find the most appropriate method of BA estimation by examining the association between Work Ability Index (WAI) and the differences of each algorithm's estimates from chronological age (CA). The WAI was found to be a measure that reflects an individual's current health status rather than the deterioration caused by a serious dependency with the age. Experiments were conducted on 200 Korean male participants using a BA estimation system developed principally under the concept of non-invasive, simple to operate and human function-based. Using the empirical data, BA estimation as well as various analyses including correlation analysis and discriminant function analysis was performed. As a result, it had been confirmed by the empirical data that Klemera and Doubal's method with uncorrelated variables from principal component analysis produces relatively reliable and acceptable BA estimates. 2009 Elsevier Ireland Ltd. All rights reserved.

  15. Proposed Core Competencies and Empirical Validation Procedure in Competency Modeling: Confirmation and Classification.

    PubMed

    Baczyńska, Anna K; Rowiński, Tomasz; Cybis, Natalia

    2016-01-01

    Competency models provide insight into key skills which are common to many positions in an organization. Moreover, there is a range of competencies that is used by many companies. Researchers have developed core competency terminology to underline their cross-organizational value. The article presents a theoretical model of core competencies consisting of two main higher-order competencies called performance and entrepreneurship. Each of them consists of three elements: the performance competency includes cooperation, organization of work and goal orientation, while entrepreneurship includes innovativeness, calculated risk-taking and pro-activeness. However, there is lack of empirical validation of competency concepts in organizations and this would seem crucial for obtaining reliable results from organizational research. We propose a two-step empirical validation procedure: (1) confirmation factor analysis, and (2) classification of employees. The sample consisted of 636 respondents (M = 44.5; SD = 15.1). Participants were administered a questionnaire developed for the study purpose. The reliability, measured by Cronbach's alpha, ranged from 0.60 to 0.83 for six scales. Next, we tested the model using a confirmatory factor analysis. The two separate, single models of performance and entrepreneurial orientations fit quite well to the data, while a complex model based on the two single concepts needs further research. In the classification of employees based on the two higher order competencies we obtained four main groups of employees. Their profiles relate to those found in the literature, including so-called niche finders and top performers. Some proposal for organizations is discussed.

  16. Estimation of Reliability Coefficients Using the Test Information Function and Its Modifications.

    ERIC Educational Resources Information Center

    Samejima, Fumiko

    1994-01-01

    The reliability coefficient is predicted from the test information function (TIF) or two modified TIF formulas and a specific trait distribution. Examples illustrate the variability of the reliability coefficient across different trait distributions, and results are compared with empirical reliability coefficients. (SLD)

  17. Chronic Fatigue Syndrome and Myalgic Encephalomyelitis: Toward An Empirical Case Definition

    PubMed Central

    Jason, Leonard A.; Kot, Bobby; Sunnquist, Madison; Brown, Abigail; Evans, Meredyth; Jantke, Rachel; Williams, Yolonda; Furst, Jacob; Vernon, Suzanne D.

    2015-01-01

    Current case definitions of Myalgic Encephalomyelitis (ME) and chronic fatigue syndrome (CFS) have been based on consensus methods, but empirical methods could be used to identify core symptoms and thereby improve the reliability. In the present study, several methods (i.e., continuous scores of symptoms, theoretically and empirically derived cut off scores of symptoms) were used to identify core symptoms best differentiating patients from controls. In addition, data mining with decision trees was conducted. Our study found a small number of core symptoms that have good sensitivity and specificity, and these included fatigue, post-exertional malaise, a neurocognitive symptom, and unrefreshing sleep. Outcomes from these analyses suggest that using empirically selected symptoms can help guide the creation of a more reliable case definition. PMID:26029488

  18. Development and psychometric evaluation of the Undergraduate Clinical Education Environment Measure (UCEEM).

    PubMed

    Strand, Pia; Sjöborg, Karolina; Stalmeijer, Renée; Wichmann-Hansen, Gitte; Jakobsson, Ulf; Edgren, Gudrun

    2013-12-01

    There is a paucity of instruments designed to evaluate the multiple dimensions of the workplace as an educational environment for undergraduate medical students. The aim was to develop and psychometrically evaluate an instrument to measure how undergraduate medical students perceive the clinical workplace environment, based on workplace learning theories and empirical findings. Development of the instrument relied on established standards including theoretical and empirical grounding, systematic item development and expert review at various stages to ensure content validity. Qualitative and quantitative methods were employed using a series of steps from conceptualization through psychometric analysis of scores in a Swedish medical student population. The final result was a 25-item instrument with two overarching dimensions, experiential learning and social participation, and four subscales that coincided well with theory and empirical findings: Opportunities to learn in and through work & quality of supervision; Preparedness for student entry; Workplace interaction patterns & student inclusion; and Equal treatment. Evidence from various sources supported content validity, construct validity and reliability of the instrument. The Undergraduate Clinical Education Environment Measure represents a valid, reliable and feasible multidimensional instrument for evaluation of the clinical workplace as a learning environment for undergraduate medical students. Further validation in different populations using various psychometric methods is needed.

  19. A Simple and Reliable Method of Design for Standalone Photovoltaic Systems

    NASA Astrophysics Data System (ADS)

    Srinivasarao, Mantri; Sudha, K. Rama; Bhanu, C. V. K.

    2017-06-01

    Standalone photovoltaic (SAPV) systems are seen as a promoting method of electrifying areas of developing world that lack power grid infrastructure. Proliferations of these systems require a design procedure that is simple, reliable and exhibit good performance over its life time. The proposed methodology uses simple empirical formulae and easily available parameters to design SAPV systems, that is, array size with energy storage. After arriving at the different array size (area), performance curves are obtained for optimal design of SAPV system with high amount of reliability in terms of autonomy at a specified value of loss of load probability (LOLP). Based on the array to load ratio (ALR) and levelized energy cost (LEC) through life cycle cost (LCC) analysis, it is shown that the proposed methodology gives better performance, requires simple data and is more reliable when compared with conventional design using monthly average daily load and insolation.

  20. APPLICATION OF TRAVEL TIME RELIABILITY FOR PERFORMANCE ORIENTED OPERATIONAL PLANNING OF EXPRESSWAYS

    NASA Astrophysics Data System (ADS)

    Mehran, Babak; Nakamura, Hideki

    Evaluation of impacts of congestion improvement scheme s on travel time reliability is very significant for road authorities since travel time reliability repr esents operational performance of expressway segments. In this paper, a methodology is presented to estimate travel tim e reliability prior to implementation of congestion relief schemes based on travel time variation modeling as a function of demand, capacity, weather conditions and road accident s. For subject expressway segmen ts, traffic conditions are modeled over a whole year considering demand and capacity as random variables. Patterns of demand and capacity are generated for each five minute interval by appl ying Monte-Carlo simulation technique, and accidents are randomly generated based on a model that links acci dent rate to traffic conditions. A whole year analysis is performed by comparing de mand and available capacity for each scenario and queue length is estimated through shockwave analysis for each time in terval. Travel times are estimated from refined speed-flow relationships developed for intercity expressways and buffer time index is estimated consequently as a measure of travel time reliability. For validation, estimated reliability indices are compared with measured values from empirical data, and it is shown that the proposed method is suitable for operational evaluation and planning purposes.

  1. Quantitative Analysis of the Rubric as an Assessment Tool: An Empirical Study of Student Peer-Group Rating

    ERIC Educational Resources Information Center

    Hafner, John C.; Hafner, Patti M.

    2003-01-01

    Although the rubric has emerged as one of the most popular assessment tools in progressive educational programs, there is an unfortunate dearth of information in the literature quantifying the actual effectiveness of the rubric as an assessment tool "in the hands of the students." This study focuses on the validity and reliability of the rubric as…

  2. Validity and reliability of portfolio assessment of competency in a baccalaureate dental hygiene program

    NASA Astrophysics Data System (ADS)

    Gadbury-Amyot, Cynthia C.

    This study examined validity and reliability of portfolio assessment using Messick's (1996, 1995) unified framework of construct validity. Theoretical and empirical evidence was sought for six aspects of construct validity. The sample included twenty student portfolios. Each portfolio were evaluated by seven faculty raters using a primary trait analysis scoring rubric. There was a significant relationship (r = .81--.95; p < .01) between the seven subscales in the scoring rubric demonstrating measurement of a common construct. Item analysis was conducted to examine convergent and discriminant empirical relationships of the 35 items in the scoring rubric. There was a significant relationship between all items ( p < .01), and all but one item was more strongly correlated with its own subscale than with other subscales. However, correlations of items across subscales were predominantly moderate in strength indicating that items did not strongly discriminate between subscales. A fully crossed, two facet generalizability (G) study design was used to examine reliability. Analysis of variance demonstrated that the greatest source of variance was the scoring rubric itself, accounting for 78% of the total variance. The smallest source of variance was the interaction between portfolio and rubric (1.15%) indicating that while the seven subscales varied in difficulty level, the relative standing of individual portfolios was maintained across subscales. Faculty rater variance accounted for only 1.28% of total variance. A phi coefficient of .86, analogous to a reliability coefficient in classical test theory, was obtained in the Decision study by increasing the subscales to fourteen and decreasing faculty raters to three. There was a significant relationship between portfolios and grade point average (r = .70; p < .01), and the National Dental Hygiene Board Examination (r = .60; p < .01). The relationship between portfolios and the Central Regional Dental Testing Service examination was both weak and nonsignificant (r = .19; p > .05). An open-ended survey was used to elicit student feedback on portfolio development. A majority of the students (76%) perceived value in the development of programmatic portfolios. In conclusion, the pattern of findings from this study suggest that portfolios can serve as a valid and reliable measure for assessing student competency.

  3. An Empirical Research on the Correlation between Human Capital and Career Success of Knowledge Workers in Enterprise

    NASA Astrophysics Data System (ADS)

    Guo, Wenchen; Xiao, Hongjun; Yang, Xi

    Human capital plays an important part in employability of knowledge workers, also it is the important intangible assets of company. This paper explores the correlation between human capital and career success of knowledge workers. Based on literature retrieval, we identified measuring tool of career success and modified further; measuring human capital with self-developed scale of high reliability and validity. After exploratory factor analysis, we suggest that human capital contents four dimensions, including education, work experience, learning ability and training; career success contents three dimensions, including perceived internal competitiveness of organization, perceived external competitiveness of organization and career satisfaction. The result of empirical analysis indicates that there is a positive correlation between human capital and career success, and human capital is an excellent predictor of career success beyond demographics variables.

  4. Ethical Implications of Validity-vs.-Reliability Trade-Offs in Educational Research

    ERIC Educational Resources Information Center

    Fendler, Lynn

    2016-01-01

    In educational research that calls itself empirical, the relationship between validity and reliability is that of trade-off: the stronger the bases for validity, the weaker the bases for reliability (and vice versa). Validity and reliability are widely regarded as basic criteria for evaluating research; however, there are ethical implications of…

  5. Proposed Core Competencies and Empirical Validation Procedure in Competency Modeling: Confirmation and Classification

    PubMed Central

    Baczyńska, Anna K.; Rowiński, Tomasz; Cybis, Natalia

    2016-01-01

    Competency models provide insight into key skills which are common to many positions in an organization. Moreover, there is a range of competencies that is used by many companies. Researchers have developed core competency terminology to underline their cross-organizational value. The article presents a theoretical model of core competencies consisting of two main higher-order competencies called performance and entrepreneurship. Each of them consists of three elements: the performance competency includes cooperation, organization of work and goal orientation, while entrepreneurship includes innovativeness, calculated risk-taking and pro-activeness. However, there is lack of empirical validation of competency concepts in organizations and this would seem crucial for obtaining reliable results from organizational research. We propose a two-step empirical validation procedure: (1) confirmation factor analysis, and (2) classification of employees. The sample consisted of 636 respondents (M = 44.5; SD = 15.1). Participants were administered a questionnaire developed for the study purpose. The reliability, measured by Cronbach’s alpha, ranged from 0.60 to 0.83 for six scales. Next, we tested the model using a confirmatory factor analysis. The two separate, single models of performance and entrepreneurial orientations fit quite well to the data, while a complex model based on the two single concepts needs further research. In the classification of employees based on the two higher order competencies we obtained four main groups of employees. Their profiles relate to those found in the literature, including so-called niche finders and top performers. Some proposal for organizations is discussed. PMID:27014111

  6. Psychometric Evaluation of a Coping Strategies Inventory Short-Form (CSI-SF) in the Jackson Heart Study Cohort

    PubMed Central

    Addison, Clifton C.; Campbell-Jenkins, Brenda W.; Sarpong, Daniel F.; Kibler, Jeffery; Singh, Madhu; Dubbert, Patricia; Wilson, Gregory; Payne, Thomas; Taylor, Herman

    2007-01-01

    This study sought to establish the psychometric properties of a Coping Strategies Inventory Short Form (CSI-SF) by examining coping skills in the Jackson Heart Study cohort. We used exploratory and confirmatory factor analysis, Pearson’s correlation, and Cronbach Alpha to examine reliability and validity in the CSI-SF that solicited responses from 5302 African American men and women between the ages of 35 and 84. One item was dropped from the 16-item CSI-SF, making it a 15-item survey. No significant effects were found for age and gender, strengthening the generalizability of the CSI-SF. The internal consistency reliability analysis revealed reliability between alpha = 0.58–0.72 for all of the scales, and all of the fit indices used to examine the CSI-SF provided support for its use as an adequate measure of coping. This study provides empirical support for utilizing this instrument in future efforts to understand the role of coping in moderating health outcomes. PMID:18180539

  7. The Reliability and Stability of an Inferred Phylogenetic Tree from Empirical Data.

    PubMed

    Katsura, Yukako; Stanley, Craig E; Kumar, Sudhir; Nei, Masatoshi

    2017-03-01

    The reliability of a phylogenetic tree obtained from empirical data is usually measured by the bootstrap probability (Pb) of interior branches of the tree. If the bootstrap probability is high for most branches, the tree is considered to be reliable. If some interior branches show relatively low bootstrap probabilities, we are not sure that the inferred tree is really reliable. Here, we propose another quantity measuring the reliability of the tree called the stability of a subtree. This quantity refers to the probability of obtaining a subtree (Ps) of an inferred tree obtained. We then show that if the tree is to be reliable, both Pb and Ps must be high. We also show that Ps is given by a bootstrap probability of the subtree with the closest outgroup sequence, and computer program RESTA for computing the Pb and Ps values will be presented. © The Author 2017. Published by Oxford University Press on behalf of the Society for Molecular Biology and Evolution.

  8. Statistical prediction of space motion sickness

    NASA Technical Reports Server (NTRS)

    Reschke, Millard F.

    1990-01-01

    Studies designed to empirically examine the etiology of motion sickness to develop a foundation for enhancing its prediction are discussed. Topics addressed include early attempts to predict space motion sickness, multiple test data base that uses provocative and vestibular function tests, and data base subjects; reliability of provocative tests of motion sickness susceptibility; prediction of space motion sickness using linear discriminate analysis; and prediction of space motion sickness susceptibility using the logistic model.

  9. A study of multiplex data bus techniques for the space shuttle

    NASA Technical Reports Server (NTRS)

    Kearney, R. J.; Kalange, M. A.

    1972-01-01

    A comprehensive technology base for the design of a multiplexed data bus subsystem is provided. Extensive analyses, both analytical and empirical, were performed. Subjects covered are classified under the following headings: requirements identification and analysis; transmission media studies; signal design and detection studies; synchronization, timing, and control studies; user-subsystem interface studies; operational reliability analyses; design of candidate data bus configurations; and evaluation of candidate data bus designs.

  10. Multimodal Pressure-Flow Analysis: Application of Hilbert Huang Transform in Cerebral Blood Flow Regulation

    NASA Astrophysics Data System (ADS)

    Lo, Men-Tzung; Hu, Kun; Liu, Yanhui; Peng, C.-K.; Novak, Vera

    2008-12-01

    Quantification of nonlinear interactions between two nonstationary signals presents a computational challenge in different research fields, especially for assessments of physiological systems. Traditional approaches that are based on theories of stationary signals cannot resolve nonstationarity-related issues and, thus, cannot reliably assess nonlinear interactions in physiological systems. In this review we discuss a new technique called multimodal pressure flow (MMPF) method that utilizes Hilbert-Huang transformation to quantify interaction between nonstationary cerebral blood flow velocity (BFV) and blood pressure (BP) for the assessment of dynamic cerebral autoregulation (CA). CA is an important mechanism responsible for controlling cerebral blood flow in responses to fluctuations in systemic BP within a few heart-beats. The MMPF analysis decomposes BP and BFV signals into multiple empirical modes adaptively so that the fluctuations caused by a specific physiologic process can be represented in a corresponding empirical mode. Using this technique, we showed that dynamic CA can be characterized by specific phase delays between the decomposed BP and BFV oscillations, and that the phase shifts are significantly reduced in hypertensive, diabetics and stroke subjects with impaired CA. Additionally, the new technique can reliably assess CA using both induced BP/BFV oscillations during clinical tests and spontaneous BP/BFV fluctuations during resting conditions.

  11. Can Reliability of Multiple Component Measuring Instruments Depend on Response Option Presentation Mode?

    ERIC Educational Resources Information Center

    Menold, Natalja; Raykov, Tenko

    2016-01-01

    This article examines the possible dependency of composite reliability on presentation format of the elements of a multi-item measuring instrument. Using empirical data and a recent method for interval estimation of group differences in reliability, we demonstrate that the reliability of an instrument need not be the same when polarity of the…

  12. Interrater Reliability in Large-Scale Assessments--Can Teachers Score National Tests Reliably without External Controls?

    ERIC Educational Resources Information Center

    Pantzare, Anna Lind

    2015-01-01

    In most large-scale assessment systems a set of rather expensive external quality controls are implemented in order to guarantee the quality of interrater reliability. This study empirically examines if teachers' ratings of national tests in mathematics can be reliable without using monitoring, training, or other methods of external quality…

  13. Empirical Recommendations for Improving the Stability of the Dot-Probe Task in Clinical Research

    PubMed Central

    Price, Rebecca B.; Kuckertz, Jennie M.; Siegle, Greg J.; Ladouceur, Cecile D.; Silk, Jennifer S.; Ryan, Neal D.; Dahl, Ronald E.; Amir, Nader

    2014-01-01

    The dot-probe task has been widely used in research to produce an index of biased attention based on reaction times (RTs). Despite its popularity, very few published studies have examined psychometric properties of the task, including test-retest reliability, and no previous study has examined reliability in clinically anxious samples or systematically explored the effects of task design and analysis decisions on reliability. In the current analysis, we utilized dot-probe data from three studies where attention bias towards threat-related faces was assessed at multiple (≥5) timepoints. Two of the studies were similar (adults with Social Anxiety Disorder, similar design features) while one was much more disparate (pediatric healthy volunteers, distinct task design). We explored the effects of analysis choices (e.g., bias score calculation formula, methods for outlier handling) on reliability and searched for convergence of findings across the three studies. We found that, when considering the three studies concurrently, the most reliable RT bias index utilized data from dot-bottom trials, comparing congruent to incongruent trials, with rescaled outliers, particularly after averaging across more than one assessment point. Although reliability of RT bias indices was moderate to low under most circumstances, within-session variability in bias (attention bias variability; ABV), a recently proposed RT index, was more reliable across sessions. Several eyetracking-based indices of attention bias (available in the pediatric healthy sample only) showed reliability that matched the optimal RT index (ABV). On the basis of these findings, we make specific recommendations to researchers using the dot probe, particularly those wishing to investigate individual differences and/or single-patient applications. PMID:25419646

  14. Score Reliability of Adolescent Alcohol Screening Measures: A Meta-Analytic Inquiry

    ERIC Educational Resources Information Center

    Shields, Alan L.; Campfield, Delia C.; Miller, Christopher S.; Howell, Ryan T.; Wallace, Kimberly; Weiss, Roger D.

    2008-01-01

    This study describes the reliability reporting practices in empirical studies using eight adolescent alcohol screening tools and characterizes and explores variability in internal consistency estimates across samples. Of 119 observed administrations of these instruments, 40 (34%) reported usable reliability information. The Personal Experience…

  15. How to be rational about empirical success in ongoing science: The case of the quantum nose and its critics.

    PubMed

    Barwich, Ann-Sophie

    2018-06-01

    Empirical success is a central criterion for scientific decision-making. Yet its understanding in philosophical studies of science deserves renewed attention: Should philosophers think differently about the advancement of science when they deal with the uncertainty of outcome in ongoing research in comparison with historical episodes? This paper argues that normative appeals to empirical success in the evaluation of competing scientific explanations can result in unreliable conclusions, especially when we are looking at the changeability of direction in unsettled investigations. The challenges we encounter arise from the inherent dynamics of disciplinary and experimental objectives in research practice. In this paper we discuss how these dynamics inform the evaluation of empirical success by analyzing three of its requirements: data accommodation, instrumental reliability, and predictive power. We conclude that the assessment of empirical success in developing inquiry is set against the background of a model's interactive success and prospective value in an experimental context. Our argument is exemplified by the analysis of an apparent controversy surrounding the model of a quantum nose in research on olfaction. Notably, the public narrative of this controversy rests on a distorted perspective on measures of empirical success. Copyright © 2018 The Author. Published by Elsevier Ltd.. All rights reserved.

  16. Turbulence study in the vicinity of piano key weir: relevance, instrumentation, parameters and methods

    NASA Astrophysics Data System (ADS)

    Tiwari, Harinarayan; Sharma, Nayan

    2017-05-01

    This research paper focuses on the need of turbulence, instruments reliable to capture turbulence, different turbulence parameters and some advance methodology which can decompose various turbulence structures at different levels near hydraulic structures. Small-scale turbulence research has valid prospects in open channel flow. The relevance of the study is amplified as we introduce any hydraulic structure in the channel which disturbs the natural flow and creates discontinuity. To recover this discontinuity, the piano key weir (PKW) might be used with sloped keys. Constraints of empirical results in the vicinity of PKW necessitate extensive laboratory experiments with fair and reliable instrumentation techniques. Acoustic Doppler velocimeter was established to be best suited within range of some limitations using principal component analysis. Wavelet analysis is proposed to decompose the underlying turbulence structure in a better way.

  17. Coefficient Alpha: A Reliability Coefficient for the 21st Century?

    ERIC Educational Resources Information Center

    Yang, Yanyun; Green, Samuel B.

    2011-01-01

    Coefficient alpha is almost universally applied to assess reliability of scales in psychology. We argue that researchers should consider alternatives to coefficient alpha. Our preference is for structural equation modeling (SEM) estimates of reliability because they are informative and allow for an empirical evaluation of the assumptions…

  18. Fault identification of rotor-bearing system based on ensemble empirical mode decomposition and self-zero space projection analysis

    NASA Astrophysics Data System (ADS)

    Jiang, Fan; Zhu, Zhencai; Li, Wei; Zhou, Gongbo; Chen, Guoan

    2014-07-01

    Accurately identifying faults in rotor-bearing systems by analyzing vibration signals, which are nonlinear and nonstationary, is challenging. To address this issue, a new approach based on ensemble empirical mode decomposition (EEMD) and self-zero space projection analysis is proposed in this paper. This method seeks to identify faults appearing in a rotor-bearing system using simple algebraic calculations and projection analyses. First, EEMD is applied to decompose the collected vibration signals into a set of intrinsic mode functions (IMFs) for features. Second, these extracted features under various mechanical health conditions are used to design a self-zero space matrix according to space projection analysis. Finally, the so-called projection indicators are calculated to identify the rotor-bearing system's faults with simple decision logic. Experiments are implemented to test the reliability and effectiveness of the proposed approach. The results show that this approach can accurately identify faults in rotor-bearing systems.

  19. A 16-year examination of domestic violence among Asians and Asian Americans in the empirical knowledge base: a content analysis.

    PubMed

    Yick, Alice G; Oomen-Early, Jody

    2008-08-01

    Until recently, research studies have implied that domestic violence does not affect Asian American and immigrant communities, or even Asians abroad, because ethnicity or culture has not been addressed. In this content analysis, the authors examined trends in publications in leading scholarly journals on violence relating to Asian women and domestic violence. A coding schema was developed, with two raters coding the data with high interrater reliability. Sixty articles were published over the 16 years studied, most atheoretical and focusing on individual levels of analysis. The terms used in discussing domestic violence reflected a feminist perspective. Three quarters of the studies were empirical, with most guided by logical positivism using quantitative designs. Most targeted specific Asian subgroups (almost a third focused on Asian Indians) rather than categorizing Asians as a general ethnic category. The concept of "Asian culture" was most often assessed by discussing Asian family structure. Future research is discussed in light of the findings.

  20. Development of reliable pavement models.

    DOT National Transportation Integrated Search

    2011-05-01

    The current report proposes a framework for estimating the reliability of a given pavement structure as analyzed by : the Mechanistic-Empirical Pavement Design Guide (MEPDG). The methodology proposes using a previously fit : response surface, in plac...

  1. Measuring stakeholder participation in evaluation: an empirical validation of the Participatory Evaluation Measurement Instrument (PEMI).

    PubMed

    Daigneault, Pierre-Marc; Jacob, Steve; Tremblay, Joël

    2012-08-01

    Stakeholder participation is an important trend in the field of program evaluation. Although a few measurement instruments have been proposed, they either have not been empirically validated or do not cover the full content of the concept. This study consists of a first empirical validation of a measurement instrument that fully covers the content of participation, namely the Participatory Evaluation Measurement Instrument (PEMI). It specifically examines (1) the intercoder reliability of scores derived by two research assistants on published evaluation cases; (2) the convergence between the scores of coders and those of key respondents (i.e., authors); and (3) the convergence between the authors' scores on the PEMI and the Evaluation Involvement Scale (EIS). A purposive sample of 40 cases drawn from the evaluation literature was used to assess reliability. One author per case in this sample was then invited to participate in a survey; 25 fully usable questionnaires were received. Stakeholder participation was measured on nominal and ordinal scales. Cohen's κ, the intraclass correlation coefficient, and Spearman's ρ were used to assess reliability and convergence. Reliability results ranged from fair to excellent. Convergence between coders' and authors' scores ranged from poor to good. Scores derived from the PEMI and the EIS were moderately associated. Evidence from this study is strong in the case of intercoder reliability and ranges from weak to strong in the case of convergent validation. Globally, this suggests that the PEMI can produce scores that are both reliable and valid.

  2. Genetic Diversity Analysis of Highly Incomplete SNP Genotype Data with Imputations: An Empirical Assessment

    PubMed Central

    Fu, Yong-Bi

    2014-01-01

    Genotyping by sequencing (GBS) recently has emerged as a promising genomic approach for assessing genetic diversity on a genome-wide scale. However, concerns are not lacking about the uniquely large unbalance in GBS genotype data. Although some genotype imputation has been proposed to infer missing observations, little is known about the reliability of a genetic diversity analysis of GBS data, with up to 90% of observations missing. Here we performed an empirical assessment of accuracy in genetic diversity analysis of highly incomplete single nucleotide polymorphism genotypes with imputations. Three large single-nucleotide polymorphism genotype data sets for corn, wheat, and rice were acquired, and missing data with up to 90% of missing observations were randomly generated and then imputed for missing genotypes with three map-independent imputation methods. Estimating heterozygosity and inbreeding coefficient from original, missing, and imputed data revealed variable patterns of bias from assessed levels of missingness and genotype imputation, but the estimation biases were smaller for missing data without genotype imputation. The estimates of genetic differentiation were rather robust up to 90% of missing observations but became substantially biased when missing genotypes were imputed. The estimates of topology accuracy for four representative samples of interested groups generally were reduced with increased levels of missing genotypes. Probabilistic principal component analysis based imputation performed better in terms of topology accuracy than those analyses of missing data without genotype imputation. These findings are not only significant for understanding the reliability of the genetic diversity analysis with respect to large missing data and genotype imputation but also are instructive for performing a proper genetic diversity analysis of highly incomplete GBS or other genotype data. PMID:24626289

  3. Empirical methods for assessing meaningful neuropsychological change following epilepsy surgery.

    PubMed

    Sawrie, S M; Chelune, G J; Naugle, R I; Lüders, H O

    1996-11-01

    Traditional methods for assessing the neurocognitive effects of epilepsy surgery are confounded by practice effects, test-retest reliability issues, and regression to the mean. This study employs 2 methods for assessing individual change that allow direct comparison of changes across both individuals and test measures. Fifty-one medically intractable epilepsy patients completed a comprehensive neuropsychological battery twice, approximately 8 months apart, prior to any invasive monitoring or surgical intervention. First, a Reliable Change (RC) index score was computed for each test score to take into account the reliability of that measure, and a cutoff score was empirically derived to establish the limits of statistically reliable change. These indices were subsequently adjusted for expected practice effects. The second approach used a regression technique to establish "change norms" along a common metric that models both expected practice effects and regression to the mean. The RC index scores provide the clinician with a statistical means of determining whether a patient's retest performance is "significantly" changed from baseline. The regression norms for change allow the clinician to evaluate the magnitude of a given patient's change on 1 or more variables along a common metric that takes into account the reliability and stability of each test measure. Case data illustrate how these methods provide an empirically grounded means for evaluating neurocognitive outcomes following medical interventions such as epilepsy surgery.

  4. Developing and testing the CHORDS: Characteristics of Responsible Drinking Survey.

    PubMed

    Barry, Adam E; Goodson, Patricia

    2011-01-01

    Report on the development and psychometric testing of a theoretically and evidence-grounded instrument, the Characteristics of Responsible Drinking Survey (CHORDS). Instrument subjected to four phases of pretesting (cognitive validity, cognitive and motivational qualities, pilot test, and item evaluation) and a final posttest implementation. Large public university in Texas. Randomly selected convenience sample (n  =  729) of currently enrolled students. This 78-item questionnaire measures individuals' responsible drinking beliefs, motivations, intentions, and behaviors. Cronbach α, split-half reliability, principal components analysis and Spearman ρ were conducted to investigate reliability, stability, and validity. Measures in the CHORDS exhibited high internal consistency reliability and strong correlations of split-half reliability. Factor analyses indicated five distinct scales were present, as proposed in the theoretical model. Subscale composite scores also exhibited a correlation to alcohol consumption behaviors, indicating concurrent validity. The CHORDS represents the first instrument specifically designed to assess responsible drinking beliefs and behaviors. It was found to elicit valid and reliable data among a college student sample. This instrument holds much promise for practitioners who desire to empirically investigate dimensions of responsible drinking.

  5. Financial forecasts accuracy in Brazil's social security system.

    PubMed

    Silva, Carlos Patrick Alves da; Puty, Claudio Alberto Castelo Branco; Silva, Marcelino Silva da; Carvalho, Solon Venâncio de; Francês, Carlos Renato Lisboa

    2017-01-01

    Long-term social security statistical forecasts produced and disseminated by the Brazilian government aim to provide accurate results that would serve as background information for optimal policy decisions. These forecasts are being used as support for the government's proposed pension reform that plans to radically change the Brazilian Constitution insofar as Social Security is concerned. However, the reliability of official results is uncertain since no systematic evaluation of these forecasts has ever been published by the Brazilian government or anyone else. This paper aims to present a study of the accuracy and methodology of the instruments used by the Brazilian government to carry out long-term actuarial forecasts. We base our research on an empirical and probabilistic analysis of the official models. Our empirical analysis shows that the long-term Social Security forecasts are systematically biased in the short term and have significant errors that render them meaningless in the long run. Moreover, the low level of transparency in the methods impaired the replication of results published by the Brazilian Government and the use of outdated data compromises forecast results. In the theoretical analysis, based on a mathematical modeling approach, we discuss the complexity and limitations of the macroeconomic forecast through the computation of confidence intervals. We demonstrate the problems related to error measurement inherent to any forecasting process. We then extend this exercise to the computation of confidence intervals for Social Security forecasts. This mathematical exercise raises questions about the degree of reliability of the Social Security forecasts.

  6. Financial forecasts accuracy in Brazil’s social security system

    PubMed Central

    2017-01-01

    Long-term social security statistical forecasts produced and disseminated by the Brazilian government aim to provide accurate results that would serve as background information for optimal policy decisions. These forecasts are being used as support for the government’s proposed pension reform that plans to radically change the Brazilian Constitution insofar as Social Security is concerned. However, the reliability of official results is uncertain since no systematic evaluation of these forecasts has ever been published by the Brazilian government or anyone else. This paper aims to present a study of the accuracy and methodology of the instruments used by the Brazilian government to carry out long-term actuarial forecasts. We base our research on an empirical and probabilistic analysis of the official models. Our empirical analysis shows that the long-term Social Security forecasts are systematically biased in the short term and have significant errors that render them meaningless in the long run. Moreover, the low level of transparency in the methods impaired the replication of results published by the Brazilian Government and the use of outdated data compromises forecast results. In the theoretical analysis, based on a mathematical modeling approach, we discuss the complexity and limitations of the macroeconomic forecast through the computation of confidence intervals. We demonstrate the problems related to error measurement inherent to any forecasting process. We then extend this exercise to the computation of confidence intervals for Social Security forecasts. This mathematical exercise raises questions about the degree of reliability of the Social Security forecasts. PMID:28859172

  7. Studying the Effect of Deposition Conditions on the Performance and Reliability of MEMS Gas Sensors

    PubMed Central

    Sadek, Khaled; Moussa, Walied

    2007-01-01

    In this paper, the reliability of a micro-electro-mechanical system (MEMS)-based gas sensor has been investigated using Three Dimensional (3D) coupled multiphysics Finite Element (FE) analysis. The coupled field analysis involved a two-way sequential electrothermal fields coupling and a one-way sequential thermal-structural fields coupling. An automated substructuring code was developed to reduce the computational cost involved in simulating this complicated coupled multiphysics FE analysis by up to 76 percent. The substructured multiphysics model was then used to conduct a parametric study of the MEMS-based gas sensor performance in response to the variations expected in the thermal and mechanical characteristics of thin films layers composing the sensing MEMS device generated at various stages of the microfabrication process. Whenever possible, the appropriate deposition variables were correlated in the current work to the design parameters, with good accuracy, for optimum operation conditions of the gas sensor. This is used to establish a set of design rules, using linear and nonlinear empirical relations, which can be utilized in real-time at the design and development decision-making stages of similar gas sensors to enable the microfabrication of these sensors with reliable operation.

  8. Statistical model selection for better prediction and discovering science mechanisms that affect reliability

    DOE PAGES

    Anderson-Cook, Christine M.; Morzinski, Jerome; Blecker, Kenneth D.

    2015-08-19

    Understanding the impact of production, environmental exposure and age characteristics on the reliability of a population is frequently based on underlying science and empirical assessment. When there is incomplete science to prescribe which inputs should be included in a model of reliability to predict future trends, statistical model/variable selection techniques can be leveraged on a stockpile or population of units to improve reliability predictions as well as suggest new mechanisms affecting reliability to explore. We describe a five-step process for exploring relationships between available summaries of age, usage and environmental exposure and reliability. The process involves first identifying potential candidatemore » inputs, then second organizing data for the analysis. Third, a variety of models with different combinations of the inputs are estimated, and fourth, flexible metrics are used to compare them. As a result, plots of the predicted relationships are examined to distill leading model contenders into a prioritized list for subject matter experts to understand and compare. The complexity of the model, quality of prediction and cost of future data collection are all factors to be considered by the subject matter experts when selecting a final model.« less

  9. Initial empirical analysis of nuclear power plant organization and its effect on safety performance

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Olson, J.; McLaughlin, S.D.; Osborn, R.N.

    This report contains an analysis of the relationship between selected aspects of organizational structure and the safety-related performance of nuclear power plants. The report starts by identifying and operationalizing certain key dimensions of organizational structure that may be expected to be related to plant safety performance. Next, indicators of plant safety performance are created by combining existing performance measures into more reliable indicators. Finally, the indicators of plant safety performance using correlational and discriminant analysis. The overall results show that plants with better developed coordination mechanisms, shorter vertical hierarchies, and a greater number of departments tend to perform more safely.

  10. Recoding low-level simulator data into a record of meaningful task performance: the integrated task modeling environment (ITME).

    PubMed

    King, Robert; Parker, Simon; Mouzakis, Kon; Fletcher, Winston; Fitzgerald, Patrick

    2007-11-01

    The Integrated Task Modeling Environment (ITME) is a user-friendly software tool that has been developed to automatically recode low-level data into an empirical record of meaningful task performance. The present research investigated and validated the performance of the ITME software package by conducting complex simulation missions and comparing the task analyses produced by ITME with taskanalyses produced by experienced video analysts. A very high interrater reliability (> or = .94) existed between experienced video analysts and the ITME for the task analyses produced for each mission. The mean session time:analysis time ratio was 1:24 using video analysis techniques and 1:5 using the ITME. It was concluded that the ITME produced task analyses that were as reliable as those produced by experienced video analysts, and significantly reduced the time cost associated with these analyses.

  11. Cross-validating a bidimensional mathematics anxiety scale.

    PubMed

    Haiyan Bai

    2011-03-01

    The psychometric properties of a 14-item bidimensional Mathematics Anxiety Scale-Revised (MAS-R) were empirically cross-validated with two independent samples consisting of 647 secondary school students. An exploratory factor analysis on the scale yielded strong construct validity with a clear two-factor structure. The results from a confirmatory factor analysis indicated an excellent model-fit (χ(2) = 98.32, df = 62; normed fit index = .92, comparative fit index = .97; root mean square error of approximation = .04). The internal consistency (.85), test-retest reliability (.71), interfactor correlation (.26, p < .001), and positive discrimination power indicated that MAS-R is a psychometrically reliable and valid instrument for measuring mathematics anxiety. Math anxiety, as measured by MAS-R, correlated negatively with student achievement scores (r = -.38), suggesting that MAS-R may be a useful tool for classroom teachers and other educational personnel tasked with identifying students at risk of reduced math achievement because of anxiety.

  12. Image Restoration Theory: An Empirical Study of Corporate Apology Tactics Employed by the U.S. Air Force Academy

    DTIC Science & Technology

    2006-07-06

    measurements are applied to metrically defined units and these are used to characterize and compare documents” ( Denzin & Lincoln , 1994, p. 464). Stacks...errors in data interpretation; incorrect sampling; generalization; and inter-coder reliability, calling its validity into question. Denzin and Lincoln ...a content analysis may be “unable to capture the context within which a written text has meaning” ( Denzin & Lincoln , 1994, p. 464). However, the

  13. A Pilot Study Examining the Test-Retest and Internal Consistency Reliability of the ABLLS-R

    ERIC Educational Resources Information Center

    Partington, James W.; Bailey, Autumn; Partington, Scott W.

    2018-01-01

    The literature contains a variety of assessment tools for measuring the skills of individuals with autism or other developmental delays, but most lack adequate empirical evidence supporting their reliability and validity. The current pilot study sought to examine the reliability of scores obtained from the Assessment of Basic Language and Learning…

  14. Multimodal Pressure Flow Analysis: Application of Hilbert Huang Transform in Cerebral Blood Flow Regulation

    PubMed Central

    Lo, Men-Tzung; Hu, Kun; Liu, Yanhui; Peng, C.-K.; Novak, Vera

    2008-01-01

    Quantification of nonlinear interactions between two nonstationary signals presents a computational challenge in different research fields, especially for assessments of physiological systems. Traditional approaches that are based on theories of stationary signals cannot resolve nonstationarity-related issues and, thus, cannot reliably assess nonlinear interactions in physiological systems. In this review we discuss a new technique “Multi-Modal Pressure Flow method (MMPF)” that utilizes Hilbert-Huang transformation to quantify dynamic cerebral autoregulation (CA) by studying interaction between nonstationary cerebral blood flow velocity (BFV) and blood pressure (BP). CA is an important mechanism responsible for controlling cerebral blood flow in responses to fluctuations in systemic BP within a few heart-beats. The influence of CA is traditionally assessed from the relationship between the well-pronounced systemic BP and BFV oscillations induced by clinical tests. Reliable noninvasive assessment of dynamic CA, however, remains a challenge in clinical and diagnostic medicine. In this brief review we: 1) present an overview of transfer function analysis (TFA) that is traditionally used to quantify CA; 2) describe the a MMPF method and its modifications; 3) introduce a newly developed automatic algorithm and engineering aspects of the improved MMPF method; and 4) review clinical applications of MMPF and its sensitivity for detection of CA abnormalities in clinical studies. The MMPF analysis decomposes complex nonstationary BP and BFV signals into multiple empirical modes adaptively so that the fluctuations caused by a specific physiologic process can be represented in a corresponding empirical mode. Using this technique, we recently showed that dynamic CA can be characterized by specific phase delays between the decomposed BP and BFV oscillations, and that the phase shifts are significantly reduced in hypertensive, diabetics and stroke subjects with impaired CA. In addition, the new technique enables reliable assessment of CA using both data collected during clinical test and spontaneous BP/BFV fluctuations during baseline resting conditions. PMID:18725996

  15. A decision-tree approach to the assessment of posttraumatic stress disorder: Engineering empirically rigorous and ecologically valid assessment measures.

    PubMed

    Stewart, Regan W; Tuerk, Peter W; Metzger, Isha W; Davidson, Tatiana M; Young, John

    2016-02-01

    Structured diagnostic interviews are widely considered to be the optimal method of assessing symptoms of posttraumatic stress; however, few clinicians report using structured assessments to guide clinical practice. One commonly cited impediment to these assessment approaches is the amount of time required for test administration and interpretation. Empirically keyed methods to reduce the administration time of structured assessments may be a viable solution to increase the use of standardized and reliable diagnostic tools. Thus, the present research conducted an initial feasibility study using a sample of treatment-seeking military veterans (N = 1,517) to develop a truncated assessment protocol based on the Clinician-Administered Posttraumatic Stress Disorder (PTSD) Scale (CAPS). Decision-tree analysis was utilized to identify a subset of predictor variables among the CAPS items that were most predictive of a diagnosis of PTSD. The algorithm-driven, atheoretical sequence of questions reduced the number of items administered by more than 75% and classified the validation sample at 92% accuracy. These results demonstrated the feasibility of developing a protocol to assess PTSD in a way that imposes little assessment burden while still providing a reliable categorization. (c) 2016 APA, all rights reserved).

  16. Conceptual and Empirical Approaches to Financial Decision-making by Older Adults: Results from a Financial Decision-making Rating Scale.

    PubMed

    Lichtenberg, Peter A; Ocepek-Welikson, Katja; Ficker, Lisa J; Gross, Evan; Rahman-Filipiak, Analise; Teresi, Jeanne A

    2018-01-01

    The objectives of this study were threefold: (1) to empirically test the conceptual model proposed by the Lichtenberg Financial Decision-making Rating Scale (LFDRS); (2) to examine the psychometric properties of the LFDRS contextual factors in financial decision-making by investigating both the reliability and convergent validity of the subscales and total scale, and (3) extending previous work on the scale through the collection of normative data on financial decision-making. A convenience sample of 200 independent function and community dwelling older adults underwent cognitive and financial management testing and were interviewed using the LFDRS. Confirmatory factor analysis, internal consistency measures, and hierarchical regression were used in a sample of 200 community-dwelling older adults, all of whom were making or had recently made a significant financial decision. Results confirmed the scale's reliability and supported the conceptual model. Convergent validity analyses indicate that as hypothesized, cognition is a significant predictor of risk scores. Financial management scores, however, were not predictive of decision-making risk scores. The psychometric properties of the LFDRS support the scale's use as it was proposed. The LFDRS instructions and scale are provided for clinicians to use in financial capacity assessments.

  17. Student mathematical imagination instruments: construction, cultural adaptation and validity

    NASA Astrophysics Data System (ADS)

    Dwijayanti, I.; Budayasa, I. K.; Siswono, T. Y. E.

    2018-03-01

    Imagination has an important role as the center of sensorimotor activity of the students. The purpose of this research is to construct the instrument of students’ mathematical imagination in understanding concept of algebraic expression. The researcher performs validity using questionnaire and test technique and data analysis using descriptive method. Stages performed include: 1) the construction of the embodiment of the imagination; 2) determine the learning style questionnaire; 3) construct instruments; 4) translate to Indonesian as well as adaptation of learning style questionnaire content to student culture; 5) perform content validation. The results stated that the constructed instrument is valid by content validation and empirical validation so that it can be used with revisions. Content validation involves Indonesian linguists, english linguists and mathematics material experts. Empirical validation is done through a legibility test (10 students) and shows that in general the language used can be understood. In addition, a questionnaire test (86 students) was analyzed using a biserial point correlation technique resulting in 16 valid items with a reliability test using KR 20 with medium reability criteria. While the test instrument test (32 students) to find all items are valid and reliability test using KR 21 with reability is 0,62.

  18. Participation in Decision Making as a Property of Complex Adaptive Systems: Developing and Testing a Measure

    PubMed Central

    Anderson, Ruth A.; Hsieh, Pi-Ching; Su, Hui Fang; Landerman, Lawrence R.; McDaniel, Reuben R.

    2013-01-01

    Objectives. To (1) describe participation in decision-making as a systems-level property of complex adaptive systems and (2) present empirical evidence of reliability and validity of a corresponding measure. Method. Study 1 was a mail survey of a single respondent (administrators or directors of nursing) in each of 197 nursing homes. Study 2 was a field study using random, proportionally stratified sampling procedure that included 195 organizations with 3,968 respondents. Analysis. In Study 1, we analyzed the data to reduce the number of scale items and establish initial reliability and validity. In Study 2, we strengthened the psychometric test using a large sample. Results. Results demonstrated validity and reliability of the participation in decision-making instrument (PDMI) while measuring participation of workers in two distinct job categories (RNs and CNAs). We established reliability at the organizational level aggregated items scores. We established validity of the multidimensional properties using convergent and discriminant validity and confirmatory factor analysis. Conclusions. Participation in decision making, when modeled as a systems-level property of organization, has multiple dimensions and is more complex than is being traditionally measured. Managers can use this model to form decision teams that maximize the depth and breadth of expertise needed and to foster connection among them. PMID:24349771

  19. Development and validation of an instrument for evaluating inquiry-based tasks in science textbooks

    NASA Astrophysics Data System (ADS)

    Yang, Wenyuan; Liu, Enshan

    2016-12-01

    This article describes the development and validation of an instrument that can be used for content analysis of inquiry-based tasks. According to the theories of educational evaluation and qualities of inquiry, four essential functions that inquiry-based tasks should serve are defined: (1) assisting in the construction of understandings about scientific concepts, (2) providing students opportunities to use inquiry process skills, (3) being conducive to establishing understandings about scientific inquiry, and (4) giving students opportunities to develop higher order thinking skills. An instrument - the Inquiry-Based Tasks Analysis Inventory (ITAI) - was developed to judge whether inquiry-based tasks perform these functions well. To test the reliability and validity of the ITAI, 4 faculty members were invited to use the ITAI to collect data from 53 inquiry-based tasks in the 3 most widely adopted senior secondary biology textbooks in Mainland China. The results indicate that (1) the inter-rater reliability reached 87.7%, (2) the grading criteria have high discriminant validity, (3) the items possess high convergent validity, and (4) the Cronbach's alpha reliability coefficient reached 0.792. The study concludes that the ITAI is valid and reliable. Because of its solid foundations in theoretical and empirical argumentation, the ITAI is trustworthy.

  20. Participation in decision making as a property of complex adaptive systems: developing and testing a measure.

    PubMed

    Anderson, Ruth A; Plowman, Donde; Corazzini, Kirsten; Hsieh, Pi-Ching; Su, Hui Fang; Landerman, Lawrence R; McDaniel, Reuben R

    2013-01-01

    Objectives. To (1) describe participation in decision-making as a systems-level property of complex adaptive systems and (2) present empirical evidence of reliability and validity of a corresponding measure. Method. Study 1 was a mail survey of a single respondent (administrators or directors of nursing) in each of 197 nursing homes. Study 2 was a field study using random, proportionally stratified sampling procedure that included 195 organizations with 3,968 respondents. Analysis. In Study 1, we analyzed the data to reduce the number of scale items and establish initial reliability and validity. In Study 2, we strengthened the psychometric test using a large sample. Results. Results demonstrated validity and reliability of the participation in decision-making instrument (PDMI) while measuring participation of workers in two distinct job categories (RNs and CNAs). We established reliability at the organizational level aggregated items scores. We established validity of the multidimensional properties using convergent and discriminant validity and confirmatory factor analysis. Conclusions. Participation in decision making, when modeled as a systems-level property of organization, has multiple dimensions and is more complex than is being traditionally measured. Managers can use this model to form decision teams that maximize the depth and breadth of expertise needed and to foster connection among them.

  1. How to assess and compare inter-rater reliability, agreement and correlation of ratings: an exemplary analysis of mother-father and parent-teacher expressive vocabulary rating pairs

    PubMed Central

    Stolarova, Margarita; Wolf, Corinna; Rinker, Tanja; Brielmann, Aenne

    2014-01-01

    This report has two main purposes. First, we combine well-known analytical approaches to conduct a comprehensive assessment of agreement and correlation of rating-pairs and to dis-entangle these often confused concepts, providing a best-practice example on concrete data and a tutorial for future reference. Second, we explore whether a screening questionnaire developed for use with parents can be reliably employed with daycare teachers when assessing early expressive vocabulary. A total of 53 vocabulary rating pairs (34 parent–teacher and 19 mother–father pairs) collected for two-year-old children (12 bilingual) are evaluated. First, inter-rater reliability both within and across subgroups is assessed using the intra-class correlation coefficient (ICC). Next, based on this analysis of reliability and on the test-retest reliability of the employed tool, inter-rater agreement is analyzed, magnitude and direction of rating differences are considered. Finally, Pearson correlation coefficients of standardized vocabulary scores are calculated and compared across subgroups. The results underline the necessity to distinguish between reliability measures, agreement and correlation. They also demonstrate the impact of the employed reliability on agreement evaluations. This study provides evidence that parent–teacher ratings of children's early vocabulary can achieve agreement and correlation comparable to those of mother–father ratings on the assessed vocabulary scale. Bilingualism of the evaluated child decreased the likelihood of raters' agreement. We conclude that future reports of agreement, correlation and reliability of ratings will benefit from better definition of terms and stricter methodological approaches. The methodological tutorial provided here holds the potential to increase comparability across empirical reports and can help improve research practices and knowledge transfer to educational and therapeutic settings. PMID:24994985

  2. Statistical validity of using ratio variables in human kinetics research.

    PubMed

    Liu, Yuanlong; Schutz, Robert W

    2003-09-01

    The purposes of this study were to investigate the validity of the simple ratio and three alternative deflation models and examine how the variation of the numerator and denominator variables affects the reliability of a ratio variable. A simple ratio and three alternative deflation models were fitted to four empirical data sets, and common criteria were applied to determine the best model for deflation. Intraclass correlation was used to examine the component effect on the reliability of a ratio variable. The results indicate that the validity, of a deflation model depends on the statistical characteristics of the particular component variables used, and an optimal deflation model for all ratio variables may not exist. Therefore, it is recommended that different models be fitted to each empirical data set to determine the best deflation model. It was found that the reliability of a simple ratio is affected by the coefficients of variation and the within- and between-trial correlations between the numerator and denominator variables. It was recommended that researchers should compute the reliability of the derived ratio scores and not assume that strong reliabilities in the numerator and denominator measures automatically lead to high reliability in the ratio measures.

  3. Beyond alpha: an empirical examination of the effects of different sources of measurement error on reliability estimates for measures of individual differences constructs.

    PubMed

    Schmidt, Frank L; Le, Huy; Ilies, Remus

    2003-06-01

    On the basis of an empirical study of measures of constructs from the cognitive domain, the personality domain, and the domain of affective traits, the authors of this study examine the implications of transient measurement error for the measurement of frequently studied individual differences variables. The authors clarify relevant reliability concepts as they relate to transient error and present a procedure for estimating the coefficient of equivalence and stability (L. J. Cronbach, 1947), the only classical reliability coefficient that assesses all 3 major sources of measurement error (random response, transient, and specific factor errors). The authors conclude that transient error exists in all 3 trait domains and is especially large in the domain of affective traits. Their findings indicate that the nearly universal use of the coefficient of equivalence (Cronbach's alpha; L. J. Cronbach, 1951), which fails to assess transient error, leads to overestimates of reliability and undercorrections for biases due to measurement error.

  4. Online Patient Education for Chronic Disease Management: Consumer Perspectives.

    PubMed

    Win, Khin Than; Hassan, Naffisah Mohd; Oinas-Kukkonen, Harri; Probst, Yasmine

    2016-04-01

    Patient education plays an important role in chronic disease management. The aim of this study is to identify patients' preferences in regard to the design features of effective online patient education (OPE) and the benefits. A review of the existing literature was conducted in order to identify the benefits of OPE and its essential design features. These design features were empirically tested by conducting survey with patients and caregivers. Reliability analysis, construct validity and regression analysis were performed for data analysis. The results identified patient-tailored information, interactivity, content credibility, clear presentation of content, use of multimedia and interpretability as the essential design features of online patient education websites for chronic disease management.

  5. Performance-Based Service Quality Model: An Empirical Study on Japanese Universities

    ERIC Educational Resources Information Center

    Sultan, Parves; Wong, Ho

    2010-01-01

    Purpose: This paper aims to develop and empirically test the performance-based higher education service quality model. Design/methodology/approach: The study develops 67-item instrument for measuring performance-based service quality with a particular focus on the higher education sector. Scale reliability is confirmed using the Cronbach's alpha.…

  6. Evaluating the intersection of a regional wildlife connectivity network with highways

    Treesearch

    Samuel A. Cushman; Jesse S. Lewis; Erin L. Landguth

    2013-01-01

    Reliable predictions of regional-scale population connectivity are needed to prioritize conservation actions. However, there have been few examples of regional connectivity models that are empirically derived and validated. The central goals of this paper were to (1) evaluate the effectiveness of factorial least cost path corridor mapping on an empirical...

  7. Carbon deposition model for oxygen-hydrocarbon combustion. Task 6: Data analysis and formulation of an empirical model

    NASA Technical Reports Server (NTRS)

    Makel, Darby B.; Rosenberg, Sanders D.

    1990-01-01

    The formation and deposition of carbon (soot) was studied in the Carbon Deposition Model for Oxygen-Hydrocarbon Combustion Program. An empirical, 1-D model for predicting soot formation and deposition in LO2/hydrocarbon gas generators/preburners was derived. The experimental data required to anchor the model were identified and a test program to obtain the data was defined. In support of the model development, cold flow mixing experiments using a high injection density injector were performed. The purpose of this investigation was to advance the state-of-the-art in LO2/hydrocarbon gas generator design by developing a reliable engineering model of gas generator operation. The model was formulated to account for the influences of fluid dynamics, chemical kinetics, and gas generator hardware design on soot formation and deposition.

  8. The Use of Empirical Data Sources in HRA

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Bruce Hallbert; David Gertman; Julie Marble

    This paper presents a review of available information related to human performance to support Human Reliability Analysis (HRA) performed for nuclear power plants (NPPs). A number of data sources are identified as potentially useful. These include NPP licensee event reports (LERs), augmented inspection team (AIT) reports, operator requalification data, results from the literature in experimental psychology, and the Aviation Safety Reporting System (ASRSs). The paper discusses how utilizing such information improves our capability to model and quantify human performance. In particular the paper discusses how information related to performance shaping factors (PSFs) can be extracted from empirical data to determinemore » their size effect, their relative effects, as well as their interactions. The paper concludes that appropriate use of existing sources can help addressing some of the important issues we are currently facing in HRA.« less

  9. A Poisson process approximation for generalized K-5 confidence regions

    NASA Technical Reports Server (NTRS)

    Arsham, H.; Miller, D. R.

    1982-01-01

    One-sided confidence regions for continuous cumulative distribution functions are constructed using empirical cumulative distribution functions and the generalized Kolmogorov-Smirnov distance. The band width of such regions becomes narrower in the right or left tail of the distribution. To avoid tedious computation of confidence levels and critical values, an approximation based on the Poisson process is introduced. This aproximation provides a conservative confidence region; moreover, the approximation error decreases monotonically to 0 as sample size increases. Critical values necessary for implementation are given. Applications are made to the areas of risk analysis, investment modeling, reliability assessment, and analysis of fault tolerant systems.

  10. Validation of the Implementation Leadership Scale (ILS) with Supervisors' Self-Ratings.

    PubMed

    Torres, Elisa M; Ehrhart, Mark G; Beidas, Rinad S; Farahnak, Lauren R; Finn, Natalie K; Aarons, Gregory A

    2018-01-01

    Although often discussed, there is a lack of empirical research on the role of leadership in the management and delivery of health services. The implementation leadership scale (ILS) assesses the degree to which leaders are knowledgeable, proactive, perseverant, and supportive during evidence-based practice (EBP) implementation. The purpose of this study was to examine the psychometric properties of the ILS for leaders' self-ratings using a sample of mental health clinic supervisors (N = 119). Supervisors (i.e., leaders) completed surveys including self-ratings of their implementation leadership. Confirmatory factor analysis, reliability, and validity of the ILS were evaluated. The ILS factor structure was supported in the sample of supervisors. Results demonstrated internal consistency reliability and validity. Cronbach alpha's ranged from 0.92 to 0.96 for the ILS subscales and 0.95 for the ILS overall scale. The factor structure replication and reliability of the ILS in a sample of supervisors demonstrates its applicability with employees across organizational levels.

  11. Racism as a determinant of health: a protocol for conducting a systematic review and meta-analysis.

    PubMed

    Paradies, Yin; Priest, Naomi; Ben, Jehonathan; Truong, Mandy; Gupta, Arpana; Pieterse, Alex; Kelaher, Margaret; Gee, Gilbert

    2013-09-23

    Racism is increasingly recognized as a key determinant of health. A growing body of epidemiological evidence shows strong associations between self-reported racism and poor health outcomes across diverse minority groups in developed countries. While the relationship between racism and health has received increasing attention over the last two decades, a comprehensive meta-analysis focused on the health effects of racism has yet to be conducted. The aim of this review protocol is to provide a structure from which to conduct a systematic review and meta-analysis of studies that assess the relationship between racism and health. This research will consist of a systematic review and meta-analysis. Studies will be considered for review if they are empirical studies reporting quantitative data on the association between racism and health for adults and/or children of all ages from any racial/ethnic/cultural groups. Outcome measures will include general health and well-being, physical health, mental health, healthcare use and health behaviors. Scientific databases (for example, Medline) will be searched using a comprehensive search strategy and reference lists will be manually searched for relevant studies. In addition, use of online search engines (for example, Google Scholar), key websites, and personal contact with experts will also be undertaken. Screening of search results and extraction of data from included studies will be independently conducted by at least two authors, including assessment of inter-rater reliability. Studies included in the review will be appraised for quality using tools tailored to each study design. Summary statistics of study characteristics and findings will be compiled and findings synthesized in a narrative summary as well as a meta-analysis. This review aims to examine associations between reported racism and health outcomes. This comprehensive and systematic review and meta-analysis of empirical research will provide a rigorous and reliable evidence base for future research, policy and practice, including information on the extent of available evidence for a range of racial/ethnic minority groups.

  12. An empirical look at the Defense Mechanism Test (DMT): reliability and construct validity.

    PubMed

    Ekehammar, Bo; Zuber, Irena; Konstenius, Marja-Liisa

    2005-07-01

    Although the Defense Mechanism Test (DMT) has been in use for almost half a century, there are still quite contradictory views about whether it is a reliable instrument, and if so, what it really measures. Thus, based on data from 39 female students, we first examined DMT inter-coder reliability by analyzing the agreement among trained judges in their coding of the same DMT protocols. Second, we constructed a "parallel" photographic picture that retained all structural characteristic of the original and analyzed DMT parallel-test reliability. Third, we examined the construct validity of the DMT by (a) employing three self-report defense-mechanism inventories and analyzing the intercorrelations between DMT defense scores and corresponding defenses in these instruments, (b) studying the relationships between DMT responses and scores on trait and state anxiety, and (c) relating DMT-defense scores to measures of self-esteem. The main results showed that the DMT can be coded with high reliability by trained coders, that the parallel-test reliability is unsatisfactory compared to traditional psychometric standards, that there is a certain generalizability in the number of perceptual distortions that people display from one picture to another, and that the construct validation provided meager empirical evidence for the conclusion that the DMT measures what it purports to measure, that is, psychological defense mechanisms.

  13. Combining empirical approaches and error modelling to enhance predictive uncertainty estimation in extrapolation for operational flood forecasting. Tests on flood events on the Loire basin, France.

    NASA Astrophysics Data System (ADS)

    Berthet, Lionel; Marty, Renaud; Bourgin, François; Viatgé, Julie; Piotte, Olivier; Perrin, Charles

    2017-04-01

    An increasing number of operational flood forecasting centres assess the predictive uncertainty associated with their forecasts and communicate it to the end users. This information can match the end-users needs (i.e. prove to be useful for an efficient crisis management) only if it is reliable: reliability is therefore a key quality for operational flood forecasts. In 2015, the French flood forecasting national and regional services (Vigicrues network; www.vigicrues.gouv.fr) implemented a framework to compute quantitative discharge and water level forecasts and to assess the predictive uncertainty. Among the possible technical options to achieve this goal, a statistical analysis of past forecasting errors of deterministic models has been selected (QUOIQUE method, Bourgin, 2014). It is a data-based and non-parametric approach based on as few assumptions as possible about the forecasting error mathematical structure. In particular, a very simple assumption is made regarding the predictive uncertainty distributions for large events outside the range of the calibration data: the multiplicative error distribution is assumed to be constant, whatever the magnitude of the flood. Indeed, the predictive distributions may not be reliable in extrapolation. However, estimating the predictive uncertainty for these rare events is crucial when major floods are of concern. In order to improve the forecasts reliability for major floods, an attempt at combining the operational strength of the empirical statistical analysis and a simple error modelling is done. Since the heteroscedasticity of forecast errors can considerably weaken the predictive reliability for large floods, this error modelling is based on the log-sinh transformation which proved to reduce significantly the heteroscedasticity of the transformed error in a simulation context, even for flood peaks (Wang et al., 2012). Exploratory tests on some operational forecasts issued during the recent floods experienced in France (major spring floods in June 2016 on the Loire river tributaries and flash floods in fall 2016) will be shown and discussed. References Bourgin, F. (2014). How to assess the predictive uncertainty in hydrological modelling? An exploratory work on a large sample of watersheds, AgroParisTech Wang, Q. J., Shrestha, D. L., Robertson, D. E. and Pokhrel, P (2012). A log-sinh transformation for data normalization and variance stabilization. Water Resources Research, , W05514, doi:10.1029/2011WR010973

  14. Gravity-darkening exponents in semi-detached binary systems from their photometric observations. II.

    NASA Astrophysics Data System (ADS)

    Djurašević, G.; Rovithis-Livaniou, H.; Rovithis, P.; Georgiades, N.; Erkapić, S.; Pavlović, R.

    2006-01-01

    This second part of our study concerning gravity-darkening presents the results for 8 semi-detached close binary systems. From the light-curve analysis of these systems the exponent of the gravity-darkening (GDE) for the Roche lobe filling components has been empirically derived. The method used for the light-curve analysis is based on Roche geometry, and enables simultaneous estimation of the systems' parameters and the gravity-darkening exponents. Our analysis is restricted to the black-body approximation which can influence in some degree the parameter estimation. The results of our analysis are: 1) For four of the systems, namely: TX UMa, β Per, AW Cam and TW Cas, there is a very good agreement between empirically estimated and theoretically predicted values for purely convective envelopes. 2) For the AI Dra system, the estimated value of gravity-darkening exponent is greater, and for UX Her, TW And and XZ Pup lesser than corresponding theoretical predictions, but for all mentioned systems the obtained values of the gravity-darkening exponent are quite close to the theoretically expected values. 3) Our analysis has proved generally that with the correction of the previously estimated mass ratios of the components within some of the analysed systems, the theoretical predictions of the gravity-darkening exponents for stars with convective envelopes are highly reliable. The anomalous values of the GDE found in some earlier studies of these systems can be considered as the consequence of the inappropriate method used to estimate the GDE. 4) The empirical estimations of GDE given in Paper I and in the present study indicate that in the light-curve analysis one can apply the recent theoretical predictions of GDE with high confidence for stars with both convective and radiative envelopes.

  15. Assessment of Prevalence of Persons with Down Syndrome: A Theory-Based Demographic Model

    ERIC Educational Resources Information Center

    de Graaf, Gert; Vis, Jeroen C.; Haveman, Meindert; van Hove, Geert; de Graaf, Erik A. B.; Tijssen, Jan G. P.; Mulder, Barbara J. M.

    2011-01-01

    Background: The Netherlands are lacking reliable empirical data in relation to the development of birth and population prevalence of Down syndrome. For the UK and Ireland there are more historical empirical data available. A theory-based model is developed for predicting Down syndrome prevalence in the Netherlands from the 1950s onwards. It is…

  16. Predicting the Magnetic Properties of ICMEs: A Pragmatic View

    NASA Astrophysics Data System (ADS)

    Riley, P.; Linker, J.; Ben-Nun, M.; Torok, T.; Ulrich, R. K.; Russell, C. T.; Lai, H.; de Koning, C. A.; Pizzo, V. J.; Liu, Y.; Hoeksema, J. T.

    2017-12-01

    The southward component of the interplanetary magnetic field plays a crucial role in being able to successfully predict space weather phenomena. Yet, thus far, it has proven extremely difficult to forecast with any degree of accuracy. In this presentation, we describe an empirically-based modeling framework for estimating Bz values during the passage of interplanetary coronal mass ejections (ICMEs). The model includes: (1) an empirically-based estimate of the magnetic properties of the flux rope in the low corona (including helicity and field strength); (2) an empirically-based estimate of the dynamic properties of the flux rope in the high corona (including direction, speed, and mass); and (3) a physics-based estimate of the evolution of the flux rope during its passage to 1 AU driven by the output from (1) and (2). We compare model output with observations for a selection of events to estimate the accuracy of this approach. Importantly, we pay specific attention to the uncertainties introduced by the components within the framework, separating intrinsic limitations from those that can be improved upon, either by better observations or more sophisticated modeling. Our analysis suggests that current observations/modeling are insufficient for this empirically-based framework to provide reliable and actionable prediction of the magnetic properties of ICMEs. We suggest several paths that may lead to better forecasts.

  17. Conceptual and Empirical Approaches to Financial Decision-making by Older Adults: Results from a Financial Decision-Making Rating Scale

    PubMed Central

    Lichtenberg, Peter A.; Ocepek-Welikson, Katja; Ficker, Lisa J.; Gross, Evan; Rahman-Filipiak, Analise; Teresi, Jeanne A.

    2017-01-01

    Objectives The objectives of this study were threefold: (1) to empirically test the conceptual model proposed by the Lichtenberg Financial Decision Rating Scale (LFDRS); (2) to examine the psychometric properties of the LFDRS contextual factors in financial decision-making by investigating both the reliability and convergent validity of the subscales and total scale, and (3) extending previous work on the scale through the collection of normative data on financial decision-making. Methods A convenience sample of 200 independent function and community dwelling older adults underwent cognitive and financial management testing and were interviewed using the LFDRS. Confirmatory factor analysis, internal consistency measures, and hierarchical regression were used in a sample of 200 community-dwelling older adults, all of whom were making or had recently made a significant financial decision. Results Results confirmed the scale’s reliability and supported the conceptual model. Convergent validity analyses indicate that as hypothesized, cognition is a significant predictor of risk scores. Financial management scores, however, were not predictive of decision-making risk scores. Conclusions The psychometric properties of the LFDRS support the scale’s use as it was proposed in Lichtenberg et al., 2015. Clinical Implications The LFDRS instructions and scale are provided for clinicians to use in financial capacity assessments. PMID:29077531

  18. A Tentative Study on the Evaluation of Community Health Service Quality*

    NASA Astrophysics Data System (ADS)

    Ma, Zhi-qiang; Zhu, Yong-yue

    Community health service is the key point of health reform in China. Based on pertinent studies, this paper constructed an indicator system for the community health service quality evaluation from such five perspectives as visible image, reliability, responsiveness, assurance and sympathy, according to service quality evaluation scale designed by Parasuraman, Zeithaml and Berry. A multilevel fuzzy synthetical evaluation model was constructed to evaluate community health service by fuzzy mathematics theory. The applicability and maneuverability of the evaluation indicator system and evaluation model were verified by empirical analysis.

  19. Existing generating assets squeezed as new project starts slow

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Jones, R.B.; Tiffany, E.D.

    Most forecasting reports concentrate on political or regulatory events to predict future industry trends. Frequently overlooked are the more empirical performance trends of the principal power generation technologies. Solomon and Associates queried its many power plant performance databases and crunched some numbers to identify those trends. Areas of investigation included reliability, utilization (net output factor and net capacity factor) and cost (operating costs). An in-depth analysis for North America and Europe is presented in this article, by region and by regeneration technology. 4 figs., 2 tabs.

  20. Do health technology assessments comply with QUOROM diagram guidance? An empirical study.

    PubMed

    Hind, Daniel; Booth, Andrew

    2007-11-20

    The Quality of Reporting of Meta-analyses (QUOROM) statement provides guidance for improving the quality of reporting of systematic reviews and meta-analyses. To make the process of study selection transparent it recommends "a flow diagram providing information about the number of RCTs identified, included, and excluded and the reasons for excluding them". We undertook an empirical study to identify the extent of compliance in the UK Health Technology Assessment (HTA) programme. We searched Medline to retrieve all systematic reviews of therapeutic interventions in the HTA monograph series published from 2001 to 2005. Two researchers recorded whether each study contained a meta-analysis of controlled trials, whether a QUOROM flow diagram was presented and, if so, whether it expressed the relationship between the number of citations and the number of studies. We used Cohen's kappa to test inter-rater reliability. 87 systematic reviews were retrieved. There was good and excellent inter-rater reliability for, respectively, whether a review contained a meta-analysis and whether each diagram contained a citation-to-study relationship. 49% of systematic reviews used a study selection flow diagram. When only systematic reviews containing a meta-analysis were analysed, compliance was only 32%. Only 20 studies (23% of all systematic reviews; 43% of those having a study selection diagram) had a diagram which expressed the relationship between citations and studies. Compliance with the recommendations of the QUOROM statement is not universal in systematic reviews or meta-analyses. Flow diagrams make the conduct of study selection transparent only if the relationship between citations and studies is clearly expressed. Reviewers should understand what they are counting: citations, papers, studies and trials are fundamentally different concepts which should not be confused in a diagram.

  1. Assessing the Conditional Reliability of State Assessments

    ERIC Educational Resources Information Center

    May, Henry; Cole, Russell; Haimson, Josh; Perez-Johnson, Irma

    2010-01-01

    The purpose of this study is to provide empirical benchmarks of the conditional reliabilities of state tests for samples of the student population defined by ability level. Given that many educational interventions are targeted for samples of low performing students, schools, or districts, the primary goal of this research is to determine how…

  2. Examining Readability Estimates' Predictions of Students' Oral Reading Rate: Spache, Lexile, and Forcast

    ERIC Educational Resources Information Center

    Ardoin, Scott P.; Williams, Jessica C.; Christ, Theodore J.; Klubnik, Cynthia; Wellborn, Claire

    2010-01-01

    Beyond reliability and validity, measures used to model student growth must consist of multiple probes that are equivalent in level of difficulty to establish consistent measurement conditions across time. Although existing evidence supports the reliability of curriculum-based measurement in reading (CBMR), few studies have empirically evaluated…

  3. Reliability and Validity of a Turkish Version of the DELES

    ERIC Educational Resources Information Center

    Ozkok, Alev; Walker, Scott L.; Buyukozturk, Sener

    2009-01-01

    The primary aim of this study was to examine the reliability and validity of a Turkish version of the Distance Education Learning Environment Survey (DELES) in post-secondary distance education. The second aim was to investigate empirically the conceptualisation of the distance education learning environment as a singular latent construct, within…

  4. An Assessment of Reliability and Validity of a Rubric for Grading APA-Style Introductions

    ERIC Educational Resources Information Center

    Stellmack, Mark A.; Konheim-Kalkstein, Yasmine L.; Manor, Julia E.; Massey, Abigail R.; Schmitz, Julie Ann P.

    2009-01-01

    This article describes the empirical evaluation of the reliability and validity of a grading rubric for grading APA-style introductions of undergraduate students. Levels of interrater agreement and intrarater agreement were not extremely high but were similar to values reported in the literature for comparably structured rubrics. Rank-order…

  5. Application of Stein and related parametric empirical Bayes estimators to the nuclear plant reliability data system

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Hill, J.R.; Heger, A.S.; Koen, B.V.

    1984-04-01

    This report is the result of a preliminary feasibility study of the applicability of Stein and related parametric empirical Bayes (PEB) estimators to the Nuclear Plant Reliability Data System (NPRDS). A new estimator is derived for the means of several independent Poisson distributions with different sampling times. This estimator is applied to data from NPRDS in an attempt to improve failure rate estimation. Theoretical and Monte Carlo results indicate that the new PEB estimator can perform significantly better than the standard maximum likelihood estimator if the estimation of the individual means can be combined through the loss function or throughmore » a parametric class of prior distributions.« less

  6. Trend extraction using empirical mode decomposition and statistical empirical mode decomposition: Case study: Kuala Lumpur stock market

    NASA Astrophysics Data System (ADS)

    Jaber, Abobaker M.

    2014-12-01

    Two nonparametric methods for prediction and modeling of financial time series signals are proposed. The proposed techniques are designed to handle non-stationary and non-linearity behave and to extract meaningful signals for reliable prediction. Due to Fourier Transform (FT), the methods select significant decomposed signals that will be employed for signal prediction. The proposed techniques developed by coupling Holt-winter method with Empirical Mode Decomposition (EMD) and it is Extending the scope of empirical mode decomposition by smoothing (SEMD). To show performance of proposed techniques, we analyze daily closed price of Kuala Lumpur stock market index.

  7. Empirical validation of the English version of the Fear of Cancer Recurrence Inventory.

    PubMed

    Lebel, Sophie; Simard, Sebastien; Harris, Cheryl; Feldstain, Andrea; Beattie, Sara; McCallum, Megan; Lefebvre, Monique; Savard, Josée; Devins, Gerald M

    2016-02-01

    Cancer patients report that help in managing fear of cancer recurrence (FCR) is one of their greatest unmet needs. Research on FCR has been limited by the very few validated, multi-dimensional measures of this construct. One exception is the Fear of Cancer Recurrence Inventory (FCRI), originally developed and empirically validated in French. The present study validated the English version of the FCRI. The FCRI was translated into English using a forward-backward translation procedure and pilot-tested with 17 English-speaking cancer patients. Cross-cultural equivalency of the French and English versions was established by administering both forms to 42 bilingual cancer patients. Last, 350 English-speaking breast, colon, prostate, or lung cancer patients were asked to complete the FCRI. A subsample (n = 135) was mailed the FCRI again one month later to evaluate test-retest reliability. The English translation of the FCRI was well accepted by participants. There was no item-bias when comparing bilingual participants' answers on both versions. A confirmatory factor analysis supported the hypothesized seven-factor structure. The English version has high internal consistency (α = .96 for the total scale and .71-.94 for the subscales) and test-retest reliability (r = .88 for the total scale and 56-.87 for the subscales). The English version of the FCRI is a reliable and valid measure of FCR applicable to breast, colon, prostate, and lung cancer patients. Its multi-dimensional nature makes it an attractive research and clinical tool to further our knowledge of FCR.

  8. Measuring the gradualist approach to internationalization: Empirical evidence from the wine sector

    PubMed Central

    2018-01-01

    The objective of this paper is to fill a gap in the literature on internationalization, in relation to the absence of objective and measurable performance indicators for the process of how firms sequentially enter external markets. To that end, this research develops a quantitative tool for use as a performance indicator of gradualness for firms entering external markets at a sectoral level. The performance indicator is based on firms’ export volumes, number of years operating in the export market, geographic areas targeted for export and when exports began to each area. The indicator is tested empirically in the wine sector. The main contribution of this study is the creation of a reliable international priority index, which can serve more widely as a valuable tool because of its potential use in other industry sectors and geographic areas, and which would allow the analysis of how geographically differentiated internationalization strategies develop. PMID:29727461

  9. Measuring the gradualist approach to internationalization: Empirical evidence from the wine sector.

    PubMed

    Clavel San Emeterio, Mónica; Fernández-Ortiz, Rubén; Arteaga-Ortiz, Jesús; Dorta-González, Pablo

    2018-01-01

    The objective of this paper is to fill a gap in the literature on internationalization, in relation to the absence of objective and measurable performance indicators for the process of how firms sequentially enter external markets. To that end, this research develops a quantitative tool for use as a performance indicator of gradualness for firms entering external markets at a sectoral level. The performance indicator is based on firms' export volumes, number of years operating in the export market, geographic areas targeted for export and when exports began to each area. The indicator is tested empirically in the wine sector. The main contribution of this study is the creation of a reliable international priority index, which can serve more widely as a valuable tool because of its potential use in other industry sectors and geographic areas, and which would allow the analysis of how geographically differentiated internationalization strategies develop.

  10. Validity and reliability of the "German Utilization Questionnaire-Dissemination and Use of Research" to measure attitude, availability, and support toward implementation of research in nursing practice.

    PubMed

    Haslinger-Baumann, Elisabeth; Lang, Gert; Müller, Gerhard

    2014-01-01

    In nursing practice, research results have to undergo a systematic process of transformation. Currently in Austria, there is no empirical data available concerning the actual implementation of research results. An English validated questionnaire was translated into German and tested for validity and reliability. A survey of 178 registered nurses (n = 178) was conducted in a multicenter, quantitative, cross-sectional study in Austria in 2011. Cronbach's alpha values (.82-.92) were calculated for 4 variables ("use," "attitude," "availability," "support") after the reduction of 7 irrelevant items. Exploratory factor analysis was calculated with Kaiser-Meyer-Olkin (KMO) ranging from .78 to .92; the total variance ranged from 46% to 56%. A validated German questionnaire concerning the implementation of research results is now available for the nursing practice.

  11. Gravity Tides Extracted from Relative Gravimeter Data by Combining Empirical Mode Decomposition and Independent Component Analysis

    NASA Astrophysics Data System (ADS)

    Yu, Hongjuan; Guo, Jinyun; Kong, Qiaoli; Chen, Xiaodong

    2018-04-01

    The static observation data from a relative gravimeter contain noise and signals such as gravity tides. This paper focuses on the extraction of the gravity tides from the static relative gravimeter data for the first time applying the combined method of empirical mode decomposition (EMD) and independent component analysis (ICA), called the EMD-ICA method. The experimental results from the CG-5 gravimeter (SCINTREX Limited Ontario Canada) data show that the gravity tides time series derived by EMD-ICA are consistent with the theoretical reference (Longman formula) and the RMS of their differences only reaches 4.4 μGal. The time series of the gravity tides derived by EMD-ICA have a strong correlation with the theoretical time series and the correlation coefficient is greater than 0.997. The accuracy of the gravity tides estimated by EMD-ICA is comparable to the theoretical model and is slightly higher than that of independent component analysis (ICA). EMD-ICA could overcome the limitation of ICA having to process multiple observations and slightly improve the extraction accuracy and reliability of gravity tides from relative gravimeter data compared to that estimated with ICA.

  12. Validation of the Narcissistic Admiration and Rivalry Questionnaire Short Scale (NARQ-S) in convenience and representative samples.

    PubMed

    Leckelt, Marius; Wetzel, Eunike; Gerlach, Tanja M; Ackerman, Robert A; Miller, Joshua D; Chopik, William J; Penke, Lars; Geukes, Katharina; Küfner, Albrecht C P; Hutteman, Roos; Richter, David; Renner, Karl-Heinz; Allroggen, Marc; Brecheen, Courtney; Campbell, W Keith; Grossmann, Igor; Back, Mitja D

    2018-01-01

    Due to increased empirical interest in narcissism across the social sciences, there is a need for inventories that can be administered quickly while also reliably measuring both the agentic and antagonistic aspects of grandiose narcissism. In this study, we sought to validate the factor structure, provide representative descriptive data and reliability estimates, assess the reliability across the trait spectrum, and examine the nomological network of the short version of the Narcissistic Admiration and Rivalry Questionnaire (NARQ-S; Back et al., 2013). We used data from a large convenience sample (total N = 11,937) as well as data from a large representative sample (total N = 4,433) that included responses to other narcissism measures as well as related constructs, including the other Dark Triad traits, Big Five personality traits, and self-esteem. Confirmatory factor analysis and item response theory were used to validate the factor structure and estimate the reliability across the latent trait spectrum, respectively. Results suggest that the NARQ-S shows a robust factor structure and is a reliable and valid short measure of the agentic and antagonistic aspects of grandiose narcissism. We also discuss future directions and applications of the NARQ-S as a short and comprehensive measure of grandiose narcissism. (PsycINFO Database Record (c) 2018 APA, all rights reserved).

  13. A Chinese Mandarin translation and validation of the Myocardial Infarction Dimensional Assessment Scale (MIDAS).

    PubMed

    Wang, W; Lopez, V; Thompson, D R

    2006-09-01

    To evaluate the validity, reliability, and cultural relevance of the Chinese Mandarin version of Myocardial Infarction Dimensional Assessment Scale (MIDAS) as a disease-specific quality of life measure. The cultural relevance and content validity of the Chinese Mandarin version of the MIDAS (CM-MIDAS) was evaluated by an expert panel. Measurement performance was tested on 180 randomly selected Chinese MI patents. Thirty participants from the primary group completed the CM-MIDAS for test-retest reliability after 2 weeks. Reliability, validity and discriminatory power of the CM-MIDAS were calculated. Two items were modified as suggested by the expert panel. The overall CM-MIDAS had acceptable internal consistency with Cronbach's alpha coefficient 0.93 for the scale and 0.71-0.94 for the seven domains. Test-retest reliability by intraclass correlations was 0.85 for the overall scale and 0.74-0.94 for the seven domains. There was acceptable concurrent validity with significant (p < 0.05) correlations between the CM-MDAS and the Chinese Version of the Short Form 36. The principal components analysis extracted seven factors that explained 67.18% of the variance with high factor loading indicating good construct validity. Empirical data support CM-MIDAS as a valid and reliable disease-specific quality of life measure for Chinese Mandarin speaking patients with myocardial infarction.

  14. A Reliable and Valid Weighted Scoring Instrument for Use in Grading APA-Style Empirical Research Report

    ERIC Educational Resources Information Center

    Greenberg, Kathleen Puglisi

    2012-01-01

    The scoring instrument described in this article is based on a deconstruction of the seven sections of an American Psychological Association (APA)-style empirical research report into a set of learning outcomes divided into content-, expression-, and format-related categories. A double-weighting scheme used to score the report yields a final grade…

  15. An Evaluation of Empirical Bayes's Estimation of Value-Added Teacher Performance Measures

    ERIC Educational Resources Information Center

    Guarino, Cassandra M.; Maxfield, Michelle; Reckase, Mark D.; Thompson, Paul N.; Wooldridge, Jeffrey M.

    2015-01-01

    Empirical Bayes's (EB) estimation has become a popular procedure used to calculate teacher value added, often as a way to make imprecise estimates more reliable. In this article, we review the theory of EB estimation and use simulated and real student achievement data to study the ability of EB estimators to properly rank teachers. We compare the…

  16. Assessing the Reliability of Material Flow Analysis Results: The Cases of Rhenium, Gallium, and Germanium in the United States Economy.

    PubMed

    Meylan, Grégoire; Reck, Barbara K; Rechberger, Helmut; Graedel, Thomas E; Schwab, Oliver

    2017-10-17

    Decision-makers traditionally expect "hard facts" from scientific inquiry, an expectation that the results of material flow analyses (MFAs) can hardly meet. MFA limitations are attributable to incompleteness of flowcharts, limited data quality, and model assumptions. Moreover, MFA results are, for the most part, based less on empirical observation but rather on social knowledge construction processes. Developing, applying, and improving the means of evaluating and communicating the reliability of MFA results is imperative. We apply two recently proposed approaches for making quantitative statements on MFA reliability to national minor metals systems: rhenium, gallium, and germanium in the United States in 2012. We discuss the reliability of results in policy and management contexts. The first approach consists of assessing data quality based on systematic characterization of MFA data and the associated meta-information and quantifying the "information content" of MFAs. The second is a quantification of data inconsistencies indicated by the "degree of data reconciliation" between the data and the model. A high information content and a low degree of reconciliation indicate reliable or certain MFA results. This article contributes to reliability and uncertainty discourses in MFA, exemplifying the usefulness of the approaches in policy and management, and to raw material supply discussions by providing country-level information on three important minor metals often considered critical.

  17. Validation of the German prostate-specific module.

    PubMed

    Bestmann, Beate; Rohde, Volker; Siebmann, Jens-Ulrich; Galalae, Razvan; Weidner, Wolfgang; Küchler, Thomas

    2006-02-01

    Theoretically, all patients newly diagnosed with prostate cancer are faced with a choice of treatment options: radical prostatectomy or radio therapy. Although these different treatments may have no differences in terms of survival, they may have very different consequences on the subsequent quality of life (QoL). Prerequisite to analyze QoL is a reliable and valid instrument to assess these differences not only in terms of general QoL (EORTC QLQ-C30) but prostate specific symptoms with a prostate specific module as well. Therefore, the aim of this study was a psychometric evaluation (validation) of the prostate-specific module (PSM). Five historical cohort studies were put together for an empirical meta-analysis. The main objective was to analyze the module's psychometric properties. The total sample consisted of 1,185 patients, of whom 950 completed the QoL questionnaires (EORTC QLQ-C30 and a prostate specific module developed by Kuechler et al.). First step of analysis was a principal component analysis that revealed the following scales: urinary problems, incontinence, erectile dysfunction, sexual problems, problems with partner, pain, heat, nutrition, and psychic strain. The module showed good reliability and concurrent validity and very good construct validity, since the module is able to discriminate between different treatment regimes, tumor stages and age. The German PSM is a reliable, valid and applicable tool for QoL in patients with prostate cancer.

  18. Development and Validation of the Controller Acceptance Rating Scale (CARS): Results of Empirical Research

    NASA Technical Reports Server (NTRS)

    Lee, Katharine K.; Kerns, Karol; Bone, Randall

    2001-01-01

    The measurement of operational acceptability is important for the development, implementation, and evolution of air traffic management decision support tools. The Controller Acceptance Rating Scale was developed at NASA Ames Research Center for the development and evaluation of the Passive Final Approach Spacing Tool. CARS was modeled after a well-known pilot evaluation rating instrument, the Cooper-Harper Scale, and has since been used in the evaluation of the User Request Evaluation Tool, developed by MITRE's Center for Advanced Aviation System Development. In this paper, we provide a discussion of the development of CARS and an analysis of the empirical data collected with CARS to examine construct validity. Results of intraclass correlations indicated statistically significant reliability for the CARS. From the subjective workload data that were collected in conjunction with the CARS, it appears that the expected set of workload attributes was correlated with the CARS. As expected, the analysis also showed that CARS was a sensitive indicator of the impact of decision support tools on controller operations. Suggestions for future CARS development and its improvement are also provided.

  19. The revised NEUROGES-ELAN system: An objective and reliable interdisciplinary analysis tool for nonverbal behavior and gesture.

    PubMed

    Lausberg, Hedda; Sloetjes, Han

    2016-09-01

    As visual media spread to all domains of public and scientific life, nonverbal behavior is taking its place as an important form of communication alongside the written and spoken word. An objective and reliable method of analysis for hand movement behavior and gesture is therefore currently required in various scientific disciplines, including psychology, medicine, linguistics, anthropology, sociology, and computer science. However, no adequate common methodological standards have been developed thus far. Many behavioral gesture-coding systems lack objectivity and reliability, and automated methods that register specific movement parameters often fail to show validity with regard to psychological and social functions. To address these deficits, we have combined two methods, an elaborated behavioral coding system and an annotation tool for video and audio data. The NEUROGES-ELAN system is an effective and user-friendly research tool for the analysis of hand movement behavior, including gesture, self-touch, shifts, and actions. Since its first publication in 2009 in Behavior Research Methods, the tool has been used in interdisciplinary research projects to analyze a total of 467 individuals from different cultures, including subjects with mental disease and brain damage. Partly on the basis of new insights from these studies, the system has been revised methodologically and conceptually. The article presents the revised version of the system, including a detailed study of reliability. The improved reproducibility of the revised version makes NEUROGES-ELAN a suitable system for basic empirical research into the relation between hand movement behavior and gesture and cognitive, emotional, and interactive processes and for the development of automated movement behavior recognition methods.

  20. The German Version of the Herth Hope Index (HHI-D): Development and Psychometric Properties.

    PubMed

    Geiser, Franziska; Zajackowski, Katharina; Conrad, Rupert; Imbierowicz, Katrin; Wegener, Ingo; Herth, Kaye A; Urbach, Anne Sarah

    2015-01-01

    The importance of hope is evident in clinical oncological care. Hope is associated with psychological and also physical functioning. However, there is still a dearth of empirical research on hope as a multidimensional concept. The Herth Hope Index is a reliable and valid instrument for the measurement of hope and is available in many languages. Until now no authorized German translation has been published and validated. After translation, the questionnaire was completed by 192 patients with different tumor entities in radiation therapy. Reliability, concurrent validity, and factor structure of the questionnaire were determined. Correlations were high with depression and anxiety as well as optimism and pessimism. As expected, correlations with coping styles were moderate. Internal consistency and test-retest reliability were satisfactory. We could not replicate the original 3-factor model. Application of the scree plot criterion in an exploratory factor analysis resulted in a single-factor structure. The Herth Hope Index - German Version (HHI-D) is a short, reliable, and valid instrument for the assessment of hope in patient populations. We recommend using only the HHI-D total score until further research gives more insights into possible factorial solutions and subscales. © 2015 S. Karger GmbH, Freiburg.

  1. Development and preliminary validation of a questionnaire to measure satisfaction with home care in Greece: an exploratory factor analysis of polychoric correlations

    PubMed Central

    2010-01-01

    Background The primary aim of this study was to develop and psychometrically test a Greek-language instrument for measuring satisfaction with home care. The first empirical evidence about the level of satisfaction with these services in Greece is also provided. Methods The questionnaire resulted from literature search, on-site observation and cognitive interviews. It was applied in 2006 to a sample of 201 enrollees of five home care programs in the city of Thessaloniki and contains 31 items that measure satisfaction with individual service attributes and are expressed on a 5-point Likert scale. The latter has been usually considered in practice as an interval scale, although it is in principle ordinal. We thus treated the variable as an ordinal one, but also employed the traditional approach in order to compare the findings. Our analysis was therefore based on ordinal measures such as the polychoric correlation, Kendall's Tau b coefficient and ordinal Cronbach's alpha. Exploratory factor analysis was followed by an assessment of internal consistency reliability, test-retest reliability, construct validity and sensitivity. Results Analyses with ordinal and interval scale measures produced in essence very similar results and identified four multi-item scales. Three of these were found to be reliable and valid: socioeconomic change, staff skills and attitudes and service appropriateness. A fourth dimension -service planning- had lower internal consistency reliability and yet very satisfactory test-retest reliability, construct validity and floor and ceiling effects. The global satisfaction scale created was also quite reliable. Overall, participants were satisfied -yet not very satisfied- with home care services. More room for improvement seems to exist for the socio-economic and planning aspects of care and less for staff skills and attitudes and appropriateness of provided services. Conclusions The methods developed seem to be a promising tool for the measurement of home care satisfaction in Greece. PMID:20602759

  2. Directional variance adjustment: bias reduction in covariance matrices based on factor analysis with an application to portfolio optimization.

    PubMed

    Bartz, Daniel; Hatrick, Kerr; Hesse, Christian W; Müller, Klaus-Robert; Lemm, Steven

    2013-01-01

    Robust and reliable covariance estimates play a decisive role in financial and many other applications. An important class of estimators is based on factor models. Here, we show by extensive Monte Carlo simulations that covariance matrices derived from the statistical Factor Analysis model exhibit a systematic error, which is similar to the well-known systematic error of the spectrum of the sample covariance matrix. Moreover, we introduce the Directional Variance Adjustment (DVA) algorithm, which diminishes the systematic error. In a thorough empirical study for the US, European, and Hong Kong stock market we show that our proposed method leads to improved portfolio allocation.

  3. Directional Variance Adjustment: Bias Reduction in Covariance Matrices Based on Factor Analysis with an Application to Portfolio Optimization

    PubMed Central

    Bartz, Daniel; Hatrick, Kerr; Hesse, Christian W.; Müller, Klaus-Robert; Lemm, Steven

    2013-01-01

    Robust and reliable covariance estimates play a decisive role in financial and many other applications. An important class of estimators is based on factor models. Here, we show by extensive Monte Carlo simulations that covariance matrices derived from the statistical Factor Analysis model exhibit a systematic error, which is similar to the well-known systematic error of the spectrum of the sample covariance matrix. Moreover, we introduce the Directional Variance Adjustment (DVA) algorithm, which diminishes the systematic error. In a thorough empirical study for the US, European, and Hong Kong stock market we show that our proposed method leads to improved portfolio allocation. PMID:23844016

  4. Medicine is not science: guessing the future, predicting the past.

    PubMed

    Miller, Clifford

    2014-12-01

    Irregularity limits human ability to know, understand and predict. A better understanding of irregularity may improve the reliability of knowledge. Irregularity and its consequences for knowledge are considered. Reliable predictive empirical knowledge of the physical world has always been obtained by observation of regularities, without needing science or theory. Prediction from observational knowledge can remain reliable despite some theories based on it proving false. A naïve theory of irregularity is outlined. Reducing irregularity and/or increasing regularity can increase the reliability of knowledge. Beyond long experience and specialization, improvements include implementing supporting knowledge systems of libraries of appropriately classified prior cases and clinical histories and education about expertise, intuition and professional judgement. A consequence of irregularity and complexity is that classical reductionist science cannot provide reliable predictions of the behaviour of complex systems found in nature, including of the human body. Expertise, expert judgement and their exercise appear overarching. Diagnosis involves predicting the past will recur in the current patient applying expertise and intuition from knowledge and experience of previous cases and probabilistic medical theory. Treatment decisions are an educated guess about the future (prognosis). Benefits of the improvements suggested here are likely in fields where paucity of feedback for practitioners limits development of reliable expert diagnostic intuition. Further analysis, definition and classification of irregularity is appropriate. Observing and recording irregularities are initial steps in developing irregularity theory to improve the reliability and extent of knowledge, albeit some forms of irregularity present inherent difficulties. © 2014 John Wiley & Sons, Ltd.

  5. Exploring Equivalent Forms Reliability Using a Key Stage 2 Reading Test

    ERIC Educational Resources Information Center

    Benton, Tom

    2013-01-01

    This article outlines an empirical investigation into equivalent forms reliability using a case study of a national curriculum reading test. Within the situation being studied, there has been a genuine attempt to create several equivalent forms and so it is of interest to compare the actual behaviour of the relationship between these forms to the…

  6. On the Accuracy of Probabilistic Bucking Load Prediction

    NASA Technical Reports Server (NTRS)

    Arbocz, Johann; Starnes, James H.; Nemeth, Michael P.

    2001-01-01

    The buckling strength of thin-walled stiffened or unstiffened, metallic or composite shells is of major concern in aeronautical and space applications. The difficulty to predict the behavior of axially compressed thin-walled cylindrical shells continues to worry design engineers as we enter the third millennium. Thanks to extensive research programs in the late sixties and early seventies and the contributions of many eminent scientists, it is known that buckling strength calculations are affected by the uncertainties in the definition of the parameters of the problem such as definition of loads, material properties, geometric variables, edge support conditions, and the accuracy of the engineering models and analysis tools used in the design phase. The NASA design criteria monographs from the late sixties account for these design uncertainties by the use of a lump sum safety factor. This so-called 'empirical knockdown factor gamma' usually results in overly conservative design. Recently new reliability based probabilistic design procedure for buckling critical imperfect shells have been proposed. It essentially consists of a stochastic approach which introduces an improved 'scientific knockdown factor lambda(sub a)', that is not as conservative as the traditional empirical one. In order to incorporate probabilistic methods into a High Fidelity Analysis Approach one must be able to assess the accuracy of the various steps that must be executed to complete a reliability calculation. In the present paper the effect of size of the experimental input sample on the predicted value of the scientific knockdown factor lambda(sub a) calculated by the First-Order, Second-Moment Method is investigated.

  7. A study of the longevity and operational reliability of Goddard Spacecraft, 1960-1980

    NASA Technical Reports Server (NTRS)

    Shockey, E. F.

    1981-01-01

    Compiled data regarding the design lives and lifetimes actually achieved by 104 orbiting satellites launched by the Goddard Spaceflight Center between the years 1960 and 1980 is analyzed. Historical trends over the entire 21 year period are reviewed, and the more recent data is subjected to an examination of several key parameters. An empirical reliability function is derived, and compared with various mathematical models. Data from related studies is also discussed. The results provide insight into the reliability history of Goddard spacecraft an guidance for estimating the reliability of future programs.

  8. Home Healthcare Nurses' Job Satisfaction Scale: refinement and psychometric testing.

    PubMed

    Ellenbecker, Carol H; Byleckie, James J

    2005-10-01

    This paper describes a study to further develop and test the psychometric properties of the Home Healthcare Nurses' Job Satisfaction Scale, including reliability and construct and criterion validity. Numerous scales have been developed to measure nurses' job satisfaction. Only one, the Home Healthcare Nurses' Job Satisfaction Scale, has been designed specifically to measure job satisfaction of home healthcare nurses. The Home Healthcare Nurses' Job Satisfaction Scale is based on a theoretical model that integrates the findings of empirical research related to job satisfaction. A convenience sample of 340 home healthcare nurses completed the Home Healthcare Nurses' Job Satisfaction Scale and the Mueller and McCloskey Satisfaction Scale, which was used to test criterion validity. Factor analysis was used for testing and refinement of the theory-based assignment of items to constructs. Reliability was assessed by Cronbach's alpha internal consistency reliability coefficients. The data were collected in 2003. Nine factors contributing to home healthcare nurses' job satisfaction emerged from the factor analysis and were strongly supported by the underlying theory. Factor loadings were all above 0.4. Cronbach's alpha coefficients for each of the nine subscales ranged from 0.64 to 0.83; the alpha for the global scale was 0.89. The correlations between the Home Healthcare Nurses' Job Satisfaction Scale and Mueller and McCloskey Satisfaction Scale was 0.79, indicating good criterion-related validity. The Home Healthcare Nurses' Job Satisfaction Scale has potential as a reliable and valid scale for measurement of job satisfaction of home healthcare nurses.

  9. Human Factors in Financial Trading: An Analysis of Trading Incidents.

    PubMed

    Leaver, Meghan; Reader, Tom W

    2016-09-01

    This study tests the reliability of a system (FINANS) to collect and analyze incident reports in the financial trading domain and is guided by a human factors taxonomy used to describe error in the trading domain. Research indicates the utility of applying human factors theory to understand error in finance, yet empirical research is lacking. We report on the development of the first system for capturing and analyzing human factors-related issues in operational trading incidents. In the first study, 20 incidents are analyzed by an expert user group against a referent standard to establish the reliability of FINANS. In the second study, 750 incidents are analyzed using distribution, mean, pathway, and associative analysis to describe the data. Kappa scores indicate that categories within FINANS can be reliably used to identify and extract data on human factors-related problems underlying trading incidents. Approximately 1% of trades (n = 750) lead to an incident. Slip/lapse (61%), situation awareness (51%), and teamwork (40%) were found to be the most common problems underlying incidents. For the most serious incidents, problems in situation awareness and teamwork were most common. We show that (a) experts in the trading domain can reliably and accurately code human factors in incidents, (b) 1% of trades incur error, and (c) poor teamwork skills and situation awareness underpin the most critical incidents. This research provides data crucial for ameliorating risk within financial trading organizations, with implications for regulation and policy. © 2016, Human Factors and Ergonomics Society.

  10. Determination of a Limited Scope Network's Lightning Detection Efficiency

    NASA Technical Reports Server (NTRS)

    Rompala, John T.; Blakeslee, R.

    2008-01-01

    This paper outlines a modeling technique to map lightning detection efficiency variations over a region surveyed by a sparse array of ground based detectors. A reliable flash peak current distribution (PCD) for the region serves as the technique's base. This distribution is recast as an event probability distribution function. The technique then uses the PCD together with information regarding: site signal detection thresholds, type of solution algorithm used, and range attenuation; to formulate the probability that a flash at a specified location will yield a solution. Applying this technique to the full region produces detection efficiency contour maps specific to the parameters employed. These contours facilitate a comparative analysis of each parameter's effect on the network's detection efficiency. In an alternate application, this modeling technique gives an estimate of the number, strength, and distribution of events going undetected. This approach leads to a variety of event density contour maps. This application is also illustrated. The technique's base PCD can be empirical or analytical. A process for formulating an empirical PCD specific to the region and network being studied is presented. A new method for producing an analytical representation of the empirical PCD is also introduced.

  11. Human figure drawings in the evaluation of severe adolescent suicidal behavior.

    PubMed

    Zalsman, G; Netanel, R; Fischel, T; Freudenstein, O; Landau, E; Orbach, I; Weizman, A; Pfeffer, C R; Apter, A

    2000-08-01

    To evaluate the reliability of using certain indicators derived from human figure drawings to distinguish between suicidal and nonsuicidal adolescents. Ninety consecutive admissions to an adolescent inpatient unit were assessed. Thirty-nine patients were admitted because of suicidal behavior and 51 for other reasons. All subjects were given the Human Figure Drawing (HFD) test. HFD was evaluated according to the method of Pfeffer and Richman, and the degree of suicidal behavior was rated by the Child Suicide Potential Scale. The internal reliability was satisfactory. HFD indicators correlated significantly with quantitative measures of suicidal behavior; of these indicators specifically, overall impression of the evaluator enabled the prediction of suicidal behavior and the distinction between suicidal and nonsuicidal inpatients (p < .001). A group of graphic indicators derived from a discriminant analysis formed a function, which was able to identify 84.6% of the suicidal and 76.6% of the nonsuicidal adolescents correctly. Many of the items had a regressive quality. The HFD is an example of a simple projective test that may have empirical reliability. It may be useful for the assessment of severe suicidal behavior in adolescents.

  12. Software reliability through fault-avoidance and fault-tolerance

    NASA Technical Reports Server (NTRS)

    Vouk, Mladen A.; Mcallister, David F.

    1993-01-01

    Strategies and tools for the testing, risk assessment and risk control of dependable software-based systems were developed. Part of this project consists of studies to enable the transfer of technology to industry, for example the risk management techniques for safety-concious systems. Theoretical investigations of Boolean and Relational Operator (BRO) testing strategy were conducted for condition-based testing. The Basic Graph Generation and Analysis tool (BGG) was extended to fully incorporate several variants of the BRO metric. Single- and multi-phase risk, coverage and time-based models are being developed to provide additional theoretical and empirical basis for estimation of the reliability and availability of large, highly dependable software. A model for software process and risk management was developed. The use of cause-effect graphing for software specification and validation was investigated. Lastly, advanced software fault-tolerance models were studied to provide alternatives and improvements in situations where simple software fault-tolerance strategies break down.

  13. On the reliability of self-reported health: evidence from Albanian data.

    PubMed

    Vaillant, Nicolas; Wolff, François-Charles

    2012-06-01

    This paper investigates the reliability of self-assessed measures of health using panel data collected in Albania by the World Bank in 2002, 2003 and 2004 through the Living Standard Measurement Study project. As the survey includes questions on a self-assessed measure of health and on more objective health problems, both types of information are combined with a view to understanding how respondents change their answers to the self-reported measures over time. Estimates from random effects ordered Probit models show that differences in self-reported subjective health between individuals are much more marked than those over time, suggesting a strong state dependence in subjective health status. The empirical analysis also reveals respondent consistency, from both a subjective and an objective viewpoint. Self-reported health is much more influenced by permanent shocks than by more transitory illness or injury. Copyright © 2012 Ministry of Health, Saudi Arabia. Published by Elsevier Ltd. All rights reserved.

  14. Kinematics of mechanical and adhesional micromanipulation under a scanning electron microscope

    NASA Astrophysics Data System (ADS)

    Saito, Shigeki; Miyazaki, Hideki T.; Sato, Tomomasa; Takahashi, Kunio

    2002-11-01

    In this paper, the kinematics of mechanical and adhesional micromanipulation using a needle-shaped tool under a scanning electron microscope is analyzed. A mode diagram is derived to indicate the possible micro-object behavior for the specified operational conditions. Based on the diagram, a reasonable method for pick and place operation is proposed. The keys to successful analysis are to introduce adhesional and rolling-resistance factors into the kinematic system consisting of a sphere, a needle-shaped tool, and a substrate, and to consider the time dependence of these factors due to the electron-beam (EB) irradiation. Adhesional force and the lower limit of maximum rolling resistance are evaluated quantitatively in theoretical and experimental ways. This analysis shows that it is possible to control the fracture of either the tool-sphere or substrate-sphere interface of the system selectively by the tool-loading angle and that such a selective fracture of the interfaces enables reliable pick or place operation even under EB irradiation. Although the conventional micromanipulation was not repeatable because the technique was based on an empirically effective method, this analysis should provide us with a guideline to reliable micromanipulation.

  15. IMatter: validation of the NHS Scotland Employee Engagement Index.

    PubMed

    Snowden, Austyn; MacArthur, Ewan

    2014-11-08

    Employee engagement is a fundamental component of quality healthcare. In order to provide empirical data of engagement in NHS Scotland an Employee Engagement Index was co-constructed with staff. 'iMatter' consists of 25 Likert questions developed iteratively from the literature and a series of validation events with NHS Scotland staff. The aim of this study was to test the face, content and construct validity of iMatter. Cross sectional survey of NHS Scotland staff. In January 2013 iMatter was sent to 2300 staff across all disciplines in NHS Scotland. 1280 staff completed it. Demographic data were collected. Internal consistency of the scale was calculated. Construct validity consisted of concurrent application of factor analysis and Rasch analysis. Face and content validity were checked using 3 focus groups. The sample was representative of the NHSScotland population. iMatter showed very strong reliability (α = 0.958). Factor analysis revealed a four-factor structure consistent with the following interpretation: iMatter showed evidence of high reliability and validity. It is a popular measure of staff engagement in NHS Scotland. Implications for practice focus on the importance of coproduction in psychometric development.

  16. Barriers and Negative Nudges: Exploring Challenges in Food Journaling

    PubMed Central

    Cordeiro, Felicia; Epstein, Daniel A.; Thomaz, Edison; Bales, Elizabeth; Jagannathan, Arvind K.; Abowd, Gregory D.; Fogarty, James

    2016-01-01

    Although food journaling is understood to be both important and difficult, little work has empirically documented the specific challenges people experience with food journals. We identify key challenges in a qualitative study combining a survey of 141 current and lapsed food journalers with analysis of 5,526 posts in community forums for three mobile food journals. Analyzing themes in this data, we find and discuss barriers to reliable food entry, negative nudges caused by current techniques, and challenges with social features. Our results motivate research exploring a wider range of approaches to food journal design and technology. PMID:26894233

  17. Barriers and Negative Nudges: Exploring Challenges in Food Journaling.

    PubMed

    Cordeiro, Felicia; Epstein, Daniel A; Thomaz, Edison; Bales, Elizabeth; Jagannathan, Arvind K; Abowd, Gregory D; Fogarty, James

    2015-04-01

    Although food journaling is understood to be both important and difficult, little work has empirically documented the specific challenges people experience with food journals. We identify key challenges in a qualitative study combining a survey of 141 current and lapsed food journalers with analysis of 5,526 posts in community forums for three mobile food journals. Analyzing themes in this data, we find and discuss barriers to reliable food entry, negative nudges caused by current techniques, and challenges with social features. Our results motivate research exploring a wider range of approaches to food journal design and technology.

  18. Quantifying the process and outcomes of person-centered planning.

    PubMed

    Holburn, S; Jacobson, J W; Vietze, P M; Schwartz, A A; Sersen, E

    2000-09-01

    Although person-centered planning is a popular approach in the field of developmental disabilities, there has been little systematic assessment of its process and outcomes. To measure person-centered planning, we developed three instruments designed to assess its various aspects. We then constructed variables comprising both a Process and an Outcome Index using a combined rational-empirical method. Test-retest reliability and measures of internal consistency appeared adequate. Variable correlations and factor analysis were generally consistent with our conceptualization and resulting item and variable classifications. Practical implications for intervention integrity, program evaluation, and organizational performance are discussed.

  19. Patients' perceptions of service quality dimensions: an empirical examination of health care in New Zealand.

    PubMed

    Clemes, M D; Ozanne, L K; Laurensen, W L

    2001-01-01

    The 1984 liberalization of the New Zealand economy has resulted in a health care sector that has become very competitive (Zwier and Clarke, 1999). The private sector is now able to supply health care services and, as a result, a greater value is being placed on patient satisfaction (Zwier and Clarke, 1999). However, despite the increasing focus on customer satisfaction, research into health care patients' perceptions of the dimensions of service quality is scarce. This can be problematic, as quality of care is an essential issue in the strategic marketing of health care services (Turner and Pol, 1995). This study takes a step towards addressing this deficiency by identifying patients' perceptions of the dimensions of service quality in health care. The findings of this study are based on the empirical analysis of a sample of 389 respondents interviewed by telephone. The findings indicate that the service quality dimensions identified in this health care specific study differ in number and dimensional structure from the widely adopted service quality dimensions first identified by Parasuraman, Berry and Zeithaml (1988): reliability, responsiveness, assurance, empathy and tangibles. The service quality dimensions identified in this study were: reliability, tangibles, assurance, empathy, food, access, outcome, admission, discharge and responsiveness. In addition, health care patients perceive the service quality dimensions relating to the core product in health care delivery (for example, outcome and reliability) as more important than the service quality dimensions relating to the peripheral product in health care delivery (for example, food, access and tangibles). Finally, the results of this study suggest that patients with different geographic, demographic, and behavioristic characteristics have different needs and wants during health care delivery and therefore perceive different service quality dimensions as important.

  20. Development and Validation of the Five-by-Five Resilience Scale.

    PubMed

    DeSimone, Justin A; Harms, P D; Vanhove, Adam J; Herian, Mitchel N

    2017-09-01

    This article introduces a new measure of resilience and five related protective factors. The Five-by-Five Resilience Scale (5×5RS) is developed on the basis of theoretical and empirical considerations. Two samples ( N = 475 and N = 613) are used to assess the factor structure, reliability, convergent validity, and criterion-related validity of the 5×5RS. Confirmatory factor analysis supports a bifactor model. The 5×5RS demonstrates adequate internal consistency as evidenced by Cronbach's alpha and empirical reliability estimates. The 5×5RS correlates positively with the Connor-Davidson Resilience Scale (CD-RISC), a commonly used measure of resilience. The 5×5RS exhibits similar criterion-related validity to the CD-RISC as evidenced by positive correlations with satisfaction with life, meaning in life, and secure attachment style as well as negative correlations with rumination and anxious or avoidant attachment styles. 5×5RS scores are positively correlated with healthy behaviors such as exercise and negatively correlated with sleep difficulty and symptomology of anxiety and depression. The 5×5RS incrementally explains variance in some criteria above and beyond the CD-RISC. Item responses are modeled using the graded response model. Information estimates demonstrate the ability of the 5×5RS to assess individuals within at least one standard deviation of the mean on relevant latent traits.

  1. An instrument measuring prospective mathematics teacher self-regulated learning: validity and reliability

    NASA Astrophysics Data System (ADS)

    Nugroho, A. A.; Juniati, D.; Siswono, T. Y. E.

    2018-03-01

    Self Regulated Learning (SRL) is an individual's ability to achieve academic goals by controlling behavior, motivate yourself and use cognitive in learning, so it is important for a teacher especially teachers of mathematics related to the ability of management, design, implementation of learning and evaluation of learning outcomes. The purpose of the research is to develop an instrument to describe the SRL of a prospective mathematics teacher. Data were collected by (1) the study of the theory of SRL produced the indicator SRL used to design the questionnaire SRL; (2) analysis of the questionnaire SRL obtained from several References; and (3) development stage of the SRL questionnaire through validity test of content and empirical validation. The study involved 2 content experts in mathematics, 1 linguist, and 92 prospective mathematics teachers. The results of the research on content validity test based on Indonesian expert and 2 content experts indicate that the content can assess the indicator of the SRL and feasible to be used, in the test of legibility of two prospective mathematics teacher concluded that the instrument has a language that can be understood by the prospective teacher of mathematics and on empirical validation involving 92 prospective mathematics teacher generate data that of 65 statements there are 3 invalid statements. Reliability calculation shows high category that values 0,93. The conclusion is the SRL instrument developed for the prospective mathematics teacher.

  2. An Evaluation of Empirical Bayes' Estimation of Value- Added Teacher Performance Measures. Working Paper #31. Revised

    ERIC Educational Resources Information Center

    Guarino, Cassandra M.; Maxfield, Michelle; Reckase, Mark D.; Thompson, Paul; Wooldridge, Jeffrey M.

    2014-01-01

    Empirical Bayes' (EB) estimation is a widely used procedure to calculate teacher value-added. It is primarily viewed as a way to make imprecise estimates more reliable. In this paper we review the theory of EB estimation and use simulated data to study its ability to properly rank teachers. We compare the performance of EB estimators with that of…

  3. Visual modeling in an analysis of multidimensional data

    NASA Astrophysics Data System (ADS)

    Zakharova, A. A.; Vekhter, E. V.; Shklyar, A. V.; Pak, A. J.

    2018-01-01

    The article proposes an approach to solve visualization problems and the subsequent analysis of multidimensional data. Requirements to the properties of visual models, which were created to solve analysis problems, are described. As a perspective direction for the development of visual analysis tools for multidimensional and voluminous data, there was suggested an active use of factors of subjective perception and dynamic visualization. Practical results of solving the problem of multidimensional data analysis are shown using the example of a visual model of empirical data on the current state of studying processes of obtaining silicon carbide by an electric arc method. There are several results of solving this problem. At first, an idea of possibilities of determining the strategy for the development of the domain, secondly, the reliability of the published data on this subject, and changes in the areas of attention of researchers over time.

  4. Assessing performance of an Electronic Health Record (EHR) using Cognitive Task Analysis.

    PubMed

    Saitwal, Himali; Feng, Xuan; Walji, Muhammad; Patel, Vimla; Zhang, Jiajie

    2010-07-01

    Many Electronic Health Record (EHR) systems fail to provide user-friendly interfaces due to the lack of systematic consideration of human-centered computing issues. Such interfaces can be improved to provide easy to use, easy to learn, and error-resistant EHR systems to the users. To evaluate the usability of an EHR system and suggest areas of improvement in the user interface. The user interface of the AHLTA (Armed Forces Health Longitudinal Technology Application) was analyzed using the Cognitive Task Analysis (CTA) method called GOMS (Goals, Operators, Methods, and Selection rules) and an associated technique called KLM (Keystroke Level Model). The GOMS method was used to evaluate the AHLTA user interface by classifying each step of a given task into Mental (Internal) or Physical (External) operators. This analysis was performed by two analysts independently and the inter-rater reliability was computed to verify the reliability of the GOMS method. Further evaluation was performed using KLM to estimate the execution time required to perform the given task through application of its standard set of operators. The results are based on the analysis of 14 prototypical tasks performed by AHLTA users. The results show that on average a user needs to go through 106 steps to complete a task. To perform all 14 tasks, they would spend about 22 min (independent of system response time) for data entry, of which 11 min are spent on more effortful mental operators. The inter-rater reliability analysis performed for all 14 tasks was 0.8 (kappa), indicating good reliability of the method. This paper empirically reveals and identifies the following finding related to the performance of AHLTA: (1) large number of average total steps to complete common tasks, (2) high average execution time and (3) large percentage of mental operators. The user interface can be improved by reducing (a) the total number of steps and (b) the percentage of mental effort, required for the tasks. 2010 Elsevier Ireland Ltd. All rights reserved.

  5. Racism as a determinant of health: a protocol for conducting a systematic review and meta-analysis

    PubMed Central

    2013-01-01

    Background Racism is increasingly recognized as a key determinant of health. A growing body of epidemiological evidence shows strong associations between self-reported racism and poor health outcomes across diverse minority groups in developed countries. While the relationship between racism and health has received increasing attention over the last two decades, a comprehensive meta-analysis focused on the health effects of racism has yet to be conducted. The aim of this review protocol is to provide a structure from which to conduct a systematic review and meta-analysis of studies that assess the relationship between racism and health. Methods This research will consist of a systematic review and meta-analysis. Studies will be considered for review if they are empirical studies reporting quantitative data on the association between racism and health for adults and/or children of all ages from any racial/ethnic/cultural groups. Outcome measures will include general health and well-being, physical health, mental health, healthcare use and health behaviors. Scientific databases (for example, Medline) will be searched using a comprehensive search strategy and reference lists will be manually searched for relevant studies. In addition, use of online search engines (for example, Google Scholar), key websites, and personal contact with experts will also be undertaken. Screening of search results and extraction of data from included studies will be independently conducted by at least two authors, including assessment of inter-rater reliability. Studies included in the review will be appraised for quality using tools tailored to each study design. Summary statistics of study characteristics and findings will be compiled and findings synthesized in a narrative summary as well as a meta-analysis. Discussion This review aims to examine associations between reported racism and health outcomes. This comprehensive and systematic review and meta-analysis of empirical research will provide a rigorous and reliable evidence base for future research, policy and practice, including information on the extent of available evidence for a range of racial/ethnic minority groups PMID:24059279

  6. Wavelet analysis for wind fields estimation.

    PubMed

    Leite, Gladeston C; Ushizima, Daniela M; Medeiros, Fátima N S; de Lima, Gilson G

    2010-01-01

    Wind field analysis from synthetic aperture radar images allows the estimation of wind direction and speed based on image descriptors. In this paper, we propose a framework to automate wind direction retrieval based on wavelet decomposition associated with spectral processing. We extend existing undecimated wavelet transform approaches, by including à trous with B(3) spline scaling function, in addition to other wavelet bases as Gabor and Mexican-hat. The purpose is to extract more reliable directional information, when wind speed values range from 5 to 10 ms(-1). Using C-band empirical models, associated with the estimated directional information, we calculate local wind speed values and compare our results with QuikSCAT scatterometer data. The proposed approach has potential application in the evaluation of oil spills and wind farms.

  7. Compulsive sexual behavior inventory: a preliminary study of reliability and validity.

    PubMed

    Coleman, E; Miner, M; Ohlerking, F; Raymond, N

    2001-01-01

    This preliminary study was designed to develop empirically a scale of compulsive sexual behavior (CSB) and to test its reliability and validity in a sample of individuals with nonparaphilic CSB (N = 15), in a sample of pedophiles (N = 35) in treatment for sexual offending, and in a sample of normal controls (N = 42). Following a factor analysis and a varimax rotation, those items with factor loadings on the rotated factors of greater than .60 were retained. Three factors were identified, which appeared to measure control, abuse, and violence. Cronbach's alphas indicated that the subscales have good reliability. The 28-item scale was then tested for validity by a linear discriminant function analysis. The scale successfully discriminated the nonparaphilic CSB sample and the pedophiles from controls. Further analysis indicated that this scale is a valid measure of CSB in that there were significant differences between the three groups on the control subscale. Pedophiles scored significantly lower than the other two groups on the abuse subscale, with the other two groups not scoring significantly differently from one another. This indicated that pedophiles were more abusive than the nonparaphilic CSB individuals or the controls. Pedophiles scored significantly lower than controls on the violence subscale. Nonparaphilic individuals with compulsive sexual behavior scored slightly lower on the violence subscale, although not significantly different. As a preliminary study, there are several limitations to this study, which should be addressed, in further studies with larger sample sizes.

  8. Measuring strategies for learning regulation in medical education: scale reliability and dimensionality in a Swedish sample.

    PubMed

    Edelbring, Samuel

    2012-08-15

    The degree of learners' self-regulated learning and dependence on external regulation influence learning processes in higher education. These regulation strategies are commonly measured by questionnaires developed in other settings than in which they are being used, thereby requiring renewed validation. The aim of this study was to psychometrically evaluate the learning regulation strategy scales from the Inventory of Learning Styles with Swedish medical students (N = 206). The regulation scales were evaluated regarding their reliability, scale dimensionality and interrelations. The primary evaluation focused on dimensionality and was performed with Mokken scale analysis. To assist future scale refinement, additional item analysis, such as item-to-scale correlations, was performed. Scale scores in the Swedish sample displayed good reliability in relation to published results: Cronbach's alpha: 0.82, 0.72, and 0.65 for self-regulation, external regulation and lack of regulation scales respectively. The dimensionalities in scales were adequate for self-regulation and its subscales, whereas external regulation and lack of regulation displayed less unidimensionality. The established theoretical scales were largely replicated in the exploratory analysis. The item analysis identified two items that contributed little to their respective scales. The results indicate that these scales have an adequate capacity for detecting the three theoretically proposed learning regulation strategies in the medical education sample. Further construct validity should be sought by interpreting scale scores in relation to specific learning activities. Using established scales for measuring students' regulation strategies enables a broad empirical base for increasing knowledge on regulation strategies in relation to different disciplinary settings and contributes to theoretical development.

  9. Decision-making in healthcare: a practical application of partial least square path modelling to coverage of newborn screening programmes

    PubMed Central

    2012-01-01

    Background Decision-making in healthcare is complex. Research on coverage decision-making has focused on comparative studies for several countries, statistical analyses for single decision-makers, the decision outcome and appraisal criteria. Accounting for decision processes extends the complexity, as they are multidimensional and process elements need to be regarded as latent constructs (composites) that are not observed directly. The objective of this study was to present a practical application of partial least square path modelling (PLS-PM) to evaluate how it offers a method for empirical analysis of decision-making in healthcare. Methods Empirical approaches that applied PLS-PM to decision-making in healthcare were identified through a systematic literature search. PLS-PM was used as an estimation technique for a structural equation model that specified hypotheses between the components of decision processes and the reasonableness of decision-making in terms of medical, economic and other ethical criteria. The model was estimated for a sample of 55 coverage decisions on the extension of newborn screening programmes in Europe. Results were evaluated by standard reliability and validity measures for PLS-PM. Results After modification by dropping two indicators that showed poor measures in the measurement models’ quality assessment and were not meaningful for newborn screening, the structural equation model estimation produced plausible results. The presence of three influences was supported: the links between both stakeholder participation or transparency and the reasonableness of decision-making; and the effect of transparency on the degree of scientific rigour of assessment. Reliable and valid measurement models were obtained to describe the composites of ‘transparency’, ‘participation’, ‘scientific rigour’ and ‘reasonableness’. Conclusions The structural equation model was among the first applications of PLS-PM to coverage decision-making. It allowed testing of hypotheses in situations where there are links between several non-observable constructs. PLS-PM was compatible in accounting for the complexity of coverage decisions to obtain a more realistic perspective for empirical analysis. The model specification can be used for hypothesis testing by using larger sample sizes and for data in the full domain of health technologies. PMID:22856325

  10. Decision-making in healthcare: a practical application of partial least square path modelling to coverage of newborn screening programmes.

    PubMed

    Fischer, Katharina E

    2012-08-02

    Decision-making in healthcare is complex. Research on coverage decision-making has focused on comparative studies for several countries, statistical analyses for single decision-makers, the decision outcome and appraisal criteria. Accounting for decision processes extends the complexity, as they are multidimensional and process elements need to be regarded as latent constructs (composites) that are not observed directly. The objective of this study was to present a practical application of partial least square path modelling (PLS-PM) to evaluate how it offers a method for empirical analysis of decision-making in healthcare. Empirical approaches that applied PLS-PM to decision-making in healthcare were identified through a systematic literature search. PLS-PM was used as an estimation technique for a structural equation model that specified hypotheses between the components of decision processes and the reasonableness of decision-making in terms of medical, economic and other ethical criteria. The model was estimated for a sample of 55 coverage decisions on the extension of newborn screening programmes in Europe. Results were evaluated by standard reliability and validity measures for PLS-PM. After modification by dropping two indicators that showed poor measures in the measurement models' quality assessment and were not meaningful for newborn screening, the structural equation model estimation produced plausible results. The presence of three influences was supported: the links between both stakeholder participation or transparency and the reasonableness of decision-making; and the effect of transparency on the degree of scientific rigour of assessment. Reliable and valid measurement models were obtained to describe the composites of 'transparency', 'participation', 'scientific rigour' and 'reasonableness'. The structural equation model was among the first applications of PLS-PM to coverage decision-making. It allowed testing of hypotheses in situations where there are links between several non-observable constructs. PLS-PM was compatible in accounting for the complexity of coverage decisions to obtain a more realistic perspective for empirical analysis. The model specification can be used for hypothesis testing by using larger sample sizes and for data in the full domain of health technologies.

  11. What makes an accurate and reliable subject-specific finite element model? A case study of an elephant femur

    PubMed Central

    Panagiotopoulou, O.; Wilshin, S. D.; Rayfield, E. J.; Shefelbine, S. J.; Hutchinson, J. R.

    2012-01-01

    Finite element modelling is well entrenched in comparative vertebrate biomechanics as a tool to assess the mechanical design of skeletal structures and to better comprehend the complex interaction of their form–function relationships. But what makes a reliable subject-specific finite element model? To approach this question, we here present a set of convergence and sensitivity analyses and a validation study as an example, for finite element analysis (FEA) in general, of ways to ensure a reliable model. We detail how choices of element size, type and material properties in FEA influence the results of simulations. We also present an empirical model for estimating heterogeneous material properties throughout an elephant femur (but of broad applicability to FEA). We then use an ex vivo experimental validation test of a cadaveric femur to check our FEA results and find that the heterogeneous model matches the experimental results extremely well, and far better than the homogeneous model. We emphasize how considering heterogeneous material properties in FEA may be critical, so this should become standard practice in comparative FEA studies along with convergence analyses, consideration of element size, type and experimental validation. These steps may be required to obtain accurate models and derive reliable conclusions from them. PMID:21752810

  12. The Motivational Value Systems Questionnaire (MVSQ): Psychometric Analysis Using a Forced Choice Thurstonian IRT Model

    PubMed Central

    Merk, Josef; Schlotz, Wolff; Falter, Thomas

    2017-01-01

    This study presents a new measure of value systems, the Motivational Value Systems Questionnaire (MVSQ), which is based on a theory of value systems by psychologist Clare W. Graves. The purpose of the instrument is to help people identify their personal hierarchies of value systems and thus become more aware of what motivates and demotivates them in work-related contexts. The MVSQ is a forced-choice (FC) measure, making it quicker to complete and more difficult to intentionally distort, but also more difficult to assess its psychometric properties due to ipsativity of FC data compared to rating scales. To overcome limitations of ipsative data, a Thurstonian IRT (TIRT) model was fitted to the questionnaire data, based on a broad sample of N = 1,217 professionals and students. Comparison of normative (IRT) scale scores and ipsative scores suggested that MVSQ IRT scores are largely freed from restrictions due to ipsativity and thus allow interindividual comparison of scale scores. Empirical reliability was estimated using a sample-based simulation approach which showed acceptable and good estimates and, on average, slightly higher test-retest reliabilities. Further, validation studies provided evidence on both construct validity and criterion-related validity. Scale score correlations and associations of scores with both age and gender were largely in line with theoretically- and empirically-based expectations, and results of a multitrait-multimethod analysis supports convergent and discriminant construct validity. Criterion validity was assessed by examining the relation of value system preferences to departmental affiliation which revealed significant relations in line with prior hypothesizing. These findings demonstrate the good psychometric properties of the MVSQ and support its application in the assessment of value systems in work-related contexts. PMID:28979228

  13. The Motivational Value Systems Questionnaire (MVSQ): Psychometric Analysis Using a Forced Choice Thurstonian IRT Model.

    PubMed

    Merk, Josef; Schlotz, Wolff; Falter, Thomas

    2017-01-01

    This study presents a new measure of value systems, the Motivational Value Systems Questionnaire (MVSQ), which is based on a theory of value systems by psychologist Clare W. Graves. The purpose of the instrument is to help people identify their personal hierarchies of value systems and thus become more aware of what motivates and demotivates them in work-related contexts. The MVSQ is a forced-choice (FC) measure, making it quicker to complete and more difficult to intentionally distort, but also more difficult to assess its psychometric properties due to ipsativity of FC data compared to rating scales. To overcome limitations of ipsative data, a Thurstonian IRT (TIRT) model was fitted to the questionnaire data, based on a broad sample of N = 1,217 professionals and students. Comparison of normative (IRT) scale scores and ipsative scores suggested that MVSQ IRT scores are largely freed from restrictions due to ipsativity and thus allow interindividual comparison of scale scores. Empirical reliability was estimated using a sample-based simulation approach which showed acceptable and good estimates and, on average, slightly higher test-retest reliabilities. Further, validation studies provided evidence on both construct validity and criterion-related validity. Scale score correlations and associations of scores with both age and gender were largely in line with theoretically- and empirically-based expectations, and results of a multitrait-multimethod analysis supports convergent and discriminant construct validity. Criterion validity was assessed by examining the relation of value system preferences to departmental affiliation which revealed significant relations in line with prior hypothesizing. These findings demonstrate the good psychometric properties of the MVSQ and support its application in the assessment of value systems in work-related contexts.

  14. The Implementation Leadership Scale (ILS): development of a brief measure of unit level implementation leadership.

    PubMed

    Aarons, Gregory A; Ehrhart, Mark G; Farahnak, Lauren R

    2014-04-14

    In healthcare and allied healthcare settings, leadership that supports effective implementation of evidenced-based practices (EBPs) is a critical concern. However, there are no empirically validated measures to assess implementation leadership. This paper describes the development, factor structure, and initial reliability and convergent and discriminant validity of a very brief measure of implementation leadership: the Implementation Leadership Scale (ILS). Participants were 459 mental health clinicians working in 93 different outpatient mental health programs in Southern California, USA. Initial item development was supported as part of a two United States National Institutes of Health (NIH) studies focused on developing implementation leadership training and implementation measure development. Clinician work group/team-level data were randomly assigned to be utilized for an exploratory factor analysis (n = 229; k = 46 teams) or for a confirmatory factor analysis (n = 230; k = 47 teams). The confirmatory factor analysis controlled for the multilevel, nested data structure. Reliability and validity analyses were then conducted with the full sample. The exploratory factor analysis resulted in a 12-item scale with four subscales representing proactive leadership, knowledgeable leadership, supportive leadership, and perseverant leadership. Confirmatory factor analysis supported an a priori higher order factor structure with subscales contributing to a single higher order implementation leadership factor. The scale demonstrated excellent internal consistency reliability as well as convergent and discriminant validity. The ILS is a brief and efficient measure of unit level leadership for EBP implementation. The availability of the ILS will allow researchers to assess strategic leadership for implementation in order to advance understanding of leadership as a predictor of organizational context for implementation. The ILS also holds promise as a tool for leader and organizational development to improve EBP implementation.

  15. The implementation leadership scale (ILS): development of a brief measure of unit level implementation leadership

    PubMed Central

    2014-01-01

    Background In healthcare and allied healthcare settings, leadership that supports effective implementation of evidenced-based practices (EBPs) is a critical concern. However, there are no empirically validated measures to assess implementation leadership. This paper describes the development, factor structure, and initial reliability and convergent and discriminant validity of a very brief measure of implementation leadership: the Implementation Leadership Scale (ILS). Methods Participants were 459 mental health clinicians working in 93 different outpatient mental health programs in Southern California, USA. Initial item development was supported as part of a two United States National Institutes of Health (NIH) studies focused on developing implementation leadership training and implementation measure development. Clinician work group/team-level data were randomly assigned to be utilized for an exploratory factor analysis (n = 229; k = 46 teams) or for a confirmatory factor analysis (n = 230; k = 47 teams). The confirmatory factor analysis controlled for the multilevel, nested data structure. Reliability and validity analyses were then conducted with the full sample. Results The exploratory factor analysis resulted in a 12-item scale with four subscales representing proactive leadership, knowledgeable leadership, supportive leadership, and perseverant leadership. Confirmatory factor analysis supported an a priori higher order factor structure with subscales contributing to a single higher order implementation leadership factor. The scale demonstrated excellent internal consistency reliability as well as convergent and discriminant validity. Conclusions The ILS is a brief and efficient measure of unit level leadership for EBP implementation. The availability of the ILS will allow researchers to assess strategic leadership for implementation in order to advance understanding of leadership as a predictor of organizational context for implementation. The ILS also holds promise as a tool for leader and organizational development to improve EBP implementation. PMID:24731295

  16. Fluorescence Intrinsic Characterization of Excitation-Emission Matrix Using Multi-Dimensional Ensemble Empirical Mode Decomposition

    PubMed Central

    Chang, Chi-Ying; Chang, Chia-Chi; Hsiao, Tzu-Chien

    2013-01-01

    Excitation-emission matrix (EEM) fluorescence spectroscopy is a noninvasive method for tissue diagnosis and has become important in clinical use. However, the intrinsic characterization of EEM fluorescence remains unclear. Photobleaching and the complexity of the chemical compounds make it difficult to distinguish individual compounds due to overlapping features. Conventional studies use principal component analysis (PCA) for EEM fluorescence analysis, and the relationship between the EEM features extracted by PCA and diseases has been examined. The spectral features of different tissue constituents are not fully separable or clearly defined. Recently, a non-stationary method called multi-dimensional ensemble empirical mode decomposition (MEEMD) was introduced; this method can extract the intrinsic oscillations on multiple spatial scales without loss of information. The aim of this study was to propose a fluorescence spectroscopy system for EEM measurements and to describe a method for extracting the intrinsic characteristics of EEM by MEEMD. The results indicate that, although PCA provides the principal factor for the spectral features associated with chemical compounds, MEEMD can provide additional intrinsic features with more reliable mapping of the chemical compounds. MEEMD has the potential to extract intrinsic fluorescence features and improve the detection of biochemical changes. PMID:24240806

  17. When Assessing Intra-Familial Relationships, Are Sociologists, Psychoanalysts and Psychiatrists Really Considering Different Constructs? An Empirical Study.

    PubMed

    Falissard, Bruno; Barry, Caroline; Hassler, Christine; Letrait, Muriel; Macher, Guillaume; Marty, François; Ramos, Elsa; Revah-Lévy, Anne; Robert, Philippe; de Singly, François

    2015-01-01

    This paper aimed to look for the existence of a common core when envisaging intra-familial interactions as perceived by adolescents, which could be shared by sociology, psychoanalysis and child and adolescent psychiatry. An empirical study based on a mixed-method design collected the responses of 194 adolescents to the instruction "In the next half hour, would you please write as freely as you wish about your relationships in your family, explaining how things are". All answers were then analyzed and 18 dimensions related to 3 different theoretical frameworks were rated blind using numerical scores by two independent raters from each discipline. Inter-rater reliability was good. A parallel analysis evidenced a strong underlying factor explaining a large amount of variance (>50%). This factor is bipolar, it reflects the level of positivity/negativity in the adolescent's point of view concerning his/her intra-familial relationships. A second factor can marginally be considered (10% of the variance). The 2-factor analysis found one factor related to positive feelings and the other to negative feelings. This finding of unidimensionality supports family study as an intervention science.

  18. Aircraft directional stability and vertical tail design: A review of semi-empirical methods

    NASA Astrophysics Data System (ADS)

    Ciliberti, Danilo; Della Vecchia, Pierluigi; Nicolosi, Fabrizio; De Marco, Agostino

    2017-11-01

    Aircraft directional stability and control are related to vertical tail design. The safety, performance, and flight qualities of an aircraft also depend on a correct empennage sizing. Specifically, the vertical tail is responsible for the aircraft yaw stability and control. If these characteristics are not well balanced, the entire aircraft design may fail. Stability and control are often evaluated, especially in the preliminary design phase, with semi-empirical methods, which are based on the results of experimental investigations performed in the past decades, and occasionally are merged with data provided by theoretical assumptions. This paper reviews the standard semi-empirical methods usually applied in the estimation of airplane directional stability derivatives in preliminary design, highlighting the advantages and drawbacks of these approaches that were developed from wind tunnel tests performed mainly on fighter airplane configurations of the first decades of the past century, and discussing their applicability on current transport aircraft configurations. Recent investigations made by the authors have shown the limit of these methods, proving the existence of aerodynamic interference effects in sideslip conditions which are not adequately considered in classical formulations. The article continues with a concise review of the numerical methods for aerodynamics and their applicability in aircraft design, highlighting how Reynolds-Averaged Navier-Stokes (RANS) solvers are well-suited to attain reliable results in attached flow conditions, with reasonable computational times. From the results of RANS simulations on a modular model of a representative regional turboprop airplane layout, the authors have developed a modern method to evaluate the vertical tail and fuselage contributions to aircraft directional stability. The investigation on the modular model has permitted an effective analysis of the aerodynamic interference effects by moving, changing, and expanding the available airplane components. Wind tunnel tests over a wide range of airplane configurations have been used to validate the numerical approach. The comparison between the proposed method and the standard semi-empirical methods available in literature proves the reliability of the innovative approach, according to the available experimental data collected in the wind tunnel test campaign.

  19. [Construction of competency model of 'excellent doctor' in Chinese medicine].

    PubMed

    Jin, Aning; Tian, Yongquan; Zhao, Taiyang

    2014-05-01

    To evaluate outstanding and ordinary persons from personal characteristics using competency as the important criteria, which is the future direction of medical education reform. We carried on a behavior event interview about famous doctors of old traditional Chinese medicine, compiled competency dictionary, proceed control prediction test. SPSS and AMOS were used to be data analysis tools on statistics. We adopted the model of peer assessment and contrast to carry out empirical research. This project has carried on exploratory factor analysis and confirmatory factor analysis, established a "5A" competency model which include moral ability, thinking ability, communication ability, learning and practical ability. Competency model of "excellent doctor" in Chinese medicine has been validated, with good reliability and validity, and embodies the characteristics of traditional Chinese medicine personnel training, with theoretical and practical significance for excellence in medicine physician training.

  20. Challenges and Advances in Validating Enzyme Design Proposals: The Case of the Kemp Eliminase Catalysis†

    PubMed Central

    Frushicheva, Maria P.; Cao, Jie; Warshel, Arieh

    2011-01-01

    One of the fundamental challenges in biotechnology and biochemistry is the ability to design effective enzymes. Despite recent progress, most of the advances on this front have been made by placing the reacting fragments in the proper places, rather than by optimizing the preorganization of the environment, which is the key factor in enzyme catalysis. Thus, rational improvement of the preorganization would require approaches capable of evaluating reliably the actual catalytic effect. This work considers the catalytic effects in different Kemp eliminases as a benchmark for a computer aided enzyme design. It is shown that the empirical valence bond provides a powerful screening tool, with significant advantage over current alternative strategies. The insights provided by the empirical valence bond calculations are discussed emphasizing the ability to analyze the difference between the linear free energy relationships obtained in solution to those found in the enzymes. We also point out the trade off between reliability and speed of the calculations and try to determine what it takes to obtain reliable computer aided screening. PMID:21443179

  1. Challenges and advances in validating enzyme design proposals: the case of kemp eliminase catalysis.

    PubMed

    Frushicheva, Maria P; Cao, Jie; Warshel, Arieh

    2011-05-10

    One of the fundamental challenges in biotechnology and biochemistry is the ability to design effective enzymes. Despite recent progress, most of the advances on this front have been made by placing the reacting fragments in the proper places, rather than by optimizing the preorganization of the environment, which is the key factor in enzyme catalysis. Thus, rational improvement of the preorganization would require approaches capable of evaluating reliably the actual catalytic effect. This work considers the catalytic effects in different Kemp eliminases as a benchmark for a computer-aided enzyme design. It is shown that the empirical valence bond provides a powerful screening tool, with significant advantages over current alternative strategies. The insights provided by the empirical valence bond calculations are discussed with an emphasis on the ability to analyze the difference between the linear free energy relationships obtained in solution and those found in the enzymes. We also point out the trade-off between the reliability and speed of the calculations and try to determine what it takes to realize reliable computer-aided screening.

  2. Public Reception of Climate Science: Coherence, Reliability, and Independence.

    PubMed

    Hahn, Ulrike; Harris, Adam J L; Corner, Adam

    2016-01-01

    Possible measures to mitigate climate change require global collective actions whose impacts will be felt by many, if not all. Implementing such actions requires successful communication of the reasons for them, and hence the underlying climate science, to a degree that far exceeds typical scientific issues which do not require large-scale societal response. Empirical studies have identified factors, such as the perceived level of consensus in scientific opinion and the perceived reliability of scientists, that can limit people's trust in science communicators and their subsequent acceptance of climate change claims. Little consideration has been given, however, to recent formal results within philosophy concerning the relationship between truth, the reliability of evidence sources, the coherence of multiple pieces of evidence/testimonies, and the impact of (non-)independence between sources of evidence. This study draws on these results to evaluate exactly what has (and, more important, has not yet) been established in the empirical literature about the factors that bias the public's reception of scientific communications about climate change. Copyright © 2015 Cognitive Science Society, Inc.

  3. LD-SPatt: large deviations statistics for patterns on Markov chains.

    PubMed

    Nuel, G

    2004-01-01

    Statistics on Markov chains are widely used for the study of patterns in biological sequences. Statistics on these models can be done through several approaches. Central limit theorem (CLT) producing Gaussian approximations are one of the most popular ones. Unfortunately, in order to find a pattern of interest, these methods have to deal with tail distribution events where CLT is especially bad. In this paper, we propose a new approach based on the large deviations theory to assess pattern statistics. We first recall theoretical results for empiric mean (level 1) as well as empiric distribution (level 2) large deviations on Markov chains. Then, we present the applications of these results focusing on numerical issues. LD-SPatt is the name of GPL software implementing these algorithms. We compare this approach to several existing ones in terms of complexity and reliability and show that the large deviations are more reliable than the Gaussian approximations in absolute values as well as in terms of ranking and are at least as reliable as compound Poisson approximations. We then finally discuss some further possible improvements and applications of this new method.

  4. Empirical evaluation of the Process Overview Measure for assessing situation awareness in process plants.

    PubMed

    Lau, Nathan; Jamieson, Greg A; Skraaning, Gyrd

    2016-03-01

    The Process Overview Measure is a query-based measure developed to assess operator situation awareness (SA) from monitoring process plants. A companion paper describes how the measure has been developed according to process plant properties and operator cognitive work. The Process Overview Measure demonstrated practicality, sensitivity, validity and reliability in two full-scope simulator experiments investigating dramatically different operational concepts. Practicality was assessed based on qualitative feedback of participants and researchers. The Process Overview Measure demonstrated sensitivity and validity by revealing significant effects of experimental manipulations that corroborated with other empirical results. The measure also demonstrated adequate inter-rater reliability and practicality for measuring SA in full-scope simulator settings based on data collected on process experts. Thus, full-scope simulator studies can employ the Process Overview Measure to reveal the impact of new control room technology and operational concepts on monitoring process plants. Practitioner Summary: The Process Overview Measure is a query-based measure that demonstrated practicality, sensitivity, validity and reliability for assessing operator situation awareness (SA) from monitoring process plants in representative settings.

  5. Validity and Reliability of the "Behavior Problems Inventory," the "Aberrant Behavior Checklist," and the "Repetitive Behavior Scale--Revised" among Infants and Toddlers at Risk for Intellectual or Developmental Disabilities: A Multi-Method Assessment Approach

    ERIC Educational Resources Information Center

    Rojahn, Johannes; Schroeder, Stephen R.; Mayo-Ortega, Liliana; Oyama-Ganiko, Rosao; LeBlanc, Judith; Marquis, Janet; Berke, Elizabeth

    2013-01-01

    Reliable and valid assessment of aberrant behaviors is essential in empirically verifying prevention and intervention for individuals with intellectual or developmental disabilities (IDD). Few instruments exist which assess behavior problems in infants. The current longitudinal study examined the performance of three behavior-rating scales for…

  6. Detecting long-term growth trends using tree rings: a critical evaluation of methods.

    PubMed

    Peters, Richard L; Groenendijk, Peter; Vlam, Mart; Zuidema, Pieter A

    2015-05-01

    Tree-ring analysis is often used to assess long-term trends in tree growth. A variety of growth-trend detection methods (GDMs) exist to disentangle age/size trends in growth from long-term growth changes. However, these detrending methods strongly differ in approach, with possible implications for their output. Here, we critically evaluate the consistency, sensitivity, reliability and accuracy of four most widely used GDMs: conservative detrending (CD) applies mathematical functions to correct for decreasing ring widths with age; basal area correction (BAC) transforms diameter into basal area growth; regional curve standardization (RCS) detrends individual tree-ring series using average age/size trends; and size class isolation (SCI) calculates growth trends within separate size classes. First, we evaluated whether these GDMs produce consistent results applied to an empirical tree-ring data set of Melia azedarach, a tropical tree species from Thailand. Three GDMs yielded similar results - a growth decline over time - but the widely used CD method did not detect any change. Second, we assessed the sensitivity (probability of correct growth-trend detection), reliability (100% minus probability of detecting false trends) and accuracy (whether the strength of imposed trends is correctly detected) of these GDMs, by applying them to simulated growth trajectories with different imposed trends: no trend, strong trends (-6% and +6% change per decade) and weak trends (-2%, +2%). All methods except CD, showed high sensitivity, reliability and accuracy to detect strong imposed trends. However, these were considerably lower in the weak or no-trend scenarios. BAC showed good sensitivity and accuracy, but low reliability, indicating uncertainty of trend detection using this method. Our study reveals that the choice of GDM influences results of growth-trend studies. We recommend applying multiple methods when analysing trends and encourage performing sensitivity and reliability analysis. Finally, we recommend SCI and RCS, as these methods showed highest reliability to detect long-term growth trends. © 2014 John Wiley & Sons Ltd.

  7. Neuroscience and Ethics.

    PubMed

    Liao, S Matthew

    2017-03-01

    A number of people believe that results from neuroscience have the potential to settle seemingly intractable debates concerning the nature, practice, and reliability of moral judgments. In particular, Joshua Greene has argued that evidence from neuroscience can be used to advance the long-standing debate between consequentialism and deontology. This paper first argues that charitably interpreted, Greene's neuroscientific evidence can contribute to substantive ethical discussions by being part of an epistemic debunking argument. It then argues that taken as an epistemic debunking argument, Greene's argument falls short in undermining deontological judgments. Lastly, it proposes that accepting Greene's methodology at face value, neuroimaging results may in fact call into question the reliability of consequentialist judgments. The upshot is that Greene's empirical results do not undermine deontology and that Greene's project points toward a way by which empirical evidence such as neuroscientific evidence can play a role in normative debates.

  8. Cross-cultural adaptation and validation of the Chinese Comfort, Afford, Respect, and Expect scale of caring nurse-patient interaction competence.

    PubMed

    Chung, Hui-Chun; Hsieh, Tsung-Cheng; Chen, Yueh-Chih; Chang, Shu-Chuan; Hsu, Wen-Lin

    2017-11-29

    To investigate the construct validity and reliability of the Chinese Comfort, Afford, Respect, and Expect scale, which can be used to determine clinical nurses' competence. The results can also serve to promote nursing competence and improve patient satisfaction. Nurse-patient interaction is critical for improving nursing care quality. However, to date, no relevant validated instrument has been proposed for assessing caring nurse-patient interaction competence in clinical practice. This study adapted and validated the Chinese version of the caring nurse-patient interaction scale. A cross-cultural adaptation and validation study. A psychometric analysis of the four major constructs of the Chinese Comfort, Afford, Respect, and Expect scale was conducted on a sample of 356 nurses from a medical centre in China. Item analysis and exploratory factor analysis were adopted to extract the main components, both the internal consistency and correlation coefficients were used to examine reliability and a confirmatory factor analysis was adopted to verify the construct validity. The goodness-of-fit results of the model were strong. The standardised factor loadings of the Chinese Comfort, Afford, Respect, and Expect scale ranged from 0.73-0.95, indicating that the validity and reliability of this instrument were favourable. Moreover, the 12 extracted items explained 95.9% of the measured content of the Chinese Comfort, Afford, Respect, and Expect scale. The results serve as empirical evidence regarding the validity and reliability of the Chinese Comfort, Afford, Respect, and Expect scale. Hospital nurses increasingly demand help from patients and their family members in identifying health problems and assisting with medical decision-making. Therefore, enhancing nurses' competence in nurse-patient interactions is crucial for nursing and hospital managers to improve nursing care quality. The Chinese caring nurse-patient interaction scale can serve as an effective tool for nursing and hospital managers to evaluate the caring nurse-patient interaction confidence of nurses and improve inpatient satisfaction and quality of care. © 2017 John Wiley & Sons Ltd.

  9. Test Reliability at the Individual Level

    PubMed Central

    Hu, Yueqin; Nesselroade, John R.; Erbacher, Monica K.; Boker, Steven M.; Burt, S. Alexandra; Keel, Pamela K.; Neale, Michael C.; Sisk, Cheryl L.; Klump, Kelly

    2016-01-01

    Reliability has a long history as one of the key psychometric properties of a test. However, a given test might not measure people equally reliably. Test scores from some individuals may have considerably greater error than others. This study proposed two approaches using intraindividual variation to estimate test reliability for each person. A simulation study suggested that the parallel tests approach and the structural equation modeling approach recovered the simulated reliability coefficients. Then in an empirical study, where forty-five females were measured daily on the Positive and Negative Affect Schedule (PANAS) for 45 consecutive days, separate estimates of reliability were generated for each person. Results showed that reliability estimates of the PANAS varied substantially from person to person. The methods provided in this article apply to tests measuring changeable attributes and require repeated measures across time on each individual. This article also provides a set of parallel forms of PANAS. PMID:28936107

  10. Benchmarking DFT and semi-empirical methods for a reliable and cost-efficient computational screening of benzofulvene derivatives as donor materials for small-molecule organic solar cells.

    PubMed

    Tortorella, Sara; Talamo, Maurizio Mastropasqua; Cardone, Antonio; Pastore, Mariachiara; De Angelis, Filippo

    2016-02-24

    A systematic computational investigation on the optical properties of a group of novel benzofulvene derivatives (Martinelli 2014 Org. Lett. 16 3424-7), proposed as possible donor materials in small molecule organic photovoltaic (smOPV) devices, is presented. A benchmark evaluation against experimental results on the accuracy of different exchange and correlation functionals and semi-empirical methods in predicting both reliable ground state equilibrium geometries and electronic absorption spectra is carried out. The benchmark of the geometry optimization level indicated that the best agreement with x-ray data is achieved by using the B3LYP functional. Concerning the optical gap prediction, we found that, among the employed functionals, MPW1K provides the most accurate excitation energies over the entire set of benzofulvenes. Similarly reliable results were also obtained for range-separated hybrid functionals (CAM-B3LYP and wB97XD) and for global hybrid methods incorporating a large amount of non-local exchange (M06-2X and M06-HF). Density functional theory (DFT) hybrids with a moderate (about 20-30%) extent of Hartree-Fock exchange (HFexc) (PBE0, B3LYP and M06) were also found to deliver HOMO-LUMO energy gaps which compare well with the experimental absorption maxima, thus representing a valuable alternative for a prompt and predictive estimation of the optical gap. The possibility of using completely semi-empirical approaches (AM1/ZINDO) is also discussed.

  11. What can Written-Words Tell us About Lexical Retrieval in Speech Production?

    PubMed Central

    Navarrete, Eduardo; Mahon, Bradford Z.; Lorenzoni, Anna; Peressotti, Francesca

    2016-01-01

    In recent decades, researchers have exploited semantic context effects in picture naming tasks in order to investigate the mechanisms involved in the retrieval of words from the mental lexicon. In the blocked naming paradigm, participants name target pictures that are either blocked or not blocked by semantic category. In the continuous naming task, participants name a sequence of target pictures that are drawn from multiple semantic categories. Semantic context effects in both tasks are a highly reliable phenomenon. The empirical evidence is, however, sparse and inconsistent when the target stimuli are printed-words instead of pictures. In the first part of the present study we review the empirical evidence regarding semantic context effects with written-word stimuli in the blocked and continuous naming tasks. In the second part, we empirically test whether semantic context effects are transferred from picture naming trials to word reading trials, and from word reading trials to picture naming trials. The results indicate a transfer of semantic context effects from picture naming to subsequently read within-category words. There is no transfer of semantic effects from target words that were read to subsequently named within-category pictures. These results replicate previous findings (Navarrete et al., 2010) and are contrary to predictions from a recent theoretical analysis by Belke (2013). The empirical evidence reported in the literature together with the present results, are discussed in relation to current accounts of semantic context effects in speech production. PMID:26779090

  12. Certified dual-corrected radiation patterns of phased antenna arrays by offline–online order reduction of finite-element models

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Sommer, A., E-mail: a.sommer@lte.uni-saarland.de; Farle, O., E-mail: o.farle@lte.uni-saarland.de; Dyczij-Edlinger, R., E-mail: edlinger@lte.uni-saarland.de

    2015-10-15

    This paper presents a fast numerical method for computing certified far-field patterns of phased antenna arrays over broad frequency bands as well as wide ranges of steering and look angles. The proposed scheme combines finite-element analysis, dual-corrected model-order reduction, and empirical interpolation. To assure the reliability of the results, improved a posteriori error bounds for the radiated power and directive gain are derived. Both the reduced-order model and the error-bounds algorithm feature offline–online decomposition. A real-world example is provided to demonstrate the efficiency and accuracy of the suggested approach.

  13. Empirically defining rapid response to intensive treatment to maximize prognostic utility for bulimia nervosa and purging disorder.

    PubMed

    MacDonald, Danielle E; Trottier, Kathryn; McFarlane, Traci; Olmsted, Marion P

    2015-05-01

    Rapid response (RR) to eating disorder treatment has been reliably identified as a predictor of post-treatment and sustained remission, but its definition has varied widely. Although signal detection methods have been used to empirically define RR thresholds in outpatient settings, RR to intensive treatment has not been investigated. This study investigated the optimal definition of RR to day hospital treatment for bulimia nervosa and purging disorder. Participants were 158 patients who completed ≥6 weeks of day hospital treatment. Receiver operating characteristic (ROC) analysis was used to create four definitions of RR that could differentiate between remission and nonremission at the end of treatment. Definitions were based on binge/vomit episode frequency or percent reduction from pre-treatment, during either the first four or first two weeks of treatment. All definitions were associated with higher remission rates in rapid compared to nonrapid responders. Only one definition (i.e., ≤3 episodes in the first four weeks of treatment) predicted sustained remission (versus relapse) at 6- and 12-month follow-up. These findings provide an empirically derived definition of RR to intensive eating disorder treatment, and provide further evidence that early change is an important prognostic indicator. Copyright © 2015 Elsevier Ltd. All rights reserved.

  14. A Behaviorally Specific, Empirical Alternative to Bullying: Aggravated Peer Victimization.

    PubMed

    Finkelhor, David; Shattuck, Anne; Turner, Heather; Hamby, Sherry

    2016-11-01

    To test a behaviorally specific measure of serious peer victimization, called aggravated peer victimization (APV), using empirically derived aggravating elements of episodes (injury, weapon, bias content, sexual content, multiple perpetrators, and multiple contexts) and compare this measure with the conventional Olweus bullying (OB) measure, which uses repetition and power imbalance as its seriousness criteria. The data for this study come from The National Survey of Children's Exposure to Violence 2014, a study conducted via telephone interviews with a nationally representative sample. This analysis uses the 1,949 youth ages 10-17 from that survey. The APV measure identified twice as many youth with serious episodes involving injury, weapons, sexual assaults, and bias content as the OB measure. In terms of demographic and social characteristics, the groups were very similar. However, the APV explained significantly more of the variation in distress than the OB (R 2  = .19 vs. .12). An empirical approach to identifying the most serious incidents of peer victimization has advantages in identifying more of the youth suffering the effects of peer victimization. Moreover, its behaviorally specific criteria also bypass the difficult challenge of trying to reliably assess what is truly bullying with its ambiguous definitional element of power imbalance. Copyright © 2016 Society for Adolescent Health and Medicine. Published by Elsevier Inc. All rights reserved.

  15. Creating an automated chiller fault detection and diagnostics tool using a data fault library.

    PubMed

    Bailey, Margaret B; Kreider, Jan F

    2003-07-01

    Reliable, automated detection and diagnosis of abnormal behavior within vapor compression refrigeration cycle (VCRC) equipment is extremely desirable for equipment owners and operators. The specific type of VCRC equipment studied in this paper is a 70-ton helical rotary, air-cooled chiller. The fault detection and diagnostic (FDD) tool developed as part of this research analyzes chiller operating data and detects faults through recognizing trends or patterns existing within the data. The FDD method incorporates a neural network (NN) classifier to infer the current state given a vector of observables. Therefore the FDD method relies upon the availability of normal and fault empirical data for training purposes and therefore a fault library of empirical data is assembled. This paper presents procedures for conducting sophisticated fault experiments on chillers that simulate air-cooled condenser, refrigerant, and oil related faults. The experimental processes described here are not well documented in literature and therefore will provide the interested reader with a useful guide. In addition, the authors provide evidence, based on both thermodynamics and empirical data analysis, that chiller performance is significantly degraded during fault operation. The chiller's performance degradation is successfully detected and classified by the NN FDD classifier as discussed in the paper's final section.

  16. Wavelet Analysis for Wind Fields Estimation

    PubMed Central

    Leite, Gladeston C.; Ushizima, Daniela M.; Medeiros, Fátima N. S.; de Lima, Gilson G.

    2010-01-01

    Wind field analysis from synthetic aperture radar images allows the estimation of wind direction and speed based on image descriptors. In this paper, we propose a framework to automate wind direction retrieval based on wavelet decomposition associated with spectral processing. We extend existing undecimated wavelet transform approaches, by including à trous with B3 spline scaling function, in addition to other wavelet bases as Gabor and Mexican-hat. The purpose is to extract more reliable directional information, when wind speed values range from 5 to 10 ms−1. Using C-band empirical models, associated with the estimated directional information, we calculate local wind speed values and compare our results with QuikSCAT scatterometer data. The proposed approach has potential application in the evaluation of oil spills and wind farms. PMID:22219699

  17. A stable systemic risk ranking in China's banking sector: Based on principal component analysis

    NASA Astrophysics Data System (ADS)

    Fang, Libing; Xiao, Binqing; Yu, Honghai; You, Qixing

    2018-02-01

    In this paper, we compare five popular systemic risk rankings, and apply principal component analysis (PCA) model to provide a stable systemic risk ranking for the Chinese banking sector. Our empirical results indicate that five methods suggest vastly different systemic risk rankings for the same bank, while the combined systemic risk measure based on PCA provides a reliable ranking. Furthermore, according to factor loadings of the first component, PCA combined ranking is mainly based on fundamentals instead of market price data. We clearly find that price-based rankings are not as practical a method as fundamentals-based ones. This PCA combined ranking directly shows systemic risk contributions of each bank for banking supervision purpose and reminds banks to prevent and cope with the financial crisis in advance.

  18. Effect of Small Numbers of Test Results on Accuracy of Hoek-Brown Strength Parameter Estimations: A Statistical Simulation Study

    NASA Astrophysics Data System (ADS)

    Bozorgzadeh, Nezam; Yanagimura, Yoko; Harrison, John P.

    2017-12-01

    The Hoek-Brown empirical strength criterion for intact rock is widely used as the basis for estimating the strength of rock masses. Estimations of the intact rock H-B parameters, namely the empirical constant m and the uniaxial compressive strength σc, are commonly obtained by fitting the criterion to triaxial strength data sets of small sample size. This paper investigates how such small sample sizes affect the uncertainty associated with the H-B parameter estimations. We use Monte Carlo (MC) simulation to generate data sets of different sizes and different combinations of H-B parameters, and then investigate the uncertainty in H-B parameters estimated from these limited data sets. We show that the uncertainties depend not only on the level of variability but also on the particular combination of parameters being investigated. As particular combinations of H-B parameters can informally be considered to represent specific rock types, we discuss that as the minimum number of required samples depends on rock type it should correspond to some acceptable level of uncertainty in the estimations. Also, a comparison of the results from our analysis with actual rock strength data shows that the probability of obtaining reliable strength parameter estimations using small samples may be very low. We further discuss the impact of this on ongoing implementation of reliability-based design protocols and conclude with suggestions for improvements in this respect.

  19. A General Reliability Model for Ni-BaTiO3-Based Multilayer Ceramic Capacitors

    NASA Technical Reports Server (NTRS)

    Liu, Donhang

    2014-01-01

    The evaluation of multilayer ceramic capacitors (MLCCs) with Ni electrode and BaTiO3 dielectric material for potential space project applications requires an in-depth understanding of their reliability. A general reliability model for Ni-BaTiO3 MLCC is developed and discussed. The model consists of three parts: a statistical distribution; an acceleration function that describes how a capacitor's reliability life responds to the external stresses, and an empirical function that defines contribution of the structural and constructional characteristics of a multilayer capacitor device, such as the number of dielectric layers N, dielectric thickness d, average grain size, and capacitor chip size A. Application examples are also discussed based on the proposed reliability model for Ni-BaTiO3 MLCCs.

  20. A General Reliability Model for Ni-BaTiO3-Based Multilayer Ceramic Capacitors

    NASA Technical Reports Server (NTRS)

    Liu, Donhang

    2014-01-01

    The evaluation for potential space project applications of multilayer ceramic capacitors (MLCCs) with Ni electrode and BaTiO3 dielectric material requires an in-depth understanding of the MLCCs reliability. A general reliability model for Ni-BaTiO3 MLCCs is developed and discussed in this paper. The model consists of three parts: a statistical distribution; an acceleration function that describes how a capacitors reliability life responds to external stresses; and an empirical function that defines the contribution of the structural and constructional characteristics of a multilayer capacitor device, such as the number of dielectric layers N, dielectric thickness d, average grain size r, and capacitor chip size A. Application examples are also discussed based on the proposed reliability model for Ni-BaTiO3 MLCCs.

  1. Adapting the SERVQUAL scale to hospital services: an empirical investigation.

    PubMed Central

    Babakus, E; Mangold, W G

    1992-01-01

    Defining and measuring the quality of service has been a major challenge for health care marketers. A comprehensive service quality measurement scale (SERVQUAL) is empirically evaluated for its potential usefulness in a hospital service environment. Active participation by hospital management helped to address practical and user-related aspects of the assessment. The completed expectations and perceptions scales met various criteria for reliability and validity. Suggestions are provided for the managerial use of the scale, and a number of future research issues are identified. PMID:1737708

  2. [A short form of the positions on nursing diagnosis scale: development and psychometric testing].

    PubMed

    Romero-Sánchez, José Manuel; Paloma-Castro, Olga; Paramio-Cuevas, Juan Carlos; Pastor-Montero, Sonia María; O'Ferrall-González, Cristina; Gabaldón-Bravo, Eva Maria; González-Domínguez, Maria Eugenia; Castro-Yuste, Cristina; Frandsen, Anna J; Martínez-Sabater, Antonio

    2013-06-01

    The Positions on Nursing Diagnosis (PND) is a scale that uses the semantic differential technique to measure nurses' attitudes towards the nursing diagnosis concept. The aim of this study was to develop a shortened form of the Spanish version of this scale and evaluate its psychometric properties and efficiency. A double theoretical-empirical approach was used to obtain a short form of the PND, the PND-7-SV, which would be equivalent to the original. Using a cross-sectional survey design, the reliability (internal consistency and test-retest reliability), construct (exploratory factor analysis, known-groups technique and discriminant validity) and criterion-related validity (concurrent validity), sensitivity to change and efficiency of the PND-7-SV were assessed in a sample of 476 Spanish nursing students. The results endorsed the utility of the PND-7-SV to measure attitudes toward nursing diagnosis in an equivalent manner to the complete form of the scale and in a shorter time.

  3. Student engagement and its relationship with early high school dropout.

    PubMed

    Archambault, Isabelle; Janosz, Michel; Fallu, Jean-Sébastien; Pagani, Linda S

    2009-06-01

    Although the concept of school engagement figures prominently in most school dropout theories, there has been little empirical research conducted on its nature and course and, more importantly, the association with dropout. Information on the natural development of school engagement would greatly benefit those interested in preventing student alienation during adolescence. Using a longitudinal sample of 11,827 French-Canadian high school students, we tested behavioral, affective, cognitive indices of engagement both separately and as a global construct. We then assessed their contribution as prospective predictors of school dropout using factor analysis and structural equation modeling. Global engagement reliably predicted school dropout. Among its three specific dimensions, only behavioral engagement made a significant contribution in the prediction equation. Our findings confirm the robustness of the overall multidimensional construct of school engagement, which reflects both cognitive and psychosocial characteristics, and underscore the importance attributed to basic participation and compliance issues in reliably estimating risk of not completing basic schooling during adolescence.

  4. Validation of the knowledge, attitude and perceived practice of asthma instrument among community pharmacists using Rasch analysis.

    PubMed

    Akram, Waqas; Hussein, Maryam S E; Ahmad, Sohail; Mamat, Mohd N; Ismail, Nahlah E

    2015-10-01

    There is no instrument which collectively assesses the knowledge, attitude and perceived practice of asthma among community pharmacists. Therefore, this study aimed to validate the instrument which measured the knowledge, attitude and perceived practice of asthma among community pharmacists by producing empirical evidence of validity and reliability of the items using Rasch model (Bond & Fox software®) for dichotomous and polytomous data. This baseline study recruited 33 community pharmacists from Penang, Malaysia. The results showed that all PTMEA Corr were in positive values, where an item was able to distinguish between the ability of respondents. Based on the MNSQ infit and outfit range (0.60-1.40), out of 55 items, 2 items from the instrument were suggested to be removed. The findings indicated that the instrument fitted with Rasch measurement model and showed the acceptable reliability values of 0.88 and 0.83 and 0.79 for knowledge, attitude and perceived practice respectively.

  5. Religious Priming: A Meta-Analysis With a Focus on Prosociality.

    PubMed

    Shariff, Azim F; Willard, Aiyana K; Andersen, Teresa; Norenzayan, Ara

    2016-02-01

    Priming has emerged as a valuable tool within the psychological study of religion, allowing for tests of religion's causal effect on a number of psychological outcomes, such as prosocial behavior. As the literature has grown, questions about the reliability and boundary conditions of religious priming have arisen. We use a combination of traditional effect-size analyses, p-curve analyses, and adjustments for publication bias to evaluate the robustness of four types of religious priming (Analyses 1-3), review the empirical evidence for religion's effect specifically on prosocial behavior (Analyses 4-5), and test whether religious-priming effects generalize to individuals who report little or no religiosity (Analyses 6-7). Results across 93 studies and 11,653 participants show that religious priming has robust effects across a variety of outcome measures-prosocial measures included. Religious priming does not, however, reliably affect non-religious participants-suggesting that priming depends on the cognitive activation of culturally transmitted religious beliefs. © 2015 by the Society for Personality and Social Psychology, Inc.

  6. Changes in Differentiation-Relatedness During Psychoanalysis.

    PubMed

    Calamaras, Martha R; Reviere, Susan L; Gallagher, Kathryn E; Kaslow, Nadine J

    2016-01-01

    This study sought to determine (a) if the Differentiation-Relatedness Scale of Self and Object Representations (D-RS), a coding model used with the Object Relations Inventory (Blatt, Wein, Chevron, & Quinlan, 1979 ) could be reliably applied to transcripts of psychoanalyses, and (b) if levels of differentiation-relatedness improve over the course of psychoanalysis. Participants were 4 creative writers who underwent psychoanalysis as part of a longitudinal research project focused on the processes and outcomes of psychoanalysis. Transcripts from the beginning and termination phases of psychoanalysis were coded by 2 independent raters for global, low, and high levels of self and other differentiation-relatedness and compared. There was good interrater agreement, suggesting that, like other forms of narrative material, psychoanalysis transcripts can be reliably rated for levels of object relations. Analysands showed an increase in global levels of differentiation-relatedness from a predominance of emergent ambivalent constancy (M = 6.2) at the beginning of analysis to consolidated, constant representations of self and other (M = 7.5) at the end of analysis. These preliminary findings contribute significantly to the empirical literature with regard to the measurement of self and object representations and change in these representations over the course of psychoanalysis.

  7. Empirical research in medical ethics: how conceptual accounts on normative-empirical collaboration may improve research practice.

    PubMed

    Salloch, Sabine; Schildmann, Jan; Vollmann, Jochen

    2012-04-13

    The methodology of medical ethics during the last few decades has shifted from a predominant use of normative-philosophical analyses to an increasing involvement of empirical methods. The articles which have been published in the course of this so-called 'empirical turn' can be divided into conceptual accounts of empirical-normative collaboration and studies which use socio-empirical methods to investigate ethically relevant issues in concrete social contexts. A considered reference to normative research questions can be expected from good quality empirical research in medical ethics. However, a significant proportion of empirical studies currently published in medical ethics lacks such linkage between the empirical research and the normative analysis. In the first part of this paper, we will outline two typical shortcomings of empirical studies in medical ethics with regard to a link between normative questions and empirical data: (1) The complete lack of normative analysis, and (2) cryptonormativity and a missing account with regard to the relationship between 'is' and 'ought' statements. Subsequently, two selected concepts of empirical-normative collaboration will be presented and how these concepts may contribute to improve the linkage between normative and empirical aspects of empirical research in medical ethics will be demonstrated. Based on our analysis, as well as our own practical experience with empirical research in medical ethics, we conclude with a sketch of concrete suggestions for the conduct of empirical research in medical ethics. High quality empirical research in medical ethics is in need of a considered reference to normative analysis. In this paper, we demonstrate how conceptual approaches of empirical-normative collaboration can enhance empirical research in medical ethics with regard to the link between empirical research and normative analysis.

  8. Empirical research in medical ethics: How conceptual accounts on normative-empirical collaboration may improve research practice

    PubMed Central

    2012-01-01

    Background The methodology of medical ethics during the last few decades has shifted from a predominant use of normative-philosophical analyses to an increasing involvement of empirical methods. The articles which have been published in the course of this so-called 'empirical turn' can be divided into conceptual accounts of empirical-normative collaboration and studies which use socio-empirical methods to investigate ethically relevant issues in concrete social contexts. Discussion A considered reference to normative research questions can be expected from good quality empirical research in medical ethics. However, a significant proportion of empirical studies currently published in medical ethics lacks such linkage between the empirical research and the normative analysis. In the first part of this paper, we will outline two typical shortcomings of empirical studies in medical ethics with regard to a link between normative questions and empirical data: (1) The complete lack of normative analysis, and (2) cryptonormativity and a missing account with regard to the relationship between 'is' and 'ought' statements. Subsequently, two selected concepts of empirical-normative collaboration will be presented and how these concepts may contribute to improve the linkage between normative and empirical aspects of empirical research in medical ethics will be demonstrated. Based on our analysis, as well as our own practical experience with empirical research in medical ethics, we conclude with a sketch of concrete suggestions for the conduct of empirical research in medical ethics. Summary High quality empirical research in medical ethics is in need of a considered reference to normative analysis. In this paper, we demonstrate how conceptual approaches of empirical-normative collaboration can enhance empirical research in medical ethics with regard to the link between empirical research and normative analysis. PMID:22500496

  9. LOFT Debriefings: An Analysis of Instructor Techniques and Crew Participation

    NASA Technical Reports Server (NTRS)

    Dismukes, R. Key; Jobe, Kimberly K.; McDonnell, Lori K.

    1997-01-01

    This study analyzes techniques instructors use to facilitate crew analysis and evaluation of their Line-Oriented Flight Training (LOFT) performance. A rating instrument called the Debriefing Assessment Battery (DAB) was developed which enables raters to reliably assess instructor facilitation techniques and characterize crew participation. Thirty-six debriefing sessions conducted at five U.S. airlines were analyzed to determine the nature of instructor facilitation and crew participation. Ratings obtained using the DAB corresponded closely with descriptive measures of instructor and crew performance. The data provide empirical evidence that facilitation can be an effective tool for increasing the depth of crew participation and self-analysis of CRM performance. Instructor facilitation skill varied dramatically, suggesting a need for more concrete hands-on training in facilitation techniques. Crews were responsive but fell short of actively leading their own debriefings. Ways to improve debriefing effectiveness are suggested.

  10. Teleconsultation in school settings: linking classroom teachers and behavior analysts through web-based technology.

    PubMed

    Frieder, Jessica E; Peterson, Stephanie M; Woodward, Judy; Crane, Jaelee; Garner, Marlane

    2009-01-01

    This paper describes a technically driven, collaborative approach to assessing the function of problem behavior using web-based technology. A case example is provided to illustrate the process used in this pilot project. A school team conducted a functional analysis with a child who demonstrated challenging behaviors in a preschool setting. Behavior analysts at a university setting provided the school team with initial workshop trainings, on-site visits, e-mail and phone communication, as well as live web-based feedback on functional analysis sessions. The school personnel implemented the functional analysis with high fidelity and scored the data reliably. Outcomes of the project suggest that there is great potential for collaboration via the use of web-based technologies for ongoing assessment and development of effective interventions. However, an empirical evaluation of this model should be conducted before wide-scale adoption is recommended.

  11. Spatiotemporal Bayesian analysis of Lyme disease in New York state, 1990-2000.

    PubMed

    Chen, Haiyan; Stratton, Howard H; Caraco, Thomas B; White, Dennis J

    2006-07-01

    Mapping ordinarily increases our understanding of nontrivial spatial and temporal heterogeneities in disease rates. However, the large number of parameters required by the corresponding statistical models often complicates detailed analysis. This study investigates the feasibility of a fully Bayesian hierarchical regression approach to the problem and identifies how it outperforms two more popular methods: crude rate estimates (CRE) and empirical Bayes standardization (EBS). In particular, we apply a fully Bayesian approach to the spatiotemporal analysis of Lyme disease incidence in New York state for the period 1990-2000. These results are compared with those obtained by CRE and EBS in Chen et al. (2005). We show that the fully Bayesian regression model not only gives more reliable estimates of disease rates than the other two approaches but also allows for tractable models that can accommodate more numerous sources of variation and unknown parameters.

  12. Examination of the reliability of the crash modification factors using empirical Bayes method with resampling technique.

    PubMed

    Wang, Jung-Han; Abdel-Aty, Mohamed; Wang, Ling

    2017-07-01

    There have been plenty of studies intended to use different methods, for example, empirical Bayes before-after methods, to get accurate estimation of CMFs. All of them have different assumptions toward crash count if there was no treatment. Additionally, another major assumption is that multiple sites share the same true CMF. Under this assumption, the CMF at an individual intersection is randomly drawn from a normally distributed population of CMFs at all intersections. Since CMFs are non-zero values, the population of all CMFs might not follow normal distributions, and even if it does, the true mean of CMFs at some intersections may be different from that at others. Therefore, a bootstrap method based on before-after empirical Bayes theory was proposed to estimate CMFs, but it did not make distributional assumptions. This bootstrap procedure has the added benefit of producing a measure of CMF stability. Furthermore, based on the bootstrapped CMF, a new CMF precision rating method was proposed to evaluate the reliability of CMFs. This study chose 29 urban four-legged intersections as treated sites, and their controls were changed from stop-controlled to signal-controlled. Meanwhile, 124 urban four-legged stop-controlled intersections were selected as reference sites. At first, different safety performance functions (SPFs) were applied to five crash categories, and it was found that each crash category had different optimal SPF form. Then, the CMFs of these five crash categories were estimated using the bootstrap empirical Bayes method. The results of the bootstrapped method showed that signalization significantly decreased Angle+Left-Turn crashes, and its CMF had the highest precision. While, the CMF for Rear-End crashes was unreliable. For KABCO, KABC, and KAB crashes, their CMFs were proved to be reliable for the majority of intersections, but the estimated effect of signalization may be not accurate at some sites. Copyright © 2017 Elsevier Ltd. All rights reserved.

  13. Olympic Scoring of English Compositions

    ERIC Educational Resources Information Center

    Follman, John; Panther, Edward

    1974-01-01

    Examines empirically the efficacy of utilizing Olympic diving and gymnastic scoring systems for grading graduate students' English compositions. Results indicated that such scoring rules do not produce ratings different in reliability or in level from conventional letter grades. (ED)

  14. Reliability reporting across studies using the Buss Durkee Hostility Inventory.

    PubMed

    Vassar, Matt; Hale, William

    2009-01-01

    Empirical research on anger and hostility has pervaded the academic literature for more than 50 years. Accurate measurement of anger/hostility and subsequent interpretation of results requires that the instruments yield strong psychometric properties. For consistent measurement, reliability estimates must be calculated with each administration, because changes in sample characteristics may alter the scale's ability to generate reliable scores. Therefore, the present study was designed to address reliability reporting practices for a widely used anger assessment, the Buss Durkee Hostility Inventory (BDHI). Of the 250 published articles reviewed, 11.2% calculated and presented reliability estimates for the data at hand, 6.8% cited estimates from a previous study, and 77.1% made no mention of score reliability. Mean alpha estimates of scores for BDHI subscales generally fell below acceptable standards. Additionally, no detectable pattern was found between reporting practices and publication year or journal prestige. Areas for future research are also discussed.

  15. The Development of an Empirical Model of Mental Health Stigma in Adolescents.

    PubMed

    Silke, Charlotte; Swords, Lorraine; Heary, Caroline

    2016-08-30

    Research on mental health stigma in adolescents is hampered by a lack of empirical investigation into the theoretical conceptualisation of stigma, as well as by the lack of validated stigma measures. This research aims to develop a model of public stigma toward depression in adolescents and to use this model to empirically examine whether stigma is composed of three separate dimensions (Stereotypes, Prejudice and Discrimination), as is theoretically proposed. Adolescents completed self-report measures assessing their stigmatising responses toward a fictional peer with depression. An exploratory factor analysis (EFA; N=332) was carried out on 58-items, which proposed to measure aspects of stigma. A confirmatory factor analysis (CFA; N=236) was then carried out to evaluate the validity of the observed stigma model. Finally, higher-order CFAs were conducted in order to assess whether the observed model supported the tripartite conceptualisation of stigma. The EFA returned a seven-factor model of stigma. These factors were designated as Dangerousness, Warmth & Competency, Responsibility, Negative Attributes, Prejudice, Classroom Discrimination and Friendship Discrimination. The CFA supported the goodness-of-fit of this seven-factor model. The higher-order CFAs indicated that these seven factors represented the latent constructs of, Stereotypes, Prejudice and Discrimination, which in turn represented Stigma. Overall, results support the tripartite conceptualisation of stigma and suggest that measurements of mental health stigma in adolescents should include assessments of all three dimensions. These results also highlight the importance of establishing valid and reliable measures for assessing stigma in adolescents. Copyright © 2016 Elsevier Ireland Ltd. All rights reserved.

  16. Sediment yield estimation in mountain catchments of the Camastra reservoir, southern Italy: a comparison among different empirical methods

    NASA Astrophysics Data System (ADS)

    Lazzari, Maurizio; Danese, Maria; Gioia, Dario; Piccarreta, Marco

    2013-04-01

    Sedimentary budget estimation is an important topic for both scientific and social community, because it is crucial to understand both dynamics of orogenic belts and many practical problems, such as soil conservation and sediment accumulation in reservoir. Estimations of sediment yield or denudation rates in southern-central Italy are generally obtained by simple empirical relationships based on statistical regression between geomorphic parameters of the drainage network and the measured suspended sediment yield at the outlet of several drainage basins or through the use of models based on sediment delivery ratio or on soil loss equations. In this work, we perform a study of catchment dynamics and an estimation of sedimentary yield for several mountain catchments of the central-western sector of the Basilicata region, southern Italy. Sediment yield estimation has been obtained through both an indirect estimation of suspended sediment yield based on the Tu index (mean annual suspension sediment yield, Ciccacci et al., 1980) and the application of the Rusle (Renard et al., 1997) and the USPED (Mitasova et al., 1996) empirical methods. The preliminary results indicate a reliable difference between the RUSLE and USPED methods and the estimation based on the Tu index; a critical data analysis of results has been carried out considering also the present-day spatial distribution of erosion, transport and depositional processes in relation to the maps obtained from the application of those different empirical methods. The studied catchments drain an artificial reservoir (i.e. the Camastra dam), where a detailed evaluation of the amount of historical sediment storage has been collected. Sediment yield estimation obtained by means of the empirical methods have been compared and checked with historical data of sediment accumulation measured in the artificial reservoir of the Camastra dam. The validation of such estimations of sediment yield at the scale of large catchments using sediment storage in reservoirs provides a good opportunity: i) to test the reliability of the empirical methods used to estimate the sediment yield; ii) to investigate the catchment dynamics and its spatial and temporal evolution in terms of erosion, transport and deposition. References Ciccacci S., Fredi F., Lupia Palmieri E., Pugliese F., 1980. Contributo dell'analisi geomorfica quantitativa alla valutazione dell'entita dell'erosione nei bacini fluviali. Bollettino della Società Geologica Italiana 99: 455-516. Mitasova H, Hofierka J, Zlocha M, Iverson LR. 1996. Modeling topographic potential for erosion and deposition using GIS. International Journal of Geographical Information Systems 10: 629-641. Renard K.G., Foster G.R., Weesies G.A., McCool D.K., Yoder D.C., 1997. Predicting soil erosion by water: a guide to conservation planning with the Revised Universal Soil Loss Equation (RUSLE), USDA-ARS, Agricultural Handbook No. 703.

  17. A method of bias correction for maximal reliability with dichotomous measures.

    PubMed

    Penev, Spiridon; Raykov, Tenko

    2010-02-01

    This paper is concerned with the reliability of weighted combinations of a given set of dichotomous measures. Maximal reliability for such measures has been discussed in the past, but the pertinent estimator exhibits a considerable bias and mean squared error for moderate sample sizes. We examine this bias, propose a procedure for bias correction, and develop a more accurate asymptotic confidence interval for the resulting estimator. In most empirically relevant cases, the bias correction and mean squared error correction can be performed simultaneously. We propose an approximate (asymptotic) confidence interval for the maximal reliability coefficient, discuss the implementation of this estimator, and investigate the mean squared error of the associated asymptotic approximation. We illustrate the proposed methods using a numerical example.

  18. An empirical study of flight control software reliability

    NASA Technical Reports Server (NTRS)

    Dunham, J. R.; Pierce, J. L.

    1986-01-01

    The results of a laboratory experiment in flight control software reliability are reported. The experiment tests a small sample of implementations of a pitch axis control law for a PA28 aircraft with over 14 million pitch commands with varying levels of additive input and feedback noise. The testing which uses the method of n-version programming for error detection surfaced four software faults in one implementation of the control law. The small number of detected faults precluded the conduct of the error burst analyses. The pitch axis problem provides data for use in constructing a model in the prediction of the reliability of software in systems with feedback. The study is undertaken to find means to perform reliability evaluations of flight control software.

  19. The reliability and validity of a child and adolescent participation in decision-making questionnaire.

    PubMed

    O'Hare, L; Santin, O; Winter, K; McGuinness, C

    2016-09-01

    There is a growing impetus across the research, policy and practice communities for children and young people to participate in decisions that affect their lives. Furthermore, there is a dearth of general instruments that measure children and young people's views on their participation in decision-making. This paper presents the reliability and validity of the Child and Adolescent Participation in Decision-Making Questionnaire (CAP-DMQ) and specifically looks at a population of looked-after children, where a lack of participation in decision-making is an acute issue. The participants were 151 looked after children and adolescents between 10-23 years of age who completed the 10 item CAP-DMQ. Of the participants 113 were in receipt of an advocacy service that had an aim of increasing participation in decision-making with the remaining participants not having received this service. The results showed that the CAP-DMQ had good reliability (Cronbach's alpha = 0.94) and showed promising uni-dimensional construct validity through an exploratory factor analysis. The items in the CAP-DMQ also demonstrated good content validity by overlapping with prominent models of child and adolescent participation (Lundy 2007) and decision-making (Halpern 2014). A regression analysis showed that age and gender were not significant predictors of CAP-DMQ scores but receipt of advocacy was a significant predictor of scores (effect size d = 0.88), thus showing appropriate discriminant criterion validity. Overall, the CAP-DMQ showed good reliability and validity. Therefore, the measure has excellent promise for theoretical investigation in the area of child and adolescent participation in decision-making and equally shows empirical promise for use as a measure in evaluating services, which have increasing the participation of children and adolescents in decision-making as an intended outcome. © 2016 John Wiley & Sons Ltd.

  20. ESTIMATING IMPERVIOUS COVER FROM REGIONALLY AVAILABLE DATA

    EPA Science Inventory

    The objective of this study is to compare and evaluate the reliability of different approaches for estimating impervious cover including three empirical formulations for estimating impervious cover from population density data, estimation from categorized land cover data, and to ...

  1. Determination of dynamic modulus master curves for Oklahoma HMA mixtures.

    DOT National Transportation Integrated Search

    2007-12-01

    The Mechanistic-Empirical Pavement Design Guide (M-EPDG) uses a hierarchical approach with three : levels of material characterization for asphalt materials. The first level provides the highest design : reliability and each succeeding level is a dro...

  2. Testing Standard Reliability Criteria

    ERIC Educational Resources Information Center

    Sherry, David

    2017-01-01

    Maul's paper, "Rethinking Traditional Methods of Survey Validation" (Andrew Maul), contains two stages. First he presents empirical results that cast doubt on traditional methods for validating psychological measurement instruments. These results motivate the second stage, a critique of current conceptions of psychological measurement…

  3. Magnetic Memory of Two Lunar Samples, 15405 and 15445

    NASA Astrophysics Data System (ADS)

    Kletetschka, G.; Kamenikova, K.; Fuller, M.; Cizkova, K.

    2016-08-01

    We reanalyzed Apollo's literature records using empirical scaling law methods to search for the presence of Lunar iron carriers capable of recording reliable magnetic paleofields. Lunar rocks have a large spectrum of paleofields (1-100uT).

  4. On the Reliability of Source Time Functions Estimated Using Empirical Green's Function Methods

    NASA Astrophysics Data System (ADS)

    Gallegos, A. C.; Xie, J.; Suarez Salas, L.

    2017-12-01

    The Empirical Green's Function (EGF) method (Hartzell, 1978) has been widely used to extract source time functions (STFs). In this method, seismograms generated by collocated events with different magnitudes are deconvolved. Under a fundamental assumption that the STF of the small event is a delta function, the deconvolved Relative Source Time Function (RSTF) yields the large event's STF. While this assumption can be empirically justified by examination of differences in event size and frequency content of the seismograms, there can be a lack of rigorous justification of the assumption. In practice, a small event might have a finite duration when the RSTF is retrieved and interpreted as the large event STF with a bias. In this study, we rigorously analyze this bias using synthetic waveforms generated by convolving a realistic Green's function waveform with pairs of finite-duration triangular or parabolic STFs. The RSTFs are found using a time-domain based matrix deconvolution. We find when the STFs of smaller events are finite, the RSTFs are a series of narrow non-physical spikes. Interpreting these RSTFs as a series of high-frequency source radiations would be very misleading. The only reliable and unambiguous information we can retrieve from these RSTFs is the difference in durations and the moment ratio of the two STFs. We can apply a Tikhonov smoothing to obtain a single-pulse RSTF, but its duration is dependent on the choice of weighting, which may be subjective. We then test the Multi-Channel Deconvolution (MCD) method (Plourde & Bostock, 2017) which assumes that both STFs have finite durations to be solved for. A concern about the MCD method is that the number of unknown parameters is larger, which would tend to make the problem rank-deficient. Because the kernel matrix is dependent on the STFs to be solved for under a positivity constraint, we can only estimate the rank-deficiency with a semi-empirical approach. Based on the results so far, we find that the rank-deficiency makes it improbable to solve for both STFs. To solve for the larger STF we need to assume the shape of the small STF to be known a priori. Thus, the reliability of the estimated large STF depends on the difference between the assumed and true shapes of the small STF. We will show how the reliability varies with realistic scenarios.

  5. Scenario analysis of freight vehicle accident risks in Taiwan.

    PubMed

    Tsai, Ming-Chih; Su, Chien-Chih

    2004-07-01

    This study develops a quantitative risk model by utilizing Generalized Linear Interactive Model (GLIM) to analyze the major freight vehicle accidents in Taiwan. Eight scenarios are established by interacting three categorical variables of driver ages, vehicle types and road types, each of which contains two levels. The database that consists of 2043 major accidents occurring between 1994 and 1998 in Taiwan is utilized to fit and calibrate the model parameters. The empirical results indicate that accident rates of freight vehicles in Taiwan were high in the scenarios involving trucks and non-freeway systems, while; accident consequences were severe in the scenarios involving mature drivers or non-freeway systems. Empirical evidences also show that there is no significant relationship between accident rates and accident consequences. This is to stress that safety studies that describe risk merely as accident rates rather than the combination of accident rates and consequences by definition might lead to biased risk perceptions. Finally, the study recommends using number of vehicle as an alternative of traffic exposure in commercial vehicle risk analysis. The merits of this would be that it is simple and thus reliable; meanwhile, the resulted risk that is termed as fatalities per vehicle could provide clear and direct policy implications for insurance practices and safety regulations.

  6. Time-frequency analysis : mathematical analysis of the empirical mode decomposition.

    DOT National Transportation Integrated Search

    2009-01-01

    Invented over 10 years ago, empirical mode : decomposition (EMD) provides a nonlinear : time-frequency analysis with the ability to successfully : analyze nonstationary signals. Mathematical : Analysis of the Empirical Mode Decomposition : is a...

  7. Treatment decisions under ambiguity.

    PubMed

    Berger, Loïc; Bleichrodt, Han; Eeckhoudt, Louis

    2013-05-01

    Many health risks are ambiguous in the sense that reliable and credible information about these risks is unavailable. In health economics, ambiguity is usually handled through sensitivity analysis, which implicitly assumes that people are neutral towards ambiguity. However, empirical evidence suggests that people are averse to ambiguity and react strongly to it. This paper studies the effects of ambiguity aversion on two classical medical decision problems. If there is ambiguity regarding the diagnosis of a patient, ambiguity aversion increases the decision maker's propensity to opt for treatment. On the other hand, in the case of ambiguity regarding the effects of treatment, ambiguity aversion leads to a reduction in the propensity to choose treatment. Copyright © 2013 Elsevier B.V. All rights reserved.

  8. Basic research in evolution and ecology enhances forensics.

    PubMed

    Tomberlin, Jeffery K; Benbow, M Eric; Tarone, Aaron M; Mohr, Rachel M

    2011-02-01

    In 2009, the National Research Council recommended that the forensic sciences strengthen their grounding in basic empirical research to mitigate against criticism and improve accuracy and reliability. For DNA-based identification, this goal was achieved under the guidance of the population genetics community. This effort resulted in DNA analysis becoming the 'gold standard' of the forensic sciences. Elsewhere, we proposed a framework for streamlining research in decomposition ecology, which promotes quantitative approaches to collecting and applying data to forensic investigations involving decomposing human remains. To extend the ecological aspects of this approach, this review focuses on forensic entomology, although the framework can be extended to other areas of decomposition. Published by Elsevier Ltd.

  9. Patterns of Cognitive Strengths and Weaknesses: Identification Rates, Agreement, and Validity for Learning Disabilities Identification

    PubMed Central

    Miciak, Jeremy; Fletcher, Jack M.; Stuebing, Karla; Vaughn, Sharon; Tolar, Tammy D.

    2014-01-01

    Purpose Few empirical investigations have evaluated LD identification methods based on a pattern of cognitive strengths and weaknesses (PSW). This study investigated the reliability and validity of two proposed PSW methods: the concordance/discordance method (C/DM) and cross battery assessment (XBA) method. Methods Cognitive assessment data for 139 adolescents demonstrating inadequate response to intervention was utilized to empirically classify participants as meeting or not meeting PSW LD identification criteria using the two approaches, permitting an analysis of: (1) LD identification rates; (2) agreement between methods; and (3) external validity. Results LD identification rates varied between the two methods depending upon the cut point for low achievement, with low agreement for LD identification decisions. Comparisons of groups that met and did not meet LD identification criteria on external academic variables were largely null, raising questions of external validity. Conclusions This study found low agreement and little evidence of validity for LD identification decisions based on PSW methods. An alternative may be to use multiple measures of academic achievement to guide intervention. PMID:24274155

  10. Development and validation of the Hogan Grief Reaction Checklist.

    PubMed

    Hogan, N S; Greenfield, D B; Schmidt, L A

    2001-01-01

    The purpose of this article is to provide data on a recently developed instrument to measure the multidimensional nature of the bereavement process. In contrast to widely used grief instruments that have been developed using rational methods of instrument construction, the Hogan Grief Reaction Checklist (HGRC) was developed empirically from data collected from bereaved adults who had experienced the death of a loved one. Factor analysis of the HGRC revealed 6 factors in the normal trajectory of the grieving process: Despair, Panic Behavior, Blame and Anger, Detachment, Disorganization, and Personal Growth. Additional data are provided that support reliability and validity of the HGRC as well as its ability to discriminate variability in the grieving process as a function of cause of death and time lapsed since death. Empirical support is also provided for Personal Growth as an integral component of the bereavement process. The article concludes by considering the substantive as well as psychometric findings of this research for such issues as traumatic grief, anticipatory grief, change in the bereaved person's self-schema, and spiritual and existential growth.

  11. Distinguishing perceived competence and self-efficacy: an example from exercise.

    PubMed

    Rodgers, Wendy M; Markland, David; Selzler, Anne-Marie; Murray, Terra C; Wilson, Philip M

    2014-12-01

    This article examined the conceptual and statistical distinction between perceived competence and self-efficacy. Although they are frequently used interchangeably, it is possible that distinguishing them might assist researchers in better understanding their roles in developing enduring adaptive behavior patterns. Perceived competence is conceived in the theoretical framework of self-determination theory and self-efficacy is conceived in the theoretical framework of social-cognitive theory. The purpose of this study was to empirically distinguish perceived competence from self-efficacy for exercise. Two studies evaluated the independence of perceived competence and self-efficacy in the context of exercise. Using 2 extant instruments with validity and reliability evidence in exercise contexts, the distinctiveness of the 2 constructs was assessed in 2 separate samples (n = 357 middle-aged sedentary adults; n = 247 undergraduate students). Confirmatory factor analysis supported the conceptual and empirical distinction of the 2 constructs. This study supports the conceptual and statistical distinction of perceived competence from perceived self-efficacy. Applications of these results provide a rationale for more precise future theorizing regarding their respective roles in supporting initiation and maintenance of health behaviors.

  12. Empirically Derived Dehydration Scoring and Decision Tree Models for Children With Diarrhea: Assessment and Internal Validation in a Prospective Cohort Study in Dhaka, Bangladesh

    PubMed Central

    Glavis-Bloom, Justin; Modi, Payal; Nasrin, Sabiha; Rege, Soham; Chu, Chieh; Schmid, Christopher H; Alam, Nur H

    2015-01-01

    Introduction: Diarrhea remains one of the most common and most deadly conditions affecting children worldwide. Accurately assessing dehydration status is critical to determining treatment course, yet no clinical diagnostic models for dehydration have been empirically derived and validated for use in resource-limited settings. Methods: In the Dehydration: Assessing Kids Accurately (DHAKA) prospective cohort study, a random sample of children under 5 with acute diarrhea was enrolled between February and June 2014 in Bangladesh. Local nurses assessed children for clinical signs of dehydration on arrival, and then serial weights were obtained as subjects were rehydrated. For each child, the percent weight change with rehydration was used to classify subjects with severe dehydration (>9% weight change), some dehydration (3–9%), or no dehydration (<3%). Clinical variables were then entered into logistic regression and recursive partitioning models to develop the DHAKA Dehydration Score and DHAKA Dehydration Tree, respectively. Models were assessed for their accuracy using the area under their receiver operating characteristic curve (AUC) and for their reliability through repeat clinical exams. Bootstrapping was used to internally validate the models. Results: A total of 850 children were enrolled, with 771 included in the final analysis. Of the 771 children included in the analysis, 11% were classified with severe dehydration, 45% with some dehydration, and 44% with no dehydration. Both the DHAKA Dehydration Score and DHAKA Dehydration Tree had significant AUCs of 0.79 (95% CI = 0.74, 0.84) and 0.76 (95% CI = 0.71, 0.80), respectively, for the diagnosis of severe dehydration. Additionally, the DHAKA Dehydration Score and DHAKA Dehydration Tree had significant positive likelihood ratios of 2.0 (95% CI = 1.8, 2.3) and 2.5 (95% CI = 2.1, 2.8), respectively, and significant negative likelihood ratios of 0.23 (95% CI = 0.13, 0.40) and 0.28 (95% CI = 0.18, 0.44), respectively, for the diagnosis of severe dehydration. Both models demonstrated 90% agreement between independent raters and good reproducibility using bootstrapping. Conclusion: This study is the first to empirically derive and internally validate accurate and reliable clinical diagnostic models for dehydration in a resource-limited setting. After external validation, frontline providers may use these new tools to better manage acute diarrhea in children. PMID:26374802

  13. Empirically Derived Dehydration Scoring and Decision Tree Models for Children With Diarrhea: Assessment and Internal Validation in a Prospective Cohort Study in Dhaka, Bangladesh.

    PubMed

    Levine, Adam C; Glavis-Bloom, Justin; Modi, Payal; Nasrin, Sabiha; Rege, Soham; Chu, Chieh; Schmid, Christopher H; Alam, Nur H

    2015-08-18

    Diarrhea remains one of the most common and most deadly conditions affecting children worldwide. Accurately assessing dehydration status is critical to determining treatment course, yet no clinical diagnostic models for dehydration have been empirically derived and validated for use in resource-limited settings. In the Dehydration: Assessing Kids Accurately (DHAKA) prospective cohort study, a random sample of children under 5 with acute diarrhea was enrolled between February and June 2014 in Bangladesh. Local nurses assessed children for clinical signs of dehydration on arrival, and then serial weights were obtained as subjects were rehydrated. For each child, the percent weight change with rehydration was used to classify subjects with severe dehydration (>9% weight change), some dehydration (3-9%), or no dehydration (<3%). Clinical variables were then entered into logistic regression and recursive partitioning models to develop the DHAKA Dehydration Score and DHAKA Dehydration Tree, respectively. Models were assessed for their accuracy using the area under their receiver operating characteristic curve (AUC) and for their reliability through repeat clinical exams. Bootstrapping was used to internally validate the models. A total of 850 children were enrolled, with 771 included in the final analysis. Of the 771 children included in the analysis, 11% were classified with severe dehydration, 45% with some dehydration, and 44% with no dehydration. Both the DHAKA Dehydration Score and DHAKA Dehydration Tree had significant AUCs of 0.79 (95% CI = 0.74, 0.84) and 0.76 (95% CI = 0.71, 0.80), respectively, for the diagnosis of severe dehydration. Additionally, the DHAKA Dehydration Score and DHAKA Dehydration Tree had significant positive likelihood ratios of 2.0 (95% CI = 1.8, 2.3) and 2.5 (95% CI = 2.1, 2.8), respectively, and significant negative likelihood ratios of 0.23 (95% CI = 0.13, 0.40) and 0.28 (95% CI = 0.18, 0.44), respectively, for the diagnosis of severe dehydration. Both models demonstrated 90% agreement between independent raters and good reproducibility using bootstrapping. This study is the first to empirically derive and internally validate accurate and reliable clinical diagnostic models for dehydration in a resource-limited setting. After external validation, frontline providers may use these new tools to better manage acute diarrhea in children. © Levine et al.

  14. Collection of materials and performance data for Texas flexible pavements and overlays : project summary.

    DOT National Transportation Integrated Search

    2015-08-31

    Proper calibration of mechanistic-empirical : (M-E) design and rehabilitation performance : models to meet Texas conditions is essential : for cost-effective flexible pavement designs. : Such a calibration effort would require a : reliable source of ...

  15. Analysis of acoustic emission signals and monitoring of machining processes

    PubMed

    Govekar; Gradisek; Grabec

    2000-03-01

    Monitoring of a machining process on the basis of sensor signals requires a selection of informative inputs in order to reliably characterize and model the process. In this article, a system for selection of informative characteristics from signals of multiple sensors is presented. For signal analysis, methods of spectral analysis and methods of nonlinear time series analysis are used. With the aim of modeling relationships between signal characteristics and the corresponding process state, an adaptive empirical modeler is applied. The application of the system is demonstrated by characterization of different parameters defining the states of a turning machining process, such as: chip form, tool wear, and onset of chatter vibration. The results show that, in spite of the complexity of the turning process, the state of the process can be well characterized by just a few proper characteristics extracted from a representative sensor signal. The process characterization can be further improved by joining characteristics from multiple sensors and by application of chaotic characteristics.

  16. MEPROCS framework for Craniofacial Superimposition: Validation study.

    PubMed

    Ibáñez, O; Vicente, R; Navega, D; Campomanes-Álvarez, C; Cattaneo, C; Jankauskas, R; Huete, M I; Navarro, F; Hardiman, R; Ruiz, E; Imaizumi, K; Cavalli, F; Veselovskaya, E; Humpire, D; Cardoso, J; Collini, F; Mazzarelli, D; Gibelli, D; Damas, S

    2016-11-01

    Craniofacial Superimposition (CFS) involves the process of overlaying a skull with a number of ante-mortem images of an individual and the analysis of their morphological correspondence. The lack of unified working protocols and the absence of commonly accepted standards, led to contradictory consensus regarding its reliability. One of the more important aims of 'New Methodologies and Protocols of Forensic Identification by Craniofacial Superimposition (MEPROCS)' project was to propose a common framework for CFS, what can be considered the first international standard in the field. The framework aimed to serve as a roadmap for avoiding particular assumptions that could bias the process. At the same time, it provides some empirical support to certain practices, technological means, and morphological criteria expected to facilitate the application of the CFS task and to improve its reliability. In order to confirm the utility and potential benefits of the framework use, there is a need to empirically evaluate it in CFS identification scenarios as close as possible to the reality. Thus, the purpose of this study is to validate the CFS framework developed. For that aim 12 participants were asked to report about a variable number of CFS following all the recommendations of the framework. The results are analysed and discussed according to the framework understanding and fulfilment, the participants' performance, and the correlation between expected decisions and those given by the participants. In view of the quantitative results and qualitative examination criteria we can conclude that those who follow the MEPROCS recommendations improve their performance. Copyright © 2016 Elsevier Ireland Ltd. All rights reserved.

  17. An exploratory study examining the influence of translation on the validity and reliability of qualitative data in nursing research.

    PubMed

    Twinn, S

    1997-08-01

    Although the complexity of undertaking qualitative research with non-English speaking informants has become increasingly recognized, few empirical studies exist which explore the influence of translation on the findings of the study. The aim of this exploratory study was therefore to examine the influence of translation on the reliability and validity of the findings of a qualitative research study. In-depth interviews were undertaken in Cantonese with a convenience sample of six women to explore their perceptions of factors influencing their uptake of Pap smears. Data analysis involved three stages. The first stage involved the translation and transcription of all the interviews into English independently by two translators as well as transcription into Chinese by a third researcher. The second stage involved content analysis of the three data sets to develop categories and themes and the third stage involved a comparison of the categories and themes generated from the Chinese and English data sets. Despite no significant differences in the major categories generated from the Chinese and English data, some minor differences were identified in the themes generated from the data. More significantly the results of the study demonstrated some important issues to consider when using translation in qualitative research, in particular the complexity of managing data when no equivalent word exists in the target language and the influence of the grammatical style on the analysis. In addition the findings raise questions about the significance of the conceptual framework of the research design and sampling to the validity of the study. The importance of using only one translator to maximize the reliability of the study was also demonstrated. In addition the author suggests the findings demonstrate particular problems in using translation in phenomenological research designs.

  18. An experimental investigation of fault tolerant software structures in an avionics application

    NASA Technical Reports Server (NTRS)

    Caglayan, Alper K.; Eckhardt, Dave E., Jr.

    1989-01-01

    The objective of this experimental investigation is to compare the functional performance and software reliability of competing fault tolerant software structures utilizing software diversity. In this experiment, three versions of the redundancy management software for a skewed sensor array have been developed using three diverse failure detection and isolation algorithms and incorporated into various N-version, recovery block and hybrid software structures. The empirical results show that, for maximum functional performance improvement in the selected application domain, the results of diverse algorithms should be voted before being processed by multiple versions without enforced diversity. Results also suggest that when the reliability gain with an N-version structure is modest, recovery block structures are more feasible since higher reliability can be obtained using an acceptance check with a modest reliability.

  19. Measuring metacognitive ability based on science literacy in dynamic electricity topic

    NASA Astrophysics Data System (ADS)

    Warni; Sunyono; Rosidin

    2018-01-01

    This study aims to produce an instrument of metacognition ability assessment based on science literacy on theoretically and empirically feasible dynamic electrical material. The feasibility of the assessment instrument includes theoretical validity on material, construction, and language aspects, as well as empirical validity, reliability, difficulty, distinguishing, and distractor indices. The development of assessment instruments refers to the Dick and Carey development model which includes the preliminary study stage, initial product development, validation and revision, and piloting. The instrument was tested to 32 students of class IX in SMP Negeri 20 Bandar Lampung, using the design of One Group Pretest-Postest Design. The result shows that the metacognition ability assessment instrument based on science literacy is feasible theoretically with theoretical validity percentage of 95.44% and empirical validity of 43.75% for the high category, 43.75% for the medium category, and 12.50 % for low category questions; Reliability of assessment instruments of 0.83 high categories; Difficulty level of difficult item is about 31.25% and medium category is equal to 68.75%. Item that has very good distinguishing power is 12.50%, 62.50% for good stage, and medium category is 25.00%; As well as the duplexing function on a matter of multiple choice is 80.00% including good category and 20.00% for medium category.

  20. Prediction of shear wave velocity using empirical correlations and artificial intelligence methods

    NASA Astrophysics Data System (ADS)

    Maleki, Shahoo; Moradzadeh, Ali; Riabi, Reza Ghavami; Gholami, Raoof; Sadeghzadeh, Farhad

    2014-06-01

    Good understanding of mechanical properties of rock formations is essential during the development and production phases of a hydrocarbon reservoir. Conventionally, these properties are estimated from the petrophysical logs with compression and shear sonic data being the main input to the correlations. This is while in many cases the shear sonic data are not acquired during well logging, which may be for cost saving purposes. In this case, shear wave velocity is estimated using available empirical correlations or artificial intelligent methods proposed during the last few decades. In this paper, petrophysical logs corresponding to a well drilled in southern part of Iran were used to estimate the shear wave velocity using empirical correlations as well as two robust artificial intelligence methods knows as Support Vector Regression (SVR) and Back-Propagation Neural Network (BPNN). Although the results obtained by SVR seem to be reliable, the estimated values are not very precise and considering the importance of shear sonic data as the input into different models, this study suggests acquiring shear sonic data during well logging. It is important to note that the benefits of having reliable shear sonic data for estimation of rock formation mechanical properties will compensate the possible additional costs for acquiring a shear log.

  1. Confirmatory factors analysis of science teacher leadership in the Thailand world-class standard schools

    NASA Astrophysics Data System (ADS)

    Thawinkarn, Dawruwan

    2018-01-01

    This research aims to analyze factors of science teacher leadership in the Thailand World-Class Standard Schools. The research instrument was a five scale rating questionnaire with reliability 0.986. The sample group included 500 science teachers from World-Class Standard Schools who had been selected by using the stratified random sampling technique. Factor analysis of science teacher leadership in the Thailand World-Class Standard Schools was conducted by using M plus for Windows. The results are as follows: The results of confirmatory factor analysis on science teacher leadership in the Thailand World-Class Standard Schools revealed that the model significantly correlated with the empirical data. The consistency index value was x2 = 105.655, df = 88, P-Value = 0.086, TLI = 0.997, CFI = 0.999, RMSEA = 0.022, and SRMR = 0.019. The value of factor loading of science teacher leadership was positive, with statistical significance at the level of 0.01. The value of six factors was between 0.880-0.996. The highest factor loading was the professional learning community, followed by child-centered instruction, participation in development, the role model in teaching, transformational leaders, and self-development with factor loading at 0.996, 0.928, 0.911, 0.907, 0.901, and 0.871, respectively. The reliability of each factor was 99.1%, 86.0%, 83.0%, 82.2%, 81.0%, and 75.8%, respectively.

  2. The Depression Anxiety Stress Scales (DASS): normative data and latent structure in a large non-clinical sample.

    PubMed

    Crawford, John R; Henry, Julie D

    2003-06-01

    To provide UK normative data for the Depression Anxiety and Stress Scale (DASS) and test its convergent, discriminant and construct validity. Cross-sectional, correlational and confirmatory factor analysis (CFA). The DASS was administered to a non-clinical sample, broadly representative of the general adult UK population (N = 1,771) in terms of demographic variables. Competing models of the latent structure of the DASS were derived from theoretical and empirical sources and evaluated using confirmatory factor analysis. Correlational analysis was used to determine the influence of demographic variables on DASS scores. The convergent and discriminant validity of the measure was examined through correlating the measure with two other measures of depression and anxiety (the HADS and the sAD), and a measure of positive and negative affectivity (the PANAS). The best fitting model (CFI =.93) of the latent structure of the DASS consisted of three correlated factors corresponding to the depression, anxiety and stress scales with correlated error permitted between items comprising the DASS subscales. Demographic variables had only very modest influences on DASS scores. The reliability of the DASS was excellent, and the measure possessed adequate convergent and discriminant validity Conclusions: The DASS is a reliable and valid measure of the constructs it was intended to assess. The utility of this measure for UK clinicians is enhanced by the provision of large sample normative data.

  3. The Best of Both Worlds: Building on the COPUS and RTOP Observation Protocols to Easily and Reliably Measure Various Levels of Reformed Instructional Practice

    PubMed Central

    Lund, Travis J.; Pilarz, Matthew; Velasco, Jonathan B.; Chakraverty, Devasmita; Rosploch, Kaitlyn; Undersander, Molly; Stains, Marilyne

    2015-01-01

    Researchers, university administrators, and faculty members are increasingly interested in measuring and describing instructional practices provided in science, technology, engineering, and mathematics (STEM) courses at the college level. Specifically, there is keen interest in comparing instructional practices between courses, monitoring changes over time, and mapping observed practices to research-based teaching. While increasingly common observation protocols (Reformed Teaching Observation Protocol [RTOP] and Classroom Observation Protocol in Undergraduate STEM [COPUS]) at the postsecondary level help achieve some of these goals, they also suffer from weaknesses that limit their applicability. In this study, we leverage the strengths of these protocols to provide an easy method that enables the reliable and valid characterization of instructional practices. This method was developed empirically via a cluster analysis using observations of 269 individual class periods, corresponding to 73 different faculty members, 28 different research-intensive institutions, and various STEM disciplines. Ten clusters, called COPUS profiles, emerged from this analysis; they represent the most common types of instructional practices enacted in the classrooms observed for this study. RTOP scores were used to validate the alignment of the 10 COPUS profiles with reformed teaching. Herein, we present a detailed description of the cluster analysis method, the COPUS profiles, and the distribution of the COPUS profiles across various STEM courses at research-intensive universities. PMID:25976654

  4. Probabilistic Assessment of High-Throughput Wireless Sensor Networks

    PubMed Central

    Kim, Robin E.; Mechitov, Kirill; Sim, Sung-Han; Spencer, Billie F.; Song, Junho

    2016-01-01

    Structural health monitoring (SHM) using wireless smart sensors (WSS) has the potential to provide rich information on the state of a structure. However, because of their distributed nature, maintaining highly robust and reliable networks can be challenging. Assessing WSS network communication quality before and after finalizing a deployment is critical to achieve a successful WSS network for SHM purposes. Early studies on WSS network reliability mostly used temporal signal indicators, composed of a smaller number of packets, to assess the network reliability. However, because the WSS networks for SHM purpose often require high data throughput, i.e., a larger number of packets are delivered within the communication, such an approach is not sufficient. Instead, in this study, a model that can assess, probabilistically, the long-term performance of the network is proposed. The proposed model is based on readily-available measured data sets that represent communication quality during high-throughput data transfer. Then, an empirical limit-state function is determined, which is further used to estimate the probability of network communication failure. Monte Carlo simulation is adopted in this paper and applied to a small and a full-bridge wireless networks. By performing the proposed analysis in complex sensor networks, an optimized sensor topology can be achieved. PMID:27258270

  5. "vocd": A Theoretical and Empirical Evaluation

    ERIC Educational Resources Information Center

    McCarthy, Philip M.; Jarvis, Scott

    2007-01-01

    A reliable index of lexical diversity (LD) has remained stubbornly elusive for over 60 years. Meanwhile, researchers in fields as varied as "stylistics," "neuropathology," "language acquisition," and even "forensics" continue to use flawed LD indices--often ignorant that their results are questionable and in…

  6. Kerogen maturation and incipient graphitization of hydrocarbon source rocks in the Arkoma Basin, Oklahoma and Arkansas: A combined petrographic and Raman spectrometric study

    USGS Publications Warehouse

    Spotl, C.; Houseknecht, D.W.; Jaques, R.C.

    1998-01-01

    Dispersed kerogen of the Woodford-Chattanooga and Atoka Formations from the subsurface of the Arkoma Basin show a wide range of thermal maturities (0.38 to 6.1% R(o)) indicating thermal conditions ranging from diagenesis to incipient rock metamorphism. Raman spectral analysis reveals systematic changes of both the first- and second-order spectrum with increasing thermal maturity. These changes include a pronounced increase in the D/O peak height ratio accompanied by a narrowing of the D peak, a gradual decrease in the D/O peak width ratio, and a shift of both peaks toward higher wave numbers. Second-order Raman peaks, though less intensive, also show systematic peak shifting as a function of R(o). These empirical results underscore the high potential of Raman spectrometry as a fast and reliable geothermometer of mature to supermature hydrocarbon source rocks, and as an indicator of thermal maturity levels within the anchizone.Dispersed kerogen of the Woodford-Chattanooga and Atoka Formations from the subsurface of the Arkoma Basin show a wide range of thermal maturities (0.38 to 6.1% Ro) indicating thermal conditions ranging from diagenesis to incipient rock metamorphism. Raman spectral analysis reveals systematic changes of both the first- and second-order spectrum with increasing thermal maturity. These changes include a pronounced increase in the D/O peak height ratio accompanied by a narrowing of the D peak, a gradual decrease in the D/O peak width ratio, and a shift of both peaks toward higher wave numbers. Second-order Raman peaks, though less intensive, also show systematic peak shifting as a function of Ro. These empirical results underscore the high potential of Raman spectrometry as a fast and reliable geothermometer of mature to supermature hydrocarbon source rocks, and as an indicator of thermal maturity levels within the anchizone.

  7. Achieving Reliable Communication in Dynamic Emergency Responses

    PubMed Central

    Chipara, Octav; Plymoth, Anders N.; Liu, Fang; Huang, Ricky; Evans, Brian; Johansson, Per; Rao, Ramesh; Griswold, William G.

    2011-01-01

    Emergency responses require the coordination of first responders to assess the condition of victims, stabilize their condition, and transport them to hospitals based on the severity of their injuries. WIISARD is a system designed to facilitate the collection of medical information and its reliable dissemination during emergency responses. A key challenge in WIISARD is to deliver data with high reliability as first responders move and operate in a dynamic radio environment fraught with frequent network disconnections. The initial WIISARD system employed a client-server architecture and an ad-hoc routing protocol was used to exchange data. The system had low reliability when deployed during emergency drills. In this paper, we identify the underlying causes of unreliability and propose a novel peer-to-peer architecture that in combination with a gossip-based communication protocol achieves high reliability. Empirical studies show that compared to the initial WIISARD system, the redesigned system improves reliability by as much as 37% while reducing the number of transmitted packets by 23%. PMID:22195075

  8. Exploring connectivity with large-scale Granger causality on resting-state functional MRI.

    PubMed

    DSouza, Adora M; Abidin, Anas Z; Leistritz, Lutz; Wismüller, Axel

    2017-08-01

    Large-scale Granger causality (lsGC) is a recently developed, resting-state functional MRI (fMRI) connectivity analysis approach that estimates multivariate voxel-resolution connectivity. Unlike most commonly used multivariate approaches, which establish coarse-resolution connectivity by aggregating voxel time-series avoiding an underdetermined problem, lsGC estimates voxel-resolution, fine-grained connectivity by incorporating an embedded dimension reduction. We investigate application of lsGC on realistic fMRI simulations, modeling smoothing of neuronal activity by the hemodynamic response function and repetition time (TR), and empirical resting-state fMRI data. Subsequently, functional subnetworks are extracted from lsGC connectivity measures for both datasets and validated quantitatively. We also provide guidelines to select lsGC free parameters. Results indicate that lsGC reliably recovers underlying network structure with area under receiver operator characteristic curve (AUC) of 0.93 at TR=1.5s for a 10-min session of fMRI simulations. Furthermore, subnetworks of closely interacting modules are recovered from the aforementioned lsGC networks. Results on empirical resting-state fMRI data demonstrate recovery of visual and motor cortex in close agreement with spatial maps obtained from (i) visuo-motor fMRI stimulation task-sequence (Accuracy=0.76) and (ii) independent component analysis (ICA) of resting-state fMRI (Accuracy=0.86). Compared with conventional Granger causality approach (AUC=0.75), lsGC produces better network recovery on fMRI simulations. Furthermore, it cannot recover functional subnetworks from empirical fMRI data, since quantifying voxel-resolution connectivity is not possible as consequence of encountering an underdetermined problem. Functional network recovery from fMRI data suggests that lsGC gives useful insight into connectivity patterns from resting-state fMRI at a multivariate voxel-resolution. Copyright © 2017 Elsevier B.V. All rights reserved.

  9. Measuring teacher self-report on classroom practices: Construct validity and reliability of the Classroom Strategies Scale-Teacher Form.

    PubMed

    Reddy, Linda A; Dudek, Christopher M; Fabiano, Gregory A; Peters, Stephanie

    2015-12-01

    This article presents information about the construct validity and reliability of a new teacher self-report measure of classroom instructional and behavioral practices (the Classroom Strategies Scales-Teacher Form; CSS-T). The theoretical underpinnings and empirical basis for the instructional and behavioral management scales are presented. Information is provided about the construct validity, internal consistency, test-retest reliability, and freedom from item-bias of the scales. Given previous investigations with the CSS Observer Form, it was hypothesized that internal consistency would be adequate and that confirmatory factor analyses (CFA) of CSS-T data from 293 classrooms would offer empirical support for the CSS-T's Total, Composite and subscales, and yield a similar factor structure to that of the CSS Observer Form. Goodness-of-fit indices of χ2/df, Root Mean Square Error of Approximation, Goodness of Fit Index, and Adjusted Goodness of Fit Index suggested satisfactory fit of proposed CFA models whereas the Comparative Fit Index did not. Internal consistency estimates of .93 and .94 were obtained for the Instructional Strategies and Behavioral Strategies Total scales respectively. Adequate test-retest reliability was found for instructional and behavioral total scales (r = .79, r = .84, percent agreement 93% and 93%). The CSS-T evidences freedom from item bias on important teacher demographics (age, educational degree, and years of teaching experience). Implications of results are discussed. (c) 2015 APA, all rights reserved).

  10. A preliminary study of mechanistic approach in pavement design to accommodate climate change effects

    NASA Astrophysics Data System (ADS)

    Harnaeni, S. R.; Pramesti, F. P.; Budiarto, A.; Setyawan, A.

    2018-03-01

    Road damage is caused by some factors, including climate changes, overload, and inappropriate procedure for material and development process. Meanwhile, climate change is a phenomenon which cannot be avoided. The effects observed include air temperature rise, sea level rise, rainfall changes, and the intensity of extreme weather phenomena. Previous studies had shown the impacts of climate changes on road damage. Therefore, several measures to anticipate the damage should be considered during the planning and construction in order to reduce the cost of road maintenance. There are three approaches generally applied in the design of flexible pavement thickness, namely mechanistic approach, mechanistic-empirical (ME) approach and empirical approach. The advantages of applying mechanistic approach or mechanistic-empirical (ME) approaches are its efficiency and reliability in the design of flexible pavement thickness as well as its capacity to accommodate climate changes in compared to empirical approach. However, generally, the design of flexible pavement thickness in Indonesia still applies empirical approach. This preliminary study aimed to emphasize the importance of the shifting towards a mechanistic approach in the design of flexible pavement thickness.

  11. Human Factors in Financial Trading

    PubMed Central

    Leaver, Meghan; Reader, Tom W.

    2016-01-01

    Objective This study tests the reliability of a system (FINANS) to collect and analyze incident reports in the financial trading domain and is guided by a human factors taxonomy used to describe error in the trading domain. Background Research indicates the utility of applying human factors theory to understand error in finance, yet empirical research is lacking. We report on the development of the first system for capturing and analyzing human factors–related issues in operational trading incidents. Method In the first study, 20 incidents are analyzed by an expert user group against a referent standard to establish the reliability of FINANS. In the second study, 750 incidents are analyzed using distribution, mean, pathway, and associative analysis to describe the data. Results Kappa scores indicate that categories within FINANS can be reliably used to identify and extract data on human factors–related problems underlying trading incidents. Approximately 1% of trades (n = 750) lead to an incident. Slip/lapse (61%), situation awareness (51%), and teamwork (40%) were found to be the most common problems underlying incidents. For the most serious incidents, problems in situation awareness and teamwork were most common. Conclusion We show that (a) experts in the trading domain can reliably and accurately code human factors in incidents, (b) 1% of trades incur error, and (c) poor teamwork skills and situation awareness underpin the most critical incidents. Application This research provides data crucial for ameliorating risk within financial trading organizations, with implications for regulation and policy. PMID:27142394

  12. Mixed Phylogenetic Signal in Fish Toxicity Data across Chemical Classes

    EPA Science Inventory

    Chemical use in society is growing rapidly and is one of the five major pressures on biodiversity worldwide. Since empirical toxicity studies of pollutants generally focus on a handful of model organisms, reliable approaches are needed to assess sensitivity to chemicals across th...

  13. Sex Differences in Locus of Control, Helplessness, Hopelessness, and Depression.

    ERIC Educational Resources Information Center

    Kolotkin, Richard A.; And Others

    This experiment investigated: (1) relationships among locus of control, attributional style, and depression; (2) if a depressogenic attributional style could be empirically isolated; and (3) if reliable relationships existed between attribution and depression when depression was operationalized using different instruments. Subjects completed the…

  14. An Empirical Framework for ePortfolio Assessment

    ERIC Educational Resources Information Center

    Kelly-Riley, Diane; Elliot, Norbert; Rudniy, Alex

    2016-01-01

    This research focuses on ePortfolio assessment strategies that yield important accountability and reporting information. Under foundational categories of reliability, validity, and fairness, we present methods of gathering evidence from ePortfolio scores and their relationship to demographic information (gender, race/ethnicity, and socio-economic…

  15. Empirical Evidence for Childhood Depression.

    ERIC Educational Resources Information Center

    Lachar, David

    Although several theoretical positions deal with the concept of childhood depression, accurate measurement of depression can only occur if valid and reliable measures are available. Current efforts emphasize direct questioning of the child and quantification of parents' observations. One scale used to study childhood depression, the Personality…

  16. Brain Migration Revisited

    ERIC Educational Resources Information Center

    Vinokur, Annie

    2006-01-01

    The "brain drain/brain gain" debate has been going on for the past 40 years, with irresolvable theoretical disputes and unenforceable policy recommendations that economists commonly ascribe to the lack of reliable empirical data. The recent report of the World Bank, "International migration, remittances and the brain drain", documents the…

  17. Five systems of psychiatric classification for preschool children: do differences in validity, usefulness and reliability make for competitive or complimentary constellations?

    PubMed

    Postert, Christian; Averbeck-Holocher, Marlies; Beyer, Thomas; Müller, Jörg; Furniss, Tilman

    2009-03-01

    DSM-IV and ICD-10 have limitations in the diagnostic classification of psychiatric disorders at preschool age (0-5 years). The publication of the Diagnostic Classification 0-3 (DC:0-3) in 1994, its basically revised second edition (DC:0-3R) in 2005 and the Research Diagnostic Criteria-Preschool Age (RDC-PA) in 2004 have provided several modifications of these manuals. Taking into account the growing empirical evidence highlighting the need for a diagnostic classification system for psychiatric disorders in preschool children, the main categorical classification systems in preschool psychiatry will be presented and discussed. The paper will focus on issues of validity, usefulness and reliability in DSM-IV, ICD-10, RDC-PA, DC:0-3, and DC:0-3R. The reasons for including or excluding postulated psychiatric disorder categories for preschool children with variable degrees of empirical evidence into the different diagnostic systems will be discussed.

  18. The influence of tie strength on evolutionary games on networks: An empirical investigation

    NASA Astrophysics Data System (ADS)

    Buesser, Pierre; Peña, Jorge; Pestelacci, Enea; Tomassini, Marco

    2011-11-01

    Extending previous work on unweighted networks, we present here a systematic numerical investigation of standard evolutionary games on weighted networks. In the absence of any reliable model for generating weighted social networks, we attribute weights to links in a few ways supported by empirical data ranging from totally uncorrelated to weighted bipartite networks. The results of the extensive simulation work on standard complex network models show that, except in a case that does not seem to be common in social networks, taking the tie strength into account does not change in a radical manner the long-run steady-state behavior of the studied games. Besides model networks, we also included a real-life case drawn from a coauthorship network. In this case also, taking the weights into account only changes the results slightly with respect to the raw unweighted graph, although to draw more reliable conclusions on real social networks many more cases should be studied as these weighted networks become available.

  19. Screening utility of the social anxiety screening scale in Spanish speaking adolescents.

    PubMed

    Piqueras, José Antonio; Olivares, José; Hidalgo, María Dolores

    2012-07-01

    The aim of this study was to analyse the screening utility of the Social Anxiety Screening Scale (SASS/EDAS) in a sample of 227 adolescents with social anxiety disorder and 156 Without it (14-17 years). Results showed that the EDAS subscales (Avoidance, Distress and Interference) scores were reliable in terms of internal consistency (alpha > .80). All the subscales discriminated between adolescents with and without the disorder. They also showed a positive and significant correlation with other empirically validated measures of social anxiety. The three subscales indicated relevant sensitivity (69.16-84.14%), specificity (63.46-66.03%) and areas under the curve (.74-.81%). Binary logistic regression analyses indicated the adequate predictive utility of EDAS subscales, with the Distress subscale as the best diagnostic predictor. The data provide empirical evidence of the usefulness of EDAS as a screener for adolescent social anxiety disorder in terms of reliability, convergent and discriminant validity, diagnostic accuracy and clinical usefulness.

  20. Cross-cultural adaptation and construct validity of the Korean version of a physical activity measure for community-dwelling elderly.

    PubMed

    Choi, Bongsam

    2018-01-01

    [Purpose] This study aimed to cross-cultural adapt and validate the Korean version of an physical activity measure (K-PAM) for community-dwelling elderly. [Subjects and Methods] One hundred and thirty eight community-dwelling elderlies, 32 males and 106 female, participated in the study. All participants were asked to fill out a fifty-one item questionnaire measuring perceived difficulty in the activities of daily living (ADL) for the elderly. One-parameter model of item response theory (Rasch analysis) was applied to determine the construct validity and to inspect item-level psychometric properties of 51 ADL items of the K-PAM. [Results] Person separation reliability (analogous to Cronbach's alpha) for internal consistency was ranging 0.93 to 0.94. A total of 16 items was misfit to the Rasch model. After misfit item deletion, 35 ADL items of the K-PAM were placed in an empirically meaningful hierarchy from easy to hard. The item-person map analysis delineated that the item difficulty was well matched for the elderlies with moderate and low ability except for high ceilings. [Conclusion] Cross-cultural adapted K-PAM was shown to be sufficient for establishing construct validity and stable psychometric properties confirmed by person separation reliability and fit statistics.

  1. Improving 1D Site Specific Velocity Profiles for the Kik-Net Network

    NASA Astrophysics Data System (ADS)

    Holt, James; Edwards, Benjamin; Pilz, Marco; Fäh, Donat; Rietbrock, Andreas

    2017-04-01

    Ground motion predication equations (GMPEs) form the cornerstone of modern seismic hazard assessments. When produced to a high standard they provide reliable estimates of ground motion/spectral acceleration for a given site and earthquake scenario. This information is crucial for engineers to optimise design and for regulators who enforce legal minimum safe design capacities. Classically, GMPEs were built upon the assumption that variability around the median model could be treated as aleatory. As understanding improved, it was noted that the propagation could be segregated into the response of the average path from the source and the response of the site. This is because the heterogeneity of the near-surface lithology is significantly different from that of the bulk path. It was then suggested that the semi-ergodic approach could be taken if the site response could be determined, moving uncertainty away from aleatory to epistemic. The determination of reliable site-specific response models is therefore becoming increasingly critical for ground motion models used in engineering practice. Today it is common practice to include proxies for site response within the scope of a GMPE, such as Vs30 or site classification, in an effort to reduce the overall uncertainty of the predication at a given site. However, these proxies are not always reliable enough to give confident ground motion estimates, due to the complexity of the near-surface. Other approaches of quantifying the response of the site include detailed numerical simulations (1/2/3D - linear, EQL, non-linear etc.). However, in order to be reliable, they require highly detailed and accurate velocity and, for non-linear analyses, material property models. It is possible to obtain this information through invasive methods, but is expensive, and not feasible for most projects. Here we propose an alternative method to derive reliable velocity profiles (and their uncertainty), calibrated using almost 20 years of recorded data from the Kik-Net network. First, using a reliable subset of sites, the empirical surface to borehole (S/B) ratio is calculated in the frequency domain using all events recorded at that site. In a subsequent step, we use numerical simulation to produce 1D SH transfer function curves using a suite of stochastic velocity models. Comparing the resulting amplification with the empirical S/B ratio we find optimal 1D velocity models and their uncertainty. The method will be tested to determine the level of initial information required to obtain a reliable Vs profile (e.g., starting Vs model, only Vs30, site-class, H/V ratio etc.) and then applied and tested against data from other regions using site-to-reference or empirical spectral model amplification.

  2. Time irreversibility and intrinsics revealing of series with complex network approach

    NASA Astrophysics Data System (ADS)

    Xiong, Hui; Shang, Pengjian; Xia, Jianan; Wang, Jing

    2018-06-01

    In this work, we analyze time series on the basis of the visibility graph algorithm that maps the original series into a graph. By taking into account the all-round information carried by the signals, the time irreversibility and fractal behavior of series are evaluated from a complex network perspective, and considered signals are further classified from different aspects. The reliability of the proposed analysis is supported by numerical simulations on synthesized uncorrelated random noise, short-term correlated chaotic systems and long-term correlated fractal processes, and by the empirical analysis on daily closing prices of eleven worldwide stock indices. Obtained results suggest that finite size has a significant effect on the evaluation, and that there might be no direct relation between the time irreversibility and long-range correlation of series. Similarity and dissimilarity between stock indices are also indicated from respective regional and global perspectives, showing the existence of multiple features of underlying systems.

  3. Making big data useful for health care: a summary of the inaugural mit critical data conference.

    PubMed

    Badawi, Omar; Brennan, Thomas; Celi, Leo Anthony; Feng, Mengling; Ghassemi, Marzyeh; Ippolito, Andrea; Johnson, Alistair; Mark, Roger G; Mayaud, Louis; Moody, George; Moses, Christopher; Naumann, Tristan; Pimentel, Marco; Pollard, Tom J; Santos, Mauro; Stone, David J; Zimolzak, Andrew

    2014-08-22

    With growing concerns that big data will only augment the problem of unreliable research, the Laboratory of Computational Physiology at the Massachusetts Institute of Technology organized the Critical Data Conference in January 2014. Thought leaders from academia, government, and industry across disciplines-including clinical medicine, computer science, public health, informatics, biomedical research, health technology, statistics, and epidemiology-gathered and discussed the pitfalls and challenges of big data in health care. The key message from the conference is that the value of large amounts of data hinges on the ability of researchers to share data, methodologies, and findings in an open setting. If empirical value is to be from the analysis of retrospective data, groups must continuously work together on similar problems to create more effective peer review. This will lead to improvement in methodology and quality, with each iteration of analysis resulting in more reliability.

  4. Adaptive filtering and maximum entropy spectra with application to changes in atmospheric angular momentum

    NASA Technical Reports Server (NTRS)

    Penland, Cecile; Ghil, Michael; Weickmann, Klaus M.

    1991-01-01

    The spectral resolution and statistical significance of a harmonic analysis obtained by low-order MEM can be improved by subjecting the data to an adaptive filter. This adaptive filter consists of projecting the data onto the leading temporal empirical orthogonal functions obtained from singular spectrum analysis (SSA). The combined SSA-MEM method is applied both to a synthetic time series and a time series of AAM data. The procedure is very effective when the background noise is white and less so when the background noise is red. The latter case obtains in the AAM data. Nevertheless, reliable evidence for intraseasonal and interannual oscillations in AAM is detected. The interannual periods include a quasi-biennial one and an LF one, of 5 years, both related to the El Nino/Southern Oscillation. In the intraseasonal band, separate oscillations of about 48.5 and 51 days are ascertained.

  5. A measurement model for real estate bubble size based on the panel data analysis: An empirical case study

    PubMed Central

    Liu, Fengyun; Liu, Deqiang; Malekian, Reza; Li, Zhixiong; Wang, Deqing

    2017-01-01

    Employing the fundamental value of real estate determined by the economic fundamentals, a measurement model for real estate bubble size is established based on the panel data analysis. Using this model, real estate bubble sizes in various regions in Japan in the late 1980s and in recent China are examined. Two panel models for Japan provide results, which are consistent with the reality in the 1980s where a commercial land price bubble appeared in most area and was much larger than that of residential land. This provides evidence of the reliability of our model, overcoming the limit of existing literature with this method. The same models for housing prices in China at both the provincial and city levels show that contrary to the concern of serious housing price bubble in China, over-valuing in recent China is much smaller than that in 1980s Japan. PMID:28273141

  6. Preliminary study of soil permeability properties using principal component analysis

    NASA Astrophysics Data System (ADS)

    Yulianti, M.; Sudriani, Y.; Rustini, H. A.

    2018-02-01

    Soil permeability measurement is undoubtedly important in carrying out soil-water research such as rainfall-runoff modelling, irrigation water distribution systems, etc. It is also known that acquiring reliable soil permeability data is rather laborious, time-consuming, and costly. Therefore, it is desirable to develop the prediction model. Several studies of empirical equations for predicting permeability have been undertaken by many researchers. These studies derived the models from areas which soil characteristics are different from Indonesian soil, which suggest a possibility that these permeability models are site-specific. The purpose of this study is to identify which soil parameters correspond strongly to soil permeability and propose a preliminary model for permeability prediction. Principal component analysis (PCA) was applied to 16 parameters analysed from 37 sites consist of 91 samples obtained from Batanghari Watershed. Findings indicated five variables that have strong correlation with soil permeability, and we recommend a preliminary permeability model, which is potential for further development.

  7. A measurement model for real estate bubble size based on the panel data analysis: An empirical case study.

    PubMed

    Liu, Fengyun; Liu, Deqiang; Malekian, Reza; Li, Zhixiong; Wang, Deqing

    2017-01-01

    Employing the fundamental value of real estate determined by the economic fundamentals, a measurement model for real estate bubble size is established based on the panel data analysis. Using this model, real estate bubble sizes in various regions in Japan in the late 1980s and in recent China are examined. Two panel models for Japan provide results, which are consistent with the reality in the 1980s where a commercial land price bubble appeared in most area and was much larger than that of residential land. This provides evidence of the reliability of our model, overcoming the limit of existing literature with this method. The same models for housing prices in China at both the provincial and city levels show that contrary to the concern of serious housing price bubble in China, over-valuing in recent China is much smaller than that in 1980s Japan.

  8. Development and validation of the Alcohol Myopia Scale.

    PubMed

    Lac, Andrew; Berger, Dale E

    2013-09-01

    Alcohol myopia theory conceptualizes the ability of alcohol to narrow attention and how this demand on mental resources produces the impairments of self-inflation, relief, and excess. The current research was designed to develop and validate a scale based on this framework. People who were alcohol users rated items representing myopic experiences arising from drinking episodes in the past month. In Study 1 (N = 260), the preliminary 3-factor structure was supported by exploratory factor analysis. In Study 2 (N = 289), the 3-factor structure was substantiated with confirmatory factor analysis, and it was superior in fit to an empirically indefensible 1-factor structure. The final 14-item scale was evaluated with internal consistency reliability, discriminant validity, convergent validity, criterion validity, and incremental validity. The alcohol myopia scale (AMS) illuminates conceptual underpinnings of this theory and yields insights for understanding the tunnel vision that arises from intoxication.

  9. Making Big Data Useful for Health Care: A Summary of the Inaugural MIT Critical Data Conference

    PubMed Central

    2014-01-01

    With growing concerns that big data will only augment the problem of unreliable research, the Laboratory of Computational Physiology at the Massachusetts Institute of Technology organized the Critical Data Conference in January 2014. Thought leaders from academia, government, and industry across disciplines—including clinical medicine, computer science, public health, informatics, biomedical research, health technology, statistics, and epidemiology—gathered and discussed the pitfalls and challenges of big data in health care. The key message from the conference is that the value of large amounts of data hinges on the ability of researchers to share data, methodologies, and findings in an open setting. If empirical value is to be from the analysis of retrospective data, groups must continuously work together on similar problems to create more effective peer review. This will lead to improvement in methodology and quality, with each iteration of analysis resulting in more reliability. PMID:25600172

  10. Rethinking big data: A review on the data quality and usage issues

    NASA Astrophysics Data System (ADS)

    Liu, Jianzheng; Li, Jie; Li, Weifeng; Wu, Jiansheng

    2016-05-01

    The recent explosive publications of big data studies have well documented the rise of big data and its ongoing prevalence. Different types of ;big data; have emerged and have greatly enriched spatial information sciences and related fields in terms of breadth and granularity. Studies that were difficult to conduct in the past time due to data availability can now be carried out. However, big data brings lots of ;big errors; in data quality and data usage, which cannot be used as a substitute for sound research design and solid theories. We indicated and summarized the problems faced by current big data studies with regard to data collection, processing and analysis: inauthentic data collection, information incompleteness and noise of big data, unrepresentativeness, consistency and reliability, and ethical issues. Cases of empirical studies are provided as evidences for each problem. We propose that big data research should closely follow good scientific practice to provide reliable and scientific ;stories;, as well as explore and develop techniques and methods to mitigate or rectify those 'big-errors' brought by big data.

  11. Distinguishing between forensic science and forensic pseudoscience: testing of validity and reliability, and approaches to forensic voice comparison.

    PubMed

    Morrison, Geoffrey Stewart

    2014-05-01

    In this paper it is argued that one should not attempt to directly assess whether a forensic analysis technique is scientifically acceptable. Rather one should first specify what one considers to be appropriate principles governing acceptable practice, then consider any particular approach in light of those principles. This paper focuses on one principle: the validity and reliability of an approach should be empirically tested under conditions reflecting those of the case under investigation using test data drawn from the relevant population. Versions of this principle have been key elements in several reports on forensic science, including forensic voice comparison, published over the last four-and-a-half decades. The aural-spectrographic approach to forensic voice comparison (also known as "voiceprint" or "voicegram" examination) and the currently widely practiced auditory-acoustic-phonetic approach are considered in light of this principle (these two approaches do not appear to be mutually exclusive). Approaches based on data, quantitative measurements, and statistical models are also considered in light of this principle. © 2013.

  12. Detection Copy Number Variants from NGS with Sparse and Smooth Constraints.

    PubMed

    Zhang, Yue; Cheung, Yiu-Ming; Xu, Bo; Su, Weifeng

    2017-01-01

    It is known that copy number variations (CNVs) are associated with complex diseases and particular tumor types, thus reliable identification of CNVs is of great potential value. Recent advances in next generation sequencing (NGS) data analysis have helped manifest the richness of CNV information. However, the performances of these methods are not consistent. Reliably finding CNVs in NGS data in an efficient way remains a challenging topic, worthy of further investigation. Accordingly, we tackle the problem by formulating CNVs identification into a quadratic optimization problem involving two constraints. By imposing the constraints of sparsity and smoothness, the reconstructed read depth signal from NGS is anticipated to fit the CNVs patterns more accurately. An efficient numerical solution tailored from alternating direction minimization (ADM) framework is elaborated. We demonstrate the advantages of the proposed method, namely ADM-CNV, by comparing it with six popular CNV detection methods using synthetic, simulated, and empirical sequencing data. It is shown that the proposed approach can successfully reconstruct CNV patterns from raw data, and achieve superior or comparable performance in detection of the CNVs compared to the existing counterparts.

  13. Should the SCOPA-COG be modified? A Rasch analysis perspective.

    PubMed

    Forjaz, M J; Frades-Payo, B; Rodriguez-Blazquez, C; Ayala, A; Martinez-Martin, P

    2010-02-01

    The SCales for Outcomes in PArkinson's disease-Cognition (SCOPA-COG) is a specific measure of cognitive function for Parkinson's disease (PD) patients. Previous studies, under the frame of the classic test theory, indicate satisfactory psychometric properties. The Rasch model, an item response theory approach, provides new information about the scale, as well as results in a linear scale. This study aims at analysing the SCOPA-COG according to the Rasch model and, on the basis of results, suggesting modification to the SCOPA-COG. Fit to the Rasch model was analysed using a sample of 384 PD patients. A good fit was obtained after rescoring for disordered thresholds. The person separation index, a reliability measure, was 0.83. Differential item functioning was observed by age for three items and by gender for one item. The SCOPA-COG is a unidimensional measure of global cognitive function in PD patients, with good scale targeting and no empirical evidence for use of the subscale scores. Its adequate reliability and internal construct validity were supported. The SCOPA-COG, with the proposed scoring scheme, generates true linear interval scores.

  14. [Job rotation in anaesthesiological care: impacts on knowledge and learning].

    PubMed

    Jordan, Silvia; Brauner, Elisabeth

    2008-04-01

    Job rotation in anaesthesiological care is explored in this article. Based on discussions of 'high reliability organizations', and based on current theory in social and organizational psychology, we empirically investigated the impact of job rotation on knowledge management and learning. The study was conducted at the University Hospital of Innsbruck during the job training period of novice nurse anaesthetists. Qualitative interviews were conducted as well as a questionnaire administered. Data were collected between September 2003 and June 2005. Observational and interview data were combined with a questionnaire study to assess transactive memory and cooperation using a mixed-method design. Qualitative data were analyzed using GABEK, a computer-supported content analysis system; quantitative data were analyzed statistically with SPSS. Results indicate that job rotation had both positive and negative effects on knowledge and learning. On the one hand, job rotation can foster flexibility and awareness of a person's own fallibility and lack of knowledge. On the other hand, trust, group cohesion, and social meta-knowledge can be hampered by constant rotation. Consequently, stabilization and change need equal consideration when designing work in high reliability organizations.

  15. Reliability of sensor-based real-time workflow recognition in laparoscopic cholecystectomy.

    PubMed

    Kranzfelder, Michael; Schneider, Armin; Fiolka, Adam; Koller, Sebastian; Reiser, Silvano; Vogel, Thomas; Wilhelm, Dirk; Feussner, Hubertus

    2014-11-01

    Laparoscopic cholecystectomy is a very common minimally invasive surgical procedure that may be improved by autonomous or cooperative assistance support systems. Model-based surgery with a precise definition of distinct procedural tasks (PT) of the operation was implemented and tested to depict and analyze the process of this procedure. Reliability of real-time workflow recognition in laparoscopic cholecystectomy ([Formula: see text] cases) was evaluated by continuous sensor-based data acquisition. Ten PTs were defined including begin/end preparation calots' triangle, clipping/cutting cystic artery and duct, begin/end gallbladder dissection, begin/end hemostasis, gallbladder removal, and end of operation. Data acquisition was achieved with continuous instrument detection, room/table light status, intra-abdominal pressure, table tilt, irrigation/aspiration volume and coagulation/cutting current application. Two independent observers recorded start and endpoint of each step by analysis of the sensor data. The data were cross-checked with laparoscopic video recordings serving as gold standard for PT identification. Bland-Altman analysis revealed for 95% of cases a difference of annotation results within the limits of agreement ranging from [Formula: see text]309 s (PT 7) to +368 s (PT 5). Laparoscopic video and sensor data matched to a greater or lesser extent within the different procedural tasks. In the majority of cases, the observer results exceeded those obtained from the laparoscopic video. Empirical knowledge was required to detect phase transit. A set of sensors used to monitor laparoscopic cholecystectomy procedures was sufficient to enable expert observers to reliably identify each PT. In the future, computer systems may automate the task identification process provided a more robust data inflow is available.

  16. Validation of the Spanish version of the Amsterdam Preoperative Anxiety and Information Scale (APAIS).

    PubMed

    Vergara-Romero, Manuel; Morales-Asencio, José Miguel; Morales-Fernández, Angelines; Canca-Sanchez, Jose Carlos; Rivas-Ruiz, Francisco; Reinaldo-Lapuerta, Jose Antonio

    2017-06-07

    Preoperative anxiety is a frequent and challenging problem with deleterious effects on the development of surgical procedures and postoperative outcomes. To prevent and treat preoperative anxiety effectively, the level of anxiety of patients needs to be assessed through valid and reliable measuring instruments. One such measurement tool is the Amsterdam Preoperative Anxiety and Information Scale (APAIS), of which a Spanish version has not been validated yet. To perform a Spanish cultural adaptation and empirical validation of the APAIS for assessing preoperative anxiety in the Spanish population. A two-step forward/back translation of the APAIS scale was performed to ensure a reliable Spanish cultural adaptation. The final Spanish version of the APAIS questionnaire was administered to 529 patients between the ages of 18 to 70 undergoing elective surgery at hospitals of the Agencia Sanitaria Costa del Sol (Spain). Cronbach's alpha, homogeneity index, intra-class correlation coefficient, and confirmatory factor analysis were calculated to assess internal consistency and criteria and construct validity. Confirmatory factor analysis showed that a one-factor model was better fitted than a two-factor model, with good fitting patterns (root mean square error of approximation: 0.05, normed-fit index: 0.99, goodness-of-fit statistic: 0.99). The questionnaire showed high internal consistency (Cronbach's alpha: 0.84) and a good correlation with the Goldberg Anxiety Scale (CCI: 0.62 (95% CI: 0.55 to 0.68). The Spanish version of the APAIS is a valid and reliable preoperative anxiety measurement tool and shows psychometric properties similar to those obtained by similar previous studies.

  17. Quantitative analysis of the rubric as an assessment tool: an empirical study of student peer-group rating

    NASA Astrophysics Data System (ADS)

    Hafner, John C.; Hafner, Patti M.

    2003-12-01

    Although the rubric has emerged as one of the most popular assessment tools in progressive educational programs, there is an unfortunate dearth of information in the literature quantifying the actual effectiveness of the rubric as an assessment tool in the hands of the students. This study focuses on the validity and reliability of the rubric as an assessment tool for student peer-group evaluation in an effort to further explore the use and effectiveness of the rubric. A total of 1577 peer-group ratings using a rubric for an oral presentation was used in this 3-year study involving 107 college biology students. A quantitative analysis of the rubric used in this study shows that it is used consistently by both students and the instructor across the study years. Moreover, the rubric appears to be 'gender neutral' and the students' academic strength has no significant bearing on the way that they employ the rubric. A significant, one-to-one relationship (slope = 1.0) between the instructor's assessment and the students' rating is seen across all years using the rubric. A generalizability study yields estimates of inter-rater reliability of moderate values across all years and allows for the estimation of variance components. Taken together, these data indicate that the general form and evaluative criteria of the rubric are clear and that the rubric is a useful assessment tool for peer-group (and self-) assessment by students. To our knowledge, these data provide the first statistical documentation of the validity and reliability of the rubric for student peer-group assessment.

  18. The PMA Scale: A Measure of Physicians' Motivation to Adopt Medical Devices.

    PubMed

    Hatz, Maximilian H M; Sonnenschein, Tim; Blankart, Carl Rudolf

    2017-04-01

    Studies have often stated that individual-level determinants are important drivers for the adoption of medical devices. Empirical evidence supporting this claim is, however, scarce. At the individual level, physicians' adoption motivation was often considered important in the context of adoption decisions, but a clear notion of its dimensions and corresponding measurement scales is not available. To develop and subsequently validate a scale to measure the motivation to adopt medical devices of hospital-based physicians. The development and validation of the physician-motivation-adoption (PMA) scale were based on a literature search, internal expert meetings, a pilot study with physicians, and a three-stage online survey. The data collected in the online survey were analyzed using exploratory factor analysis (EFA), and the PMA scale was revised according to the results. Confirmatory factor analysis (CFA) was conducted to test the results from the EFA in the third stage. Reliability and validity tests and subgroup analyses were also conducted. Overall, 457 questionnaires were completed by medical personnel of the National Health Service England. The EFA favored a six-factor solution to appropriately describe physicians' motivation. The CFA confirmed the results from the EFA. Our tests indicated good reliability and validity of the PMA scale. This is the first reliable and valid scale to measure physicians' adoption motivation. Future adoption studies assessing the individual level should include the PMA scale to obtain more information about the role of physicians' motivation in the broader adoption context. Copyright © 2017 International Society for Pharmacoeconomics and Outcomes Research (ISPOR). Published by Elsevier Inc. All rights reserved.

  19. The analysis of predictability of recent alpha decay formulae and the alpha partial half-lives of some exotic nuclei

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Dasgupta-Schubert, N.; Reyes, M. A.; Tamez, V. A.

    2009-04-20

    Alpha decay is one of the two main decay modes of the heaviest nuclei, (SHE), and constitutes one of the dominant decay modes of highly neutron deficient medium mass nuclei ('exotics'). Thus identifying and characterizing the alpha decay chains form a crucial part of the identification of SHE. We report the extension of the previously developed method for the detailed and systematic investigation of the reliability of the three main extant analytical formulae of alpha decay half-lives: the generalized liquid drop model based formula of Royer et al. (FR), the Sobiczewski modified semi-empirical Viola-Seaborg formula (VSS) and the recent phenomenologicalmore » formula of Sobiczewski and Parkhomenko (SP)« less

  20. Examining perceptions of academic stress and its sources among university students: The Perception of Academic Stress Scale

    PubMed Central

    Bedewy, Dalia

    2015-01-01

    The development of a scale to measure perceived sources of academic stress among university students. Based on empirical evidence and recent literature review, we developed an 18-item scale to measure perceptions of academic stress and its sources. Experts (n = 12) participated in the content validation process of the instrument before it was administered to (n = 100) students. The developed instrument has internal consistency reliability of 0.7 (Cronbach’s alpha), there was evidence for content validity, and factor analysis resulted in four correlated and theoretically meaningful factors. We developed and tested a scale to measure academic stress and its sources. This scale takes 5 minutes to complete. PMID:28070363

  1. The multiaxial system of DSM-III: where did it come from and where should it go? II. Empirical studies, innovations, and recommendations.

    PubMed

    Williams, J B

    1985-02-01

    In a companion report, the development of the DSM-III multiaxial system was described, as well as critiques that have been made of each of its five axes. In this report, empirical studies that provide information about the reliability and validity of the system are reviewed. In addition, suggestions that have been made for changes in the system are critically examined. Finally, this material is synthesized into specific proposals for improvements in the multiaxial system that will be considered by the recently constituted Work Group to Revise DSM-III.

  2. Measuring and Monitoring Conceptions of Research

    ERIC Educational Resources Information Center

    Zhang, Ran; Zwaal, Wichard; Otting, Hans

    2018-01-01

    This study assessed the validity and reliability of the Meyer, Shanahan, and Laugksch's Conceptions of Research Inventory using data collected from 227 undergraduate hotel management students in the Netherlands. The results of a series of exploratory and confirmatory factor analyses showed substantial empirical support for the five-factor…

  3. Comparison of ensemble post-processing approaches, based on empirical and dynamical error modelisation of rainfall-runoff model forecasts

    NASA Astrophysics Data System (ADS)

    Chardon, J.; Mathevet, T.; Le Lay, M.; Gailhard, J.

    2012-04-01

    In the context of a national energy company (EDF : Electricité de France), hydro-meteorological forecasts are necessary to ensure safety and security of installations, meet environmental standards and improve water ressources management and decision making. Hydrological ensemble forecasts allow a better representation of meteorological and hydrological forecasts uncertainties and improve human expertise of hydrological forecasts, which is essential to synthesize available informations, coming from different meteorological and hydrological models and human experience. An operational hydrological ensemble forecasting chain has been developed at EDF since 2008 and is being used since 2010 on more than 30 watersheds in France. This ensemble forecasting chain is characterized ensemble pre-processing (rainfall and temperature) and post-processing (streamflow), where a large human expertise is solicited. The aim of this paper is to compare 2 hydrological ensemble post-processing methods developed at EDF in order improve ensemble forecasts reliability (similar to Monatanari &Brath, 2004; Schaefli et al., 2007). The aim of the post-processing methods is to dress hydrological ensemble forecasts with hydrological model uncertainties, based on perfect forecasts. The first method (called empirical approach) is based on a statistical modelisation of empirical error of perfect forecasts, by streamflow sub-samples of quantile class and lead-time. The second method (called dynamical approach) is based on streamflow sub-samples of quantile class and streamflow variation, and lead-time. On a set of 20 watersheds used for operational forecasts, results show that both approaches are necessary to ensure a good post-processing of hydrological ensemble, allowing a good improvement of reliability, skill and sharpness of ensemble forecasts. The comparison of the empirical and dynamical approaches shows the limits of the empirical approach which is not able to take into account hydrological dynamic and processes, i. e. sample heterogeneity. For a same streamflow range corresponds different processes such as rising limbs or recession, where uncertainties are different. The dynamical approach improves reliability, skills and sharpness of forecasts and globally reduces confidence intervals width. When compared in details, the dynamical approach allows a noticeable reduction of confidence intervals during recessions where uncertainty is relatively lower and a slight increase of confidence intervals during rising limbs or snowmelt where uncertainty is greater. The dynamic approach, validated by forecaster's experience that considered the empirical approach not discriminative enough, improved forecaster's confidence and communication of uncertainties. Montanari, A. and Brath, A., (2004). A stochastic approach for assessing the uncertainty of rainfall-runoff simulations. Water Resources Research, 40, W01106, doi:10.1029/2003WR002540. Schaefli, B., Balin Talamba, D. and Musy, A., (2007). Quantifying hydrological modeling errors through a mixture of normal distributions. Journal of Hydrology, 332, 303-315.

  4. Advancing the detection of steady-state visual evoked potentials in brain-computer interfaces.

    PubMed

    Abu-Alqumsan, Mohammad; Peer, Angelika

    2016-06-01

    Spatial filtering has proved to be a powerful pre-processing step in detection of steady-state visual evoked potentials and boosted typical detection rates both in offline analysis and online SSVEP-based brain-computer interface applications. State-of-the-art detection methods and the spatial filters used thereby share many common foundations as they all build upon the second order statistics of the acquired Electroencephalographic (EEG) data, that is, its spatial autocovariance and cross-covariance with what is assumed to be a pure SSVEP response. The present study aims at highlighting the similarities and differences between these methods. We consider the canonical correlation analysis (CCA) method as a basis for the theoretical and empirical (with real EEG data) analysis of the state-of-the-art detection methods and the spatial filters used thereby. We build upon the findings of this analysis and prior research and propose a new detection method (CVARS) that combines the power of the canonical variates and that of the autoregressive spectral analysis in estimating the signal and noise power levels. We found that the multivariate synchronization index method and the maximum contrast combination method are variations of the CCA method. All three methods were found to provide relatively unreliable detections in low signal-to-noise ratio (SNR) regimes. CVARS and the minimum energy combination methods were found to provide better estimates for different SNR levels. Our theoretical and empirical results demonstrate that the proposed CVARS method outperforms other state-of-the-art detection methods when used in an unsupervised fashion. Furthermore, when used in a supervised fashion, a linear classifier learned from a short training session is able to estimate the hidden user intention, including the idle state (when the user is not attending to any stimulus), rapidly, accurately and reliably.

  5. A stochastic evolutionary model generating a mixture of exponential distributions

    NASA Astrophysics Data System (ADS)

    Fenner, Trevor; Levene, Mark; Loizou, George

    2016-02-01

    Recent interest in human dynamics has stimulated the investigation of the stochastic processes that explain human behaviour in various contexts, such as mobile phone networks and social media. In this paper, we extend the stochastic urn-based model proposed in [T. Fenner, M. Levene, G. Loizou, J. Stat. Mech. 2015, P08015 (2015)] so that it can generate mixture models, in particular, a mixture of exponential distributions. The model is designed to capture the dynamics of survival analysis, traditionally employed in clinical trials, reliability analysis in engineering, and more recently in the analysis of large data sets recording human dynamics. The mixture modelling approach, which is relatively simple and well understood, is very effective in capturing heterogeneity in data. We provide empirical evidence for the validity of the model, using a data set of popular search engine queries collected over a period of 114 months. We show that the survival function of these queries is closely matched by the exponential mixture solution for our model.

  6. An empirical study of experiential value and lifestyles and their effects on satisfaction in adolescents: an example using online gaming.

    PubMed

    Shieh, Kwei-Fen; Cheng, Ming-Sung

    2007-01-01

    This study tests a consumer behavioral model on Taiwanese adolescents and young adults engaging in online gaming. The major focus is on how these consumers transfer the value of their experiences and lifestyles to satisfaction, which may assist in the further exploration of the specific consumption behavior characteristics of adolescents and young adults, particularly with regard to their social functioning and deep-seated psychological needs. Using a two-stage sampling design process, data were collected on a total of 211 consumers, with the statistical analysis methods adopted for this study including a reliability test, confirmatory factor analysis, and LISREL analysis. Our results indicate that causal relationships hold in certain experiential value and lifestyle constructs. In particular, two experiential value constructs (social function, empathy and escapism) and two lifestyle constructs (pursuit of recreation and taste for life, reference group) play major roles that affect satisfaction among adolescents and young adults in online gaming in Taiwan.

  7. The Scientific Status of Projective Techniques.

    PubMed

    Lilienfeld, S O; Wood, J M; Garb, H N

    2000-11-01

    Although projective techniques continue to be widely used in clinical and forensic settings, their scientific status remains highly controversial. In this monograph, we review the current state of the literature concerning the psychometric properties (norms, reliability, validity, incremental validity, treatment utility) of three major projective instruments: Rorschach Inkblot Test, Thematic Apperception Test (TAT), and human figure drawings. We conclude that there is empirical support for the validity of a small number of indexes derived from the Rorschach and TAT. However, the substantial majority of Rorschach and TAT indexes are not empirically supported. The validity evidence for human figure drawings is even more limited. With a few exceptions, projective indexes have not consistently demonstrated incremental validity above and beyond other psychometric data. In addition, we summarize the results of a new meta-analysis intended to examine the capacity of these three instruments to detect child sexual abuse. Although some projective instruments were better than chance at detecting child sexual abuse, there were virtually no replicated findings across independent investigative teams. This meta-analysis also provides the first clear evidence of substantial file drawer effects in the projectives literature, as the effect sizes from published studies markedly exceeded those from unpublished studies. We conclude with recommendations regarding the (a) construction of projective techniques with adequate validity, (b) forensic and clinical use of projective techniques, and (c) education and training of future psychologists regarding projective techniques. © 2000 Association for Psychological Science.

  8. Estimate of the Reliability in Geological Forecasts for Tunnels: Toward a Structured Approach

    NASA Astrophysics Data System (ADS)

    Perello, Paolo

    2011-11-01

    In tunnelling, a reliable geological model often allows providing an effective design and facing the construction phase without unpleasant surprises. A geological model can be considered reliable when it is a valid support to correctly foresee the rock mass behaviour, therefore preventing unexpected events during the excavation. The higher the model reliability, the lower the probability of unforeseen rock mass behaviour. Unfortunately, owing to different reasons, geological models are affected by uncertainties and a fully reliable knowledge of the rock mass is, in most cases, impossible. Therefore, estimating to which degree a geological model is reliable, becomes a primary requirement in order to save time and money and to adopt the appropriate construction strategy. The definition of the geological model reliability is often achieved by engineering geologists through an unstructured analytical process and variable criteria. This paper focusses on geological models for projects of linear underground structures and represents an effort to analyse and include in a conceptual framework the factors influencing such models. An empirical parametric procedure is then developed with the aim of obtaining an index called "geological model rating (GMR)", which can be used to provide a more standardised definition of a geological model reliability.

  9. Improved RMR Rock Mass Classification Using Artificial Intelligence Algorithms

    NASA Astrophysics Data System (ADS)

    Gholami, Raoof; Rasouli, Vamegh; Alimoradi, Andisheh

    2013-09-01

    Rock mass classification systems such as rock mass rating (RMR) are very reliable means to provide information about the quality of rocks surrounding a structure as well as to propose suitable support systems for unstable regions. Many correlations have been proposed to relate measured quantities such as wave velocity to rock mass classification systems to limit the associated time and cost of conducting the sampling and mechanical tests conventionally used to calculate RMR values. However, these empirical correlations have been found to be unreliable, as they usually overestimate or underestimate the RMR value. The aim of this paper is to compare the results of RMR classification obtained from the use of empirical correlations versus machine-learning methodologies based on artificial intelligence algorithms. The proposed methods were verified based on two case studies located in northern Iran. Relevance vector regression (RVR) and support vector regression (SVR), as two robust machine-learning methodologies, were used to predict the RMR for tunnel host rocks. RMR values already obtained by sampling and site investigation at one tunnel were taken into account as the output of the artificial networks during training and testing phases. The results reveal that use of empirical correlations overestimates the predicted RMR values. RVR and SVR, however, showed more reliable results, and are therefore suggested for use in RMR classification for design purposes of rock structures.

  10. Enlight: A Comprehensive Quality and Therapeutic Potential Evaluation Tool for Mobile and Web-Based eHealth Interventions

    PubMed Central

    Faber, Keren; Mathur, Nandita; Kane, John M; Muench, Fred

    2017-01-01

    Background Studies of criteria-based assessment tools have demonstrated the feasibility of objectively evaluating eHealth interventions independent of empirical testing. However, current tools have not included some quality constructs associated with intervention outcome, such as persuasive design, behavior change, or therapeutic alliance. In addition, the generalizability of such tools has not been explicitly examined. Objective The aim is to introduce the development and further analysis of the Enlight suite of measures, developed to incorporate the aforementioned concepts and address generalizability aspects. Methods As a first step, a comprehensive systematic review was performed to identify relevant quality rating criteria in line with the PRISMA statement. These criteria were then categorized to create Enlight. The second step involved testing Enlight on 42 mobile apps and 42 Web-based programs (delivery mediums) targeting modifiable behaviors related to medical illness or mental health (clinical aims). Results A total of 476 criteria from 99 identified sources were used to build Enlight. The rating measures were divided into two sections: quality assessments and checklists. Quality assessments included usability, visual design, user engagement, content, therapeutic persuasiveness, therapeutic alliance, and general subjective evaluation. The checklists included credibility, privacy explanation, basic security, and evidence-based program ranking. The quality constructs exhibited excellent interrater reliability (intraclass correlations=.77-.98, median .91) and internal consistency (Cronbach alphas=.83-.90, median .88), with similar results when separated into delivery mediums or clinical aims. Conditional probability analysis revealed that 100% of the programs that received a score of fair or above (≥3.0) in therapeutic persuasiveness or therapeutic alliance received the same range of scores in user engagement and content—a pattern that did not appear in the opposite direction. Preliminary concurrent validity analysis pointed to positive correlations of combined quality scores with selected variables. The combined score that did not include therapeutic persuasiveness and therapeutic alliance descriptively underperformed the other combined scores. Conclusions This paper provides empirical evidence supporting the importance of persuasive design and therapeutic alliance within the context of a program’s evaluation. Reliability metrics and preliminary concurrent validity analysis indicate the potential of Enlight in examining eHealth programs regardless of delivery mediums and clinical aims. PMID:28325712

  11. Mokken scale analysis of mental health and well-being questionnaire item responses: a non-parametric IRT method in empirical research for applied health researchers

    PubMed Central

    2012-01-01

    Background Mokken scaling techniques are a useful tool for researchers who wish to construct unidimensional tests or use questionnaires that comprise multiple binary or polytomous items. The stochastic cumulative scaling model offered by this approach is ideally suited when the intention is to score an underlying latent trait by simple addition of the item response values. In our experience, the Mokken model appears to be less well-known than for example the (related) Rasch model, but is seeing increasing use in contemporary clinical research and public health. Mokken's method is a generalisation of Guttman scaling that can assist in the determination of the dimensionality of tests or scales, and enables consideration of reliability, without reliance on Cronbach's alpha. This paper provides a practical guide to the application and interpretation of this non-parametric item response theory method in empirical research with health and well-being questionnaires. Methods Scalability of data from 1) a cross-sectional health survey (the Scottish Health Education Population Survey) and 2) a general population birth cohort study (the National Child Development Study) illustrate the method and modeling steps for dichotomous and polytomous items respectively. The questionnaire data analyzed comprise responses to the 12 item General Health Questionnaire, under the binary recoding recommended for screening applications, and the ordinal/polytomous responses to the Warwick-Edinburgh Mental Well-being Scale. Results and conclusions After an initial analysis example in which we select items by phrasing (six positive versus six negatively worded items) we show that all items from the 12-item General Health Questionnaire (GHQ-12) – when binary scored – were scalable according to the double monotonicity model, in two short scales comprising six items each (Bech’s “well-being” and “distress” clinical scales). An illustration of ordinal item analysis confirmed that all 14 positively worded items of the Warwick-Edinburgh Mental Well-being Scale (WEMWBS) met criteria for the monotone homogeneity model but four items violated double monotonicity with respect to a single underlying dimension. Software availability and commands used to specify unidimensionality and reliability analysis and graphical displays for diagnosing monotone homogeneity and double monotonicity are discussed, with an emphasis on current implementations in freeware. PMID:22686586

  12. Mokken scale analysis of mental health and well-being questionnaire item responses: a non-parametric IRT method in empirical research for applied health researchers.

    PubMed

    Stochl, Jan; Jones, Peter B; Croudace, Tim J

    2012-06-11

    Mokken scaling techniques are a useful tool for researchers who wish to construct unidimensional tests or use questionnaires that comprise multiple binary or polytomous items. The stochastic cumulative scaling model offered by this approach is ideally suited when the intention is to score an underlying latent trait by simple addition of the item response values. In our experience, the Mokken model appears to be less well-known than for example the (related) Rasch model, but is seeing increasing use in contemporary clinical research and public health. Mokken's method is a generalisation of Guttman scaling that can assist in the determination of the dimensionality of tests or scales, and enables consideration of reliability, without reliance on Cronbach's alpha. This paper provides a practical guide to the application and interpretation of this non-parametric item response theory method in empirical research with health and well-being questionnaires. Scalability of data from 1) a cross-sectional health survey (the Scottish Health Education Population Survey) and 2) a general population birth cohort study (the National Child Development Study) illustrate the method and modeling steps for dichotomous and polytomous items respectively. The questionnaire data analyzed comprise responses to the 12 item General Health Questionnaire, under the binary recoding recommended for screening applications, and the ordinal/polytomous responses to the Warwick-Edinburgh Mental Well-being Scale. After an initial analysis example in which we select items by phrasing (six positive versus six negatively worded items) we show that all items from the 12-item General Health Questionnaire (GHQ-12)--when binary scored--were scalable according to the double monotonicity model, in two short scales comprising six items each (Bech's "well-being" and "distress" clinical scales). An illustration of ordinal item analysis confirmed that all 14 positively worded items of the Warwick-Edinburgh Mental Well-being Scale (WEMWBS) met criteria for the monotone homogeneity model but four items violated double monotonicity with respect to a single underlying dimension.Software availability and commands used to specify unidimensionality and reliability analysis and graphical displays for diagnosing monotone homogeneity and double monotonicity are discussed, with an emphasis on current implementations in freeware.

  13. Enlight: A Comprehensive Quality and Therapeutic Potential Evaluation Tool for Mobile and Web-Based eHealth Interventions.

    PubMed

    Baumel, Amit; Faber, Keren; Mathur, Nandita; Kane, John M; Muench, Fred

    2017-03-21

    Studies of criteria-based assessment tools have demonstrated the feasibility of objectively evaluating eHealth interventions independent of empirical testing. However, current tools have not included some quality constructs associated with intervention outcome, such as persuasive design, behavior change, or therapeutic alliance. In addition, the generalizability of such tools has not been explicitly examined. The aim is to introduce the development and further analysis of the Enlight suite of measures, developed to incorporate the aforementioned concepts and address generalizability aspects. As a first step, a comprehensive systematic review was performed to identify relevant quality rating criteria in line with the PRISMA statement. These criteria were then categorized to create Enlight. The second step involved testing Enlight on 42 mobile apps and 42 Web-based programs (delivery mediums) targeting modifiable behaviors related to medical illness or mental health (clinical aims). A total of 476 criteria from 99 identified sources were used to build Enlight. The rating measures were divided into two sections: quality assessments and checklists. Quality assessments included usability, visual design, user engagement, content, therapeutic persuasiveness, therapeutic alliance, and general subjective evaluation. The checklists included credibility, privacy explanation, basic security, and evidence-based program ranking. The quality constructs exhibited excellent interrater reliability (intraclass correlations=.77-.98, median .91) and internal consistency (Cronbach alphas=.83-.90, median .88), with similar results when separated into delivery mediums or clinical aims. Conditional probability analysis revealed that 100% of the programs that received a score of fair or above (≥3.0) in therapeutic persuasiveness or therapeutic alliance received the same range of scores in user engagement and content-a pattern that did not appear in the opposite direction. Preliminary concurrent validity analysis pointed to positive correlations of combined quality scores with selected variables. The combined score that did not include therapeutic persuasiveness and therapeutic alliance descriptively underperformed the other combined scores. This paper provides empirical evidence supporting the importance of persuasive design and therapeutic alliance within the context of a program's evaluation. Reliability metrics and preliminary concurrent validity analysis indicate the potential of Enlight in examining eHealth programs regardless of delivery mediums and clinical aims. ©Amit Baumel, Keren Faber, Nandita Mathur, John M Kane, Fred Muench. Originally published in the Journal of Medical Internet Research (http://www.jmir.org), 21.03.2017.

  14. Validation of the Mindful Coping Scale

    ERIC Educational Resources Information Center

    Tharaldsen, Kjersti B.; Bru, Edvin

    2011-01-01

    The aim of this research is to develop and validate a self-report measure of mindfulness and coping, the mindful coping scale (MCS). Dimensions of mindful coping were theoretically deduced from mindfulness theory and coping theory. The MCS was empirically evaluated by use of factor analyses, reliability testing and nomological network validation.…

  15. Print and Internet Catalog Shopping: Assessing Attitudes and Intentions.

    ERIC Educational Resources Information Center

    Vijayasarathy, Leo R.; Jones, Joseph M.

    2000-01-01

    Findings of an empirical study that compared individuals' attitudes and intentions to shop using print and Internet catalogs suggest that individuals perceived differences between the two catalog media on the shopping factors of reliability, tangibility, and consumer risk. Product value, pre-order information, post-selection information, shopping…

  16. 75 FR 13515 - Office of Innovation and Improvement (OII); Overview Information; Ready-to-Learn Television...

    Federal Register 2010, 2011, 2012, 2013, 2014

    2010-03-22

    ... on rigorous scientifically based research methods to assess the effectiveness of a particular... activities and programs; and (B) Includes research that-- (i) Employs systematic, empirical methods that draw... or observational methods that provide reliable and valid data across evaluators and observers, across...

  17. Toward Evidence-Informed Policy and Practice in Child Welfare

    ERIC Educational Resources Information Center

    Littell, Julia H.; Shlonsky, Aron

    2010-01-01

    Drawing on the authors' experience in the international Campbell Collaboration, this essay presents a principled and pragmatic approach to evidence-informed decisions about child welfare. This approach takes into account the growing body of empirical evidence on the reliability and validity of various methods of research synthesis. It also…

  18. Mental Disorder Among Homeless Persons in the United States: An Overview of Recent Empirical Literature.

    ERIC Educational Resources Information Center

    Robertson, Marjorie J.

    1986-01-01

    Reviews literature on the homeless reporting higher rates of psychiatric disorder, psychological distress, and previous psychiatric hospitalization compared to the general population. However, understandardized methodology and lack of consistent findings across studies prohibit reliable prevalence estimates of mental disorder among the homeless.…

  19. A Psychometric Evaluation of the Core Bereavement Items

    ERIC Educational Resources Information Center

    Holland, Jason M.; Nam, Ilsung; Neimeyer, Robert A.

    2013-01-01

    Despite being a routinely administered assessment of grieving, few studies have empirically examined the psychometric properties of the Core Bereavement Items (CBI). The present study investigated the factor structure, internal reliability, and concurrent validity of the CBI in a large, diverse sample of bereaved young adults (N = 1,366).…

  20. Sparsity guided empirical wavelet transform for fault diagnosis of rolling element bearings

    NASA Astrophysics Data System (ADS)

    Wang, Dong; Zhao, Yang; Yi, Cai; Tsui, Kwok-Leung; Lin, Jianhui

    2018-02-01

    Rolling element bearings are widely used in various industrial machines, such as electric motors, generators, pumps, gearboxes, railway axles, turbines, and helicopter transmissions. Fault diagnosis of rolling element bearings is beneficial to preventing any unexpected accident and reducing economic loss. In the past years, many bearing fault detection methods have been developed. Recently, a new adaptive signal processing method called empirical wavelet transform attracts much attention from readers and engineers and its applications to bearing fault diagnosis have been reported. The main problem of empirical wavelet transform is that Fourier segments required in empirical wavelet transform are strongly dependent on the local maxima of the amplitudes of the Fourier spectrum of a signal, which connotes that Fourier segments are not always reliable and effective if the Fourier spectrum of the signal is complicated and overwhelmed by heavy noises and other strong vibration components. In this paper, sparsity guided empirical wavelet transform is proposed to automatically establish Fourier segments required in empirical wavelet transform for fault diagnosis of rolling element bearings. Industrial bearing fault signals caused by single and multiple railway axle bearing defects are used to verify the effectiveness of the proposed sparsity guided empirical wavelet transform. Results show that the proposed method can automatically discover Fourier segments required in empirical wavelet transform and reveal single and multiple railway axle bearing defects. Besides, some comparisons with three popular signal processing methods including ensemble empirical mode decomposition, the fast kurtogram and the fast spectral correlation are conducted to highlight the superiority of the proposed method.

  1. Systematic Review of Methods in Low-Consensus Fields: Supporting Commensuration through `Construct-Centered Methods Aggregation' in the Case of Climate Change Vulnerability Research.

    PubMed

    Delaney, Aogán; Tamás, Peter A; Crane, Todd A; Chesterman, Sabrina

    2016-01-01

    There is increasing interest in using systematic review to synthesize evidence on the social and environmental effects of and adaptations to climate change. Use of systematic review for evidence in this field is complicated by the heterogeneity of methods used and by uneven reporting. In order to facilitate synthesis of results and design of subsequent research a method, construct-centered methods aggregation, was designed to 1) provide a transparent, valid and reliable description of research methods, 2) support comparability of primary studies and 3) contribute to a shared empirical basis for improving research practice. Rather than taking research reports at face value, research designs are reviewed through inductive analysis. This involves bottom-up identification of constructs, definitions and operationalizations; assessment of concepts' commensurability through comparison of definitions; identification of theoretical frameworks through patterns of construct use; and integration of transparently reported and valid operationalizations into ideal-type research frameworks. Through the integration of reliable bottom-up inductive coding from operationalizations and top-down coding driven from stated theory with expert interpretation, construct-centered methods aggregation enabled both resolution of heterogeneity within identically named constructs and merging of differently labeled but identical constructs. These two processes allowed transparent, rigorous and contextually sensitive synthesis of the research presented in an uneven set of reports undertaken in a heterogenous field. If adopted more broadly, construct-centered methods aggregation may contribute to the emergence of a valid, empirically-grounded description of methods used in primary research. These descriptions may function as a set of expectations that improves the transparency of reporting and as an evolving comprehensive framework that supports both interpretation of existing and design of future research.

  2. Systems Toxicology: Real World Applications and Opportunities.

    PubMed

    Hartung, Thomas; FitzGerald, Rex E; Jennings, Paul; Mirams, Gary R; Peitsch, Manuel C; Rostami-Hodjegan, Amin; Shah, Imran; Wilks, Martin F; Sturla, Shana J

    2017-04-17

    Systems Toxicology aims to change the basis of how adverse biological effects of xenobiotics are characterized from empirical end points to describing modes of action as adverse outcome pathways and perturbed networks. Toward this aim, Systems Toxicology entails the integration of in vitro and in vivo toxicity data with computational modeling. This evolving approach depends critically on data reliability and relevance, which in turn depends on the quality of experimental models and bioanalysis techniques used to generate toxicological data. Systems Toxicology involves the use of large-scale data streams ("big data"), such as those derived from omics measurements that require computational means for obtaining informative results. Thus, integrative analysis of multiple molecular measurements, particularly acquired by omics strategies, is a key approach in Systems Toxicology. In recent years, there have been significant advances centered on in vitro test systems and bioanalytical strategies, yet a frontier challenge concerns linking observed network perturbations to phenotypes, which will require understanding pathways and networks that give rise to adverse responses. This summary perspective from a 2016 Systems Toxicology meeting, an international conference held in the Alps of Switzerland, describes the limitations and opportunities of selected emerging applications in this rapidly advancing field. Systems Toxicology aims to change the basis of how adverse biological effects of xenobiotics are characterized, from empirical end points to pathways of toxicity. This requires the integration of in vitro and in vivo data with computational modeling. Test systems and bioanalytical technologies have made significant advances, but ensuring data reliability and relevance is an ongoing concern. The major challenge facing the new pathway approach is determining how to link observed network perturbations to phenotypic toxicity.

  3. Systematic Review of Methods in Low-Consensus Fields: Supporting Commensuration through `Construct-Centered Methods Aggregation’ in the Case of Climate Change Vulnerability Research

    PubMed Central

    Crane, Todd A.; Chesterman, Sabrina

    2016-01-01

    There is increasing interest in using systematic review to synthesize evidence on the social and environmental effects of and adaptations to climate change. Use of systematic review for evidence in this field is complicated by the heterogeneity of methods used and by uneven reporting. In order to facilitate synthesis of results and design of subsequent research a method, construct-centered methods aggregation, was designed to 1) provide a transparent, valid and reliable description of research methods, 2) support comparability of primary studies and 3) contribute to a shared empirical basis for improving research practice. Rather than taking research reports at face value, research designs are reviewed through inductive analysis. This involves bottom-up identification of constructs, definitions and operationalizations; assessment of concepts’ commensurability through comparison of definitions; identification of theoretical frameworks through patterns of construct use; and integration of transparently reported and valid operationalizations into ideal-type research frameworks. Through the integration of reliable bottom-up inductive coding from operationalizations and top-down coding driven from stated theory with expert interpretation, construct-centered methods aggregation enabled both resolution of heterogeneity within identically named constructs and merging of differently labeled but identical constructs. These two processes allowed transparent, rigorous and contextually sensitive synthesis of the research presented in an uneven set of reports undertaken in a heterogenous field. If adopted more broadly, construct-centered methods aggregation may contribute to the emergence of a valid, empirically-grounded description of methods used in primary research. These descriptions may function as a set of expectations that improves the transparency of reporting and as an evolving comprehensive framework that supports both interpretation of existing and design of future research. PMID:26901409

  4. Systems Toxicology: Real World Applications and Opportunities

    PubMed Central

    2017-01-01

    Systems Toxicology aims to change the basis of how adverse biological effects of xenobiotics are characterized from empirical end points to describing modes of action as adverse outcome pathways and perturbed networks. Toward this aim, Systems Toxicology entails the integration of in vitro and in vivo toxicity data with computational modeling. This evolving approach depends critically on data reliability and relevance, which in turn depends on the quality of experimental models and bioanalysis techniques used to generate toxicological data. Systems Toxicology involves the use of large-scale data streams (“big data”), such as those derived from omics measurements that require computational means for obtaining informative results. Thus, integrative analysis of multiple molecular measurements, particularly acquired by omics strategies, is a key approach in Systems Toxicology. In recent years, there have been significant advances centered on in vitro test systems and bioanalytical strategies, yet a frontier challenge concerns linking observed network perturbations to phenotypes, which will require understanding pathways and networks that give rise to adverse responses. This summary perspective from a 2016 Systems Toxicology meeting, an international conference held in the Alps of Switzerland, describes the limitations and opportunities of selected emerging applications in this rapidly advancing field. Systems Toxicology aims to change the basis of how adverse biological effects of xenobiotics are characterized, from empirical end points to pathways of toxicity. This requires the integration of in vitro and in vivo data with computational modeling. Test systems and bioanalytical technologies have made significant advances, but ensuring data reliability and relevance is an ongoing concern. The major challenge facing the new pathway approach is determining how to link observed network perturbations to phenotypic toxicity. PMID:28362102

  5. Helicobacter pylori therapy: a paradigm shift

    PubMed Central

    Graham, David Y; Dore, Maria Pina

    2016-01-01

    SUMMARY Helicobacter pylori (H. Pylori) is a leading cause of gastroduodenal disease, including gastric cancer. H. pylori eradication therapies and their efficacy are summarized. A number of current treatment regimens will reliably yield >90% or 95% cure rates with susceptible strains. None has proven to be superior. We show how to predict the efficacy of a regimen in any population provided one knows the prevalence of antibiotic resistance. As with other infectious diseases, therapy should always be susceptibility-based. Susceptibility testing should be demanded. We provide recommendations for empiric therapies when the only option and describe how to distinguish studies providing misinformation from those providing reliable and interpretable data. When treated as an infectious disease, high H. pylori cure rates are relatively simple to reliably achieve. PMID:27077447

  6. Effects of sampling conditions on DNA-based estimates of American black bear abundance

    USGS Publications Warehouse

    Laufenberg, Jared S.; Van Manen, Frank T.; Clark, Joseph D.

    2013-01-01

    DNA-based capture-mark-recapture techniques are commonly used to estimate American black bear (Ursus americanus) population abundance (N). Although the technique is well established, many questions remain regarding study design. In particular, relationships among N, capture probability of heterogeneity mixtures A and B (pA and pB, respectively, or p, collectively), the proportion of each mixture (π), number of capture occasions (k), and probability of obtaining reliable estimates of N are not fully understood. We investigated these relationships using 1) an empirical dataset of DNA samples for which true N was unknown and 2) simulated datasets with known properties that represented a broader array of sampling conditions. For the empirical data analysis, we used the full closed population with heterogeneity data type in Program MARK to estimate N for a black bear population in Great Smoky Mountains National Park, Tennessee. We systematically reduced the number of those samples used in the analysis to evaluate the effect that changes in capture probabilities may have on parameter estimates. Model-averaged N for females and males were 161 (95% CI = 114–272) and 100 (95% CI = 74–167), respectively (pooled N = 261, 95% CI = 192–419), and the average weekly p was 0.09 for females and 0.12 for males. When we reduced the number of samples of the empirical data, support for heterogeneity models decreased. For the simulation analysis, we generated capture data with individual heterogeneity covering a range of sampling conditions commonly encountered in DNA-based capture-mark-recapture studies and examined the relationships between those conditions and accuracy (i.e., probability of obtaining an estimated N that is within 20% of true N), coverage (i.e., probability that 95% confidence interval includes true N), and precision (i.e., probability of obtaining a coefficient of variation ≤20%) of estimates using logistic regression. The capture probability for the larger of 2 mixture proportions of the population (i.e., pA or pB, depending on the value of π) was most important for predicting accuracy and precision, whereas capture probabilities of both mixture proportions (pA and pB) were important to explain variation in coverage. Based on sampling conditions similar to parameter estimates from the empirical dataset (pA = 0.30, pB = 0.05, N = 250, π = 0.15, and k = 10), predicted accuracy and precision were low (60% and 53%, respectively), whereas coverage was high (94%). Increasing pB, the capture probability for the predominate but most difficult to capture proportion of the population, was most effective to improve accuracy under those conditions. However, manipulation of other parameters may be more effective under different conditions. In general, the probabilities of obtaining accurate and precise estimates were best when p≥ 0.2. Our regression models can be used by managers to evaluate specific sampling scenarios and guide development of sampling frameworks or to assess reliability of DNA-based capture-mark-recapture studies.

  7. Identification student’s misconception of heat and temperature using three-tier diagnostic test

    NASA Astrophysics Data System (ADS)

    Suliyanah; Putri, H. N. P. A.; Rohmawati, L.

    2018-03-01

    The objective of this research is to develop a Three-Tier Diagnostic Test (TTDT) to identify the student's misconception of heat and temperature. Stages of development include: analysis, planning, design, development, evaluation and revise. The results of this study show that (1) the quality of the three-tier type diagnostic test instrument developed has been expressed well with the following details: (a) Internal validity of 88.19% belonging to the valid category. (b) External validity of empirical construct validity test using Pearson Product Moment obtained 0.43 is classified and result of empirical construct validity test obtained false positives 6.1% and false negatives 5.9% then the instrument was valid. (c) Test reliability by using Cronbach’s Alpha of 0.98 which means acceptable. (d) The 80% difficulty level test is quite difficult. (2) Student misconceptions on the temperature of heat and displacement materials based on the II test the highest (84%), the lowest (21%), and the non-misconceptions (7%). (3) The highest cause of misconception among students is associative thinking (22%) and the lowest is caused by incomplete or incomplete reasoning (11%). Three-Tier Diagnostic Test (TTDT) could identify the student's misconception of heat and temperature.

  8. Site-specific landslide assessment in Alpine area using a reliable integrated monitoring system

    NASA Astrophysics Data System (ADS)

    Romeo, Saverio; Di Matteo, Lucio; Kieffer, Daniel Scott

    2016-04-01

    Rockfalls are one of major cause of landslide fatalities around the world. The present work discusses the reliability of integrated monitoring of displacements in a rockfall within the Alpine region (Salzburg Land - Austria), taking into account also the effect of the ongoing climate change. Due to the unpredictability of the frequency and magnitude, that threatens human lives and infrastructure, frequently it is necessary to implement an efficient monitoring system. For this reason, during the last decades, integrated monitoring systems of unstable slopes were widely developed and used (e.g., extensometers, cameras, remote sensing, etc.). In this framework, Remote Sensing techniques, such as GBInSAR technique (Groung-Based Interferometric Synthetic Aperture Radar), have emerged as efficient and powerful tools for deformation monitoring. GBInSAR measurements can be useful to achieve an early warning system using surface deformation parameters as ground displacement or inverse velocity (for semi-empirical forecasting methods). In order to check the reliability of GBInSAR and to monitor the evolution of landslide, it is very important to integrate different techniques. Indeed, a multi-instrumental approach is essential to investigate movements both in surface and in depth and the use of different monitoring techniques allows to perform a cross analysis of the data and to minimize errors, to check the data quality and to improve the monitoring system. During 2013, an intense and complete monitoring campaign has been conducted on the Ingelsberg landslide. By analyzing both historical temperature series (HISTALP) recorded during the last century and those from local weather stations, temperature values (Autumn-Winter, Winter and Spring) are clearly increased in Bad Hofgastein area as well as in Alpine region. As consequence, in the last decades the rockfall events have been shifted from spring to summer due to warmer winters. It is interesting to point out that temperature values recorded in the valley and on the slope show a good relationship indicating that the climatic monitoring is reliable. In addition, the landslide displacement monitoring is reliable as well: the comparison between displacements in depth by extensometers and in surface by GBInSAR - referred to March-December 2013 - shows ad high reliability as confirmed by the inter-rater reliability analysis (Pearson correlation coefficient higher than 0.9). In conclusion, the reliability of the monitoring system confirms that data can be useful to improve the knowledge on rockfall kinematic and to develop an accurate early warning system useful for civil protection issues.

  9. The Bi-directional Relationship between Source Characteristics and Message Content

    PubMed Central

    Collins, Peter J.; Hahn, Ulrike; von Gerber, Ylva; Olsson, Erik J.

    2018-01-01

    Much of what we believe we know, we know through the testimony of others (Coady, 1992). While there has been long-standing evidence that people are sensitive to the characteristics of the sources of testimony, for example in the context of persuasion, researchers have only recently begun to explore the wider implications of source reliability considerations for the nature of our beliefs. Likewise, much remains to be established concerning what factors influence source reliability. In this paper, we examine, both theoretically and empirically, the implications of using message content as a cue to source reliability. We present a set of experiments examining the relationship between source information and message content in people's responses to simple communications. The results show that people spontaneously revise their beliefs in the reliability of the source on the basis of the expectedness of a source's claim and, conversely, adjust message impact by perceived reliability; hence source reliability and message content have a bi-directional relationship. The implications are discussed for a variety of psychological, philosophical and political issues such as belief polarization and dual-route models of persuasion. PMID:29441029

  10. Development and pilot-test of the Workplace Readiness Questionnaire, a theory-based instrument to measure small workplaces’ readiness to implement wellness programs

    PubMed Central

    Hannon, Peggy A.; Helfrich, Christian D.; Chan, K. Gary; Allen, Claire L.; Hammerback, Kristen; Kohn, Marlana J.; Parrish, Amanda T.; Weiner, Bryan J.; Harris, Jeffrey R.

    2016-01-01

    Purpose To develop a theory-based questionnaire to assess readiness for change in small workplaces adopting wellness programs. Design In developing our scale, we first tested items via “think-aloud” interviews. We tested the revised items in a cross-sectional quantitative telephone survey. Setting Small workplaces (20–250 employees) in low-wage industries. Subjects Decision-makers representing small workplaces in King County, Washington (think-aloud interviews, n=9) and the United States (telephone survey, n=201). Measures We generated items for each construct in Weiner’s theory of organizational readiness for change. We also measured workplace characteristics and current implementation of workplace wellness programs. Analysis We assessed reliability by coefficient alpha for each of the readiness questionnaire subscales. We tested the association of all subscales with employers’ current implementation of wellness policies, programs, and communications, and conducted a path analysis to test the associations in the theory of organizational readiness to change. Results Each of the readiness subscales exhibited acceptable internal reliability (coefficient alpha range = .75–.88) and was positively associated with wellness program implementation (p <.05). The path analysis was consistent with the theory of organizational readiness to change, except change efficacy did not predict change-related effort. Conclusion We developed a new questionnaire to assess small workplaces’ readiness to adopt and implement evidence-based wellness programs. Our findings also provide empirical validation of Weiner’s theory of readiness for change. PMID:26389975

  11. A decomposition model and voxel selection framework for fMRI analysis to predict neural response of visual stimuli.

    PubMed

    Raut, Savita V; Yadav, Dinkar M

    2018-03-28

    This paper presents an fMRI signal analysis methodology using geometric mean curve decomposition (GMCD) and mutual information-based voxel selection framework. Previously, the fMRI signal analysis has been conducted using empirical mean curve decomposition (EMCD) model and voxel selection on raw fMRI signal. The erstwhile methodology loses frequency component, while the latter methodology suffers from signal redundancy. Both challenges are addressed by our methodology in which the frequency component is considered by decomposing the raw fMRI signal using geometric mean rather than arithmetic mean and the voxels are selected from EMCD signal using GMCD components, rather than raw fMRI signal. The proposed methodologies are adopted for predicting the neural response. Experimentations are conducted in the openly available fMRI data of six subjects, and comparisons are made with existing decomposition models and voxel selection frameworks. Subsequently, the effect of degree of selected voxels and the selection constraints are analyzed. The comparative results and the analysis demonstrate the superiority and the reliability of the proposed methodology.

  12. Medical student quality-of-life in the clerkships: a scale validation study.

    PubMed

    Brannick, Michael T; Horn, Gregory T; Schnaus, Michael J; Wahi, Monika M; Goldin, Steven B

    2015-04-01

    Many aspects of medical school are stressful for students. To empirically assess student reactions to clerkship programs, or to assess efforts to improve such programs, educators must measure the overall well-being of the students reliably and validly. The purpose of the study was to develop and validate a measure designed to achieve these goals. The authors developed a measure of quality of life for medical students by sampling (public domain) items tapping general happiness, fatigue, and anxiety. A quality-of-life scale was developed by factor analyzing responses to the items from students in two different clerkships from 2005 to 2008. Reliability was assessed using Cronbach's alpha. Validity was assessed by factor analysis, convergence with additional theoretically relevant scales, and sensitivity to change over time. The refined nine-item measure is a Likert scaled survey of quality-of-life items comprised of two domains: exhaustion and general happiness. The resulting scale demonstrated good reliability and factorial validity at two time points for each of the two samples. The quality-of-life measure also correlated with measures of depression and the amount of sleep reported during the clerkships. The quality-of-life measure appeared more sensitive to changes over time than did the depression measure. The measure is short and can be easily administered in a survey. The scale appears useful for program evaluation and more generally as an outcome variable in medical educational research.

  13. The psychometric properties of the 5-item gratitude questionnaire in Chinese adolescents.

    PubMed

    Zeng, Y; Ling, Y; Huebner, E S; He, Y; Lei, X

    2017-05-01

    WHAT IS KNOWN ON THE SUBJECT?: The GQ-6 is one of the most widely used self-report questionnaires to evaluate the level of gratitude among adults. The GQ-5 appears suitable for adolescents. WHAT THIS PAPER ADDS TO EXISTING KNOWLEDGE?: We developed a Chinese version of the GQ-5 and examined evidence for its reliability and validity. Results demonstrated adequate reliability and validity, indicating that it is appropriate for the assessment of gratitude in Chinese adolescents. In addition, Chinese early adolescent females reported higher gratitude than adolescent males. WHAT ARE THE IMPLICATIONS FOR PRACTICE?: Screening adolescents who have lower levels of gratitude through the GQ-5 could help identify students who may benefit from empirically validated interventions to promote higher levels of gratitude in an effort to promote positive psychosocial and academic outcomes. Background This study was conducted to evaluate the psychometric properties of the Chinese version of the 5-item Gratitude Questionnaire (GQ-5). Method The sample consisted of 2093 middle school students (46.8% males) in mainland China. Confirmatory factor analysis and multigroup confirmatory factor analysis were performed to examine the factor structure and the measurement equivalence across gender. The convergent validity, Cronbach's α and mean interitem correlations of the GQ-5 were also evaluated. Results The results provided evidence of internal consistency reliability through a Cronbach's α of 0.812 and a mean interitem correlation of 0.463 for the total sample. The results also supported a one-dimensional factor structure. In addition, convergent validity was assessed by statistically significant positive correlations between the GQ-5 and the two subscales of the Children's Hope Scale (CHS) and the Brief Multidimensional Students' Life Satisfaction Scale (BMSLSS) total score. Finally, multigroup confirmatory factor analysis also demonstrated measurement equivalence across gender. Subsequent analyses of latent mean revealed gender differences in early adolescent male and female students. Conclusions The Chinese version of the GQ-5 appears to be a reliable and valid measure of gratitude among Chinese early adolescents. Early adolescent female students reported higher gratitude than early adolescent male students. © 2017 John Wiley & Sons Ltd.

  14. Content validity of the DSM-IV borderline and narcissistic personality disorder criteria sets.

    PubMed

    Blais, M A; Hilsenroth, M J; Castlebury, F D

    1997-01-01

    This study sought to empirically evaluate the content validity of the newly revised DSM-IV narcissistic personality disorder (NPD) and borderline personality disorder (BPD) criteria sets. Using the essential features of each disorder as construct definitions, factor analysis was used to determine how adequately the criteria sets covered the constructs. In addition, this empirical investigation sought to: 1) help define the dimensions underlying these polythetic disorders; 2) identify core features of each diagnosis; and 3) highlight the characteristics that may be most useful in diagnosing these two disorders. Ninety-one outpatients meeting DSM-IV criteria for a personality disorder (PD) were identified through a retrospective analysis of chart information. Records of these 91 patients were independently rated on all of the BPD and NPD symptom criteria for the DSM-IV. Acceptable interrater reliability (kappa estimates) was obtained for both presence or absence of a PD and symptom criteria for BPD and NPD. The factor analysis, performed separately for each disorder, identified a three-factor solution for both the DSM-IV BPD and NPD criteria sets. The results of this study provide strong support for the content validity of the NPD criteria set and moderate support for the content validly of the BPD criteria set. Three domains were found to comprise the BPD criteria set, with the essential features of interpersonal and identity instability forming one domain, and impulsivity and affective instability each identified as separate domains. Factor analysis of the NPD criteria set found three factors basically corresponding to the essential features of grandiosity, lack of empathy, and need for admiration. Therefore, the NPD criteria set adequately covers the essential or defining features of the disorder.

  15. Evaluating the intersection of a regional wildlife connectivity network with highways.

    PubMed

    Cushman, Samuel A; Lewis, Jesse S; Landguth, Erin L

    2013-01-01

    Reliable predictions of regional-scale population connectivity are needed to prioritize conservation actions. However, there have been few examples of regional connectivity models that are empirically derived and validated. The central goals of this paper were to (1) evaluate the effectiveness of factorial least cost path corridor mapping on an empirical resistance surface in reflecting the frequency of highway crossings by American black bear, (2) predict the location and predicted intensity of use of movement corridors for American black bear, and (3) identify where these corridors cross major highways and rank the intensity of these crossings. We used factorial least cost path modeling coupled with resistant kernel analysis to predict a network of movement corridors across a 30.2 million hectare analysis area in Montana and Idaho, USA. Factorial least cost path corridor mapping was associated with the locations of actual bear highway crossings. We identified corridor-highway intersections and ranked these based on corridor strength. We found that a major wildlife crossing overpass structure was located close to one of the most intense predicted corridors, and that the vast majority of the predicted corridor network was "protected" under federal management. However, narrow, linear corridors connecting the Greater Yellowstone Ecosystem to the rest of the analysis area had limited protection by federal ownership, making these additionally vulnerable to habitat loss and fragmentation. Factorial least cost path modeling coupled with resistant kernel analysis provides detailed, synoptic information about connectivity across populations that vary in distribution and density in complex landscapes. Specifically, our results could be used to quantify the structure of the connectivity network, identify critical linkage nodes and core areas, map potential barriers and fracture zones, and prioritize locations for mitigation, restoration and conservation actions.

  16. Development and construct validity of the Classroom Strategies Scale-Observer Form.

    PubMed

    Reddy, Linda A; Fabiano, Gregory; Dudek, Christopher M; Hsu, Louis

    2013-12-01

    Research on progress monitoring has almost exclusively focused on student behavior and not on teacher practices. This article presents the development and validation of a new teacher observational assessment (Classroom Strategies Scale) of classroom instructional and behavioral management practices. The theoretical underpinnings and empirical basis for the instructional and behavioral management scales are presented. The Classroom Strategies Scale (CSS) evidenced overall good reliability estimates including internal consistency, interrater reliability, test-retest reliability, and freedom from item bias on important teacher demographics (age, educational degree, years of teaching experience). Confirmatory factor analyses (CFAs) of CSS data from 317 classrooms were carried out to assess the level of empirical support for (a) a 4 first-order factor theory concerning teachers' instructional practices, and (b) a 4 first-order factor theory concerning teachers' behavior management practice. Several fit indices indicated acceptable fit of the (a) and (b) CFA models to the data, as well as acceptable fit of less parsimonious alternative CFA models that included 1 or 2 second-order factors. Information-theory-based indices generally suggested that the (a) and (b) CFA models fit better than some more parsimonious alternative CFA models that included constraints on relations of first-order factors. Overall, CFA first-order and higher order factor results support the CSS-Observer Total, Composite, and subscales. Suggestions for future measurement development efforts are outlined. PsycINFO Database Record (c) 2013 APA, all rights reserved.

  17. Empirical Analysis of Optical Attenuator Performance in Quantum Key Distribution Systems Using a Particle Model

    DTIC Science & Technology

    2012-03-01

    EMPIRICAL ANALYSIS OF OPTICAL ATTENUATOR PERFORMANCE IN QUANTUM KEY DISTRIBUTION SYSTEMS USING A...DISTRIBUTION IS UNLIMITED AFIT/GCS/ENG/12-01 EMPIRICAL ANALYSIS OF OPTICAL ATTENUATOR PERFORMANCE IN QUANTUM KEY DISTRIBUTION SYSTEMS USING ...challenging as the complexity of actual implementation specifics are considered. Two components common to most quantum key distribution

  18. A critique of silvicultural approaches to managing defoliating insects in North America

    Treesearch

    R.M. ​Muzika; A.M. Liebhold

    2000-01-01

    A variety of silvicultural techniques have been suggested for managing forest defoliating insects. The objectives focus on minimizing defoliation or minimizing damage from defoliation. The theoretical foundations of many approaches have been built upon observation and correlation, and very little reliable empirical evidence exists to support the objectives of...

  19. A Descriptive, Multiyear Examination of Positive Behavior Support

    ERIC Educational Resources Information Center

    Dunlap, Glen; Carr, Edward G.; Horner, Robert H.; Koegel, Robert L.; Sailor, Wayne; Clarke, Shelley; Koegel, Lynn Kern; Albin, Richard W.; Vaughn, Bobbie J.; McLaughlin, Darlene Magito; James, Kim Mullen; Todd, Anne W.; Newton, J. Stephen; Lucyshyn, Joseph; Griggs, Peter; Bohanon, Hank; Choi, Jeong Hoon; Vismara, Laurie; Minjarez, Mendy Boettcher; Buschbacher, Pamelazita; Fox, Lise

    2010-01-01

    A major goal of positive behavior support (PBS) is to produce broad-based, long-term improvements in adaptive behavior; however, the empirical base, at present, is mainly composed of relatively short-term studies carried out in circumscribed contexts. Therefore, a need exists for reliable data that can inform the field regarding the comprehensive…

  20. Who Should Mark What? A Study of Factors Affecting Marking Accuracy in a Biology Examination

    ERIC Educational Resources Information Center

    Suto, Irenka; Nadas, Rita; Bell, John

    2011-01-01

    Accurate marking is crucial to the reliability and validity of public examinations, in England and internationally. Factors contributing to accuracy have been conceptualised as affecting either marking task demands or markers' personal expertise. The aim of this empirical study was to develop this conceptualisation through investigating the…

  1. Analytical Rubrics in Higher Education: A Repository of Empirical Data

    ERIC Educational Resources Information Center

    Hack, Catherine

    2015-01-01

    The use of rubrics for grading and feedback in higher education has increased in response to requirements for consistency and transparency across a diverse range of assessment tasks. There is a growing evidence base demonstrating the reliability of rubrics across different markers and instances. The number of studies describing the impact of…

  2. Learning Capability and Business Performance: A Non-Financial and Financial Assessment

    ERIC Educational Resources Information Center

    Ma Prieto, Isabel; Revilla, Elena

    2006-01-01

    Purpose: There has been little research that includes reliable deductions about the positive influence of learning capability on business performance. For this reason, the main objective of the present study is to empirically explore the link between learning capability in organizations and business performance evaluated in both financial and…

  3. Child and Adolescent Behaviorally Based Disorders: A Critical Review of Reliability and Validity

    ERIC Educational Resources Information Center

    Mallett, Christopher A.

    2014-01-01

    Objectives: The purpose of this study was to investigate the historical construction and empirical support of two child and adolescent behaviorally based mental health disorders: oppositional defiant and conduct disorders. Method: The study utilized a historiography methodology to review, from 1880 to 2012, these disorders' inclusion in…

  4. Establishing Validity of the Consensus Auditory-Perceptual Evaluation of Voice (CAPE-V)

    ERIC Educational Resources Information Center

    Zraick, Richard I.; Kempster, Gail B.; Connor, Nadine P.; Thibeault, Susan; Klaben, Bernice K.; Bursac, Zoran; Thrush, Carol R.; Glaze, Leslie E.

    2011-01-01

    Purpose: The Consensus Auditory-Perceptual Evaluation of Voice (CAPE-V) was developed to provide a protocol and form for clinicians to use when assessing the voice quality of adults with voice disorders (Kempster, Gerratt, Verdolini Abbott, Barkmeier-Kramer, & Hillman, 2009). This study examined the reliability and the empirical validity of the…

  5. Improving the Validity of Quantitative Measures in Applied Linguistics Research

    ERIC Educational Resources Information Center

    Purpura, James E.; Brown, James Dean; Schoonen, Rob

    2015-01-01

    In empirical applied linguistics research it is essential that the key variables are operationalized in a valid and reliable way, and that the scores are treated appropriately, allowing for a proper testing of the hypotheses under investigation. The current article addresses several theoretical and practical issues regarding the use of measurement…

  6. Patterns of Cognitive Strengths and Weaknesses: Identification Rates, Agreement, and Validity for Learning Disabilities Identification

    ERIC Educational Resources Information Center

    Miciak, Jeremy; Fletcher, Jack M.; Stuebing, Karla K.; Vaughn, Sharon; Tolar, Tammy D.

    2014-01-01

    Few empirical investigations have evaluated learning disabilities (LD) identification methods based on a pattern of cognitive strengths and weaknesses (PSW). This study investigated the reliability and validity of two proposed PSW methods: the concordance/discordance method (C/DM) and cross battery assessment (XBA) method. Cognitive assessment…

  7. Measurement Properties of Two Innovative Item Formats in a Computer-Based Test

    ERIC Educational Resources Information Center

    Wan, Lei; Henly, George A.

    2012-01-01

    Many innovative item formats have been proposed over the past decade, but little empirical research has been conducted on their measurement properties. This study examines the reliability, efficiency, and construct validity of two innovative item formats--the figural response (FR) and constructed response (CR) formats used in a K-12 computerized…

  8. Empirically Derived Optimal Growth Equations For Hardwoods and Softwoods in Arkansas

    Treesearch

    Don C. Bragg

    2002-01-01

    Accurate growth projections are critical to reliable forest models, and ecologically based simulators can improve siivicultural predictions because of their sensitivity to change and their capacity to produce long-term forecasts. Potential relative increment (PRI) optimal diameter growth equations for loblolly pine, shortleaf pine, sweetgum, and white oak were fit to...

  9. Toward a Psychotherapy Integration Approach for Complex Post Traumatic Stress Disorder: A Critical Literature Review

    ERIC Educational Resources Information Center

    Confer, Jacob Russell

    2013-01-01

    The symptoms, assessment, and treatments of Post Traumatic Stress Disorder (PTSD) have been empirically investigated to the extent that there is a breadth of valid and reliable instruments investigating this psychopathological syndrome. There, too, exists a substantial evidence base for various treatment models demonstrating effectiveness in…

  10. Kids and Credibility: An Empirical Examination of Youth, Digital Media Use, and Information Credibility

    ERIC Educational Resources Information Center

    Flanagin, Andrew J.; Metzger, Miriam J.

    2010-01-01

    How well do children navigate the ocean of information that is available online? The enormous variety of Web-based resources represents both opportunities and challenges for Internet-savvy kids, offering extraordinary potential for learning and social connection but little guidance on assessing the reliability of online information. This book…

  11. Impact of Using an Educational Robot-Based Learning System on Students' Motivation in Elementary Education

    ERIC Educational Resources Information Center

    Chin, Kai-Yi; Hong, Zeng-Wei; Chen, Yen-Lin

    2014-01-01

    Educational robotics has been regarded as an effective instructional tool over the past decade. Many studies have tested the role of robots in supporting educational classroom activities. However, reliable empirical evidence confirming the effectiveness of educational robots remains limited. Therefore, this study developed an educational…

  12. Competencies in Geriatric Nursing: Empirical Evidence from a Computer-Based Large-Scale Assessment Calibration Study

    ERIC Educational Resources Information Center

    Kaspar, Roman; Döring, Ottmar; Wittmann, Eveline; Hartig, Johannes; Weyland, Ulrike; Nauerth, Annette; Möllers, Michaela; Rechenbach, Simone; Simon, Julia; Worofka, Iberé

    2016-01-01

    Valid and reliable standardized assessment of nursing competencies is needed to monitor the quality of vocational education and training (VET) in nursing and evaluate learning outcomes for care work trainees with increasingly heterogeneous learning backgrounds. To date, however, the modeling of professional competencies has not yet evolved into…

  13. An Empirically Keyed Scale for Measuring Managerial Attitudes toward Women Executives.

    ERIC Educational Resources Information Center

    Dubno, Peter; And Others

    1979-01-01

    A scale (Managerial Attitudes toward Women Executives Scale -- MATWES) provides reliability and validity measures regarding managerial attitudes toward women executives. It employs a projective test for item generation and uses a panel of women executives as Q-sorters to select items. The Scale and its value in minimizing researcher bias in its…

  14. Measuring the Reliability of Picture Story Exercises like the TAT

    PubMed Central

    Gruber, Nicole; Kreuzpointner, Ludwig

    2013-01-01

    As frequently reported, psychometric assessments on Picture Story Exercises, especially variations of the Thematic Apperception Test, mostly reveal inadequate scores for internal consistency. We demonstrate that the reason for this apparent shortcoming is not caused by the coding system itself but from the incorrect use of internal consistency coefficients, especially Cronbach’s α. This problem could be eliminated by using the category-scores as items instead of the picture-scores. In addition to a theoretical explanation we prove mathematically why the use of category-scores produces an adequate internal consistency estimation and examine our idea empirically with the origin data set of the Thematic Apperception Test by Heckhausen and two additional data sets. We found generally higher values when using the category-scores as items instead of picture-scores. From an empirical and theoretical point of view, the estimated reliability is also superior to each category within a picture as item measuring. When comparing our suggestion with a multifaceted Rasch-model we provide evidence that our procedure better fits the underlying principles of PSE. PMID:24348902

  15. [Risk epidemiology and child protection statistics in early childhood – a pilot study in southern Germany].

    PubMed

    Thurn, Leonore; Besier, Tanja; Ziegenhain, Ute; Jud, Andreas; Kindler, Heinz; Fischer, Dieter; Fegert, Jörg M; Künster, Anne Katrin

    2017-07-01

    In contrast to many other countries in Europe, Germany lacks sufficient empirical data on the incidence/prevalence of child maltreatment and its risk factors. This pilot study generated systematic data on the prevalence of child abuse and neglect and its risk factors in Germany. Using a newly developed questionnaire (ESM1 and ESM2) on child abuse and neglect as well as indicators for risk factors, we conducted a survey on 35 child daycare centers in a county in southern Germany, the goal being to generate reliable data. The questionnaire and the Strength and Difficulties Questionnaire (SDQ) was completed by daycare professionals for every child between 4 and under 7 years who was attending the participating daycare center (1,122 children). Child maltreatment was reported in 13.2 % of the cases, and risk factors for future child maltreatment were detected in 38.4 % cases. This study demonstrates that systematic data collection concerning child protection is feasible in child daycare centers. In the future, we recommend that local child protection networks be modified on the basis of reliable empirical data.

  16. SSM/I and ECMWF Wind Vector Comparison

    NASA Technical Reports Server (NTRS)

    Wentz, Frank J.; Ashcroft, Peter D.

    1996-01-01

    Wentz was the first to convincingly show that satellite microwave radiometers have the potential to measure the oceanic wind vector. The most compelling evidence for this conclusion was the monthly wind vector maps derived solely from a statistical analysis of Special Sensor Microwave Imager (SSM/I) observations. In a qualitative sense, these maps clearly showed the general circulation over the world's oceans. In this report we take a closer look at the SSM/I monthly wind vector maps and compare them to European Center for Medium-Range Weather Forecasts (ECMWF) wind fields. This investigation leads both to an empirical comparison of SSM/I calculated wind vectors with ECMWF wind vectors, and to an examination of possible reasons that the SSM/I calculated wind vector direction would be inherently more reliable at some locations than others.

  17. The typological approach in child and family psychology: a review of theory, methods, and research.

    PubMed

    Mandara, Jelani

    2003-06-01

    The purpose of this paper was to review the theoretical underpinnings, major concepts, and methods of the typological approach. It was argued that the typological approach offers a systematic, empirically rigorous and reliable way to synthesize the nomothetic variable-centered approach with the idiographic case-centered approach. Recent advances in cluster analysis validation make it a promising method for uncovering natural typologies. This paper also reviewed findings from personality and family studies that have revealed 3 prototypical personalities and parenting styles: Adjusted/Authoritative, Overcontrolled/Authoritarian, and Undercontrolled/Permissive. These prototypes are theorized to be synonymous with attractor basins in psychological state space. The connection between family types and personality structure as well as future directions of typological research were also discussed.

  18. Posttraumatic idioms of distress among Darfur refugees: Hozun and Majnun.

    PubMed

    Rasmussen, Andrew; Katoni, Basila; Keller, Allen S; Wilkinson, John

    2011-09-01

    Although psychosocial programming is seen as essential to the humanitarian response to the Darfur conflict, aid groups lack culturally-appropriate assessment instruments for monitoring and evaluation. The current study used an emic-etic integrated approach to: (i) create a culturally-appropriate measure of distress (Study 1), and (ii) test the measure in structured interviews of 848 Darfuris living in two refugee camps in Chad (Study 2). Traditional healers identified two trauma-related idioms, hozun and majnun, which shared features with but were not identical to posttraumatic stress disorder and depression. Measures of these constructs were reliable and correlated with trauma, loss, and functional impairment. Exploratory factor analysis resulted in empirical symptom clusters conceptually parallel to general Western psychiatric constructs. Findings are discussed in terms of their implications for psychosocial programming.

  19. Obtaining Reliable Predictions of Terrestrial Energy Coupling From Real-Time Solar Wind Measurements

    NASA Technical Reports Server (NTRS)

    Weimer, Daniel R.

    2002-01-01

    Measurements of the interplanetary magnetic field (IMF) from the ACE (Advanced Composition Explorer), Wind, IMP-8 (Interplanetary Monitoring Platform), and Geotail spacecraft have revealed that the IMF variations are contained in phase planes that are tilted with respect to the propagation direction, resulting in continuously variable changes in propagation times between spacecraft, and therefore, to the Earth. Techniques for using 'minimum variance analysis' have been developed in order to be able to measure the phase front tilt angles, and better predict the actual propagation times from the L1 orbit to the Earth, using only the real-time IMF measurements from one spacecraft. The use of empirical models with the IMF measurements at L1 from ACE (or future satellites) for predicting 'space weather' effects has also been demonstrated.

  20. Psychometric properties of the Brief Symptom Inventory-18 in a Spanish breast cancer sample.

    PubMed

    Galdón, Ma José; Durá, Estrella; Andreu, Yolanda; Ferrando, Maite; Murgui, Sergio; Pérez, Sandra; Ibañez, Elena

    2008-12-01

    The objective of this work was to study the psychometric and structural properties of the Brief Symptom Inventory-18 (BSI-18) in a sample of breast cancer patients (N=175). Confirmatory factor analyses were conducted. Two models were tested: the theoretical model with the original structure (three-dimensional), and the empirical model (a four-factor structure) obtained through exploratory factor analysis initially performed by the authors of the BSI-18. The eligible structure was the original proposal consisting of three dimensions: somatization, depression, and anxiety scores. These measures also showed good internal consistency. The results of this study support the reliability and structural validity of the BSI-18 as a standardized instrument for screening purposes in breast cancer patients, with the added benefits of simplicity and ease of application.

  1. Typology of Couples Entering Alcohol Behavioral Couple Therapy: An Empirical Approach and Test of Predictive Validity on Treatment Response

    PubMed Central

    Ladd, Benjamin O.; McCrady, Barbara S.

    2016-01-01

    The current study aimed to examine whether classification of couples in which one partner has an alcohol problem is similar to that reported in the general couples literature. Typologies of couples seeking Alcohol Behavioral Couple Therapy (ABCT) were developed via hierarchical cluster analysis using behavioral codes of couple interactions during their first ABCT session. Four couples types based on in-session behavior were established reliably, labeled Avoider, Validator, Hostile, and Ambivalent-Detached. These couple types resembled couples types found in previous research. Couple type was associated with baseline relationship satisfaction, but not alcohol use. Results suggest heterogeneity in couples with alcohol problems presenting to treatment; further study is needed to investigate the function of alcohol within these different types. PMID:25808432

  2. Development and Validation of the Conceptual Assessment of Natural Selection (CANS)

    PubMed Central

    Kalinowski, Steven T.; Leonard, Mary J.; Taper, Mark L.

    2016-01-01

    We developed and validated the Conceptual Assessment of Natural Selection (CANS), a multiple-choice test designed to assess how well college students understand the central principles of natural selection. The expert panel that reviewed the CANS concluded its questions were relevant to natural selection and generally did a good job sampling the specific concepts they were intended to assess. Student interviews confirmed questions on the CANS provided accurate reflections of how students think about natural selection. And, finally, statistical analysis of student responses using item response theory showed that the CANS did a very good job of estimating how well students understood natural selection. The empirical reliability of the CANS was substantially higher than the Force Concept Inventory, a highly regarded test in physics that has a similar purpose. PMID:27856552

  3. Patent information retrieval: approaching a method and analysing nanotechnology patent collaborations.

    PubMed

    Ozcan, Sercan; Islam, Nazrul

    2017-01-01

    Many challenges still remain in the processing of explicit technological knowledge documents such as patents. Given the limitations and drawbacks of the existing approaches, this research sets out to develop an improved method for searching patent databases and extracting patent information to increase the efficiency and reliability of nanotechnology patent information retrieval process and to empirically analyse patent collaboration. A tech-mining method was applied and the subsequent analysis was performed using Thomson data analyser software. The findings show that nations such as Korea and Japan are highly collaborative in sharing technological knowledge across academic and corporate organisations within their national boundaries, and China presents, in some cases, a great illustration of effective patent collaboration and co-inventorship. This study also analyses key patent strengths by country, organisation and technology.

  4. Item-level and subscale-level factoring of Biggs' Learning Process Questionnaire (LPQ) in a mainland Chinese sample.

    PubMed

    Sachs, J; Gao, L

    2000-09-01

    The learning process questionnaire (LPQ) has been the source of intensive cross-cultural study. However, an item-level factor analysis of all the LPQ items simultaneously has never been reported. Rather, items within each subscale have been factor analysed to establish subscale unidimensionality and justify the use of composite subscale scores. It was of major interest to see if the six logically constructed items groups of the LPQ would be supported by empirical evidence. Additionally, it was of interest to compare the consistency of the reliability and correlational structure of the LPQ subscales in our study with those of previous cross-cultural studies. Confirmatory factor analysis was used to fit the six-factor item level model and to fit five representative subscale level factor models. A total of 1070 students between the ages of 15 to 18 years was drawn from a representative selection of 29 classes from within 15 secondary schools in Guangzhou, China. Males and females were almost equally represented. The six-factor item level model of the LPQ seemed to fit reasonably well, thus supporting the six dimensional structure of the LPQ and justifying the use of composite subscale scores for each LPQ dimension. However, the reliability of many of these subscales was low. Furthermore, only two subscale-level factor models showed marginally acceptable fit. Substantive considerations supported an oblique three-factor model. Because the LPQ subscales often show low internal consistency reliability, experimental and correlational studies that have used these subscales as dependent measures have been disappointing. It is suggested that some LPQ items should be revised and other items added to improve the inventory's overall psychometric properties.

  5. The Employment Precariousness Scale (EPRES): psychometric properties of a new tool for epidemiological studies among waged and salaried workers.

    PubMed

    Vives, Alejandra; Amable, Marcelo; Ferrer, Montserrat; Moncada, Salvador; Llorens, Clara; Muntaner, Carles; Benavides, Fernando G; Benach, Joan

    2010-08-01

    Despite the fact that labour market flexibility has resulted in an expansion of precarious employment in industrialised countries, to date there is limited empirical evidence concerning its health consequences. The Employment Precariousness Scale (EPRES) is a newly developed, theory-based, multidimensional questionnaire specifically devised for epidemiological studies among waged and salaried workers. To assess the acceptability, reliability and construct validity of EPRES in a sample of waged and salaried workers in Spain. A sample of 6968 temporary and permanent workers from a population-based survey carried out in 2004-2005 was analysed. The survey questionnaire was interviewer administered and included the six EPRES subscales, and measures of the psychosocial work environment (COPSOQ ISTAS21) and perceived general and mental health (SF-36). A high response rate to all EPRES items indicated good acceptability; Cronbach's alpha coefficients, over 0.70 for all subscales and the global score, demonstrated good internal consistency reliability; exploratory factor analysis using principal axis analysis and varimax rotation confirmed the six-subscale structure and the theoretical allocation of all items. Patterns across known groups and correlation coefficients with psychosocial work environment measures and perceived health demonstrated the expected relations, providing evidence of construct validity. Our results provide evidence in support of the psychometric properties of EPRES, which appears to be a promising tool for the measurement of employment precariousness in public health research.

  6. Measuring stress in medical education: validation of the Korean version of the higher education stress inventory with medical students.

    PubMed

    Shim, Eun-Jung; Jeon, Hong Jin; Kim, Hana; Lee, Kwang-Min; Jung, Dooyoung; Noh, Hae-Lim; Roh, Myoung-Sun; Hahm, Bong-Jin

    2016-11-24

    Medical students face a variety of stressors associated with their education; if not promptly identified and adequately dealt with, it may bring about several negative consequences in terms of mental health and academic performance. This study examined psychometric properties of the Korean version of the Higher Education Stress Inventory (K-HESI). The reliability and validity of the K-HESI were examined in a large scale multi-site survey involving 7110 medical students. The K-HESI, Beck Depression Inventory (BDI) and questions regarding quality of life (QOL) and self-rated physical health (SPH) were administered. Exploratory factor analysis of the K-HESI identified seven factors: Low commitment; financial concerns; teacher-student relationship; worries about future profession; non-supportive climate; workload; and dissatisfaction with education. A subsequent confirmatory factor analysis supported the 7-factor model. Internal consistency of the K-HESI was satisfactory (Cronbach's α = .78). Convergent validity was demonstrated by its positive association with the BDI. Known group validity was supported by the K-HESI's ability to detect significant differences on the overall and subscale scores of K-HESI according to different levels of QOL and SPH. The K-HESI is a psychometrically valid tool that comprehensively assesses various relevant stressors related to medical education. Evidence-based stress management in medical education empirically guided by the regular assessment of stress using reliable and valid measure is warranted.

  7. Measuring acuity of the approximate number system reliably and validly: the evaluation of an adaptive test procedure

    PubMed Central

    Lindskog, Marcus; Winman, Anders; Juslin, Peter; Poom, Leo

    2013-01-01

    Two studies investigated the reliability and predictive validity of commonly used measures and models of Approximate Number System acuity (ANS). Study 1 investigated reliability by both an empirical approach and a simulation of maximum obtainable reliability under ideal conditions. Results showed that common measures of the Weber fraction (w) are reliable only when using a substantial number of trials, even under ideal conditions. Study 2 compared different purported measures of ANS acuity as for convergent and predictive validity in a within-subjects design and evaluated an adaptive test using the ZEST algorithm. Results showed that the adaptive measure can reduce the number of trials needed to reach acceptable reliability. Only direct tests with non-symbolic numerosity discriminations of stimuli presented simultaneously were related to arithmetic fluency. This correlation remained when controlling for general cognitive ability and perceptual speed. Further, the purported indirect measure of ANS acuity in terms of the Numeric Distance Effect (NDE) was not reliable and showed no sign of predictive validity. The non-symbolic NDE for reaction time was significantly related to direct w estimates in a direction contrary to the expected. Easier stimuli were found to be more reliable, but only harder (7:8 ratio) stimuli contributed to predictive validity. PMID:23964256

  8. Stochastic modeling of hourly rainfall times series in Campania (Italy)

    NASA Astrophysics Data System (ADS)

    Giorgio, M.; Greco, R.

    2009-04-01

    Occurrence of flowslides and floods in small catchments is uneasy to predict, since it is affected by a number of variables, such as mechanical and hydraulic soil properties, slope morphology, vegetation coverage, rainfall spatial and temporal variability. Consequently, landslide risk assessment procedures and early warning systems still rely on simple empirical models based on correlation between recorded rainfall data and observed landslides and/or river discharges. Effectiveness of such systems could be improved by reliable quantitative rainfall prediction, which can allow gaining larger lead-times. Analysis of on-site recorded rainfall height time series represents the most effective approach for a reliable prediction of local temporal evolution of rainfall. Hydrological time series analysis is a widely studied field in hydrology, often carried out by means of autoregressive models, such as AR, ARMA, ARX, ARMAX (e.g. Salas [1992]). Such models gave the best results when applied to the analysis of autocorrelated hydrological time series, like river flow or level time series. Conversely, they are not able to model the behaviour of intermittent time series, like point rainfall height series usually are, especially when recorded with short sampling time intervals. More useful for this issue are the so-called DRIP (Disaggregated Rectangular Intensity Pulse) and NSRP (Neymann-Scott Rectangular Pulse) model [Heneker et al., 2001; Cowpertwait et al., 2002], usually adopted to generate synthetic point rainfall series. In this paper, the DRIP model approach is adopted, in which the sequence of rain storms and dry intervals constituting the structure of rainfall time series is modeled as an alternating renewal process. Final aim of the study is to provide a useful tool to implement an early warning system for hydrogeological risk management. Model calibration has been carried out with hourly rainfall hieght data provided by the rain gauges of Campania Region civil protection agency meteorological warning network. ACKNOWLEDGEMENTS The research was co-financed by the Italian Ministry of University, by means of the PRIN 2006 PRIN program, within the research project entitled ‘Definition of critical rainfall thresholds for destructive landslides for civil protection purposes'. REFERENCES Cowpertwait, P.S.P., Kilsby, C.G. and O'Connell, P.E., 2002. A space-time Neyman-Scott model of rainfall: Empirical analysis of extremes, Water Resources Research, 38(8):1-14. Salas, J.D., 1992. Analysis and modeling of hydrological time series, in D.R. Maidment, ed., Handbook of Hydrology, McGraw-Hill, New York. Heneker, T.M., Lambert, M.F. and Kuczera G., 2001. A point rainfall model for risk-based design, Journal of Hydrology, 247(1-2):54-71.

  9. Evaluation of the Performance of Routine Information System Management (PRISM) framework: evidence from Uganda.

    PubMed

    Hotchkiss, David R; Aqil, Anwer; Lippeveld, Theo; Mukooyo, Edward

    2010-07-03

    Sound policy, resource allocation and day-to-day management decisions in the health sector require timely information from routine health information systems (RHIS). In most low- and middle-income countries, the RHIS is viewed as being inadequate in providing quality data and continuous information that can be used to help improve health system performance. In addition, there is limited evidence on the effectiveness of RHIS strengthening interventions in improving data quality and use. The purpose of this study is to evaluate the usefulness of the newly developed Performance of Routine Information System Management (PRISM) framework, which consists of a conceptual framework and associated data collection and analysis tools to assess, design, strengthen and evaluate RHIS. The specific objectives of the study are: a) to assess the reliability and validity of the PRISM instruments and b) to assess the validity of the PRISM conceptual framework. Facility- and worker-level data were collected from 110 health care facilities in twelve districts in Uganda in 2004 and 2007 using records reviews, structured interviews and self-administered questionnaires. The analysis procedures include Cronbach's alpha to assess internal consistency of selected instruments, test-retest analysis to assess the reliability and sensitivity of the instruments, and bivariate and multivariate statistical techniques to assess validity of the PRISM instruments and conceptual framework. Cronbach's alpha analysis suggests high reliability (0.7 or greater) for the indices measuring a promotion of a culture of information, RHIS tasks self-efficacy and motivation. The study results also suggest that a promotion of a culture of information influences RHIS tasks self-efficacy, RHIS tasks competence and motivation, and that self-efficacy and the presence of RHIS staff have a direct influence on the use of RHIS information, a key aspect of RHIS performance. The study results provide some empirical support for the reliability and validity of the PRISM instruments and the validity of the PRISM conceptual framework, suggesting that the PRISM approach can be effectively used by RHIS policy makers and practitioners to assess the RHIS and evaluate RHIS strengthening interventions. However, additional studies with larger sample sizes are needed to further investigate the value of the PRISM instruments in exploring the linkages between RHIS data quality and use, and health systems performance.

  10. The spaces in between: science, ocean, empire.

    PubMed

    Reidy, Michael S; Rozwadowski, Helen M

    2014-06-01

    Historians of science have richly documented the interconnections between science and empire in the nineteenth century. These studies primarily begin with Britain, Europe, or the United States at the center and have focused almost entirely on lands far off in the periphery--India or Australia, for instance. The spaces in between have received scant attention. Because use of the ocean in this period was infused with the doctrine of the freedom of the seas, the ocean was constructed as a space amenable to control by any nation that could master its surface and use its resources effectively. Oceans transformed in the mid-nineteenth century from highway to destination, becoming--among other things--the focus of sustained scientific interest for the first time in history. Use of the sea rested on reliable knowledge of the ocean. Particularly significant were the graphical representations of knowledge that could be passed from scientists to publishers to captains or other agents of empire. This process also motivated early government patronage of science and crystallized scientists' rising authority in society. The advance of science, the creation of empire, and the construction of the ocean were mutually sustaining.

  11. An empirical Bayes safety evaluation of tram/streetcar signal and lane priority measures in Melbourne.

    PubMed

    Naznin, Farhana; Currie, Graham; Sarvi, Majid; Logan, David

    2016-01-01

    Streetcars/tram systems are growing worldwide, and many are given priority to increase speed and reliability performance in mixed traffic conditions. Research related to the road safety impact of tram priority is limited. This study explores the road safety impacts of tram priority measures including lane and intersection/signal priority measures. A before-after crash study was conducted using the empirical Bayes (EB) method to provide more accurate crash impact estimates by accounting for wider crash trends and regression to the mean effects. Before-after crash data for 29 intersections with tram signal priority and 23 arterials with tram lane priority in Melbourne, Australia, were analyzed to evaluate the road safety impact of tram priority. The EB before-after analysis results indicated a statistically significant adjusted crash reduction rate of 16.4% after implementation of tram priority measures. Signal priority measures were found to reduce crashes by 13.9% and lane priority by 19.4%. A disaggregate level simple before-after analysis indicated reductions in total and serious crashes as well as vehicle-, pedestrian-, and motorcycle-involved crashes. In addition, reductions in on-path crashes, pedestrian-involved crashes, and collisions among vehicles moving in the same and opposite directions and all other specific crash types were found after tram priority implementation. Results suggest that streetcar/tram priority measures result in safety benefits for all road users, including vehicles, pedestrians, and cyclists. Policy implications and areas for future research are discussed.

  12. A Spanish Validation of the Canadian Adolescent Gambling Inventory (CAGI).

    PubMed

    Jiménez-Murcia, Susana; Granero, Roser; Stinchfield, Randy; Tremblay, Joël; Del Pino-Gutiérrez, Amparo; Moragas, Laura; Savvidou, Lamprini G; Fernández-Aranda, Fernando; Aymamí, Neus; Gómez-Peña, Mónica; Tárrega, Salomé; Gunnard, Katarina; Martín-Romera, Virginia; Steward, Trevor; Mestre-Bach, Gemma; Menchón, José M

    2017-01-01

    Aims: Large-scale epidemiological studies show a significant prevalence of gambling disorder (GD) during adolescence and emerging adulthood, and highlight the need to identify gambling-related behaviors at early ages. However, there are only a handful of screening instruments for this population and many studies measuring youth gambling problems use adult instruments that may not be developmentally appropriate. The aim of this study was to validate a Spanish version of the Canadian Adolescent Gambling Inventory (CAGI) among late adolescent and young adults and to explore its psychometric properties. Methods: The sample (16-29 years old) included a clinical group ( n = 55) with GD patients and a control group ( n = 340). Results: Exploratory factor analysis yielded one factor as the best model. This 24-item scale demonstrated satisfactory reliability (internal consistency, Cronbach's alpha, α = 0.91), satisfactory convergent validity as measured by correlation with South Oaks Gambling Screen ( r = 0.74), and excellent classification accuracy (AUC = 0.99; sensitivity = 0.98; and specificity = 0.99). Conclusion: Our results provide empirical support for our validation of the Spanish version of the CAGI. We uphold that the Spanish CAGI can be used as a brief, reliable, and valid instrument to assess gambling problems in Spanish youth.

  13. The development and psychometric properties of the American sign language proficiency assessment (ASL-PA).

    PubMed

    Maller, S; Singleton, J; Supalla, S; Wix, T

    1999-01-01

    We describe the procedures for constructing an instrument designed to evaluate children's proficiency in American Sign Language (ASL). The American Sign Language Proficiency Assessment (ASL-PA) is a much-needed tool that potentially could be used by researchers, language specialists, and qualified school personnel. A half-hour ASL sample is collected on video from a target child (between ages 6 and 12) across three separate discourse settings and is later analyzed and scored by an assessor who is highly proficient in ASL. After the child's language sample is scored, he or she can be assigned an ASL proficiency rating of Level 1, 2, or 3. At this phase in its development, substantial evidence of reliability and validity has been obtained for the ASL-PA using a sample of 80 profoundly deaf children (ages 6-12) of varying ASL skill levels. The article first explains the item development and administration of the ASL-PA instrument, then describes the empirical item analysis, standard setting procedures, and evidence of reliability and validity. The ASL-PA is a promising instrument for assessing elementary school-age children's ASL proficiency. Plans for further development are also discussed.

  14. An item response theory analysis of the Olweus Bullying scale.

    PubMed

    Breivik, Kyrre; Olweus, Dan

    2014-12-02

    In the present article, we used IRT (graded response) modeling as a useful technology for a detailed and refined study of the psychometric properties of the various items of the Olweus Bullying scale and the scale itself. The sample consisted of a very large number of Norwegian 4th-10th grade students (n = 48 926). The IRT analyses revealed that the scale was essentially unidimensional and had excellent reliability in the upper ranges of the latent bullying tendency trait, as intended and desired. Gender DIF effects were identified with regard to girls' use of indirect bullying by social exclusion and boys' use of physical bullying by hitting and kicking but these effects were small and worked in opposite directions, having negligible effects at the scale level. Also scale scores adjusted for DIF effects differed very little from non-adjusted scores. In conclusion, the empirical data were well characterized by the chosen IRT model and the Olweus Bullying scale was considered well suited for the conduct of fair and reliable comparisons involving different gender-age groups. Information Aggr. Behav. 9999:XX-XX, 2014. © 2014 Wiley Periodicals, Inc. © 2014 Wiley Periodicals, Inc.

  15. An item response theory analysis of the Olweus Bullying scale.

    PubMed

    Breivik, Kyrre; Olweus, Dan

    2015-01-01

    In the present article, we used IRT (graded response) modeling as a useful technology for a detailed and refined study of the psychometric properties of the various items of the Olweus Bullying scale and the scale itself. The sample consisted of a very large number of Norwegian 4th-10th grade students (n = 48 926). The IRT analyses revealed that the scale was essentially unidimensional and had excellent reliability in the upper ranges of the latent bullying tendency trait, as intended and desired. Gender DIF effects were identified with regard to girls' use of indirect bullying by social exclusion and boys' use of physical bullying by hitting and kicking but these effects were small and worked in opposite directions, having negligible effects at the scale level. Also scale scores adjusted for DIF effects differed very little from non-adjusted scores. In conclusion, the empirical data were well characterized by the chosen IRT model and the Olweus Bullying scale was considered well suited for the conduct of fair and reliable comparisons involving different gender-age groups. Information Aggr. Behav. 41:1-13, 2015. © 2014 Wiley Periodicals, Inc. © 2014 Wiley Periodicals, Inc.

  16. Trends and associated uncertainty in the global mean temperature record

    NASA Astrophysics Data System (ADS)

    Poppick, A. N.; Moyer, E. J.; Stein, M.

    2016-12-01

    Physical models suggest that the Earth's mean temperature warms in response to changing CO2 concentrations (and hence increased radiative forcing); given physical uncertainties in this relationship, the historical temperature record is a source of empirical information about global warming. A persistent thread in many analyses of the historical temperature record, however, is the reliance on methods that appear to deemphasize both physical and statistical assumptions. Examples include regression models that treat time rather than radiative forcing as the relevant covariate, and time series methods that account for natural variability in nonparametric rather than parametric ways. We show here that methods that deemphasize assumptions can limit the scope of analysis and can lead to misleading inferences, particularly in the setting considered where the data record is relatively short and the scale of temporal correlation is relatively long. A proposed model that is simple but physically informed provides a more reliable estimate of trends and allows a broader array of questions to be addressed. In accounting for uncertainty, we also illustrate how parametric statistical models that are attuned to the important characteristics of natural variability can be more reliable than ostensibly more flexible approaches.

  17. Quantifying interactions between real oscillators with information theory and phase models: application to cardiorespiratory coupling.

    PubMed

    Zhu, Yenan; Hsieh, Yee-Hsee; Dhingra, Rishi R; Dick, Thomas E; Jacono, Frank J; Galán, Roberto F

    2013-02-01

    Interactions between oscillators can be investigated with standard tools of time series analysis. However, these methods are insensitive to the directionality of the coupling, i.e., the asymmetry of the interactions. An elegant alternative was proposed by Rosenblum and collaborators [M. G. Rosenblum, L. Cimponeriu, A. Bezerianos, A. Patzak, and R. Mrowka, Phys. Rev. E 65, 041909 (2002); M. G. Rosenblum and A. S. Pikovsky, Phys. Rev. E 64, 045202 (2001)] which consists in fitting the empirical phases to a generic model of two weakly coupled phase oscillators. This allows one to obtain the interaction functions defining the coupling and its directionality. A limitation of this approach is that a solution always exists in the least-squares sense, even in the absence of coupling. To preclude spurious results, we propose a three-step protocol: (1) Determine if a statistical dependency exists in the data by evaluating the mutual information of the phases; (2) if so, compute the interaction functions of the oscillators; and (3) validate the empirical oscillator model by comparing the joint probability of the phases obtained from simulating the model with that of the empirical phases. We apply this protocol to a model of two coupled Stuart-Landau oscillators and show that it reliably detects genuine coupling. We also apply this protocol to investigate cardiorespiratory coupling in anesthetized rats. We observe reciprocal coupling between respiration and heartbeat and that the influence of respiration on the heartbeat is generally much stronger than vice versa. In addition, we find that the vagus nerve mediates coupling in both directions.

  18. Estimating canopy cover from standard forest inventory measurements in western Oregon

    Treesearch

    Anne McIntosh; Andrew Gray; Steven. Garman

    2012-01-01

    Reliable measures of canopy cover are important in the management of public and private forests. However, direct sampling of canopy cover is both labor- and time-intensive. More efficient methods for estimating percent canopy cover could be empirically derived relationships between more readily measured stand attributes and canopy cover or, alternatively, the use of...

  19. Using Item Data for Evaluating Criterion Reference Measures with an Empirical Investigation of Index Consistency.

    ERIC Educational Resources Information Center

    Meredith, Keith E.; Sabers, Darrell L.

    Data required for evaluating a Criterion Referenced Measurement (CRM) is described with a matrix. The information within the matrix consists of the "pass-fail" decisions of two CRMs. By differentially defining these two CRMs, different concepts of reliability and validity can be examined. Indices suggested for analyzing the matrix are listed with…

  20. Psychometric Evaluation of the Chinese Version of the Existential Anxiety Questionnaire in a Sample of Chinese Adolescents Living in Hong Kong

    ERIC Educational Resources Information Center

    To, Siu-ming; Chan, Wallace Chi-ho

    2016-01-01

    Background: While Western academia has increasingly recognized the importance of studying existential anxiety among adolescents, psychometrically valid and reliable tools for measuring this construct remain unavailable in Chinese societies. Objective: This research investigated the empirical viability of the construct of existential anxiety in…

  1. The Effect of Achievement Test Selection on Identification of Learning Disabilities within a Patterns of Strengths and Weaknesses Framework

    ERIC Educational Resources Information Center

    Miciak, Jeremy; Taylor, W. Pat; Denton, Carolyn A.; Fletcher, Jack M.

    2015-01-01

    Few empirical investigations have evaluated learning disabilities (LD) identification methods based on a pattern of cognitive strengths and weaknesses (PSW). This study investigated the reliability of LD classification decisions of the concordance/discordance method (C/DM) across different psychoeducational assessment batteries. C/DM criteria were…

  2. Appraising the reliability of visual impact assessment methods

    Treesearch

    Nickolaus R. Feimer; Kenneth H. Craik; Richard C. Smardon; Stephen R.J. Sheppard

    1979-01-01

    This paper presents the research approach and selected results of an empirical investigation aimed at the evaluation of selected observer-based visual impact assessment (VIA) methods. The VIA methods under examination were chosen to cover a range of VIA methods currently in use in both applied and research settings. Variation in three facets of VIA methods were...

  3. Reliability of the Attitudes Toward Women Scale (AWS) and the Personal Attributes Questionnaire (PAQ).

    ERIC Educational Resources Information Center

    Yoder, Jan D.; And Others

    In recent years, there has been a dramatic increase in the volume of empirical research directed toward the issue of sex-roles, including the development of evaluative instruments such as the Attitudes Toward Women Scale (AWS) and the Personal Attributes Questionnaire (PAQ). The United States Military Academy's Project Athena, designed to examine…

  4. Motivation for Knowledge Sharing by Expert Participants in Company-Hosted Online User Communities

    ERIC Educational Resources Information Center

    Cheng, Jingli

    2014-01-01

    Company-hosted online user communities are increasingly popular as firms continue to search for ways to provide their customers with high quality and reliable support in a low cost and scalable way. Yet, empirical understanding of motivations for knowledge sharing in this type of online communities is lacking, especially with regard to an…

  5. A Cartoon-Based Measure of PTSD Symptomatology in Children Exposed to a Disaster

    ERIC Educational Resources Information Center

    Elklit, Ask; Nielsen, Louise Hjort; Lasgaard, Mathias; Duch, Christina

    2013-01-01

    Research on childhood posttraumatic stress disorder (PTSD) is sparse. This is partly due to the limited availability of empirically validated measures for children who are insecure readers. The present study examined the reliability and validity of a cartoon-based measure of PTSD symptoms in children exposed to a disaster. Cartoons were generated…

  6. Wanting to Learn: A Necessary Condition for the Effectiveness of Instructional Design

    ERIC Educational Resources Information Center

    Gropper, George L.

    2015-01-01

    There are inevitable challenges standing in the way of the success of any approach to Instructional Design. Some are being faced up to. Some are not. Not least among them is the absence of empirically established principles undergirding Instructional Design prescriptions. The challenge it poses is to their reliability and validity. Nothing that…

  7. What the Research Tells Us about the Impact of Induction and Mentoring Programs for Beginning Teachers

    ERIC Educational Resources Information Center

    Ingersoll, Richard; Strong, Michael

    2012-01-01

    This chapter summarizes a comprehensive and critical review that the authors recently completed of empirical studies that evaluate the effects of induction on various outcomes. The review's objective was to provide researchers, policy makers, and educators with a reliable and current assessment of what is known and not known about the…

  8. Empirical Approaches to Measuring the Intelligibility of Different Varieties of English in Predicting Listener Comprehension

    ERIC Educational Resources Information Center

    Kang, Okim; Thomson, Ron I.; Moran, Meghan

    2018-01-01

    This study compared five research-based intelligibility measures as they were applied to six varieties of English. The objective was to determine which approach to measuring intelligibility would be most reliable for predicting listener comprehension, as measured through a listening comprehension test similar to the Test of English as a Foreign…

  9. The "Public Opinion Survey of Human Attributes-Stuttering" (POSHA-S): Summary Framework and Empirical Comparisons

    ERIC Educational Resources Information Center

    St. Louis, Kenneth O.

    2011-01-01

    Purpose: The "Public Opinion Survey of Human Attributes-Stuttering" ("POSHA-S") was developed to make available worldwide a standard measure of public attitudes toward stuttering that is practical, reliable, valid, and translatable. Mean data from past field studies as comparisons for interpretation of "POSHA-S" results are reported. Method: Means…

  10. A comprehensive review of the psychometric properties of the Drug Abuse Screening Test.

    PubMed

    Yudko, Errol; Lozhkina, Olga; Fouts, Adriana

    2007-03-01

    This article reviews the reliability and the validity of the (10-, 20-, and 28-item) Drug Abuse Screening Test (DAST). The reliability and the validity of the adolescent version of the DAST are also reviewed. An extensive literature review was conducted using the Medline and Psychinfo databases from the years 1982 to 2005. All articles that addressed the reliability and the validity of the DAST were examined. Publications in which the DAST was used as a screening tool but had no data on its psychometric properties were not included. Descriptive information about each version of the test, as well as discussion of the empirical literature that has explored measures of the reliability and the validity of the DAST, has been included. The DAST tended to have moderate to high levels of test-retest, interitem, and item-total reliabilities. The DAST also tended to have moderate to high levels of validity, sensitivity, and specificity. In general, all versions of the DAST yield satisfactory measures of reliability and validity for use as clinical or research tools. Furthermore, these tests are easy to administer and have been used in a variety of populations.

  11. Goal setting as an outcome measure: A systematic review.

    PubMed

    Hurn, Jane; Kneebone, Ian; Cropley, Mark

    2006-09-01

    Goal achievement has been considered to be an important measure of outcome by clinicians working with patients in physical and neurological rehabilitation settings. This systematic review was undertaken to examine the reliability, validity and sensitivity of goal setting and goal attainment scaling approaches when used with working age and older people. To review the reliability, validity and sensitivity of both goal setting and goal attainment scaling when employed as an outcome measure within a physical and neurological working age and older person rehabilitation environment, by examining the research literature covering the 36 years since goal-setting theory was proposed. Data sources included a computer-aided literature search of published studies examining the reliability, validity and sensitivity of goal setting/goal attainment scaling, with further references sourced from articles obtained through this process. There is strong evidence for the reliability, validity and sensitivity of goal attainment scaling. Empirical support was found for the validity of goal setting but research demonstrating its reliability and sensitivity is limited. Goal attainment scaling appears to be a sound measure for use in physical rehabilitation settings with working age and older people. Further work needs to be carried out with goal setting to establish its reliability and sensitivity as a measurement tool.

  12. Empirical dual energy calibration (EDEC) for cone-beam computed tomography.

    PubMed

    Stenner, Philip; Berkus, Timo; Kachelriess, Marc

    2007-09-01

    Material-selective imaging using dual energy CT (DECT) relies heavily on well-calibrated material decomposition functions. These require the precise knowledge of the detected x-ray spectra, and even if they are exactly known the reliability of DECT will suffer from scattered radiation. We propose an empirical method to determine the proper decomposition function. In contrast to other decomposition algorithms our empirical dual energy calibration (EDEC) technique requires neither knowledge of the spectra nor of the attenuation coefficients. The desired material-selective raw data p1 and p2 are obtained as functions of the measured attenuation data q1 and q2 (one DECT scan = two raw data sets) by passing them through a polynomial function. The polynomial's coefficients are determined using a general least squares fit based on thresholded images of a calibration phantom. The calibration phantom's dimension should be of the same order of magnitude as the test object, but other than that no assumptions on its exact size or positioning are made. Once the decomposition coefficients are determined DECT raw data can be decomposed by simply passing them through the polynomial. To demonstrate EDEC simulations of an oval CTDI phantom, a lung phantom, a thorax phantom and a mouse phantom were carried out. The method was further verified by measuring a physical mouse phantom, a half-and-half-cylinder phantom and a Yin-Yang phantom with a dedicated in vivo dual source micro-CT scanner. The raw data were decomposed into their components, reconstructed, and the pixel values obtained were compared to the theoretical values. The determination of the calibration coefficients with EDEC is very robust and depends only slightly on the type of calibration phantom used. The images of the test phantoms (simulations and measurements) show a nearly perfect agreement with the theoretical micro values and density values. Since EDEC is an empirical technique it inherently compensates for scatter components. The empirical dual energy calibration technique is a pragmatic, simple, and reliable calibration approach that produces highly quantitative DECT images.

  13. Global computer-assisted appraisal of osteoporosis risk in Asian women: an innovative study.

    PubMed

    Chang, Shu F; Hong, Chin M; Yang, Rong S

    2011-05-01

    To develop a computer-assisted appraisal system of osteoporosis that can predict osteoporosis health risk in community-dwelling women and to use it in an empirical analysis of the risk in Asian women. As the literature indicates, health risk assessment tools are generally applied in clinical practice for patient diagnosis. However, few studies have explored how to assist community-dwelling women to understand the risk of osteoporosis without invasive data. A longitudinal, evidence-based study. The first stage of this study is to establish a system that combines expertise in nursing, medicine and information technology. This part includes information from random samples (n = 700), including data on bone mineral density, osteoporosis risk factors, knowledge, beliefs and behaviour, which are used as the health risk appraisal system database. The second stage is to apply an empirical study. The relative risks of osteoporosis of the participants (n = 300) were determined with the system. The participants that were classified as at-risk were randomly grouped into experimental and control groups. Each group was treated using different nursing intervention methods. The sensitivity and specificity of the analytical tools was 75%. In empirical study, analysis results indicate that the prevalence of osteoporosis was 14.0%. Data indicate that strategic application of multiple nursing interventions can promote osteoporosis prevention knowledge in high-risk women and enhance the effectiveness of preventive action. The system can also provide people in remote areas or with insufficient medical resources a simple and effective means of managing health risk and implement the idea of self-evaluation and self-caring among community-dwelling women at home to achieve the final goal of early detection and early treatment of osteoporosis. This study developed a useful approach for providing Asia women with a reliable, valid, convenient and economical self-health management model. Health care professionals can explore the use of advanced information systems and nursing interventions to increase the effectiveness of osteoporosis prevention programmes for women. © 2011 Blackwell Publishing Ltd.

  14. The Development and Validation of the Online Shopping Addiction Scale.

    PubMed

    Zhao, Haiyan; Tian, Wei; Xin, Tao

    2017-01-01

    We report the development and validation of a scale to measure online shopping addiction. Inspired by previous theories and research on behavioral addiction, the Griffiths's widely accepted six-factor component model was referred to and an 18-item scale was constructed, with each component measured by three items. The results of exploratory factor analysis, based on Sample 1 (999 college students) and confirmatory factor analysis, based on Sample 2 (854 college students) showed the Griffiths's substantive six-factor structure underlay the online shopping addiction scale. Cronbach's alpha suggested that the resulting scale was highly reliable. Concurrent validity, based on Sample 3 (328 college students), was also satisfactory as indicated by correlations between the scale and measures of similar constructs. Finally, self-perceived online shopping addiction can be predicted to a relatively high degree. The present 18-item scale is a solid theory-based instrument to empirically measure online shopping addiction and can be used for understanding the phenomena among young adults.

  15. The Development and Validation of the Online Shopping Addiction Scale

    PubMed Central

    Zhao, Haiyan; Tian, Wei; Xin, Tao

    2017-01-01

    We report the development and validation of a scale to measure online shopping addiction. Inspired by previous theories and research on behavioral addiction, the Griffiths's widely accepted six-factor component model was referred to and an 18-item scale was constructed, with each component measured by three items. The results of exploratory factor analysis, based on Sample 1 (999 college students) and confirmatory factor analysis, based on Sample 2 (854 college students) showed the Griffiths's substantive six-factor structure underlay the online shopping addiction scale. Cronbach's alpha suggested that the resulting scale was highly reliable. Concurrent validity, based on Sample 3 (328 college students), was also satisfactory as indicated by correlations between the scale and measures of similar constructs. Finally, self-perceived online shopping addiction can be predicted to a relatively high degree. The present 18-item scale is a solid theory-based instrument to empirically measure online shopping addiction and can be used for understanding the phenomena among young adults. PMID:28559864

  16. An empirical evaluation of three vibrational spectroscopic methods for detection of aflatoxins in maize.

    PubMed

    Lee, Kyung-Min; Davis, Jessica; Herrman, Timothy J; Murray, Seth C; Deng, Youjun

    2015-04-15

    Three commercially available vibrational spectroscopic techniques, including Raman, Fourier transform near infrared reflectance (FT-NIR), and Fourier transform infrared (FTIR) were evaluated to help users determine the spectroscopic method best suitable for aflatoxin analysis in maize (Zea mays L.) grain based on their relative efficiency and predictive ability. Spectral differences of Raman and FTIR spectra were more marked and pronounced among aflatoxin contamination groups than those of FT-NIR spectra. From the observations and findings in our current and previous studies, Raman and FTIR spectroscopic methods are superior to FT-NIR method in terms of predictive power and model performance for aflatoxin analysis and they are equally effective and accurate in predicting aflatoxin concentration in maize. The present study is considered as the first attempt to assess how spectroscopic techniques with different physical processes can influence and improve accuracy and reliability for rapid screening of aflatoxin contaminated maize samples. Copyright © 2014 Elsevier Ltd. All rights reserved.

  17. The parametric modified limited penetrable visibility graph for constructing complex networks from time series

    NASA Astrophysics Data System (ADS)

    Li, Xiuming; Sun, Mei; Gao, Cuixia; Han, Dun; Wang, Minggang

    2018-02-01

    This paper presents the parametric modified limited penetrable visibility graph (PMLPVG) algorithm for constructing complex networks from time series. We modify the penetrable visibility criterion of limited penetrable visibility graph (LPVG) in order to improve the rationality of the original penetrable visibility and preserve the dynamic characteristics of the time series. The addition of view angle provides a new approach to characterize the dynamic structure of the time series that is invisible in the previous algorithm. The reliability of the PMLPVG algorithm is verified by applying it to three types of artificial data as well as the actual data of natural gas prices in different regions. The empirical results indicate that PMLPVG algorithm can distinguish the different time series from each other. Meanwhile, the analysis results of natural gas prices data using PMLPVG are consistent with the detrended fluctuation analysis (DFA). The results imply that the PMLPVG algorithm may be a reasonable and significant tool for identifying various time series in different fields.

  18. Validating Quantitative Measurement Using Qualitative Data: Combining Rasch Scaling and Latent Semantic Analysis in Psychiatry

    NASA Astrophysics Data System (ADS)

    Lange, Rense

    2015-02-01

    An extension of concurrent validity is proposed that uses qualitative data for the purpose of validating quantitative measures. The approach relies on Latent Semantic Analysis (LSA) which places verbal (written) statements in a high dimensional semantic space. Using data from a medical / psychiatric domain as a case study - Near Death Experiences, or NDE - we established concurrent validity by connecting NDErs qualitative (written) experiential accounts with their locations on a Rasch scalable measure of NDE intensity. Concurrent validity received strong empirical support since the variance in the Rasch measures could be predicted reliably from the coordinates of their accounts in the LSA derived semantic space (R2 = 0.33). These coordinates also predicted NDErs age with considerable precision (R2 = 0.25). Both estimates are probably artificially low due to the small available data samples (n = 588). It appears that Rasch scalability of NDE intensity is a prerequisite for these findings, as each intensity level is associated (at least probabilistically) with a well- defined pattern of item endorsements.

  19. edgeR: a Bioconductor package for differential expression analysis of digital gene expression data.

    PubMed

    Robinson, Mark D; McCarthy, Davis J; Smyth, Gordon K

    2010-01-01

    It is expected that emerging digital gene expression (DGE) technologies will overtake microarray technologies in the near future for many functional genomics applications. One of the fundamental data analysis tasks, especially for gene expression studies, involves determining whether there is evidence that counts for a transcript or exon are significantly different across experimental conditions. edgeR is a Bioconductor software package for examining differential expression of replicated count data. An overdispersed Poisson model is used to account for both biological and technical variability. Empirical Bayes methods are used to moderate the degree of overdispersion across transcripts, improving the reliability of inference. The methodology can be used even with the most minimal levels of replication, provided at least one phenotype or experimental condition is replicated. The software may have other applications beyond sequencing data, such as proteome peptide count data. The package is freely available under the LGPL licence from the Bioconductor web site (http://bioconductor.org).

  20. Analysis of Network Address Shuffling as a Moving Target Defense

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Carroll, Thomas E.; Crouse, Michael B.; Fulp, Errin W.

    2014-06-10

    Address shuffling is a type of moving target defense that prevents an attacker from reliably contacting a system by periodically remapping network addresses. Although limited testing has demonstrated it to be effective, little research has been conducted to examine the theoretical limits of address shuffling. As a result, it is difficult to understand how effective shuffling is and under what circumstances it is a viable moving target defense. This paper introduces probabilistic models that can provide insight into the performance of address shuffling. These models quantify the probability of attacker success in terms of network size, quantity of addresses scanned,more » quantity of vulnerable systems, and the frequency of shuffling. Theoretical analysis will show that shuffling is an acceptable defense if there is a small population of vulnerable systems within a large network address space, however shuffling has a cost for legitimate users. These results will also be shown empirically using simulation and actual traffic traces.« less

  1. Nonlinear mode decomposition: A noise-robust, adaptive decomposition method

    NASA Astrophysics Data System (ADS)

    Iatsenko, Dmytro; McClintock, Peter V. E.; Stefanovska, Aneta

    2015-09-01

    The signals emanating from complex systems are usually composed of a mixture of different oscillations which, for a reliable analysis, should be separated from each other and from the inevitable background of noise. Here we introduce an adaptive decomposition tool—nonlinear mode decomposition (NMD)—which decomposes a given signal into a set of physically meaningful oscillations for any wave form, simultaneously removing the noise. NMD is based on the powerful combination of time-frequency analysis techniques—which, together with the adaptive choice of their parameters, make it extremely noise robust—and surrogate data tests used to identify interdependent oscillations and to distinguish deterministic from random activity. We illustrate the application of NMD to both simulated and real signals and demonstrate its qualitative and quantitative superiority over other approaches, such as (ensemble) empirical mode decomposition, Karhunen-Loève expansion, and independent component analysis. We point out that NMD is likely to be applicable and useful in many different areas of research, such as geophysics, finance, and the life sciences. The necessary matlab codes for running NMD are freely available for download.

  2. Morphological cladistic analysis of eight popular Olive (Olea europaea L.) cultivars grown in Saudi Arabia using Numerical Taxonomic System for personal computer to detect phyletic relationship and their proximate fruit composition

    PubMed Central

    Al-Ruqaie, I.; Al-Khalifah, N.S.; Shanavaskhan, A.E.

    2015-01-01

    Varietal identification of olives is an intrinsic and empirical exercise owing to the large number of synonyms and homonyms, intensive exchange of genotypes, presence of varietal clones and lack of proper certification in nurseries. A comparative study of morphological characters of eight olive cultivars grown in Saudi Arabia was carried out and analyzed using NTSYSpc (Numerical Taxonomy System for personal computer) system segregated smaller fruits in one clade and the rest in two clades. Koroneiki, a Greek cultivar with a small sized fruit shared arm with Spanish variety Arbosana. Morphologic analysis using NTSYSpc revealed that biometrics of leaves, fruits and seeds are reliable morphologic characters to distinguish between varieties, except for a few morphologically very similar olive cultivars. The proximate analysis showed significant variations in the protein, fiber, crude fat, ash and moisture content of different cultivars. The study also showed that neither the size of fruit nor the fruit pulp thickness is a limiting factor determining crude fat content of olives. PMID:26858547

  3. Empirical research on the influencing factors of the occupational stress for construction workers

    NASA Astrophysics Data System (ADS)

    LV, Xing; WU, Xiang; CI, Huipeng; LIU, Qing; YAO, Yongzheng

    2017-04-01

    Employee’s occupational stress and safety performance are highly related, which has been generally recognized by the researchers. We did this research to understand the status of the stress for construction workers, and explore the influence factors of pressure source with characteristics of construction industry. Based on the results of previous studies, we designed questionnaire to collect the influence factors on occupational stressors. The study selected workers from certain construction units at the grass-roots level as sample source. According to the results of the questionnaire, we redesigned the interview outline, and did the semi-structured interviews on workers randomly selected. Finally, we developed a scale which combined the characteristics of construction projects in China. Using SPSS software for factor analysis, reliability analysis, and descriptive statistical analysis, the results show that there are six factors affecting the workers’ occupational stress, including The Work Itself, Family-Work, Career Development, Organization Style, Interpersonal Relationship and Role Management Style. The work itself is the main sources of occupational stress. The results can be used by the construction company to provide guidance for workers to control and manage occupational stress.

  4. EGG: Empirical Galaxy Generator

    NASA Astrophysics Data System (ADS)

    Schreiber, C.; Elbaz, D.; Pannella, M.; Merlin, E.; Castellano, M.; Fontana, A.; Bourne, N.; Boutsia, K.; Cullen, F.; Dunlop, J.; Ferguson, H. C.; Michałowski, M. J.; Okumura, K.; Santini, P.; Shu, X. W.; Wang, T.; White, C.

    2018-04-01

    The Empirical Galaxy Generator (EGG) generates fake galaxy catalogs and images with realistic positions, morphologies and fluxes from the far-ultraviolet to the far-infrared. The catalogs are generated by egg-gencat and stored in binary FITS tables (column oriented). Another program, egg-2skymaker, is used to convert the generated catalog into ASCII tables suitable for ingestion by SkyMaker (ascl:1010.066) to produce realistic high resolution images (e.g., Hubble-like), while egg-gennoise and egg-genmap can be used to generate the low resolution images (e.g., Herschel-like). These tools can be used to test source extraction codes, or to evaluate the reliability of any map-based science (stacking, dropout identification, etc.).

  5. Sistematización de las diferencias entre las estimaciones de modelos de extinción y las observaciones en la Vía Láctea

    NASA Astrophysics Data System (ADS)

    Perren, G.; Vázquez, R. A.; Navone, H.

    This paper analyses the reliability of the reddening estimates, extended to the entire sky, from two new Galaxy models built by Amores & Lépine (2005), using as a source of empirical data the database of open star clusters WEBDA. We also used the 100 um maps by Schlegel et al. (1998). It is concluded that the predictions of the Amores & Lépine models have a good correlation with empirical values until a relatively close distance to the Sun, while the Schlegel et al. model do not match the reddening estimation within the Milky Way. FULL TEXT IN SPANISH

  6. Semi-empirical quantum evaluation of peptide - MHC class II binding

    NASA Astrophysics Data System (ADS)

    González, Ronald; Suárez, Carlos F.; Bohórquez, Hugo J.; Patarroyo, Manuel A.; Patarroyo, Manuel E.

    2017-01-01

    Peptide presentation by the major histocompatibility complex (MHC) is a key process for triggering a specific immune response. Studying peptide-MHC (pMHC) binding from a structural-based approach has potential for reducing the costs of investigation into vaccine development. This study involved using two semi-empirical quantum chemistry methods (PM7 and FMO-DFTB) for computing the binding energies of peptides bonded to HLA-DR1 and HLA-DR2. We found that key stabilising water molecules involved in the peptide binding mechanism were required for finding high correlation with IC50 experimental values. Our proposal is computationally non-intensive, and is a reliable alternative for studying pMHC binding interactions.

  7. Inter-rater reliability of an observation-based ergonomics assessment checklist for office workers.

    PubMed

    Pereira, Michelle Jessica; Straker, Leon Melville; Comans, Tracy Anne; Johnston, Venerina

    2016-12-01

    To establish the inter-rater reliability of an observation-based ergonomics assessment checklist for computer workers. A 37-item (38-item if a laptop was part of the workstation) comprehensive observational ergonomics assessment checklist comparable to government guidelines and up to date with empirical evidence was developed. Two trained practitioners assessed full-time office workers performing their usual computer-based work and evaluated the suitability of workstations used. Practitioners assessed each participant consecutively. The order of assessors was randomised, and the second assessor was blinded to the findings of the first. Unadjusted kappa coefficients between the raters were obtained for the overall checklist and subsections that were formed from question-items relevant to specific workstation equipment. Twenty-seven office workers were recruited. The inter-rater reliability between two trained practitioners achieved moderate to good reliability for all except one checklist component. This checklist has mostly moderate to good reliability between two trained practitioners. Practitioner Summary: This reliable ergonomics assessment checklist for computer workers was designed using accessible government guidelines and supplemented with up-to-date evidence. Employers in Queensland (Australia) can fulfil legislative requirements by using this reliable checklist to identify and subsequently address potential risk factors for work-related injury to provide a safe working environment.

  8. Inter-agency communication and operations capabilities during a hospital functional exercise: reliability and validity of a measurement tool.

    PubMed

    Savoia, Elena; Biddinger, Paul D; Burstein, Jon; Stoto, Michael A

    2010-01-01

    As proxies for actual emergencies, drills and exercises can raise awareness, stimulate improvements in planning and training, and provide an opportunity to examine how different components of the public health system would combine to respond to a challenge. Despite these benefits, there remains a substantial need for widely accepted and prospectively validated tools to evaluate agencies' and hospitals' performance during such events. Unfortunately, to date, few studies have focused on addressing this need. The purpose of this study was to assess the validity and reliability of a qualitative performance assessment tool designed to measure hospitals' communication and operational capabilities during a functional exercise. The study population included 154 hospital personnel representing nine hospitals that participated in a functional exercise in Massachusetts in June 2008. A 25-item questionnaire was developed to assess the following three hospital functional capabilities: (1) inter-agency communication; (2) communication with the public; and (3) disaster operations. Analyses were conducted to examine internal consistency, associations among scales, the empirical structure of the items, and inter-rater agreement. Twenty-two questions were retained in the final instrument, which demonstrated reliability with alpha coefficients of 0.83 or higher for all scales. A three-factor solution from the principal components analysis accounted for 57% of the total variance, and the factor structure was consistent with the original hypothesized domains. Inter-rater agreement between participants' self reported scores and external evaluators' scores ranged from moderate to good. The resulting 22-item performance measurement tool reliably measured hospital capabilities in a functional exercise setting, with preliminary evidence of concurrent and criterion-related validity.

  9. Structural Patterns in Empirical Research Articles: A Cross-Disciplinary Study

    ERIC Educational Resources Information Center

    Lin, Ling; Evans, Stephen

    2012-01-01

    This paper presents an analysis of the major generic structures of empirical research articles (RAs), with a particular focus on disciplinary variation and the relationship between the adjacent sections in the introductory and concluding parts. The findings were derived from a close "manual" analysis of 433 recent empirical RAs from high-impact…

  10. The Outcome and Assessment Information Set (OASIS): A Review of Validity and Reliability

    PubMed Central

    O’CONNOR, MELISSA; DAVITT, JOAN K.

    2015-01-01

    The Outcome and Assessment Information Set (OASIS) is the patient-specific, standardized assessment used in Medicare home health care to plan care, determine reimbursement, and measure quality. Since its inception in 1999, there has been debate over the reliability and validity of the OASIS as a research tool and outcome measure. A systematic literature review of English-language articles identified 12 studies published in the last 10 years examining the validity and reliability of the OASIS. Empirical findings indicate the validity and reliability of the OASIS range from low to moderate but vary depending on the item studied. Limitations in the existing research include: nonrepresentative samples; inconsistencies in methods used, items tested, measurement, and statistical procedures; and the changes to the OASIS itself over time. The inconsistencies suggest that these results are tentative at best; additional research is needed to confirm the value of the OASIS for measuring patient outcomes, research, and quality improvement. PMID:23216513

  11. Statistical significance test for transition matrices of atmospheric Markov chains

    NASA Technical Reports Server (NTRS)

    Vautard, Robert; Mo, Kingtse C.; Ghil, Michael

    1990-01-01

    Low-frequency variability of large-scale atmospheric dynamics can be represented schematically by a Markov chain of multiple flow regimes. This Markov chain contains useful information for the long-range forecaster, provided that the statistical significance of the associated transition matrix can be reliably tested. Monte Carlo simulation yields a very reliable significance test for the elements of this matrix. The results of this test agree with previously used empirical formulae when each cluster of maps identified as a distinct flow regime is sufficiently large and when they all contain a comparable number of maps. Monte Carlo simulation provides a more reliable way to test the statistical significance of transitions to and from small clusters. It can determine the most likely transitions, as well as the most unlikely ones, with a prescribed level of statistical significance.

  12. SMART empirical approaches for predicting field performance of PV modules from results of reliability tests

    NASA Astrophysics Data System (ADS)

    Hardikar, Kedar Y.; Liu, Bill J. J.; Bheemreddy, Venkata

    2016-09-01

    Gaining an understanding of degradation mechanisms and their characterization are critical in developing relevant accelerated tests to ensure PV module performance warranty over a typical lifetime of 25 years. As newer technologies are adapted for PV, including new PV cell technologies, new packaging materials, and newer product designs, the availability of field data over extended periods of time for product performance assessment cannot be expected within the typical timeframe for business decisions. In this work, to enable product design decisions and product performance assessment for PV modules utilizing newer technologies, Simulation and Mechanism based Accelerated Reliability Testing (SMART) methodology and empirical approaches to predict field performance from accelerated test results are presented. The method is demonstrated for field life assessment of flexible PV modules based on degradation mechanisms observed in two accelerated tests, namely, Damp Heat and Thermal Cycling. The method is based on design of accelerated testing scheme with the intent to develop relevant acceleration factor models. The acceleration factor model is validated by extensive reliability testing under different conditions going beyond the established certification standards. Once the acceleration factor model is validated for the test matrix a modeling scheme is developed to predict field performance from results of accelerated testing for particular failure modes of interest. Further refinement of the model can continue as more field data becomes available. While the demonstration of the method in this work is for thin film flexible PV modules, the framework and methodology can be adapted to other PV products.

  13. Random-effects meta-analysis: the number of studies matters.

    PubMed

    Guolo, Annamaria; Varin, Cristiano

    2017-06-01

    This paper investigates the impact of the number of studies on meta-analysis and meta-regression within the random-effects model framework. It is frequently neglected that inference in random-effects models requires a substantial number of studies included in meta-analysis to guarantee reliable conclusions. Several authors warn about the risk of inaccurate results of the traditional DerSimonian and Laird approach especially in the common case of meta-analysis involving a limited number of studies. This paper presents a selection of likelihood and non-likelihood methods for inference in meta-analysis proposed to overcome the limitations of the DerSimonian and Laird procedure, with a focus on the effect of the number of studies. The applicability and the performance of the methods are investigated in terms of Type I error rates and empirical power to detect effects, according to scenarios of practical interest. Simulation studies and applications to real meta-analyses highlight that it is not possible to identify an approach uniformly superior to alternatives. The overall recommendation is to avoid the DerSimonian and Laird method when the number of meta-analysis studies is modest and prefer a more comprehensive procedure that compares alternative inferential approaches. R code for meta-analysis according to all of the inferential methods examined in the paper is provided.

  14. Calculation of skin-friction coefficients for low Reynolds number turbulent boundary layer flows. M.S. Thesis - California Univ. at Davis

    NASA Technical Reports Server (NTRS)

    Barr, P. K.

    1980-01-01

    An analysis is presented of the reliability of various generally accepted empirical expressions for the prediction of the skin-friction coefficient C/sub f/ of turbulent boundary layers at low Reynolds numbers in zero-pressure-gradient flows on a smooth flat plate. The skin-friction coefficients predicted from these expressions were compared to the skin-friction coefficients of experimental profiles that were determined from a graphical method formulated from the law of the wall. These expressions are found to predict values that are consistently different than those obtained from the graphical method over the range 600 Re/sub theta 2000. A curve-fitted empirical relationship was developed from the present data and yields a better estimated value of C/sub f/ in this range. The data, covering the range 200 Re/sub theta 7000, provide insight into the nature of transitional flows. They show that fully developed turbulent boundary layers occur at Reynolds numbers Re/sub theta/ down to 425. Below this level there appears to be a well-ordered evolutionary process from the laminar to the turbulent profiles. These profiles clearly display the development of the turbulent core region and the shrinking of the laminar sublayer with increasing values of Re/sub theta/.

  15. Comparison of disease prevalence in two populations in the presence of misclassification.

    PubMed

    Tang, Man-Lai; Qiu, Shi-Fang; Poon, Wai-Yin

    2012-11-01

    Comparing disease prevalence in two groups is an important topic in medical research, and prevalence rates are obtained by classifying subjects according to whether they have the disease. Both high-cost infallible gold-standard classifiers or low-cost fallible classifiers can be used to classify subjects. However, statistical analysis that is based on data sets with misclassifications leads to biased results. As a compromise between the two classification approaches, partially validated sets are often used in which all individuals are classified by fallible classifiers, and some of the individuals are validated by the accurate gold-standard classifiers. In this article, we develop several reliable test procedures and approximate sample size formulas for disease prevalence studies based on the difference between two disease prevalence rates with two independent partially validated series. Empirical studies show that (i) the Score test produces close-to-nominal level and is preferred in practice; and (ii) the sample size formula based on the Score test is also fairly accurate in terms of the empirical power and type I error rate, and is hence recommended. A real example from an aplastic anemia study is used to illustrate the proposed methodologies. © 2012 WILEY-VCH Verlag GmbH & Co. KGaA, Weinheim.

  16. Crowd-sourcing relative preferences for ecosystem services in ...

    EPA Pesticide Factsheets

    Analysis of ecosystem service tradeoffs among project scenarios is more reliable when valuation data are available. Empirical valuation data are expensive and difficult to collect. As a possible alternative or supplement to empirical data, we downloaded and classified images from social media sites (SMS) Panoramio (n= 639), Instagram (n=2086), and Flickr (n=6644) for the AOC and 100 m buffer. We classified each image from the perspective of the beneficiary (photographer) according to US EPA’s Final Ecosystem Goods and Services (FEGS) classification system. After removing images not in the AOC or with bad links, 58 (10%), 361 (24%), and 1035 (16%) of the images from Panoramio, Instagram, and Flickr, respectively, depicted an ecosystem service. The most frequently occurring non ecosystem services depicted were bridges, ships, indoor scenes, and people. Across SMS with repeat images removed, AOC services were percentage ranked as follows: recreational scene viewers, 28%; boaters, 24%, flora and fauna viewers, 20%; other recreation experiencers (e.g., dog walkers, beach goers, bikers, trail and greenspace users), 15%; angling, 7%; learners, 1%, inspirational/sacred experiencers, 1%. Across SMS, recreational services were ranked as follows: scene viewing, 28%; boating, 24%, birding, 16%; fauna and flora viewing, 8%; trail and greenspace use, 7%; angling 7%; dog walking, 3%; swimming, beach and ice use, 3%; biking, 1%; excursion rail, 1%. Potential biases include S

  17. An analysis code for the Rapid Engineering Estimation of Momentum and Energy Losses (REMEL)

    NASA Technical Reports Server (NTRS)

    Dechant, Lawrence J.

    1994-01-01

    Nonideal behavior has traditionally been modeled by defining efficiency (a comparison between actual and isentropic processes), and subsequent specification by empirical or heuristic methods. With the increasing complexity of aeropropulsion system designs, the reliability of these more traditional methods is uncertain. Computational fluid dynamics (CFD) and experimental methods can provide this information but are expensive in terms of human resources, cost, and time. This report discusses an alternative to empirical and CFD methods by applying classical analytical techniques and a simplified flow model to provide rapid engineering estimates of these losses based on steady, quasi-one-dimensional governing equations including viscous and heat transfer terms (estimated by Reynold's analogy). A preliminary verification of REMEL has been compared with full Navier-Stokes (FNS) and CFD boundary layer computations for several high-speed inlet and forebody designs. Current methods compare quite well with more complex method results and solutions compare very well with simple degenerate and asymptotic results such as Fanno flow, isentropic variable area flow, and a newly developed, combined variable area duct with friction flow solution. These solution comparisons may offer an alternative to transitional and CFD-intense methods for the rapid estimation of viscous and heat transfer losses in aeropropulsion systems.

  18. Learning temporal rules to forecast instability in continuously monitored patients

    PubMed Central

    Dubrawski, Artur; Wang, Donghan; Hravnak, Marilyn; Clermont, Gilles; Pinsky, Michael R

    2017-01-01

    Inductive machine learning, and in particular extraction of association rules from data, has been successfully used in multiple application domains, such as market basket analysis, disease prognosis, fraud detection, and protein sequencing. The appeal of rule extraction techniques stems from their ability to handle intricate problems yet produce models based on rules that can be comprehended by humans, and are therefore more transparent. Human comprehension is a factor that may improve adoption and use of data-driven decision support systems clinically via face validity. In this work, we explore whether we can reliably and informatively forecast cardiorespiratory instability (CRI) in step-down unit (SDU) patients utilizing data from continuous monitoring of physiologic vital sign (VS) measurements. We use a temporal association rule extraction technique in conjunction with a rule fusion protocol to learn how to forecast CRI in continuously monitored patients. We detail our approach and present and discuss encouraging empirical results obtained using continuous multivariate VS data from the bedside monitors of 297 SDU patients spanning 29 346 hours (3.35 patient-years) of observation. We present example rules that have been learned from data to illustrate potential benefits of comprehensibility of the extracted models, and we analyze the empirical utility of each VS as a potential leading indicator of an impending CRI event. PMID:27274020

  19. Accuracy and reliability of forensic latent fingerprint decisions

    PubMed Central

    Ulery, Bradford T.; Hicklin, R. Austin; Buscaglia, JoAnn; Roberts, Maria Antonia

    2011-01-01

    The interpretation of forensic fingerprint evidence relies on the expertise of latent print examiners. The National Research Council of the National Academies and the legal and forensic sciences communities have called for research to measure the accuracy and reliability of latent print examiners’ decisions, a challenging and complex problem in need of systematic analysis. Our research is focused on the development of empirical approaches to studying this problem. Here, we report on the first large-scale study of the accuracy and reliability of latent print examiners’ decisions, in which 169 latent print examiners each compared approximately 100 pairs of latent and exemplar fingerprints from a pool of 744 pairs. The fingerprints were selected to include a range of attributes and quality encountered in forensic casework, and to be comparable to searches of an automated fingerprint identification system containing more than 58 million subjects. This study evaluated examiners on key decision points in the fingerprint examination process; procedures used operationally include additional safeguards designed to minimize errors. Five examiners made false positive errors for an overall false positive rate of 0.1%. Eighty-five percent of examiners made at least one false negative error for an overall false negative rate of 7.5%. Independent examination of the same comparisons by different participants (analogous to blind verification) was found to detect all false positive errors and the majority of false negative errors in this study. Examiners frequently differed on whether fingerprints were suitable for reaching a conclusion. PMID:21518906

  20. Psychometric properties of a new measure of fathers' involvement in the management of pediatric chronic diseases.

    PubMed

    Wysocki, Tim; Gavin, Leslie

    2004-01-01

    Despite theoretical and empirical reasons suggesting its potential importance, paternal involvement in the management of pediatric chronic diseases has rarely been a primary focus of research on family adaptation to pediatric disease. This may be due to a lack of appropriate tools to measure relevant behaviors. This study assessed the reliability and validity of the Dads' Active Disease Support scale (DADS), which was designed to measure male and female caregivers' estimates of the amount and helpfulness of paternal involvement in managing six pediatric chronic diseases. A sample of 224 heterosexual couples completed the DADS and measures of mother, child, and family functioning, yielding 190 who completed data sets. Of these, 91 mothers and 88 fathers completed the DADS again after a 1-month interval. Results confirmed DADS internal consistency, test-retest reliability, and interparent agreement. Confirmatory factor analysis supported its construct validity: A two-factor model (amount and helpfulness of fathers' involvement) best accounted for participants' responses. Significant correlations with scores on the Family Assessment Device supported DADS convergent validity. DADS scores of mothers and fathers suggest substantial room for improvement in both the amount and the helpfulness of paternal involvement in disease management. Mothers provided significantly higher ratings of the helpfulness of fathers' involvement than did fathers. The DADS appears to be a reliable and valid measure for studies of the associations between paternal involvement in disease management and child, maternal, and family adaptation to pediatric chronic medical conditions.

  1. Accuracy and reliability of forensic latent fingerprint decisions.

    PubMed

    Ulery, Bradford T; Hicklin, R Austin; Buscaglia, Joann; Roberts, Maria Antonia

    2011-05-10

    The interpretation of forensic fingerprint evidence relies on the expertise of latent print examiners. The National Research Council of the National Academies and the legal and forensic sciences communities have called for research to measure the accuracy and reliability of latent print examiners' decisions, a challenging and complex problem in need of systematic analysis. Our research is focused on the development of empirical approaches to studying this problem. Here, we report on the first large-scale study of the accuracy and reliability of latent print examiners' decisions, in which 169 latent print examiners each compared approximately 100 pairs of latent and exemplar fingerprints from a pool of 744 pairs. The fingerprints were selected to include a range of attributes and quality encountered in forensic casework, and to be comparable to searches of an automated fingerprint identification system containing more than 58 million subjects. This study evaluated examiners on key decision points in the fingerprint examination process; procedures used operationally include additional safeguards designed to minimize errors. Five examiners made false positive errors for an overall false positive rate of 0.1%. Eighty-five percent of examiners made at least one false negative error for an overall false negative rate of 7.5%. Independent examination of the same comparisons by different participants (analogous to blind verification) was found to detect all false positive errors and the majority of false negative errors in this study. Examiners frequently differed on whether fingerprints were suitable for reaching a conclusion.

  2. Empirical Assessment of the Mean Block Volume of Rock Masses Intersected by Four Joint Sets

    NASA Astrophysics Data System (ADS)

    Morelli, Gian Luca

    2016-05-01

    The estimation of a representative value for the rock block volume ( V b) is of huge interest in rock engineering in regards to rock mass characterization purposes. However, while mathematical relationships to precisely estimate this parameter from the spacing of joints can be found in literature for rock masses intersected by three dominant joint sets, corresponding relationships do not actually exist when more than three sets occur. In these cases, a consistent assessment of V b can only be achieved by directly measuring the dimensions of several representative natural rock blocks in the field or by means of more sophisticated 3D numerical modeling approaches. However, Palmström's empirical relationship based on the volumetric joint count J v and on a block shape factor β is commonly used in the practice, although strictly valid only for rock masses intersected by three joint sets. Starting from these considerations, the present paper is primarily intended to investigate the reliability of a set of empirical relationships linking the block volume with the indexes most commonly used to characterize the degree of jointing in a rock mass (i.e. the J v and the mean value of the joint set spacings) specifically applicable to rock masses intersected by four sets of persistent discontinuities. Based on the analysis of artificial 3D block assemblies generated using the software AutoCAD, the most accurate best-fit regression has been found between the mean block volume (V_{{{{b}}_{{m}} }}) of tested rock mass samples and the geometric mean value of the spacings of the joint sets delimiting blocks; thus, indicating this mean value as a promising parameter for the preliminary characterization of the block size. Tests on field outcrops have demonstrated that the proposed empirical methodology has the potential of predicting the mean block volume of multiple-set jointed rock masses with an acceptable accuracy for common uses in most practical rock engineering applications.

  3. Non-suicidal self-injury and life stress: A systematic meta-analysis and theoretical elaboration

    PubMed Central

    Liu, Richard T.; Cheek, Shayna M.; Nestor, Bridget A.

    2016-01-01

    Recent years have seen a considerable growth of interest in the study of life stress and non-suicidal self-injury (NSSI). The current article presents a systematic review of the empirical literature on this association. In addition to providing a comprehensive meta-analysis, the current article includes a qualitative review of the findings for which there were too few cases (i.e., < 3) for reliable approximations of effect sizes. Across the studies included in the meta-analysis, a significant but modest relation between life stress and NSSI was found (pooled OR = 1.81 [95% CI = 1.49–2.21]). After an adjustment was made for publication bias, the estimated effect size was smaller but still significant (pooled OR = 1.33 [95% CI = 1.08–1.63]). This relation was moderated by sample type, NSSI measure type, and length of period covered by the NSSI measure. The empirical literature is characterized by several methodological limitations, particularly the frequent use of cross-sectional analyses involving temporal overlap between assessments of life stress and NSSI, leaving unclear the precise nature of the relation between these two phenomena (e.g., whether life stress may be a cause, concomitant, or consequence of NSSI). Theoretically informed research utilizing multi-wave designs, assessing life stress and NSSI over relatively brief intervals, and featuring interview-based assessments of these constructs holds promise for advancing our understanding of their relation. The current review concludes with a theoretical elaboration of the association between NSSI and life stress, with the aim of providing a conceptual framework to guide future study in this area. PMID:27267345

  4. Scaling and efficiency determine the irreversible evolution of a market

    PubMed Central

    Baldovin, F.; Stella, A. L.

    2007-01-01

    In setting up a stochastic description of the time evolution of a financial index, the challenge consists in devising a model compatible with all stylized facts emerging from the analysis of financial time series and providing a reliable basis for simulating such series. Based on constraints imposed by market efficiency and on an inhomogeneous-time generalization of standard simple scaling, we propose an analytical model which accounts simultaneously for empirical results like the linear decorrelation of successive returns, the power law dependence on time of the volatility autocorrelation function, and the multiscaling associated to this dependence. In addition, our approach gives a justification and a quantitative assessment of the irreversible character of the index dynamics. This irreversibility enters as a key ingredient in a novel simulation strategy of index evolution which demonstrates the predictive potential of the model.

  5. Typology of Couples Entering Alcohol Behavioral Couple Therapy: An Empirical Approach and Test of Predictive Validity on Treatment Response.

    PubMed

    Ladd, Benjamin O; McCrady, Barbara S

    2016-01-01

    This study aimed to examine whether classification of couples in which one partner has an alcohol problem is similar to that reported in the general couples literature. Typologies of couples seeking alcohol behavioral couple therapy (ABCT) were developed via hierarchical cluster analysis using behavioral codes of couple interactions during their first ABCT session. Four couples types based on in-session behavior were established reliably, labeled avoider, validator, hostile, and ambivalent-detached. These couple types resembled couples types found in previous research. Couple type was associated with baseline relationship satisfaction, but not alcohol use. Results suggest heterogeneity in couples with alcohol problems presenting to treatment; further study is needed to investigate the function of alcohol within these different types. © 2015 American Association for Marriage and Family Therapy.

  6. A non-Gaussian approach to risk measures

    NASA Astrophysics Data System (ADS)

    Bormetti, Giacomo; Cisana, Enrica; Montagna, Guido; Nicrosini, Oreste

    2007-03-01

    Reliable calculations of financial risk require that the fat-tailed nature of prices changes is included in risk measures. To this end, a non-Gaussian approach to financial risk management is presented, modelling the power-law tails of the returns distribution in terms of a Student- t distribution. Non-Gaussian closed-form solutions for value-at-risk and expected shortfall are obtained and standard formulae known in the literature under the normality assumption are recovered as a special case. The implications of the approach for risk management are demonstrated through an empirical analysis of financial time series from the Italian stock market and in comparison with the results of the most widely used procedures of quantitative finance. Particular attention is paid to quantify the size of the errors affecting the market risk measures obtained according to different methodologies, by employing a bootstrap technique.

  7. Testing jumps via false discovery rate control.

    PubMed

    Yen, Yu-Min

    2013-01-01

    Many recently developed nonparametric jump tests can be viewed as multiple hypothesis testing problems. For such multiple hypothesis tests, it is well known that controlling type I error often makes a large proportion of erroneous rejections, and such situation becomes even worse when the jump occurrence is a rare event. To obtain more reliable results, we aim to control the false discovery rate (FDR), an efficient compound error measure for erroneous rejections in multiple testing problems. We perform the test via the Barndorff-Nielsen and Shephard (BNS) test statistic, and control the FDR with the Benjamini and Hochberg (BH) procedure. We provide asymptotic results for the FDR control. From simulations, we examine relevant theoretical results and demonstrate the advantages of controlling the FDR. The hybrid approach is then applied to empirical analysis on two benchmark stock indices with high frequency data.

  8. A harmonic linear dynamical system for prominent ECG feature extraction.

    PubMed

    Thi, Ngoc Anh Nguyen; Yang, Hyung-Jeong; Kim, SunHee; Do, Luu Ngoc

    2014-01-01

    Unsupervised mining of electrocardiography (ECG) time series is a crucial task in biomedical applications. To have efficiency of the clustering results, the prominent features extracted from preprocessing analysis on multiple ECG time series need to be investigated. In this paper, a Harmonic Linear Dynamical System is applied to discover vital prominent features via mining the evolving hidden dynamics and correlations in ECG time series. The discovery of the comprehensible and interpretable features of the proposed feature extraction methodology effectively represents the accuracy and the reliability of clustering results. Particularly, the empirical evaluation results of the proposed method demonstrate the improved performance of clustering compared to the previous main stream feature extraction approaches for ECG time series clustering tasks. Furthermore, the experimental results on real-world datasets show scalability with linear computation time to the duration of the time series.

  9. Determination of Turboprop Reduction Gearbox System Fatigue Life and Reliability

    NASA Technical Reports Server (NTRS)

    Zaretsky, Erwin V.; Lewicki, David G.; Savage, Michael; Vlcek, Brian L.

    2007-01-01

    Two computational models to determine the fatigue life and reliability of a commercial turboprop gearbox are compared with each other and with field data. These models are (1) Monte Carlo simulation of randomly selected lives of individual bearings and gears comprising the system and (2) two-parameter Weibull distribution function for bearings and gears comprising the system using strict-series system reliability to combine the calculated individual component lives in the gearbox. The Monte Carlo simulation included the virtual testing of 744,450 gearboxes. Two sets of field data were obtained from 64 gearboxes that were first-run to removal for cause, were refurbished and placed back in service, and then were second-run until removal for cause. A series of equations were empirically developed from the Monte Carlo simulation to determine the statistical variation in predicted life and Weibull slope as a function of the number of gearboxes failed. The resultant L(sub 10) life from the field data was 5,627 hr. From strict-series system reliability, the predicted L(sub 10) life was 774 hr. From the Monte Carlo simulation, the median value for the L(sub 10) gearbox lives equaled 757 hr. Half of the gearbox L(sub 10) lives will be less than this value and the other half more. The resultant L(sub 10) life of the second-run (refurbished) gearboxes was 1,334 hr. The apparent load-life exponent p for the roller bearings is 5.2. Were the bearing lives to be recalculated with a load-life exponent p equal to 5.2, the predicted L(sub 10) life of the gearbox would be equal to the actual life obtained in the field. The component failure distribution of the gearbox from the Monte Carlo simulation was nearly identical to that using the strict-series system reliability analysis, proving the compatibility of these methods.

  10. Advancing the detection of steady-state visual evoked potentials in brain-computer interfaces

    NASA Astrophysics Data System (ADS)

    Abu-Alqumsan, Mohammad; Peer, Angelika

    2016-06-01

    Objective. Spatial filtering has proved to be a powerful pre-processing step in detection of steady-state visual evoked potentials and boosted typical detection rates both in offline analysis and online SSVEP-based brain-computer interface applications. State-of-the-art detection methods and the spatial filters used thereby share many common foundations as they all build upon the second order statistics of the acquired Electroencephalographic (EEG) data, that is, its spatial autocovariance and cross-covariance with what is assumed to be a pure SSVEP response. The present study aims at highlighting the similarities and differences between these methods. Approach. We consider the canonical correlation analysis (CCA) method as a basis for the theoretical and empirical (with real EEG data) analysis of the state-of-the-art detection methods and the spatial filters used thereby. We build upon the findings of this analysis and prior research and propose a new detection method (CVARS) that combines the power of the canonical variates and that of the autoregressive spectral analysis in estimating the signal and noise power levels. Main results. We found that the multivariate synchronization index method and the maximum contrast combination method are variations of the CCA method. All three methods were found to provide relatively unreliable detections in low signal-to-noise ratio (SNR) regimes. CVARS and the minimum energy combination methods were found to provide better estimates for different SNR levels. Significance. Our theoretical and empirical results demonstrate that the proposed CVARS method outperforms other state-of-the-art detection methods when used in an unsupervised fashion. Furthermore, when used in a supervised fashion, a linear classifier learned from a short training session is able to estimate the hidden user intention, including the idle state (when the user is not attending to any stimulus), rapidly, accurately and reliably.

  11. Simple, empirical approach to predict neutron capture cross sections from nuclear masses

    NASA Astrophysics Data System (ADS)

    Couture, A.; Casten, R. F.; Cakirli, R. B.

    2017-12-01

    Background: Neutron capture cross sections are essential to understanding the astrophysical s and r processes, the modeling of nuclear reactor design and performance, and for a wide variety of nuclear forensics applications. Often, cross sections are needed for nuclei where experimental measurements are difficult. Enormous effort, over many decades, has gone into attempting to develop sophisticated statistical reaction models to predict these cross sections. Such work has met with some success but is often unable to reproduce measured cross sections to better than 40 % , and has limited predictive power, with predictions from different models rapidly differing by an order of magnitude a few nucleons from the last measurement. Purpose: To develop a new approach to predicting neutron capture cross sections over broad ranges of nuclei that accounts for their values where known and which has reliable predictive power with small uncertainties for many nuclei where they are unknown. Methods: Experimental neutron capture cross sections were compared to empirical mass observables in regions of similar structure. Results: We present an extremely simple method, based solely on empirical mass observables, that correlates neutron capture cross sections in the critical energy range from a few keV to a couple hundred keV. We show that regional cross sections are compactly correlated in medium and heavy mass nuclei with the two-neutron separation energy. These correlations are easily amenable to predict unknown cross sections, often converting the usual extrapolations to more reliable interpolations. It almost always reproduces existing data to within 25 % and estimated uncertainties are below about 40 % up to 10 nucleons beyond known data. Conclusions: Neutron capture cross sections display a surprisingly strong connection to the two-neutron separation energy, a nuclear structure property. The simple, empirical correlations uncovered provide model-independent predictions of neutron capture cross sections, extending far from stability, including for nuclei of the highest sensitivity to r -process nucleosynthesis.

  12. Cross-Cultural Adaptation and Validation of the Commitment-to-School Scale Using a Sample of Junior High School Youth in Ghana

    ERIC Educational Resources Information Center

    Ansong, David; Chowa, Gina A.; Masa, Rainier D.

    2016-01-01

    Empirical evidence from developed countries suggests that students' commitment to school is fundamental to their academic success. However, in developing countries, validated measures of student commitment to school do not exist. The current study helps fill this research gap by examining the validity and reliability of a commitment-to-school…

  13. Measuring Plagiarism: Researching What Students Do, Not What They Say They Do

    ERIC Educational Resources Information Center

    Walker, John

    2010-01-01

    Student plagiarism in colleges and universities has become a controversial issue in recent years. A key problem has been the lack of reliable empirical data on the frequency, nature and extent of plagiarism in student assignments. The aim of the study described here was to provide this data. Patterns of plagiarism were tracked in two university…

  14. Further Empirical Data on the Psychoeducational Profile-Revised (PEP-R): Reliability and Validation with the Vineland Adaptive Behavior Scales

    ERIC Educational Resources Information Center

    Villa, Susanna; Micheli, Enrico; Villa, Laura; Pastore, Valentina; Crippa, Alessandro; Molteni, Massimo

    2010-01-01

    The PEP-R (psychoeducational profile revised) is an instrument that has been used in many countries to assess abilities and formulate treatment programs for children with autism and related developmental disorders. To the end to provide further information on the PEP-R's psychometric properties, a large sample (N = 137) of children presenting…

  15. Intrinsic Motivation and Engagement as "Active Ingredients" in Garden-Based Education: Examining Models and Measures Derived from Self-Determination Theory

    ERIC Educational Resources Information Center

    Skinner, Ellen A.; Chi, Una

    2012-01-01

    Building on self-determination theory, this study presents a model of intrinsic motivation and engagement as "active ingredients" in garden-based education. The model was used to create reliable and valid measures of key constructs, and to guide the empirical exploration of motivational processes in garden-based learning. Teacher- and…

  16. The Peer Attitudes toward Children Who Stutter Scale: Reliability, Known Groups Validity, and Negativity of Elementary School-Age Children's Attitudes

    ERIC Educational Resources Information Center

    Langevin, Marilyn

    2009-01-01

    Psychometric properties of the Peer Attitudes Toward Children who Stutter (PATCS) scale (Langevin, M., & Hagler, P. (2004). Development of a scale to measure peer attitudes toward children who stutter. In A.K. Bothe (Ed.), Evidence-based treatment of stuttering: empirical bases and clinical applications (pp. 139-171). Mahwah, NJ: Lawrence…

  17. Factorial Validity and Reliability of the Sense of Community in Online Courses Scale

    ERIC Educational Resources Information Center

    Randolph, Justus J.; Crawford, Linda M.

    2013-01-01

    The alarmingly high rate of attrition in online courses results in many negative consequences for students, faculty, online institutions, and for society as a whole. One reason theorized for this attrition is a lack of a sense of community in online courses; however, there is much theoretical and empirical debate on what factors contribute to that…

  18. Collegiality in Business Schools: Development of a Collegiality Measure and Evaluations of its Implications

    ERIC Educational Resources Information Center

    Miles, Morgan P.; Shepherd, C. David; Rose, Jacob M.; Dibben, Mark

    2015-01-01

    Purpose: While collegiality is often discussed and touted as a critical aspect of academia, there is little research that empirically examines collegiality in university business schools. One cause of the paucity of research is the lack of a reliable scale to measure collegiality (Sabharwal, 2011). The purpose of this paper is to develop a scale…

  19. Using small area estimation and Lidar-derived variables for multivariate prediction of forest attributes

    Treesearch

    F. Mauro; Vicente Monleon; H. Temesgen

    2015-01-01

    Small area estimation (SAE) techniques have been successfully applied in forest inventories to provide reliable estimates for domains where the sample size is small (i.e. small areas). Previous studies have explored the use of either Area Level or Unit Level Empirical Best Linear Unbiased Predictors (EBLUPs) in a univariate framework, modeling each variable of interest...

  20. Reliability and Validity Evidence for Achievement Goal Models in High School Physical Education Settings

    ERIC Educational Resources Information Center

    Guan, Jianmin; McBride, Ron; Xiang, Ping

    2007-01-01

    Although empirical research in academic areas provides support for both a 3-factor as well as a 4-factor achievement goal model, both models were proposed and tested with a collegiate sample. Little is known about the generalizability of either model with high school level samples. This study was designed to examine whether the 3-factor model…

  1. Scrutinizing A Survey-Based Measure of Science and Mathematics Teacher Knowledge: Relationship to Observations of Teaching Practice

    ERIC Educational Resources Information Center

    Talbot, Robert M., III

    2017-01-01

    There is a clear need for valid and reliable instrumentation that measures teacher knowledge. However, the process of investigating and making a case for instrument validity is not a simple undertaking; rather, it is a complex endeavor. This paper presents the empirical case of one aspect of such an instrument validation effort. The particular…

  2. DOE Office of Scientific and Technical Information (OSTI.GOV)

    Simpson, L.; Britt, J.; Birkmire, R.

    ITN Energy Systems, Inc., and Global Solar Energy, Inc., assisted by NREL's PV Manufacturing R&D program, have continued to advance CIGS production technology by developing trajectory-oriented predictive/control models, fault-tolerance control, control platform development, in-situ sensors, and process improvements. Modeling activities included developing physics-based and empirical models for CIGS and sputter-deposition processing, implementing model-based control, and applying predictive models to the construction of new evaporation sources and for control. Model-based control is enabled by implementing reduced or empirical models into a control platform. Reliability improvement activities include implementing preventive maintenance schedules; detecting failed sensors/equipment and reconfiguring to tinue processing; and systematicmore » development of fault prevention and reconfiguration strategies for the full range of CIGS PV production deposition processes. In-situ sensor development activities have resulted in improved control and indicated the potential for enhanced process status monitoring and control of the deposition processes. Substantial process improvements have been made, including significant improvement in CIGS uniformity, thickness control, efficiency, yield, and throughput. In large measure, these gains have been driven by process optimization, which in turn have been enabled by control and reliability improvements due to this PV Manufacturing R&D program.« less

  3. Practical guidance on representing the heteroscedasticity of residual errors of hydrological predictions

    NASA Astrophysics Data System (ADS)

    McInerney, David; Thyer, Mark; Kavetski, Dmitri; Kuczera, George

    2016-04-01

    Appropriate representation of residual errors in hydrological modelling is essential for accurate and reliable probabilistic streamflow predictions. In particular, residual errors of hydrological predictions are often heteroscedastic, with large errors associated with high runoff events. Although multiple approaches exist for representing this heteroscedasticity, few if any studies have undertaken a comprehensive evaluation and comparison of these approaches. This study fills this research gap by evaluating a range of approaches for representing heteroscedasticity in residual errors. These approaches include the 'direct' weighted least squares approach and 'transformational' approaches, such as logarithmic, Box-Cox (with and without fitting the transformation parameter), logsinh and the inverse transformation. The study reports (1) theoretical comparison of heteroscedasticity approaches, (2) empirical evaluation of heteroscedasticity approaches using a range of multiple catchments / hydrological models / performance metrics and (3) interpretation of empirical results using theory to provide practical guidance on the selection of heteroscedasticity approaches. Importantly, for hydrological practitioners, the results will simplify the choice of approaches to represent heteroscedasticity. This will enhance their ability to provide hydrological probabilistic predictions with the best reliability and precision for different catchment types (e.g. high/low degree of ephemerality).

  4. A robust approach for ECG-based analysis of cardiopulmonary coupling.

    PubMed

    Zheng, Jiewen; Wang, Weidong; Zhang, Zhengbo; Wu, Dalei; Wu, Hao; Peng, Chung-Kang

    2016-07-01

    Deriving respiratory signal from a surface electrocardiogram (ECG) measurement has advantage of simultaneously monitoring of cardiac and respiratory activities. ECG-based cardiopulmonary coupling (CPC) analysis estimated by heart period variability and ECG-derived respiration (EDR) shows promising applications in medical field. The aim of this paper is to provide a quantitative analysis of the ECG-based CPC, and further improve its performance. Two conventional strategies were tested to obtain EDR signal: R-S wave amplitude and area of the QRS complex. An adaptive filter was utilized to extract the common component of inter-beat interval (RRI) and EDR, generating enhanced versions of EDR signal. CPC is assessed through probing the nonlinear phase interactions between RRI series and respiratory signal. Respiratory oscillations presented in both RRI series and respiratory signals were extracted by ensemble empirical mode decomposition for coupling analysis via phase synchronization index. The results demonstrated that CPC estimated from conventional EDR series exhibits constant and proportional biases, while that estimated from enhanced EDR series is more reliable. Adaptive filtering can improve the accuracy of the ECG-based CPC estimation significantly and achieve robust CPC analysis. The improved ECG-based CPC estimation may provide additional prognostic information for both sleep medicine and autonomic function analysis. Copyright © 2016 IPEM. Published by Elsevier Ltd. All rights reserved.

  5. Confirming, Validating, and Norming the Factor Structure of Systemic Therapy Inventory of Change Initial and Intersession.

    PubMed

    Pinsof, William M; Zinbarg, Richard E; Shimokawa, Kenichi; Latta, Tara A; Goldsmith, Jacob Z; Knobloch-Fedders, Lynne M; Chambers, Anthony L; Lebow, Jay L

    2015-09-01

    Progress or feedback research tracks and feeds back client progress data throughout the course of psychotherapy. In the effort to empirically ground psychotherapeutic practice, feedback research is both a complement and alternative to empirically supported manualized treatments. Evidence suggests that tracking and feeding back progress data with individual or nonsystemic feedback systems improves outcomes in individual and couple therapy. The research reported in this article pertains to the STIC(®) (Systemic Therapy Inventory of Change)-the first client-report feedback system designed to empirically assess and track change within client systems from multisystemic and multidimensional perspectives in individual, couple, and family therapy. Clients complete the STIC Initial before the first session and the shorter STIC Intersession before every subsequent session. This study tested and its results supported the hypothesized factor structure of the six scales that comprise both STIC forms in a clinical outpatient sample and in a normal, random representative sample of the U.S. This study also tested the STIC's concurrent validity and found that its 6 scales and 40 of its 41 subscales differentiated the clinical and normal samples. Lastly, the study derived clinical cut-offs for each scale and subscale to determine whether and how much a client's score falls in the normal or clinical range. Beyond supporting the factorial and concurrent validity of both STIC forms, this research supported the reliabilities of the six scales (Omegahierarchical ) as well as the reliabilities of most subscales (alpha and rate-rerate). This article delineates clinical implications and directions for future research. © 2015 Family Process Institute.

  6. Evaluation of the existing triple point path models with new experimental data: proposal of an original empirical formulation

    NASA Astrophysics Data System (ADS)

    Boutillier, J.; Ehrhardt, L.; De Mezzo, S.; Deck, C.; Magnan, P.; Naz, P.; Willinger, R.

    2018-03-01

    With the increasing use of improvised explosive devices (IEDs), the need for better mitigation, either for building integrity or for personal security, increases in importance. Before focusing on the interaction of the shock wave with a target and the potential associated damage, knowledge must be acquired regarding the nature of the blast threat, i.e., the pressure-time history. This requirement motivates gaining further insight into the triple point (TP) path, in order to know precisely which regime the target will encounter (simple reflection or Mach reflection). Within this context, the purpose of this study is to evaluate three existing TP path empirical models, which in turn are used in other empirical models for the determination of the pressure profile. These three TP models are the empirical function of Kinney, the Unified Facilities Criteria (UFC) curves, and the model of the Natural Resources Defense Council (NRDC). As discrepancies are observed between these models, new experimental data were obtained to test their reliability and a new promising formulation is proposed for scaled heights of burst ranging from 24.6-172.9 cm/kg^{1/3}.

  7. Assessing the value of transgenic crops.

    PubMed

    Lacey, Hugh

    2002-10-01

    In the current controversy about the value of transgenic crops, matters open to empirical inquiry are centrally at issue. One such matter is a key premise in a common argument (that I summarize) that transgenic crops should be considered to have universal value. The premise is that there are no alternative forms of agriculture available to enable the production of sufficient food to feed the world. The proponents of agroecology challenge it, claiming that agroecology provides an alternative, and they deny the claim that it is well founded on empirical evidence. It is, therefore, a matter of both social and scientific importance that this premise and the criticisms of it be investigated rigorously and empirically, so that the benefits and disadvantages of transgenic-intensive agriculture and agroecology can be compared in a reliable way. Conducting adequate investigation about the potential contribution of agroecology requires that the cultural conditions of its practice (and, thus, of the practices and movements of small-scale farmers in the "third world") be strengthened--and this puts the interests of investigation into tension with the socio-economic interests driving the development of transgenics. General issues about relationship between ethical argument and empirical (scientific) investigation are raised throughout the article.

  8. Validation of the Internet Gaming Disorder Scale - Short-Form (IGDS9-SF) in an Italian-speaking sample.

    PubMed

    Monacis, Lucia; Palo, Valeria de; Griffiths, Mark D; Sinatra, Maria

    2016-12-01

    Background and aims The inclusion of Internet Gaming Disorder (IGD) in Section III of the fifth edition of the Diagnostic and Statistical Manual of Mental Disorders has increased the interest of researchers in the development of new standardized psychometric tools for the assessment of such a disorder. To date, the nine-item Internet Gaming Disorder Scale - Short-Form (IGDS9-SF) has only been validated in English, Portuguese, and Slovenian languages. Therefore, the aim of this investigation was to examine the psychometric properties of the IGDS9-SF in an Italian-speaking sample. Methods A total of 757 participants were recruited to the present study. Confirmatory factor analysis and multi-group analyses were applied to assess the construct validity. Reliability analyses comprised the average variance extracted, the standard error of measurement, and the factor determinacy coefficient. Convergent and criterion validities were established through the associations with other related constructs. The receiver operating characteristic curve analysis was used to determine an empirical cut-off point. Results Findings confirmed the single-factor structure of the instrument, its measurement invariance at the configural level, and the convergent and criterion validities. Satisfactory levels of reliability and a cut-off point of 21 were obtained. Discussion and conclusions The present study provides validity evidence for the use of the Italian version of the IGDS9-SF and may foster research into gaming addiction in the Italian context.

  9. Detecting individual memories through the neural decoding of memory states and past experience.

    PubMed

    Rissman, Jesse; Greely, Henry T; Wagner, Anthony D

    2010-05-25

    A wealth of neuroscientific evidence indicates that our brains respond differently to previously encountered than to novel stimuli. There has been an upswell of interest in the prospect that functional MRI (fMRI), when coupled with multivariate data analysis techniques, might allow the presence or absence of individual memories to be detected from brain activity patterns. This could have profound implications for forensic investigations and legal proceedings, and thus the merits and limitations of such an approach are in critical need of empirical evaluation. We conducted two experiments to investigate whether neural signatures of recognition memory can be reliably decoded from fMRI data. In Exp. 1, participants were scanned while making explicit recognition judgments for studied and novel faces. Multivoxel pattern analysis (MVPA) revealed a robust ability to classify whether a given face was subjectively experienced as old or new, as well as whether recognition was accompanied by recollection, strong familiarity, or weak familiarity. Moreover, a participant's subjective mnemonic experiences could be reliably decoded even when the classifier was trained on the brain data from other individuals. In contrast, the ability to classify a face's objective old/new status, when holding subjective status constant, was severely limited. This important boundary condition was further evidenced in Exp. 2, which demonstrated that mnemonic decoding is poor when memory is indirectly (implicitly) probed. Thus, although subjective memory states can be decoded quite accurately under controlled experimental conditions, fMRI has uncertain utility for objectively detecting an individual's past experiences.

  10. Montreal Battery of Evaluation of Amusia: Validity evidence and norms for adolescents in Belo Horizonte, Minas Gerais, Brazil

    PubMed Central

    Nunes-Silva, Marília; Haase, Vitor Geraldi

    2012-01-01

    The Montreal Battery of Evaluation of Amusia (MBEA) is a battery of tests that assesses six music processing components: scale, contour, interval, rhythm, metric, and music memory. The present study sought to verify the psychometric characteristics of the MBEA in a sample of 150 adolescents aged 14-18 years in the city of Belo Horizonte, Minas Gerais, Brazil, and to develop specific norms for this population. We used statistical procedures that explored the dimensional structure of the MBEA and its items, evaluating their adequacy from empirical data, verifying their reliability, and providing evidence of validity. The results for the difficult levels for each test indicated a trend toward higher scores, corroborating previous studies. From the analysis of the criterion groups, almost all of the items were considered discriminatory. The global score of the MBEA was shown to be valid and reliable (rK-R20=0.896) for assessing the musical ability of normal teenagers. Based on the analysis of the items, we proposed a short version of the MBEA. Further studies with larger samples and amusic individuals are necessary to provide evidence of the validity of the MBEA in the Brazilian milieu. The present study brings to the Brazilian context a tool for diagnosing deficits in musical skills and will serve as a basis for comparisons with single case studies and studies of populations with specific neuropsychological syndromes. PMID:29213804

  11. Validation of the Internet Gaming Disorder Scale – Short-Form (IGDS9-SF) in an Italian-speaking sample

    PubMed Central

    Monacis, Lucia; de Palo, Valeria; Griffiths, Mark D.; Sinatra, Maria

    2016-01-01

    Background and aims The inclusion of Internet Gaming Disorder (IGD) in Section III of the fifth edition of the Diagnostic and Statistical Manual of Mental Disorders has increased the interest of researchers in the development of new standardized psychometric tools for the assessment of such a disorder. To date, the nine-item Internet Gaming Disorder Scale – Short-Form (IGDS9-SF) has only been validated in English, Portuguese, and Slovenian languages. Therefore, the aim of this investigation was to examine the psychometric properties of the IGDS9-SF in an Italian-speaking sample. Methods A total of 757 participants were recruited to the present study. Confirmatory factor analysis and multi-group analyses were applied to assess the construct validity. Reliability analyses comprised the average variance extracted, the standard error of measurement, and the factor determinacy coefficient. Convergent and criterion validities were established through the associations with other related constructs. The receiver operating characteristic curve analysis was used to determine an empirical cut-off point. Results Findings confirmed the single-factor structure of the instrument, its measurement invariance at the configural level, and the convergent and criterion validities. Satisfactory levels of reliability and a cut-off point of 21 were obtained. Discussion and conclusions The present study provides validity evidence for the use of the Italian version of the IGDS9-SF and may foster research into gaming addiction in the Italian context. PMID:27876422

  12. Development and Pilot Test of the Workplace Readiness Questionnaire, a Theory-Based Instrument to Measure Small Workplaces' Readiness to Implement Wellness Programs.

    PubMed

    Hannon, Peggy A; Helfrich, Christian D; Chan, K Gary; Allen, Claire L; Hammerback, Kristen; Kohn, Marlana J; Parrish, Amanda T; Weiner, Bryan J; Harris, Jeffrey R

    2017-01-01

    To develop a theory-based questionnaire to assess readiness for change in small workplaces adopting wellness programs. In developing our scale, we first tested items via "think-aloud" interviews. We tested the revised items in a cross-sectional quantitative telephone survey. The study setting comprised small workplaces (20-250 employees) in low-wage industries. Decision-makers representing small workplaces in King County, Washington (think-aloud interviews, n = 9), and the United States (telephone survey, n = 201) served as study subjects. We generated items for each construct in Weiner's theory of organizational readiness for change. We also measured workplace characteristics and current implementation of workplace wellness programs. We assessed reliability by coefficient alpha for each of the readiness questionnaire subscales. We tested the association of all subscales with employers' current implementation of wellness policies, programs, and communications, and conducted a path analysis to test the associations in the theory of organizational readiness to change. Each of the readiness subscales exhibited acceptable internal reliability (coefficient alpha range, .75-.88) and was positively associated with wellness program implementation ( p < .05). The path analysis was consistent with the theory of organizational readiness to change, except change efficacy did not predict change-related effort. We developed a new questionnaire to assess small workplaces' readiness to adopt and implement evidence-based wellness programs. Our findings also provide empirical validation of Weiner's theory of readiness for change.

  13. Adequacy of laser diffraction for soil particle size analysis

    PubMed Central

    Fisher, Peter; Aumann, Colin; Chia, Kohleth; O'Halloran, Nick; Chandra, Subhash

    2017-01-01

    Sedimentation has been a standard methodology for particle size analysis since the early 1900s. In recent years laser diffraction is beginning to replace sedimentation as the prefered technique in some industries, such as marine sediment analysis. However, for the particle size analysis of soils, which have a diverse range of both particle size and shape, laser diffraction still requires evaluation of its reliability. In this study, the sedimentation based sieve plummet balance method and the laser diffraction method were used to measure the particle size distribution of 22 soil samples representing four contrasting Australian Soil Orders. Initially, a precise wet riffling methodology was developed capable of obtaining representative samples within the recommended obscuration range for laser diffraction. It was found that repeatable results were obtained even if measurements were made at the extreme ends of the manufacturer’s recommended obscuration range. Results from statistical analysis suggested that the use of sample pretreatment to remove soil organic carbon (and possible traces of calcium-carbonate content) made minor differences to the laser diffraction particle size distributions compared to no pretreatment. These differences were found to be marginally statistically significant in the Podosol topsoil and Vertosol subsoil. There are well known reasons why sedimentation methods may be considered to ‘overestimate’ plate-like clay particles, while laser diffraction will ‘underestimate’ the proportion of clay particles. In this study we used Lin’s concordance correlation coefficient to determine the equivalence of laser diffraction and sieve plummet balance results. The results suggested that the laser diffraction equivalent thresholds corresponding to the sieve plummet balance cumulative particle sizes of < 2 μm, < 20 μm, and < 200 μm, were < 9 μm, < 26 μm, < 275 μm respectively. The many advantages of laser diffraction for soil particle size analysis, and the empirical results of this study, suggest that deployment of laser diffraction as a standard test procedure can provide reliable results, provided consistent sample preparation is used. PMID:28472043

  14. A Lean Six Sigma approach to the improvement of the selenium analysis method.

    PubMed

    Cloete, Bronwyn C; Bester, André

    2012-11-02

    Reliable results represent the pinnacle assessment of quality of an analytical laboratory, and therefore variability is considered to be a critical quality problem associated with the selenium analysis method executed at Western Cape Provincial Veterinary Laboratory (WCPVL). The elimination and control of variability is undoubtedly of significant importance because of the narrow margin of safety between toxic and deficient doses of the trace element for good animal health. A quality methodology known as Lean Six Sigma was believed to present the most feasible solution for overcoming the adverse effect of variation, through steps towards analytical process improvement. Lean Six Sigma represents a form of scientific method type, which is empirical, inductive and deductive, and systematic, which relies on data, and is fact-based. The Lean Six Sigma methodology comprises five macro-phases, namely Define, Measure, Analyse, Improve and Control (DMAIC). Both qualitative and quantitative laboratory data were collected in terms of these phases. Qualitative data were collected by using quality-tools, namely an Ishikawa diagram, a Pareto chart, Kaizen analysis and a Failure Mode Effect analysis tool. Quantitative laboratory data, based on the analytical chemistry test method, were collected through a controlled experiment. The controlled experiment entailed 13 replicated runs of the selenium test method, whereby 11 samples were repetitively analysed, whilst Certified Reference Material (CRM) was also included in 6 of the runs. Laboratory results obtained from the controlled experiment was analysed by using statistical methods, commonly associated with quality validation of chemistry procedures. Analysis of both sets of data yielded an improved selenium analysis method, believed to provide greater reliability of results, in addition to a greatly reduced cycle time and superior control features. Lean Six Sigma may therefore be regarded as a valuable tool in any laboratory, and represents both a management discipline, and a standardised approach to problem solving and process optimisation.

  15. GTest: a software tool for graphical assessment of empirical distributions' Gaussianity.

    PubMed

    Barca, E; Bruno, E; Bruno, D E; Passarella, G

    2016-03-01

    In the present paper, the novel software GTest is introduced, designed for testing the normality of a user-specified empirical distribution. It has been implemented with two unusual characteristics; the first is the user option of selecting four different versions of the normality test, each of them suited to be applied to a specific dataset or goal, and the second is the inferential paradigm that informs the output of such tests: it is basically graphical and intrinsically self-explanatory. The concept of inference-by-eye is an emerging inferential approach which will find a successful application in the near future due to the growing need of widening the audience of users of statistical methods to people with informal statistical skills. For instance, the latest European regulation concerning environmental issues introduced strict protocols for data handling (data quality assurance, outliers detection, etc.) and information exchange (areal statistics, trend detection, etc.) between regional and central environmental agencies. Therefore, more and more frequently, laboratory and field technicians will be requested to utilize complex software applications for subjecting data coming from monitoring, surveying or laboratory activities to specific statistical analyses. Unfortunately, inferential statistics, which actually influence the decisional processes for the correct managing of environmental resources, are often implemented in a way which expresses its outcomes in a numerical form with brief comments in a strict statistical jargon (degrees of freedom, level of significance, accepted/rejected H0, etc.). Therefore, often, the interpretation of such outcomes is really difficult for people with poor statistical knowledge. In such framework, the paradigm of the visual inference can contribute to fill in such gap, providing outcomes in self-explanatory graphical forms with a brief comment in the common language. Actually, the difficulties experienced by colleagues and their request for an effective tool for addressing such difficulties motivated us in adopting the inference-by-eye paradigm and implementing an easy-to-use, quick and reliable statistical tool. GTest visualizes its outcomes as a modified version of the Q-Q plot. The application has been developed in Visual Basic for Applications (VBA) within MS Excel 2010, which demonstrated to have all the characteristics of robustness and reliability needed. GTest provides true graphical normality tests which are as reliable as any statistical quantitative approach but much easier to understand. The Q-Q plots have been integrated with the outlining of an acceptance region around the representation of the theoretical distribution, defined in accordance with the alpha level of significance and the data sample size. The test decision rule is the following: if the empirical scatterplot falls completely within the acceptance region, then it can be concluded that the empirical distribution fits the theoretical one at the given alpha level. A comprehensive case study has been carried out with simulated and real-world data in order to check the robustness and reliability of the software.

  16. Rating the methodological quality of single-subject designs and n-of-1 trials: introducing the Single-Case Experimental Design (SCED) Scale.

    PubMed

    Tate, Robyn L; McDonald, Skye; Perdices, Michael; Togher, Leanne; Schultz, Regina; Savage, Sharon

    2008-08-01

    Rating scales that assess methodological quality of clinical trials provide a means to critically appraise the literature. Scales are currently available to rate randomised and non-randomised controlled trials, but there are none that assess single-subject designs. The Single-Case Experimental Design (SCED) Scale was developed for this purpose and evaluated for reliability. Six clinical researchers who were trained and experienced in rating methodological quality of clinical trials developed the scale and participated in reliability studies. The SCED Scale is an 11-item rating scale for single-subject designs, of which 10 items are used to assess methodological quality and use of statistical analysis. The scale was developed and refined over a 3-year period. Content validity was addressed by identifying items to reduce the main sources of bias in single-case methodology as stipulated by authorities in the field, which were empirically tested against 85 published reports. Inter-rater reliability was assessed using a random sample of 20/312 single-subject reports archived in the Psychological Database of Brain Impairment Treatment Efficacy (PsycBITE). Inter-rater reliability for the total score was excellent, both for individual raters (overall ICC = 0.84; 95% confidence interval 0.73-0.92) and for consensus ratings between pairs of raters (overall ICC = 0.88; 95% confidence interval 0.78-0.95). Item reliability was fair to excellent for consensus ratings between pairs of raters (range k = 0.48 to 1.00). The results were replicated with two independent novice raters who were trained in the use of the scale (ICC = 0.88, 95% confidence interval 0.73-0.95). The SCED Scale thus provides a brief and valid evaluation of methodological quality of single-subject designs, with the total score demonstrating excellent inter-rater reliability using both individual and consensus ratings. Items from the scale can also be used as a checklist in the design, reporting and critical appraisal of single-subject designs, thereby assisting to improve standards of single-case methodology.

  17. Hazard Assessment in a Big Data World

    NASA Astrophysics Data System (ADS)

    Kossobokov, Vladimir; Nekrasova, Anastasia

    2017-04-01

    Open data in a Big Data World provides unprecedented opportunities for enhancing scientific studies and better understanding of the Earth System. At the same time, it opens wide avenues for deceptive associations in inter- and transdisciplinary data misleading to erroneous predictions, which are unacceptable for implementation. Even the advanced tools of data analysis may lead to wrong assessments when inappropriately used to describe the phenomenon under consideration. A (self-) deceptive conclusion could be avoided by verification of candidate models in experiments on empirical data and in no other way. Seismology is not an exception. Moreover, seismic evidences accumulated to-date demonstrate clearly that most of the empirical relations commonly accepted in early history of instrumental seismology can be proved erroneous when subjected to objective hypothesis testing. In many cases of seismic hazard assessment (SHA), either probabilistic or deterministic, term-less or short-term, the claims of a high potential of a model forecasts are based on a flawed application of statistics and, therefore, are hardly suitable for communication to decision makers, which situation creates numerous deception points and resulted controversies. So far, most, if not all, the standard probabilistic methods to assess seismic hazard and associated risks are based on subjective, commonly unrealistic, and even erroneous assumptions about seismic recurrence and none of the proposed short-term precursory signals showed sufficient evidence to be used as a reliable precursor of catastrophic earthquakes. Accurate testing against real observations must be done in advance claiming seismically hazardous areas and/or times. The set of errors of the first and second kind in such a comparison permits evaluating the SHA method effectiveness and determining the optimal choice of parameters in regard to a user-defined cost-benefit function. The information obtained in testing experiments may supply us with realistic estimates of confidence and accuracy of SHA predictions. If proved reliable, but not necessarily perfect, forecast/prediction related recommendations on the level of risks in regard to engineering design, insurance, and emergency management can be used for efficient decision making.

  18. Can high resolution 3D topographic surveys provide reliable grain size estimates in gravel bed rivers?

    NASA Astrophysics Data System (ADS)

    Pearson, E.; Smith, M. W.; Klaar, M. J.; Brown, L. E.

    2017-09-01

    High resolution topographic surveys such as those provided by Structure-from-Motion (SfM) contain a wealth of information that is not always exploited in the generation of Digital Elevation Models (DEMs). In particular, several authors have related sub-metre scale topographic variability (or 'surface roughness') to sediment grain size by deriving empirical relationships between the two. In fluvial applications, such relationships permit rapid analysis of the spatial distribution of grain size over entire river reaches, providing improved data to drive three-dimensional hydraulic models, allowing rapid geomorphic monitoring of sub-reach river restoration projects, and enabling more robust characterisation of riverbed habitats. However, comparison of previously published roughness-grain-size relationships shows substantial variability between field sites. Using a combination of over 300 laboratory and field-based SfM surveys, we demonstrate the influence of inherent survey error, irregularity of natural gravels, particle shape, grain packing structure, sorting, and form roughness on roughness-grain-size relationships. Roughness analysis from SfM datasets can accurately predict the diameter of smooth hemispheres, though natural, irregular gravels result in a higher roughness value for a given diameter and different grain shapes yield different relationships. A suite of empirical relationships is presented as a decision tree which improves predictions of grain size. By accounting for differences in patch facies, large improvements in D50 prediction are possible. SfM is capable of providing accurate grain size estimates, although further refinement is needed for poorly sorted gravel patches, for which c-axis percentiles are better predicted than b-axis percentiles.

  19. Understanding the Reliability of Solder Joints Used in Advanced Structural and Electronics Applications: Part 2 - Reliability Performance.

    DOE PAGES

    Vianco, Paul T.

    2017-03-01

    Whether structural or electronic, all solder joints must provide the necessary level of reliability for the application. The Part 1 report examined the effects of filler metal properties and the soldering process on joint reliability. Filler metal solderability and mechanical properties, as well as the extents of base material dissolution and interface reaction that occur during the soldering process, were shown to affect reliability performance. The continuation of this discussion is presented in this Part 2 report, which highlights those factors that directly affect solder joint reliability. There is the growth of an intermetallic compound (IMC) reaction layer at themore » solder/base material interface by means of solid-state diffusion processes. In terms of mechanical response by the solder joint, fatigue remains as the foremost concern for long-term performance. Thermal mechanical fatigue (TMF), a form of low-cycle fatigue (LCF), occurs when temperature cycling is combined with mismatched values of the coefficient of thermal expansion (CTE) between materials comprising the solder joint “system.” Vibration environments give rise to high-cycle fatigue (HCF) degradation. Although accelerated aging studies provide valuable empirical data, too many variants of filler metals, base materials, joint geometries, and service environments are forcing design engineers to embrace computational modeling to predict the long-term reliability of solder joints.« less

  20. Quantifying interactions between real oscillators with information theory and phase models: Application to cardiorespiratory coupling

    NASA Astrophysics Data System (ADS)

    Zhu, Yenan; Hsieh, Yee-Hsee; Dhingra, Rishi R.; Dick, Thomas E.; Jacono, Frank J.; Galán, Roberto F.

    2013-02-01

    Interactions between oscillators can be investigated with standard tools of time series analysis. However, these methods are insensitive to the directionality of the coupling, i.e., the asymmetry of the interactions. An elegant alternative was proposed by Rosenblum and collaborators [M. G. Rosenblum, L. Cimponeriu, A. Bezerianos, A. Patzak, and R. Mrowka, Phys. Rev. EPLEEE81063-651X10.1103/PhysRevE.65.041909 65, 041909 (2002); M. G. Rosenblum and A. S. Pikovsky, Phys. Rev. EPLEEE81063-651X10.1103/PhysRevE.64.045202 64, 045202 (2001)] which consists in fitting the empirical phases to a generic model of two weakly coupled phase oscillators. This allows one to obtain the interaction functions defining the coupling and its directionality. A limitation of this approach is that a solution always exists in the least-squares sense, even in the absence of coupling. To preclude spurious results, we propose a three-step protocol: (1) Determine if a statistical dependency exists in the data by evaluating the mutual information of the phases; (2) if so, compute the interaction functions of the oscillators; and (3) validate the empirical oscillator model by comparing the joint probability of the phases obtained from simulating the model with that of the empirical phases. We apply this protocol to a model of two coupled Stuart-Landau oscillators and show that it reliably detects genuine coupling. We also apply this protocol to investigate cardiorespiratory coupling in anesthetized rats. We observe reciprocal coupling between respiration and heartbeat and that the influence of respiration on the heartbeat is generally much stronger than vice versa. In addition, we find that the vagus nerve mediates coupling in both directions.

  1. Psychosocial stressors and the prognosis of major depression: a test of Axis IV

    PubMed Central

    Gilman, Stephen E.; Trinh, Nhi-Ha; Smoller, Jordan W.; Fava, Maurizio; Murphy, Jane M.; Breslau, Joshua

    2013-01-01

    Background Axis IV is for reporting “psychosocial and environmental problems that may affect the diagnosis, treatment, and prognosis of mental disorders.” No studies have examined the prognostic value of Axis IV in DSM-IV. Method We analyzed data from 2,497 participants in the National Epidemiologic Survey on Alcohol and Related Conditions with major depressive episode (MDE). We hypothesized that psychosocial stressors predict a poor prognosis of MDE. Secondarily, we hypothesized that psychosocial stressors predict a poor prognosis of anxiety and substance use disorders. Stressors were defined according to DSM-IV’s taxonomy, and empirically using latent class analysis. Results Primary support group problems, occupational problems, and childhood adversity increased the risks of depressive episodes and suicidal ideation by 20–30%. Associations of the empirically derived classes of stressors with depression were larger in magnitude. Economic stressors conferred a 1.5-fold increase in risk for a depressive episode (CI=1.2–1.9); financial and interpersonal instability conferred a 1.3-fold increased risk of recurrent depression (CI=1.1–1.6). These two classes of stressors also predicted the recurrence of anxiety and substance use disorders. Stressors were not related to suicidal ideation independent from depression severity. Conclusions Psychosocial and environmental problems are associated with the prognosis of MDE and other Axis I disorders. Though DSM-IV’s taxonomy of stressors stands to be improved, these results provide empirical support for the prognostic value of Axis IV. Future work is needed to determine the reliability of Axis IV assessments in clinical practice, and the usefulness of this information to improving the clinical course of mental disorders. PMID:22640506

  2. Development and validation of a music performance anxiety inventory for gifted adolescent musicians.

    PubMed

    Osborne, Margaret S; Kenny, Dianna T

    2005-01-01

    Music performance anxiety (MPA) is a distressing experience for musicians of all ages, yet the empirical investigation of MPA in adolescents has received little attention to date. No measures specifically targeting MPA in adolescents have been empirically validated. This article presents findings of an initial study into the psychometric properties and validation of the Music Performance Anxiety Inventory for Adolescents (MPAI-A), a new self-report measure of MPA for this group. Data from 381 elite young musicians aged 12-19 years was used to investigate the factor structure, internal reliability, construct and divergent validity of the MPAI-A. Cronbach's alpha for the full measure was .91. Factor analysis identified three factors, which together accounted for 53% of the variance. Construct validity was demonstrated by significant positive relationships with social phobia (measured using the Social Phobia Anxiety Inventory [Beidel, D. C., Turner, S. M., & Morris, T. L. (1995). A new inventory to assess childhood social anxiety and phobia: The Social Phobia and Anxiety Inventory for Children. Psychological Assessment, 7(1), 73-79; Beidel, D. C., Turner, S. M., & Morris, T. L. (1998). Social Phobia and Anxiety Inventory for Children (SPAI-C). North Tonawanda, NY: Multi-Health Systems Inc.]) and trait anxiety (measured using the State Trait Anxiety Inventory [Spielberger, C. D. (1983). State-Trait Anxiety Inventory STAI (Form Y). Palo Alto, CA: Consulting Psychologists Press, Inc.]). The MPAI-A demonstrated convergent validity by a moderate to strong positive correlation with an adult measure of MPA. Discriminant validity was established by a weaker positive relationship with depression, and no relationship with externalizing behavior problems. It is hoped that the MPAI-A, as the first empirically validated measure of adolescent musicians' performance anxiety, will enhance and promote phenomenological and treatment research in this area.

  3. Analysis of copper and brass coins of the early roman empire.

    PubMed

    Carter, G F

    1966-01-14

    X-ray fluorescence analysis of 14 copper and brass coins of the early Roman Empire shows differences in composition between coins minted in Rome and in France. Concentrations of tin, lead, and antimony are nearly always less than in coins minted before 29 B.C. or after 54 A.D. Older coins were not melted to make copper coins of the early empire.

  4. Assessing differential expression in two-color microarrays: a resampling-based empirical Bayes approach.

    PubMed

    Li, Dongmei; Le Pape, Marc A; Parikh, Nisha I; Chen, Will X; Dye, Timothy D

    2013-01-01

    Microarrays are widely used for examining differential gene expression, identifying single nucleotide polymorphisms, and detecting methylation loci. Multiple testing methods in microarray data analysis aim at controlling both Type I and Type II error rates; however, real microarray data do not always fit their distribution assumptions. Smyth's ubiquitous parametric method, for example, inadequately accommodates violations of normality assumptions, resulting in inflated Type I error rates. The Significance Analysis of Microarrays, another widely used microarray data analysis method, is based on a permutation test and is robust to non-normally distributed data; however, the Significance Analysis of Microarrays method fold change criteria are problematic, and can critically alter the conclusion of a study, as a result of compositional changes of the control data set in the analysis. We propose a novel approach, combining resampling with empirical Bayes methods: the Resampling-based empirical Bayes Methods. This approach not only reduces false discovery rates for non-normally distributed microarray data, but it is also impervious to fold change threshold since no control data set selection is needed. Through simulation studies, sensitivities, specificities, total rejections, and false discovery rates are compared across the Smyth's parametric method, the Significance Analysis of Microarrays, and the Resampling-based empirical Bayes Methods. Differences in false discovery rates controls between each approach are illustrated through a preterm delivery methylation study. The results show that the Resampling-based empirical Bayes Methods offer significantly higher specificity and lower false discovery rates compared to Smyth's parametric method when data are not normally distributed. The Resampling-based empirical Bayes Methods also offers higher statistical power than the Significance Analysis of Microarrays method when the proportion of significantly differentially expressed genes is large for both normally and non-normally distributed data. Finally, the Resampling-based empirical Bayes Methods are generalizable to next generation sequencing RNA-seq data analysis.

  5. The Validity and Reliability of the Violence Risk Scale-Sexual Offender Version: Assessing Sex Offender Risk and Evaluating Therapeutic Change

    ERIC Educational Resources Information Center

    Olver, Mark E.; Wong, Stephen C. P.; Nicholaichuk, Terry; Gordon, Audrey

    2007-01-01

    The Violence Risk Scale-Sexual Offender version (VRS-SO) is a rating scale designed to assess risk and predict sexual recidivism, to measure and link treatment changes to sexual recidivism, and to inform the delivery of sexual offender treatment. The VRS-SO comprises 7 static and 17 dynamic items empirically or conceptually linked to sexual…

  6. Statistical models for the distribution of modulus of elasticity and modulus of rupture in lumber with implications for reliability calculations

    Treesearch

    Steve P. Verrill; Frank C. Owens; David E. Kretschmann; Rubin Shmulsky

    2017-01-01

    It is common practice to assume that a two-parameter Weibull probability distribution is suitable for modeling lumber properties. Verrill and co-workers demonstrated theoretically and empirically that the modulus of rupture (MOR) distribution of visually graded or machine stress rated (MSR) lumber is not distributed as a Weibull. Instead, the tails of the MOR...

  7. A reliability analysis tool for SpaceWire network

    NASA Astrophysics Data System (ADS)

    Zhou, Qiang; Zhu, Longjiang; Fei, Haidong; Wang, Xingyou

    2017-04-01

    A SpaceWire is a standard for on-board satellite networks as the basis for future data-handling architectures. It is becoming more and more popular in space applications due to its technical advantages, including reliability, low power and fault protection, etc. High reliability is the vital issue for spacecraft. Therefore, it is very important to analyze and improve the reliability performance of the SpaceWire network. This paper deals with the problem of reliability modeling and analysis with SpaceWire network. According to the function division of distributed network, a reliability analysis method based on a task is proposed, the reliability analysis of every task can lead to the system reliability matrix, the reliability result of the network system can be deduced by integrating these entire reliability indexes in the matrix. With the method, we develop a reliability analysis tool for SpaceWire Network based on VC, where the computation schemes for reliability matrix and the multi-path-task reliability are also implemented. By using this tool, we analyze several cases on typical architectures. And the analytic results indicate that redundancy architecture has better reliability performance than basic one. In practical, the dual redundancy scheme has been adopted for some key unit, to improve the reliability index of the system or task. Finally, this reliability analysis tool will has a directive influence on both task division and topology selection in the phase of SpaceWire network system design.

  8. Development of guidance for states transitioning to new safety analysis tools

    NASA Astrophysics Data System (ADS)

    Alluri, Priyanka

    With about 125 people dying on US roads each day, the US Department of Transportation heightened the awareness of critical safety issues with the passage of SAFETEA-LU (Safe Accountable Flexible Efficient Transportation Equity Act---a Legacy for Users) legislation in 2005. The legislation required each of the states to develop a Strategic Highway Safety Plan (SHSP) and incorporate data-driven approaches to prioritize and evaluate program outcomes: Failure to do so resulted in funding sanctioning. In conjunction with the legislation, research efforts have also been progressing toward the development of new safety analysis tools such as IHSDM (Interactive Highway Safety Design Model), SafetyAnalyst, and HSM (Highway Safety Manual). These software and analysis tools are comparatively more advanced in statistical theory and level of accuracy, and have a tendency to be more data intensive. A review of the 2009 five-percent reports and excerpts from the nationwide survey revealed astonishing facts about the continuing use of traditional methods including crash frequencies and rates for site selection and prioritization. The intense data requirements and statistical complexity of advanced safety tools are considered as a hindrance to their adoption. In this context, this research aims at identifying the data requirements and data availability for SafetyAnalyst and HSM by working with both the tools. This research sets the stage for working with the Empirical Bayes approach by highlighting some of the biases and issues associated with the traditional methods of selecting projects such as greater emphasis on traffic volume and regression-to-mean phenomena. Further, the not-so-obvious issue with shorter segment lengths, which effect the results independent of the methods used, is also discussed. The more reliable and statistically acceptable Empirical Bayes methodology requires safety performance functions (SPFs), regression equations predicting the relation between crashes and exposure for a subset of roadway network. These SPFs, specific to a region and the analysis period are often unavailable. Calibration of already existing default national SPFs to the state's data could be a feasible solution, but, how well the state's data is represented is a legitimate question. With this background, SPFs were generated for various classifications of segments in Georgia and compared against the national default SPFs used in SafetyAnalyst calibrated to Georgia data. Dwelling deeper into the development of SPFs, the influence of actual and estimated traffic data on the fit of the equations is also studied questioning the accuracy and reliability of traffic estimations. In addition to SafetyAnalyst, HSM aims at performing quantitative safety analysis. Applying HSM methodology to two-way two-lane rural roads, the effect of using multiple CMFs (Crash Modification Factors) is studied. Lastly, data requirements, methodology, constraints, and results are compared between SafetyAnalyst and HSM.

  9. A succinct rating scale for radiology report quality

    PubMed Central

    Yang, Chengwu; Ouyang, Tao; Peterson, Christine M; Sarwani, Nabeel I; Tappouni, Rafel; Bruno, Michael

    2014-01-01

    Context: Poorly written radiology reports are common among residents and are a significant challenge for radiology education. While training may improve report quality, a professionally developed reliable and valid scale to measure report quality does not exist. Objectives: To develop a measurement tool for report quality, the quality of report scale, with rigorous validation through empirical data. Methods: A research team of an experienced psychometrician and six senior radiologists conducted qualitative and quantitative studies. Five items were identified for the quality of report scale, each measuring a distinct aspect of report quality. Two dedicated training sessions were designed and implemented to help residents generate high-quality reports. In a blinded fashion, the quality of report scale was applied to 804 randomly selected reports issued before (n = 403) and after (n = 401) training. Full-scale psychometrical assessments were implemented onto the quality of report scale’s item- and scale-scores from the reports. The quality of report scale scores were correlated with report professionalism and attendings’ preference and were compared pre-/post-training. Results: The quality of report scale showed sound psychometrical properties, with high validity and reliability. Reports with higher quality of report scale score were more professional and preferable by attendings. Training improved the quality of report scale score, empirically validating the quality of report scale further. Conclusion: While succinct and practitioner friendly, the quality of report scale is a reliable and valid measure of radiology report quality and has the potential to be easily adapted to other fields such as pathology, where similar training would be beneficial. PMID:26770756

  10. Assessing positive and negative experiences: validation of a new measure of well-being in an Italian population.

    PubMed

    Corno, Giulia; Molinari, Guadalupe; Baños, Rosa Maria

    2016-01-01

    The aim of this study is to explore the psychometric properties of an affect scale, the Scale of Positive and Negative Experience (SPANE), in an Italian-speaking population. The results of this study demonstrate that the Italian version of the SPANE has psychometric properties similar to those shown by the original and previous versions, and it presents satisfactory reliability and factorial validity. The results of the Confirmatory Factor Analysis support the expected two-factor structure, positive and negative feeling, which characterized the previous versions. As expected, measures of negative affect, anxiety, negative future expectances, and depression correlated positively with the negative experiences SPANE subscale, and negatively with the positive experiences SPANE subscale. Results of this study demonstrate that the Italian version of the SPANE has psychometric properties similar to those shown by the original and previous versions, and it presents satisfactory reliability and factorial validity. The use of this instrument provides clinically useful information about a person’s overall emotional experience and it is an indicator of well-being. Although further studies are required to confirm the psychometric characteristics of the scale, the SPANE Italian version is expected to improve theoretical and empirical research on the well-being of the Italian population.

  11. The validity and scalability of the Theory of Mind Scale with toddlers and preschoolers.

    PubMed

    Hiller, Rachel M; Weber, Nathan; Young, Robyn L

    2014-12-01

    Despite the importance of theory of mind (ToM) for typical development, there remain 2 key issues affecting our ability to draw robust conclusions. One is the continued focus on false belief as the sole measure of ToM. The second is the lack of empirically validated measures of ToM as a broad construct. Our key aim was to examine the validity and reliability of the 5-item ToM scale (Peterson, Wellman, & Liu, 2005). In particular, we extended on previous research of this scale by assessing its scalability and validity for use with children from 2 years of age. Sixty-eight typically developing children (aged 24 to 61 months) were assessed on the scale's 5 tasks, along with a sixth Sally-Anne false-belief task. Our data replicated the scalability of the 5 tasks for a Rasch-but not Guttman-scale. Guttman analysis showed that a 4-item scale may be more suitable for this age range. Further, the tasks showed good internal consistency and validity for use with children as young as 2 years of age. Overall, the measure provides a valid and reliable tool for the assessment of ToM, and in particular, the longitudinal assessment of this ability as a construct. (c) 2014 APA, all rights reserved.

  12. A Spanish Validation of the Canadian Adolescent Gambling Inventory (CAGI)

    PubMed Central

    Jiménez-Murcia, Susana; Granero, Roser; Stinchfield, Randy; Tremblay, Joël; del Pino-Gutiérrez, Amparo; Moragas, Laura; Savvidou, Lamprini G.; Fernández-Aranda, Fernando; Aymamí, Neus; Gómez-Peña, Mónica; Tárrega, Salomé; Gunnard, Katarina; Martín-Romera, Virginia; Steward, Trevor; Mestre-Bach, Gemma; Menchón, José M.

    2017-01-01

    Aims: Large-scale epidemiological studies show a significant prevalence of gambling disorder (GD) during adolescence and emerging adulthood, and highlight the need to identify gambling-related behaviors at early ages. However, there are only a handful of screening instruments for this population and many studies measuring youth gambling problems use adult instruments that may not be developmentally appropriate. The aim of this study was to validate a Spanish version of the Canadian Adolescent Gambling Inventory (CAGI) among late adolescent and young adults and to explore its psychometric properties. Methods: The sample (16–29 years old) included a clinical group (n = 55) with GD patients and a control group (n = 340). Results: Exploratory factor analysis yielded one factor as the best model. This 24-item scale demonstrated satisfactory reliability (internal consistency, Cronbach’s alpha, α = 0.91), satisfactory convergent validity as measured by correlation with South Oaks Gambling Screen (r = 0.74), and excellent classification accuracy (AUC = 0.99; sensitivity = 0.98; and specificity = 0.99). Conclusion: Our results provide empirical support for our validation of the Spanish version of the CAGI. We uphold that the Spanish CAGI can be used as a brief, reliable, and valid instrument to assess gambling problems in Spanish youth. PMID:28223961

  13. Development and validation of a brief trauma screening measure for children: The Child Trauma Screen.

    PubMed

    Lang, Jason M; Connell, Christian M

    2017-05-01

    Childhood exposure to trauma, including violence and abuse, is a major public health concern that has resulted in increased efforts to promote trauma-informed child-serving systems. Trauma screening is an important component of such trauma-informed systems, yet widespread use of trauma screening is rare in part due to the lack of brief, validated trauma screening measures for children. We describe development and validation of the Child Trauma Screen (CTS), a 10-item screening measure of trauma exposure and posttraumatic stress disorder (PTSD) symptoms for children consistent with the DSM-5 definition of PTSD. Study 1 describes measure development incorporating analysis to derive items based on existing measures from 1,065 children and caregivers together with stakeholder input to finalize item selection. Study 2 describes validation of the CTS with a clinical sample of 74 children and their caregivers. Results support the CTS as an empirically derived, reliable measure to screen children for trauma exposure and PTSD symptoms with strong convergent, divergent, and criterion validity. The CTS is a promising measure for rapidly and reliably screening children for trauma exposure and PTSD symptoms. Future research is needed to confirm validation and to examine feasibility and utility of its use across various child-serving systems. (PsycINFO Database Record (c) 2017 APA, all rights reserved).

  14. A model of scientific attitudes assessment by observation in physics learning based scientific approach: case study of dynamic fluid topic in high school

    NASA Astrophysics Data System (ADS)

    Yusliana Ekawati, Elvin

    2017-01-01

    This study aimed to produce a model of scientific attitude assessment in terms of the observations for physics learning based scientific approach (case study of dynamic fluid topic in high school). Development of instruments in this study adaptation of the Plomp model, the procedure includes the initial investigation, design, construction, testing, evaluation and revision. The test is done in Surakarta, so that the data obtained are analyzed using Aiken formula to determine the validity of the content of the instrument, Cronbach’s alpha to determine the reliability of the instrument, and construct validity using confirmatory factor analysis with LISREL 8.50 program. The results of this research were conceptual models, instruments and guidelines on scientific attitudes assessment by observation. The construct assessment instruments include components of curiosity, objectivity, suspended judgment, open-mindedness, honesty and perseverance. The construct validity of instruments has been qualified (rated load factor > 0.3). The reliability of the model is quite good with the Alpha value 0.899 (> 0.7). The test showed that the model fits the theoretical models are supported by empirical data, namely p-value 0.315 (≥ 0.05), RMSEA 0.027 (≤ 0.08)

  15. Empirical equations for viscosity and specific heat capacity determination of paraffin PCM and fatty acid PCM

    NASA Astrophysics Data System (ADS)

    Barreneche, C.; Ferrer, G.; Palacios, A.; Solé, A.; Inés Fernández, A.; Cabeza, L. F.

    2017-10-01

    Phase change materials (PCM) used in thermal energy storage (TES) systems have been presented, over recent years, as one of the most effective options in energy storage. Paraffin and fatty acids are some of the most used PCM in TES systems, as they have high phase change enthalpy and in addition they do not present subcooling nor hysteresis and have proper cycling stability. The simulations and design of TES systems require the knowledge of the thermophysical properties of PCM. Thermal conductivity, viscosity, specific heat capacity (Cp) can be experimentally determined, but these are material and time consuming tasks. To avoid or to reduce them, and to have reliable data without the need of experimentation, thermal properties can be calculated by empirical equations. In this study, five different equations are given to calculate the viscosity and specific heat capacity of fatty acid PCM and paraffin PCM. Two of these equations concern, respectively, the empirical calculation of the viscosity and liquid Cp of the whole paraffin PCM family, while the other three equations presented are for the corresponding calculation of viscosity, solid Cp, liquid Cp of the whole fatty acid family of PCM. Therefore, this study summarize the work performed to obtain the main empirical equations to measure the above mentioned properties for whole fatty acid PCM family and whole paraffin PCM family. Moreover, empirical equations have been obtained to calculate these properties for other materials of these PCM groups and these empirical equations can be extrapolated for PCM with higher or lower phase change temperatures within a lower relative error 4%.

  16. An Investigation of Document Partitions.

    ERIC Educational Resources Information Center

    Shaw, W. M., Jr.

    1986-01-01

    Empirical significance of document partitions is investigated as a function of index term-weight and similarity thresholds. Results show the same empirically preferred partitions can be detected by two independent strategies: an analysis of cluster-based retrieval analysis and an analysis of regularities in the underlying structure of the document…

  17. Regional Morphology Analysis Package (RMAP): Empirical Orthogonal Function Analysis, Background and Examples

    DTIC Science & Technology

    2007-10-01

    1984. Complex principal component analysis : Theory and examples. Journal of Climate and Applied Meteorology 23: 1660-1673. Hotelling, H. 1933...Sediments 99. ASCE: 2,566-2,581. Von Storch, H., and A. Navarra. 1995. Analysis of climate variability. Applications of statistical techniques. Berlin...ERDC TN-SWWRP-07-9 October 2007 Regional Morphology Empirical Analysis Package (RMAP): Orthogonal Function Analysis , Background and Examples by

  18. GAP-REACH

    PubMed Central

    Lewis-Fernández, Roberto; Raggio, Greer A.; Gorritz, Magdaliz; Duan, Naihua; Marcus, Sue; Cabassa, Leopoldo J.; Humensky, Jennifer; Becker, Anne E.; Alarcón, Renato D.; Oquendo, María A.; Hansen, Helena; Like, Robert C.; Weiss, Mitchell; Desai, Prakash N.; Jacobsen, Frederick M.; Foulks, Edward F.; Primm, Annelle; Lu, Francis; Kopelowicz, Alex; Hinton, Ladson; Hinton, Devon E.

    2015-01-01

    Growing awareness of health and health care disparities highlights the importance of including information about race, ethnicity, and culture (REC) in health research. Reporting of REC factors in research publications, however, is notoriously imprecise and unsystematic. This article describes the development of a checklist to assess the comprehensiveness and the applicability of REC factor reporting in psychiatric research publications. The 16-itemGAP-REACH© checklist was developed through a rigorous process of expert consensus, empirical content analysis in a sample of publications (N = 1205), and interrater reliability (IRR) assessment (N = 30). The items assess each section in the conventional structure of a health research article. Data from the assessment may be considered on an item-by-item basis or as a total score ranging from 0% to 100%. The final checklist has excellent IRR (κ = 0.91). The GAP-REACH may be used by multiple research stakeholders to assess the scope of REC reporting in a research article. PMID:24080673

  19. NO RELATIONSHIP BETWEEN INTELLIGENCE AND FACIAL ATTRACTIVENESS IN A LARGE, GENETICALLY INFORMATIVE SAMPLE

    PubMed Central

    Mitchem, Dorian G.; Zietsch, Brendan P.; Wright, Margaret J.; Martin, Nicholas G.; Hewitt, John K.; Keller, Matthew C.

    2015-01-01

    Theories in both evolutionary and social psychology suggest that a positive correlation should exist between facial attractiveness and general intelligence, and several empirical observations appear to corroborate this expectation. Using highly reliable measures of facial attractiveness and IQ in a large sample of identical and fraternal twins and their siblings, we found no evidence for a phenotypic correlation between these traits. Likewise, neither the genetic nor the environmental latent factor correlations were statistically significant. We supplemented our analyses of new data with a simple meta-analysis that found evidence of publication bias among past studies of the relationship between facial attractiveness and intelligence. In view of these results, we suggest that previously published reports may have overestimated the strength of the relationship and that the theoretical bases for the predicted attractiveness-intelligence correlation may need to be reconsidered. PMID:25937789

  20. Spatio-temporal water quality mapping from satellite images using geographically and temporally weighted regression

    NASA Astrophysics Data System (ADS)

    Chu, Hone-Jay; Kong, Shish-Jeng; Chang, Chih-Hua

    2018-03-01

    The turbidity (TB) of a water body varies with time and space. Water quality is traditionally estimated via linear regression based on satellite images. However, estimating and mapping water quality require a spatio-temporal nonstationary model, while TB mapping necessitates the use of geographically and temporally weighted regression (GTWR) and geographically weighted regression (GWR) models, both of which are more precise than linear regression. Given the temporal nonstationary models for mapping water quality, GTWR offers the best option for estimating regional water quality. Compared with GWR, GTWR provides highly reliable information for water quality mapping, boasts a relatively high goodness of fit, improves the explanation of variance from 44% to 87%, and shows a sufficient space-time explanatory power. The seasonal patterns of TB and the main spatial patterns of TB variability can be identified using the estimated TB maps from GTWR and by conducting an empirical orthogonal function (EOF) analysis.

  1. Quantifying the Behavior of Stock Correlations Under Market Stress

    PubMed Central

    Preis, Tobias; Kenett, Dror Y.; Stanley, H. Eugene; Helbing, Dirk; Ben-Jacob, Eshel

    2012-01-01

    Understanding correlations in complex systems is crucial in the face of turbulence, such as the ongoing financial crisis. However, in complex systems, such as financial systems, correlations are not constant but instead vary in time. Here we address the question of quantifying state-dependent correlations in stock markets. Reliable estimates of correlations are absolutely necessary to protect a portfolio. We analyze 72 years of daily closing prices of the 30 stocks forming the Dow Jones Industrial Average (DJIA). We find the striking result that the average correlation among these stocks scales linearly with market stress reflected by normalized DJIA index returns on various time scales. Consequently, the diversification effect which should protect a portfolio melts away in times of market losses, just when it would most urgently be needed. Our empirical analysis is consistent with the interesting possibility that one could anticipate diversification breakdowns, guiding the design of protected portfolios. PMID:23082242

  2. Three-Dimensional Model of Strengths: Examination of Invariance Across Gender, Age, Education Levels, and Marriage Status.

    PubMed

    Duan, Wenjie; Ho, Samuel Mun Yin

    2017-02-01

    Strengths are positive qualities that significantly contributed to well-being of individuals and community. Therefore, a reliable and valid measure of strengths for research and practice is needed. The Brief Strengths Scale (BSS) is a newly developed tool for measuring the three-dimensional strengths model (i.e., temperance, intellectual, and interpersonal strength). However, empirical support for the measurement invariance of the BSS has not been obtained. This study examined the three-factor structure of BSS across gender, age, education, and marriage groups in a community sample (n = 375) using multi-group confirmatory factor analysis. After removing one item of each subscale from the original version, the revised model provided a good fit to the data at different subgroups. The revised nine-item BSS indicated that measurement invariance across gender and age groups was achieved. In addition, the measurement was more influenced by social-cultural factors than biological factors.

  3. Patient's Communication Perceived Self-efficacy Scale (PCSS): construction and validation of a new measure in a socio-cognitive perspective.

    PubMed

    Capone, Vincenza; Petrillo, Giovanna

    2014-06-01

    In two studies we constructed and validated the Patient's Communication Perceived Self-efficacy Scale (PCSS) designed to assess patients' beliefs about their capability to successfully manage problematic situations related to communication with doctor. The 20-item scale was administered to 179 outpatients (study 1). An Exploratory Factor Analysis revealed a three-factor solution. In study 2, the 16-item scale was administered to 890 outpatients. Exploratory and Confirmatory Factor Analyses supported the 3-factor solution (Provide and Collect information, Express concerns and doubts, Verify information) that showed good psychometric properties and was invariant for gender. PCSS is an easily administered, reliable, and valid test of patients' communication self-efficacy beliefs. It can be applied optimally in the empirical study of factors influencing doctor-patient communication and used in training aimed at strengthening patients' communication skills. Copyright © 2014 Elsevier Ireland Ltd. All rights reserved.

  4. Development of measures of organizational leadership for health promotion.

    PubMed

    Barrett, Linda; Plotnikoff, Ronald C; Raine, Kim; Anderson, Donna

    2005-04-01

    This study describes the development and psychometric evaluation of four scales measuring leadership for health promotion at an organizational level in the baseline survey (n=144) of the Alberta Heart Health Project. Content validity was established through a series of focus groups and expert opinion appraisals, pilot testing of a draft based on capacity assessment instruments developed by other provinces involved in the Canadian Heart Health Initiative, and the literature. Psychometric analyses provided empirical evidence of the construct validity and reliability of the organizational leadership scales. Principal component analysis verified the unidimensionality of the leadership scales of (a) Practices for Organizational Learning, (b) Wellness Planning, (c) Workplace Climate, and (d) Organization Member Development. Scale alpha coefficients ranged between .79 and .91 thus establishing good to high scale internal consistencies. These measures can be used by both researchers and practitioners for the assessment of organizational leadership for health promotion and heart health promotion.

  5. Changing skewness: an early warning signal of regime shifts in ecosystems.

    PubMed

    Guttal, Vishwesha; Jayaprakash, Ciriyam

    2008-05-01

    Empirical evidence for large-scale abrupt changes in ecosystems such as lakes and vegetation of semi-arid regions is growing. Such changes, called regime shifts, can lead to degradation of ecological services. We study simple ecological models that show a catastrophic transition as a control parameter is varied and propose a novel early warning signal that exploits two ubiquitous features of ecological systems: nonlinearity and large external fluctuations. Either reduced resilience or increased external fluctuations can tip ecosystems to an alternative stable state. It is shown that changes in asymmetry in the distribution of time series data, quantified by changing skewness, is a model-independent and reliable early warning signal for both routes to regime shifts. Furthermore, using model simulations that mimic field measurements and a simple analysis of real data from abrupt climate change in the Sahara, we study the feasibility of skewness calculations using data available from routine monitoring.

  6. User acceptance of mobile commerce: an empirical study in Macau

    NASA Astrophysics Data System (ADS)

    Lai, Ivan K. W.; Lai, Donny C. F.

    2014-06-01

    This study aims to examine the positive and negative factors that can significantly explain user acceptance of mobile commerce (m-commerce) in Macau. A technology acceptance model for m-commerce with five factors is constructed. The proposed model is tested using data collected from 219 respondents. Confirmatory factor analysis is performed to examine the reliability and validity of the model, and structural equation modelling is performed to access the relationship between behaviour intention and each factor. The acceptance of m-commerce is influenced by factors including performance expectancy, social influence, facilitating conditions and privacy concern; while effort expectancy is insignificant in this case. The results of the study are useful for m-commerce service providers to adjust their strategies for promoting m-commerce services. This study contributes to the practice by providing a user technology acceptance model for m-commerce that can be used as a foundation for future research.

  7. Measuring patient participation in surgical treatment decision-making from healthcare professionals' perspective.

    PubMed

    Heggland, Liv-Helen; Mikkelsen, Aslaug; Øgaard, Torvald; Hausken, Kjell

    2014-02-01

    To develop, empirical test, and validate an instrument measuring patient participation in surgical treatment decision-making from healthcare professionals' perspective. Since the advent of New Public Management in many Western countries, patient participation in healthcare decision-making has been considered to be a best practice. A common notion is that well-educated and well-informed public want to choose their own treatments and providers and want to ask questions about the quality of their health services. Survey. A self-report-measuring instrument was designed and administered to 620 healthcare professionals. Items were developed, validated and tested by 451 nurses and physicians working in six surgical wards in a University Hospital in Norway. A 16-item scale with the following four dimensions was developed: information dissemination, formulation of options, integration of information and control. Factor analysis procedures and reliability testing were performed. A one-way, between-groups analysis of variance was conducted to compare doctors' and nurses' opinions on four dimensions of patient participation in surgical treatment decision-making. This article shows that patient participation in surgical treatment decision-making can be measured by a 16-item scale and four distinct dimensions. The analysis demonstrated a reasonable level of construct validity and reliability. Nurses and physicians have a positive attitude towards patient participation overall, but the two groups differ in the extent to which they accept the idea of patient participation in treatment decision-making. The instrument can be a tool for managers and healthcare professionals in the implementation of patient participation in clinical practice. Data from the instrument can be useful to identify health services being provided and what areas that could strengthen patient participation. © 2013 Blackwell Publishing Ltd.

  8. Health e-mavens: identifying active online health information users.

    PubMed

    Sun, Ye; Liu, Miao; Krakow, Melinda

    2016-10-01

    Given the rapid increase of Internet use for effective health communication, it is important for health practitioners to be able to identify and mobilize active users of online health information across various web-based health intervention programmes. We propose the concept 'health e-mavens' to characterize individuals actively engaged in online health information seeking and sharing activities. This study aimed to address three goals: (i) to test the factor structure of health e-mavenism, (ii) to assess the reliability and validity of this construct and (iii) to determine what predictors are associated with health e-mavenism. This study was a secondary analysis of nationally representative data from the 2010 Health Tracking Survey. We assessed the factor structure of health e-mavenism using confirmatory factor analysis and examined socio-demographic variables, health-related factors and use of technology as potential predictors of health e-mavenism through ordered regression analysis. Confirmatory factor analyses showed that a second-order two-factor structure best captured the health e-maven construct. Health e-mavenism comprised two second-order factors, each encompassing two first-order dimensions: information acquisition (consisting of information tracking and consulting) and information transmission (consisting of information posting and sharing). Both first-order and second-order factors exhibited good reliabilities. Several factors were found to be significant predictors of health e-mavenism. This study offers a starting point for further inquiries about health e-mavens. It is a fruitful construct for health promotion research in the age of new media technologies. We conclude with specific recommendations to further develop the health e-maven concept through continued empirical research. © 2015 The Authors. Health Expectations. Published by John Wiley & Sons Ltd.

  9. The best of both worlds: Building on the COPUS and RTOP observation protocols to easily and reliably measure various levels of reformed instructional practice.

    PubMed

    Lund, Travis J; Pilarz, Matthew; Velasco, Jonathan B; Chakraverty, Devasmita; Rosploch, Kaitlyn; Undersander, Molly; Stains, Marilyne

    2015-01-01

    Researchers, university administrators, and faculty members are increasingly interested in measuring and describing instructional practices provided in science, technology, engineering, and mathematics (STEM) courses at the college level. Specifically, there is keen interest in comparing instructional practices between courses, monitoring changes over time, and mapping observed practices to research-based teaching. While increasingly common observation protocols (Reformed Teaching Observation Protocol [RTOP] and Classroom Observation Protocol in Undergraduate STEM [COPUS]) at the postsecondary level help achieve some of these goals, they also suffer from weaknesses that limit their applicability. In this study, we leverage the strengths of these protocols to provide an easy method that enables the reliable and valid characterization of instructional practices. This method was developed empirically via a cluster analysis using observations of 269 individual class periods, corresponding to 73 different faculty members, 28 different research-intensive institutions, and various STEM disciplines. Ten clusters, called COPUS profiles, emerged from this analysis; they represent the most common types of instructional practices enacted in the classrooms observed for this study. RTOP scores were used to validate the alignment of the 10 COPUS profiles with reformed teaching. Herein, we present a detailed description of the cluster analysis method, the COPUS profiles, and the distribution of the COPUS profiles across various STEM courses at research-intensive universities. © 2015 T. J. Lund et al. CBE—Life Sciences Education © 2015 The American Society for Cell Biology. This article is distributed by The American Society for Cell Biology under license from the author(s). It is available to the public under an Attribution–Noncommercial–Share Alike 3.0 Unported Creative Commons License (http://creativecommons.org/licenses/by-nc-sa/3.0).

  10. Critical Realism and Empirical Bioethics: A Methodological Exposition.

    PubMed

    McKeown, Alex

    2017-09-01

    This paper shows how critical realism can be used to integrate empirical data and philosophical analysis within 'empirical bioethics'. The term empirical bioethics, whilst appearing oxymoronic, simply refers to an interdisciplinary approach to the resolution of practical ethical issues within the biological and life sciences, integrating social scientific, empirical data with philosophical analysis. It seeks to achieve a balanced form of ethical deliberation that is both logically rigorous and sensitive to context, to generate normative conclusions that are practically applicable to the problem, challenge, or dilemma. Since it incorporates both philosophical and social scientific components, empirical bioethics is a field that is consistent with the use of critical realism as a research methodology. The integration of philosophical and social scientific approaches to ethics has been beset with difficulties, not least because of the irreducibly normative, rather than descriptive, nature of ethical analysis and the contested relation between fact and value. However, given that facts about states of affairs inform potential courses of action and their consequences, there is a need to overcome these difficulties and successfully integrate data with theory. Previous approaches have been formulated to overcome obstacles in combining philosophical and social scientific perspectives in bioethical analysis; however each has shortcomings. As a mature interdisciplinary approach critical realism is well suited to empirical bioethics, although it has hitherto not been widely used. Here I show how it can be applied to this kind of research and explain how it represents an improvement on previous approaches.

  11. Application and Evaluation of an Expert Judgment Elicitation Procedure for Correlations.

    PubMed

    Zondervan-Zwijnenburg, Mariëlle; van de Schoot-Hubeek, Wenneke; Lek, Kimberley; Hoijtink, Herbert; van de Schoot, Rens

    2017-01-01

    The purpose of the current study was to apply and evaluate a procedure to elicit expert judgments about correlations, and to update this information with empirical data. The result is a face-to-face group elicitation procedure with as its central element a trial roulette question that elicits experts' judgments expressed as distributions. During the elicitation procedure, a concordance probability question was used to provide feedback to the experts on their judgments. We evaluated the elicitation procedure in terms of validity and reliability by means of an application with a small sample of experts. Validity means that the elicited distributions accurately represent the experts' judgments. Reliability concerns the consistency of the elicited judgments over time. Four behavioral scientists provided their judgments with respect to the correlation between cognitive potential and academic performance for two separate populations enrolled at a specific school in the Netherlands that provides special education to youth with severe behavioral problems: youth with autism spectrum disorder (ASD), and youth with diagnoses other than ASD. Measures of face-validity, feasibility, convergent validity, coherence, and intra-rater reliability showed promising results. Furthermore, the current study illustrates the use of the elicitation procedure and elicited distributions in a social science application. The elicited distributions were used as a prior for the correlation, and updated with data for both populations collected at the school of interest. The current study shows that the newly developed elicitation procedure combining the trial roulette method with the elicitation of correlations is a promising tool, and that the results of the procedure are useful as prior information in a Bayesian analysis.

  12. Development and Validity Testing of an Arthritis Self-Management Assessment Tool.

    PubMed

    Oh, HyunSoo; Han, SunYoung; Kim, SooHyun; Seo, WhaSook

    Because of the chronic, progressive nature of arthritis and the substantial effects it has on quality of life, patients may benefit from self-management. However, no valid, reliable self-management assessment tool has been devised for patients with arthritis. This study was conducted to develop a comprehensive self-management assessment tool for patients with arthritis, that is, the Arthritis Self-Management Assessment Tool (ASMAT). To develop a list of qualified items corresponding to the conceptual definitions and attributes of arthritis self-management, a measurement model was established on the basis of theoretical and empirical foundations. Content validity testing was conducted to evaluate whether listed items were suitable for assessing arthritis self-management. Construct validity and reliability of the ASMAT were tested. Construct validity was examined using confirmatory factor analysis and nomological validity. The 32-item ASMAT was developed with a sample composed of patients in a clinic in South Korea. Content validity testing validated the 32 items, which comprised medical (10 items), behavioral (13 items), and psychoemotional (9 items) management subscales. Construct validity testing of the ASMAT showed that the 32 items properly corresponded with conceptual constructs of arthritis self-management, and were suitable for assessing self-management ability in patients with arthritis. Reliability was also well supported. The ASMAT devised in the present study may aid the evaluation of patient self-management ability and the effectiveness of self-management interventions. The authors believe the developed tool may also aid the identification of problems associated with the adoption of self-management practice, and thus improve symptom management, independence, and quality of life of patients with arthritis.

  13. The assessment of emergency physicians by a regulatory authority.

    PubMed

    Lockyer, Jocelyn M; Violato, Claudio; Fidler, Herta

    2006-12-01

    To determine whether it is possible to develop a feasible, valid, and reliable multisource feedback program (360 degree evaluation) for emergency physicians. Surveys with 16, 20, 30, and 31 items were developed to assess emergency physicians by 25 patients, eight coworkers, eight medical colleagues, and self, respectively, using five-point scales along with an "unable to assess" category. Items addressed key competencies related to communication skills, professionalism, collegiality, and self-management. Data from 187 physicians who identified themselves as emergency physicians were available. The mean number of respondents per physician was 21.6 (SD +/- 3.87) (93%) for patients, 7.6 (SD +/- 0.89) (96%) for coworkers, and 7.7 (SD +/- 0.61) (95%) for medical colleagues, suggesting it was a feasible tool. Only the patient survey had four items with "unable to assess" percentages > or = 15%. The factor analysis indicated there were two factors on the patient questionnaire (communication/professionalism and patient education), two on the coworker survey (communication/collegiality and professionalism), and four on the medical colleague questionnaire (clinical performance, professionalism, self-management, and record management) that accounted for 80.0%, 62.5%, and 71.9% of the variance on the surveys, respectively. The factors were consistent with the intent of the instruments, providing empirical evidence of validity for the instruments. Reliability was established for the instruments (Cronbach's alpha > 0.94) and for each physician (generalizability coefficients were 0.68 for patients, 0.85 for coworkers, and 0.84 for medical colleagues). The psychometric examination of the data suggests that the instruments developed to assess emergency physicians were feasible and provide evidence for validity and reliability.

  14. Probabilistic analysis of tsunami hazards

    USGS Publications Warehouse

    Geist, E.L.; Parsons, T.

    2006-01-01

    Determining the likelihood of a disaster is a key component of any comprehensive hazard assessment. This is particularly true for tsunamis, even though most tsunami hazard assessments have in the past relied on scenario or deterministic type models. We discuss probabilistic tsunami hazard analysis (PTHA) from the standpoint of integrating computational methods with empirical analysis of past tsunami runup. PTHA is derived from probabilistic seismic hazard analysis (PSHA), with the main difference being that PTHA must account for far-field sources. The computational methods rely on numerical tsunami propagation models rather than empirical attenuation relationships as in PSHA in determining ground motions. Because a number of source parameters affect local tsunami runup height, PTHA can become complex and computationally intensive. Empirical analysis can function in one of two ways, depending on the length and completeness of the tsunami catalog. For site-specific studies where there is sufficient tsunami runup data available, hazard curves can primarily be derived from empirical analysis, with computational methods used to highlight deficiencies in the tsunami catalog. For region-wide analyses and sites where there are little to no tsunami data, a computationally based method such as Monte Carlo simulation is the primary method to establish tsunami hazards. Two case studies that describe how computational and empirical methods can be integrated are presented for Acapulco, Mexico (site-specific) and the U.S. Pacific Northwest coastline (region-wide analysis).

  15. The relationship of baby boomers’ participation motivation in leisure sports with recovery resilience and life satisfaction

    PubMed Central

    Koo, Jae-Eun; Lee, Gwang-Uk

    2013-01-01

    This study aimed to provide basic materials for resolving the problems of baby boomers, emerging as a social issue by identifying the effect of baby boomers’ participation motivation in leisure sports activities on recovery resilience and life satisfaction empirically. Using the convenience sampling method, the subjects were conducted by baby boomer’s 323 person lived in Seoul and Gyeong-in, 2012, excluding the missing question paper of 27 person. For accomplishing this purpose of the study, the survey questionnaires were used to collect data. Collected data was processed by factor analysis, reliability analysis, multiple regression, SPSS for Win V 18.0 program. From the analysis of this study, the following conclusion were obtained: First, among participation motivation factors of baby boomers in leisure sports activities, psychological stability and health pursuit had a significant effect on all factors of recovery resilience, while among motivation of personal relationships had a significant effect on the sub-factors of recovery resilience; empathy, optimism, and self-efficacy. Second, among participation motivation factors of baby boomers in leisure sports activities, psychological stability, personal relationships, and health pursuit had a significant effect on life satisfaction. PMID:24278870

  16. Enhancing the quality and credibility of qualitative analysis.

    PubMed

    Patton, M Q

    1999-12-01

    Varying philosophical and theoretical orientations to qualitative inquiry remind us that issues of quality and credibility intersect with audience and intended research purposes. This overview examines ways of enhancing the quality and credibility of qualitative analysis by dealing with three distinct but related inquiry concerns: rigorous techniques and methods for gathering and analyzing qualitative data, including attention to validity, reliability, and triangulation; the credibility, competence, and perceived trustworthiness of the qualitative researcher; and the philosophical beliefs of evaluation users about such paradigm-based preferences as objectivity versus subjectivity, truth versus perspective, and generalizations versus extrapolations. Although this overview examines some general approaches to issues of credibility and data quality in qualitative analysis, it is important to acknowledge that particular philosophical underpinnings, specific paradigms, and special purposes for qualitative inquiry will typically include additional or substitute criteria for assuring and judging quality, validity, and credibility. Moreover, the context for these considerations has evolved. In early literature on evaluation methods the debate between qualitative and quantitative methodologists was often strident. In recent years the debate has softened. A consensus has gradually emerged that the important challenge is to match appropriately the methods to empirical questions and issues, and not to universally advocate any single methodological approach for all problems.

  17. A half century of scalloping in the work habits of the United States Congress.

    PubMed Central

    Critchfield, Thomas S; Haley, Rebecca; Sabo, Benjamin; Colbert, Jorie; Macropoulis, Georgette

    2003-01-01

    It has been suggested that the work environment of the United States Congress bears similarity to a fixed-interval reinforcement schedule. Consistent with this notion, Weisberg and Waldrop (1972) described a positively accelerating pattern in annual congressional bill production (selected years from 1947 to 1968) that is reminiscent of the scalloped response pattern often attributed to fixed-interval schedules, but their analysis is now dated and does not bear on the functional relations that might yield scalloping. The present study described annual congressional bill production over a period of 52 years and empirically evaluated predictions derived from four hypotheses about the mechanisms that underlie scalloping. Scalloping occurred reliably in every year. The data supported several predictions about congressional productivity based on fixed-interval schedule performance, but did not consistently support any of three alternative accounts. These findings argue for the external validity of schedule-controlled operant behavior as measured in the laboratory. The present analysis also illustrates a largely overlooked role for applied behavior analysis: that of shedding light on the functional properties of behavior in uncontrolled settings of considerable interest to the public. PMID:14768667

  18. Sensitivity Analysis of Empirical Results on Civil War Onset

    ERIC Educational Resources Information Center

    Hegre, Havard; Sambanis, Nicholas

    2006-01-01

    In the literature on civil war onset, several empirical results are not robust or replicable across studies. Studies use different definitions of civil war and analyze different time periods, so readers cannot easily determine if differences in empirical results are due to those factors or if most empirical results are just not robust. The authors…

  19. How much detail is needed in modeling a transcranial magnetic stimulation figure-8 coil: Measurements and brain simulations

    PubMed Central

    Mandija, Stefano; Sommer, Iris E. C.; van den Berg, Cornelis A. T.; Neggers, Sebastiaan F. W.

    2017-01-01

    Background Despite TMS wide adoption, its spatial and temporal patterns of neuronal effects are not well understood. Although progress has been made in predicting induced currents in the brain using realistic finite element models (FEM), there is little consensus on how a magnetic field of a typical TMS coil should be modeled. Empirical validation of such models is limited and subject to several limitations. Methods We evaluate and empirically validate models of a figure-of-eight TMS coil that are commonly used in published modeling studies, of increasing complexity: simple circular coil model; coil with in-plane spiral winding turns; and finally one with stacked spiral winding turns. We will assess the electric fields induced by all 3 coil models in the motor cortex using a computer FEM model. Biot-Savart models of discretized wires were used to approximate the 3 coil models of increasing complexity. We use a tailored MR based phase mapping technique to get a full 3D validation of the incident magnetic field induced in a cylindrical phantom by our TMS coil. FEM based simulations on a meshed 3D brain model consisting of five tissues types were performed, using two orthogonal coil orientations. Results Substantial differences in the induced currents are observed, both theoretically and empirically, between highly idealized coils and coils with correctly modeled spiral winding turns. Thickness of the coil winding turns affect minimally the induced electric field, and it does not influence the predicted activation. Conclusion TMS coil models used in FEM simulations should include in-plane coil geometry in order to make reliable predictions of the incident field. Modeling the in-plane coil geometry is important to correctly simulate the induced electric field and to correctly make reliable predictions of neuronal activation PMID:28640923

  20. Comparing Usual Care With a Warfarin Initiation Protocol After Mechanical Heart Valve Replacement.

    PubMed

    Roberts, Gregory; Razooqi, Rasha; Quinn, Stephen

    2017-03-01

    The immediate postoperative warfarin sensitivity for patients receiving heart valve prostheses is increased. Established warfarin initiation protocols may lack clinical applicability, resulting in dosing based on clinical judgment. To compare current practice for warfarin initiation with a known warfarin initiation protocol, with doses proportionally reduced to account for the increased postoperative sensitivity. We compared the Mechanical Heart Valve Warfarin Initiation Protocol (Protocol group) with current practice (clinical judgment-Empirical group) for patients receiving mechanical heart valves in an observational before-and-after format. End points were the time to achieve a stable therapeutic international normalized ratio (INR), doses held in the first 6 days, and overanticoagulation in the first 6 days. The Protocol group (n = 37) achieved a stable INR more rapidly than the Empirical group (n = 77; median times 5.1 and 8.7 days, respectively; P = 0.002). Multivariable analysis indicated that the Protocol group (hazard ratio [HR] = 2.22; P = 0.005) and men (HR = 1.76; P = 0.043) more rapidly achieved a stable therapeutic INR. Age, serum albumin, amiodarone, presence of severe heart failure, and surgery type had no impact. Protocol patients had fewer doses held (1.1% vs 10.1%, P < 0.001) and no difference in overanticoagulation (2.7% vs 9.1%, P = 0.27). The Mechanical Heart Valve Warfarin Initiation Protocol provided a reliable approach to initiating warfarin in patients receiving mechanical aortic or mitral valves.

  1. Studying Regional Wave Source Time Functions Using the Empirical Green's Function Method: Application to Central Asia

    NASA Astrophysics Data System (ADS)

    Xie, J.; Schaff, D. P.; Chen, Y.; Schult, F.

    2013-12-01

    Reliably estimated source time functions (STFs) from high-frequency regional waveforms, such as Lg, Pn and Pg, provide important input for seismic source studies, explosion detection and discrimination, and minimization of parameter trade-off in attenuation studies. We have searched for candidate pairs of larger and small earthquakes in and around China that share the same focal mechanism but significantly differ in magnitudes, so that the empirical Green's function (EGF) method can be applied to study the STFs of the larger events. We conducted about a million deconvolutions using waveforms from 925 earthquakes, and screened the deconvolved traces to exclude those that are from event pairs that involved different mechanisms. Only 2,700 traces passed this screening and could be further analyzed using the EGF method. We have developed a series of codes for speeding up the final EGF analysis by implementing automations and user-graphic interface procedures. The codes have been fully tested with a subset of screened data and we are currently applying them to all the screened data. We will present a large number of deconvolved STFs retrieved using various phases (Lg, Pn, Sn and Pg and coda) with information on any directivities, any possible dependence of pulse durations on the wave types, on scaling relations for the pulse durations and event sizes, and on the estimated source static stress drops.

  2. Analysis of the Nonlinear Trends and Non-Stationary Oscillations of Regional Precipitation in Xinjiang, Northwestern China, Using Ensemble Empirical Mode Decomposition

    PubMed Central

    Guo, Bin; Chen, Zhongsheng; Guo, Jinyun; Liu, Feng; Chen, Chuanfa; Liu, Kangli

    2016-01-01

    Changes in precipitation could have crucial influences on the regional water resources in arid regions such as Xinjiang. It is necessary to understand the intrinsic multi-scale variations of precipitation in different parts of Xinjiang in the context of climate change. In this study, based on precipitation data from 53 meteorological stations in Xinjiang during 1960–2012, we investigated the intrinsic multi-scale characteristics of precipitation variability using an adaptive method named ensemble empirical mode decomposition (EEMD). Obvious non-linear upward trends in precipitation were found in the north, south, east and the entire Xinjiang. Changes in precipitation in Xinjiang exhibited significant inter-annual scale (quasi-2 and quasi-6 years) and inter-decadal scale (quasi-12 and quasi-23 years). Moreover, the 2–3-year quasi-periodic fluctuation was dominant in regional precipitation and the inter-annual variation had a considerable effect on the regional-scale precipitation variation in Xinjiang. We also found that there were distinctive spatial differences in variation trends and turning points of precipitation in Xinjiang. The results of this study indicated that compared to traditional decomposition methods, the EEMD method, without using any a priori determined basis functions, could effectively extract the reliable multi-scale fluctuations and reveal the intrinsic oscillation properties of climate elements. PMID:27007388

  3. General Purpose Electronegativity Relaxation Charge Models Applied to CoMFA and CoMSIA Study of GSK-3 Inhibitors.

    PubMed

    Tsareva, Daria A; Osolodkin, Dmitry I; Shulga, Dmitry A; Oliferenko, Alexander A; Pisarev, Sergey A; Palyulin, Vladimir A; Zefirov, Nikolay S

    2011-03-14

    Two fast empirical charge models, Kirchhoff Charge Model (KCM) and Dynamic Electronegativity Relaxation (DENR), had been developed in our laboratory previously for widespread use in drug design research. Both models are based on the electronegativity relaxation principle (Adv. Quantum Chem. 2006, 51, 139-156) and parameterized against ab initio dipole/quadrupole moments and molecular electrostatic potentials, respectively. As 3D QSAR studies comprise one of the most important fields of applied molecular modeling, they naturally have become the first topic to test our charges and thus, indirectly, the assumptions laid down to the charge model theories in a case study. Here these charge models are used in CoMFA and CoMSIA methods and tested on five glycogen synthase kinase 3 (GSK-3) inhibitor datasets, relevant to our current studies, and one steroid dataset. For comparison, eight other different charge models, ab initio through semiempirical and empirical, were tested on the same datasets. The complex analysis including correlation and cross-validation, charges robustness and predictability, as well as visual interpretability of 3D contour maps generated was carried out. As a result, our new electronegativity relaxation-based models both have shown stable results, which in conjunction with other benefits discussed render them suitable for building reliable 3D QSAR models. Copyright © 2011 WILEY-VCH Verlag GmbH & Co. KGaA, Weinheim.

  4. Reliability Practice at NASA Goddard Space Flight Center

    NASA Technical Reports Server (NTRS)

    Pruessner, Paula S.; Li, Ming

    2008-01-01

    This paper describes in brief the Reliability and Maintainability (R&M) Programs performed directly by the reliability branch at Goddard Space Flight Center (GSFC). The mission assurance requirements flow down is explained. GSFC practices for PRA, reliability prediction/fault tree analysis/reliability block diagram, FMEA, part stress and derating analysis, worst case analysis, trend analysis, limit life items are presented. Lessons learned are summarized and recommendations on improvement are identified.

  5. Development of ergonomics audits for bagging, haul truck and maintenance and repair operations in mining.

    PubMed

    Dempsey, Patrick G; Pollard, Jonisha; Porter, William L; Mayton, Alan; Heberger, John R; Gallagher, Sean; Reardon, Leanna; Drury, Colin G

    2017-12-01

    The development and testing of ergonomics and safety audits for small and bulk bag filling, haul truck and maintenance and repair operations in coal preparation and mineral processing plants found at surface mine sites is described. The content for the audits was derived from diverse sources of information on ergonomics and safety deficiencies including: analysis of injury, illness and fatality data and reports; task analysis; empirical laboratory studies of particular tasks; field studies and observations at mine sites; and maintenance records. These diverse sources of information were utilised to establish construct validity of the modular audits that were developed for use by mine safety personnel. User and interrater reliability testing was carried out prior to finalising the audits. The audits can be implemented using downloadable paper versions or with a free mobile NIOSH-developed Android application called ErgoMine. Practitioner Summary: The methodology used to develop ergonomics audits for three types of mining operations is described. Various sources of audit content are compared and contrasted to serve as a guide for developing ergonomics audits for other occupational contexts.

  6. Welding current and melting rate in GMAW of aluminium

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Pandey, S.; Rao, U.R.K.; Aghakhani, M.

    1996-12-31

    Studies on GMAW of aluminium and its alloy 5083, revealed that the welding current and melting rate were affected by any change in wire feed rate, arc voltage, nozzle to plate distance, welding speed and torch angle. Empirical models have been presented to determine accurately the welding current and melting rate for any set of these parameters. These results can be utilized for determining accurately the heat input into the workpiece from which reliable predictions can be made about the mechanical and the metallurgical properties of a welded joint. The analysis of the model also helps in providing a vitalmore » information about the static V-I characteristics of the welding power source. The models were developed using a two-level fractional factorial design. The adequacy of the model was tested by the use of analysis of variance technique and the significance of the coefficients was tested by the student`s t test. The estimated and observed values of the welding current and melting rate have been shown on a scatter diagram and the interaction effects of different parameters involved have been presented in graphical forms.« less

  7. The Hurst exponent in energy futures prices

    NASA Astrophysics Data System (ADS)

    Serletis, Apostolos; Rosenberg, Aryeh Adam

    2007-07-01

    This paper extends the work in Elder and Serletis [Long memory in energy futures prices, Rev. Financial Econ., forthcoming, 2007] and Serletis et al. [Detrended fluctuation analysis of the US stock market, Int. J. Bifurcation Chaos, forthcoming, 2007] by re-examining the empirical evidence for random walk type behavior in energy futures prices. In doing so, it uses daily data on energy futures traded on the New York Mercantile Exchange, over the period from July 2, 1990 to November 1, 2006, and a statistical physics approach-the ‘detrending moving average’ technique-providing a reliable framework for testing the information efficiency in financial markets as shown by Alessio et al. [Second-order moving average and scaling of stochastic time series, Eur. Phys. J. B 27 (2002) 197-200] and Carbone et al. [Time-dependent hurst exponent in financial time series. Physica A 344 (2004) 267-271; Analysis of clusters formed by the moving average of a long-range correlated time series. Phys. Rev. E 69 (2004) 026105]. The results show that energy futures returns display long memory and that the particular form of long memory is anti-persistence.

  8. Characteristics of German hospitals adopting health IT systems - results from an empirical study.

    PubMed

    Liebe, Jan-David; Egbert, Nicole; Frey, Andreas; Hübner, Ursula

    2011-01-01

    Hospital characteristics that facilitate IT adoption have been described by the literature extensively, however with controversial results. The aim of this study therefore is to draw a set of the most important variables from previous studies and include them in a combined analysis for testing their contribution as single factors and their interactions. Total number of IT systems installed and number of clinical IT systems in the hospital were used as criterion variables. Data from a national survey of German hospitals served as basis. Based on a stepwise multiple regression analysis four variables were identified to significantly explain the degree of IT adoption (60% explained variance): 1) hospital size, 2) IT department, 3) reference customer and 4) ownership (private vs. public). Our results replicate previous findings with regard to hospital size and ownership. In addition our study emphasizes the importance of a reliable internal structure for IT projects (existence of an IT department) and the culture of testing and installing most recent IT products (being a reference customer). None of the interactions between factors was significant.

  9. A tool to automatically analyze electromagnetic tracking data from high dose rate brachytherapy of breast cancer patients.

    PubMed

    Götz, Th I; Lahmer, G; Strnad, V; Bert, Ch; Hensel, B; Tomé, A M; Lang, E W

    2017-01-01

    During High Dose Rate Brachytherapy (HDR-BT) the spatial position of the radiation source inside catheters implanted into a female breast is determined via electromagnetic tracking (EMT). Dwell positions and dwell times of the radiation source are established, relative to the patient's anatomy, from an initial X-ray-CT-image. During the irradiation treatment, catheter displacements can occur due to patient movements. The current study develops an automatic analysis tool of EMT data sets recorded with a solenoid sensor to assure concordance of the source movement with the treatment plan. The tool combines machine learning techniques such as multi-dimensional scaling (MDS), ensemble empirical mode decomposition (EEMD), singular spectrum analysis (SSA) and particle filter (PF) to precisely detect and quantify any mismatch between the treatment plan and actual EMT measurements. We demonstrate that movement artifacts as well as technical signal distortions can be removed automatically and reliably, resulting in artifact-free reconstructed signals. This is a prerequisite for a highly accurate determination of any deviations of dwell positions from the treatment plan.

  10. HuH-7 reference genome profile: complex karyotype composed of massive loss of heterozygosity.

    PubMed

    Kasai, Fumio; Hirayama, Noriko; Ozawa, Midori; Satoh, Motonobu; Kohara, Arihiro

    2018-05-17

    Human cell lines represent a valuable resource as in vitro experimental models. A hepatoma cell line, HuH-7 (JCRB0403), has been used extensively in various research fields and a number of studies using this line have been published continuously since it was established in 1982. However, an accurate genome profile, which can be served as a reliable reference, has not been available. In this study, we performed M-FISH, SNP microarray and amplicon sequencing to characterize the cell line. Single cell analysis of metaphases revealed a high level of heterogeneity with a mode of 60 chromosomes. Cytogenetic results demonstrated chromosome abnormalities involving every chromosome in addition to a massive loss of heterozygosity, which accounts for 55.3% of the genome, consistent with the homozygous variants seen in the sequence analysis. We provide empirical data that the HuH-7 cell line is composed of highly heterogeneous cell populations, suggesting that besides cell line authentication, the quality of cell lines needs to be taken into consideration in the future use of tumor cell lines.

  11. An Empirical Taxonomy of Hospital Governing Board Roles

    PubMed Central

    Lee, Shoou-Yih D; Alexander, Jeffrey A; Wang, Virginia; Margolin, Frances S; Combes, John R

    2008-01-01

    Objective To develop a taxonomy of governing board roles in U.S. hospitals. Data Sources 2005 AHA Hospital Governance Survey, 2004 AHA Annual Survey of Hospitals, and Area Resource File. Study Design A governing board taxonomy was developed using cluster analysis. Results were validated and reviewed by industry experts. Differences in hospital and environmental characteristics across clusters were examined. Data Extraction Methods One-thousand three-hundred thirty-four hospitals with complete information on the study variables were included in the analysis. Principal Findings Five distinct clusters of hospital governing boards were identified. Statistical tests showed that the five clusters had high internal reliability and high internal validity. Statistically significant differences in hospital and environmental conditions were found among clusters. Conclusions The developed taxonomy provides policy makers, health care executives, and researchers a useful way to describe and understand hospital governing board roles. The taxonomy may also facilitate valid and systematic assessment of governance performance. Further, the taxonomy could be used as a framework for governing boards themselves to identify areas for improvement and direction for change. PMID:18355260

  12. A tool to automatically analyze electromagnetic tracking data from high dose rate brachytherapy of breast cancer patients

    PubMed Central

    Lahmer, G.; Strnad, V.; Bert, Ch.; Hensel, B.; Tomé, A. M.; Lang, E. W.

    2017-01-01

    During High Dose Rate Brachytherapy (HDR-BT) the spatial position of the radiation source inside catheters implanted into a female breast is determined via electromagnetic tracking (EMT). Dwell positions and dwell times of the radiation source are established, relative to the patient’s anatomy, from an initial X-ray-CT-image. During the irradiation treatment, catheter displacements can occur due to patient movements. The current study develops an automatic analysis tool of EMT data sets recorded with a solenoid sensor to assure concordance of the source movement with the treatment plan. The tool combines machine learning techniques such as multi-dimensional scaling (MDS), ensemble empirical mode decomposition (EEMD), singular spectrum analysis (SSA) and particle filter (PF) to precisely detect and quantify any mismatch between the treatment plan and actual EMT measurements. We demonstrate that movement artifacts as well as technical signal distortions can be removed automatically and reliably, resulting in artifact-free reconstructed signals. This is a prerequisite for a highly accurate determination of any deviations of dwell positions from the treatment plan. PMID:28934238

  13. Consumer involvement in seafood as family meals in Norway: an application of the expectancy-value approach.

    PubMed

    Olsen, S O

    2001-04-01

    A theoretical model of involvement in consumption of food products was tested in a representative survey of Norwegian households for the particular case of consuming seafood as a common family meal. The empirical study is based on using structural equation approach to test construct validity of measures and the empirical fit of the theoretical model. Attitudes, negative feelings, social norms and moral obligation were proved to be important, reliable and different constructs and explained 63% of the variation in seafood involvement. Negative feelings and moral obligation was the most important antecedents of involvement. Both our proposed model and modified model with seafood involvement as a mediator fit well with the data and proved our expectations in a promising way. Copyright 2001 Academic Press.

  14. Development of an empirical typology of African American family functioning.

    PubMed

    Mandara, Jelani; Murray, Carolyn B

    2002-09-01

    This study empirically identified types of African American families. Adolescents (N = 111) were assessed on family functioning. With cluster analytic methods, 3 types of families were identified. The cohesive-authoritative type was above average on parental education and income, averaged about 2 children, exhibited a high quality of family functioning and high self-esteem in adolescents. The conflictive-authoritarian type had average parental education and income, an average of 2.7 children, exhibited controlling and rigid discipline, and placed a high emphasis on achievement. The defensive-neglectful type was predominately headed by single mothers with below average education and income and averaged about 3 children. Such families displayed chaotic family processes, and adolescents tended to suffer from low self-esteem. The typology exhibited good reliability. The implications of the typology are discussed.

  15. The DSM diagnostic criteria for pedophilia.

    PubMed

    Blanchard, Ray

    2010-04-01

    This paper contains the author's report on pedophilia, submitted on June 2, 2008, to the work group charged with revising the diagnoses concerning sexual and gender identity disorders for the fifth edition of the American Psychiatric Association's Diagnostic and Statistical Manual of Mental Disorders (DSM). The author reviews the previously published criticisms and empirical research concerning the diagnostic criteria for pedophilia and presents criticism and relevant research of his own. The review shows that the DSM diagnostic criteria for pedophilia have repeatedly been criticized as unsatisfactory on logical or conceptual grounds, and that published empirical studies on the reliability and validity of these criteria have produced ambiguous results. It therefore seems that the current (i.e., DSM-IV-TR) diagnostic criteria need to be examined with an openness to major changes in the DSM-V.

  16. Changes in Yearly Birth Prevalence Rates of Children with Down Syndrome in the Period 1986-2007 in the Netherlands

    ERIC Educational Resources Information Center

    de Graaf, G.; Haveman, M.; Hochstenbach, R.; Engelen, J.; Gerssen-Schoorl, K.; Poddighe, P.; Smeets, D.; van Hove, G.

    2011-01-01

    Background: The Netherlands are lacking reliable national empirical data in relation to the development of birth prevalence of Down syndrome. Our study aims at assessing valid national live birth prevalence rates for the period 1986-2007. Method: On the basis of the annual child/adult ratio of Down syndrome diagnoses in five out of the eight Dutch…

  17. Influence of impact conditions on plasma generation during hypervelocity impact by aluminum projectile

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Song, Weidong, E-mail: swdgh@bit.edu.cn; Lv, Yangtao; Li, Jianqiao

    2016-07-15

    For describing hypervelocity impact (relative low-speed as related to space debris and much lower than travelling speed of meteoroids) phenomenon associated with plasma generation, a self-developed 3D code was advanced to numerically simulate projectiles impacting on a rigid wall. The numerical results were combined with a new ionization model which was developed in an early study to calculate the ionized materials during the impact. The calculated results of ionization were compared with the empirical formulas concluded by experiments in references and a good agreement was obtained. Then based on the reliable 3D numerical code, a series of impacts with differentmore » projectile configurations were simulated to investigate the influence of impact conditions on hypervelocity impact generated plasma. It was found that the form of empirical formula needed to be modified. A new empirical formula with a critical impact velocity was advanced to describe the velocity dependence of plasma generation and the parameters of the modified formula were ensured by the comparison between the numerical predictions and the empirical formulas. For different projectile configurations, the changes of plasma charges with time are different but the integrals of charges on time almost stayed in the same level.« less

  18. Modeling the erythemal surface diffuse irradiance fraction for Badajoz, Spain

    NASA Astrophysics Data System (ADS)

    Sanchez, Guadalupe; Serrano, Antonio; Cancillo, María Luisa

    2017-10-01

    Despite its important role on the human health and numerous biological processes, the diffuse component of the erythemal ultraviolet irradiance (UVER) is scarcely measured at standard radiometric stations and therefore needs to be estimated. This study proposes and compares 10 empirical models to estimate the UVER diffuse fraction. These models are inspired from mathematical expressions originally used to estimate total diffuse fraction, but, in this study, they are applied to the UVER case and tested against experimental measurements. In addition to adapting to the UVER range the various independent variables involved in these models, the total ozone column has been added in order to account for its strong impact on the attenuation of ultraviolet radiation. The proposed models are fitted to experimental measurements and validated against an independent subset. The best-performing model (RAU3) is based on a model proposed by Ruiz-Arias et al. (2010) and shows values of r2 equal to 0.91 and relative root-mean-square error (rRMSE) equal to 6.1 %. The performance achieved by this entirely empirical model is better than those obtained by previous semi-empirical approaches and therefore needs no additional information from other physically based models. This study expands on previous research to the ultraviolet range and provides reliable empirical models to accurately estimate the UVER diffuse fraction.

  19. Filtration of human EEG recordings from physiological artifacts with empirical mode method

    NASA Astrophysics Data System (ADS)

    Grubov, Vadim V.; Runnova, Anastasiya E.; Khramova, Marina V.

    2017-03-01

    In the paper we propose the new method for dealing with noise and physiological artifacts in experimental human EEG recordings. The method is based on analysis of EEG signals with empirical mode decomposition (Hilbert-Huang transform). We consider noises and physiological artifacts on EEG as specific oscillatory patterns that cause problems during EEG analysis and can be detected with additional signals recorded simultaneously with EEG (ECG, EMG, EOG, etc.) We introduce the algorithm of the method with following steps: empirical mode decomposition of EEG signal, choosing of empirical modes with artifacts, removing empirical modes with artifacts, reconstruction of the initial EEG signal. We test the method on filtration of experimental human EEG signals from eye-moving artifacts and show high efficiency of the method.

  20. The First Empirical Determination of the Fe10+ and Fe13+ Freeze-in Distances in the Solar Corona

    NASA Astrophysics Data System (ADS)

    Boe, Benjamin; Habbal, Shadia; Druckmüller, Miloslav; Landi, Enrico; Kourkchi, Ehsan; Ding, Adalbert; Starha, Pavel; Hutton, Joseph

    2018-06-01

    Heavy ions are markers of the physical processes responsible for the density and temperature distribution throughout the fine-scale magnetic structures that define the shape of the solar corona. One of their properties, whose empirical determination has remained elusive, is the “freeze-in” distance (R f ) where they reach fixed ionization states that are adhered to during their expansion with the solar wind. We present the first empirical inference of R f for {Fe}}{10+} and {Fe}}{13+} derived from multi-wavelength imaging observations of the corresponding Fe XI ({Fe}}{10+}) 789.2 nm and Fe XIV ({Fe}}{13+}) 530.3 nm emission acquired during the 2015 March 20 total solar eclipse. We find that the two ions freeze-in at different heliocentric distances. In polar coronal holes (CHs) R f is around 1.45 R ⊙ for {Fe}}{10+} and below 1.25 R ⊙ for {Fe}}{13+}. Along open field lines in streamer regions, R f ranges from 1.4 to 2 R ⊙ for {Fe}}{10+} and from 1.5 to 2.2 R ⊙ for {Fe}}{13+}. These first empirical R f values: (1) reflect the differing plasma parameters between CHs and streamers and structures within them, including prominences and coronal mass ejections; (2) are well below the currently quoted values derived from empirical model studies; and (3) place doubt on the reliability of plasma diagnostics based on the assumption of ionization equilibrium beyond 1.2 R ⊙.

  1. Identification of Child Pedestrian Training Objectives: The Role of Task Analysis and Empirical Research.

    ERIC Educational Resources Information Center

    van der Molen, Hugo H.

    1984-01-01

    Describes a study designed to demonstrate that child pedestrian training objectives may be identified systematically through various task analysis methods, making use of different types of empirical information. Early approaches to analysis of pedestrian tasks are reviewed, and an outline of the Traffic Research Centre's pedestrian task analysis…

  2. Parricide: An Empirical Analysis of 24 Years of U.S. Data

    ERIC Educational Resources Information Center

    Heide, Kathleen M.; Petee, Thomas A.

    2007-01-01

    Empirical analysis of homicides in which children have killed parents has been limited. The most comprehensive statistical analysis involving parents as victims was undertaken by Heide and used Supplementary Homicide Report (SHR) data for the 10-year period 1977 to 1986. This article provides an updated examination of characteristics of victims,…

  3. Development of Alabama traffic factors for use in mechanistic-empirical pavement design.

    DOT National Transportation Integrated Search

    2015-02-01

    The pavement engineering community is moving toward design practices that use mechanistic-empirical (M-E) approaches to the design and analysis of pavement structures. This effort is : embodied in the Mechanistic-Empirical Pavement Design Guide (MEPD...

  4. The role of empirical Bayes methodology as a leading principle in modern medical statistics.

    PubMed

    van Houwelingen, Hans C

    2014-11-01

    This paper reviews and discusses the role of Empirical Bayes methodology in medical statistics in the last 50 years. It gives some background on the origin of the empirical Bayes approach and its link with the famous Stein estimator. The paper describes the application in four important areas in medical statistics: disease mapping, health care monitoring, meta-analysis, and multiple testing. It ends with a warning that the application of the outcome of an empirical Bayes analysis to the individual "subjects" is a delicate matter that should be handled with prudence and care. © 2014 WILEY-VCH Verlag GmbH & Co. KGaA, Weinheim.

  5. Examination of Anomalous World Experience: A Report on Reliability.

    PubMed

    Conerty, Joseph; Skodlar, Borut; Pienkos, Elizabeth; Zadravek, Tina; Byrom, Greg; Sass, Louis

    2017-01-01

    The EAWE (Examination of Anomalous World Experience) is a newly developed, semi-structured interview that aims to capture anomalies of subjectivity, common in schizophrenia spectrum disorders, that pertain to experiences of the lived world, including space, time, people, language, atmosphere, and certain existential attitudes. By contrast, previous empirical studies of subjective experience in schizophrenia have focused largely on disturbances in self-experience. To assess the reliability of the EAWE, including internal consistency and interrater reliability. In the course of developing the EAWE, two distinct studies were conducted, one in the United States and the other in Slovenia. Thirteen patients diagnosed with schizophrenia spectrum or mood disorders were recruited for the US study. Fifteen such patients were recruited for the Slovenian study. Two live interviewers conducted the EAWE in the US. The Slovenian interviews were completed by one live interviewer with a second rater reviewing audiorecordings of the interview. Internal consistency and interrater reliability were calculated independently for each study, utilizing Cronbach's α, Spearman's ρ, and Cohen's κ. Each study yielded high internal consistency (Cronbach's α >0.82) and high interrater reliability for total EAWE scores (ρ > 0.83; average κ values were at least 0.78 for each study, with EAWE domain-specific κ not lower than 0.73). The EAWE, containing world-oriented inquiries into anomalies in subjective experience, has adequate reliability for use in a clinical or research setting. © 2017 S. Karger AG, Basel.

  6. Validation of the Spanish adaptation of the Consensus Auditory-Perceptual Evaluation of Voice (CAPE-V).

    PubMed

    Núñez-Batalla, Faustino; Morato-Galán, Marta; García-López, Isabel; Ávila-Menéndez, Arántzazu

    2015-01-01

    The Consensus Auditory-Perceptual Evaluation of Voice (CAPE-V) was developed.to promote a standardised approach to evaluating and documenting auditory perceptual judgments of vocal quality. This tool was originally developed in English language and its Spanish version is still inexistent. The aim of this study was to develop a Spanish adaptation of CAPE-V and to examine the reliability and empirical validity of this Spanish version. To adapt the CAPE-V protocol to the Spanish language, we proposed 6 phrases phonetically designed according to the CAPE-V requirements. Prospective instrument validation was performed. The validity of the Spanish version of the CAPE-V was examined in 4 ways: intra-rater reliability, inter-rater reliability and CAPE-V versus GRABS judgments. Inter-rater reliability coefficients for the CAPE-V ranged from 0.93 for overall severity to 0.54 for intensity; intra-rater reliability ranged from 0.98 for overall severity to 0.85 for intensity. The comparison of judgments between GRABS and CAPE-V ranged from 0.86 for overall severity to 0.61 for breathiness. The present study supports the use of the Spanish version of CAPE-V because of its validity and reliability. Copyright © 2014 Elsevier España, S.L.U. and Sociedad Española de Otorrinolaringología y Patología Cérvico-Facial. All rights reserved.

  7. Projection-Based Reduced Order Modeling for Spacecraft Thermal Analysis

    NASA Technical Reports Server (NTRS)

    Qian, Jing; Wang, Yi; Song, Hongjun; Pant, Kapil; Peabody, Hume; Ku, Jentung; Butler, Charles D.

    2015-01-01

    This paper presents a mathematically rigorous, subspace projection-based reduced order modeling (ROM) methodology and an integrated framework to automatically generate reduced order models for spacecraft thermal analysis. Two key steps in the reduced order modeling procedure are described: (1) the acquisition of a full-scale spacecraft model in the ordinary differential equation (ODE) and differential algebraic equation (DAE) form to resolve its dynamic thermal behavior; and (2) the ROM to markedly reduce the dimension of the full-scale model. Specifically, proper orthogonal decomposition (POD) in conjunction with discrete empirical interpolation method (DEIM) and trajectory piece-wise linear (TPWL) methods are developed to address the strong nonlinear thermal effects due to coupled conductive and radiative heat transfer in the spacecraft environment. Case studies using NASA-relevant satellite models are undertaken to verify the capability and to assess the computational performance of the ROM technique in terms of speed-up and error relative to the full-scale model. ROM exhibits excellent agreement in spatiotemporal thermal profiles (<0.5% relative error in pertinent time scales) along with salient computational acceleration (up to two orders of magnitude speed-up) over the full-scale analysis. These findings establish the feasibility of ROM to perform rational and computationally affordable thermal analysis, develop reliable thermal control strategies for spacecraft, and greatly reduce the development cycle times and costs.

  8. Automatic Detection of Previously-Unseen Application States for Deployment Environment Testing and Analysis

    PubMed Central

    Murphy, Christian; Vaughan, Moses; Ilahi, Waseem; Kaiser, Gail

    2010-01-01

    For large, complex software systems, it is typically impossible in terms of time and cost to reliably test the application in all possible execution states and configurations before releasing it into production. One proposed way of addressing this problem has been to continue testing and analysis of the application in the field, after it has been deployed. A practical limitation of many such automated approaches is the potentially high performance overhead incurred by the necessary instrumentation. However, it may be possible to reduce this overhead by selecting test cases and performing analysis only in previously-unseen application states, thus reducing the number of redundant tests and analyses that are run. Solutions for fault detection, model checking, security testing, and fault localization in deployed software may all benefit from a technique that ignores application states that have already been tested or explored. In this paper, we present a solution that ensures that deployment environment tests are only executed in states that the application has not previously encountered. In addition to discussing our implementation, we present the results of an empirical study that demonstrates its effectiveness, and explain how the new approach can be generalized to assist other automated testing and analysis techniques intended for the deployment environment. PMID:21197140

  9. Nonstationary Extreme Value Analysis in a Changing Climate: A Software Package

    NASA Astrophysics Data System (ADS)

    Cheng, L.; AghaKouchak, A.; Gilleland, E.

    2013-12-01

    Numerous studies show that climatic extremes have increased substantially in the second half of the 20th century. For this reason, analysis of extremes under a nonstationary assumption has received a great deal of attention. This paper presents a software package developed for estimation of return levels, return periods, and risks of climatic extremes in a changing climate. This MATLAB software package offers tools for analysis of climate extremes under both stationary and non-stationary assumptions. The Nonstationary Extreme Value Analysis (hereafter, NEVA) provides an efficient and generalized framework for analyzing extremes using Bayesian inference. NEVA estimates the extreme value parameters using a Differential Evolution Markov Chain (DE-MC) which utilizes the genetic algorithm Differential Evolution (DE) for global optimization over the real parameter space with the Markov Chain Monte Carlo (MCMC) approach and has the advantage of simplicity, speed of calculation and convergence over conventional MCMC. NEVA also offers the confidence interval and uncertainty bounds of estimated return levels based on the sampled parameters. NEVA integrates extreme value design concepts, data analysis tools, optimization and visualization, explicitly designed to facilitate analysis extremes in geosciences. The generalized input and output files of this software package make it attractive for users from across different fields. Both stationary and nonstationary components of the package are validated for a number of case studies using empirical return levels. The results show that NEVA reliably describes extremes and their return levels.

  10. A Comparison of Full and Empirical Bayes Techniques for Inferring Sea Level Changes from Tide Gauge Records

    NASA Astrophysics Data System (ADS)

    Piecuch, C. G.; Huybers, P. J.; Tingley, M.

    2016-12-01

    Sea level observations from coastal tide gauges are some of the longest instrumental records of the ocean. However, these data can be noisy, biased, and gappy, featuring missing values, and reflecting land motion and local effects. Coping with these issues in a formal manner is a challenging task. Some studies use Bayesian approaches to estimate sea level from tide gauge records, making inference probabilistically. Such methods are typically empirically Bayesian in nature: model parameters are treated as known and assigned point values. But, in reality, parameters are not perfectly known. Empirical Bayes methods thus neglect a potentially important source of uncertainty, and so may overestimate the precision (i.e., underestimate the uncertainty) of sea level estimates. We consider whether empirical Bayes methods underestimate uncertainty in sea level from tide gauge data, comparing to a full Bayes method that treats parameters as unknowns to be solved for along with the sea level field. We develop a hierarchical algorithm that we apply to tide gauge data on the North American northeast coast over 1893-2015. The algorithm is run in full Bayes mode, solving for the sea level process and parameters, and in empirical mode, solving only for the process using fixed parameter values. Error bars on sea level from the empirical method are smaller than from the full Bayes method, and the relative discrepancies increase with time; the 95% credible interval on sea level values from the empirical Bayes method in 1910 and 2010 is 23% and 56% narrower, respectively, than from the full Bayes approach. To evaluate the representativeness of the credible intervals, empirical Bayes and full Bayes methods are applied to corrupted data of a known surrogate field. Using rank histograms to evaluate the solutions, we find that the full Bayes method produces generally reliable error bars, whereas the empirical Bayes method gives too-narrow error bars, such that the 90% credible interval only encompasses 70% of true process values. Results demonstrate that parameter uncertainty is an important source of process uncertainty, and advocate for the fully Bayesian treatment of tide gauge records in ocean circulation and climate studies.

  11. CARES - CERAMICS ANALYSIS AND RELIABILITY EVALUATION OF STRUCTURES

    NASA Technical Reports Server (NTRS)

    Nemeth, N. N.

    1994-01-01

    The beneficial properties of structural ceramics include their high-temperature strength, light weight, hardness, and corrosion and oxidation resistance. For advanced heat engines, ceramics have demonstrated functional abilities at temperatures well beyond the operational limits of metals. This is offset by the fact that ceramic materials tend to be brittle. When a load is applied, their lack of significant plastic deformation causes the material to crack at microscopic flaws, destroying the component. CARES calculates the fast-fracture reliability or failure probability of macroscopically isotropic ceramic components. These components may be subjected to complex thermomechanical loadings. The program uses results from a commercial structural analysis program (MSC/NASTRAN or ANSYS) to evaluate component reliability due to inherent surface and/or volume type flaws. A multiple material capability allows the finite element model reliability to be a function of many different ceramic material statistical characterizations. The reliability analysis uses element stress, temperature, area, and volume output, which are obtained from two dimensional shell and three dimensional solid isoparametric or axisymmetric finite elements. CARES utilizes the Batdorf model and the two-parameter Weibull cumulative distribution function to describe the effects of multi-axial stress states on material strength. The shear-sensitive Batdorf model requires a user-selected flaw geometry and a mixed-mode fracture criterion. Flaws intersecting the surface and imperfections embedded in the volume can be modeled. The total strain energy release rate theory is used as a mixed mode fracture criterion for co-planar crack extension. Out-of-plane crack extension criteria are approximated by a simple equation with a semi-empirical constant that can model the maximum tangential stress theory, the minimum strain energy density criterion, the maximum strain energy release rate theory, or experimental results. For comparison, Griffith's maximum tensile stress theory, the principle of independent action, and the Weibull normal stress averaging models are also included. Weibull material strength parameters, the Batdorf crack density coefficient, and other related statistical quantities are estimated from four-point bend bar or uniform uniaxial tensile specimen fracture strength data. Parameter estimation can be performed for single or multiple failure modes by using the least-squares analysis or the maximum likelihood method. A more limited program, CARES/PC (COSMIC number LEW-15248) runs on a personal computer and estimates ceramic material properties from three-point bend bar data. CARES/PC does not perform fast fracture reliability estimation. CARES is written in FORTRAN 77 and has been implemented on DEC VAX series computers under VMS and on IBM 370 series computers under VM/CMS. On a VAX, CARES requires 10Mb of main memory. Five MSC/NASTRAN example problems and two ANSYS example problems are provided. There are two versions of CARES supplied on the distribution tape, CARES1 and CARES2. CARES2 contains sub-elements and CARES1 does not. CARES is available on a 9-track 1600 BPI VAX FILES-11 format magnetic tape (standard media) or in VAX BACKUP format on a TK50 tape cartridge. The program requires a FORTRAN 77 compiler and about 12Mb memory. CARES was developed in 1990. DEC, VAX and VMS are trademarks of Digital Equipment Corporation. IBM 370 is a trademark of International Business Machines. MSC/NASTRAN is a trademark of MacNeal-Schwendler Corporation. ANSYS is a trademark of Swanson Analysis Systems, Inc.

  12. Perceived Parental Attitudes of Gender Expansiveness: Development and Preliminary Factor Structure of a Self-Report Youth Questionnaire

    PubMed Central

    Hidalgo, Marco A.; Chen, Diane; Garofalo, Robert; Forbes, Catherine

    2017-01-01

    Abstract Purpose: Parental acceptance of gender identity/expression in lesbian, gay, bisexual, transgender, and queer/questioning (LGBTQ+) youth moderates the effects of minority stress on mental health outcomes. Given this association, mental health clinicians of gender-expansive adolescents often assess the degree to which these youth perceive their parents/primary caregivers as accepting or nonaffirming of their gender identity and expression. While existing measures may reliably assess youth's perceptions of general family support, no known tool aids in the assessment an adolescent's perceived parental support related to adolescent gender-expansive experiences. Methods: To provide both clinicians and researchers with an empirically derived tool, the current study used factor analysis to explore an underlying factor structure of a brief questionnaire developed by subject-matter experts and pertaining to multiple aspects of perceived parental support in gender-expansive adolescents and young adults. Respondents were gender-expansive adolescents and young adults seeking care in an interdisciplinary gender-health clinic within a pediatric academic medical center in the Midwestern United States. Results: Exploratory factor analysis resulted in a 14-item questionnaire comprised of two subscales assessing perceived parental nonaffirmation and perceived parental acceptance. Internal consistency and construct validity results provided support for this new questionnaire. Conclusion: This study provides preliminary evidence of the factor structure, reliability and validity of the Parental Attitudes of Gender Expansiveness Scale for Youth (PAGES-Y). These findings demonstrate both the clinical and research utility of the PAGES-Y, a tool that can yield a more nuanced understanding of family-related risk and protective factors in gender-expansive adolescents. PMID:29159312

  13. Perceived Parental Attitudes of Gender Expansiveness: Development and Preliminary Factor Structure of a Self-Report Youth Questionnaire.

    PubMed

    Hidalgo, Marco A; Chen, Diane; Garofalo, Robert; Forbes, Catherine

    2017-01-01

    Purpose: Parental acceptance of gender identity/expression in lesbian, gay, bisexual, transgender, and queer/questioning (LGBTQ+) youth moderates the effects of minority stress on mental health outcomes. Given this association, mental health clinicians of gender-expansive adolescents often assess the degree to which these youth perceive their parents/primary caregivers as accepting or nonaffirming of their gender identity and expression. While existing measures may reliably assess youth's perceptions of general family support, no known tool aids in the assessment an adolescent's perceived parental support related to adolescent gender-expansive experiences. Methods: To provide both clinicians and researchers with an empirically derived tool, the current study used factor analysis to explore an underlying factor structure of a brief questionnaire developed by subject-matter experts and pertaining to multiple aspects of perceived parental support in gender-expansive adolescents and young adults. Respondents were gender-expansive adolescents and young adults seeking care in an interdisciplinary gender-health clinic within a pediatric academic medical center in the Midwestern United States. Results: Exploratory factor analysis resulted in a 14-item questionnaire comprised of two subscales assessing perceived parental nonaffirmation and perceived parental acceptance. Internal consistency and construct validity results provided support for this new questionnaire. Conclusion: This study provides preliminary evidence of the factor structure, reliability and validity of the Parental Attitudes of Gender Expansiveness Scale for Youth (PAGES-Y). These findings demonstrate both the clinical and research utility of the PAGES-Y, a tool that can yield a more nuanced understanding of family-related risk and protective factors in gender-expansive adolescents.

  14. Chemical Explosion Experiments to Improve Nuclear Test Monitoring [Developing a New Paradigm for Nuclear Test Monitoring with the Source Physics Experiments (SPE)

    DOE PAGES

    Snelson, Catherine M.; Abbott, Robert E.; Broome, Scott T.; ...

    2013-07-02

    A series of chemical explosions, called the Source Physics Experiments (SPE), is being conducted under the auspices of the U.S. Department of Energy’s National Nuclear Security Administration (NNSA) to develop a new more physics-based paradigm for nuclear test monitoring. Currently, monitoring relies on semi-empirical models to discriminate explosions from earthquakes and to estimate key parameters such as yield. While these models have been highly successful monitoring established test sites, there is concern that future tests could occur in media and at scale depths of burial outside of our empirical experience. This is highlighted by North Korean tests, which exhibit poormore » performance of a reliable discriminant, mb:Ms (Selby et al., 2012), possibly due to source emplacement and differences in seismic responses for nascent and established test sites. The goal of SPE is to replace these semi-empirical relationships with numerical techniques grounded in a physical basis and thus applicable to any geologic setting or depth.« less

  15. Technology-Facilitated Sexual Violence: A Literature Review of Empirical Research.

    PubMed

    Henry, Nicola; Powell, Anastasia

    2018-04-01

    Technology-facilitated sexual violence (TFSV) refers to a range of behaviors where digital technologies are used to facilitate both virtual and face-to-face sexually based harms. Such behaviors include online sexual harassment, gender- and sexuality-based harassment, cyberstalking, image-based sexual exploitation, and the use of a carriage service to coerce a victim into an unwanted sexual act. This article reviews the current state of knowledge on these different dimensions, drawing on existing empirical studies. While there is a growing body of research into technology-facilitated harms perpetrated against children and adolescents, there is a dearth of qualitative and quantitative research on TFSV against adults. Moreover, few of the existing studies provide reliable data on the nature, scope, and impacts of TFSV. Preliminary studies, however, indicate that some harms, much like sexual violence more broadly, may be predominantly gender-, sexuality-, and age-based, with young women being overrepresented as victims in some categories. This review collects the empirical evidence to date regarding the prevalence and gender-based nature of TFSV against adults and discusses the implications for policy and programs, as well as suggestions for future research.

  16. Reliability and convergence of three concepts of narcissistic personality.

    PubMed

    Perry, J D; Perry, J C

    1996-01-01

    UNTIL recent years, the personality disorders have been relatively unexplored compared to other psychiatric diagnoses. Over 15 years ago, there was little agreement on the diagnosis of borderline personality disorder (Perry and Klerman 1978), but efforts to specify the constructs and respective criteria for the borderline diagnosis spurred a plethora of systematic research. The result is that, next to antisocial personality disorder, borderline has become one of the best-documented and validated personality disorders (Perry and Vaillant 1989). One important shift has been that good descriptive studies have gradually led to studies of etiological factors, such as childhood physical and sexual abuse, and severe neglect (Herman et al. 1989; Perry and Herman 1992), which in turn have led to empirically based treatment approaches (Herman 1992; Perry et al. 1990). Despite inclusion in The Diagnostic and Statistical Manual of Mental Disorders (DSM-III and DSM-III-R), narcissistic personality is still at the beginning of this process of description, empirical testing, and validation (Gunderson et al. 1991). This study empirically examines three descriptions of narcissistic personality in order to look for common underlying dimensions that may have etiological and treatment significance.

  17. Lifetime Reliability Prediction of Ceramic Structures Under Transient Thermomechanical Loads

    NASA Technical Reports Server (NTRS)

    Nemeth, Noel N.; Jadaan, Osama J.; Gyekenyesi, John P.

    2005-01-01

    An analytical methodology is developed to predict the probability of survival (reliability) of ceramic components subjected to harsh thermomechanical loads that can vary with time (transient reliability analysis). This capability enables more accurate prediction of ceramic component integrity against fracture in situations such as turbine startup and shutdown, operational vibrations, atmospheric reentry, or other rapid heating or cooling situations (thermal shock). The transient reliability analysis methodology developed herein incorporates the following features: fast-fracture transient analysis (reliability analysis without slow crack growth, SCG); transient analysis with SCG (reliability analysis with time-dependent damage due to SCG); a computationally efficient algorithm to compute the reliability for components subjected to repeated transient loading (block loading); cyclic fatigue modeling using a combined SCG and Walker fatigue law; proof testing for transient loads; and Weibull and fatigue parameters that are allowed to vary with temperature or time. Component-to-component variation in strength (stochastic strength response) is accounted for with the Weibull distribution, and either the principle of independent action or the Batdorf theory is used to predict the effect of multiaxial stresses on reliability. The reliability analysis can be performed either as a function of the component surface (for surface-distributed flaws) or component volume (for volume-distributed flaws). The transient reliability analysis capability has been added to the NASA CARES/ Life (Ceramic Analysis and Reliability Evaluation of Structures/Life) code. CARES/Life was also updated to interface with commercially available finite element analysis software, such as ANSYS, when used to model the effects of transient load histories. Examples are provided to demonstrate the features of the methodology as implemented in the CARES/Life program.

  18. An Empirical Bayes Approach to Mantel-Haenszel DIF Analysis.

    ERIC Educational Resources Information Center

    Zwick, Rebecca; Thayer, Dorothy T.; Lewis, Charles

    1999-01-01

    Developed an empirical Bayes enhancement to Mantel-Haenszel (MH) analysis of differential item functioning (DIF) in which it is assumed that the MH statistics are normally distributed and that the prior distribution of underlying DIF parameters is also normal. (Author/SLD)

  19. Social class, marginality and self-assessed health: a cross-sectional analysis of the health gradient in Mexico

    PubMed Central

    Valle, Adolfo Martinez

    2009-01-01

    Background Examining the association between social inequality and health is not new. However, there is little empirical evidence of this association in the Latin American literature, much less from the Mexican scholars. Its research, including the one conducted in Mexico, has mostly followed a theoretical approach and has not been able to provide strong empirical evidence of their important theoretical and conceptual contributions, mainly because reliable, complete and valid data are unavailable. Methods To empirically examine the gradient effect of social class on self-rated health in Mexico, a secondary cross-sectional mixed-level analysis was designed. Using individual level data from the Second National Health Survey (ENSA II), social class categories were specified following a stratification approach according to the occupation and education indicators available from ENSA II. Two types of categories were made, one for t urban and one for the rural labor force. Two indicators of perceived health status were used as health outcomes: self-assessed health and reported morbidity. Furthermore, the marginality index, an indicator of relative deprivation was used to examine its contextual effect at the state and regional level. The analysis was conducted using logistic multivariate models. Results The cross-sectional analysis showed a gradient effect of social class for good assessed-health. Relative to the low urban class, the odds ratio (OR) for a good perception of health for individuals belonging to the high urban class was 2.9 (95% confidence interval: 2.1–3.9). The OR for the middle high class was 2.8 (95% confidence interval: 2.4–3.4), while the OR for the middle low class was 1.8 (95% confidence interval: 1.6–2.1). However, for the rural labour force an OR of 1.5 was only significant between the high class who considered their health as good relative to the low class (95% confidence interval: 1.02–2.2). At the aggregate level, the results also showed individuals living in deprived regions were less likely to report their health as good than individuals living in relatively less deprived ones, OR = 0.6 (95% confidence interval: 0.4–0.7). Conclusion Overall, the findings of this study provided empirical evidence that social inequality negatively influences health through a differential exposure and an unequal distribution of resources across the class spectrum: the lower the social class, the poorer the perception of health. The results also showed that living in more deprived regions had a further negative effect on health. From a policy perspective, the gradient effects of social class suggest that non-targeted policies should be designed to address both material conditions at the individual level as well as deprived living conditions at higher levels of aggregation to improve health across the social spectrum. PMID:19236708

  20. The Limits of Outcomes Analysis; A Comment on "Sex Discrimination in Higher Employment: An Empirical Analysis of the Case Law."

    ERIC Educational Resources Information Center

    Lee, Barbara A.

    1990-01-01

    Questions assumptions by Schoenfeld and Zirkel in a study reviewing gender discrimination cases against institutions of higher education. Critiques the methodology used in that study, cautions about the overall utility of "outcomes analysis," and reports more promising routes of empirical legal research. (15 references) (MLF)

Top